Datasets:
Languages:
English
Multilinguality:
monolingual
Size Categories:
n<1K
Language Creators:
found
Source Datasets:
original
Tags:
karpathy,whisper,openai
WEBVTT | |
00:00.000 --> 00:03.040 | |
The following is a conversation with Vladimir Vapnik. | |
00:03.040 --> 00:05.280 | |
He's the coinventor of the Support Vector Machines, | |
00:05.280 --> 00:07.920 | |
Support Vector Clustering, VC Theory, | |
00:07.920 --> 00:11.200 | |
and many foundational ideas in statistical learning. | |
00:11.200 --> 00:13.640 | |
He was born in the Soviet Union and worked | |
00:13.640 --> 00:16.320 | |
at the Institute of Control Sciences in Moscow. | |
00:16.320 --> 00:20.640 | |
Then in the United States, he worked at AT&T, NEC Labs, | |
00:20.640 --> 00:24.280 | |
Facebook Research, and now as a professor at Columbia | |
00:24.280 --> 00:25.960 | |
University. | |
00:25.960 --> 00:30.320 | |
His work has been cited over 170,000 times. | |
00:30.320 --> 00:31.840 | |
He has some very interesting ideas | |
00:31.840 --> 00:34.800 | |
about artificial intelligence and the nature of learning, | |
00:34.800 --> 00:37.600 | |
especially on the limits of our current approaches | |
00:37.600 --> 00:40.440 | |
and the open problems in the field. | |
00:40.440 --> 00:42.520 | |
This conversation is part of MIT course | |
00:42.520 --> 00:44.440 | |
on artificial general intelligence | |
00:44.440 --> 00:46.840 | |
and the Artificial Intelligence Podcast. | |
00:46.840 --> 00:49.600 | |
If you enjoy it, please subscribe on YouTube | |
00:49.600 --> 00:53.040 | |
or rate it on iTunes or your podcast provider of choice | |
00:53.040 --> 00:55.320 | |
or simply connect with me on Twitter | |
00:55.320 --> 01:00.200 | |
or other social networks at Lex Friedman, spelled F R I D. | |
01:00.200 --> 01:04.800 | |
And now here's my conversation with Vladimir Vapnik. | |
01:04.800 --> 01:08.840 | |
Einstein famously said that God doesn't play dice. | |
01:08.840 --> 01:10.000 | |
Yeah. | |
01:10.000 --> 01:12.880 | |
You have studied the world through the eyes of statistics. | |
01:12.880 --> 01:17.320 | |
So let me ask you, in terms of the nature of reality, | |
01:17.320 --> 01:21.360 | |
fundamental nature of reality, does God play dice? | |
01:21.360 --> 01:26.200 | |
We don't know some factors, and because we | |
01:26.200 --> 01:30.520 | |
don't know some factors, which could be important, | |
01:30.520 --> 01:38.000 | |
it looks like God play dice, but we should describe it. | |
01:38.000 --> 01:42.080 | |
In philosophy, they distinguish between two positions, | |
01:42.080 --> 01:45.480 | |
positions of instrumentalism, where | |
01:45.480 --> 01:48.720 | |
you're creating theory for prediction | |
01:48.720 --> 01:51.400 | |
and position of realism, where you're | |
01:51.400 --> 01:54.640 | |
trying to understand what God's big. | |
01:54.640 --> 01:56.800 | |
Can you describe instrumentalism and realism | |
01:56.800 --> 01:58.400 | |
a little bit? | |
01:58.400 --> 02:06.320 | |
For example, if you have some mechanical laws, what is that? | |
02:06.320 --> 02:11.480 | |
Is it law which is true always and everywhere? | |
02:11.480 --> 02:14.880 | |
Or it is law which allows you to predict | |
02:14.880 --> 02:22.920 | |
the position of moving element, what you believe. | |
02:22.920 --> 02:28.480 | |
You believe that it is God's law, that God created the world, | |
02:28.480 --> 02:33.160 | |
which obeyed to this physical law, | |
02:33.160 --> 02:36.240 | |
or it is just law for predictions? | |
02:36.240 --> 02:38.400 | |
And which one is instrumentalism? | |
02:38.400 --> 02:39.880 | |
For predictions. | |
02:39.880 --> 02:45.400 | |
If you believe that this is law of God, and it's always | |
02:45.400 --> 02:50.040 | |
true everywhere, that means that you're a realist. | |
02:50.040 --> 02:55.480 | |
So you're trying to really understand that God's thought. | |
02:55.480 --> 03:00.040 | |
So the way you see the world as an instrumentalist? | |
03:00.040 --> 03:03.240 | |
You know, I'm working for some models, | |
03:03.240 --> 03:06.960 | |
model of machine learning. | |
03:06.960 --> 03:12.760 | |
So in this model, we can see setting, | |
03:12.760 --> 03:16.440 | |
and we try to solve, resolve the setting, | |
03:16.440 --> 03:18.240 | |
to solve the problem. | |
03:18.240 --> 03:20.760 | |
And you can do it in two different ways, | |
03:20.760 --> 03:23.840 | |
from the point of view of instrumentalists. | |
03:23.840 --> 03:27.120 | |
And that's what everybody does now, | |
03:27.120 --> 03:31.560 | |
because they say that the goal of machine learning | |
03:31.560 --> 03:36.800 | |
is to find the rule for classification. | |
03:36.800 --> 03:40.920 | |
That is true, but it is an instrument for prediction. | |
03:40.920 --> 03:46.160 | |
But I can say the goal of machine learning | |
03:46.160 --> 03:50.040 | |
is to learn about conditional probability. | |
03:50.040 --> 03:54.440 | |
So how God played use, and is He play? | |
03:54.440 --> 03:55.960 | |
What is probability for one? | |
03:55.960 --> 03:59.960 | |
What is probability for another given situation? | |
03:59.960 --> 04:02.600 | |
But for prediction, I don't need this. | |
04:02.600 --> 04:04.240 | |
I need the rule. | |
04:04.240 --> 04:08.480 | |
But for understanding, I need conditional probability. | |
04:08.480 --> 04:11.800 | |
So let me just step back a little bit first to talk about. | |
04:11.800 --> 04:13.960 | |
You mentioned, which I read last night, | |
04:13.960 --> 04:21.280 | |
the parts of the 1960 paper by Eugene Wigner, | |
04:21.280 --> 04:23.520 | |
unreasonable effectiveness of mathematics | |
04:23.520 --> 04:24.880 | |
and natural sciences. | |
04:24.880 --> 04:29.400 | |
Such a beautiful paper, by the way. | |
04:29.400 --> 04:34.480 | |
It made me feel, to be honest, to confess my own work | |
04:34.480 --> 04:38.400 | |
in the past few years on deep learning, heavily applied. | |
04:38.400 --> 04:40.320 | |
It made me feel that I was missing out | |
04:40.320 --> 04:43.960 | |
on some of the beauty of nature in the way | |
04:43.960 --> 04:45.560 | |
that math can uncover. | |
04:45.560 --> 04:50.360 | |
So let me just step away from the poetry of that for a second. | |
04:50.360 --> 04:53.040 | |
How do you see the role of math in your life? | |
04:53.040 --> 04:54.080 | |
Is it a tool? | |
04:54.080 --> 04:55.840 | |
Is it poetry? | |
04:55.840 --> 04:56.960 | |
Where does it sit? | |
04:56.960 --> 05:01.400 | |
And does math for you have limits of what it can describe? | |
05:01.400 --> 05:08.280 | |
Some people saying that math is language which use God. | |
05:08.280 --> 05:10.280 | |
So I believe in that. | |
05:10.280 --> 05:12.000 | |
Speak to God or use God. | |
05:12.000 --> 05:12.760 | |
Or use God. | |
05:12.760 --> 05:14.080 | |
Use God. | |
05:14.080 --> 05:15.560 | |
Yeah. | |
05:15.560 --> 05:25.680 | |
So I believe that this article about unreasonable | |
05:25.680 --> 05:29.960 | |
effectiveness of math is that if you're | |
05:29.960 --> 05:33.960 | |
looking in mathematical structures, | |
05:33.960 --> 05:37.720 | |
they know something about reality. | |
05:37.720 --> 05:42.480 | |
And the most scientists from natural science, | |
05:42.480 --> 05:48.440 | |
they're looking on equation and trying to understand reality. | |
05:48.440 --> 05:51.280 | |
So the same in machine learning. | |
05:51.280 --> 05:57.560 | |
If you're trying very carefully look on all equations | |
05:57.560 --> 06:00.640 | |
which define conditional probability, | |
06:00.640 --> 06:05.680 | |
you can understand something about reality more | |
06:05.680 --> 06:08.160 | |
than from your fantasy. | |
06:08.160 --> 06:12.480 | |
So math can reveal the simple underlying principles | |
06:12.480 --> 06:13.880 | |
of reality, perhaps. | |
06:13.880 --> 06:16.880 | |
You know, what means simple? | |
06:16.880 --> 06:20.320 | |
It is very hard to discover them. | |
06:20.320 --> 06:23.800 | |
But then when you discover them and look at them, | |
06:23.800 --> 06:27.440 | |
you see how beautiful they are. | |
06:27.440 --> 06:33.560 | |
And it is surprising why people did not see that before. | |
06:33.560 --> 06:37.480 | |
You're looking on equation and derive it from equations. | |
06:37.480 --> 06:43.360 | |
For example, I talked yesterday about least squirmated. | |
06:43.360 --> 06:48.120 | |
And people had a lot of fantasy have to improve least squirmated. | |
06:48.120 --> 06:52.360 | |
But if you're going step by step by solving some equations, | |
06:52.360 --> 06:57.680 | |
you suddenly will get some term which, | |
06:57.680 --> 07:01.040 | |
after thinking, you understand that it described | |
07:01.040 --> 07:04.360 | |
position of observation point. | |
07:04.360 --> 07:08.240 | |
In least squirmated, we throw out a lot of information. | |
07:08.240 --> 07:11.760 | |
We don't look in composition of point of observations. | |
07:11.760 --> 07:14.600 | |
We're looking only on residuals. | |
07:14.600 --> 07:19.400 | |
But when you understood that, that's a very simple idea. | |
07:19.400 --> 07:22.320 | |
But it's not too simple to understand. | |
07:22.320 --> 07:25.680 | |
And you can derive this just from equations. | |
07:25.680 --> 07:28.120 | |
So some simple algebra, a few steps | |
07:28.120 --> 07:31.040 | |
will take you to something surprising | |
07:31.040 --> 07:34.360 | |
that when you think about, you understand. | |
07:34.360 --> 07:41.120 | |
And that is proof that human intuition not to reach | |
07:41.120 --> 07:42.640 | |
and very primitive. | |
07:42.640 --> 07:48.520 | |
And it does not see very simple situations. | |
07:48.520 --> 07:51.760 | |
So let me take a step back in general. | |
07:51.760 --> 07:54.480 | |
Yes, right? | |
07:54.480 --> 07:58.840 | |
But what about human as opposed to intuition and ingenuity? | |
08:01.600 --> 08:02.960 | |
Moments of brilliance. | |
08:02.960 --> 08:09.480 | |
So do you have to be so hard on human intuition? | |
08:09.480 --> 08:11.840 | |
Are there moments of brilliance in human intuition? | |
08:11.840 --> 08:17.520 | |
They can leap ahead of math, and then the math will catch up? | |
08:17.520 --> 08:19.400 | |
I don't think so. | |
08:19.400 --> 08:23.560 | |
I think that the best human intuition, | |
08:23.560 --> 08:26.440 | |
it is putting in axioms. | |
08:26.440 --> 08:28.600 | |
And then it is technical. | |
08:28.600 --> 08:31.880 | |
See where the axioms take you. | |
08:31.880 --> 08:34.920 | |
But if they correctly take axioms, | |
08:34.920 --> 08:41.400 | |
but it axiom polished during generations of scientists. | |
08:41.400 --> 08:45.040 | |
And this is integral wisdom. | |
08:45.040 --> 08:47.480 | |
So that's beautifully put. | |
08:47.480 --> 08:54.040 | |
But if you maybe look at when you think of Einstein | |
08:54.040 --> 08:58.960 | |
and special relativity, what is the role of imagination | |
08:58.960 --> 09:04.480 | |
coming first there in the moment of discovery of an idea? | |
09:04.480 --> 09:06.440 | |
So there is obviously a mix of math | |
09:06.440 --> 09:10.800 | |
and out of the box imagination there. | |
09:10.800 --> 09:12.600 | |
That I don't know. | |
09:12.600 --> 09:18.080 | |
Whatever I did, I exclude any imagination. | |
09:18.080 --> 09:21.080 | |
Because whatever I saw in machine learning that | |
09:21.080 --> 09:26.440 | |
come from imagination, like features, like deep learning, | |
09:26.440 --> 09:29.320 | |
they are not relevant to the problem. | |
09:29.320 --> 09:31.960 | |
When you're looking very carefully | |
09:31.960 --> 09:34.280 | |
for mathematical equations, you're | |
09:34.280 --> 09:38.000 | |
deriving very simple theory, which goes far by | |
09:38.000 --> 09:42.040 | |
no theory at school than whatever people can imagine. | |
09:42.040 --> 09:44.760 | |
Because it is not good fantasy. | |
09:44.760 --> 09:46.720 | |
It is just interpretation. | |
09:46.720 --> 09:48.000 | |
It is just fantasy. | |
09:48.000 --> 09:51.320 | |
But it is not what you need. | |
09:51.320 --> 09:56.960 | |
You don't need any imagination to derive, say, | |
09:56.960 --> 10:00.040 | |
main principle of machine learning. | |
10:00.040 --> 10:02.760 | |
When you think about learning and intelligence, | |
10:02.760 --> 10:04.560 | |
maybe thinking about the human brain | |
10:04.560 --> 10:09.200 | |
and trying to describe mathematically the process of learning | |
10:09.200 --> 10:13.160 | |
that is something like what happens in the human brain, | |
10:13.160 --> 10:17.200 | |
do you think we have the tools currently? | |
10:17.200 --> 10:19.000 | |
Do you think we will ever have the tools | |
10:19.000 --> 10:22.680 | |
to try to describe that process of learning? | |
10:22.680 --> 10:25.800 | |
It is not description of what's going on. | |
10:25.800 --> 10:27.360 | |
It is interpretation. | |
10:27.360 --> 10:29.400 | |
It is your interpretation. | |
10:29.400 --> 10:32.080 | |
Your vision can be wrong. | |
10:32.080 --> 10:36.160 | |
You know, when a guy invent microscope, | |
10:36.160 --> 10:40.560 | |
Levin Cook for the first time, only he got this instrument | |
10:40.560 --> 10:45.440 | |
and nobody, he kept secrets about microscope. | |
10:45.440 --> 10:49.080 | |
But he wrote reports in London Academy of Science. | |
10:49.080 --> 10:52.040 | |
In his report, when he looked into the blood, | |
10:52.040 --> 10:54.480 | |
he looked everywhere, on the water, on the blood, | |
10:54.480 --> 10:56.320 | |
on the spin. | |
10:56.320 --> 11:04.040 | |
But he described blood like fight between queen and king. | |
11:04.040 --> 11:08.120 | |
So he saw blood cells, red cells, | |
11:08.120 --> 11:12.400 | |
and he imagines that it is army fighting each other. | |
11:12.400 --> 11:16.960 | |
And it was his interpretation of situation. | |
11:16.960 --> 11:19.760 | |
And he sent this report in Academy of Science. | |
11:19.760 --> 11:22.640 | |
They very carefully looked because they believed | |
11:22.640 --> 11:25.160 | |
that he is right, he saw something. | |
11:25.160 --> 11:28.240 | |
But he gave wrong interpretation. | |
11:28.240 --> 11:32.280 | |
And I believe the same can happen with brain. | |
11:32.280 --> 11:35.280 | |
Because the most important part, you know, | |
11:35.280 --> 11:38.840 | |
I believe in human language. | |
11:38.840 --> 11:43.000 | |
In some proverb, it's so much wisdom. | |
11:43.000 --> 11:50.240 | |
For example, people say that it is better than 1,000 days | |
11:50.240 --> 11:53.960 | |
of diligent studies one day with great teacher. | |
11:53.960 --> 11:59.480 | |
But if I will ask you what teacher does, nobody knows. | |
11:59.480 --> 12:01.400 | |
And that is intelligence. | |
12:01.400 --> 12:07.320 | |
And what we know from history, and now from mass | |
12:07.320 --> 12:12.080 | |
and machine learning, that teacher can do a lot. | |
12:12.080 --> 12:14.400 | |
So what, from a mathematical point of view, | |
12:14.400 --> 12:16.080 | |
is the great teacher? | |
12:16.080 --> 12:17.240 | |
I don't know. | |
12:17.240 --> 12:18.880 | |
That's an awful question. | |
12:18.880 --> 12:25.120 | |
Now, what we can say what teacher can do, | |
12:25.120 --> 12:29.440 | |
he can introduce some invariance, some predicate | |
12:29.440 --> 12:32.280 | |
for creating invariance. | |
12:32.280 --> 12:33.520 | |
How he doing it? | |
12:33.520 --> 12:34.080 | |
I don't know. | |
12:34.080 --> 12:37.560 | |
Because teacher knows reality and can describe | |
12:37.560 --> 12:41.200 | |
from this reality a predicate invariance. | |
12:41.200 --> 12:43.480 | |
But he knows that when you're using invariant, | |
12:43.480 --> 12:47.960 | |
he can decrease number of observations 100 times. | |
12:47.960 --> 12:52.960 | |
But maybe try to pull that apart a little bit. | |
12:52.960 --> 12:58.120 | |
I think you mentioned a piano teacher saying to the student, | |
12:58.120 --> 12:59.880 | |
play like a butterfly. | |
12:59.880 --> 13:03.720 | |
I played piano, I played guitar for a long time. | |
13:03.720 --> 13:09.800 | |
Yeah, maybe it's romantic, poetic. | |
13:09.800 --> 13:13.160 | |
But it feels like there's a lot of truth in that statement. | |
13:13.160 --> 13:15.440 | |
There is a lot of instruction in that statement. | |
13:15.440 --> 13:17.320 | |
And so can you pull that apart? | |
13:17.320 --> 13:19.760 | |
What is that? | |
13:19.760 --> 13:22.520 | |
The language itself may not contain this information. | |
13:22.520 --> 13:24.160 | |
It's not blah, blah, blah. | |
13:24.160 --> 13:25.640 | |
It does not blah, blah, blah, yeah. | |
13:25.640 --> 13:26.960 | |
It affects you. | |
13:26.960 --> 13:27.600 | |
It's what? | |
13:27.600 --> 13:28.600 | |
It affects you. | |
13:28.600 --> 13:29.800 | |
It affects your playing. | |
13:29.800 --> 13:30.640 | |
Yes, it does. | |
13:30.640 --> 13:33.640 | |
But it's not the language. | |
13:33.640 --> 13:38.000 | |
It feels like what is the information being exchanged there? | |
13:38.000 --> 13:39.760 | |
What is the nature of information? | |
13:39.760 --> 13:41.880 | |
What is the representation of that information? | |
13:41.880 --> 13:44.000 | |
I believe that it is sort of predicate. | |
13:44.000 --> 13:45.400 | |
But I don't know. | |
13:45.400 --> 13:48.880 | |
That is exactly what intelligence in machine learning | |
13:48.880 --> 13:50.080 | |
should be. | |
13:50.080 --> 13:53.200 | |
Because the rest is just mathematical technique. | |
13:53.200 --> 13:57.920 | |
I think that what was discovered recently | |
13:57.920 --> 14:03.280 | |
is that there is two mechanisms of learning. | |
14:03.280 --> 14:06.040 | |
One called strong convergence mechanism | |
14:06.040 --> 14:08.560 | |
and weak convergence mechanism. | |
14:08.560 --> 14:11.200 | |
Before, people use only one convergence. | |
14:11.200 --> 14:15.840 | |
In weak convergence mechanism, you can use predicate. | |
14:15.840 --> 14:19.360 | |
That's what play like butterfly. | |
14:19.360 --> 14:23.640 | |
And it will immediately affect your playing. | |
14:23.640 --> 14:26.360 | |
You know, there is English proverb. | |
14:26.360 --> 14:27.320 | |
Great. | |
14:27.320 --> 14:31.680 | |
If it looks like a duck, swims like a duck, | |
14:31.680 --> 14:35.200 | |
and quack like a duck, then it is probably duck. | |
14:35.200 --> 14:36.240 | |
Yes. | |
14:36.240 --> 14:40.400 | |
But this is exact about predicate. | |
14:40.400 --> 14:42.920 | |
Looks like a duck, what it means. | |
14:42.920 --> 14:46.720 | |
So you saw many ducks that you're training data. | |
14:46.720 --> 14:56.480 | |
So you have description of how looks integral looks ducks. | |
14:56.480 --> 14:59.360 | |
Yeah, the visual characteristics of a duck. | |
14:59.360 --> 15:00.840 | |
Yeah, but you won't. | |
15:00.840 --> 15:04.200 | |
And you have model for the cognition ducks. | |
15:04.200 --> 15:07.880 | |
So you would like that theoretical description | |
15:07.880 --> 15:12.720 | |
from model coincide with empirical description, which | |
15:12.720 --> 15:14.520 | |
you saw on Territax there. | |
15:14.520 --> 15:18.440 | |
So about looks like a duck, it is general. | |
15:18.440 --> 15:21.480 | |
But what about swims like a duck? | |
15:21.480 --> 15:23.560 | |
You should know that duck swims. | |
15:23.560 --> 15:26.960 | |
You can say it play chess like a duck, OK? | |
15:26.960 --> 15:28.880 | |
Duck doesn't play chess. | |
15:28.880 --> 15:35.560 | |
And it is completely legal predicate, but it is useless. | |
15:35.560 --> 15:41.040 | |
So half teacher can recognize not useless predicate. | |
15:41.040 --> 15:44.640 | |
So up to now, we don't use this predicate | |
15:44.640 --> 15:46.680 | |
in existing machine learning. | |
15:46.680 --> 15:47.200 | |
And you think that's not so useful? | |
15:47.200 --> 15:50.600 | |
So why we need billions of data? | |
15:50.600 --> 15:55.560 | |
But in this English proverb, they use only three predicate. | |
15:55.560 --> 15:59.080 | |
Looks like a duck, swims like a duck, and quack like a duck. | |
15:59.080 --> 16:02.040 | |
So you can't deny the fact that swims like a duck | |
16:02.040 --> 16:08.520 | |
and quacks like a duck has humor in it, has ambiguity. | |
16:08.520 --> 16:12.600 | |
Let's talk about swim like a duck. | |
16:12.600 --> 16:16.520 | |
It does not say jumps like a duck. | |
16:16.520 --> 16:17.680 | |
Why? | |
16:17.680 --> 16:20.760 | |
Because it's not relevant. | |
16:20.760 --> 16:25.880 | |
But that means that you know ducks, you know different birds, | |
16:25.880 --> 16:27.600 | |
you know animals. | |
16:27.600 --> 16:32.440 | |
And you derive from this that it is relevant to say swim like a duck. | |
16:32.440 --> 16:36.680 | |
So underneath, in order for us to understand swims like a duck, | |
16:36.680 --> 16:41.200 | |
it feels like we need to know millions of other little pieces | |
16:41.200 --> 16:43.000 | |
of information. | |
16:43.000 --> 16:44.280 | |
We pick up along the way. | |
16:44.280 --> 16:45.120 | |
You don't think so. | |
16:45.120 --> 16:48.480 | |
There doesn't need to be this knowledge base. | |
16:48.480 --> 16:52.600 | |
In those statements, carries some rich information | |
16:52.600 --> 16:57.280 | |
that helps us understand the essence of duck. | |
16:57.280 --> 17:01.920 | |
How far are we from integrating predicates? | |
17:01.920 --> 17:06.000 | |
You know that when you consider complete theory, | |
17:06.000 --> 17:09.320 | |
machine learning, so what it does, | |
17:09.320 --> 17:12.400 | |
you have a lot of functions. | |
17:12.400 --> 17:17.480 | |
And then you're talking, it looks like a duck. | |
17:17.480 --> 17:20.720 | |
You see your training data. | |
17:20.720 --> 17:31.040 | |
From training data, you recognize like expected duck should look. | |
17:31.040 --> 17:37.640 | |
Then you remove all functions, which does not look like you think | |
17:37.640 --> 17:40.080 | |
it should look from training data. | |
17:40.080 --> 17:45.800 | |
So you decrease amount of function from which you pick up one. | |
17:45.800 --> 17:48.320 | |
Then you give a second predicate. | |
17:48.320 --> 17:51.840 | |
And then, again, decrease the set of function. | |
17:51.840 --> 17:55.800 | |
And after that, you pick up the best function you can find. | |
17:55.800 --> 17:58.120 | |
It is standard machine learning. | |
17:58.120 --> 18:03.280 | |
So why you need not too many examples? | |
18:03.280 --> 18:06.600 | |
Because your predicates aren't very good, or you're not. | |
18:06.600 --> 18:09.200 | |
That means that predicate very good. | |
18:09.200 --> 18:12.520 | |
Because every predicate is invented | |
18:12.520 --> 18:17.720 | |
to decrease a divisible set of functions. | |
18:17.720 --> 18:20.320 | |
So you talk about admissible set of functions, | |
18:20.320 --> 18:22.440 | |
and you talk about good functions. | |
18:22.440 --> 18:24.280 | |
So what makes a good function? | |
18:24.280 --> 18:28.600 | |
So admissible set of function is set of function | |
18:28.600 --> 18:32.760 | |
which has small capacity, or small diversity, | |
18:32.760 --> 18:36.960 | |
small VC dimension example, which contain good function. | |
18:36.960 --> 18:38.760 | |
So by the way, for people who don't know, | |
18:38.760 --> 18:42.440 | |
VC, you're the V in the VC. | |
18:42.440 --> 18:50.440 | |
So how would you describe to a lay person what VC theory is? | |
18:50.440 --> 18:51.440 | |
How would you describe VC? | |
18:51.440 --> 18:56.480 | |
So when you have a machine, so a machine | |
18:56.480 --> 19:00.240 | |
capable to pick up one function from the admissible set | |
19:00.240 --> 19:02.520 | |
of function. | |
19:02.520 --> 19:07.640 | |
But set of admissibles function can be big. | |
19:07.640 --> 19:11.600 | |
They contain all continuous functions and it's useless. | |
19:11.600 --> 19:15.280 | |
You don't have so many examples to pick up function. | |
19:15.280 --> 19:17.280 | |
But it can be small. | |
19:17.280 --> 19:24.560 | |
Small, we call it capacity, but maybe better called diversity. | |
19:24.560 --> 19:27.160 | |
So not very different function in the set | |
19:27.160 --> 19:31.280 | |
is infinite set of function, but not very diverse. | |
19:31.280 --> 19:34.280 | |
So it is small VC dimension. | |
19:34.280 --> 19:39.360 | |
When VC dimension is small, you need small amount | |
19:39.360 --> 19:41.760 | |
of training data. | |
19:41.760 --> 19:47.360 | |
So the goal is to create admissible set of functions | |
19:47.360 --> 19:53.200 | |
which have small VC dimension and contain good function. | |
19:53.200 --> 19:58.160 | |
Then you will be able to pick up the function | |
19:58.160 --> 20:02.400 | |
using small amount of observations. | |
20:02.400 --> 20:06.760 | |
So that is the task of learning. | |
20:06.760 --> 20:11.360 | |
It is creating a set of admissible functions | |
20:11.360 --> 20:13.120 | |
that has a small VC dimension. | |
20:13.120 --> 20:17.320 | |
And then you've figured out a clever way of picking up. | |
20:17.320 --> 20:22.440 | |
No, that is goal of learning, which I formulated yesterday. | |
20:22.440 --> 20:25.760 | |
Statistical learning theory does not | |
20:25.760 --> 20:30.360 | |
involve in creating admissible set of function. | |
20:30.360 --> 20:35.520 | |
In classical learning theory, everywhere, 100% in textbook, | |
20:35.520 --> 20:39.200 | |
the set of function admissible set of function is given. | |
20:39.200 --> 20:41.760 | |
But this is science about nothing, | |
20:41.760 --> 20:44.040 | |
because the most difficult problem | |
20:44.040 --> 20:50.120 | |
to create admissible set of functions, given, say, | |
20:50.120 --> 20:53.080 | |
a lot of functions, continuum set of functions, | |
20:53.080 --> 20:54.960 | |
create admissible set of functions, | |
20:54.960 --> 20:58.760 | |
that means that it has finite VC dimension, | |
20:58.760 --> 21:02.280 | |
small VC dimension, and contain good function. | |
21:02.280 --> 21:05.280 | |
So this was out of consideration. | |
21:05.280 --> 21:07.240 | |
So what's the process of doing that? | |
21:07.240 --> 21:08.240 | |
I mean, it's fascinating. | |
21:08.240 --> 21:13.200 | |
What is the process of creating this admissible set of functions? | |
21:13.200 --> 21:14.920 | |
That is invariant. | |
21:14.920 --> 21:15.760 | |
That's invariance. | |
21:15.760 --> 21:17.280 | |
Can you describe invariance? | |
21:17.280 --> 21:22.440 | |
Yeah, you're looking of properties of training data. | |
21:22.440 --> 21:30.120 | |
And properties means that you have some function, | |
21:30.120 --> 21:36.520 | |
and you just count what is the average value of function | |
21:36.520 --> 21:38.960 | |
on training data. | |
21:38.960 --> 21:43.040 | |
You have a model, and what is the expectation | |
21:43.040 --> 21:44.960 | |
of this function on the model. | |
21:44.960 --> 21:46.720 | |
And they should coincide. | |
21:46.720 --> 21:51.800 | |
So the problem is about how to pick up functions. | |
21:51.800 --> 21:53.200 | |
It can be any function. | |
21:53.200 --> 21:59.280 | |
In fact, it is true for all functions. | |
21:59.280 --> 22:05.000 | |
But because when I talking set, say, | |
22:05.000 --> 22:09.920 | |
duck does not jumping, so you don't ask question, jump like a duck. | |
22:09.920 --> 22:13.360 | |
Because it is trivial, it does not jumping, | |
22:13.360 --> 22:15.560 | |
it doesn't help you to recognize jump. | |
22:15.560 --> 22:19.000 | |
But you know something, which question to ask, | |
22:19.000 --> 22:23.840 | |
when you're asking, it swims like a jump, like a duck. | |
22:23.840 --> 22:26.840 | |
But looks like a duck, it is general situation. | |
22:26.840 --> 22:34.440 | |
Looks like, say, guy who have this illness, this disease, | |
22:34.440 --> 22:42.280 | |
it is legal, so there is a general type of predicate | |
22:42.280 --> 22:46.440 | |
looks like, and special type of predicate, | |
22:46.440 --> 22:50.040 | |
which related to this specific problem. | |
22:50.040 --> 22:53.440 | |
And that is intelligence part of all this business. | |
22:53.440 --> 22:55.440 | |
And that we are teachers in world. | |
22:55.440 --> 22:58.440 | |
Incorporating those specialized predicates. | |
22:58.440 --> 23:04.840 | |
What do you think about deep learning as neural networks, | |
23:04.840 --> 23:11.440 | |
these arbitrary architectures as helping accomplish some of the tasks | |
23:11.440 --> 23:14.440 | |
you're thinking about, their effectiveness or lack thereof, | |
23:14.440 --> 23:19.440 | |
what are the weaknesses and what are the possible strengths? | |
23:19.440 --> 23:22.440 | |
You know, I think that this is fantasy. | |
23:22.440 --> 23:28.440 | |
Everything which like deep learning, like features. | |
23:28.440 --> 23:32.440 | |
Let me give you this example. | |
23:32.440 --> 23:38.440 | |
One of the greatest book, this Churchill book about history of Second World War. | |
23:38.440 --> 23:47.440 | |
And he's starting this book describing that in all time, when war is over, | |
23:47.440 --> 23:54.440 | |
so the great kings, they gathered together, | |
23:54.440 --> 23:57.440 | |
almost all of them were relatives, | |
23:57.440 --> 24:02.440 | |
and they discussed what should be done, how to create peace. | |
24:02.440 --> 24:04.440 | |
And they came to agreement. | |
24:04.440 --> 24:13.440 | |
And when happens First World War, the general public came in power. | |
24:13.440 --> 24:17.440 | |
And they were so greedy that robbed Germany. | |
24:17.440 --> 24:21.440 | |
And it was clear for everybody that it is not peace. | |
24:21.440 --> 24:28.440 | |
That peace will last only 20 years, because they were not professionals. | |
24:28.440 --> 24:31.440 | |
It's the same I see in machine learning. | |
24:31.440 --> 24:37.440 | |
There are mathematicians who are looking for the problem from a very deep point of view, | |
24:37.440 --> 24:39.440 | |
a mathematical point of view. | |
24:39.440 --> 24:45.440 | |
And there are computer scientists who mostly does not know mathematics. | |
24:45.440 --> 24:48.440 | |
They just have interpretation of that. | |
24:48.440 --> 24:53.440 | |
And they invented a lot of blah, blah, blah interpretations like deep learning. | |
24:53.440 --> 24:55.440 | |
Why you need deep learning? | |
24:55.440 --> 24:57.440 | |
Mathematics does not know deep learning. | |
24:57.440 --> 25:00.440 | |
Mathematics does not know neurons. | |
25:00.440 --> 25:02.440 | |
It is just function. | |
25:02.440 --> 25:06.440 | |
If you like to say piecewise linear function, say that, | |
25:06.440 --> 25:10.440 | |
and do it in class of piecewise linear function. | |
25:10.440 --> 25:12.440 | |
But they invent something. | |
25:12.440 --> 25:20.440 | |
And then they try to prove the advantage of that through interpretations, | |
25:20.440 --> 25:22.440 | |
which mostly wrong. | |
25:22.440 --> 25:25.440 | |
And when not enough they appeal to brain, | |
25:25.440 --> 25:27.440 | |
which they know nothing about that. | |
25:27.440 --> 25:29.440 | |
Nobody knows what's going on in the brain. | |
25:29.440 --> 25:34.440 | |
So I think that more reliable look on maths. | |
25:34.440 --> 25:36.440 | |
This is a mathematical problem. | |
25:36.440 --> 25:38.440 | |
Do your best to solve this problem. | |
25:38.440 --> 25:43.440 | |
Try to understand that there is not only one way of convergence, | |
25:43.440 --> 25:45.440 | |
which is strong way of convergence. | |
25:45.440 --> 25:49.440 | |
There is a weak way of convergence, which requires predicate. | |
25:49.440 --> 25:52.440 | |
And if you will go through all this stuff, | |
25:52.440 --> 25:55.440 | |
you will see that you don't need deep learning. | |
25:55.440 --> 26:00.440 | |
Even more, I would say one of the theorem, | |
26:00.440 --> 26:02.440 | |
which is called representor theorem. | |
26:02.440 --> 26:10.440 | |
It says that optimal solution of mathematical problem, | |
26:10.440 --> 26:20.440 | |
which described learning, is on shadow network, not on deep learning. | |
26:20.440 --> 26:22.440 | |
And a shallow network, yeah. | |
26:22.440 --> 26:24.440 | |
The ultimate problem is there. | |
26:24.440 --> 26:25.440 | |
Absolutely. | |
26:25.440 --> 26:29.440 | |
So in the end, what you're saying is exactly right. | |
26:29.440 --> 26:35.440 | |
The question is, you have no value for throwing something on the table, | |
26:35.440 --> 26:38.440 | |
playing with it, not math. | |
26:38.440 --> 26:41.440 | |
It's like in your old network where you said throwing something in the bucket | |
26:41.440 --> 26:45.440 | |
or the biological example and looking at kings and queens | |
26:45.440 --> 26:47.440 | |
or the cells or the microscope. | |
26:47.440 --> 26:52.440 | |
You don't see value in imagining the cells or kings and queens | |
26:52.440 --> 26:56.440 | |
and using that as inspiration and imagination | |
26:56.440 --> 26:59.440 | |
for where the math will eventually lead you. | |
26:59.440 --> 27:06.440 | |
You think that interpretation basically deceives you in a way that's not productive. | |
27:06.440 --> 27:14.440 | |
I think that if you're trying to analyze this business of learning | |
27:14.440 --> 27:18.440 | |
and especially discussion about deep learning, | |
27:18.440 --> 27:21.440 | |
it is discussion about interpretation. | |
27:21.440 --> 27:26.440 | |
It's discussion about things, about what you can say about things. | |
27:26.440 --> 27:29.440 | |
That's right, but aren't you surprised by the beauty of it? | |
27:29.440 --> 27:36.440 | |
Not mathematical beauty, but the fact that it works at all. | |
27:36.440 --> 27:39.440 | |
Or are you criticizing that very beauty, | |
27:39.440 --> 27:45.440 | |
our human desire to interpret, | |
27:45.440 --> 27:49.440 | |
to find our silly interpretations in these constructs? | |
27:49.440 --> 27:51.440 | |
Let me ask you this. | |
27:51.440 --> 27:55.440 | |
Are you surprised? | |
27:55.440 --> 27:57.440 | |
Does it inspire you? | |
27:57.440 --> 28:00.440 | |
How do you feel about the success of a system like AlphaGo | |
28:00.440 --> 28:03.440 | |
at beating the game of Go? | |
28:03.440 --> 28:09.440 | |
Using neural networks to estimate the quality of a board | |
28:09.440 --> 28:11.440 | |
and the quality of the board? | |
28:11.440 --> 28:14.440 | |
That is your interpretation quality of the board. | |
28:14.440 --> 28:17.440 | |
Yes. | |
28:17.440 --> 28:20.440 | |
It's not our interpretation. | |
28:20.440 --> 28:23.440 | |
The fact is, a neural network system doesn't matter. | |
28:23.440 --> 28:27.440 | |
A learning system that we don't mathematically understand | |
28:27.440 --> 28:29.440 | |
that beats the best human player. | |
28:29.440 --> 28:31.440 | |
It does something that was thought impossible. | |
28:31.440 --> 28:35.440 | |
That means that it's not very difficult problem. | |
28:35.440 --> 28:41.440 | |
We've empirically discovered that this is not a very difficult problem. | |
28:41.440 --> 28:43.440 | |
That's true. | |
28:43.440 --> 28:49.440 | |
Maybe I can't argue. | |
28:49.440 --> 28:52.440 | |
Even more, I would say, | |
28:52.440 --> 28:54.440 | |
that if they use deep learning, | |
28:54.440 --> 28:59.440 | |
it is not the most effective way of learning theory. | |
28:59.440 --> 29:03.440 | |
Usually, when people use deep learning, | |
29:03.440 --> 29:09.440 | |
they're using zillions of training data. | |
29:09.440 --> 29:13.440 | |
But you don't need this. | |
29:13.440 --> 29:15.440 | |
I describe the challenge. | |
29:15.440 --> 29:22.440 | |
Can we do some problems with deep learning method | |
29:22.440 --> 29:27.440 | |
with deep net using 100 times less training data? | |
29:27.440 --> 29:33.440 | |
Even more, some problems deep learning cannot solve | |
29:33.440 --> 29:37.440 | |
because it's not necessary. | |
29:37.440 --> 29:40.440 | |
They create admissible set of functions. | |
29:40.440 --> 29:45.440 | |
Deep architecture means to create admissible set of functions. | |
29:45.440 --> 29:49.440 | |
You cannot say that you're creating good admissible set of functions. | |
29:49.440 --> 29:52.440 | |
It's your fantasy. | |
29:52.440 --> 29:54.440 | |
It does not come from mass. | |
29:54.440 --> 29:58.440 | |
But it is possible to create admissible set of functions | |
29:58.440 --> 30:01.440 | |
because you have your training data. | |
30:01.440 --> 30:08.440 | |
Actually, for mathematicians, when you consider a variant, | |
30:08.440 --> 30:11.440 | |
you need to use law of large numbers. | |
30:11.440 --> 30:17.440 | |
When you're making training in existing algorithm, | |
30:17.440 --> 30:20.440 | |
you need uniform law of large numbers, | |
30:20.440 --> 30:22.440 | |
which is much more difficult. | |
30:22.440 --> 30:24.440 | |
You see dimension and all this stuff. | |
30:24.440 --> 30:32.440 | |
Nevertheless, if you use both weak and strong way of convergence, | |
30:32.440 --> 30:34.440 | |
you can decrease a lot of training data. | |
30:34.440 --> 30:39.440 | |
You could do the three, the Swims like a duck and Quacks like a duck. | |
30:39.440 --> 30:47.440 | |
Let's step back and think about human intelligence in general. | |
30:47.440 --> 30:52.440 | |
Clearly, that has evolved in a nonmathematical way. | |
30:52.440 --> 31:00.440 | |
As far as we know, God, or whoever, | |
31:00.440 --> 31:05.440 | |
didn't come up with a model in place in our brain of admissible functions. | |
31:05.440 --> 31:06.440 | |
It kind of evolved. | |
31:06.440 --> 31:07.440 | |
I don't know. | |
31:07.440 --> 31:08.440 | |
Maybe you have a view on this. | |
31:08.440 --> 31:15.440 | |
Alan Turing in the 50s in his paper asked and rejected the question, | |
31:15.440 --> 31:16.440 | |
can machines think? | |
31:16.440 --> 31:18.440 | |
It's not a very useful question. | |
31:18.440 --> 31:23.440 | |
But can you briefly entertain this useless question? | |
31:23.440 --> 31:25.440 | |
Can machines think? | |
31:25.440 --> 31:28.440 | |
So talk about intelligence and your view of it. | |
31:28.440 --> 31:29.440 | |
I don't know that. | |
31:29.440 --> 31:34.440 | |
I know that Turing described imitation. | |
31:34.440 --> 31:41.440 | |
If computer can imitate human being, let's call it intelligent. | |
31:41.440 --> 31:45.440 | |
And he understands that it is not thinking computer. | |
31:45.440 --> 31:46.440 | |
Yes. | |
31:46.440 --> 31:49.440 | |
He completely understands what he's doing. | |
31:49.440 --> 31:53.440 | |
But he's set up a problem of imitation. | |
31:53.440 --> 31:57.440 | |
So now we understand that the problem is not in imitation. | |
31:57.440 --> 32:04.440 | |
I'm not sure that intelligence is just inside of us. | |
32:04.440 --> 32:06.440 | |
It may be also outside of us. | |
32:06.440 --> 32:09.440 | |
I have several observations. | |
32:09.440 --> 32:15.440 | |
So when I prove some theorem, it's a very difficult theorem. | |
32:15.440 --> 32:22.440 | |
But in a couple of years, in several places, people proved the same theorem. | |
32:22.440 --> 32:26.440 | |
Say, soil lemma after us was done. | |
32:26.440 --> 32:29.440 | |
Then another guy proved the same theorem. | |
32:29.440 --> 32:32.440 | |
In the history of science, it's happened all the time. | |
32:32.440 --> 32:35.440 | |
For example, geometry. | |
32:35.440 --> 32:37.440 | |
It's happened simultaneously. | |
32:37.440 --> 32:43.440 | |
First it did Lobachevsky and then Gauss and Boyai and other guys. | |
32:43.440 --> 32:48.440 | |
It happened simultaneously in 10 years period of time. | |
32:48.440 --> 32:51.440 | |
And I saw a lot of examples like that. | |
32:51.440 --> 32:56.440 | |
And many mathematicians think that when they develop something, | |
32:56.440 --> 33:01.440 | |
they develop something in general which affects everybody. | |
33:01.440 --> 33:07.440 | |
So maybe our model that intelligence is only inside of us is incorrect. | |
33:07.440 --> 33:09.440 | |
It's our interpretation. | |
33:09.440 --> 33:15.440 | |
Maybe there exists some connection with world intelligence. | |
33:15.440 --> 33:16.440 | |
I don't know. | |
33:16.440 --> 33:19.440 | |
You're almost like plugging in into... | |
33:19.440 --> 33:20.440 | |
Yeah, exactly. | |
33:20.440 --> 33:22.440 | |
...and contributing to this... | |
33:22.440 --> 33:23.440 | |
Into a big network. | |
33:23.440 --> 33:26.440 | |
...into a big, maybe in your own network. | |
33:26.440 --> 33:27.440 | |
No, no, no. | |
33:27.440 --> 33:34.440 | |
On the flip side of that, maybe you can comment on big O complexity | |
33:34.440 --> 33:40.440 | |
and how you see classifying algorithms by worst case running time | |
33:40.440 --> 33:42.440 | |
in relation to their input. | |
33:42.440 --> 33:45.440 | |
So that way of thinking about functions. | |
33:45.440 --> 33:47.440 | |
Do you think P equals NP? | |
33:47.440 --> 33:49.440 | |
Do you think that's an interesting question? | |
33:49.440 --> 33:51.440 | |
Yeah, it is an interesting question. | |
33:51.440 --> 34:01.440 | |
But let me talk about complexity and about worst case scenario. | |
34:01.440 --> 34:03.440 | |
There is a mathematical setting. | |
34:03.440 --> 34:07.440 | |
When I came to the United States in 1990, | |
34:07.440 --> 34:09.440 | |
people did not know this theory. | |
34:09.440 --> 34:12.440 | |
They did not know statistical learning theory. | |
34:12.440 --> 34:17.440 | |
So in Russia it was published to monographs or monographs, | |
34:17.440 --> 34:19.440 | |
but in America they didn't know. | |
34:19.440 --> 34:22.440 | |
Then they learned. | |
34:22.440 --> 34:25.440 | |
And somebody told me that if it's worst case theory, | |
34:25.440 --> 34:27.440 | |
and they will create real case theory, | |
34:27.440 --> 34:30.440 | |
but till now it did not. | |
34:30.440 --> 34:33.440 | |
Because it is a mathematical tool. | |
34:33.440 --> 34:38.440 | |
You can do only what you can do using mathematics, | |
34:38.440 --> 34:45.440 | |
which has a clear understanding and clear description. | |
34:45.440 --> 34:52.440 | |
And for this reason we introduced complexity. | |
34:52.440 --> 34:54.440 | |
And we need this. | |
34:54.440 --> 35:01.440 | |
Because actually it is diverse, I like this one more. | |
35:01.440 --> 35:04.440 | |
This dimension you can prove some theorems. | |
35:04.440 --> 35:12.440 | |
But we also create theory for case when you know probability measure. | |
35:12.440 --> 35:14.440 | |
And that is the best case which can happen. | |
35:14.440 --> 35:17.440 | |
It is entropy theory. | |
35:17.440 --> 35:20.440 | |
So from a mathematical point of view, | |
35:20.440 --> 35:25.440 | |
you know the best possible case and the worst possible case. | |
35:25.440 --> 35:28.440 | |
You can derive different model in medium. | |
35:28.440 --> 35:30.440 | |
But it's not so interesting. | |
35:30.440 --> 35:33.440 | |
You think the edges are interesting? | |
35:33.440 --> 35:35.440 | |
The edges are interesting. | |
35:35.440 --> 35:44.440 | |
Because it is not so easy to get a good bound, exact bound. | |
35:44.440 --> 35:47.440 | |
It's not many cases where you have. | |
35:47.440 --> 35:49.440 | |
The bound is not exact. | |
35:49.440 --> 35:54.440 | |
But interesting principles which discover the mass. | |
35:54.440 --> 35:57.440 | |
Do you think it's interesting because it's challenging | |
35:57.440 --> 36:02.440 | |
and reveals interesting principles that allow you to get those bounds? | |
36:02.440 --> 36:05.440 | |
Or do you think it's interesting because it's actually very useful | |
36:05.440 --> 36:10.440 | |
for understanding the essence of a function of an algorithm? | |
36:10.440 --> 36:15.440 | |
So it's like me judging your life as a human being | |
36:15.440 --> 36:19.440 | |
by the worst thing you did and the best thing you did | |
36:19.440 --> 36:21.440 | |
versus all the stuff in the middle. | |
36:21.440 --> 36:25.440 | |
It seems not productive. | |
36:25.440 --> 36:31.440 | |
I don't think so because you cannot describe situation in the middle. | |
36:31.440 --> 36:34.440 | |
Or it will be not general. | |
36:34.440 --> 36:38.440 | |
So you can describe edges cases. | |
36:38.440 --> 36:41.440 | |
And it is clear it has some model. | |
36:41.440 --> 36:47.440 | |
But you cannot describe model for every new case. | |
36:47.440 --> 36:53.440 | |
So you will be never accurate when you're using model. | |
36:53.440 --> 36:55.440 | |
But from a statistical point of view, | |
36:55.440 --> 37:00.440 | |
the way you've studied functions and the nature of learning | |
37:00.440 --> 37:07.440 | |
and the world, don't you think that the real world has a very long tail | |
37:07.440 --> 37:13.440 | |
that the edge cases are very far away from the mean, | |
37:13.440 --> 37:19.440 | |
the stuff in the middle, or no? | |
37:19.440 --> 37:21.440 | |
I don't know that. | |
37:21.440 --> 37:29.440 | |
I think that from my point of view, | |
37:29.440 --> 37:39.440 | |
if you will use formal statistic, uniform law of large numbers, | |
37:39.440 --> 37:47.440 | |
if you will use this invariance business, | |
37:47.440 --> 37:51.440 | |
you will need just law of large numbers. | |
37:51.440 --> 37:55.440 | |
And there's a huge difference between uniform law of large numbers | |
37:55.440 --> 37:57.440 | |
and large numbers. | |
37:57.440 --> 37:59.440 | |
Can you describe that a little more? | |
37:59.440 --> 38:01.440 | |
Or should we just take it to... | |
38:01.440 --> 38:05.440 | |
No, for example, when I'm talking about duck, | |
38:05.440 --> 38:09.440 | |
I gave three predicates and it was enough. | |
38:09.440 --> 38:14.440 | |
But if you will try to do formal distinguish, | |
38:14.440 --> 38:17.440 | |
you will need a lot of observations. | |
38:17.440 --> 38:19.440 | |
I got you. | |
38:19.440 --> 38:24.440 | |
And so that means that information about looks like a duck | |
38:24.440 --> 38:27.440 | |
contain a lot of bits of information, | |
38:27.440 --> 38:29.440 | |
formal bits of information. | |
38:29.440 --> 38:35.440 | |
So we don't know that how much bit of information | |
38:35.440 --> 38:39.440 | |
contain things from artificial intelligence. | |
38:39.440 --> 38:42.440 | |
And that is the subject of analysis. | |
38:42.440 --> 38:47.440 | |
Till now, old business, | |
38:47.440 --> 38:54.440 | |
I don't like how people consider artificial intelligence. | |
38:54.440 --> 39:00.440 | |
They consider us some codes which imitate activity of human being. | |
39:00.440 --> 39:02.440 | |
It is not science. | |
39:02.440 --> 39:04.440 | |
It is applications. | |
39:04.440 --> 39:06.440 | |
You would like to imitate God. | |
39:06.440 --> 39:09.440 | |
It is very useful and we have good problem. | |
39:09.440 --> 39:15.440 | |
But you need to learn something more. | |
39:15.440 --> 39:23.440 | |
How people can to develop predicates, | |
39:23.440 --> 39:25.440 | |
swims like a duck, | |
39:25.440 --> 39:28.440 | |
or play like butterfly or something like that. | |
39:28.440 --> 39:33.440 | |
Not the teacher tells you how it came in his mind. | |
39:33.440 --> 39:36.440 | |
How he choose this image. | |
39:36.440 --> 39:39.440 | |
That is problem of intelligence. | |
39:39.440 --> 39:41.440 | |
That is the problem of intelligence. | |
39:41.440 --> 39:44.440 | |
And you see that connected to the problem of learning? | |
39:44.440 --> 39:45.440 | |
Absolutely. | |
39:45.440 --> 39:48.440 | |
Because you immediately give this predicate | |
39:48.440 --> 39:52.440 | |
like specific predicate, swims like a duck, | |
39:52.440 --> 39:54.440 | |
or quack like a duck. | |
39:54.440 --> 39:57.440 | |
It was chosen somehow. | |
39:57.440 --> 40:00.440 | |
So what is the line of work, would you say? | |
40:00.440 --> 40:05.440 | |
If you were to formulate as a set of open problems, | |
40:05.440 --> 40:07.440 | |
that will take us there. | |
40:07.440 --> 40:09.440 | |
Play like a butterfly. | |
40:09.440 --> 40:11.440 | |
We will get a system to be able to... | |
40:11.440 --> 40:13.440 | |
Let's separate two stories. | |
40:13.440 --> 40:15.440 | |
One mathematical story. | |
40:15.440 --> 40:19.440 | |
That if you have predicate, you can do something. | |
40:19.440 --> 40:22.440 | |
And another story you have to get predicate. | |
40:22.440 --> 40:26.440 | |
It is intelligence problem. | |
40:26.440 --> 40:31.440 | |
And people even did not start understanding intelligence. | |
40:31.440 --> 40:34.440 | |
Because to understand intelligence, first of all, | |
40:34.440 --> 40:37.440 | |
try to understand what doing teachers. | |
40:37.440 --> 40:40.440 | |
How teacher teach. | |
40:40.440 --> 40:43.440 | |
Why one teacher better than another one? | |
40:43.440 --> 40:44.440 | |
Yeah. | |
40:44.440 --> 40:48.440 | |
So you think we really even haven't started on the journey | |
40:48.440 --> 40:50.440 | |
of generating the predicate? | |
40:50.440 --> 40:51.440 | |
No. | |
40:51.440 --> 40:52.440 | |
We don't understand. | |
40:52.440 --> 40:56.440 | |
We even don't understand that this problem exists. | |
40:56.440 --> 40:58.440 | |
Because did you hear? | |
40:58.440 --> 40:59.440 | |
You do. | |
40:59.440 --> 41:02.440 | |
No, I just know name. | |
41:02.440 --> 41:07.440 | |
I want to understand why one teacher better than another. | |
41:07.440 --> 41:12.440 | |
And how affect teacher student. | |
41:12.440 --> 41:17.440 | |
It is not because he repeating the problem which is in textbook. | |
41:17.440 --> 41:18.440 | |
Yes. | |
41:18.440 --> 41:20.440 | |
He make some remarks. | |
41:20.440 --> 41:23.440 | |
He make some philosophy of reasoning. | |
41:23.440 --> 41:24.440 | |
Yeah, that's a beautiful... | |
41:24.440 --> 41:31.440 | |
So it is a formulation of a question that is the open problem. | |
41:31.440 --> 41:33.440 | |
Why is one teacher better than another? | |
41:33.440 --> 41:34.440 | |
Right. | |
41:34.440 --> 41:37.440 | |
What he does better. | |
41:37.440 --> 41:38.440 | |
Yeah. | |
41:38.440 --> 41:39.440 | |
What... | |
41:39.440 --> 41:42.440 | |
Why at every level? | |
41:42.440 --> 41:44.440 | |
How do they get better? | |
41:44.440 --> 41:47.440 | |
What does it mean to be better? | |
41:47.440 --> 41:49.440 | |
The whole... | |
41:49.440 --> 41:50.440 | |
Yeah. | |
41:50.440 --> 41:53.440 | |
From whatever model I have. | |
41:53.440 --> 41:56.440 | |
One teacher can give a very good predicate. | |
41:56.440 --> 42:00.440 | |
One teacher can say swims like a dog. | |
42:00.440 --> 42:03.440 | |
And another can say jump like a dog. | |
42:03.440 --> 42:05.440 | |
And jump like a dog. | |
42:05.440 --> 42:07.440 | |
Car is zero information. | |
42:07.440 --> 42:08.440 | |
Yeah. | |
42:08.440 --> 42:13.440 | |
So what is the most exciting problem in statistical learning you've ever worked on? | |
42:13.440 --> 42:16.440 | |
Or are working on now? | |
42:16.440 --> 42:22.440 | |
I just finished this invariant story. | |
42:22.440 --> 42:24.440 | |
And I'm happy that... | |
42:24.440 --> 42:30.440 | |
I believe that it is ultimate learning story. | |
42:30.440 --> 42:37.440 | |
At least I can show that there are no another mechanism, only two mechanisms. | |
42:37.440 --> 42:44.440 | |
But they separate statistical part from intelligent part. | |
42:44.440 --> 42:48.440 | |
And I know nothing about intelligent part. | |
42:48.440 --> 42:52.440 | |
And if we will know this intelligent part, | |
42:52.440 --> 42:59.440 | |
so it will help us a lot in teaching, in learning. | |
42:59.440 --> 43:02.440 | |
You don't know it when we see it? | |
43:02.440 --> 43:06.440 | |
So for example, in my talk, the last slide was the challenge. | |
43:06.440 --> 43:11.440 | |
So you have, say, NIST digital recognition problem. | |
43:11.440 --> 43:16.440 | |
And deep learning claims that they did it very well. | |
43:16.440 --> 43:21.440 | |
Say 99.5% of correct answers. | |
43:21.440 --> 43:24.440 | |
But they use 60,000 observations. | |
43:24.440 --> 43:26.440 | |
Can you do the same? | |
43:26.440 --> 43:29.440 | |
100 times less. | |
43:29.440 --> 43:31.440 | |
But incorporating invariants. | |
43:31.440 --> 43:34.440 | |
What it means, you know, digit one, two, three. | |
43:34.440 --> 43:35.440 | |
Yeah. | |
43:35.440 --> 43:37.440 | |
Just looking at that. | |
43:37.440 --> 43:40.440 | |
Explain to me which invariant I should keep. | |
43:40.440 --> 43:43.440 | |
To use 100 examples. | |
43:43.440 --> 43:48.440 | |
Or say 100 times less examples to do the same job. | |
43:48.440 --> 43:49.440 | |
Yeah. | |
43:49.440 --> 43:55.440 | |
That last slide, unfortunately, you talk ended quickly. | |
43:55.440 --> 43:59.440 | |
The last slide was a powerful open challenge | |
43:59.440 --> 44:02.440 | |
and a formulation of the essence here. | |
44:02.440 --> 44:06.440 | |
That is the exact problem of intelligence. | |
44:06.440 --> 44:12.440 | |
Because everybody, when machine learning started, | |
44:12.440 --> 44:15.440 | |
it was developed by mathematicians, | |
44:15.440 --> 44:19.440 | |
they immediately recognized that we use much more | |
44:19.440 --> 44:22.440 | |
training data than humans needed. | |
44:22.440 --> 44:25.440 | |
But now again, we came to the same story. | |
44:25.440 --> 44:27.440 | |
Have to decrease. | |
44:27.440 --> 44:30.440 | |
That is the problem of learning. | |
44:30.440 --> 44:32.440 | |
It is not like in deep learning, | |
44:32.440 --> 44:35.440 | |
they use zealons of training data. | |
44:35.440 --> 44:38.440 | |
Because maybe zealons are not enough | |
44:38.440 --> 44:44.440 | |
if you have a good invariance. | |
44:44.440 --> 44:49.440 | |
Maybe you'll never collect some number of observations. | |
44:49.440 --> 44:53.440 | |
But now it is a question to intelligence. | |
44:53.440 --> 44:55.440 | |
Have to do that. | |
44:55.440 --> 44:58.440 | |
Because statistical part is ready. | |
44:58.440 --> 45:02.440 | |
As soon as you supply us with predicate, | |
45:02.440 --> 45:06.440 | |
we can do good job with small amount of observations. | |
45:06.440 --> 45:10.440 | |
And the very first challenges will know digit recognition. | |
45:10.440 --> 45:12.440 | |
And you know digits. | |
45:12.440 --> 45:15.440 | |
And please tell me invariance. | |
45:15.440 --> 45:16.440 | |
I think about that. | |
45:16.440 --> 45:20.440 | |
I can say for digit 3, I would introduce | |
45:20.440 --> 45:24.440 | |
concept of horizontal symmetry. | |
45:24.440 --> 45:29.440 | |
So the digit 3 has horizontal symmetry | |
45:29.440 --> 45:33.440 | |
more than say digit 2 or something like that. | |
45:33.440 --> 45:37.440 | |
But as soon as I get the idea of horizontal symmetry, | |
45:37.440 --> 45:40.440 | |
I can mathematically invent a lot of | |
45:40.440 --> 45:43.440 | |
measure of horizontal symmetry | |
45:43.440 --> 45:46.440 | |
on vertical symmetry or diagonal symmetry, | |
45:46.440 --> 45:49.440 | |
whatever, if I have a day of symmetry. | |
45:49.440 --> 45:52.440 | |
But what else? | |
45:52.440 --> 46:00.440 | |
Looking on digit, I see that it is metapredicate, | |
46:00.440 --> 46:04.440 | |
which is not shape. | |
46:04.440 --> 46:07.440 | |
It is something like symmetry, | |
46:07.440 --> 46:12.440 | |
like how dark is whole picture, something like that. | |
46:12.440 --> 46:15.440 | |
Which can self rise up predicate. | |
46:15.440 --> 46:18.440 | |
You think such a predicate could rise | |
46:18.440 --> 46:26.440 | |
out of something that is not general. | |
46:26.440 --> 46:31.440 | |
Meaning it feels like for me to be able to | |
46:31.440 --> 46:34.440 | |
understand the difference between a 2 and a 3, | |
46:34.440 --> 46:39.440 | |
I would need to have had a childhood | |
46:39.440 --> 46:45.440 | |
of 10 to 15 years playing with kids, | |
46:45.440 --> 46:50.440 | |
going to school, being yelled by parents. | |
46:50.440 --> 46:55.440 | |
All of that, walking, jumping, looking at ducks. | |
46:55.440 --> 46:58.440 | |
And now then I would be able to generate | |
46:58.440 --> 47:01.440 | |
the right predicate for telling the difference | |
47:01.440 --> 47:03.440 | |
between 2 and a 3. | |
47:03.440 --> 47:06.440 | |
Or do you think there is a more efficient way? | |
47:06.440 --> 47:10.440 | |
I know for sure that you must know | |
47:10.440 --> 47:12.440 | |
something more than digits. | |
47:12.440 --> 47:15.440 | |
That's a powerful statement. | |
47:15.440 --> 47:19.440 | |
But maybe there are several languages | |
47:19.440 --> 47:24.440 | |
of description, these elements of digits. | |
47:24.440 --> 47:27.440 | |
So I'm talking about symmetry, | |
47:27.440 --> 47:30.440 | |
about some properties of geometry, | |
47:30.440 --> 47:33.440 | |
I'm talking about something abstract. | |
47:33.440 --> 47:38.440 | |
But this is a problem of intelligence. | |
47:38.440 --> 47:42.440 | |
So in one of our articles, it is trivial to show | |
47:42.440 --> 47:46.440 | |
that every example can carry | |
47:46.440 --> 47:49.440 | |
not more than one bit of information in real. | |
47:49.440 --> 47:54.440 | |
Because when you show example | |
47:54.440 --> 47:59.440 | |
and you say this is one, you can remove, say, | |
47:59.440 --> 48:03.440 | |
a function which does not tell you one, say, | |
48:03.440 --> 48:06.440 | |
the best strategy, if you can do it perfectly, | |
48:06.440 --> 48:09.440 | |
it's remove half of the functions. | |
48:09.440 --> 48:14.440 | |
But when you use one predicate, which looks like a duck, | |
48:14.440 --> 48:18.440 | |
you can remove much more functions than half. | |
48:18.440 --> 48:20.440 | |
And that means that it contains | |
48:20.440 --> 48:25.440 | |
a lot of bit of information from a formal point of view. | |
48:25.440 --> 48:31.440 | |
But when you have a general picture | |
48:31.440 --> 48:33.440 | |
of what you want to recognize, | |
48:33.440 --> 48:36.440 | |
a general picture of the world, | |
48:36.440 --> 48:40.440 | |
can you invent this predicate? | |
48:40.440 --> 48:46.440 | |
And that predicate carries a lot of information. | |
48:46.440 --> 48:49.440 | |
Beautifully put, maybe just me, | |
48:49.440 --> 48:53.440 | |
but in all the math you show, in your work, | |
48:53.440 --> 48:57.440 | |
which is some of the most profound mathematical work | |
48:57.440 --> 49:01.440 | |
in the field of learning AI and just math in general. | |
49:01.440 --> 49:04.440 | |
I hear a lot of poetry and philosophy. | |
49:04.440 --> 49:09.440 | |
You really kind of talk about philosophy of science. | |
49:09.440 --> 49:12.440 | |
There's a poetry and music to a lot of the work you're doing | |
49:12.440 --> 49:14.440 | |
and the way you're thinking about it. | |
49:14.440 --> 49:16.440 | |
So where does that come from? | |
49:16.440 --> 49:20.440 | |
Do you escape to poetry? Do you escape to music? | |
49:20.440 --> 49:24.440 | |
I think that there exists ground truth. | |
49:24.440 --> 49:26.440 | |
There exists ground truth? | |
49:26.440 --> 49:30.440 | |
Yeah, and that can be seen everywhere. | |
49:30.440 --> 49:32.440 | |
The smart guy, philosopher, | |
49:32.440 --> 49:38.440 | |
sometimes I surprise how they deep see. | |
49:38.440 --> 49:45.440 | |
Sometimes I see that some of them are completely out of subject. | |
49:45.440 --> 49:50.440 | |
But the ground truth I see in music. | |
49:50.440 --> 49:52.440 | |
Music is the ground truth? | |
49:52.440 --> 49:53.440 | |
Yeah. | |
49:53.440 --> 50:01.440 | |
And in poetry, many poets, they believe they take dictation. | |
50:01.440 --> 50:06.440 | |
So what piece of music, | |
50:06.440 --> 50:08.440 | |
as a piece of empirical evidence, | |
50:08.440 --> 50:14.440 | |
gave you a sense that they are touching something in the ground truth? | |
50:14.440 --> 50:16.440 | |
It is structure. | |
50:16.440 --> 50:18.440 | |
The structure with the math of music. | |
50:18.440 --> 50:20.440 | |
Because when you're listening to Bach, | |
50:20.440 --> 50:22.440 | |
you see this structure. | |
50:22.440 --> 50:25.440 | |
Very clear, very classic, very simple. | |
50:25.440 --> 50:31.440 | |
And the same in Bach, when you have axioms in geometry, | |
50:31.440 --> 50:33.440 | |
you have the same feeling. | |
50:33.440 --> 50:36.440 | |
And in poetry, sometimes you see the same. | |
50:36.440 --> 50:40.440 | |
And if you look back at your childhood, | |
50:40.440 --> 50:42.440 | |
you grew up in Russia, | |
50:42.440 --> 50:46.440 | |
you maybe were born as a researcher in Russia, | |
50:46.440 --> 50:48.440 | |
you developed as a researcher in Russia, | |
50:48.440 --> 50:51.440 | |
you came to the United States in a few places. | |
50:51.440 --> 50:53.440 | |
If you look back, | |
50:53.440 --> 50:59.440 | |
what were some of your happiest moments as a researcher? | |
50:59.440 --> 51:02.440 | |
Some of the most profound moments. | |
51:02.440 --> 51:06.440 | |
Not in terms of their impact on society, | |
51:06.440 --> 51:12.440 | |
but in terms of their impact on how damn good you feel that day, | |
51:12.440 --> 51:15.440 | |
and you remember that moment. | |
51:15.440 --> 51:20.440 | |
You know, every time when you found something, | |
51:20.440 --> 51:22.440 | |
it is great. | |
51:22.440 --> 51:24.440 | |
It's a life. | |
51:24.440 --> 51:26.440 | |
Every simple thing. | |
51:26.440 --> 51:32.440 | |
But my general feeling that most of my time was wrong. | |
51:32.440 --> 51:35.440 | |
You should go again and again and again | |
51:35.440 --> 51:39.440 | |
and try to be honest in front of yourself. | |
51:39.440 --> 51:41.440 | |
Not to make interpretation, | |
51:41.440 --> 51:46.440 | |
but try to understand that it's related to ground truth. | |
51:46.440 --> 51:52.440 | |
It is not my blah, blah, blah interpretation or something like that. | |
51:52.440 --> 51:57.440 | |
But you're allowed to get excited at the possibility of discovery. | |
51:57.440 --> 52:00.440 | |
You have to double check it, but... | |
52:00.440 --> 52:04.440 | |
No, but how it's related to the other ground truth | |
52:04.440 --> 52:10.440 | |
is it just temporary or it is forever? | |
52:10.440 --> 52:13.440 | |
You know, you always have a feeling | |
52:13.440 --> 52:17.440 | |
when you found something, | |
52:17.440 --> 52:19.440 | |
how big is that? | |
52:19.440 --> 52:23.440 | |
So, 20 years ago, when we discovered statistical learning, | |
52:23.440 --> 52:25.440 | |
so nobody believed. | |
52:25.440 --> 52:31.440 | |
Except for one guy, Dudley from MIT. | |
52:31.440 --> 52:36.440 | |
And then in 20 years, it became fashion. | |
52:36.440 --> 52:39.440 | |
And the same with support vector machines. | |
52:39.440 --> 52:41.440 | |
That's kernel machines. | |
52:41.440 --> 52:44.440 | |
So with support vector machines and learning theory, | |
52:44.440 --> 52:48.440 | |
when you were working on it, | |
52:48.440 --> 52:55.440 | |
you had a sense that you had a sense of the profundity of it, | |
52:55.440 --> 52:59.440 | |
how this seems to be right. | |
52:59.440 --> 53:01.440 | |
It seems to be powerful. | |
53:01.440 --> 53:04.440 | |
Right, absolutely, immediately. | |
53:04.440 --> 53:08.440 | |
I recognize that it will last forever. | |
53:08.440 --> 53:17.440 | |
And now, when I found this invariance story, | |
53:17.440 --> 53:21.440 | |
I have a feeling that it is completely wrong. | |
53:21.440 --> 53:25.440 | |
Because I have proved that there are no different mechanisms. | |
53:25.440 --> 53:30.440 | |
Some say cosmetic improvement you can do, | |
53:30.440 --> 53:34.440 | |
but in terms of invariance, | |
53:34.440 --> 53:38.440 | |
you need both invariance and statistical learning | |
53:38.440 --> 53:41.440 | |
and they should work together. | |
53:41.440 --> 53:47.440 | |
But also, I'm happy that we can formulate | |
53:47.440 --> 53:51.440 | |
what is intelligence from that | |
53:51.440 --> 53:54.440 | |
and to separate from technical part. | |
53:54.440 --> 53:56.440 | |
And that is completely different. | |
53:56.440 --> 53:58.440 | |
Absolutely. | |
53:58.440 --> 54:00.440 | |
Well, Vladimir, thank you so much for talking today. | |
54:00.440 --> 54:01.440 | |
Thank you. | |
54:01.440 --> 54:28.440 | |
Thank you very much. | |