diff --git "a/CS_441_2023_Spring_February_07,_2023.vtt" "b/CS_441_2023_Spring_February_07,_2023.vtt" new file mode 100644--- /dev/null +++ "b/CS_441_2023_Spring_February_07,_2023.vtt" @@ -0,0 +1,5198 @@ +WEBVTT Kind: captions; Language: en-US + +NOTE +Created on 2024-02-07T20:54:31.1029159Z by ClassTranscribe + +00:01:27.710 --> 00:01:28.060 +All right. + +00:01:28.060 --> 00:01:28.990 +Good morning, everybody. + +00:01:30.560 --> 00:01:32.070 +Hope you had a good weekend. + +00:01:33.880 --> 00:01:35.350 +Form relatively. + +00:01:37.950 --> 00:01:40.110 +Alright, so I'm going to get started. + +00:01:40.110 --> 00:01:42.920 +So in the previous lectures we've + +00:01:42.920 --> 00:01:44.820 +mainly learned about how to build and + +00:01:44.820 --> 00:01:46.220 +apply single models. + +00:01:46.220 --> 00:01:48.550 +So we talked about nearest neighbor, + +00:01:48.550 --> 00:01:50.915 +logistic regression, linear regression, + +00:01:50.915 --> 00:01:51.960 +and trees. + +00:01:51.960 --> 00:01:54.609 +And so now we're going to. + +00:01:55.570 --> 00:01:57.676 +Talk about how to build collection of + +00:01:57.676 --> 00:01:59.850 +models and use them for prediction. + +00:01:59.850 --> 00:02:02.045 +So that technique is called ensembles + +00:02:02.045 --> 00:02:05.280 +and ensemble is when you build a bunch + +00:02:05.280 --> 00:02:07.420 +of models and then you average their + +00:02:07.420 --> 00:02:09.430 +predictions or you train them in a way + +00:02:09.430 --> 00:02:11.040 +that they build on top of each other. + +00:02:12.270 --> 00:02:14.020 +So some of you might remember this show + +00:02:14.020 --> 00:02:15.160 +who wants to be a millionaire? + +00:02:16.100 --> 00:02:18.520 +The idea of this show is that there's a + +00:02:18.520 --> 00:02:20.490 +contestant and they get asked a series + +00:02:20.490 --> 00:02:22.280 +of questions and they have multiple + +00:02:22.280 --> 00:02:25.030 +choice answers and if they get it right + +00:02:25.030 --> 00:02:27.020 +then like the dollar value that they + +00:02:27.020 --> 00:02:29.429 +would bring home increases, but if they + +00:02:29.430 --> 00:02:31.280 +ever get it wrong, then they go home + +00:02:31.280 --> 00:02:31.910 +with nothing. + +00:02:32.620 --> 00:02:35.150 +And they had three forms of help. + +00:02:35.150 --> 00:02:37.070 +One of the forms was that they could + +00:02:37.070 --> 00:02:39.380 +eliminate 2 of the incorrect choices. + +00:02:40.230 --> 00:02:42.517 +Another form is that they could call a + +00:02:42.517 --> 00:02:42.769 +friend. + +00:02:42.770 --> 00:02:44.610 +So they would have like people. + +00:02:44.610 --> 00:02:46.210 +They would have friends at home that + +00:02:46.210 --> 00:02:48.695 +they think have like various expertise. + +00:02:48.695 --> 00:02:51.135 +And if they see a question that they + +00:02:51.135 --> 00:02:52.450 +think is really hard and they're not + +00:02:52.450 --> 00:02:54.220 +sure of the answer, they could choose + +00:02:54.220 --> 00:02:55.946 +which friend to call to give them the + +00:02:55.946 --> 00:02:56.199 +answer. + +00:02:57.660 --> 00:03:00.120 +The third, the third form of help they + +00:03:00.120 --> 00:03:02.910 +could get is pull the audience so. + +00:03:03.680 --> 00:03:06.475 +They would ask the audience to vote on + +00:03:06.475 --> 00:03:07.520 +the correct answer. + +00:03:08.120 --> 00:03:11.120 +And the audience would all vote, and + +00:03:11.120 --> 00:03:12.530 +then they could make a decision based + +00:03:12.530 --> 00:03:13.190 +on that. + +00:03:14.020 --> 00:03:15.745 +And they could use each of these forms + +00:03:15.745 --> 00:03:17.850 +of help one time. + +00:03:18.780 --> 00:03:22.369 +What do you which of these do you think + +00:03:22.370 --> 00:03:24.270 +between pull the audience and call a + +00:03:24.270 --> 00:03:24.900 +friend? + +00:03:24.900 --> 00:03:28.369 +Which of these do you think is a is + +00:03:28.370 --> 00:03:30.590 +more likely to give the correct answer? + +00:03:33.500 --> 00:03:35.020 +Alright, so how many people think it's + +00:03:35.020 --> 00:03:36.250 +pulled the audience? + +00:03:36.250 --> 00:03:39.710 +How many people think it's for in a + +00:03:39.710 --> 00:03:40.210 +friend? + +00:03:42.060 --> 00:03:45.000 +So the audience is correct, it's pulled + +00:03:45.000 --> 00:03:45.540 +the audience. + +00:03:46.250 --> 00:03:49.975 +But they did statistics. + +00:03:49.975 --> 00:03:52.910 +They looked at analysis of the show and + +00:03:52.910 --> 00:03:55.110 +on average the audience is correct 92% + +00:03:55.110 --> 00:03:56.240 +of the time. + +00:03:57.050 --> 00:03:59.750 +And call a friend is correct 66% of the + +00:03:59.750 --> 00:04:00.150 +time. + +00:04:01.780 --> 00:04:04.500 +So that might be kind of unintuitive, + +00:04:04.500 --> 00:04:06.970 +especially the margin, because. + +00:04:08.210 --> 00:04:09.574 +When you get to call a friend, you get + +00:04:09.574 --> 00:04:11.670 +to call somebody who you think knows + +00:04:11.670 --> 00:04:13.620 +about the particular subject matter. + +00:04:13.620 --> 00:04:15.300 +So they're an expert. + +00:04:15.300 --> 00:04:16.562 +You would expect that out of. + +00:04:16.562 --> 00:04:18.200 +You would expect that they would be + +00:04:18.200 --> 00:04:20.160 +much, much more informed than an + +00:04:20.160 --> 00:04:22.770 +average audience member who is just + +00:04:22.770 --> 00:04:24.020 +there to be entertained. + +00:04:24.880 --> 00:04:28.190 +But the audience is actually much more + +00:04:28.190 --> 00:04:30.160 +accurate and that kind of that + +00:04:30.160 --> 00:04:32.330 +demonstrates the power of ensembles + +00:04:32.330 --> 00:04:34.370 +that averaging multiple weak + +00:04:34.370 --> 00:04:35.140 +predictions. + +00:04:35.830 --> 00:04:38.720 +Is often more accurate than any single + +00:04:38.720 --> 00:04:40.003 +predictor, even if that single + +00:04:40.003 --> 00:04:41.150 +predictor is pretty good. + +00:04:43.770 --> 00:04:46.464 +It's possible to construct models to + +00:04:46.464 --> 00:04:48.269 +construct ensembles in different ways. + +00:04:48.270 --> 00:04:49.930 +One of the ways is that you + +00:04:49.930 --> 00:04:51.745 +independently train a bunch of + +00:04:51.745 --> 00:04:53.540 +different models by resampling the data + +00:04:53.540 --> 00:04:55.830 +or resampling features, and then you + +00:04:55.830 --> 00:04:57.846 +average those the predictions of those + +00:04:57.846 --> 00:04:58.119 +models. + +00:04:58.810 --> 00:05:00.780 +Another is that you incrementally train + +00:05:00.780 --> 00:05:02.860 +new models that try to fix the mistakes + +00:05:02.860 --> 00:05:04.350 +of the previous models. + +00:05:04.350 --> 00:05:05.750 +So we're going to talk about both of + +00:05:05.750 --> 00:05:06.170 +those. + +00:05:06.790 --> 00:05:08.800 +And they work on different principles. + +00:05:08.800 --> 00:05:10.758 +There's different reasons why each one + +00:05:10.758 --> 00:05:13.460 +is a is a reasonable choice. + +00:05:16.420 --> 00:05:19.740 +So the theory behind ensembles really + +00:05:19.740 --> 00:05:22.260 +comes down to this theorem called the + +00:05:22.260 --> 00:05:24.480 +balance, the bias variance tradeoff. + +00:05:25.110 --> 00:05:27.040 +And this is a really fundamental + +00:05:27.040 --> 00:05:28.850 +concept in machine learning. + +00:05:29.690 --> 00:05:31.730 +And I'm not going to go through the + +00:05:31.730 --> 00:05:33.780 +derivation of it, it's at this link + +00:05:33.780 --> 00:05:34.084 +here. + +00:05:34.084 --> 00:05:34.692 +It's not. + +00:05:34.692 --> 00:05:36.280 +It's not really, it's something that + +00:05:36.280 --> 00:05:37.960 +anyone could follow along, but it does + +00:05:37.960 --> 00:05:38.980 +take a while to get through it. + +00:05:40.280 --> 00:05:41.880 +But it's a really fundamental idea in + +00:05:41.880 --> 00:05:42.740 +machine learning. + +00:05:42.740 --> 00:05:46.390 +So in terms of one way that you can + +00:05:46.390 --> 00:05:48.560 +express it is in terms of the squared + +00:05:48.560 --> 00:05:49.610 +error of prediction. + +00:05:50.620 --> 00:05:53.220 +So for regression, but there's also + +00:05:53.220 --> 00:05:55.949 +equivalent theorems for classification, + +00:05:55.949 --> 00:05:59.450 +for 01 classification or for log + +00:05:59.450 --> 00:06:00.760 +probability loss. + +00:06:01.870 --> 00:06:04.460 +And it all works out to the same thing, + +00:06:04.460 --> 00:06:06.080 +which is that you're expected test + +00:06:06.080 --> 00:06:06.670 +error. + +00:06:06.670 --> 00:06:08.599 +So what this means is that. + +00:06:09.350 --> 00:06:11.490 +If you were to randomly choose some + +00:06:11.490 --> 00:06:13.410 +number of samples from the general + +00:06:13.410 --> 00:06:14.680 +distribution of data. + +00:06:15.900 --> 00:06:18.530 +Then the expected error that you would + +00:06:18.530 --> 00:06:20.410 +get for the model that you've trained + +00:06:20.410 --> 00:06:24.230 +on your sample of data compared to what + +00:06:24.230 --> 00:06:25.260 +it should have predicted. + +00:06:26.680 --> 00:06:29.560 +Has three different components, so one + +00:06:29.560 --> 00:06:30.910 +component is the variance. + +00:06:31.590 --> 00:06:34.095 +The variance is that if UV sampled that + +00:06:34.095 --> 00:06:36.510 +same amount of data multiple times from + +00:06:36.510 --> 00:06:38.590 +the general distribution, you'd get + +00:06:38.590 --> 00:06:40.390 +different data samples and that would + +00:06:40.390 --> 00:06:41.920 +lead to different models that make + +00:06:41.920 --> 00:06:43.660 +different predictions on the same test + +00:06:43.660 --> 00:06:44.000 +data. + +00:06:44.730 --> 00:06:46.710 +So you have some variance in your + +00:06:46.710 --> 00:06:47.180 +prediction. + +00:06:47.180 --> 00:06:48.470 +That's due to the randomness of + +00:06:48.470 --> 00:06:49.600 +sampling your model. + +00:06:49.600 --> 00:06:52.049 +Or it could be due to if you have a + +00:06:52.050 --> 00:06:53.023 +randomized optimization. + +00:06:53.023 --> 00:06:54.390 +It could also be due to the + +00:06:54.390 --> 00:06:56.060 +randomization of the optimization. + +00:06:57.910 --> 00:07:00.360 +So this is a variance mainly due to + +00:07:00.360 --> 00:07:02.580 +resampling data of your model. + +00:07:03.580 --> 00:07:05.760 +Compared to your expected model. + +00:07:05.760 --> 00:07:08.919 +So this is how the sum of the average + +00:07:08.920 --> 00:07:12.310 +square distance between the predictions + +00:07:12.310 --> 00:07:15.240 +of an individual model and the average + +00:07:15.240 --> 00:07:17.080 +over all possible models that you would + +00:07:17.080 --> 00:07:18.500 +learn from sampling the data many + +00:07:18.500 --> 00:07:18.940 +times. + +00:07:20.570 --> 00:07:23.270 +Then there's a skip over here for now. + +00:07:23.270 --> 00:07:25.347 +Then there's a bias component squared. + +00:07:25.347 --> 00:07:28.690 +So the bias is if you were to sample + +00:07:28.690 --> 00:07:31.820 +the data infinite times, train your + +00:07:31.820 --> 00:07:33.375 +infinite models and average them, then + +00:07:33.375 --> 00:07:35.497 +you get this expected prediction. + +00:07:35.497 --> 00:07:37.940 +So it's the expected the average + +00:07:37.940 --> 00:07:39.850 +prediction of all of those infinite + +00:07:39.850 --> 00:07:41.240 +models that you trained with the same + +00:07:41.240 --> 00:07:41.949 +amount of data. + +00:07:43.010 --> 00:07:44.460 +And if you look at the difference + +00:07:44.460 --> 00:07:46.790 +between that and the true prediction, + +00:07:46.790 --> 00:07:48.030 +then that's your bias. + +00:07:49.220 --> 00:07:53.070 +So if you have no bias, then obviously + +00:07:53.070 --> 00:07:55.655 +if you have no bias this would be 0. + +00:07:55.655 --> 00:07:57.379 +If on average your models would + +00:07:57.380 --> 00:07:59.095 +converge to the true answer, this will + +00:07:59.095 --> 00:07:59.700 +be 0. + +00:07:59.700 --> 00:08:01.660 +But if your models tend to predict too + +00:08:01.660 --> 00:08:04.050 +high or too low on average, then this + +00:08:04.050 --> 00:08:05.110 +will be nonzero. + +00:08:06.440 --> 00:08:07.970 +And then finally there's the noise. + +00:08:07.970 --> 00:08:10.710 +So this is kind of like the irreducible + +00:08:10.710 --> 00:08:13.000 +error due to the problem that it might + +00:08:13.000 --> 00:08:14.780 +be that for the exact same input + +00:08:14.780 --> 00:08:16.060 +there's different outputs that are + +00:08:16.060 --> 00:08:17.380 +possible, like if you're trying to + +00:08:17.380 --> 00:08:20.205 +predict temperature or read characters + +00:08:20.205 --> 00:08:22.390 +or something like that, the features + +00:08:22.390 --> 00:08:24.250 +are not sufficient to completely + +00:08:24.250 --> 00:08:26.150 +identify the correct answer. + +00:08:26.970 --> 00:08:29.390 +So there's these three parts to the + +00:08:29.390 --> 00:08:29.690 +error. + +00:08:29.690 --> 00:08:31.330 +There's the variance due to limited + +00:08:31.330 --> 00:08:34.069 +data in your models due to the + +00:08:34.070 --> 00:08:35.800 +randomness in a model. + +00:08:35.800 --> 00:08:38.083 +That's either due to randomly sampling + +00:08:38.083 --> 00:08:40.040 +the data or due to your optimization. + +00:08:40.660 --> 00:08:42.340 +There's the bias, which is due to the + +00:08:42.340 --> 00:08:44.770 +inability of your model to fit the true + +00:08:44.770 --> 00:08:45.390 +solution. + +00:08:46.080 --> 00:08:48.740 +And there's a noise which is due to the + +00:08:48.740 --> 00:08:50.160 +problem characteristics or the + +00:08:50.160 --> 00:08:51.840 +inability to make a perfect prediction + +00:08:51.840 --> 00:08:52.600 +from the features. + +00:08:54.920 --> 00:08:55.410 +Yeah. + +00:08:57.940 --> 00:09:02.930 +So here, so why is a particular? + +00:09:04.210 --> 00:09:08.110 +That particular label assigned to X&Y + +00:09:08.110 --> 00:09:12.260 +bar is the average of all the labels + +00:09:12.260 --> 00:09:14.390 +that you would that could be assigned + +00:09:14.390 --> 00:09:15.260 +to ex. + +00:09:15.260 --> 00:09:18.337 +So for example, if you had imagine that + +00:09:18.337 --> 00:09:20.700 +you had the exact same, let's say your + +00:09:20.700 --> 00:09:22.600 +prediction predicting temperature based + +00:09:22.600 --> 00:09:23.640 +on the last five days. + +00:09:24.360 --> 00:09:26.480 +And you saw that exact same scenario of + +00:09:26.480 --> 00:09:29.675 +the last five days like 15 times, but + +00:09:29.675 --> 00:09:31.620 +you had different next day + +00:09:31.620 --> 00:09:32.340 +temperatures. + +00:09:32.960 --> 00:09:35.683 +So why would be like one of those next + +00:09:35.683 --> 00:09:37.190 +day temperatures and why bar is the + +00:09:37.190 --> 00:09:38.780 +average of those next day temperatures? + +00:09:39.980 --> 00:09:40.460 +Question. + +00:09:43.200 --> 00:09:44.820 +How is your model? + +00:09:44.820 --> 00:09:48.684 +So HD is a model that's trained on a + +00:09:48.684 --> 00:09:51.310 +sample on a DF sample of the + +00:09:51.310 --> 00:09:51.950 +distribution. + +00:09:53.210 --> 00:09:56.310 +And H bar is the average of all such + +00:09:56.310 --> 00:09:56.680 +models. + +00:10:03.740 --> 00:10:07.270 +So the bias and variance is illustrated + +00:10:07.270 --> 00:10:08.215 +here. + +00:10:08.215 --> 00:10:10.500 +So imagine that you're trying to shoot + +00:10:10.500 --> 00:10:11.040 +a target. + +00:10:11.700 --> 00:10:13.833 +Then if you have low bias and low + +00:10:13.833 --> 00:10:15.243 +variance, it means that all your shots + +00:10:15.243 --> 00:10:17.470 +are clustered in the center of the + +00:10:17.470 --> 00:10:17.774 +target. + +00:10:17.774 --> 00:10:20.265 +If you have low bias and high variance + +00:10:20.265 --> 00:10:22.910 +means that the average of your shots is + +00:10:22.910 --> 00:10:24.640 +in the center of your target, but the + +00:10:24.640 --> 00:10:26.260 +shots are more widely distributed. + +00:10:27.890 --> 00:10:31.360 +If you have high bias and low variance, + +00:10:31.360 --> 00:10:33.210 +it means that your shots are clustered + +00:10:33.210 --> 00:10:34.730 +tight together, but they're off the + +00:10:34.730 --> 00:10:35.160 +center. + +00:10:35.940 --> 00:10:37.580 +And if you have high bias and high + +00:10:37.580 --> 00:10:40.298 +variance, then both they're dispersed, + +00:10:40.298 --> 00:10:42.560 +dispersed, and they're off the center. + +00:10:44.230 --> 00:10:45.920 +So you can see from even from this + +00:10:45.920 --> 00:10:48.924 +illustration that obviously low bias + +00:10:48.924 --> 00:10:51.840 +and low variance is the best, but both + +00:10:51.840 --> 00:10:54.267 +variance and bias caused some error, + +00:10:54.267 --> 00:10:56.590 +and high bias and high variance has the + +00:10:56.590 --> 00:10:57.950 +greatest average error. + +00:11:02.670 --> 00:11:04.988 +You also often see a expressed in a + +00:11:04.988 --> 00:11:07.147 +plot like this, where you're looking at + +00:11:07.147 --> 00:11:09.654 +your model complexity and this is like. + +00:11:09.654 --> 00:11:10.990 +This is kind of like a classic + +00:11:10.990 --> 00:11:13.580 +overfitting plot, so this model + +00:11:13.580 --> 00:11:15.240 +complexity could for example be the + +00:11:15.240 --> 00:11:16.440 +height of your tree. + +00:11:17.540 --> 00:11:19.420 +So if you train a tree with two leaf + +00:11:19.420 --> 00:11:22.930 +nodes with just a height of 1, then + +00:11:22.930 --> 00:11:24.754 +you're going to have a very low + +00:11:24.754 --> 00:11:25.016 +variance. + +00:11:25.016 --> 00:11:26.900 +If you were to resample the data many + +00:11:26.900 --> 00:11:29.259 +times and train that short tree, you + +00:11:29.260 --> 00:11:30.790 +would very likely get a very similar + +00:11:30.790 --> 00:11:33.304 +tree every single time, so the variance + +00:11:33.304 --> 00:11:33.980 +is low. + +00:11:33.980 --> 00:11:34.870 +That's the blue curve. + +00:11:35.760 --> 00:11:37.100 +But the bias is high. + +00:11:37.100 --> 00:11:38.580 +You're unlikely to make very good + +00:11:38.580 --> 00:11:40.070 +predictions with that really short + +00:11:40.070 --> 00:11:40.880 +tree. + +00:11:40.880 --> 00:11:43.275 +Even if you averaged an infinite number + +00:11:43.275 --> 00:11:44.189 +of them, you would still. + +00:11:44.189 --> 00:11:45.570 +You would still have a lot of error. + +00:11:46.960 --> 00:11:49.520 +As you increase the depth of the tree, + +00:11:49.520 --> 00:11:51.290 +your bias drops. + +00:11:51.290 --> 00:11:53.232 +You're able to make better predictions + +00:11:53.232 --> 00:11:56.030 +on your on average. + +00:11:57.250 --> 00:11:59.340 +But the variance starts to increase. + +00:11:59.340 --> 00:12:01.030 +The trees start to look more different + +00:12:01.030 --> 00:12:01.920 +from each other. + +00:12:01.920 --> 00:12:04.780 +So if you train a full tree so that + +00:12:04.780 --> 00:12:06.990 +there's one data point per leaf node, + +00:12:06.990 --> 00:12:08.410 +then the trees are going to look pretty + +00:12:08.410 --> 00:12:10.230 +different when you resample the data + +00:12:10.230 --> 00:12:11.550 +because you'll have different data + +00:12:11.550 --> 00:12:12.080 +samples. + +00:12:13.850 --> 00:12:16.460 +So eventually, at some point you reach + +00:12:16.460 --> 00:12:19.616 +some ideal situation where the bias + +00:12:19.616 --> 00:12:21.677 +plus the bias squared plus the variance + +00:12:21.677 --> 00:12:23.940 +is minimized, and that's when you'd + +00:12:23.940 --> 00:12:25.510 +want to, like, stop if you're trying to + +00:12:25.510 --> 00:12:26.165 +choose hyperparameters. + +00:12:26.165 --> 00:12:29.530 +And if you train more complex models, + +00:12:29.530 --> 00:12:31.330 +it's going to continue to reduce the + +00:12:31.330 --> 00:12:32.925 +bias, but the increase in variance is + +00:12:32.925 --> 00:12:35.326 +going to cause your test error to + +00:12:35.326 --> 00:12:35.629 +increase. + +00:12:39.100 --> 00:12:41.404 +So if you're thinking about it in terms + +00:12:41.404 --> 00:12:45.510 +of a single model, really this, then + +00:12:45.510 --> 00:12:47.111 +you would be thinking about it in terms + +00:12:47.111 --> 00:12:49.190 +of the plot that I just showed where + +00:12:49.190 --> 00:12:50.690 +you're trying to figure out like what + +00:12:50.690 --> 00:12:52.330 +complexity, if it's a model that can + +00:12:52.330 --> 00:12:54.450 +have varying complexity trees or neural + +00:12:54.450 --> 00:12:57.327 +networks, like how complex should my + +00:12:57.327 --> 00:12:59.550 +model be in order to best. + +00:13:00.440 --> 00:13:02.285 +Find the balance between the bias and + +00:13:02.285 --> 00:13:02.950 +the variance. + +00:13:03.710 --> 00:13:05.910 +But ensembles have a different way to + +00:13:05.910 --> 00:13:08.050 +directly combat the bias and the + +00:13:08.050 --> 00:13:10.430 +variance, so I'm going to talk about a + +00:13:10.430 --> 00:13:12.470 +few ensemble methods and how they + +00:13:12.470 --> 00:13:12.920 +relate. + +00:13:16.400 --> 00:13:19.130 +The first one is called first, like. + +00:13:19.130 --> 00:13:20.580 +This is actually not one of these + +00:13:20.580 --> 00:13:22.007 +ensemble method, but it is an ensemble + +00:13:22.007 --> 00:13:22.245 +method. + +00:13:22.245 --> 00:13:23.690 +It's the simplest of these, and it's + +00:13:23.690 --> 00:13:25.219 +kind of the foundation of the ensemble + +00:13:25.220 --> 00:13:25.810 +methods. + +00:13:25.810 --> 00:13:28.010 +So it's a statistical technique called + +00:13:28.010 --> 00:13:28.710 +bootstrapping. + +00:13:29.860 --> 00:13:32.740 +Imagine that, for example, I wanted to + +00:13:32.740 --> 00:13:35.170 +know what is the average age of + +00:13:35.170 --> 00:13:36.380 +somebody in this class. + +00:13:37.610 --> 00:13:39.990 +One way that I could do it is I could + +00:13:39.990 --> 00:13:42.323 +ask each of you your ages and then I + +00:13:42.323 --> 00:13:43.840 +could average it, and then that might + +00:13:43.840 --> 00:13:45.605 +give me like an estimate for the + +00:13:45.605 --> 00:13:47.110 +average age of all the students in the + +00:13:47.110 --> 00:13:47.450 +class. + +00:13:48.720 --> 00:13:51.700 +But maybe I not only want to know the + +00:13:51.700 --> 00:13:53.850 +average age, but I also want some + +00:13:53.850 --> 00:13:56.020 +confidence range on that average age. + +00:13:56.020 --> 00:13:58.210 +And if all I do is I average all your + +00:13:58.210 --> 00:14:00.960 +ages, that doesn't tell me how likely I + +00:14:00.960 --> 00:14:02.930 +am to be within, say, like three years. + +00:14:04.000 --> 00:14:07.090 +And so one way, one way that I can + +00:14:07.090 --> 00:14:09.950 +solve that problem is with bootstrap + +00:14:09.950 --> 00:14:13.590 +estimation where I resample the data + +00:14:13.590 --> 00:14:15.530 +multiple times so I could choose. + +00:14:15.530 --> 00:14:18.800 +I could take 50 samples and sample with + +00:14:18.800 --> 00:14:21.235 +repetition so I could potentially call + +00:14:21.235 --> 00:14:22.350 +the same person twice. + +00:14:23.160 --> 00:14:24.125 +Ask your ages. + +00:14:24.125 --> 00:14:26.750 +Ask the ages of 50 individuals. + +00:14:26.750 --> 00:14:28.140 +Again, the same individual may be + +00:14:28.140 --> 00:14:28.870 +repeated. + +00:14:28.870 --> 00:14:31.530 +I take the average from that and repeat + +00:14:31.530 --> 00:14:33.810 +that many times, and then I can look at + +00:14:33.810 --> 00:14:35.579 +the variance of those estimates that I + +00:14:35.580 --> 00:14:35.800 +get. + +00:14:36.470 --> 00:14:38.050 +And then I can use the variance of + +00:14:38.050 --> 00:14:40.430 +those estimates to get a confidence + +00:14:40.430 --> 00:14:42.570 +range on my estimate of the mean. + +00:14:43.810 --> 00:14:47.080 +So bootstrap bootstrapping is a way to. + +00:14:47.190 --> 00:14:50.710 +To estimate a particular parameter, in + +00:14:50.710 --> 00:14:53.035 +this case the average age, as well as + +00:14:53.035 --> 00:14:55.040 +my variance of my estimate of that + +00:14:55.040 --> 00:14:55.690 +parameter. + +00:14:55.690 --> 00:14:58.550 +So like how far off am I would expect + +00:14:58.550 --> 00:14:58.970 +to be? + +00:15:02.560 --> 00:15:04.300 +We can apply that idea to + +00:15:04.300 --> 00:15:08.918 +classification to try to produce a more + +00:15:08.918 --> 00:15:11.266 +stable estimate of the mean or to + +00:15:11.266 --> 00:15:13.370 +produce a more stable prediction. + +00:15:13.370 --> 00:15:15.270 +In other words, to reduce the variance + +00:15:15.270 --> 00:15:17.930 +of my classifiers given a particular + +00:15:17.930 --> 00:15:18.620 +data sample. + +00:15:20.250 --> 00:15:23.010 +So the method is called bagging, which + +00:15:23.010 --> 00:15:24.890 +stands for aggregate bootstrapping. + +00:15:25.990 --> 00:15:27.390 +And the idea is pretty simple. + +00:15:28.630 --> 00:15:32.340 +For M different times capital M, So I'm + +00:15:32.340 --> 00:15:34.730 +going to train train M classifiers. + +00:15:35.430 --> 00:15:37.620 +I draw some number of samples which + +00:15:37.620 --> 00:15:39.533 +should be less than my total number of + +00:15:39.533 --> 00:15:40.800 +samples, but I'm going to draw them + +00:15:40.800 --> 00:15:41.828 +with replacement. + +00:15:41.828 --> 00:15:43.860 +Draw with replacement means I can + +00:15:43.860 --> 00:15:45.310 +choose the same sample twice. + +00:15:46.750 --> 00:15:48.410 +Then I train a classifier on those + +00:15:48.410 --> 00:15:51.120 +samples, and then at the end my final + +00:15:51.120 --> 00:15:54.290 +classifier is an average of all of my + +00:15:54.290 --> 00:15:55.620 +predictions from the individual + +00:15:55.620 --> 00:15:56.340 +classifiers. + +00:15:57.080 --> 00:15:59.040 +So if I'm doing regression, I would + +00:15:59.040 --> 00:16:01.940 +just be averaging the continuous values + +00:16:01.940 --> 00:16:04.200 +that the classifiers are aggressors + +00:16:04.200 --> 00:16:04.890 +predicted. + +00:16:04.890 --> 00:16:07.555 +If I'm doing classification, I would + +00:16:07.555 --> 00:16:10.116 +average the probabilities or average + +00:16:10.116 --> 00:16:13.056 +the most likely label from each of the + +00:16:13.056 --> 00:16:13.389 +classifiers. + +00:16:14.380 --> 00:16:16.810 +And there's lots of theory that shows + +00:16:16.810 --> 00:16:19.100 +that this increases the stability of + +00:16:19.100 --> 00:16:21.500 +the classifier and reduces reduces the + +00:16:21.500 --> 00:16:24.915 +variance, and so the average of a bunch + +00:16:24.915 --> 00:16:26.630 +of classifiers trained this way. + +00:16:27.300 --> 00:16:30.110 +Typically outperform any individual + +00:16:30.110 --> 00:16:30.840 +classifier. + +00:16:32.030 --> 00:16:33.870 +In these classifiers will be different + +00:16:33.870 --> 00:16:36.490 +from each other because there's a + +00:16:36.490 --> 00:16:37.100 +difference. + +00:16:37.100 --> 00:16:39.670 +Because the data is, a different sample + +00:16:39.670 --> 00:16:41.030 +of data is drawn to train each + +00:16:41.030 --> 00:16:41.590 +classifier. + +00:16:45.070 --> 00:16:46.790 +So that's the question. + +00:17:00.050 --> 00:17:02.463 +So not yeah, but not features, it's + +00:17:02.463 --> 00:17:03.186 +samples. + +00:17:03.186 --> 00:17:06.700 +So I have say 1000 data samples. + +00:17:07.340 --> 00:17:10.770 +And I draw say 900 data samples, but + +00:17:10.770 --> 00:17:13.467 +they're not 900 out of the thousand, + +00:17:13.467 --> 00:17:16.190 +it's 900 with repetition. + +00:17:16.190 --> 00:17:17.720 +So there might be 1 sample that I + +00:17:17.720 --> 00:17:19.596 +choose draw three times, others that I + +00:17:19.596 --> 00:17:21.259 +draw no times, others that I draw one + +00:17:21.260 --> 00:17:21.850 +time. + +00:17:21.850 --> 00:17:23.700 +So you can in terms of like + +00:17:23.700 --> 00:17:26.840 +programming, you would just do a random + +00:17:26.840 --> 00:17:31.290 +like 0 to 1 * N and then and then turn + +00:17:31.290 --> 00:17:33.397 +it into an integer and then you get + +00:17:33.397 --> 00:17:35.159 +like you get a random sample with + +00:17:35.160 --> 00:17:35.660 +replacement. + +00:17:46.940 --> 00:17:47.720 +Typically. + +00:17:47.720 --> 00:17:49.626 +So usually each of the classifiers is + +00:17:49.626 --> 00:17:50.820 +of the same form. + +00:17:50.820 --> 00:17:51.190 +Yep. + +00:17:53.550 --> 00:17:55.270 +So this is the idea behind random + +00:17:55.270 --> 00:17:57.760 +forests, which is a really powerful + +00:17:57.760 --> 00:17:59.940 +classifier, but very easy to explain at + +00:17:59.940 --> 00:18:01.500 +least once you once you know about + +00:18:01.500 --> 00:18:02.270 +decision trees. + +00:18:03.780 --> 00:18:06.040 +So in a random forest, train a + +00:18:06.040 --> 00:18:07.150 +collection of trees. + +00:18:08.140 --> 00:18:09.970 +For each tree that you're going to + +00:18:09.970 --> 00:18:11.786 +train, you sample some fraction in the + +00:18:11.786 --> 00:18:13.880 +data, for example 90% of the data. + +00:18:13.880 --> 00:18:15.620 +Sometimes people just sample all the + +00:18:15.620 --> 00:18:15.990 +data. + +00:18:16.430 --> 00:18:19.948 +Then you randomly sample some number of + +00:18:19.948 --> 00:18:20.325 +features. + +00:18:20.325 --> 00:18:23.042 +So for regression, one suggestion is to + +00:18:23.042 --> 00:18:24.648 +use 1/3 of the features. + +00:18:24.648 --> 00:18:28.003 +For classification you would use like. + +00:18:28.003 --> 00:18:30.000 +Some suggestions are to use like a + +00:18:30.000 --> 00:18:31.565 +square root of the number of features. + +00:18:31.565 --> 00:18:32.240 +So if there's. + +00:18:32.970 --> 00:18:36.260 +If there are 400 features, then you + +00:18:36.260 --> 00:18:38.290 +randomly sample 20 of them. + +00:18:38.290 --> 00:18:40.240 +Or another suggestion is to use log + +00:18:40.240 --> 00:18:40.820 +base 2. + +00:18:41.650 --> 00:18:43.389 +It's not really that critical, but you + +00:18:43.389 --> 00:18:44.820 +want you want the number of features + +00:18:44.820 --> 00:18:46.995 +that you select to be much less than + +00:18:46.995 --> 00:18:48.430 +the total number of features. + +00:18:49.110 --> 00:18:51.800 +So here previously I was talking about + +00:18:51.800 --> 00:18:53.760 +when I say sample the data, what I mean + +00:18:53.760 --> 00:18:55.870 +is like is choosing a subset of + +00:18:55.870 --> 00:18:56.790 +training samples. + +00:18:57.910 --> 00:19:00.290 +But when I say sample the features, I + +00:19:00.290 --> 00:19:02.699 +mean choose a subset of the features of + +00:19:02.699 --> 00:19:05.365 +the columns of your of your matrix if + +00:19:05.365 --> 00:19:06.914 +the rows are samples and the columns + +00:19:06.914 --> 00:19:07.350 +are features. + +00:19:09.360 --> 00:19:11.710 +So the you need to sample the features + +00:19:11.710 --> 00:19:13.210 +because otherwise if you train the tree + +00:19:13.210 --> 00:19:14.693 +you're going to get the same result if + +00:19:14.693 --> 00:19:17.720 +you're doing like minimizing the + +00:19:17.720 --> 00:19:19.440 +maximizing mutual information for + +00:19:19.440 --> 00:19:19.890 +example. + +00:19:20.700 --> 00:19:22.270 +If you were to sample all your data and + +00:19:22.270 --> 00:19:23.600 +all the features, you would just train + +00:19:23.600 --> 00:19:24.280 +the same tree. + +00:19:25.070 --> 00:19:27.660 +MN times and that would give you no + +00:19:27.660 --> 00:19:28.160 +benefit. + +00:19:28.900 --> 00:19:30.240 +All right, so you randomly sample some + +00:19:30.240 --> 00:19:31.540 +features, train a tree. + +00:19:32.240 --> 00:19:34.497 +Optionally, you can estimate your + +00:19:34.497 --> 00:19:36.020 +validation error on the data that + +00:19:36.020 --> 00:19:38.283 +wasn't used to train that tree, and you + +00:19:38.283 --> 00:19:41.140 +can use the average of those validation + +00:19:41.140 --> 00:19:44.513 +errors in order to get a estimate of + +00:19:44.513 --> 00:19:46.930 +your error for the for your final + +00:19:46.930 --> 00:19:47.480 +collection. + +00:19:50.000 --> 00:19:51.886 +And after you've trained all the trees, + +00:19:51.886 --> 00:19:54.610 +you just do that 100 times or whatever. + +00:19:54.610 --> 00:19:55.920 +It's completely independent. + +00:19:55.920 --> 00:19:58.330 +So it's just like a very if you've got + +00:19:58.330 --> 00:19:59.920 +code to train a tree, it's just a very + +00:19:59.920 --> 00:20:01.090 +small loop. + +00:20:02.370 --> 00:20:04.990 +And then at the end you average the + +00:20:04.990 --> 00:20:06.766 +prediction of all the trees. + +00:20:06.766 --> 00:20:08.930 +So usually you would train your trees + +00:20:08.930 --> 00:20:09.535 +to completion. + +00:20:09.535 --> 00:20:12.160 +So if you're doing like classification + +00:20:12.160 --> 00:20:14.850 +or in either case you would end up with + +00:20:14.850 --> 00:20:16.480 +a leaf node that contains one data + +00:20:16.480 --> 00:20:16.926 +sample. + +00:20:16.926 --> 00:20:19.060 +So you're training like very high + +00:20:19.060 --> 00:20:21.530 +variance trees, they're deep trees. + +00:20:22.650 --> 00:20:24.760 +That have low bias, they can fit the + +00:20:24.760 --> 00:20:27.580 +training data perfectly, but. + +00:20:29.470 --> 00:20:31.027 +But then you're going to average all of + +00:20:31.027 --> 00:20:31.235 +them. + +00:20:31.235 --> 00:20:34.534 +So you start out with high bias or high + +00:20:34.534 --> 00:20:36.650 +variance, low bias classifiers, and + +00:20:36.650 --> 00:20:37.743 +then you average them. + +00:20:37.743 --> 00:20:40.044 +So you end up with low bias, low + +00:20:40.044 --> 00:20:40.669 +variance classifiers. + +00:20:49.930 --> 00:20:51.310 +Yes, for each tree. + +00:20:51.310 --> 00:20:52.460 +Yeah, for each tree. + +00:20:52.630 --> 00:20:53.160 +Yeah. + +00:20:59.180 --> 00:21:02.920 +You increase the number of trees, yeah, + +00:21:02.920 --> 00:21:03.410 +so. + +00:21:04.110 --> 00:21:07.720 +If you if so, think of it this way. + +00:21:07.720 --> 00:21:12.075 +If I were to if I were to try to + +00:21:12.075 --> 00:21:14.995 +estimate the sum of your ages, then as + +00:21:14.995 --> 00:21:17.900 +I ask you your ages and add them up, my + +00:21:17.900 --> 00:21:19.463 +estimate of the variance of the + +00:21:19.463 --> 00:21:21.288 +variance on the estimate, the sum is + +00:21:21.288 --> 00:21:23.400 +going to increase linearly, right? + +00:21:23.400 --> 00:21:26.680 +It's going to keep on increasing until + +00:21:26.680 --> 00:21:30.660 +sum is 100,000 ± 10,000 or something. + +00:21:31.480 --> 00:21:33.168 +But if I'm trying to estimate the + +00:21:33.168 --> 00:21:35.700 +average of your ages and I keep on + +00:21:35.700 --> 00:21:38.250 +asking your ages, then my variance is + +00:21:38.250 --> 00:21:39.950 +going to go down South. + +00:21:39.950 --> 00:21:43.040 +The variance of the sum is North Times + +00:21:43.040 --> 00:21:47.030 +Sigma squared, but the variance of the + +00:21:47.030 --> 00:21:50.980 +average is N over Sigma I think just no + +00:21:50.980 --> 00:21:53.688 +over Sigma or sorry, Sigma over N, + +00:21:53.688 --> 00:21:56.100 +Sigma squared over N the variance of + +00:21:56.100 --> 00:21:58.513 +the average is Sigma squared over N, + +00:21:58.513 --> 00:22:01.269 +but the variance of the sum is N. + +00:22:01.330 --> 00:22:02.500 +Times Sigma squared. + +00:22:04.490 --> 00:22:06.934 +So the average reduces the variance. + +00:22:06.934 --> 00:22:08.135 +Yeah, so if I. + +00:22:08.135 --> 00:22:09.960 +So by averaging the trees I reduce the + +00:22:09.960 --> 00:22:10.160 +variance. + +00:22:14.870 --> 00:22:17.250 +So that's random forests and I will + +00:22:17.250 --> 00:22:17.840 +talk more. + +00:22:17.840 --> 00:22:20.467 +I'll give an example of use of random + +00:22:20.467 --> 00:22:22.280 +forests and I'll talk about like some + +00:22:22.280 --> 00:22:24.780 +studies about the performance of + +00:22:24.780 --> 00:22:26.750 +various classifiers including random + +00:22:26.750 --> 00:22:27.320 +forests. + +00:22:27.320 --> 00:22:29.946 +But before I do that, I want to talk + +00:22:29.946 --> 00:22:31.330 +about boosting, which is the other + +00:22:31.330 --> 00:22:31.890 +strategy. + +00:22:33.860 --> 00:22:36.080 +So I have the boosting terms here as + +00:22:36.080 --> 00:22:36.490 +well. + +00:22:37.730 --> 00:22:38.170 +All right. + +00:22:38.170 --> 00:22:41.085 +So the first version of boosting and + +00:22:41.085 --> 00:22:42.740 +one other thing I want to say about + +00:22:42.740 --> 00:22:45.350 +this is random forest was popularized + +00:22:45.350 --> 00:22:47.885 +by this paper by Bremen in 2001. + +00:22:47.885 --> 00:22:50.460 +So decision trees go back to the 90s at + +00:22:50.460 --> 00:22:53.893 +least, but they were never really, like + +00:22:53.893 --> 00:22:56.680 +I said, were they're good for helping + +00:22:56.680 --> 00:22:59.750 +for making decisions that people can + +00:22:59.750 --> 00:23:01.360 +understand, that you can communicate + +00:23:01.360 --> 00:23:02.780 +and explain like why it made this + +00:23:02.780 --> 00:23:03.130 +decision. + +00:23:03.890 --> 00:23:05.710 +And they're good for analyzing data, + +00:23:05.710 --> 00:23:07.040 +but they're not really very good + +00:23:07.040 --> 00:23:08.770 +classifiers or aggressors compared to + +00:23:08.770 --> 00:23:09.880 +other methods that are out there. + +00:23:11.210 --> 00:23:14.390 +But Bremen popularized random forests + +00:23:14.390 --> 00:23:16.530 +in 2001 and showed that the + +00:23:16.530 --> 00:23:19.050 +combinations of trees is actually super + +00:23:19.050 --> 00:23:20.380 +powerful and super useful. + +00:23:21.840 --> 00:23:23.770 +And provides like the theory for why it + +00:23:23.770 --> 00:23:25.800 +works and why you should be sampling + +00:23:25.800 --> 00:23:27.780 +different subsets of features, and the + +00:23:27.780 --> 00:23:29.160 +idea that you want the trees to be + +00:23:29.160 --> 00:23:30.000 +decorrelated. + +00:23:31.000 --> 00:23:34.130 +To make different predictions but also + +00:23:34.130 --> 00:23:34.800 +be powerful. + +00:23:37.140 --> 00:23:37.710 +Alright. + +00:23:37.710 --> 00:23:41.140 +So the other strategy is boosting and + +00:23:41.140 --> 00:23:42.910 +the first boosting paper I think was + +00:23:42.910 --> 00:23:44.630 +Shapira in 1989. + +00:23:45.500 --> 00:23:46.900 +And that's one was pretty simple. + +00:23:47.680 --> 00:23:51.090 +So the idea was that you first randomly + +00:23:51.090 --> 00:23:52.690 +choose a set of samples. + +00:23:53.470 --> 00:23:55.280 +Without replacement at this time. + +00:23:55.280 --> 00:23:57.970 +So if you've got 1000, you randomly + +00:23:57.970 --> 00:24:00.133 +choose, say, 800 of them without + +00:24:00.133 --> 00:24:00.539 +replacement. + +00:24:01.440 --> 00:24:04.320 +And you train a classifier on those + +00:24:04.320 --> 00:24:07.140 +samples, that's the weak learner, C1. + +00:24:07.760 --> 00:24:10.170 +So I've got the notation over here in + +00:24:10.170 --> 00:24:12.060 +the literature you'll see things like + +00:24:12.060 --> 00:24:15.140 +learner, hypothesis, classifier, they + +00:24:15.140 --> 00:24:16.130 +all mean the same thing. + +00:24:16.130 --> 00:24:17.560 +There's something that's some model + +00:24:17.560 --> 00:24:18.810 +that's doing some prediction. + +00:24:19.960 --> 00:24:22.530 +A weak learner is just a classifier + +00:24:22.530 --> 00:24:25.260 +that can achieve less than 50% training + +00:24:25.260 --> 00:24:27.140 +error over any training distribution. + +00:24:27.910 --> 00:24:30.120 +So almost any classifier we would + +00:24:30.120 --> 00:24:32.217 +consider is a weak learner. + +00:24:32.217 --> 00:24:34.000 +As long as you can guarantee that it + +00:24:34.000 --> 00:24:35.970 +will be able to get at least chance + +00:24:35.970 --> 00:24:38.030 +performance in a two class problem, + +00:24:38.030 --> 00:24:39.309 +then it's a weak learner. + +00:24:42.560 --> 00:24:45.286 +A strong learner is a combination of + +00:24:45.286 --> 00:24:46.182 +the weak learner. + +00:24:46.182 --> 00:24:47.852 +It's a predictor that uses a + +00:24:47.852 --> 00:24:49.230 +combination of the weak learners. + +00:24:49.230 --> 00:24:52.020 +So first you train 1 classifier in a + +00:24:52.020 --> 00:24:52.940 +subset of the data. + +00:24:53.620 --> 00:24:55.936 +Then you draw a new sample, and this + +00:24:55.936 --> 00:24:58.490 +new sample is drawn so that half the + +00:24:58.490 --> 00:24:59.310 +samples. + +00:25:00.010 --> 00:25:04.960 +Are misclassified by the 1st classifier + +00:25:04.960 --> 00:25:06.640 +and this can be drawn with replacement. + +00:25:07.460 --> 00:25:10.172 +So half of your N2 samples were + +00:25:10.172 --> 00:25:12.310 +misclassified by C1 and half of them + +00:25:12.310 --> 00:25:14.009 +were not misclassified by C1. + +00:25:14.900 --> 00:25:17.230 +And so now in this new sample of data. + +00:25:18.500 --> 00:25:21.220 +Your classifier C1 had a 5050 chance of + +00:25:21.220 --> 00:25:22.910 +getting it right by construction. + +00:25:22.980 --> 00:25:23.150 +Right. + +00:25:23.880 --> 00:25:25.640 +Then you train C2. + +00:25:27.060 --> 00:25:29.590 +To try to like do well on this new + +00:25:29.590 --> 00:25:30.560 +distribution. + +00:25:30.560 --> 00:25:32.590 +So C2 has like a more difficult job, + +00:25:32.590 --> 00:25:33.970 +it's going to focus on the things that + +00:25:33.970 --> 00:25:35.240 +C1 found more difficult. + +00:25:37.140 --> 00:25:39.250 +Then finally you take all the samples + +00:25:39.250 --> 00:25:41.830 +that C1 and C2 disagree on, and you + +00:25:41.830 --> 00:25:43.590 +train a third week learner 1/3 + +00:25:43.590 --> 00:25:45.740 +classifier just on those examples. + +00:25:46.420 --> 00:25:49.470 +And then at the end you take an average + +00:25:49.470 --> 00:25:50.500 +of those votes. + +00:25:50.500 --> 00:25:52.621 +So basically you have like you have + +00:25:52.621 --> 00:25:54.050 +like one person who's making a + +00:25:54.050 --> 00:25:54.740 +prediction. + +00:25:55.810 --> 00:25:57.946 +You take half the predictions that + +00:25:57.946 --> 00:26:00.770 +person made incorrect and half that + +00:26:00.770 --> 00:26:02.320 +were correct, and then you get a second + +00:26:02.320 --> 00:26:04.192 +person to make predictions just looking + +00:26:04.192 --> 00:26:05.690 +at that at those samples. + +00:26:06.470 --> 00:26:08.130 +Then you get a third person to be the + +00:26:08.130 --> 00:26:09.915 +tiebreaker between the first two people + +00:26:09.915 --> 00:26:11.440 +if they made if they had different + +00:26:11.440 --> 00:26:13.320 +answers, and then you take a vote of + +00:26:13.320 --> 00:26:14.790 +those three people as you're finally + +00:26:14.790 --> 00:26:15.160 +answer. + +00:26:16.780 --> 00:26:18.590 +Where you can substitute classifier for + +00:26:18.590 --> 00:26:19.290 +people. + +00:26:20.660 --> 00:26:22.100 +So this is the boosting idea. + +00:26:23.100 --> 00:26:25.120 +Now this actually became much more + +00:26:25.120 --> 00:26:27.000 +popular when it was generalized a + +00:26:27.000 --> 00:26:28.480 +little bit into this method called + +00:26:28.480 --> 00:26:31.450 +Adaboost, which stands for adaptive + +00:26:31.450 --> 00:26:31.970 +boosting. + +00:26:33.210 --> 00:26:33.650 +So. + +00:26:34.390 --> 00:26:38.710 +The in adaptive boosting, instead of + +00:26:38.710 --> 00:26:42.940 +justice directly sampling the data, you + +00:26:42.940 --> 00:26:44.730 +assign a weight to the data. + +00:26:44.730 --> 00:26:46.640 +And I'll explain in the next slide, I + +00:26:46.640 --> 00:26:48.564 +think more of what it means to like + +00:26:48.564 --> 00:26:49.860 +weight the data when you're doing + +00:26:49.860 --> 00:26:50.850 +parameter estimation. + +00:26:52.360 --> 00:26:55.200 +But you assign assign new weights to + +00:26:55.200 --> 00:26:57.357 +the data so that under that + +00:26:57.357 --> 00:27:00.036 +distribution the previous weak learner, + +00:27:00.036 --> 00:27:02.140 +the previous classifier has chance + +00:27:02.140 --> 00:27:04.150 +accuracy at that weighted distribution. + +00:27:04.920 --> 00:27:07.775 +So this was one way of doing achieving + +00:27:07.775 --> 00:27:10.010 +the same thing where you just you draw + +00:27:10.010 --> 00:27:12.390 +like whole samples so that the previous + +00:27:12.390 --> 00:27:14.150 +week learner had a 5050 chance of + +00:27:14.150 --> 00:27:16.000 +getting those samples correct. + +00:27:16.830 --> 00:27:18.540 +But you can instead assign a softer + +00:27:18.540 --> 00:27:20.510 +weight to just say that some samples + +00:27:20.510 --> 00:27:23.160 +matter more than others, so that on the + +00:27:23.160 --> 00:27:24.950 +distribution the previous classifier + +00:27:24.950 --> 00:27:26.330 +has a 5050 chance. + +00:27:27.900 --> 00:27:30.680 +Then you train a new classifier on the + +00:27:30.680 --> 00:27:31.820 +reweighted samples. + +00:27:32.440 --> 00:27:33.350 +And then you iterate. + +00:27:33.350 --> 00:27:34.800 +So then you reweigh them again and + +00:27:34.800 --> 00:27:36.340 +train a new classifier and keep doing + +00:27:36.340 --> 00:27:36.850 +that. + +00:27:36.850 --> 00:27:38.870 +And then at the end you take a weighted + +00:27:38.870 --> 00:27:41.560 +vote of all of the weak classifiers as + +00:27:41.560 --> 00:27:42.510 +your final predictor. + +00:27:43.430 --> 00:27:47.810 +So each each sample is going to each + +00:27:47.810 --> 00:27:49.600 +classifier is going to try to correct + +00:27:49.600 --> 00:27:50.760 +the mistakes of the previous + +00:27:50.760 --> 00:27:53.090 +classifiers, and then all of their + +00:27:53.090 --> 00:27:54.650 +predictions are combined. + +00:27:55.920 --> 00:27:57.240 +So I'm going to show a specific + +00:27:57.240 --> 00:27:59.650 +algorithm in a moment, but first I want + +00:27:59.650 --> 00:28:00.520 +to clarify. + +00:28:01.450 --> 00:28:03.610 +What it means to take A to do, like a + +00:28:03.610 --> 00:28:05.880 +weighted estimation or weighting your + +00:28:05.880 --> 00:28:06.720 +training samples. + +00:28:07.560 --> 00:28:09.600 +So essentially it just means that some + +00:28:09.600 --> 00:28:11.795 +samples count more than others towards + +00:28:11.795 --> 00:28:13.780 +your parameter estimation or your + +00:28:13.780 --> 00:28:14.660 +learning objective. + +00:28:15.410 --> 00:28:17.500 +So let's say that we're trying to build + +00:28:17.500 --> 00:28:19.880 +a naive Bayes classifier, and so we + +00:28:19.880 --> 00:28:21.870 +need to estimate the probability that + +00:28:21.870 --> 00:28:24.745 +some feature is equal to 0 given that + +00:28:24.745 --> 00:28:26.130 +the label is equal to 0. + +00:28:26.130 --> 00:28:28.200 +That's like one of the parameters of + +00:28:28.200 --> 00:28:28.940 +our model. + +00:28:29.960 --> 00:28:32.250 +If we have an unweighted distribution, + +00:28:32.250 --> 00:28:35.940 +then that would be a count of how many + +00:28:35.940 --> 00:28:39.290 +times the feature is equal to 0 and the + +00:28:39.290 --> 00:28:40.440 +label is equal to 0. + +00:28:41.070 --> 00:28:43.380 +Divided by a count of how many times + +00:28:43.380 --> 00:28:45.290 +the label is equal to 0, right? + +00:28:45.290 --> 00:28:47.489 +So that's probability of X&Y + +00:28:47.490 --> 00:28:49.112 +essentially divided by probability of + +00:28:49.112 --> 00:28:49.380 +Y. + +00:28:51.950 --> 00:28:53.940 +Times north on the numerator and + +00:28:53.940 --> 00:28:54.720 +denominator. + +00:28:56.520 --> 00:28:58.780 +Then if I want to take a weighted + +00:28:58.780 --> 00:29:01.430 +sample, if I wanted an estimate of a + +00:29:01.430 --> 00:29:03.490 +weighted distribution, I have a weight + +00:29:03.490 --> 00:29:04.840 +assigned to each of these training + +00:29:04.840 --> 00:29:07.570 +samples, and that's often done so that + +00:29:07.570 --> 00:29:11.140 +the weights sum up to one, but it + +00:29:11.140 --> 00:29:12.619 +doesn't have to be, but they have to be + +00:29:12.620 --> 00:29:13.240 +non negative. + +00:29:15.290 --> 00:29:16.950 +OK, so I have to wait for each of these + +00:29:16.950 --> 00:29:18.973 +samples that says how important it is. + +00:29:18.973 --> 00:29:20.940 +So when I count the number of times + +00:29:20.940 --> 00:29:25.320 +that X n = 0 and Y n = 0, then I am + +00:29:25.320 --> 00:29:27.200 +waiting those counts by won. + +00:29:27.200 --> 00:29:29.140 +So it's the sum of the weights where + +00:29:29.140 --> 00:29:31.185 +for the samples in which this condition + +00:29:31.185 --> 00:29:33.698 +is true divided by the sum of the + +00:29:33.698 --> 00:29:35.886 +weights for which YN is equal to 0. + +00:29:35.886 --> 00:29:37.649 +So that's my weighted estimate of that + +00:29:37.650 --> 00:29:38.260 +statistic. + +00:29:40.910 --> 00:29:41.470 +Right. + +00:29:41.470 --> 00:29:42.960 +So it's your turn. + +00:29:44.180 --> 00:29:46.470 +Let's say that we have this table here. + +00:29:46.470 --> 00:29:48.810 +So we've got weights on the left side, + +00:29:48.810 --> 00:29:51.850 +X in the middle, Y and the right, and + +00:29:51.850 --> 00:29:53.735 +I'm trying to estimate probability of X + +00:29:53.735 --> 00:29:55.440 += 0 given y = 0. + +00:29:56.140 --> 00:29:57.950 +So I'll give you a moment to think + +00:29:57.950 --> 00:29:58.690 +about it. + +00:29:58.690 --> 00:30:00.590 +First, what is the unweighted + +00:30:00.590 --> 00:30:03.040 +distribution and then what is the + +00:30:03.040 --> 00:30:04.380 +weighted distribution? + +00:30:12.540 --> 00:30:13.100 +Right. + +00:30:20.290 --> 00:30:21.170 +Me too. + +00:30:21.170 --> 00:30:23.410 +My daughter woke me up at 4:00 AM and I + +00:30:23.410 --> 00:30:24.700 +couldn't fall back asleep. + +00:30:39.450 --> 00:30:41.990 +I'll I will go through these are the + +00:30:41.990 --> 00:30:43.920 +examples, so I'll go through it. + +00:30:45.400 --> 00:30:45.930 +Alright. + +00:30:48.690 --> 00:30:50.650 +Going, I'll step through it in a + +00:30:50.650 --> 00:30:50.930 +moment. + +00:30:52.270 --> 00:30:53.404 +Alright, so let's do the. + +00:30:53.404 --> 00:30:55.090 +Let's do the unweighted first. + +00:30:56.800 --> 00:31:00.940 +So how many times does X equal 0 and y + +00:31:00.940 --> 00:31:01.480 += 0? + +00:31:03.440 --> 00:31:05.030 +Right, three. + +00:31:05.030 --> 00:31:06.350 +OK, so I'm going to have three on the + +00:31:06.350 --> 00:31:09.665 +numerator and how many times does y = + +00:31:09.665 --> 00:31:10.120 +0? + +00:31:12.070 --> 00:31:13.000 +OK, right. + +00:31:13.000 --> 00:31:15.710 +So unweighted is going to be 3 out of + +00:31:15.710 --> 00:31:16.500 +five, right? + +00:31:18.560 --> 00:31:20.470 +Now let's do the weighted. + +00:31:20.470 --> 00:31:22.990 +So what's the sum of the weights where + +00:31:22.990 --> 00:31:25.309 +X = 0 and y = 0? + +00:31:31.640 --> 00:31:35.026 +So there's three rows where X = 0 and y + +00:31:35.026 --> 00:31:35.619 += 0. + +00:31:36.360 --> 00:31:36.830 +Right. + +00:31:39.410 --> 00:31:40.990 +Right, yeah, three. + +00:31:40.990 --> 00:31:42.742 +So there's just these three rows, and + +00:31:42.742 --> 00:31:44.230 +there's a .1 for each of them. + +00:31:44.940 --> 00:31:46.030 +So that's .3. + +00:31:46.800 --> 00:31:49.830 +And what is the total weight for y = 0? + +00:31:51.710 --> 00:31:52.960 +Right .7. + +00:31:54.060 --> 00:31:55.960 +So the weighted distribution. + +00:31:55.960 --> 00:31:57.456 +My estimate on the weighted + +00:31:57.456 --> 00:31:58.920 +distribution is 3 out of seven. + +00:32:00.000 --> 00:32:01.120 +So that's how it works. + +00:32:01.830 --> 00:32:04.770 +And if you had so a lot of times we are + +00:32:04.770 --> 00:32:06.260 +just estimating counts like this. + +00:32:06.260 --> 00:32:08.500 +If we were training a shorter tree for + +00:32:08.500 --> 00:32:11.148 +example, then we would be estimating + +00:32:11.148 --> 00:32:13.330 +the probability of each class within + +00:32:13.330 --> 00:32:14.920 +the leaf node, which would just be by + +00:32:14.920 --> 00:32:15.380 +counting. + +00:32:17.040 --> 00:32:18.980 +Other times, if you're doing like + +00:32:18.980 --> 00:32:21.515 +logistic regression or had some other + +00:32:21.515 --> 00:32:24.000 +kind of training or neural network, + +00:32:24.000 --> 00:32:26.660 +then usually these weights would show + +00:32:26.660 --> 00:32:28.410 +up as some kind of like weight on the + +00:32:28.410 --> 00:32:29.140 +loss. + +00:32:29.140 --> 00:32:31.290 +So we're going to talk about a + +00:32:31.290 --> 00:32:32.740 +sarcastic gradient descent. + +00:32:33.750 --> 00:32:35.110 +Starting in the next class. + +00:32:35.720 --> 00:32:37.725 +And a higher weight would just be like + +00:32:37.725 --> 00:32:39.440 +a direct multiple on how much you + +00:32:39.440 --> 00:32:42.230 +adjust your model parameters. + +00:32:45.810 --> 00:32:47.920 +So here's a specific algorithm called + +00:32:47.920 --> 00:32:49.040 +Adaboost. + +00:32:49.440 --> 00:32:52.289 +A real boost, I mean, there's like a + +00:32:52.290 --> 00:32:53.816 +ton of boosting algorithms. + +00:32:53.816 --> 00:32:56.037 +There's like discrete ETA boost, real + +00:32:56.037 --> 00:32:57.695 +boost, logic boost. + +00:32:57.695 --> 00:32:59.186 +I don't know. + +00:32:59.186 --> 00:33:01.880 +There's like literally like probably 50 + +00:33:01.880 --> 00:33:02.260 +of them. + +00:33:03.670 --> 00:33:05.660 +But here's one of the mainstays. + +00:33:05.660 --> 00:33:08.930 +So you start with the weights being + +00:33:08.930 --> 00:33:09.560 +uniform. + +00:33:09.560 --> 00:33:11.700 +They're one over north with N samples. + +00:33:11.700 --> 00:33:13.240 +Then you're going to train M + +00:33:13.240 --> 00:33:14.160 +classifiers. + +00:33:14.910 --> 00:33:17.605 +You fit the classifier to obtain a + +00:33:17.605 --> 00:33:19.690 +probability estimate, the probability + +00:33:19.690 --> 00:33:22.630 +of the label being one based on the + +00:33:22.630 --> 00:33:23.620 +weighted distribution. + +00:33:24.500 --> 00:33:26.130 +So again, if you're doing trees, this + +00:33:26.130 --> 00:33:28.460 +would be the fraction of samples in + +00:33:28.460 --> 00:33:30.040 +each leaf node of the trees where the + +00:33:30.040 --> 00:33:31.000 +label is equal to 1. + +00:33:31.850 --> 00:33:33.530 +And where you'd be using a weighted + +00:33:33.530 --> 00:33:35.530 +sample to compute that fraction, just + +00:33:35.530 --> 00:33:36.580 +like we did in the last slide. + +00:33:37.750 --> 00:33:39.860 +Then the prediction of this the score + +00:33:39.860 --> 00:33:43.369 +essentially for the label one is this + +00:33:43.370 --> 00:33:44.110 +logic. + +00:33:44.110 --> 00:33:47.960 +It's the log probability of the label + +00:33:47.960 --> 00:33:50.240 +being one over the probability not + +00:33:50.240 --> 00:33:51.943 +being one, which is 1 minus the + +00:33:51.943 --> 00:33:52.892 +probability of it being one. + +00:33:52.892 --> 00:33:54.470 +This is for binary classifier. + +00:33:55.650 --> 00:33:57.570 +That's 1/2 of that logic value. + +00:33:58.780 --> 00:34:03.040 +And then I re weight the samples and I + +00:34:03.040 --> 00:34:05.330 +take the previous weight of each sample + +00:34:05.330 --> 00:34:07.240 +and I multiply it by east to the + +00:34:07.240 --> 00:34:09.440 +negative yiff FMX. + +00:34:09.440 --> 00:34:11.047 +So this again is a score. + +00:34:11.047 --> 00:34:13.370 +So this score defined this way, if it's + +00:34:13.370 --> 00:34:15.260 +greater than zero that means that. + +00:34:16.090 --> 00:34:21.220 +If Y ifm is greater than zero, here Yi + +00:34:21.220 --> 00:34:24.144 +is either -, 1 or one, so -, 1 is the + +00:34:24.144 --> 00:34:25.900 +negative label, one is the positive + +00:34:25.900 --> 00:34:26.200 +label. + +00:34:26.910 --> 00:34:28.484 +If this is greater than zero, that + +00:34:28.484 --> 00:34:30.449 +means that I'm correct, and if it's + +00:34:30.450 --> 00:34:31.969 +less than zero it means that I'm + +00:34:31.970 --> 00:34:32.960 +incorrect. + +00:34:32.960 --> 00:34:34.846 +So if I predict a score of 1, it means + +00:34:34.846 --> 00:34:36.540 +that I think it's positive. + +00:34:36.540 --> 00:34:40.597 +But if the label is -, 1, then Y ifm is + +00:34:40.597 --> 00:34:41.850 +-, 1, so. + +00:34:44.620 --> 00:34:48.350 +So this negative Y ifm, if I'm correct + +00:34:48.350 --> 00:34:49.990 +this is going to be less than one + +00:34:49.990 --> 00:34:53.450 +because this is going to be east to the + +00:34:53.450 --> 00:34:54.860 +negative sum value. + +00:34:55.970 --> 00:34:57.700 +And if I'm incorrect, this is going to + +00:34:57.700 --> 00:34:58.600 +be greater than one. + +00:34:59.270 --> 00:35:00.993 +So if I'm correct, the weight is going + +00:35:00.993 --> 00:35:03.141 +to go down, and if I'm incorrect the + +00:35:03.141 --> 00:35:04.070 +weight is going to go up. + +00:35:04.830 --> 00:35:06.650 +And if I'm like confidently correct, + +00:35:06.650 --> 00:35:07.908 +then the way it's going to go down a + +00:35:07.908 --> 00:35:08.156 +lot. + +00:35:08.156 --> 00:35:09.835 +And if I'm confidently incorrect then + +00:35:09.835 --> 00:35:10.960 +the weight is going to go up a lot. + +00:35:12.410 --> 00:35:13.480 +That's kind of intuitive. + +00:35:14.120 --> 00:35:15.470 +And then I just reweigh. + +00:35:15.470 --> 00:35:17.630 +I just sum my. + +00:35:18.910 --> 00:35:19.480 +My weight. + +00:35:19.480 --> 00:35:22.050 +I renormalize my weights, so I make it + +00:35:22.050 --> 00:35:23.460 +so that the weights sum to one by + +00:35:23.460 --> 00:35:24.479 +dividing by the sum. + +00:35:25.980 --> 00:35:27.630 +So then I just iterate, then I train + +00:35:27.630 --> 00:35:29.235 +new classifier and the way distribution + +00:35:29.235 --> 00:35:31.430 +recompute this, recompute the weights, + +00:35:31.430 --> 00:35:33.300 +do that say 20 times. + +00:35:33.910 --> 00:35:36.642 +And then at the end my classifier is. + +00:35:36.642 --> 00:35:38.607 +My total score for the classifier is + +00:35:38.607 --> 00:35:40.430 +the sum of the individual classifier + +00:35:40.430 --> 00:35:40.840 +scores. + +00:35:42.130 --> 00:35:43.300 +So it's not too complicated. + +00:35:44.220 --> 00:35:47.163 +That theory is somewhat complicated, so + +00:35:47.163 --> 00:35:49.310 +the derivation of why this is the right + +00:35:49.310 --> 00:35:51.240 +answer and what it's minimizing, and + +00:35:51.240 --> 00:35:52.500 +that it's like doing with just + +00:35:52.500 --> 00:35:54.840 +aggression, et cetera, that's all a + +00:35:54.840 --> 00:35:56.960 +little bit more complicated, but it's + +00:35:56.960 --> 00:35:58.250 +well worth reading if you're + +00:35:58.250 --> 00:35:58.660 +interested. + +00:35:58.660 --> 00:36:00.046 +So there's a link here. + +00:36:00.046 --> 00:36:02.085 +This is my favorite boosting paper, + +00:36:02.085 --> 00:36:03.780 +that out of logistic regression paper. + +00:36:04.510 --> 00:36:07.660 +But this paper is also probably a good + +00:36:07.660 --> 00:36:08.080 +one to read. + +00:36:08.080 --> 00:36:11.440 +First, the intro to boosting by friend + +00:36:11.440 --> 00:36:12.040 +and Shapiro. + +00:36:16.960 --> 00:36:18.910 +So we can use this with trees. + +00:36:18.910 --> 00:36:21.420 +We initialize the weights to be + +00:36:21.420 --> 00:36:22.190 +uniform. + +00:36:22.190 --> 00:36:24.250 +Then for each tree, usually you do like + +00:36:24.250 --> 00:36:24.840 +maybe 20. + +00:36:25.520 --> 00:36:27.740 +You train a small tree this time. + +00:36:28.880 --> 00:36:31.370 +So you want to train a small tree, + +00:36:31.370 --> 00:36:33.550 +because the idea of boosting is that + +00:36:33.550 --> 00:36:36.020 +you're going to reduce the variance by + +00:36:36.020 --> 00:36:38.270 +having each subsequent classifier fix + +00:36:38.270 --> 00:36:39.810 +the mistakes of the previous ones. + +00:36:40.880 --> 00:36:44.580 +So in random forests you have high + +00:36:44.580 --> 00:36:46.730 +variance, low bias classifiers that + +00:36:46.730 --> 00:36:49.650 +you've averaged to get low biased low + +00:36:49.650 --> 00:36:50.490 +variance classifiers. + +00:36:51.170 --> 00:36:53.560 +In boosting you have low variance, high + +00:36:53.560 --> 00:36:56.400 +bias classifiers that you incrementally + +00:36:56.400 --> 00:36:58.730 +train to end up with a low biased, low + +00:36:58.730 --> 00:36:59.580 +variance classifier. + +00:37:01.600 --> 00:37:04.470 +So you the tree to a depth, typically + +00:37:04.470 --> 00:37:05.620 +two to four. + +00:37:05.620 --> 00:37:07.960 +So often it might sound silly, but + +00:37:07.960 --> 00:37:09.690 +often you only choose one feature and + +00:37:09.690 --> 00:37:11.096 +split based on that, and you just have + +00:37:11.096 --> 00:37:13.020 +like the shortest tree possible, a tree + +00:37:13.020 --> 00:37:16.050 +with two leaf nodes, and you train 200 + +00:37:16.050 --> 00:37:16.910 +of these trees. + +00:37:17.600 --> 00:37:19.975 +That actually is surprisingly it works. + +00:37:19.975 --> 00:37:22.810 +It works quite well, but you might + +00:37:22.810 --> 00:37:23.840 +train deeper trees. + +00:37:25.890 --> 00:37:28.880 +So I've used this method for predicting + +00:37:28.880 --> 00:37:31.400 +like whether pixels belong to the + +00:37:31.400 --> 00:37:34.300 +ground or sky or et cetera, and I had + +00:37:34.300 --> 00:37:37.945 +like trees that were of death three and + +00:37:37.945 --> 00:37:39.180 +I trained 20 trees. + +00:37:40.810 --> 00:37:43.480 +You estimate you estimate logic + +00:37:43.480 --> 00:37:44.810 +prediction at each leaf node. + +00:37:44.810 --> 00:37:46.840 +So just based on the count of how many + +00:37:46.840 --> 00:37:48.860 +times each class appears in each leaf + +00:37:48.860 --> 00:37:50.780 +node, reweigh the samples and repeat. + +00:37:52.060 --> 00:37:53.780 +And then at the end you have the + +00:37:53.780 --> 00:37:55.290 +prediction is the sum of the logic + +00:37:55.290 --> 00:37:56.610 +predictions from all the trees. + +00:37:59.890 --> 00:38:02.470 +So this is a. + +00:38:03.810 --> 00:38:07.490 +There's like one study by there's a + +00:38:07.490 --> 00:38:09.590 +couple of studies by Caruana of + +00:38:09.590 --> 00:38:11.110 +comparing different machine learning + +00:38:11.110 --> 00:38:11.600 +methods. + +00:38:12.320 --> 00:38:14.720 +On a bunch of different datasets, so + +00:38:14.720 --> 00:38:16.660 +this one is from 2006. + +00:38:17.480 --> 00:38:20.300 +So these are all different data sets. + +00:38:20.300 --> 00:38:21.750 +It's not too important what they are. + +00:38:22.950 --> 00:38:24.610 +In this case, they're kind of smaller + +00:38:24.610 --> 00:38:26.470 +data sets, not too not too many + +00:38:26.470 --> 00:38:27.890 +samples, not too many features. + +00:38:28.620 --> 00:38:31.520 +And the scores are normalized so that + +00:38:31.520 --> 00:38:34.040 +one is like the best achievable score + +00:38:34.040 --> 00:38:37.130 +and I guess zero would be like chance. + +00:38:37.130 --> 00:38:39.940 +So that way you can average the + +00:38:39.940 --> 00:38:41.890 +performance across different data sets + +00:38:41.890 --> 00:38:43.300 +in a more meaningful way than if you + +00:38:43.300 --> 00:38:44.660 +were just averaging their errors. + +00:38:46.020 --> 00:38:47.760 +So here this is like a normalized + +00:38:47.760 --> 00:38:50.200 +accuracy, so higher is better. + +00:38:51.260 --> 00:38:54.700 +And then this BTDT is boosted decision + +00:38:54.700 --> 00:38:56.760 +tree, our F is random forest and north + +00:38:56.760 --> 00:38:59.020 +is neural network, Ann SVM, which we'll + +00:38:59.020 --> 00:39:01.420 +talk about Thursday night Bayes, + +00:39:01.420 --> 00:39:02.630 +logistic regression. + +00:39:02.630 --> 00:39:05.580 +So Naive Bayes is like pulling up the + +00:39:05.580 --> 00:39:06.980 +rear, not doing so well. + +00:39:06.980 --> 00:39:08.055 +It's at the very bottom. + +00:39:08.055 --> 00:39:10.236 +The district regression is just above + +00:39:10.236 --> 00:39:10.588 +that. + +00:39:10.588 --> 00:39:12.370 +Decision trees are just above that. + +00:39:13.160 --> 00:39:14.890 +And then boosted stumps. + +00:39:14.890 --> 00:39:17.130 +If you train a very shallow tree that + +00:39:17.130 --> 00:39:19.540 +only has one feature in each tree, + +00:39:19.540 --> 00:39:20.810 +that's the next best. + +00:39:20.810 --> 00:39:22.010 +It's actually pretty similar to + +00:39:22.010 --> 00:39:22.930 +logistic regression. + +00:39:24.050 --> 00:39:29.110 +K&N near neural networks SVMS. + +00:39:29.760 --> 00:39:32.860 +And then the top is boosted decision + +00:39:32.860 --> 00:39:33.940 +trees and random forests. + +00:39:34.680 --> 00:39:36.440 +And there's different versions of this, + +00:39:36.440 --> 00:39:37.903 +which is just like different ways of + +00:39:37.903 --> 00:39:39.130 +trying to calibrate your final + +00:39:39.130 --> 00:39:40.550 +prediction, which means trying to make + +00:39:40.550 --> 00:39:41.890 +it better fit of probability. + +00:39:41.890 --> 00:39:44.055 +But that's not our topic for now, so + +00:39:44.055 --> 00:39:45.290 +that's kind of ignorable. + +00:39:46.110 --> 00:39:48.350 +Main the main conclusion is that in + +00:39:48.350 --> 00:39:50.690 +this competition among classifiers. + +00:39:51.340 --> 00:39:54.690 +Boosted decision trees is #1 and + +00:39:54.690 --> 00:39:56.950 +following very close behind is random + +00:39:56.950 --> 00:39:58.810 +forests with almost the same average + +00:39:58.810 --> 00:39:59.180 +score. + +00:40:00.070 --> 00:40:01.890 +So these two ensemble methods of trees + +00:40:01.890 --> 00:40:03.070 +are the two best methods. + +00:40:04.040 --> 00:40:05.030 +According to the study. + +00:40:06.160 --> 00:40:07.990 +Then in 2008 they did another + +00:40:07.990 --> 00:40:11.110 +comparison on high dimensional data. + +00:40:12.360 --> 00:40:14.570 +So here they had the features range + +00:40:14.570 --> 00:40:17.900 +from around 700 features to 685,000 + +00:40:17.900 --> 00:40:18.870 +features. + +00:40:19.750 --> 00:40:21.540 +This is like IMDb where you're trying + +00:40:21.540 --> 00:40:25.490 +to predict the rating of movies. + +00:40:25.490 --> 00:40:28.750 +I think spam classification and other + +00:40:28.750 --> 00:40:29.210 +problems. + +00:40:30.100 --> 00:40:32.340 +And then again, they're comparing the + +00:40:32.340 --> 00:40:33.460 +different approaches. + +00:40:33.460 --> 00:40:36.675 +So again, boosted decision trees gets + +00:40:36.675 --> 00:40:38.400 +the best score on average. + +00:40:38.400 --> 00:40:41.030 +I don't know exactly how the weighting + +00:40:41.030 --> 00:40:42.480 +is done here, they can be greater than + +00:40:42.480 --> 00:40:42.580 +one. + +00:40:43.270 --> 00:40:45.410 +But boosted decision trees probably + +00:40:45.410 --> 00:40:46.963 +compared to some baseline boosted + +00:40:46.963 --> 00:40:48.610 +decision trees gets the best score on + +00:40:48.610 --> 00:40:49.340 +average. + +00:40:49.340 --> 00:40:51.650 +And random forests is number 2. + +00:40:51.650 --> 00:40:53.660 +Again, it's naive Bayes on the bottom. + +00:40:53.750 --> 00:40:54.210 + + +00:40:55.000 --> 00:40:56.420 +Logistic regression does a bit better + +00:40:56.420 --> 00:40:57.780 +and this high dimensional data. + +00:40:57.780 --> 00:40:59.420 +Again, linear classifiers are more + +00:40:59.420 --> 00:41:00.950 +powerful when you have more features, + +00:41:00.950 --> 00:41:03.980 +but still not outperforming their + +00:41:03.980 --> 00:41:05.750 +neural networks or SVM or random + +00:41:05.750 --> 00:41:06.140 +forests. + +00:41:07.950 --> 00:41:10.620 +But also, even though boosted decision + +00:41:10.620 --> 00:41:13.070 +trees did the best on average, they're + +00:41:13.070 --> 00:41:15.150 +not doing so when you have tons of + +00:41:15.150 --> 00:41:15.940 +features. + +00:41:15.940 --> 00:41:17.926 +They're random forest is doing the + +00:41:17.926 --> 00:41:18.189 +best. + +00:41:19.490 --> 00:41:22.200 +And the reason for that is that boosted + +00:41:22.200 --> 00:41:27.580 +decision trees have a weakness of that. + +00:41:27.810 --> 00:41:29.700 +High. + +00:41:29.770 --> 00:41:30.380 + + +00:41:31.500 --> 00:41:31.932 +They have. + +00:41:31.932 --> 00:41:33.480 +They have a weakness of tending to + +00:41:33.480 --> 00:41:35.100 +overfit the data if they've got too + +00:41:35.100 --> 00:41:36.210 +much flexibility. + +00:41:36.210 --> 00:41:39.049 +So if you have 600,000 features and + +00:41:39.050 --> 00:41:40.512 +you're trying to just fix the mistakes + +00:41:40.512 --> 00:41:42.930 +of the previous classifier iteratively, + +00:41:42.930 --> 00:41:44.400 +then there's a pretty good chance that + +00:41:44.400 --> 00:41:45.840 +you could fix those mistakes for the + +00:41:45.840 --> 00:41:46.365 +wrong reason. + +00:41:46.365 --> 00:41:47.970 +And so they tend to be. + +00:41:47.970 --> 00:41:49.847 +When you have a lot of features, you + +00:41:49.847 --> 00:41:52.596 +end up with high, high variance, high + +00:41:52.596 --> 00:41:55.186 +bias features that you then reduce the + +00:41:55.186 --> 00:41:57.588 +variance of, but you still end up with + +00:41:57.588 --> 00:41:59.840 +high variance, low bias features + +00:41:59.840 --> 00:42:00.710 +classifiers. + +00:42:05.030 --> 00:42:07.480 +So just to recap that boosted decision + +00:42:07.480 --> 00:42:09.150 +trees and random forests work for + +00:42:09.150 --> 00:42:10.063 +different reasons. + +00:42:10.063 --> 00:42:12.345 +Boosted trees use a lot of small trees + +00:42:12.345 --> 00:42:14.430 +to iteratively refine the prediction, + +00:42:14.430 --> 00:42:16.445 +and combining the prediction from many + +00:42:16.445 --> 00:42:18.020 +trees reduces the bias. + +00:42:18.020 --> 00:42:20.380 +But they have a danger of overfitting + +00:42:20.380 --> 00:42:22.717 +if you have too many trees, or the + +00:42:22.717 --> 00:42:24.640 +trees are too big or you have too many + +00:42:24.640 --> 00:42:25.160 +features. + +00:42:25.820 --> 00:42:28.470 +Then they may not generalize that well. + +00:42:29.740 --> 00:42:32.170 +Random forests used big trees, which + +00:42:32.170 --> 00:42:34.050 +are low bias and high variance. + +00:42:34.050 --> 00:42:36.000 +They average a lot of those tree + +00:42:36.000 --> 00:42:38.303 +predictions, which reduces the + +00:42:38.303 --> 00:42:40.170 +variance, and it's kind of hard to make + +00:42:40.170 --> 00:42:41.079 +them not work. + +00:42:41.080 --> 00:42:42.900 +They're not always like the very best + +00:42:42.900 --> 00:42:46.320 +thing you can do, but they always, as + +00:42:46.320 --> 00:42:48.240 +far as I can see and I've ever seen, + +00:42:48.240 --> 00:42:49.810 +they always work like at least pretty + +00:42:49.810 --> 00:42:50.110 +well. + +00:42:51.130 --> 00:42:52.790 +As long as you just train enough trees. + +00:42:55.870 --> 00:42:56.906 +Ensemble. + +00:42:56.906 --> 00:43:00.090 +There's other kinds of ensembles too, + +00:43:00.090 --> 00:43:01.635 +so you can average the predictions of + +00:43:01.635 --> 00:43:03.280 +any classifiers as long as they're not + +00:43:03.280 --> 00:43:04.210 +duplicates of each other. + +00:43:04.210 --> 00:43:05.323 +If they're duplicates of each other, + +00:43:05.323 --> 00:43:07.150 +you don't get any benefit, obviously, + +00:43:07.150 --> 00:43:08.260 +because they'll just make the same + +00:43:08.260 --> 00:43:08.720 +prediction. + +00:43:10.000 --> 00:43:12.170 +So you can also apply this to deep + +00:43:12.170 --> 00:43:13.510 +neural networks, for example. + +00:43:13.510 --> 00:43:15.650 +So here is something showing that + +00:43:15.650 --> 00:43:19.120 +cascades and averages on average + +00:43:19.120 --> 00:43:21.430 +ensembles of classifiers outperform + +00:43:21.430 --> 00:43:23.260 +single classifiers even when you're + +00:43:23.260 --> 00:43:25.470 +considering the computation required + +00:43:25.470 --> 00:43:26.110 +for them. + +00:43:27.550 --> 00:43:29.460 +And a cascade is when you train one + +00:43:29.460 --> 00:43:30.340 +classifier. + +00:43:31.050 --> 00:43:34.512 +And then you let it make its confident + +00:43:34.512 --> 00:43:36.180 +decisions, and then subsequent + +00:43:36.180 --> 00:43:38.240 +classifiers only make decisions about + +00:43:38.240 --> 00:43:39.280 +the less confident. + +00:43:40.500 --> 00:43:41.660 +Examples. + +00:43:41.660 --> 00:43:42.870 +And then you keep on doing that. + +00:43:46.120 --> 00:43:49.770 +Let me give you a two-minute stretch + +00:43:49.770 --> 00:43:51.430 +break before I go into a detailed + +00:43:51.430 --> 00:43:53.670 +example of using random forests. + +00:43:54.690 --> 00:43:56.620 +And you can think about this question + +00:43:56.620 --> 00:43:57.220 +if you want. + +00:43:57.920 --> 00:44:00.120 +So suppose you had an infinite size + +00:44:00.120 --> 00:44:03.100 +audience and where and they could + +00:44:03.100 --> 00:44:04.100 +choose ABCD. + +00:44:05.500 --> 00:44:07.120 +What is the situation where you're + +00:44:07.120 --> 00:44:08.845 +guaranteed to have a correct answer? + +00:44:08.845 --> 00:44:11.410 +What if, let's say, a randomly sampled + +00:44:11.410 --> 00:44:12.970 +audience member is going to report an + +00:44:12.970 --> 00:44:14.800 +answer with probability PY? + +00:44:15.770 --> 00:44:17.650 +What guarantees a correct answer? + +00:44:17.650 --> 00:44:19.930 +And let's say instead you choose a + +00:44:19.930 --> 00:44:21.850 +friend which is a random member of the + +00:44:21.850 --> 00:44:22.830 +audience in this case. + +00:44:23.570 --> 00:44:24.900 +What's the probability that your + +00:44:24.900 --> 00:44:25.930 +friend's answer is correct? + +00:44:26.560 --> 00:44:28.950 +So think about those or don't. + +00:44:28.950 --> 00:44:30.280 +It's up to you. + +00:44:30.280 --> 00:44:31.790 +I'll give you the answer in 2 minutes. + +00:45:07.040 --> 00:45:09.180 +Some people would, they would say like + +00:45:09.180 --> 00:45:11.130 +cherry or yeah. + +00:45:13.980 --> 00:45:14.270 +Yeah. + +00:45:15.730 --> 00:45:17.400 +Or they might be color blind. + +00:45:18.390 --> 00:45:18.960 +I see. + +00:45:24.750 --> 00:45:25.310 +That's true. + +00:45:29.140 --> 00:45:31.120 +It's actually pretty hard not get a + +00:45:31.120 --> 00:45:32.550 +correct answer, I would say. + +00:45:43.340 --> 00:45:46.300 +Correct decision wide away look goes + +00:45:46.300 --> 00:45:49.670 +down because you want the subsequent + +00:45:49.670 --> 00:45:51.240 +classifiers to focus more on the + +00:45:51.240 --> 00:45:52.050 +mistakes. + +00:45:52.050 --> 00:45:56.300 +So if it's incorrect then the weight + +00:45:56.300 --> 00:45:57.920 +goes up so then it matters more to the + +00:45:57.920 --> 00:45:58.730 +next classifier. + +00:46:02.730 --> 00:46:04.160 +Unclassified award goes to. + +00:46:06.000 --> 00:46:07.700 +It could go back up, yeah. + +00:46:10.830 --> 00:46:12.670 +The weights keeping being multiplied by + +00:46:12.670 --> 00:46:14.500 +that factor, so yeah. + +00:46:15.520 --> 00:46:15.870 +Yeah. + +00:46:17.280 --> 00:46:17.700 +You're welcome. + +00:46:25.930 --> 00:46:27.410 +All right, times up. + +00:46:28.930 --> 00:46:32.470 +So what is like the weakest condition? + +00:46:32.470 --> 00:46:34.270 +I should have made it a little harder. + +00:46:34.270 --> 00:46:35.900 +Obviously there's one condition, which + +00:46:35.900 --> 00:46:37.450 +is that every audience member knows the + +00:46:37.450 --> 00:46:37.820 +answer. + +00:46:37.820 --> 00:46:38.380 +That's easy. + +00:46:39.350 --> 00:46:41.160 +But what's the weakest condition that + +00:46:41.160 --> 00:46:43.090 +guarantees a correct answer? + +00:46:43.090 --> 00:46:45.725 +So what has to be true for this answer + +00:46:45.725 --> 00:46:47.330 +to be correct with an infinite audience + +00:46:47.330 --> 00:46:47.710 +size? + +00:46:52.040 --> 00:46:52.530 +Right. + +00:46:54.740 --> 00:46:56.290 +Yes, one audience member. + +00:46:56.290 --> 00:46:57.810 +No, that won't work. + +00:46:57.810 --> 00:46:59.550 +So because then the probability would + +00:46:59.550 --> 00:47:03.790 +be 0 right of the correct answer if all + +00:47:03.790 --> 00:47:05.470 +the other audience members thought it + +00:47:05.470 --> 00:47:06.280 +was a different answer. + +00:47:10.760 --> 00:47:12.740 +If this size of the audience is one, + +00:47:12.740 --> 00:47:14.936 +yeah, but you have an infinite size + +00:47:14.936 --> 00:47:15.940 +audience and the problem. + +00:47:18.270 --> 00:47:18.770 +Does anybody? + +00:47:18.770 --> 00:47:19.940 +Yeah. + +00:47:23.010 --> 00:47:24.938 +Yes, the probability of the correct + +00:47:24.938 --> 00:47:26.070 +answer has to be the highest. + +00:47:26.070 --> 00:47:27.548 +So if the probability of the correct + +00:47:27.548 --> 00:47:30.714 +answer is say 26%, but the probability + +00:47:30.714 --> 00:47:33.220 +of all the other answers is like just + +00:47:33.220 --> 00:47:35.923 +under 25%, then you'll get the correct + +00:47:35.923 --> 00:47:36.226 +answer. + +00:47:36.226 --> 00:47:38.578 +So even though almost three out of four + +00:47:38.578 --> 00:47:41.013 +of the audience members can be wrong, + +00:47:41.013 --> 00:47:41.569 +it's. + +00:47:41.570 --> 00:47:43.378 +I mean, it's possible for three out of + +00:47:43.378 --> 00:47:45.038 +four of the audience members to be + +00:47:45.038 --> 00:47:46.698 +wrong almost, but still get the correct + +00:47:46.698 --> 00:47:48.140 +answer, still be guaranteed they're + +00:47:48.140 --> 00:47:48.760 +correct answer. + +00:47:50.250 --> 00:47:52.385 +If you were to pull the infinite size + +00:47:52.385 --> 00:47:53.940 +audience, of course with the limited + +00:47:53.940 --> 00:47:55.930 +audience you also have then variance, + +00:47:55.930 --> 00:47:57.800 +so you would want a bigger margin to be + +00:47:57.800 --> 00:47:58.190 +confident. + +00:47:59.100 --> 00:48:01.480 +And if a friend is a random member of + +00:48:01.480 --> 00:48:02.660 +the audience, this is an easier + +00:48:02.660 --> 00:48:03.270 +question. + +00:48:03.270 --> 00:48:05.190 +Then what's the probability that your + +00:48:05.190 --> 00:48:06.290 +friend's answer is correct? + +00:48:09.150 --> 00:48:09.440 +Right. + +00:48:10.320 --> 00:48:11.852 +Yeah, P of A, yeah. + +00:48:11.852 --> 00:48:13.830 +So in this setting, so it's possible + +00:48:13.830 --> 00:48:15.898 +that your friend could only have a 25% + +00:48:15.898 --> 00:48:17.650 +chance of being correct, but the + +00:48:17.650 --> 00:48:19.595 +audience has a 100% chance of being + +00:48:19.595 --> 00:48:19.859 +correct. + +00:48:24.800 --> 00:48:26.830 +Alright, so I'm going to give a + +00:48:26.830 --> 00:48:29.010 +detailed example of random forests. + +00:48:29.010 --> 00:48:30.950 +If you took computational photography + +00:48:30.950 --> 00:48:32.850 +with me, then you saw this example, but + +00:48:32.850 --> 00:48:34.100 +now you will see it in a new light. + +00:48:34.950 --> 00:48:37.960 +And so this is using this is the Kinect + +00:48:37.960 --> 00:48:38.490 +algorithm. + +00:48:38.490 --> 00:48:40.220 +So you guys might remember the Kinect + +00:48:40.220 --> 00:48:42.740 +came out in around 2011. + +00:48:43.720 --> 00:48:46.080 +For gaming and then was like widely + +00:48:46.080 --> 00:48:47.590 +adopted by the robotics community + +00:48:47.590 --> 00:48:48.270 +question. + +00:48:56.480 --> 00:48:59.850 +Alright, the answer is probability of a + +00:48:59.850 --> 00:49:04.080 +can be just marginally above 25% and + +00:49:04.080 --> 00:49:06.360 +the other probabilities are marginally + +00:49:06.360 --> 00:49:07.440 +below 25%. + +00:49:09.310 --> 00:49:09.720 +Yeah. + +00:49:11.560 --> 00:49:15.050 +All right, so the Kinect came out, you + +00:49:15.050 --> 00:49:17.280 +could play lots of games with it and it + +00:49:17.280 --> 00:49:18.570 +was also used for robotics. + +00:49:18.570 --> 00:49:20.864 +But for the games anyway, one of the + +00:49:20.864 --> 00:49:22.950 +one of the key things they had to solve + +00:49:22.950 --> 00:49:23.943 +was to. + +00:49:23.943 --> 00:49:26.635 +So first the Kinect has it does some + +00:49:26.635 --> 00:49:28.120 +like structured light thing in order to + +00:49:28.120 --> 00:49:28.990 +get a depth image. + +00:49:29.660 --> 00:49:30.550 +And then? + +00:49:30.720 --> 00:49:31.330 +And. + +00:49:32.070 --> 00:49:34.040 +And then the Kinect needs to estimate + +00:49:34.040 --> 00:49:37.000 +body purpose given the depth image, so + +00:49:37.000 --> 00:49:38.940 +that it can tell if you're like dancing + +00:49:38.940 --> 00:49:40.810 +correctly or doing the sport or + +00:49:40.810 --> 00:49:44.000 +whatever corresponds to the game. + +00:49:45.020 --> 00:49:47.260 +So given this depth image, you have to + +00:49:47.260 --> 00:49:50.580 +try to predict for like what are the + +00:49:50.580 --> 00:49:52.300 +key points of the body pose. + +00:49:52.300 --> 00:49:53.050 +That's the problem. + +00:49:54.850 --> 00:49:56.840 +And they need to do it really fast too, + +00:49:56.840 --> 00:49:59.230 +because they're because they only get a + +00:49:59.230 --> 00:50:02.064 +small fraction of the GPU of the Xbox + +00:50:02.064 --> 00:50:05.222 +to do this, 2% of the GPU of the Xbox + +00:50:05.222 --> 00:50:06.740 +to do this in real time. + +00:50:09.190 --> 00:50:12.370 +So the basic algorithm is from. + +00:50:12.370 --> 00:50:15.450 +This is described in this paper by + +00:50:15.450 --> 00:50:16.640 +Microsoft Cambridge. + +00:50:17.400 --> 00:50:21.430 +And the overall the processes, you go + +00:50:21.430 --> 00:50:23.180 +from a depth image and segment it. + +00:50:23.180 --> 00:50:25.950 +Then you predict for each pixel which + +00:50:25.950 --> 00:50:28.200 +of the body parts corresponds to that + +00:50:28.200 --> 00:50:29.200 +pixel. + +00:50:29.200 --> 00:50:30.410 +Is it like the right side of the face + +00:50:30.410 --> 00:50:31.380 +or left side of the face? + +00:50:32.180 --> 00:50:34.540 +And then you take those predictions and + +00:50:34.540 --> 00:50:36.210 +combine them to get a key point + +00:50:36.210 --> 00:50:36.730 +estimate. + +00:50:38.490 --> 00:50:39.730 +So here's another view of it. + +00:50:40.400 --> 00:50:42.905 +Given RGB image, that's Jamie shot in + +00:50:42.905 --> 00:50:45.846 +the first author you then and a depth + +00:50:45.846 --> 00:50:46.223 +image. + +00:50:46.223 --> 00:50:48.120 +You don't use the RGB actually, you + +00:50:48.120 --> 00:50:49.983 +just segment out the body from the + +00:50:49.983 --> 00:50:50.199 +depth. + +00:50:50.200 --> 00:50:51.900 +It's like the near pixels. + +00:50:52.670 --> 00:50:55.185 +And then you label them into parts and + +00:50:55.185 --> 00:50:57.790 +then you assign the joint positions. + +00:51:00.690 --> 00:51:03.489 +So the reason this is kind of this is + +00:51:03.490 --> 00:51:05.050 +pretty hard because you're going to + +00:51:05.050 --> 00:51:06.470 +have a lot of different bodies and + +00:51:06.470 --> 00:51:08.370 +orientations and poses and wearing + +00:51:08.370 --> 00:51:10.500 +different kinds of clothes, and you + +00:51:10.500 --> 00:51:12.490 +want this to work for everybody because + +00:51:12.490 --> 00:51:14.400 +if it fails, then the games not any + +00:51:14.400 --> 00:51:14.710 +fun. + +00:51:15.740 --> 00:51:19.610 +And So what they did is they collected + +00:51:19.610 --> 00:51:22.995 +a lot of examples of motion capture + +00:51:22.995 --> 00:51:24.990 +they had like different people do like + +00:51:24.990 --> 00:51:26.970 +motion capture and got like real + +00:51:26.970 --> 00:51:30.190 +examples and then they took those body + +00:51:30.190 --> 00:51:33.270 +frames and rigged a synthetic models. + +00:51:33.940 --> 00:51:35.700 +And generated even more synthetic + +00:51:35.700 --> 00:51:37.550 +examples of people in the same poses. + +00:51:38.150 --> 00:51:40.020 +And on these synthetic examples, it was + +00:51:40.020 --> 00:51:41.945 +easy to label the parts because they're + +00:51:41.945 --> 00:51:42.450 +synthetic. + +00:51:42.450 --> 00:51:44.080 +So they could just like essentially + +00:51:44.080 --> 00:51:46.740 +texture the parts and then they would + +00:51:46.740 --> 00:51:48.880 +know like which pixel corresponds to + +00:51:48.880 --> 00:51:49.410 +each label. + +00:51:51.640 --> 00:51:53.930 +So the same this is showing that the + +00:51:53.930 --> 00:51:58.010 +same body part this wrist or hand here. + +00:51:58.740 --> 00:52:00.300 +Can look quite different. + +00:52:00.300 --> 00:52:02.050 +It's the same part in all of these + +00:52:02.050 --> 00:52:04.200 +images, but depending on where it is + +00:52:04.200 --> 00:52:05.700 +and how the body is posed, then the + +00:52:05.700 --> 00:52:06.820 +image looks pretty different. + +00:52:06.820 --> 00:52:09.060 +So this is a pretty challenging problem + +00:52:09.060 --> 00:52:11.590 +to know that this pixel in the center + +00:52:11.590 --> 00:52:14.520 +of the cross is the wrist. + +00:52:15.390 --> 00:52:16.090 +Where the hand? + +00:52:19.180 --> 00:52:21.070 +All right, so the thresholding of the + +00:52:21.070 --> 00:52:24.640 +depth is relatively straightforward. + +00:52:24.640 --> 00:52:27.190 +And then they need to learn to predict + +00:52:27.190 --> 00:52:30.599 +for each pixel whether which of the + +00:52:30.600 --> 00:52:32.700 +possible body parts that pixel + +00:52:32.700 --> 00:52:33.510 +corresponds to. + +00:52:34.910 --> 00:52:37.015 +And these really simple features, the + +00:52:37.015 --> 00:52:41.500 +features are either a an offset feature + +00:52:41.500 --> 00:52:43.270 +where if you're trying to predict for + +00:52:43.270 --> 00:52:46.610 +this pixel at the center, here you + +00:52:46.610 --> 00:52:49.570 +shift some number of pixels that are + +00:52:49.570 --> 00:52:51.650 +dependent, so some pixels times depth. + +00:52:52.360 --> 00:52:54.100 +In some direction, and you look at the + +00:52:54.100 --> 00:52:55.740 +depth of that corresponding pixel, + +00:52:55.740 --> 00:52:58.230 +which could be like a particular value + +00:52:58.230 --> 00:52:59.660 +to indicate that it's off the body. + +00:53:01.290 --> 00:53:03.020 +So if you're at this pixel and you use + +00:53:03.020 --> 00:53:05.205 +this feature Theta one, then you end up + +00:53:05.205 --> 00:53:05.667 +over here. + +00:53:05.667 --> 00:53:07.144 +If you're looking at this pixel then + +00:53:07.144 --> 00:53:08.770 +you end up on the head over here in + +00:53:08.770 --> 00:53:09.450 +this example. + +00:53:10.350 --> 00:53:12.440 +And then you have other features that + +00:53:12.440 --> 00:53:14.210 +are based on the difference of depths. + +00:53:14.210 --> 00:53:16.870 +So given some position, you look at 2 + +00:53:16.870 --> 00:53:19.000 +offsets and take the difference of + +00:53:19.000 --> 00:53:19.600 +those depths. + +00:53:21.300 --> 00:53:23.260 +And then you can generate like + +00:53:23.260 --> 00:53:25.020 +basically infinite numbers of those + +00:53:25.020 --> 00:53:26.010 +features, right? + +00:53:26.010 --> 00:53:27.895 +There's like a lot of combinations of + +00:53:27.895 --> 00:53:29.655 +features using different offsets that + +00:53:29.655 --> 00:53:30.485 +you could create. + +00:53:30.485 --> 00:53:32.510 +And they also have lots of data, which + +00:53:32.510 --> 00:53:34.500 +as I mentioned came from mocap and then + +00:53:34.500 --> 00:53:35.260 +synthetic data. + +00:53:36.390 --> 00:53:39.060 +And so they train, they train random + +00:53:39.060 --> 00:53:42.990 +forests based on these features on all + +00:53:42.990 --> 00:53:43.640 +this data. + +00:53:43.640 --> 00:53:45.030 +So again, they have millions of + +00:53:45.030 --> 00:53:45.900 +examples. + +00:53:45.900 --> 00:53:47.995 +They can like practically infinite + +00:53:47.995 --> 00:53:49.680 +features, but you'd sample some number + +00:53:49.680 --> 00:53:50.930 +of features and tree in a tree. + +00:53:53.210 --> 00:53:54.500 +I think I just explained that. + +00:53:56.320 --> 00:53:58.270 +Sorry, I got a little ahead of myself, + +00:53:58.270 --> 00:54:00.264 +but this is just an illustration of + +00:54:00.264 --> 00:54:03.808 +their training data, 500,000 frames and + +00:54:03.808 --> 00:54:07.414 +then they got 3D models for 15 bodies + +00:54:07.414 --> 00:54:09.990 +and then they synthesized all the + +00:54:09.990 --> 00:54:11.860 +motion capture data on all of those + +00:54:11.860 --> 00:54:14.160 +bodies to get their training and test + +00:54:14.160 --> 00:54:15.319 +in synthetic test data. + +00:54:16.200 --> 00:54:17.730 +So this is showing similar synthetic + +00:54:17.730 --> 00:54:18.110 +data. + +00:54:21.210 --> 00:54:24.110 +And then so they so they're classifier + +00:54:24.110 --> 00:54:26.500 +is a random forest, so again they just. + +00:54:26.570 --> 00:54:27.060 + + +00:54:27.830 --> 00:54:31.095 +Randomly sample a set of those possible + +00:54:31.095 --> 00:54:33.030 +features, or generate a set of features + +00:54:33.030 --> 00:54:35.700 +and randomly subsample their training + +00:54:35.700 --> 00:54:36.030 +data. + +00:54:36.900 --> 00:54:39.315 +And then train a tree to completion and + +00:54:39.315 --> 00:54:41.810 +then each tree or maybe to maximum + +00:54:41.810 --> 00:54:42.100 +depth. + +00:54:42.100 --> 00:54:43.575 +In this case you might not change the + +00:54:43.575 --> 00:54:44.820 +completion since you may have like + +00:54:44.820 --> 00:54:45.680 +millions of samples. + +00:54:46.770 --> 00:54:48.660 +But you trained to some depth and then + +00:54:48.660 --> 00:54:50.570 +each node will have some probability + +00:54:50.570 --> 00:54:52.160 +estimate for each of the classes. + +00:54:52.970 --> 00:54:54.626 +And then you generate a new tree and + +00:54:54.626 --> 00:54:56.400 +you keep on doing that independently. + +00:54:57.510 --> 00:54:59.100 +And then you at the end you're + +00:54:59.100 --> 00:55:01.282 +predictor is an average of the + +00:55:01.282 --> 00:55:03.230 +probabilities, the class probabilities + +00:55:03.230 --> 00:55:04.530 +that each of the trees predicts. + +00:55:05.970 --> 00:55:09.780 +So it may sound like at first glance + +00:55:09.780 --> 00:55:11.030 +when you look at this you might think, + +00:55:11.030 --> 00:55:13.530 +well this seems really slow you then in + +00:55:13.530 --> 00:55:14.880 +order to. + +00:55:15.410 --> 00:55:16.040 +Make a prediction. + +00:55:16.040 --> 00:55:17.936 +You have to query all of these trees + +00:55:17.936 --> 00:55:19.760 +and then sum up their responses. + +00:55:19.760 --> 00:55:21.940 +But when you're implementing an GPU, + +00:55:21.940 --> 00:55:23.658 +it's actually really fast because these + +00:55:23.658 --> 00:55:24.840 +can all be done in parallel. + +00:55:24.840 --> 00:55:26.334 +The trees don't depend on each other, + +00:55:26.334 --> 00:55:29.161 +so you can do the inference on all the + +00:55:29.161 --> 00:55:31.045 +trees simultaneously, and you can do + +00:55:31.045 --> 00:55:32.120 +inference for all the pixels + +00:55:32.120 --> 00:55:33.600 +simultaneously if you have enough + +00:55:33.600 --> 00:55:33.968 +memory. + +00:55:33.968 --> 00:55:36.919 +And so it's actually can be done in + +00:55:36.920 --> 00:55:38.225 +remarkably fast. + +00:55:38.225 --> 00:55:41.300 +So they can do this in real time using + +00:55:41.300 --> 00:55:43.506 +2% of the computational resources of + +00:55:43.506 --> 00:55:44.280 +the Xbox. + +00:55:48.160 --> 00:55:48.770 + + +00:55:49.810 --> 00:55:53.730 +And then finally they would get the, so + +00:55:53.730 --> 00:55:54.700 +I'll show it here. + +00:55:54.700 --> 00:55:56.249 +So first they are like labeling the + +00:55:56.250 --> 00:55:57.465 +pixels like this. + +00:55:57.465 --> 00:56:01.607 +So this is the, sorry, over here the + +00:56:01.607 --> 00:56:03.690 +Pixel labels can be like a little bit + +00:56:03.690 --> 00:56:05.410 +of noise, a little bit noisy, but at + +00:56:05.410 --> 00:56:07.170 +the end they don't need a pixel perfect + +00:56:07.170 --> 00:56:09.430 +segmentation or pixel perfect labeling. + +00:56:10.060 --> 00:56:11.990 +What they really care about is the + +00:56:11.990 --> 00:56:13.950 +position of the joints, the 3D position + +00:56:13.950 --> 00:56:14.790 +of the joints. + +00:56:15.710 --> 00:56:17.899 +And so based on the depth and based on + +00:56:17.900 --> 00:56:19.416 +which pixels are labeled with each + +00:56:19.416 --> 00:56:22.290 +joint, they can get the average 3D + +00:56:22.290 --> 00:56:24.420 +position of these labels. + +00:56:24.420 --> 00:56:27.280 +And then they just put it like slightly + +00:56:27.280 --> 00:56:29.070 +behind that in a joint dependent way. + +00:56:29.070 --> 00:56:31.429 +So like if that the average depth of + +00:56:31.429 --> 00:56:33.346 +these pixels on my shoulder, then that + +00:56:33.346 --> 00:56:34.860 +the center of my shoulder is going to + +00:56:34.860 --> 00:56:36.950 +be an inch and 1/2 behind that or + +00:56:36.950 --> 00:56:37.619 +something like that. + +00:56:38.450 --> 00:56:40.600 +So then you get the 3D position of my + +00:56:40.600 --> 00:56:41.030 +shoulder. + +00:56:42.480 --> 00:56:44.303 +And so even though they're pixel + +00:56:44.303 --> 00:56:46.280 +predictions might be a little noisy, + +00:56:46.280 --> 00:56:48.130 +the joint predictions are more accurate + +00:56:48.130 --> 00:56:49.550 +because they're based on a combination + +00:56:49.550 --> 00:56:50.499 +of pixel predictions. + +00:56:54.090 --> 00:56:55.595 +So here is showing the ground truth. + +00:56:55.595 --> 00:56:57.360 +This is the depth image, this is a + +00:56:57.360 --> 00:57:00.160 +pixel labels and then this is the joint + +00:57:00.160 --> 00:57:00.780 +labels. + +00:57:01.450 --> 00:57:03.850 +And then and. + +00:57:03.850 --> 00:57:06.005 +This is showing the actual predictions + +00:57:06.005 --> 00:57:07.210 +and some examples. + +00:57:09.420 --> 00:57:11.020 +And here you can see the same thing. + +00:57:11.020 --> 00:57:13.630 +So these are the input depth images. + +00:57:14.400 --> 00:57:16.480 +This is the pixel predictions on those + +00:57:16.480 --> 00:57:17.210 +depth images. + +00:57:17.860 --> 00:57:19.870 +And then this is showing the estimated + +00:57:19.870 --> 00:57:22.385 +pose from different perspectives so + +00:57:22.385 --> 00:57:24.910 +that you can see it looks kind of + +00:57:24.910 --> 00:57:25.100 +right. + +00:57:25.100 --> 00:57:26.780 +So like in this case for example, it's + +00:57:26.780 --> 00:57:28.570 +estimating that the person is standing + +00:57:28.570 --> 00:57:30.840 +with his hands like out and slightly in + +00:57:30.840 --> 00:57:31.110 +front. + +00:57:36.130 --> 00:57:38.440 +And you can see if you vary the number + +00:57:38.440 --> 00:57:41.810 +of training samples, you get like + +00:57:41.810 --> 00:57:42.670 +pretty good. + +00:57:42.670 --> 00:57:45.860 +I mean essentially what I would say is + +00:57:45.860 --> 00:57:47.239 +that you need a lot of training samples + +00:57:47.240 --> 00:57:48.980 +to do well in this task. + +00:57:49.660 --> 00:57:52.330 +So as you start to get up to 100,000 or + +00:57:52.330 --> 00:57:53.640 +a million training samples. + +00:57:54.300 --> 00:57:58.360 +Your average accuracy gets up to 60%. + +00:57:59.990 --> 00:58:02.350 +And 60% might not sound that good, but + +00:58:02.350 --> 00:58:04.339 +it's actually fine because a lot of the + +00:58:04.340 --> 00:58:05.930 +errors will just be on the margin where + +00:58:05.930 --> 00:58:08.050 +you're like whether this pixel is the + +00:58:08.050 --> 00:58:09.500 +upper arm or the shoulder. + +00:58:09.500 --> 00:58:13.110 +And so the per pixel accuracy of 60% + +00:58:13.110 --> 00:58:14.420 +gives you pretty accurate joint + +00:58:14.420 --> 00:58:15.030 +positions. + +00:58:16.680 --> 00:58:18.460 +One of the surprising things about the + +00:58:18.460 --> 00:58:21.979 +paper was that the synthetic data was + +00:58:21.980 --> 00:58:24.000 +so effective because in all past + +00:58:24.000 --> 00:58:26.322 +research, pretty much when people use + +00:58:26.322 --> 00:58:27.720 +synthetic data it didn't like + +00:58:27.720 --> 00:58:29.700 +generalize that did the test data. + +00:58:29.700 --> 00:58:30.940 +And I think the reason that it + +00:58:30.940 --> 00:58:32.580 +generalizes well in this case is that + +00:58:32.580 --> 00:58:34.830 +depth data is a lot easier to simulate + +00:58:34.830 --> 00:58:35.290 +than. + +00:58:35.930 --> 00:58:37.170 +RGB data. + +00:58:37.170 --> 00:58:39.810 +So now people have used RGB data + +00:58:39.810 --> 00:58:40.340 +somewhat. + +00:58:40.340 --> 00:58:43.440 +It's often used in autonomous vehicle + +00:58:43.440 --> 00:58:46.760 +training, but at the time it had not + +00:58:46.760 --> 00:58:47.920 +really been used effectively. + +00:58:58.700 --> 00:58:58.980 +OK. + +00:59:00.020 --> 00:59:01.500 +Is there any questions about that? + +00:59:04.850 --> 00:59:06.820 +And then the last big thing I want to + +00:59:06.820 --> 00:59:08.140 +do you're probably not. + +00:59:08.500 --> 00:59:11.210 +Emotionally ready for homework 2 yet, + +00:59:11.210 --> 00:59:12.740 +but I'll give it to you anyway. + +00:59:14.930 --> 00:59:16.510 +Is to show you homework too. + +00:59:25.020 --> 00:59:27.760 +Alright, so at least in some parts of + +00:59:27.760 --> 00:59:30.070 +this are going to be a bit familiar. + +00:59:32.020 --> 00:59:32.640 +Yeah. + +00:59:32.640 --> 00:59:33.140 +Thank you. + +00:59:34.070 --> 00:59:34.750 +I always forget. + +00:59:35.730 --> 00:59:37.640 +With that, let me get rid of that. + +00:59:38.500 --> 00:59:39.000 +OK. + +00:59:42.850 --> 00:59:43.480 +Damn it. + +00:59:51.800 --> 00:59:55.390 +Alright, let's see me in a bit. + +00:59:56.330 --> 00:59:56.840 +OK. + +00:59:57.980 --> 00:59:59.290 +All right, so there's three parts of + +00:59:59.290 --> 00:59:59.900 +this. + +00:59:59.900 --> 01:00:04.780 +The first part is looking at the + +01:00:04.780 --> 01:00:06.920 +effects of model complexity with tree + +01:00:06.920 --> 01:00:07.610 +regressors. + +01:00:08.870 --> 01:00:12.560 +So you train trees with different + +01:00:12.560 --> 01:00:13.190 +depths. + +01:00:13.800 --> 01:00:17.380 +And Oregon, random forests with + +01:00:17.380 --> 01:00:18.090 +different depths. + +01:00:19.120 --> 01:00:22.745 +And then you plot the error versus the + +01:00:22.745 --> 01:00:24.150 +versus the size. + +01:00:25.280 --> 01:00:26.440 +So it's actually. + +01:00:26.440 --> 01:00:27.350 +This is actually. + +01:00:29.290 --> 01:00:29.980 +Pretty easy. + +01:00:29.980 --> 01:00:31.720 +Code wise, it's, I'll show you. + +01:00:31.720 --> 01:00:34.240 +It's just to get to just see for + +01:00:34.240 --> 01:00:35.890 +yourself like the effects of depth. + +01:00:37.260 --> 01:00:38.830 +So in this case you don't need to + +01:00:38.830 --> 01:00:40.590 +implement the trees or the random + +01:00:40.590 --> 01:00:41.920 +forests, you can use the library. + +01:00:42.740 --> 01:00:43.940 +So, and we're going to use the + +01:00:43.940 --> 01:00:44.640 +temperature data. + +01:00:46.350 --> 01:00:48.910 +Essentially you would iterate over + +01:00:48.910 --> 01:00:51.360 +these Max depths which range from 2 to + +01:00:51.360 --> 01:00:52.020 +32. + +01:00:52.970 --> 01:00:54.890 +And then for each depth you would call + +01:00:54.890 --> 01:00:58.790 +these functions and get the error and + +01:00:58.790 --> 01:01:00.300 +then you can. + +01:01:01.500 --> 01:01:04.570 +And then you can call this code to plot + +01:01:04.570 --> 01:01:05.030 +the error. + +01:01:05.670 --> 01:01:07.610 +And then you'll look at that plot, and + +01:01:07.610 --> 01:01:08.440 +then you'll. + +01:01:09.250 --> 01:01:11.580 +Provide the plot and answer some + +01:01:11.580 --> 01:01:12.120 +questions. + +01:01:12.720 --> 01:01:16.180 +So in the report there's some questions + +01:01:16.180 --> 01:01:18.090 +for you to answer based on your + +01:01:18.090 --> 01:01:18.820 +analysis. + +01:01:20.350 --> 01:01:21.846 +They're like, given a maximum depth + +01:01:21.846 --> 01:01:26.130 +tree, which model has the lowest bias + +01:01:26.130 --> 01:01:28.089 +for regression trees, what tree depth + +01:01:28.090 --> 01:01:29.900 +achieves the minimum validation error? + +01:01:31.080 --> 01:01:33.440 +When is which model is least prone to + +01:01:33.440 --> 01:01:34.810 +overfitting, for example? + +01:01:37.480 --> 01:01:38.970 +So that's the first problem. + +01:01:40.030 --> 01:01:41.530 +The second problem, this is the one + +01:01:41.530 --> 01:01:43.485 +that's going to take you the most time, + +01:01:43.485 --> 01:01:46.950 +is using MLPS, so multilayer + +01:01:46.950 --> 01:01:48.390 +perceptrons with MNIST. + +01:01:49.590 --> 01:01:52.770 +It takes about 3 minutes to train it, + +01:01:52.770 --> 01:01:54.420 +so it's not too bad compared to your + +01:01:54.420 --> 01:01:55.360 +nearest neighbor training. + +01:01:56.310 --> 01:01:56.840 +And. + +01:01:57.680 --> 01:02:01.610 +And you need you need to basically + +01:02:01.610 --> 01:02:02.680 +like. + +01:02:02.680 --> 01:02:05.225 +We're going to use Pytorch, which is + +01:02:05.225 --> 01:02:06.800 +like a really good package for deep + +01:02:06.800 --> 01:02:07.160 +learning. + +01:02:08.180 --> 01:02:09.990 +And you need to. + +01:02:11.750 --> 01:02:15.500 +Fill out the forward and. + +01:02:16.850 --> 01:02:20.370 +And the like model specification. + +01:02:20.370 --> 01:02:23.650 +So I provide in the chips a link to a + +01:02:23.650 --> 01:02:25.500 +tutorial and you can also look up other + +01:02:25.500 --> 01:02:28.320 +tutorials that explain in the tips. + +01:02:28.320 --> 01:02:30.510 +Also gives you kind of the basic code + +01:02:30.510 --> 01:02:30.930 +structure. + +01:02:31.640 --> 01:02:33.850 +But you can see like how these things + +01:02:33.850 --> 01:02:36.030 +are coded, essentially that you define + +01:02:36.030 --> 01:02:37.280 +the layers of the network here. + +01:02:37.870 --> 01:02:40.560 +And then you define like how the data + +01:02:40.560 --> 01:02:42.030 +progresses through the network to make + +01:02:42.030 --> 01:02:45.429 +a prediction and then you and then you + +01:02:45.430 --> 01:02:46.430 +can train your network. + +01:02:48.040 --> 01:02:49.410 +Obviously we haven't talked about this + +01:02:49.410 --> 01:02:50.900 +yet, so it might not make complete + +01:02:50.900 --> 01:02:52.200 +sense yet, but it will. + +01:02:53.760 --> 01:02:55.048 +So then you're going to train a + +01:02:55.048 --> 01:02:57.019 +network, then you're going to try + +01:02:57.020 --> 01:02:58.638 +different learning rates, and then + +01:02:58.638 --> 01:03:00.230 +you're going to try to get the best + +01:03:00.230 --> 01:03:03.340 +network you can with the target of 25% + +01:03:03.340 --> 01:03:04.000 +validation error. + +01:03:05.770 --> 01:03:07.150 +And then a third problem. + +01:03:07.150 --> 01:03:09.450 +We're looking at this new data set + +01:03:09.450 --> 01:03:11.820 +called the Penguin data set, the Palmer + +01:03:11.820 --> 01:03:13.500 +Archipelago Penguin data set. + +01:03:14.410 --> 01:03:16.800 +And this is a data set of like some + +01:03:16.800 --> 01:03:18.500 +various physical measurements of the + +01:03:18.500 --> 01:03:19.970 +Penguins, whether they're male or + +01:03:19.970 --> 01:03:21.813 +female, what island they came from, and + +01:03:21.813 --> 01:03:23.140 +what kind of species it is. + +01:03:23.990 --> 01:03:25.800 +So we created a clean version of the + +01:03:25.800 --> 01:03:28.510 +data here and. + +01:03:29.670 --> 01:03:31.500 +And then we have like some starter code + +01:03:31.500 --> 01:03:32.380 +to load that data. + +01:03:33.210 --> 01:03:35.370 +And you're going to 1st like visualize + +01:03:35.370 --> 01:03:36.470 +some of the features. + +01:03:36.470 --> 01:03:40.270 +So we did one example for you if you + +01:03:40.270 --> 01:03:41.970 +look at the different species of + +01:03:41.970 --> 01:03:42.740 +Penguins. + +01:03:44.890 --> 01:03:46.880 +This is like a scatter plot of body + +01:03:46.880 --> 01:03:48.900 +mass versus flipper length for some + +01:03:48.900 --> 01:03:49.980 +different Penguins. + +01:03:49.980 --> 01:03:51.950 +So you can see that this would be like + +01:03:51.950 --> 01:03:53.880 +pretty good at distinguishing Gentoo + +01:03:53.880 --> 01:03:57.230 +from a deli and chinstrap, but not so + +01:03:57.230 --> 01:03:59.030 +good at distinguishing chinstrap in a + +01:03:59.030 --> 01:03:59.280 +deli. + +01:03:59.280 --> 01:04:00.790 +So you can do this for different + +01:04:00.790 --> 01:04:01.792 +combinations of features. + +01:04:01.792 --> 01:04:03.120 +There's not a lot of features. + +01:04:03.120 --> 01:04:03.989 +I think there's 13. + +01:04:06.080 --> 01:04:07.020 +And then? + +01:04:07.100 --> 01:04:07.730 + + +01:04:08.440 --> 01:04:10.140 +And then in the report it asks like + +01:04:10.140 --> 01:04:12.410 +some kinds of like analysis questions + +01:04:12.410 --> 01:04:14.060 +based on that feature analysis. + +01:04:15.490 --> 01:04:17.410 +Then the second question is to come up + +01:04:17.410 --> 01:04:19.889 +with a simple, really simple rule A2 + +01:04:19.890 --> 01:04:21.330 +part rule that will allow you to + +01:04:21.330 --> 01:04:22.980 +perfectly classify Gentius. + +01:04:24.330 --> 01:04:27.170 +And then the third part is to design an + +01:04:27.170 --> 01:04:29.385 +mill model to maximize your accuracy on + +01:04:29.385 --> 01:04:30.160 +this problem. + +01:04:30.160 --> 01:04:33.070 +And you can use you can use like the + +01:04:33.070 --> 01:04:35.280 +library to do cross validation. + +01:04:35.280 --> 01:04:37.610 +So essentially you can use the + +01:04:37.610 --> 01:04:39.190 +libraries for your models as well. + +01:04:39.190 --> 01:04:40.390 +So you just need to choose the + +01:04:40.390 --> 01:04:42.100 +parameters of your models and then try + +01:04:42.100 --> 01:04:43.569 +to get the best performance you can. + +01:04:47.330 --> 01:04:49.180 +Then the stretch goals are to improve + +01:04:49.180 --> 01:04:52.020 +the MNIST using MLPS to find a second + +01:04:52.020 --> 01:04:54.330 +rule for classifying Gentius. + +01:04:55.050 --> 01:04:57.660 +And then this one is positional + +01:04:57.660 --> 01:05:00.765 +encoding, which is a way of like + +01:05:00.765 --> 01:05:03.130 +encoding positions that lets networks + +01:05:03.130 --> 01:05:05.170 +work better on it, but I won't go into + +01:05:05.170 --> 01:05:06.490 +details there since we haven't talked + +01:05:06.490 --> 01:05:07.070 +about networks. + +01:05:09.040 --> 01:05:11.270 +Any questions about homework 2? + +01:05:14.740 --> 01:05:16.100 +There will be, yes. + +01:05:17.910 --> 01:05:18.190 +OK. + +01:05:29.410 --> 01:05:29.700 +No. + +01:05:29.700 --> 01:05:31.484 +It says in that you don't need to + +01:05:31.484 --> 01:05:31.893 +answer them. + +01:05:31.893 --> 01:05:34.470 +You don't need to report on them. + +01:05:34.470 --> 01:05:36.450 +So you should answer them in your head + +01:05:36.450 --> 01:05:37.936 +and you'll learn more that way, but you + +01:05:37.936 --> 01:05:39.220 +don't need to provide the answer. + +01:05:40.190 --> 01:05:40.710 +Yeah. + +01:05:43.900 --> 01:05:44.230 +Why? + +01:05:47.670 --> 01:05:48.830 +Will not make a cost. + +01:05:51.690 --> 01:05:53.280 +No, it won't hurt you either. + +01:05:54.650 --> 01:05:54.910 +Yeah. + +01:05:55.930 --> 01:05:56.740 +You're not required. + +01:05:56.740 --> 01:05:58.397 +You're only required to fill out what's + +01:05:58.397 --> 01:05:59.172 +in the template. + +01:05:59.172 --> 01:06:01.880 +So sometimes I say to do like slightly + +01:06:01.880 --> 01:06:03.406 +more than what's in the template. + +01:06:03.406 --> 01:06:05.300 +The template is basically to show that + +01:06:05.300 --> 01:06:07.226 +you've done it, so sometimes you can + +01:06:07.226 --> 01:06:08.520 +show that you've done it without + +01:06:08.520 --> 01:06:09.840 +providing all the details. + +01:06:09.840 --> 01:06:10.220 +So. + +01:06:16.180 --> 01:06:17.810 +So the question is, can you resubmit + +01:06:17.810 --> 01:06:18.570 +the assignment? + +01:06:18.570 --> 01:06:20.363 +I wouldn't really recommend it. + +01:06:20.363 --> 01:06:21.176 +You would get. + +01:06:21.176 --> 01:06:23.570 +So the way that it works is that at the + +01:06:23.570 --> 01:06:25.653 +time that the T at the, it's mainly T + +01:06:25.653 --> 01:06:27.459 +is greeting, so at the time that the + +01:06:27.460 --> 01:06:28.250 +tea is green. + +01:06:29.270 --> 01:06:31.060 +Whatever is submitted last will be + +01:06:31.060 --> 01:06:31.480 +graded. + +01:06:32.390 --> 01:06:34.930 +And whatever, like with whatever late + +01:06:34.930 --> 01:06:36.950 +days have accrued for that, for that + +01:06:36.950 --> 01:06:37.360 +submission. + +01:06:37.360 --> 01:06:40.140 +If it's late so you can resubmit, but + +01:06:40.140 --> 01:06:41.590 +then once they've graded, then it's + +01:06:41.590 --> 01:06:43.270 +graded and then you can't resubmit + +01:06:43.270 --> 01:06:43.640 +anymore. + +01:06:46.300 --> 01:06:47.150 +There were. + +01:06:47.150 --> 01:06:48.910 +We basically assume that if it's past + +01:06:48.910 --> 01:06:50.530 +the deadline and you've submitted, then + +01:06:50.530 --> 01:06:54.580 +we can grade it and so it might get and + +01:06:54.580 --> 01:06:56.750 +generally if you want to get extra + +01:06:56.750 --> 01:06:57.170 +points. + +01:06:57.900 --> 01:06:59.330 +I would just recommend a move on to + +01:06:59.330 --> 01:07:01.053 +homework two and do extra points for + +01:07:01.053 --> 01:07:02.430 +homework two rather than getting stuck + +01:07:02.430 --> 01:07:03.925 +on homework one and getting late days + +01:07:03.925 --> 01:07:06.040 +and then like having trouble getting up + +01:07:06.040 --> 01:07:07.250 +getting homework 2 done. + +01:07:13.630 --> 01:07:16.730 +All right, so the things to remember + +01:07:16.730 --> 01:07:17.420 +from this class. + +01:07:18.180 --> 01:07:20.180 +Ensembles improve accuracy and + +01:07:20.180 --> 01:07:22.325 +confidence estimates by reducing the + +01:07:22.325 --> 01:07:23.990 +bias and Oregon the variance. + +01:07:23.990 --> 01:07:25.730 +And there's like this really important + +01:07:25.730 --> 01:07:28.100 +principle that test error can be + +01:07:28.100 --> 01:07:30.690 +decomposed into variance, bias and + +01:07:30.690 --> 01:07:31.670 +irreducible noise. + +01:07:32.680 --> 01:07:33.970 +And because the trees and random + +01:07:33.970 --> 01:07:35.870 +forests are really powerful and widely + +01:07:35.870 --> 01:07:38.000 +applicable classifiers and regressors. + +01:07:39.990 --> 01:07:43.440 +So in the next class I'm going to talk + +01:07:43.440 --> 01:07:45.765 +about SVM support vector machines, + +01:07:45.765 --> 01:07:48.910 +which were very popular approach, and + +01:07:48.910 --> 01:07:50.830 +stochastic gradient descent, which is a + +01:07:50.830 --> 01:07:52.310 +method to optimize them that also + +01:07:52.310 --> 01:07:54.245 +applies to neural Nets and deep Nets. + +01:07:54.245 --> 01:07:56.300 +So thank you, I'll see you on Thursday. + +01:19:53.620 --> 01:19:54.020 +Yeah. + +01:19:56.250 --> 01:19:56.660 +Testing. + +01:19:58.350 --> 01:19:58.590 +Yeah. +