diff --git "a/CS_441_2023_Spring_January_24,_2023.vtt" "b/CS_441_2023_Spring_January_24,_2023.vtt" new file mode 100644--- /dev/null +++ "b/CS_441_2023_Spring_January_24,_2023.vtt" @@ -0,0 +1,5342 @@ +WEBVTT Kind: captions; Language: en-US + +NOTE +Created on 2024-02-07T20:52:10.2470009Z by ClassTranscribe + +00:01:22.340 --> 00:01:22.750 +Good morning. + +00:01:24.260 --> 00:01:27.280 +Alright, so I'm going to just first + +00:01:27.280 --> 00:01:29.738 +finish up what I was, what I was going + +00:01:29.738 --> 00:01:31.660 +to cover at the end of the last lecture + +00:01:31.660 --> 00:01:32.980 +about Cannon. + +00:01:33.640 --> 00:01:36.550 +And then I'll talk about probabilities + +00:01:36.550 --> 00:01:37.540 +and Naive Bayes. + +00:01:38.260 --> 00:01:39.940 +And so I wanted to give an example of + +00:01:39.940 --> 00:01:41.930 +how K&N is used in practice. + +00:01:42.530 --> 00:01:44.880 +Here's one example of using it for face + +00:01:44.880 --> 00:01:45.920 +recognition. + +00:01:46.750 --> 00:01:48.480 +A lot of times when it's used in + +00:01:48.480 --> 00:01:50.030 +practice, there's a lot of feature + +00:01:50.030 --> 00:01:51.780 +learning that goes on ahead of the + +00:01:51.780 --> 00:01:52.588 +nearest neighbor. + +00:01:52.588 --> 00:01:54.510 +So nearest neighbor itself is really + +00:01:54.510 --> 00:01:55.125 +simple. + +00:01:55.125 --> 00:01:58.530 +It's efficacy depends on learning good + +00:01:58.530 --> 00:02:00.039 +representation so that. + +00:02:00.800 --> 00:02:02.640 +Data points that are near each other + +00:02:02.640 --> 00:02:04.410 +actually have similar labels. + +00:02:05.450 --> 00:02:07.385 +Here's one example. + +00:02:07.385 --> 00:02:10.550 +They want to try to be able to + +00:02:10.550 --> 00:02:12.330 +recognize whether two faces are the + +00:02:12.330 --> 00:02:13.070 +same person. + +00:02:13.820 --> 00:02:16.460 +And so the method is that you Detect + +00:02:16.460 --> 00:02:18.940 +facial features and then use those + +00:02:18.940 --> 00:02:21.630 +feature detections to align the image + +00:02:21.630 --> 00:02:23.300 +so that the face looks more frontal. + +00:02:24.060 --> 00:02:26.480 +Then they use a CNN convolutional + +00:02:26.480 --> 00:02:29.240 +neural network to train Features that + +00:02:29.240 --> 00:02:32.600 +will be good for recognizing faces. + +00:02:32.600 --> 00:02:34.360 +And the way they did that is that they + +00:02:34.360 --> 00:02:37.950 +first collected hundreds of Faces from + +00:02:37.950 --> 00:02:40.300 +a few thousand different people. + +00:02:40.300 --> 00:02:41.680 +I think it was their employees of + +00:02:41.680 --> 00:02:42.250 +Facebook. + +00:02:43.030 --> 00:02:46.420 +And they trained a classifier to say + +00:02:46.420 --> 00:02:48.970 +which, given a face, which of these + +00:02:48.970 --> 00:02:50.960 +people does the face belong to. + +00:02:52.030 --> 00:02:54.340 +And from that, they learn a + +00:02:54.340 --> 00:02:55.210 +REPRESENTATION. + +00:02:55.210 --> 00:02:57.030 +Those classifiers aren't very useful, + +00:02:57.030 --> 00:02:59.300 +because nobody's interested in seeing + +00:02:59.300 --> 00:03:00.230 +given a face. + +00:03:00.230 --> 00:03:01.843 +Which of the Facebook employees is that + +00:03:01.843 --> 00:03:02.914 +they want to know? + +00:03:02.914 --> 00:03:04.932 +Like, is it you want to know? + +00:03:04.932 --> 00:03:07.460 +Like, organize your photo album or see + +00:03:07.460 --> 00:03:08.800 +whether you've been tagged in another + +00:03:08.800 --> 00:03:09.960 +photo or something like that? + +00:03:10.630 --> 00:03:12.050 +And so then they throw out the + +00:03:12.050 --> 00:03:13.980 +Classifier and they just use the + +00:03:13.980 --> 00:03:16.280 +feature representation that was learned + +00:03:16.280 --> 00:03:21.070 +and use nearest neighbor to identify a + +00:03:21.070 --> 00:03:22.510 +person that's been detected in a + +00:03:22.510 --> 00:03:23.090 +photograph. + +00:03:24.830 --> 00:03:26.540 +So in their paper, they showed that + +00:03:26.540 --> 00:03:28.565 +this performs similarly to humans in + +00:03:28.565 --> 00:03:30.470 +this data set called label faces in the + +00:03:30.470 --> 00:03:31.970 +wild where you're trying to recognize + +00:03:31.970 --> 00:03:32.560 +celebrities. + +00:03:34.140 --> 00:03:35.770 +But it can be used for many things. + +00:03:35.770 --> 00:03:37.516 +So you can organize photo albums, you + +00:03:37.516 --> 00:03:40.360 +can detect faces and then you try to + +00:03:40.360 --> 00:03:41.970 +match Faces across the photos. + +00:03:41.970 --> 00:03:44.175 +So then you can organize like which + +00:03:44.175 --> 00:03:46.320 +photos have a particular person. + +00:03:47.070 --> 00:03:49.950 +Again, you can't identify celebrities + +00:03:49.950 --> 00:03:51.860 +or famous people by building up a + +00:03:51.860 --> 00:03:54.919 +database of faces of famous people. + +00:03:55.870 --> 00:03:58.110 +And you can also alert, alert somebody + +00:03:58.110 --> 00:04:00.100 +if somebody else uploads a photo of + +00:04:00.100 --> 00:04:00.330 +them. + +00:04:00.330 --> 00:04:02.922 +So you can see if somebody uploads a + +00:04:02.922 --> 00:04:05.364 +photo, then you can detect faces, you + +00:04:05.364 --> 00:04:07.830 +can see what their friends network is, + +00:04:07.830 --> 00:04:10.056 +see what other which of their faces + +00:04:10.056 --> 00:04:12.220 +have been uploaded and then Detect the + +00:04:12.220 --> 00:04:14.330 +other users whose faces have been + +00:04:14.330 --> 00:04:16.580 +uploaded and ask them for permission to + +00:04:16.580 --> 00:04:17.930 +like make this photo public. + +00:04:19.750 --> 00:04:22.020 +So this algorithm is actually used by + +00:04:22.020 --> 00:04:22.560 +Facebook. + +00:04:22.560 --> 00:04:24.340 +It has been for several years. + +00:04:24.340 --> 00:04:28.640 +They're limiting some of its use more + +00:04:28.640 --> 00:04:30.544 +recently, but they've been. + +00:04:30.544 --> 00:04:32.010 +But it's been used really heavily. + +00:04:32.680 --> 00:04:34.410 +And of course they have expanded + +00:04:34.410 --> 00:04:36.365 +training data because whenever anybody + +00:04:36.365 --> 00:04:37.940 +uploads photos then they can + +00:04:37.940 --> 00:04:40.353 +automatically detect them and add them + +00:04:40.353 --> 00:04:42.360 +to the database. + +00:04:42.360 --> 00:04:45.150 +So here the use of KN is important + +00:04:45.150 --> 00:04:47.220 +because KNN doesn't require any + +00:04:47.220 --> 00:04:47.490 +training. + +00:04:47.490 --> 00:04:49.295 +So every time somebody uploads a new + +00:04:49.295 --> 00:04:50.930 +face you can update the model just by + +00:04:50.930 --> 00:04:54.430 +adding this four 4096 dimensional + +00:04:54.430 --> 00:04:56.646 +feature vector that corresponds to the + +00:04:56.646 --> 00:05:00.230 +face and then use it in like based on + +00:05:00.230 --> 00:05:02.550 +the friend networks to. + +00:05:02.910 --> 00:05:04.840 +To recognize faces that are associated + +00:05:04.840 --> 00:05:05.410 +with somebody. + +00:05:07.530 --> 00:05:11.270 +I won't take time to discuss it now, + +00:05:11.270 --> 00:05:13.473 +but it's worth thinking about some of + +00:05:13.473 --> 00:05:15.710 +the consequences of the way that the + +00:05:15.710 --> 00:05:17.888 +algorithm was trained and the way that + +00:05:17.888 --> 00:05:18.620 +it's deployed. + +00:05:18.620 --> 00:05:19.600 +So for example. + +00:05:20.510 --> 00:05:22.680 +If you think about that, it was that + +00:05:22.680 --> 00:05:24.650 +the initial Features were learned on + +00:05:24.650 --> 00:05:26.030 +Facebook employees. + +00:05:26.030 --> 00:05:27.440 +That's not a very. + +00:05:28.070 --> 00:05:29.630 +That's not very representative + +00:05:29.630 --> 00:05:32.120 +demographic of the US the employees + +00:05:32.120 --> 00:05:35.000 +tend to be younger and. + +00:05:35.490 --> 00:05:38.446 +Probably skew towards male might skew + +00:05:38.446 --> 00:05:40.210 +towards certain ethnicities. + +00:05:40.820 --> 00:05:43.210 +And so the Algorithm may be much better + +00:05:43.210 --> 00:05:45.030 +at recognizing some kinds of Faces than + +00:05:45.030 --> 00:05:46.016 +other faces. + +00:05:46.016 --> 00:05:47.628 +And then, of course, there's lots and + +00:05:47.628 --> 00:05:49.495 +lots of ethical issues that surround + +00:05:49.495 --> 00:05:51.830 +the use of face recognition and its + +00:05:51.830 --> 00:05:52.610 +applications. + +00:05:53.930 --> 00:05:55.550 +Of course, like in many ways, this is + +00:05:55.550 --> 00:05:58.150 +used to help people maintain privacy. + +00:05:58.150 --> 00:06:00.080 +But even the use of recognition at all + +00:06:00.080 --> 00:06:03.120 +raises privacy concerns, and that's why + +00:06:03.120 --> 00:06:04.860 +they've limited the use to some extent. + +00:06:06.470 --> 00:06:08.060 +So just something to think about. + +00:06:09.980 --> 00:06:13.430 +So just to recap kann, the key + +00:06:13.430 --> 00:06:16.480 +assumptions of K&N are that K nearest + +00:06:16.480 --> 00:06:18.260 +neighbors that Samples with similar + +00:06:18.260 --> 00:06:19.730 +features will have similar output + +00:06:19.730 --> 00:06:20.695 +predictions. + +00:06:20.695 --> 00:06:23.290 +And for most of the Distance measures + +00:06:23.290 --> 00:06:25.590 +you implicitly assumes that all the + +00:06:25.590 --> 00:06:27.200 +dimensions are equally important. + +00:06:27.200 --> 00:06:29.820 +So it requires some kind of scaling or + +00:06:29.820 --> 00:06:31.500 +learning to be really effective. + +00:06:33.540 --> 00:06:35.620 +The parameters are just the data + +00:06:35.620 --> 00:06:36.080 +itself. + +00:06:36.080 --> 00:06:37.870 +You don't really have to learn any kind + +00:06:37.870 --> 00:06:40.526 +of statistics of the data. + +00:06:40.526 --> 00:06:42.270 +The data are the parameters. + +00:06:43.820 --> 00:06:46.160 +The designs are mainly the choice of K + +00:06:46.160 --> 00:06:48.130 +if you have higher K then it gets + +00:06:48.130 --> 00:06:49.360 +smoother Prediction. + +00:06:50.340 --> 00:06:51.730 +You can decide how you're going to + +00:06:51.730 --> 00:06:54.400 +combine predictions if K is greater + +00:06:54.400 --> 00:06:56.750 +than one, usually it's just voting or + +00:06:56.750 --> 00:06:57.280 +averaging. + +00:06:58.610 --> 00:07:00.920 +You can try to design the features and + +00:07:00.920 --> 00:07:03.450 +that's where things can get a lot more + +00:07:03.450 --> 00:07:03.930 +creative. + +00:07:04.680 --> 00:07:06.770 +And you can choose a Distance function. + +00:07:08.900 --> 00:07:12.370 +So this K&N is useful in many cases. + +00:07:12.370 --> 00:07:14.520 +So if you have very few examples per + +00:07:14.520 --> 00:07:16.605 +class then it can be applied even if + +00:07:16.605 --> 00:07:17.320 +you just have one. + +00:07:18.080 --> 00:07:20.290 +It can also work if you have many + +00:07:20.290 --> 00:07:21.560 +Examples per class. + +00:07:22.200 --> 00:07:24.910 +It's best if the features are all + +00:07:24.910 --> 00:07:26.960 +roughly equally important, because K&N + +00:07:26.960 --> 00:07:28.540 +itself doesn't really learn which + +00:07:28.540 --> 00:07:29.449 +features are important. + +00:07:31.570 --> 00:07:33.910 +It's good if the training data is + +00:07:33.910 --> 00:07:34.585 +changing frequently. + +00:07:34.585 --> 00:07:37.520 +In the face recognition Example face, + +00:07:37.520 --> 00:07:38.830 +there's no way that Facebook will + +00:07:38.830 --> 00:07:41.160 +collect everybody's Faces up front. + +00:07:41.160 --> 00:07:43.030 +People keep on joining and leaving the + +00:07:43.030 --> 00:07:45.480 +social network, and so they and they + +00:07:45.480 --> 00:07:47.080 +don't want to have to keep retraining + +00:07:47.080 --> 00:07:49.850 +models every time somebody uploads a + +00:07:49.850 --> 00:07:52.005 +image with a new face in it or tags a + +00:07:52.005 --> 00:07:52.615 +new face. + +00:07:52.615 --> 00:07:54.990 +And so the ability to instantly update + +00:07:54.990 --> 00:07:56.330 +your model is very important. + +00:07:58.160 --> 00:07:59.850 +You can apply it to classification or + +00:07:59.850 --> 00:08:01.740 +regression whether you have discrete or + +00:08:01.740 --> 00:08:04.570 +continuous values, and its most + +00:08:04.570 --> 00:08:06.020 +powerful when you do some feature + +00:08:06.020 --> 00:08:08.180 +learning as an upfront operation. + +00:08:10.130 --> 00:08:12.210 +So there's cases where it has its + +00:08:12.210 --> 00:08:13.330 +downsides though. + +00:08:13.330 --> 00:08:15.650 +One is that if you have a lot of + +00:08:15.650 --> 00:08:18.250 +examples that are available per class, + +00:08:18.250 --> 00:08:20.360 +then usually training a Logistic + +00:08:20.360 --> 00:08:23.690 +regressor other Linear Classifier will + +00:08:23.690 --> 00:08:26.200 +outperform because it's able to learn + +00:08:26.200 --> 00:08:27.990 +the importance of different Features. + +00:08:28.950 --> 00:08:32.125 +Also, K&N requires that you store all + +00:08:32.125 --> 00:08:34.692 +the training data and that may require + +00:08:34.692 --> 00:08:38.153 +a lot of storage and it requires a lot + +00:08:38.153 --> 00:08:40.145 +of computation, and that you have to + +00:08:40.145 --> 00:08:42.200 +compare each new input to all of the + +00:08:42.200 --> 00:08:43.750 +inputs in your training data. + +00:08:43.750 --> 00:08:45.525 +So in the case of Facebook for example, + +00:08:45.525 --> 00:08:47.745 +they don't need if somebody uploads, if + +00:08:47.745 --> 00:08:49.780 +they detect a face in somebody's image, + +00:08:49.780 --> 00:08:51.520 +they don't need to compare it to the + +00:08:51.520 --> 00:08:53.410 +other, like 2 billion Facebook users. + +00:08:53.410 --> 00:08:55.176 +They just would compare it to people in + +00:08:55.176 --> 00:08:56.570 +the person's social network, which will + +00:08:56.570 --> 00:08:58.900 +be a much smaller number of Faces. + +00:08:58.970 --> 00:09:01.240 +So they're able to limit the + +00:09:01.240 --> 00:09:02.190 +computation that way. + +00:09:05.940 --> 00:09:08.760 +And then finally, to recap what we + +00:09:08.760 --> 00:09:12.180 +learned on Thursday, there's a basic + +00:09:12.180 --> 00:09:14.420 +machine learning process, which is that + +00:09:14.420 --> 00:09:16.170 +you've got training data, validation + +00:09:16.170 --> 00:09:17.260 +data and TestData. + +00:09:18.160 --> 00:09:19.980 +Given the training data, which are + +00:09:19.980 --> 00:09:22.730 +pairs of Features and labels, you fit + +00:09:22.730 --> 00:09:25.060 +the parameters of your Model. + +00:09:25.060 --> 00:09:26.950 +Then you use the validation Model to + +00:09:26.950 --> 00:09:28.670 +check how good the Model is and maybe + +00:09:28.670 --> 00:09:29.805 +check many models. + +00:09:29.805 --> 00:09:31.960 +You choose the best one and then you + +00:09:31.960 --> 00:09:33.590 +get your final estimate of performance + +00:09:33.590 --> 00:09:34.410 +on the TestData. + +00:09:36.790 --> 00:09:39.670 +We talked about KNN, which is simple + +00:09:39.670 --> 00:09:42.040 +but effective Classifier and regressor + +00:09:42.040 --> 00:09:44.140 +that predicts the label of the most + +00:09:44.140 --> 00:09:45.540 +similar training Example. + +00:09:46.770 --> 00:09:49.110 +And then we talked about kind of + +00:09:49.110 --> 00:09:51.110 +patterns of error and what causes + +00:09:51.110 --> 00:09:51.580 +errors. + +00:09:51.580 --> 00:09:53.780 +So it's important to remember that as + +00:09:53.780 --> 00:09:56.069 +you get more training, more training + +00:09:56.070 --> 00:09:57.830 +samples, you would expect that fitting + +00:09:57.830 --> 00:09:58.962 +the training data gets harder. + +00:09:58.962 --> 00:10:01.500 +So your error will tend to go up while + +00:10:01.500 --> 00:10:03.390 +your error on the TestData will get + +00:10:03.390 --> 00:10:05.535 +lower because the training data better + +00:10:05.535 --> 00:10:07.010 +represents the TestData or better + +00:10:07.010 --> 00:10:08.430 +represents the full distribution. + +00:10:09.770 --> 00:10:11.840 +And there's many reasons why at the end + +00:10:11.840 --> 00:10:13.250 +of training your Algorithm, you're + +00:10:13.250 --> 00:10:14.720 +still going to have error in most + +00:10:14.720 --> 00:10:15.220 +cases. + +00:10:15.880 --> 00:10:17.400 +It could be that the problem is + +00:10:17.400 --> 00:10:20.940 +intrinsically difficult, or it's + +00:10:20.940 --> 00:10:22.590 +impossible to have 0 error. + +00:10:22.590 --> 00:10:24.232 +It could be that you're Model has + +00:10:24.232 --> 00:10:24.845 +limited power. + +00:10:24.845 --> 00:10:27.370 +It could be that your Model has plenty + +00:10:27.370 --> 00:10:29.015 +of power, but you have limited data so + +00:10:29.015 --> 00:10:30.710 +you can't Estimate the parameters + +00:10:30.710 --> 00:10:31.290 +exactly. + +00:10:32.050 --> 00:10:33.100 +And it could be that there's + +00:10:33.100 --> 00:10:34.550 +differences in the training test + +00:10:34.550 --> 00:10:35.280 +distribution. + +00:10:37.020 --> 00:10:38.980 +And then finally it's important to + +00:10:38.980 --> 00:10:41.315 +remember that this Model fitting, the + +00:10:41.315 --> 00:10:42.980 +model design and fitting is just one + +00:10:42.980 --> 00:10:44.750 +part of a larger processing collecting + +00:10:44.750 --> 00:10:46.600 +data and fitting it into an + +00:10:46.600 --> 00:10:47.610 +application. + +00:10:47.610 --> 00:10:51.230 +So both the cases of in Facebook's case + +00:10:51.230 --> 00:10:54.160 +for example they had pre training stage + +00:10:54.160 --> 00:10:56.663 +which is like training a classifier and + +00:10:56.663 --> 00:10:58.852 +then they use that in a different, they + +00:10:58.852 --> 00:11:01.370 +use it in a different way as a nearest + +00:11:01.370 --> 00:11:05.320 +neighbor recognizer on their pool of + +00:11:05.320 --> 00:11:06.010 +user data. + +00:11:07.070 --> 00:11:10.384 +And so they're kind of building a model + +00:11:10.384 --> 00:11:11.212 +using it. + +00:11:11.212 --> 00:11:13.700 +They're building a model one way and + +00:11:13.700 --> 00:11:15.150 +then using it in a different way. + +00:11:15.150 --> 00:11:16.660 +So often that's the case that you have + +00:11:16.660 --> 00:11:17.590 +to kind of be creative. + +00:11:18.360 --> 00:11:20.580 +About how you collect data and how you + +00:11:20.580 --> 00:11:23.800 +can get the model that you need to + +00:11:23.800 --> 00:11:24.860 +solve your application. + +00:11:28.010 --> 00:11:30.033 +Alright, so now I'm going to move on to + +00:11:30.033 --> 00:11:31.640 +the main topic of today's lecture, + +00:11:31.640 --> 00:11:34.880 +which is probabilities and the night + +00:11:34.880 --> 00:11:35.935 +based Classifier. + +00:11:35.935 --> 00:11:39.690 +So the knight based Classifier is + +00:11:39.690 --> 00:11:41.220 +unlike nearest neighbor, it's not. + +00:11:41.990 --> 00:11:44.020 +Usually like the final approach that + +00:11:44.020 --> 00:11:46.080 +somebody takes, but it's sometimes a + +00:11:46.080 --> 00:11:49.460 +piece of a piece of how somebody is + +00:11:49.460 --> 00:11:51.210 +estimating probabilities as part of + +00:11:51.210 --> 00:11:51.870 +their approach. + +00:11:52.690 --> 00:11:55.610 +And it's a good introduction to + +00:11:55.610 --> 00:11:56.630 +Probabilistic models. + +00:11:59.220 --> 00:12:02.525 +So with the nearest neighbor + +00:12:02.525 --> 00:12:04.670 +classifier, that's an instance based + +00:12:04.670 --> 00:12:05.960 +Classifier, which means that you're + +00:12:05.960 --> 00:12:07.800 +assigning labels just based on matching + +00:12:07.800 --> 00:12:08.515 +other instances. + +00:12:08.515 --> 00:12:11.160 +The instances the data are the Model. + +00:12:12.260 --> 00:12:14.590 +Now we're going to start talking about + +00:12:14.590 --> 00:12:15.910 +Probabilistic models. + +00:12:15.910 --> 00:12:18.290 +In a Probabilistic Model, you choose + +00:12:18.290 --> 00:12:21.060 +the label that is most likely given the + +00:12:21.060 --> 00:12:21.630 +Features. + +00:12:21.630 --> 00:12:23.390 +So that's kind of an intuitive thing to + +00:12:23.390 --> 00:12:25.510 +do if you want to know. + +00:12:26.520 --> 00:12:28.690 +Which if you're looking at an image and + +00:12:28.690 --> 00:12:30.390 +trying to classify it into a Digit, it + +00:12:30.390 --> 00:12:32.074 +makes sense that you would assign it to + +00:12:32.074 --> 00:12:34.000 +the Digit that is most likely given the + +00:12:34.000 --> 00:12:35.940 +Features given the pixel intensities. + +00:12:36.610 --> 00:12:38.170 +But of course, like the challenge is + +00:12:38.170 --> 00:12:40.030 +modeling this probability function, how + +00:12:40.030 --> 00:12:42.590 +do you Model the probability of the + +00:12:42.590 --> 00:12:44.000 +label given the data? + +00:12:45.340 --> 00:12:47.520 +So this is just a very compact way of + +00:12:47.520 --> 00:12:48.135 +writing that. + +00:12:48.135 --> 00:12:50.270 +So I have Y star is the predicted + +00:12:50.270 --> 00:12:53.150 +label, and that's equal to the argmax + +00:12:53.150 --> 00:12:53.836 +over Y. + +00:12:53.836 --> 00:12:55.770 +So it's the Y that maximizes + +00:12:55.770 --> 00:12:56.950 +probability of Y given X. + +00:12:56.950 --> 00:12:59.250 +So you assign the label that's most + +00:12:59.250 --> 00:13:00.590 +likely given the data. + +00:13:03.170 --> 00:13:05.210 +So I just want to do a very brief + +00:13:05.210 --> 00:13:08.240 +review of some probability things. + +00:13:08.240 --> 00:13:10.730 +Hopefully this looks familiar, but it's + +00:13:10.730 --> 00:13:12.920 +still useful to refresh on it. + +00:13:13.720 --> 00:13:15.290 +So first Joint and conditional + +00:13:15.290 --> 00:13:16.260 +probability. + +00:13:16.260 --> 00:13:19.040 +If you say probability of X&Y then that + +00:13:19.040 --> 00:13:20.900 +means the probability that both of + +00:13:20.900 --> 00:13:24.180 +those values are true at the same time, + +00:13:24.180 --> 00:13:25.030 +so. + +00:13:26.330 --> 00:13:28.400 +So if you say like the probability that + +00:13:28.400 --> 00:13:29.290 +it's sunny. + +00:13:29.980 --> 00:13:32.540 +And it's rainy, then that's probably a + +00:13:32.540 --> 00:13:33.910 +very low probability, because those + +00:13:33.910 --> 00:13:35.700 +usually don't happen at the same time. + +00:13:35.700 --> 00:13:37.635 +Both X&Y are true. + +00:13:37.635 --> 00:13:40.396 +That's equal to the probability of X + +00:13:40.396 --> 00:13:42.179 +given Y times probability of Y. + +00:13:42.180 --> 00:13:45.725 +So probability of X given Y is the + +00:13:45.725 --> 00:13:48.700 +probability that X is true given the + +00:13:48.700 --> 00:13:50.956 +known values of Y times the probability + +00:13:50.956 --> 00:13:52.280 +that Y is true. + +00:13:52.970 --> 00:13:54.789 +And that's also equal to probability of + +00:13:54.790 --> 00:13:56.769 +Y given X times probability of X. + +00:13:56.770 --> 00:13:59.450 +So you can take a Joint probability and + +00:13:59.450 --> 00:14:01.580 +turn it into a conditional probability + +00:14:01.580 --> 00:14:04.370 +times the probability of their meaning + +00:14:04.370 --> 00:14:06.190 +variables, the condition variables. + +00:14:07.010 --> 00:14:08.660 +And you can apply that down a chain. + +00:14:08.660 --> 00:14:11.341 +So probability of ABC is probability of + +00:14:11.341 --> 00:14:13.531 +a given BC times probability of B given + +00:14:13.531 --> 00:14:14.900 +C times probability of C. + +00:14:17.320 --> 00:14:18.730 +And then it's important to remember + +00:14:18.730 --> 00:14:21.110 +Bayes rule, which is a way of relating + +00:14:21.110 --> 00:14:23.160 +probability of X given Y and + +00:14:23.160 --> 00:14:24.869 +probability of Y given X. + +00:14:25.520 --> 00:14:27.440 +So of X given Y. + +00:14:28.100 --> 00:14:30.516 +Is equal to probability of Y given X + +00:14:30.516 --> 00:14:32.222 +times probability of X over probability + +00:14:32.222 --> 00:14:35.090 +of Y and you can get that by saying + +00:14:35.090 --> 00:14:38.595 +probability of X given Y is probability + +00:14:38.595 --> 00:14:41.599 +of X&Y over probability of Y. + +00:14:41.600 --> 00:14:43.730 +So what was done here is you multiply + +00:14:43.730 --> 00:14:45.910 +this by probability of Y and then + +00:14:45.910 --> 00:14:47.771 +divide it by probability of Y and + +00:14:47.771 --> 00:14:49.501 +probability of X given Y times + +00:14:49.501 --> 00:14:51.519 +probability of Y is probability of X&Y. + +00:14:52.600 --> 00:14:54.390 +And then the probability of X&Y is + +00:14:54.390 --> 00:14:56.030 +broken out into probability of Y given + +00:14:56.030 --> 00:14:57.209 +X times probability of X. + +00:14:59.150 --> 00:15:01.040 +So often it's the case that you want to + +00:15:01.040 --> 00:15:03.484 +kind of switch things you the label and + +00:15:03.484 --> 00:15:06.339 +you want to know the likelihood of the + +00:15:06.339 --> 00:15:08.350 +Features, but you have like a + +00:15:08.350 --> 00:15:10.544 +likelihood for that, but you want a + +00:15:10.544 --> 00:15:11.830 +likelihood the other way of the + +00:15:11.830 --> 00:15:13.654 +probability of the label given the + +00:15:13.654 --> 00:15:13.868 +Features. + +00:15:13.868 --> 00:15:15.529 +And so you use Bayes rule to kind of + +00:15:15.530 --> 00:15:17.550 +turn the tables on your likelihood + +00:15:17.550 --> 00:15:17.950 +function. + +00:15:20.620 --> 00:15:25.810 +So using using using these rules of + +00:15:25.810 --> 00:15:26.530 +probability. + +00:15:27.210 --> 00:15:29.830 +We can show that if I want to find the + +00:15:29.830 --> 00:15:33.250 +Y that maximizes the likelihood of the + +00:15:33.250 --> 00:15:34.690 +label given the data. + +00:15:35.370 --> 00:15:38.490 +That's equivalent to finding the Y that + +00:15:38.490 --> 00:15:41.240 +maximizes the likelihood of the data + +00:15:41.240 --> 00:15:44.520 +given the label times the probability + +00:15:44.520 --> 00:15:45.210 +of the label. + +00:15:45.920 --> 00:15:47.690 +So in other words, if you wanted to + +00:15:47.690 --> 00:15:50.030 +say, well, what is the probability that + +00:15:50.030 --> 00:15:53.550 +my face is Derek given my facial + +00:15:53.550 --> 00:15:54.220 +features? + +00:15:54.950 --> 00:15:56.100 +That's the top. + +00:15:56.100 --> 00:15:58.323 +That's equivalent to saying what's the + +00:15:58.323 --> 00:16:00.400 +probability that it's me without + +00:16:00.400 --> 00:16:02.635 +looking at the Features times the + +00:16:02.635 --> 00:16:04.270 +probability of my Features given that + +00:16:04.270 --> 00:16:04.870 +it's me? + +00:16:04.870 --> 00:16:05.980 +Those are the same. + +00:16:06.330 --> 00:16:09.770 +Those the why that maximizes that is + +00:16:09.770 --> 00:16:11.150 +going to be the same so. + +00:16:12.990 --> 00:16:15.230 +And the reason for that is derived down + +00:16:15.230 --> 00:16:15.720 +here. + +00:16:15.720 --> 00:16:17.473 +So I can take Y given X. + +00:16:17.473 --> 00:16:20.686 +So argmax of Y given X is the as argmax + +00:16:20.686 --> 00:16:23.029 +of Y given X times probability of X. + +00:16:23.780 --> 00:16:26.000 +And the reason for that is just that + +00:16:26.000 --> 00:16:27.880 +probability of X doesn't depend on Y. + +00:16:27.880 --> 00:16:31.140 +So I can multiply multiply this thing + +00:16:31.140 --> 00:16:33.092 +in the argmax by anything that doesn't + +00:16:33.092 --> 00:16:35.410 +depend on Y and it's going to be + +00:16:35.410 --> 00:16:37.890 +unchanged because it's just going to. + +00:16:38.870 --> 00:16:41.460 +The way that maximizes it will be the + +00:16:41.460 --> 00:16:41.780 +same. + +00:16:43.410 --> 00:16:44.940 +So then I turn that. + +00:16:45.530 --> 00:16:47.810 +I turned that into the Joint Y&X and + +00:16:47.810 --> 00:16:48.940 +then I broke it out again. + +00:16:49.900 --> 00:16:51.300 +Right, so the reason why this is + +00:16:51.300 --> 00:16:54.430 +important is that I can choose to + +00:16:54.430 --> 00:16:57.562 +either Model directly the probability + +00:16:57.562 --> 00:17:00.659 +of the label given the data, or I can + +00:17:00.659 --> 00:17:02.231 +choose the Model the probability of the + +00:17:02.231 --> 00:17:03.129 +data given the label. + +00:17:03.910 --> 00:17:06.172 +In a Naive Bayes, we're going to Model + +00:17:06.172 --> 00:17:07.950 +probability the data given the label, + +00:17:07.950 --> 00:17:09.510 +and then in the next class we'll talk + +00:17:09.510 --> 00:17:11.425 +about logistic regression where we try + +00:17:11.425 --> 00:17:12.930 +to directly Model the probability of + +00:17:12.930 --> 00:17:14.000 +the label given the data. + +00:17:22.090 --> 00:17:24.760 +All right, so let's just. + +00:17:26.170 --> 00:17:29.400 +Do a simple probability exercise just + +00:17:29.400 --> 00:17:31.430 +to kind of make sure that. + +00:17:33.430 --> 00:17:34.730 +That we get. + +00:17:37.010 --> 00:17:38.230 +So let's say. + +00:17:39.620 --> 00:17:41.060 +Here I have a feature. + +00:17:41.060 --> 00:17:41.970 +Doesn't really matter what the + +00:17:41.970 --> 00:17:43.440 +Features, but let's say that it's + +00:17:43.440 --> 00:17:45.233 +whether something is larger than £10 + +00:17:45.233 --> 00:17:48.210 +and I collected a bunch of different + +00:17:48.210 --> 00:17:50.530 +animals, cats and dogs and measured + +00:17:50.530 --> 00:17:50.770 +them. + +00:17:51.450 --> 00:17:53.130 +And I want to train something that will + +00:17:53.130 --> 00:17:54.510 +tell me whether or not something is a + +00:17:54.510 --> 00:17:54.810 +cat. + +00:17:55.730 --> 00:17:57.370 +And so. + +00:17:58.190 --> 00:18:00.985 +Or a dog, and so I have like 40 + +00:18:00.985 --> 00:18:03.280 +different cats and 45 different dogs, + +00:18:03.280 --> 00:18:04.860 +and I measured whether or not they're + +00:18:04.860 --> 00:18:06.693 +bigger than £10. + +00:18:06.693 --> 00:18:10.270 +So first, given this empirical + +00:18:10.270 --> 00:18:12.505 +distribution, given these samples that + +00:18:12.505 --> 00:18:15.120 +I have, what's the probability that Y + +00:18:15.120 --> 00:18:15.810 +is a cat? + +00:18:22.430 --> 00:18:25.970 +So it's actually 40 / 85 because it's + +00:18:25.970 --> 00:18:26.960 +going to be. + +00:18:27.640 --> 00:18:29.030 +Let me see if I can write on this. + +00:18:36.840 --> 00:18:37.330 +OK. + +00:18:39.520 --> 00:18:40.460 +That's not what I wanted. + +00:18:43.970 --> 00:18:45.500 +If I can get the pen to work. + +00:18:48.610 --> 00:18:50.360 +OK, it doesn't work that well. + +00:18:55.010 --> 00:18:56.250 +OK, forget that. + +00:18:56.250 --> 00:18:57.420 +Alright, I'll write it on the board. + +00:18:57.420 --> 00:18:59.639 +So it's 40 / 85. + +00:19:01.780 --> 00:19:05.010 +So it's 40 / 40 + 45. + +00:19:05.920 --> 00:19:08.595 +And that's because there's 40 cats and + +00:19:08.595 --> 00:19:09.888 +there's 45 dogs. + +00:19:09.888 --> 00:19:13.040 +So I take the count of all the cats and + +00:19:13.040 --> 00:19:14.970 +divide it by the count of all the data + +00:19:14.970 --> 00:19:16.635 +in total, all the cats and dogs. + +00:19:16.635 --> 00:19:17.860 +So that's 40 / 85. + +00:19:18.580 --> 00:19:20.470 +And what's the probability that Y is a + +00:19:20.470 --> 00:19:22.810 +cat given that X is false? + +00:19:29.380 --> 00:19:31.510 +So it's right? + +00:19:31.510 --> 00:19:34.240 +So it's 15 / 20 or 3 / 4. + +00:19:34.240 --> 00:19:35.890 +And that's because given that X is + +00:19:35.890 --> 00:19:37.620 +false, I'm just in this one column + +00:19:37.620 --> 00:19:40.799 +here, so it's 15 / 15 / 20. + +00:19:42.090 --> 00:19:45.110 +And what's the probability that X is + +00:19:45.110 --> 00:19:46.650 +false given that Y is a cat? + +00:19:49.320 --> 00:19:51.570 +Right, 15 / 480 because if I know that + +00:19:51.570 --> 00:19:53.500 +Y is a Cat, then I'm in the top row, so + +00:19:53.500 --> 00:19:55.590 +it's just 15 divided by all the cats, + +00:19:55.590 --> 00:19:56.650 +so 15 / 40. + +00:19:58.320 --> 00:20:00.737 +OK, and it's important to remember that + +00:20:00.737 --> 00:20:03.119 +Y given X is different than X given Y. + +00:20:05.110 --> 00:20:08.276 +Right, so some other simple rules of + +00:20:08.276 --> 00:20:08.572 +probability. + +00:20:08.572 --> 00:20:11.150 +One is the law of total probability. + +00:20:11.150 --> 00:20:13.060 +That is, if you sum over all the values + +00:20:13.060 --> 00:20:16.020 +of a variable, then the sum of those + +00:20:16.020 --> 00:20:17.630 +probabilities is equal to 1. + +00:20:18.240 --> 00:20:20.450 +And if this were a continuous variable, + +00:20:20.450 --> 00:20:21.840 +it would just be an integral over the + +00:20:21.840 --> 00:20:23.716 +domain of X over all the values of X + +00:20:23.716 --> 00:20:26.180 +and then the integral over P of X is + +00:20:26.180 --> 00:20:26.690 +equal to 1. + +00:20:27.980 --> 00:20:29.470 +Then I've got Marginalization. + +00:20:29.470 --> 00:20:31.990 +So if I have a joint probability of two + +00:20:31.990 --> 00:20:34.150 +variables and I want to get rid of one + +00:20:34.150 --> 00:20:34.520 +of them. + +00:20:35.280 --> 00:20:37.630 +Then I take this sum over all the + +00:20:37.630 --> 00:20:39.290 +values of 1 and the variables. + +00:20:39.290 --> 00:20:41.052 +In this case it's the sum over all the + +00:20:41.052 --> 00:20:41.900 +values of X. + +00:20:42.570 --> 00:20:46.268 +Of X&Y and that's going to be equal to + +00:20:46.268 --> 00:20:46.910 +P of Y. + +00:20:53.440 --> 00:20:55.380 +And then finally independence. + +00:20:55.380 --> 00:20:59.691 +So A is independent of B if and only if + +00:20:59.691 --> 00:21:02.414 +the probability of A&B is equal to the + +00:21:02.414 --> 00:21:04.115 +probability of a times the probability + +00:21:04.115 --> 00:21:04.660 +of B. + +00:21:05.430 --> 00:21:07.974 +Or another way to write it is that + +00:21:07.974 --> 00:21:10.142 +probability that what this implies is + +00:21:10.142 --> 00:21:12.500 +that probability of a given B is equal + +00:21:12.500 --> 00:21:13.890 +to probability of a. + +00:21:13.890 --> 00:21:15.680 +So if I just divide both sides by + +00:21:15.680 --> 00:21:17.250 +probability of B then I get that. + +00:21:18.160 --> 00:21:20.855 +Or probability of B given A equals + +00:21:20.855 --> 00:21:22.010 +probability of B. + +00:21:22.010 --> 00:21:24.150 +So these things are the top one. + +00:21:24.150 --> 00:21:25.700 +Might not be something that pops into + +00:21:25.700 --> 00:21:26.420 +your head right away. + +00:21:26.420 --> 00:21:28.450 +It's not necessarily as intuitive, but + +00:21:28.450 --> 00:21:30.001 +these are pretty intuitive that + +00:21:30.001 --> 00:21:32.376 +probability of a given B equals + +00:21:32.376 --> 00:21:33.564 +probability of a. + +00:21:33.564 --> 00:21:36.050 +So in other words, whether or not a is + +00:21:36.050 --> 00:21:37.470 +true doesn't depend on B at all. + +00:21:38.720 --> 00:21:40.430 +And whether or not B is true doesn't + +00:21:40.430 --> 00:21:42.360 +depend on A at all, and then you can + +00:21:42.360 --> 00:21:44.810 +easily get to the one up there just by + +00:21:44.810 --> 00:21:47.410 +multiplying here both sides by + +00:21:47.410 --> 00:21:48.100 +probability of a. + +00:21:56.140 --> 00:21:59.180 +Alright, so in some of the slides + +00:21:59.180 --> 00:22:00.650 +there's going to be a bunch of like + +00:22:00.650 --> 00:22:02.760 +indices, so I just wanted to try to be + +00:22:02.760 --> 00:22:04.370 +consistent in the way that I use them. + +00:22:05.030 --> 00:22:07.674 +And also like usually verbally say what + +00:22:07.674 --> 00:22:10.543 +the what the variables mean, but when I + +00:22:10.543 --> 00:22:14.300 +say XI mean the ith feature so I is a + +00:22:14.300 --> 00:22:15.085 +feature index. + +00:22:15.085 --> 00:22:18.619 +When I say XNI mean the nth sample, so + +00:22:18.620 --> 00:22:20.520 +north is the sample index and Lynn + +00:22:20.520 --> 00:22:21.590 +would be the nth label. + +00:22:22.370 --> 00:22:24.993 +So if I say X and I, then that's the + +00:22:24.993 --> 00:22:26.760 +ith feature of the nth label. + +00:22:26.760 --> 00:22:29.763 +So for digits for example, would be the + +00:22:29.763 --> 00:22:33.720 +ith pixel of the nth Digit Example. + +00:22:35.070 --> 00:22:37.580 +I used this delta here to indicate with + +00:22:37.580 --> 00:22:39.900 +some expression inside to indicate that + +00:22:39.900 --> 00:22:42.780 +it returns true or returns one if the + +00:22:42.780 --> 00:22:44.850 +expression inside it is true and 0 + +00:22:44.850 --> 00:22:45.410 +otherwise. + +00:22:46.200 --> 00:22:48.110 +And I'll Use V for a feature value. + +00:22:55.320 --> 00:22:57.900 +So if I want to Estimate the + +00:22:57.900 --> 00:22:59.830 +probabilities of some function, I can + +00:22:59.830 --> 00:23:00.578 +just do it by counting. + +00:23:00.578 --> 00:23:02.760 +So if I want to say what is the + +00:23:02.760 --> 00:23:04.950 +probability that X equals some value + +00:23:04.950 --> 00:23:07.600 +and I have capital N Samples, then I + +00:23:07.600 --> 00:23:09.346 +can just take a sum over all the + +00:23:09.346 --> 00:23:11.350 +samples and count for how many of them + +00:23:11.350 --> 00:23:14.030 +XN equals V so that's kind of intuitive + +00:23:14.030 --> 00:23:14.480 +if I have. + +00:23:15.870 --> 00:23:17.750 +If I have a month full of days and I + +00:23:17.750 --> 00:23:19.280 +want to say what's the probability that + +00:23:19.280 --> 00:23:21.610 +one of those days is sunny, then I can + +00:23:21.610 --> 00:23:23.809 +just take a sum over all the I can + +00:23:23.810 --> 00:23:25.370 +count how many sunny days there were + +00:23:25.370 --> 00:23:26.908 +divided by the total number of days and + +00:23:26.908 --> 00:23:27.930 +that gives me an Estimate. + +00:23:31.930 --> 00:23:35.340 +But what if I have 100 variables? + +00:23:35.340 --> 00:23:36.380 +So if I have. + +00:23:37.310 --> 00:23:39.220 +For example, in the digits case I have + +00:23:39.220 --> 00:23:42.840 +784 different and pixel intensities. + +00:23:43.710 --> 00:23:46.350 +And there's no way I can count over all + +00:23:46.350 --> 00:23:48.222 +possible combinations of pixel + +00:23:48.222 --> 00:23:49.000 +intensities, right? + +00:23:49.000 --> 00:23:51.470 +Even if I were to turn them into binary + +00:23:51.470 --> 00:23:56.070 +values, there would be 2 to the 784 + +00:23:56.070 --> 00:23:58.107 +different combinations of pixel + +00:23:58.107 --> 00:23:58.670 +intensities. + +00:23:58.670 --> 00:24:01.635 +So you would need like data samples + +00:24:01.635 --> 00:24:03.520 +that are equal to like number of atoms + +00:24:03.520 --> 00:24:05.300 +in the universe or something like that + +00:24:05.300 --> 00:24:07.415 +in order to even begin to Estimate it. + +00:24:07.415 --> 00:24:08.900 +And that would that would only be + +00:24:08.900 --> 00:24:10.460 +giving you very few samples per + +00:24:10.460 --> 00:24:11.050 +combination. + +00:24:12.860 --> 00:24:15.407 +So obviously, like jointly modeling a + +00:24:15.407 --> 00:24:17.799 +whole bunch of different, the + +00:24:17.800 --> 00:24:19.431 +probability of a whole bunch of + +00:24:19.431 --> 00:24:20.740 +different variables is usually + +00:24:20.740 --> 00:24:23.490 +impossible, and even approximating it, + +00:24:23.490 --> 00:24:24.880 +it's very challenging. + +00:24:24.880 --> 00:24:26.260 +You have to try to solve for the + +00:24:26.260 --> 00:24:28.036 +dependency structures and then solve + +00:24:28.036 --> 00:24:30.236 +for different combinations of variables + +00:24:30.236 --> 00:24:30.699 +and. + +00:24:31.550 --> 00:24:33.740 +And then worry about the dependencies + +00:24:33.740 --> 00:24:35.040 +that aren't fully accounted for. + +00:24:35.880 --> 00:24:37.670 +And so it's just really difficult to + +00:24:37.670 --> 00:24:40.160 +estimate the probability of all your + +00:24:40.160 --> 00:24:41.810 +Features given the label. + +00:24:42.900 --> 00:24:43.610 +Jointly. + +00:24:44.440 --> 00:24:47.540 +And so that's the Naive Bayes Model + +00:24:47.540 --> 00:24:48.240 +comes in. + +00:24:48.240 --> 00:24:50.430 +It makes us greatly simplifying + +00:24:50.430 --> 00:24:51.060 +assumption. + +00:24:51.730 --> 00:24:54.132 +Which is that all of the features are + +00:24:54.132 --> 00:24:56.010 +independent given the label, so it + +00:24:56.010 --> 00:24:57.480 +doesn't mean the Features are + +00:24:57.480 --> 00:24:57.840 +independent. + +00:24:57.940 --> 00:25:00.200 +Unconditionally, but they're + +00:25:00.200 --> 00:25:02.370 +independent given the label, so. + +00:25:03.550 --> 00:25:05.716 +So because of because they're + +00:25:05.716 --> 00:25:06.149 +independent. + +00:25:06.150 --> 00:25:08.400 +Remember that probability of A&B equals + +00:25:08.400 --> 00:25:11.173 +probability of a * b times probability + +00:25:11.173 --> 00:25:12.603 +B if they're independent. + +00:25:12.603 --> 00:25:15.160 +So probability of X that's like a Joint + +00:25:15.160 --> 00:25:17.920 +X, all the Features given Y is equal to + +00:25:17.920 --> 00:25:20.501 +the product over all the features of + +00:25:20.501 --> 00:25:22.919 +probability of each feature given Y. + +00:25:24.880 --> 00:25:28.866 +And so then I can make my Classifier as + +00:25:28.866 --> 00:25:30.450 +the Y star. + +00:25:30.450 --> 00:25:32.880 +The most likely label is the one that + +00:25:32.880 --> 00:25:35.415 +maximizes this joint probability of + +00:25:35.415 --> 00:25:37.930 +probability of X given Y times + +00:25:37.930 --> 00:25:38.779 +probability of Y. + +00:25:39.810 --> 00:25:42.715 +And this thing, the joint probability + +00:25:42.715 --> 00:25:44.985 +of X given Y would be really hard to + +00:25:44.985 --> 00:25:45.240 +Estimate. + +00:25:45.240 --> 00:25:47.490 +You need tons of data, but this is not + +00:25:47.490 --> 00:25:49.120 +so hard to Estimate because you're just + +00:25:49.120 --> 00:25:50.590 +estimating the probability of 1 + +00:25:50.590 --> 00:25:51.590 +variable at a time. + +00:25:57.200 --> 00:25:59.190 +So for example if I. + +00:25:59.810 --> 00:26:01.900 +In the Digit Example, this would be + +00:26:01.900 --> 00:26:03.860 +saying that the I'm going to choose the + +00:26:03.860 --> 00:26:07.310 +label that maximizes the product of + +00:26:07.310 --> 00:26:09.220 +likelihoods of each of the pixel + +00:26:09.220 --> 00:26:09.980 +intensities. + +00:26:10.690 --> 00:26:12.555 +So I'm just going to consider each + +00:26:12.555 --> 00:26:13.170 +pixel. + +00:26:13.170 --> 00:26:15.170 +How likely is each pixel to have its + +00:26:15.170 --> 00:26:16.959 +intensity given the label? + +00:26:16.960 --> 00:26:18.230 +And then I choose the label that + +00:26:18.230 --> 00:26:20.132 +maximizes that, taking the product of + +00:26:20.132 --> 00:26:21.760 +all the all those likelihoods over the + +00:26:21.760 --> 00:26:22.140 +pixels. + +00:26:23.210 --> 00:26:23.690 +So. + +00:26:24.650 --> 00:26:26.880 +Obviously it's not a perfect Model, + +00:26:26.880 --> 00:26:28.210 +even if I know that. + +00:26:28.210 --> 00:26:30.610 +If I'm given that it's a three, knowing + +00:26:30.610 --> 00:26:32.759 +that one pixel has an intensity of 1 + +00:26:32.760 --> 00:26:33.920 +makes it more likely that the + +00:26:33.920 --> 00:26:35.815 +neighboring pixel has a likelihood of + +00:26:35.815 --> 00:26:36.240 +1. + +00:26:36.240 --> 00:26:37.630 +On the other hand, it's not a terrible + +00:26:37.630 --> 00:26:38.710 +Model either. + +00:26:38.710 --> 00:26:41.028 +If I know that it's a 3, then I have a + +00:26:41.028 --> 00:26:43.210 +pretty good idea of the expected + +00:26:43.210 --> 00:26:45.177 +intensity of each pixel, so I have a + +00:26:45.177 --> 00:26:46.503 +pretty good idea of how likely each + +00:26:46.503 --> 00:26:47.920 +pixel is to be a one or a zero. + +00:26:50.490 --> 00:26:51.780 +In the case of the temperature + +00:26:51.780 --> 00:26:53.760 +Regression will make a slightly + +00:26:53.760 --> 00:26:55.040 +different assumption. + +00:26:55.040 --> 00:26:57.736 +So here we have continuous Features and + +00:26:57.736 --> 00:26:59.320 +a continuous Prediction. + +00:27:00.030 --> 00:27:02.840 +So we're going to assume that each + +00:27:02.840 --> 00:27:05.490 +feature predicts the temperature that + +00:27:05.490 --> 00:27:07.690 +we're trying to predict the tomorrow's + +00:27:07.690 --> 00:27:10.160 +Cleveland temperature with some offset + +00:27:10.160 --> 00:27:10.673 +and variance. + +00:27:10.673 --> 00:27:13.100 +So for example, if I know yesterday's + +00:27:13.100 --> 00:27:14.670 +Cleveland temperature, then tomorrow's + +00:27:14.670 --> 00:27:16.633 +Cleveland temperature is probably about + +00:27:16.633 --> 00:27:19.300 +the same, but with some variance around + +00:27:19.300 --> 00:27:19.577 +it. + +00:27:19.577 --> 00:27:21.239 +If I know the Cleveland temperature + +00:27:21.240 --> 00:27:23.520 +from three days ago, then tomorrow's is + +00:27:23.520 --> 00:27:25.732 +also expected to be about the same but + +00:27:25.732 --> 00:27:26.525 +with higher variance. + +00:27:26.525 --> 00:27:28.596 +If I know the temperature of Austin, + +00:27:28.596 --> 00:27:30.590 +TX, then probably Cleveland is a bit + +00:27:30.590 --> 00:27:31.819 +colder with some variance. + +00:27:33.550 --> 00:27:34.940 +And so I'm going to use just that + +00:27:34.940 --> 00:27:37.100 +combination of individual predictions + +00:27:37.100 --> 00:27:38.480 +to make my final prediction. + +00:27:44.170 --> 00:27:48.680 +So here is the Naive Bayes Algorithm. + +00:27:49.540 --> 00:27:53.250 +For training, I Estimate the parameters + +00:27:53.250 --> 00:27:55.370 +for each of my likelihood functions, + +00:27:55.370 --> 00:27:57.290 +the probability of each feature given + +00:27:57.290 --> 00:27:57.910 +the label. + +00:27:58.940 --> 00:28:01.878 +And I Estimate the parameters for my + +00:28:01.878 --> 00:28:02.232 +prior. + +00:28:02.232 --> 00:28:06.640 +The prior is like the my Estimate, my + +00:28:06.640 --> 00:28:08.370 +likelihood of the label when I don't + +00:28:08.370 --> 00:28:10.180 +know anything else, just before I look + +00:28:10.180 --> 00:28:11.200 +at anything. + +00:28:11.200 --> 00:28:13.475 +So the probability of the label. + +00:28:13.475 --> 00:28:14.770 +And that's usually really easy to + +00:28:14.770 --> 00:28:15.140 +Estimate. + +00:28:17.020 --> 00:28:19.280 +And then at Prediction time, I'm going + +00:28:19.280 --> 00:28:22.970 +to solve for the label that maximizes + +00:28:22.970 --> 00:28:26.330 +the probability of X&Y or the and which + +00:28:26.330 --> 00:28:28.620 +the Naive Bayes assumption is the + +00:28:28.620 --> 00:28:31.110 +product over I of probability of XI + +00:28:31.110 --> 00:28:32.649 +given Y times probability of Y. + +00:28:36.470 --> 00:28:40.455 +The Naive Naive Bayes is that it's just + +00:28:40.455 --> 00:28:42.050 +the independence assumption. + +00:28:42.050 --> 00:28:45.150 +It's not an insult to Thomas Bayes that + +00:28:45.150 --> 00:28:46.890 +he's an idiot or something. + +00:28:46.890 --> 00:28:49.970 +It's just that we're going to make this + +00:28:49.970 --> 00:28:52.140 +very simplifying assumption. + +00:28:58.170 --> 00:29:00.550 +So all right, so the first thing we + +00:29:00.550 --> 00:29:02.710 +have to deal with is how do we Estimate + +00:29:02.710 --> 00:29:03.590 +this probability? + +00:29:03.590 --> 00:29:06.500 +We want to get some probability of each + +00:29:06.500 --> 00:29:08.050 +feature given the data. + +00:29:08.960 --> 00:29:10.990 +And the basic principles are that you + +00:29:10.990 --> 00:29:12.909 +want to choose parameters. + +00:29:12.910 --> 00:29:14.550 +First you have to have a model for your + +00:29:14.550 --> 00:29:16.610 +likelihood, and then you have to + +00:29:16.610 --> 00:29:19.394 +maximize the parameters of that model + +00:29:19.394 --> 00:29:21.908 +that you have to, sorry, Choose the + +00:29:21.908 --> 00:29:22.885 +parameters of that Model. + +00:29:22.885 --> 00:29:25.180 +That makes your training data most + +00:29:25.180 --> 00:29:25.600 +likely. + +00:29:25.600 --> 00:29:27.210 +That's the main principle. + +00:29:27.210 --> 00:29:29.780 +So if I say somebody says maximum + +00:29:29.780 --> 00:29:32.390 +likelihood estimation or Emily, that's + +00:29:32.390 --> 00:29:34.190 +like straight up maximizes the + +00:29:34.190 --> 00:29:37.865 +probability of the data given your + +00:29:37.865 --> 00:29:38.800 +parameters in your model. + +00:29:40.320 --> 00:29:42.480 +Sometimes that can result in + +00:29:42.480 --> 00:29:44.120 +overconfident estimates. + +00:29:44.120 --> 00:29:46.210 +So for example if I just have like. + +00:29:46.970 --> 00:29:47.800 +If I. + +00:29:48.430 --> 00:29:51.810 +If I have like 2 measurements, let's + +00:29:51.810 --> 00:29:53.470 +say I want to know what's the average + +00:29:53.470 --> 00:29:56.044 +weight of a bird and I just have two + +00:29:56.044 --> 00:29:58.480 +birds, and I say it's probably like a + +00:29:58.480 --> 00:29:59.585 +Gaussian distribution. + +00:29:59.585 --> 00:30:02.012 +I can Estimate a mean and a variance + +00:30:02.012 --> 00:30:05.970 +from those two birds, but that Estimate + +00:30:05.970 --> 00:30:07.105 +could be like way off. + +00:30:07.105 --> 00:30:09.100 +So often it's a good idea to have some + +00:30:09.100 --> 00:30:11.530 +kind of Prior or to prevent the + +00:30:11.530 --> 00:30:12.780 +variance from going too low. + +00:30:12.780 --> 00:30:14.740 +So if I looked at two birds and I said + +00:30:14.740 --> 00:30:16.860 +and they both happen to be like 47 + +00:30:16.860 --> 00:30:17.510 +grams. + +00:30:17.870 --> 00:30:19.965 +I probably wouldn't want to say that + +00:30:19.965 --> 00:30:22.966 +the mean is 47 and the variance is 0, + +00:30:22.966 --> 00:30:25.170 +because then I would be saying like if + +00:30:25.170 --> 00:30:27.090 +there's another bird that has 48 grams, + +00:30:27.090 --> 00:30:28.550 +that's like infinitely unlikely. + +00:30:28.550 --> 00:30:29.880 +It's a 0 probability. + +00:30:29.880 --> 00:30:31.600 +So often you want to have some kind of + +00:30:31.600 --> 00:30:34.270 +Prior over your variables as well in + +00:30:34.270 --> 00:30:37.025 +order to prevent likelihoods going to 0 + +00:30:37.025 --> 00:30:38.430 +because you just didn't have enough + +00:30:38.430 --> 00:30:40.120 +data to correctly Estimate them. + +00:30:40.930 --> 00:30:42.650 +So it's like Warren Buffett says with + +00:30:42.650 --> 00:30:43.230 +investing. + +00:30:43.850 --> 00:30:45.550 +It's not just about maximizing the + +00:30:45.550 --> 00:30:47.690 +expectation, it's also about making + +00:30:47.690 --> 00:30:48.890 +sure there are no zeros. + +00:30:48.890 --> 00:30:50.190 +Because if you have a zero and your + +00:30:50.190 --> 00:30:51.670 +product of likelihoods, the whole thing + +00:30:51.670 --> 00:30:52.090 +is 0. + +00:30:53.690 --> 00:30:55.995 +And if you have a zero, return your + +00:30:55.995 --> 00:30:57.900 +whole investment at any point, your + +00:30:57.900 --> 00:30:59.330 +whole bank account is 0. + +00:31:03.120 --> 00:31:06.550 +All right, so we have so. + +00:31:06.920 --> 00:31:08.840 +How do we Estimate P of X given Y given + +00:31:08.840 --> 00:31:09.340 +the data? + +00:31:09.340 --> 00:31:10.980 +It's always based on maximizing the + +00:31:10.980 --> 00:31:11.930 +likelihood of the data. + +00:31:12.690 --> 00:31:14.360 +Over your parameters, but you have + +00:31:14.360 --> 00:31:15.940 +different solutions depending on your + +00:31:15.940 --> 00:31:18.200 +Model and. + +00:31:18.370 --> 00:31:19.860 +I guess it just depends on your Model. + +00:31:20.520 --> 00:31:24.180 +So for binomial, a binomial is just if + +00:31:24.180 --> 00:31:25.790 +you have a binary variable, then + +00:31:25.790 --> 00:31:27.314 +there's some probability that the + +00:31:27.314 --> 00:31:29.450 +variable is 1 and 1 minus that + +00:31:29.450 --> 00:31:31.790 +probability that the variable is 0. + +00:31:31.790 --> 00:31:36.126 +So Theta Ki is the probability that X I + +00:31:36.126 --> 00:31:38.510 += 1 given y = K. + +00:31:39.510 --> 00:31:40.590 +And you can write it. + +00:31:40.590 --> 00:31:42.349 +It's kind of a weird way. + +00:31:42.350 --> 00:31:43.700 +I mean it looks like a weird way to + +00:31:43.700 --> 00:31:44.390 +write it. + +00:31:44.390 --> 00:31:46.190 +But if you think about it, if XI equals + +00:31:46.190 --> 00:31:48.760 +one, then the probability is Theta Ki. + +00:31:49.390 --> 00:31:51.630 +And if XI equals zero, then the + +00:31:51.630 --> 00:31:54.160 +probability is 1 minus Theta Ki so. + +00:31:54.800 --> 00:31:55.440 +Makes sense? + +00:31:56.390 --> 00:31:58.390 +And if I want to Estimate this, all I + +00:31:58.390 --> 00:32:00.530 +have to do is count over all my data + +00:32:00.530 --> 00:32:01.180 +Samples. + +00:32:01.180 --> 00:32:06.410 +How many times does xni equal 1 and y = + +00:32:06.410 --> 00:32:06.880 +K? + +00:32:07.530 --> 00:32:09.310 +Divided by the total number of times + +00:32:09.310 --> 00:32:10.490 +that Y and equals K. + +00:32:11.610 --> 00:32:13.290 +And then here it is in Python. + +00:32:13.290 --> 00:32:15.620 +So it's just a sum over all my data. + +00:32:15.620 --> 00:32:18.170 +I'm looking at the ith feature here, + +00:32:18.170 --> 00:32:20.377 +checking how many times these equal 1 + +00:32:20.377 --> 00:32:23.585 +and the label is equal to K divided by + +00:32:23.585 --> 00:32:25.170 +the number of times the label is equal + +00:32:25.170 --> 00:32:25.580 +to K. + +00:32:27.240 --> 00:32:28.780 +And if I have a multinomial, it's + +00:32:28.780 --> 00:32:31.100 +basically the same thing except that I + +00:32:31.100 --> 00:32:35.342 +sum over the number of times that X and + +00:32:35.342 --> 00:32:37.990 +I = V, where V could be say, zero to 10 + +00:32:37.990 --> 00:32:38.840 +or something like that. + +00:32:39.740 --> 00:32:42.490 +And otherwise it's the same. + +00:32:42.490 --> 00:32:46.040 +So I can Estimate if I have 10 + +00:32:46.040 --> 00:32:49.576 +different variables and I Estimate + +00:32:49.576 --> 00:32:52.590 +Theta KIV for all 10 variables, then + +00:32:52.590 --> 00:32:54.410 +the sum of those Theta kives should be + +00:32:54.410 --> 00:32:54.624 +one. + +00:32:54.624 --> 00:32:56.540 +So one of those is a constrained + +00:32:56.540 --> 00:32:56.910 +variable. + +00:32:58.820 --> 00:33:00.420 +And it will workout that way if you + +00:33:00.420 --> 00:33:01.270 +Estimate it this way. + +00:33:05.970 --> 00:33:08.733 +So if we have a continuous variable by + +00:33:08.733 --> 00:33:11.730 +the way, like, these can be fairly + +00:33:11.730 --> 00:33:15.360 +easily derived just by writing out the + +00:33:15.360 --> 00:33:18.720 +likelihood terms and taking a partial + +00:33:18.720 --> 00:33:21.068 +derivative with respect to the variable + +00:33:21.068 --> 00:33:22.930 +and setting that equal to 0. + +00:33:22.930 --> 00:33:24.810 +But it does take like a page of + +00:33:24.810 --> 00:33:26.940 +equations, so I decided not to subject + +00:33:26.940 --> 00:33:27.379 +you to it. + +00:33:28.260 --> 00:33:30.190 +Since since, solving for these is not + +00:33:30.190 --> 00:33:30.920 +the point right now. + +00:33:32.920 --> 00:33:34.730 +And so. + +00:33:34.800 --> 00:33:36.000 +Are. + +00:33:36.000 --> 00:33:38.620 +Let's say X is a continuous variable. + +00:33:38.620 --> 00:33:40.740 +Maybe I want to assume that XI is a + +00:33:40.740 --> 00:33:44.052 +Gaussian given some label, where the + +00:33:44.052 --> 00:33:45.770 +label is a discrete variable. + +00:33:47.220 --> 00:33:51.023 +So Gaussians, if you took hopefully you + +00:33:51.023 --> 00:33:52.625 +took probably your statistics and you + +00:33:52.625 --> 00:33:53.940 +probably ran into Gaussians all the + +00:33:53.940 --> 00:33:54.230 +time. + +00:33:54.230 --> 00:33:55.820 +Gaussians come up a lot for many + +00:33:55.820 --> 00:33:56.550 +reasons. + +00:33:56.550 --> 00:33:58.749 +One of them is that if you add a lot of + +00:33:58.750 --> 00:34:01.125 +random variables together, then if you + +00:34:01.125 --> 00:34:02.839 +add enough of them, then it will end up + +00:34:02.840 --> 00:34:03.000 +there. + +00:34:03.000 --> 00:34:04.280 +Some of them will end up being a + +00:34:04.280 --> 00:34:05.320 +Gaussian distribution. + +00:34:07.080 --> 00:34:09.415 +So there's lots of things end up being + +00:34:09.415 --> 00:34:09.700 +Gaussians. + +00:34:09.700 --> 00:34:11.500 +Gaussians is a really common noise + +00:34:11.500 --> 00:34:13.536 +model, and it also is like really easy + +00:34:13.536 --> 00:34:14.320 +to work with. + +00:34:14.320 --> 00:34:16.060 +Even though it looks complicated. + +00:34:16.060 --> 00:34:17.820 +When you take the log of it ends up + +00:34:17.820 --> 00:34:19.342 +just being a quadratic, which is easy + +00:34:19.342 --> 00:34:20.010 +to minimize. + +00:34:22.250 --> 00:34:24.460 +So there's the Gaussian expression on + +00:34:24.460 --> 00:34:24.950 +the top. + +00:34:26.550 --> 00:34:28.420 +And I. + +00:34:29.290 --> 00:34:30.610 +So let me get my. + +00:34:33.940 --> 00:34:34.490 +There it goes. + +00:34:34.490 --> 00:34:37.060 +OK, so here's the Gaussian expression + +00:34:37.060 --> 00:34:39.260 +one over square of 2π Sigma Ki. + +00:34:39.260 --> 00:34:42.075 +So the parameters here are M UI which + +00:34:42.075 --> 00:34:43.830 +is mu Ki which is the mean. + +00:34:44.980 --> 00:34:47.700 +For the KTH label and the ith feature + +00:34:47.700 --> 00:34:49.946 +in Sigma, Ki is the stair deviation for + +00:34:49.946 --> 00:34:52.080 +the Keith label and the Ith feature. + +00:34:52.900 --> 00:34:54.700 +And so the higher the standard + +00:34:54.700 --> 00:34:57.090 +deviation is, the bigger the Gaussian + +00:34:57.090 --> 00:34:57.425 +is. + +00:34:57.425 --> 00:34:59.920 +So if you look at these plots here, the + +00:34:59.920 --> 00:35:02.150 +it's kind of blurry the. + +00:35:02.770 --> 00:35:05.540 +The red curve or the actually the + +00:35:05.540 --> 00:35:07.130 +yellow curve has like the biggest + +00:35:07.130 --> 00:35:08.880 +distribution, the broadest distribution + +00:35:08.880 --> 00:35:10.510 +and it has the highest variance or + +00:35:10.510 --> 00:35:12.010 +highest standard deviation. + +00:35:14.070 --> 00:35:15.780 +So this is the MLE, the maximum + +00:35:15.780 --> 00:35:17.240 +likelihood estimate of the mean. + +00:35:17.240 --> 00:35:19.809 +It's just the sum of all the X's + +00:35:19.810 --> 00:35:21.850 +divided by the number of X's. + +00:35:21.850 --> 00:35:25.109 +Or, sorry, it's a sum over all the X's. + +00:35:26.970 --> 00:35:30.190 +For which Y n = K divided by the total + +00:35:30.190 --> 00:35:31.900 +number of times that Y n = K. + +00:35:32.790 --> 00:35:34.845 +Because I'm estimating the conditional + +00:35:34.845 --> 00:35:36.120 +conditional mean. + +00:35:36.760 --> 00:35:41.570 +So it's the sum over all the X's time. + +00:35:41.570 --> 00:35:44.060 +This will be where Y and equals K + +00:35:44.060 --> 00:35:45.670 +divided by the count of y = K. + +00:35:46.320 --> 00:35:48.050 +And they're staring deviation squared. + +00:35:48.050 --> 00:35:50.650 +Or the variance is the sum over all the + +00:35:50.650 --> 00:35:53.340 +differences of the X and the mean + +00:35:53.340 --> 00:35:56.890 +squared where Y and equals K divided by + +00:35:56.890 --> 00:35:58.890 +the number of times that y = K. + +00:35:59.640 --> 00:36:01.180 +And you have to estimate the mean + +00:36:01.180 --> 00:36:02.480 +before you Estimate the steering + +00:36:02.480 --> 00:36:02.950 +deviation. + +00:36:02.950 --> 00:36:05.100 +And if you take a statistics class, + +00:36:05.100 --> 00:36:07.980 +you'll probably like prove that this is + +00:36:07.980 --> 00:36:09.945 +an OK thing to do, that you're relying + +00:36:09.945 --> 00:36:11.720 +on one Estimate in order to get the + +00:36:11.720 --> 00:36:12.720 +other Estimate. + +00:36:12.720 --> 00:36:14.420 +But it does turn out it's OK. + +00:36:16.670 --> 00:36:20.220 +Alright, so in our homework for the + +00:36:20.220 --> 00:36:22.890 +temperature Regression, we're going to + +00:36:22.890 --> 00:36:26.095 +assume that Y minus XI is a Gaussian, + +00:36:26.095 --> 00:36:27.930 +so we have two continuous variables. + +00:36:28.900 --> 00:36:29.710 +So. + +00:36:30.940 --> 00:36:34.847 +The idea is that the temperature of + +00:36:34.847 --> 00:36:38.565 +some city on someday predicts the + +00:36:38.565 --> 00:36:41.530 +temperature of Cleveland on some other + +00:36:41.530 --> 00:36:41.850 +day. + +00:36:42.600 --> 00:36:44.600 +With some offset and some variance. + +00:36:45.830 --> 00:36:48.190 +And that is pretty easy to Model. + +00:36:48.190 --> 00:36:51.020 +So here's Sigma I is then the stair + +00:36:51.020 --> 00:36:53.770 +deviation of that offset Prediction and + +00:36:53.770 --> 00:36:54.910 +MU I is the offset. + +00:36:55.560 --> 00:36:58.230 +And I just have Y minus XI minus MU I + +00:36:58.230 --> 00:37:00.166 +squared here instead of Justice XI + +00:37:00.166 --> 00:37:02.590 +minus MU I squared, which would be if I + +00:37:02.590 --> 00:37:03.960 +just said XI is a Gaussian. + +00:37:05.170 --> 00:37:08.820 +And the mean is just why the sum of Y + +00:37:08.820 --> 00:37:11.603 +minus XI divided by north, where north + +00:37:11.603 --> 00:37:12.870 +is the total number of Samples. + +00:37:13.990 --> 00:37:14.820 +Because why? + +00:37:14.820 --> 00:37:16.618 +Is not discrete, so I'm not counting + +00:37:16.618 --> 00:37:20.100 +over certain over only values X where Y + +00:37:20.100 --> 00:37:21.625 +is equal to some value, I'm counting + +00:37:21.625 --> 00:37:22.550 +over all the values. + +00:37:23.410 --> 00:37:25.280 +And the Syrian deviation or their + +00:37:25.280 --> 00:37:28.590 +variance is Y minus XI minus MU I + +00:37:28.590 --> 00:37:29.630 +squared divided by north. + +00:37:30.480 --> 00:37:32.300 +And here's the Python. + +00:37:33.630 --> 00:37:35.840 +Here I just use the mean and steering + +00:37:35.840 --> 00:37:37.630 +deviation functions to get it, but it's + +00:37:37.630 --> 00:37:40.470 +also not a very long formula if I were + +00:37:40.470 --> 00:37:41.340 +to write it all out. + +00:37:44.020 --> 00:37:46.830 +And then X&Y were jointly Gaussian. + +00:37:46.830 --> 00:37:49.660 +So if I say that I need to jointly + +00:37:49.660 --> 00:37:52.850 +Model them, then one way to do it is + +00:37:52.850 --> 00:37:53.600 +by. + +00:37:54.460 --> 00:37:56.510 +By saying that probability of XI given + +00:37:56.510 --> 00:38:00.660 +Y is the joint probability of XI and Y. + +00:38:00.660 --> 00:38:03.070 +So now I have a 2 variable Gaussian + +00:38:03.070 --> 00:38:06.780 +with A2 variable mean and a two by two + +00:38:06.780 --> 00:38:07.900 +covariance matrix. + +00:38:08.920 --> 00:38:11.210 +Divided by the probability of Y, which + +00:38:11.210 --> 00:38:12.700 +is a 1D Gaussian. + +00:38:12.700 --> 00:38:14.636 +Just the Gaussian over probability of + +00:38:14.636 --> 00:38:14.999 +Y. + +00:38:15.000 --> 00:38:16.340 +And if you were to write out all the + +00:38:16.340 --> 00:38:18.500 +math for it would simplify into some + +00:38:18.500 --> 00:38:21.890 +other Gaussian equation, but it's + +00:38:21.890 --> 00:38:23.360 +easier to think about it this way. + +00:38:27.660 --> 00:38:28.140 +Alright. + +00:38:28.140 --> 00:38:31.660 +And then what if XI is continuous but + +00:38:31.660 --> 00:38:32.770 +it's not Gaussian? + +00:38:33.920 --> 00:38:35.750 +And why is discrete? + +00:38:35.750 --> 00:38:37.763 +There's one simple thing I can do is I + +00:38:37.763 --> 00:38:40.770 +can just first turn X into a discrete. + +00:38:40.860 --> 00:38:41.490 + + +00:38:42.280 --> 00:38:45.060 +Into a discrete function, so. + +00:38:46.810 --> 00:38:48.640 +For example if. + +00:38:49.590 --> 00:38:52.260 +Let me venture with my pen again, but. + +00:39:08.410 --> 00:39:08.810 +Can't do it. + +00:39:08.810 --> 00:39:09.170 +I want. + +00:39:15.140 --> 00:39:15.490 +OK. + +00:39:16.820 --> 00:39:20.930 +So for example, X has a range from. + +00:39:21.120 --> 00:39:22.130 +From zero to 1. + +00:39:22.810 --> 00:39:26.332 +That's the case for our intensities of + +00:39:26.332 --> 00:39:28.340 +the pixel, intensities of amnesty. + +00:39:29.180 --> 00:39:31.830 +I can just set a threshold for example + +00:39:31.830 --> 00:39:38.230 +of 0.5 and if X is greater than 05 then + +00:39:38.230 --> 00:39:40.369 +I'm going to say that it's equal to 1. + +00:39:41.030 --> 00:39:43.860 +NFX is less than five, then I'm going + +00:39:43.860 --> 00:39:45.050 +to say it's equal to 0. + +00:39:45.050 --> 00:39:46.440 +So now I turn my continuous + +00:39:46.440 --> 00:39:49.350 +distribution into a binary distribution + +00:39:49.350 --> 00:39:51.040 +and now I can just Estimate it using + +00:39:51.040 --> 00:39:52.440 +the Bernoulli equation. + +00:39:53.100 --> 00:39:54.910 +Or I could turn X into 10 different + +00:39:54.910 --> 00:39:57.280 +values by just multiplying X by 10 and + +00:39:57.280 --> 00:39:58.050 +taking the floor. + +00:39:58.050 --> 00:39:59.560 +So now the values are zero to 9. + +00:40:01.490 --> 00:40:04.150 +So that's one that's actually the one + +00:40:04.150 --> 00:40:06.110 +of the easiest way to deal with the + +00:40:06.110 --> 00:40:08.190 +continuous variable that's not + +00:40:08.190 --> 00:40:08.850 +Gaussian. + +00:40:12.900 --> 00:40:15.950 +Sometimes X will be like text, so for + +00:40:15.950 --> 00:40:18.800 +example it could be like blue, orange + +00:40:18.800 --> 00:40:19.430 +or green. + +00:40:20.080 --> 00:40:22.070 +And then you just need to Map those + +00:40:22.070 --> 00:40:25.390 +different text tokens into integers. + +00:40:25.390 --> 00:40:26.441 +So I might say blue. + +00:40:26.441 --> 00:40:28.654 +I'm going to say I'm going to Map blue + +00:40:28.654 --> 00:40:30.620 +into zero, orange into one, green into + +00:40:30.620 --> 00:40:32.580 +two, and then I can just Solve by + +00:40:32.580 --> 00:40:33.060 +counting. + +00:40:36.610 --> 00:40:38.830 +And then finally I need to also + +00:40:38.830 --> 00:40:40.380 +Estimate the probability of Y. + +00:40:41.060 --> 00:40:42.990 +One common thing to do is just to say + +00:40:42.990 --> 00:40:45.880 +that Y is equally likely to be all the + +00:40:45.880 --> 00:40:46.860 +possible labels. + +00:40:47.550 --> 00:40:49.440 +And that can be a good thing to do, + +00:40:49.440 --> 00:40:51.169 +because maybe our training distribution + +00:40:51.170 --> 00:40:52.870 +isn't even, but you don't think you're + +00:40:52.870 --> 00:40:54.310 +training distribution will be the same + +00:40:54.310 --> 00:40:55.790 +as the test distribution. + +00:40:55.790 --> 00:40:58.340 +So then you say that probability of Y + +00:40:58.340 --> 00:41:00.470 +is uniform even though it's not uniform + +00:41:00.470 --> 00:41:00.920 +in training. + +00:41:01.630 --> 00:41:03.530 +If it's uniform, you can just ignore it + +00:41:03.530 --> 00:41:05.910 +because it won't have any effect on + +00:41:05.910 --> 00:41:07.060 +which Y is most likely. + +00:41:07.980 --> 00:41:09.860 +FY is discrete and non uniform. + +00:41:09.860 --> 00:41:11.810 +You can just solve it by counting how + +00:41:11.810 --> 00:41:14.050 +many times is Y equal 1 divided by all + +00:41:14.050 --> 00:41:16.850 +my data is the probability of Y equal + +00:41:16.850 --> 00:41:17.070 +1. + +00:41:17.790 --> 00:41:19.450 +If it's continuous, you can Model it as + +00:41:19.450 --> 00:41:21.660 +a Gaussian or chop it up into bins and + +00:41:21.660 --> 00:41:23.000 +then turn it into a classification + +00:41:23.000 --> 00:41:23.360 +problem. + +00:41:25.690 --> 00:41:26.050 +Right. + +00:41:28.290 --> 00:41:31.550 +So I'll give you your minute or two, + +00:41:31.550 --> 00:41:32.230 +Stretch break. + +00:41:32.230 --> 00:41:33.650 +But I want you to think about this + +00:41:33.650 --> 00:41:34.370 +while you do that. + +00:41:35.390 --> 00:41:38.100 +So suppose I want to classify a fruit + +00:41:38.100 --> 00:41:40.230 +based on description and my Features + +00:41:40.230 --> 00:41:42.389 +are weight, color, shape and whether + +00:41:42.390 --> 00:41:44.190 +it's a hard whether the outside is + +00:41:44.190 --> 00:41:44.470 +hard. + +00:41:45.330 --> 00:41:47.960 +And so first, here's some examples of + +00:41:47.960 --> 00:41:49.100 +those Features. + +00:41:49.100 --> 00:41:50.750 +See if you can figure out which fruit + +00:41:50.750 --> 00:41:51.990 +correspond to these Features. + +00:41:52.630 --> 00:41:56.150 +And second, what might be a good set of + +00:41:56.150 --> 00:41:58.080 +models to use for probability of XI + +00:41:58.080 --> 00:41:59.730 +given fruit for those four Features? + +00:42:01.210 --> 00:42:03.620 +So you have two minutes to think about + +00:42:03.620 --> 00:42:05.630 +it and Oregon Stretch or use the + +00:42:05.630 --> 00:42:07.240 +bathroom or check your e-mail or + +00:42:07.240 --> 00:42:07.620 +whatever. + +00:44:24.040 --> 00:44:24.730 +Alright. + +00:44:26.640 --> 00:44:31.100 +So first, what is the top 1.5 pounds + +00:44:31.100 --> 00:44:31.640 +red round? + +00:44:31.640 --> 00:44:33.750 +Yes, OK, good. + +00:44:33.750 --> 00:44:34.870 +That's what I was thinking. + +00:44:34.870 --> 00:44:37.930 +What's the 2nd 115 pounds? + +00:44:39.070 --> 00:44:39.810 +Avocado. + +00:44:39.810 --> 00:44:41.260 +That's a huge avocado. + +00:44:43.770 --> 00:44:44.660 +What is it? + +00:44:46.290 --> 00:44:48.090 +Watermelon watermelons, what I was + +00:44:48.090 --> 00:44:48.450 +thinking. + +00:44:49.170 --> 00:44:52.140 +.1 pounds purple round and not hard. + +00:44:53.330 --> 00:44:54.980 +I was thinking of a Grape. + +00:44:54.980 --> 00:44:55.980 +OK, good. + +00:44:57.480 --> 00:44:58.900 +There wasn't really, there wasn't + +00:44:58.900 --> 00:45:00.160 +necessarily a right answer. + +00:45:00.160 --> 00:45:01.790 +It's just kind of what I was thinking. + +00:45:02.800 --> 00:45:05.642 +Alright, and then how do you Model the + +00:45:05.642 --> 00:45:07.700 +probability of the feature given the + +00:45:07.700 --> 00:45:08.450 +fruit for each of these? + +00:45:08.450 --> 00:45:09.550 +So let's say the weight. + +00:45:09.550 --> 00:45:11.172 +What would be a good model for + +00:45:11.172 --> 00:45:13.270 +probability of XI given the label? + +00:45:15.080 --> 00:45:17.420 +Gaussian would, Gaussian would probably + +00:45:17.420 --> 00:45:18.006 +be a good choice. + +00:45:18.006 --> 00:45:19.820 +It has each of these probably has some + +00:45:19.820 --> 00:45:21.250 +expectation, maybe a Gaussian + +00:45:21.250 --> 00:45:22.130 +distribution around it. + +00:45:24.000 --> 00:45:26.490 +Alright, what about the color red, + +00:45:26.490 --> 00:45:27.315 +green, purple? + +00:45:27.315 --> 00:45:28.440 +What could I do for that? + +00:45:31.440 --> 00:45:35.610 +So I could use a multinomial so I can + +00:45:35.610 --> 00:45:37.210 +just turn it into discrete very + +00:45:37.210 --> 00:45:39.410 +discrete numbers, integer numbers and + +00:45:39.410 --> 00:45:41.480 +then count and the shape. + +00:45:50.470 --> 00:45:52.470 +So if there's assuming that there's + +00:45:52.470 --> 00:45:54.470 +other shapes, I don't know if there are + +00:45:54.470 --> 00:45:55.880 +star fruit for example. + +00:45:56.790 --> 00:45:58.940 +And then multinomial. + +00:45:58.940 --> 00:46:00.640 +But either way I'll turn it in discrete + +00:46:00.640 --> 00:46:04.090 +variables and count and the yes nodes. + +00:46:05.540 --> 00:46:07.010 +So that will be Binomial. + +00:46:08.240 --> 00:46:08.540 +OK. + +00:46:14.840 --> 00:46:18.500 +All right, so now we know how to + +00:46:18.500 --> 00:46:20.770 +Estimate probability of X given Y. + +00:46:20.770 --> 00:46:23.065 +Now after I go through all that work on + +00:46:23.065 --> 00:46:25.178 +the training data and I get new test + +00:46:25.178 --> 00:46:25.512 +sample. + +00:46:25.512 --> 00:46:27.900 +Now I want to know what's the most + +00:46:27.900 --> 00:46:29.620 +likely label of that test sample. + +00:46:31.200 --> 00:46:31.660 +So. + +00:46:32.370 --> 00:46:33.860 +I can write this in two ways. + +00:46:33.860 --> 00:46:36.615 +One is I can write Y is the argmax over + +00:46:36.615 --> 00:46:38.735 +the product of probability of XI given + +00:46:38.735 --> 00:46:39.959 +Y times probability of Y. + +00:46:40.990 --> 00:46:44.334 +Or I can write it as the argmax of the + +00:46:44.334 --> 00:46:46.718 +log of that, which is just the argmax + +00:46:46.718 --> 00:46:48.970 +of Y of the sum over I of log of + +00:46:48.970 --> 00:46:50.904 +probability of XI given Yi plus log of + +00:46:50.904 --> 00:46:51.599 +probability of Y. + +00:46:52.570 --> 00:46:55.130 +And I can do that because the thing + +00:46:55.130 --> 00:46:57.798 +that maximizes X also maximizes log of + +00:46:57.798 --> 00:46:59.280 +X and vice versa. + +00:46:59.280 --> 00:47:01.910 +And that's actually a really useful + +00:47:01.910 --> 00:47:04.270 +property because often the logs are + +00:47:04.270 --> 00:47:05.745 +probabilities are a lot simpler. + +00:47:05.745 --> 00:47:08.790 +And for example, if I took for example + +00:47:08.790 --> 00:47:10.434 +at the Gaussian, if I take the log of + +00:47:10.434 --> 00:47:11.950 +the Gaussian, then it just becomes a + +00:47:11.950 --> 00:47:12.760 +squared term. + +00:47:13.640 --> 00:47:16.400 +And the other thing is that these + +00:47:16.400 --> 00:47:18.350 +probability of Xis might be. + +00:47:18.470 --> 00:47:21.553 +If I have a lot of them, if I have like + +00:47:21.553 --> 00:47:23.723 +500 of them and they're on average like + +00:47:23.723 --> 00:47:26.320 +.1, that would be like .1 to the 500, + +00:47:26.320 --> 00:47:27.530 +which is going to go outside in + +00:47:27.530 --> 00:47:28.690 +numerical precision. + +00:47:28.690 --> 00:47:30.740 +So if you try to Compute this product + +00:47:30.740 --> 00:47:32.290 +directly, you're probably going to get + +00:47:32.290 --> 00:47:34.470 +0 or some kind of wonky value. + +00:47:35.190 --> 00:47:37.320 +And so it's much better to take the sum + +00:47:37.320 --> 00:47:39.265 +of the logs than to take the product of + +00:47:39.265 --> 00:47:40.060 +the probabilities. + +00:47:42.650 --> 00:47:44.290 +Right, so, but I can compute the + +00:47:44.290 --> 00:47:45.830 +probability of X&Y or the log + +00:47:45.830 --> 00:47:48.004 +probability of X&Y for each value of Y + +00:47:48.004 --> 00:47:49.630 +and then choose the value with maximum + +00:47:49.630 --> 00:47:50.240 +likelihood. + +00:47:50.240 --> 00:47:51.686 +That will work in the case of the + +00:47:51.686 --> 00:47:53.409 +digits because I only have 10 digits. + +00:47:54.420 --> 00:47:56.940 +And so I can check for each possible + +00:47:56.940 --> 00:48:00.365 +Digit, how likely is the sum of log + +00:48:00.365 --> 00:48:01.958 +probability of XI given Yi plus + +00:48:01.958 --> 00:48:03.770 +probability log probability of Y. + +00:48:03.770 --> 00:48:06.980 +And then I choose the Digit Digit label + +00:48:06.980 --> 00:48:08.570 +that makes this most likely. + +00:48:11.240 --> 00:48:12.580 +That's pretty simple. + +00:48:12.580 --> 00:48:14.110 +In the case of Y is discrete. + +00:48:14.900 --> 00:48:16.415 +And again, I just want to emphasize + +00:48:16.415 --> 00:48:18.983 +that this thing of turning product of + +00:48:18.983 --> 00:48:21.070 +probabilities into a sum of log + +00:48:21.070 --> 00:48:23.250 +probabilities is really, really widely + +00:48:23.250 --> 00:48:23.760 +used. + +00:48:23.760 --> 00:48:27.610 +Almost anytime you Solve for anything + +00:48:27.610 --> 00:48:29.140 +with probabilities, it involves that + +00:48:29.140 --> 00:48:29.380 +step. + +00:48:31.840 --> 00:48:34.420 +Now if Y is continuous, it's a bit more + +00:48:34.420 --> 00:48:36.610 +complicated and I. + +00:48:37.440 --> 00:48:39.890 +So I have the derivation here for you. + +00:48:39.890 --> 00:48:42.166 +So this is for the case. + +00:48:42.166 --> 00:48:44.859 +I'm going to use as an example the case + +00:48:44.860 --> 00:48:47.470 +where I'm modeling probability of Y + +00:48:47.470 --> 00:48:51.400 +minus XI of 1 dimensional Gaussian. + +00:48:53.280 --> 00:48:56.260 +And anytime you solve this kind of + +00:48:56.260 --> 00:48:58.320 +thing you're going to go through, you + +00:48:58.320 --> 00:48:59.580 +would go through the same derivation. + +00:48:59.580 --> 00:49:00.280 +If it's not. + +00:49:00.280 --> 00:49:03.180 +Just like a simple matter of if you + +00:49:03.180 --> 00:49:05.000 +don't have discrete wise, if you have + +00:49:05.000 --> 00:49:06.360 +continuous wise, then you have to find + +00:49:06.360 --> 00:49:08.320 +the Y that actually maximizes this + +00:49:08.320 --> 00:49:10.760 +because you can't check all possible + +00:49:10.760 --> 00:49:12.310 +values of a continuous variable. + +00:49:14.180 --> 00:49:15.390 +So it's not. + +00:49:16.540 --> 00:49:17.451 +It's a lot. + +00:49:17.451 --> 00:49:18.362 +It's a lot. + +00:49:18.362 --> 00:49:20.350 +It's a fair number of equations, but + +00:49:20.350 --> 00:49:23.420 +it's not anything super complicated. + +00:49:23.420 --> 00:49:24.940 +Let me see if I can get my cursor up + +00:49:24.940 --> 00:49:25.960 +there again, OK? + +00:49:26.710 --> 00:49:29.560 +Alright, so first I take the partial + +00:49:29.560 --> 00:49:32.526 +derivative of the log probability of + +00:49:32.526 --> 00:49:34.780 +X&Y with respect to Y and set it equal + +00:49:34.780 --> 00:49:35.190 +to 0. + +00:49:35.190 --> 00:49:36.890 +So you might remember from calculus + +00:49:36.890 --> 00:49:38.720 +like if you want to find the min or Max + +00:49:38.720 --> 00:49:39.580 +of some value. + +00:49:40.290 --> 00:49:43.109 +Then take the partial with respect to + +00:49:43.110 --> 00:49:44.750 +some variable. + +00:49:44.750 --> 00:49:47.340 +You take the partial derivative with + +00:49:47.340 --> 00:49:48.800 +respect to that variable and set it + +00:49:48.800 --> 00:49:49.539 +equal to 0. + +00:49:50.680 --> 00:49:51.360 +And. + +00:49:53.080 --> 00:49:55.020 +So here I did that. + +00:49:55.020 --> 00:49:58.100 +Now I've plugged in this Gaussian + +00:49:58.100 --> 00:50:00.200 +distribution and taken the log. + +00:50:01.050 --> 00:50:02.510 +And I kind of like there's some + +00:50:02.510 --> 00:50:04.020 +invisible steps here, because there's + +00:50:04.020 --> 00:50:06.410 +some terms like the log of one over + +00:50:06.410 --> 00:50:07.940 +square of 2π Sigma. + +00:50:08.580 --> 00:50:10.069 +That just don't. + +00:50:10.069 --> 00:50:12.290 +Those terms don't matter because they + +00:50:12.290 --> 00:50:13.080 +don't involve Y. + +00:50:13.080 --> 00:50:14.743 +So the partial derivative of those + +00:50:14.743 --> 00:50:16.215 +terms with respect to Y is 0. + +00:50:16.215 --> 00:50:19.090 +So I just didn't include them. + +00:50:19.750 --> 00:50:21.815 +So these are the terms that include Y + +00:50:21.815 --> 00:50:23.590 +and I've already taken the log. + +00:50:23.590 --> 00:50:25.550 +This was originally east to the -, 1 + +00:50:25.550 --> 00:50:27.839 +half whatever is shown here, and the + +00:50:27.839 --> 00:50:30.360 +log of X of X is equal to X. + +00:50:31.840 --> 00:50:33.490 +And so I get this guy. + +00:50:34.450 --> 00:50:36.530 +Now I broke it out into different + +00:50:36.530 --> 00:50:39.320 +terms, so I did the quadratic of Y + +00:50:39.320 --> 00:50:41.190 +minus XI minus MU I ^2. + +00:50:42.420 --> 00:50:44.100 +Mainly so that I don't have to use the + +00:50:44.100 --> 00:50:45.620 +chain rule and I can keep my + +00:50:45.620 --> 00:50:46.740 +derivatives really Simple. + +00:50:47.830 --> 00:50:51.959 +So here I just broke that out to y ^2 y + +00:50:51.960 --> 00:50:54.130 +axis YMUI. + +00:50:54.130 --> 00:50:55.530 +And again, I don't need to worry about + +00:50:55.530 --> 00:50:57.779 +the MU I squared over Sigma I squared + +00:50:57.780 --> 00:50:59.750 +because it doesn't involve Y so I just + +00:50:59.750 --> 00:51:00.230 +left it out. + +00:51:02.140 --> 00:51:03.990 +I. + +00:51:04.100 --> 00:51:07.021 +Take the derivative with respect to Y. + +00:51:07.021 --> 00:51:09.468 +So the derivative of y ^2 is 2 Y. + +00:51:09.468 --> 00:51:10.976 +So this half goes away. + +00:51:10.976 --> 00:51:14.080 +Derivative of YX is just X. + +00:51:15.070 --> 00:51:18.000 +So this should be a subscript I. + +00:51:18.730 --> 00:51:21.120 +And then I did the same for these guys + +00:51:21.120 --> 00:51:21.330 +here. + +00:51:22.500 --> 00:51:25.740 +It's just basic algebra, so I just try + +00:51:25.740 --> 00:51:27.610 +to group the terms that involve Y and + +00:51:27.610 --> 00:51:29.480 +the terms that don't involve Yi, put + +00:51:29.480 --> 00:51:30.840 +the terms that don't involve Y and the + +00:51:30.840 --> 00:51:33.370 +right side, and then finally I divide + +00:51:33.370 --> 00:51:36.830 +the coefficient of Y and I get this guy + +00:51:36.830 --> 00:51:37.150 +here. + +00:51:38.030 --> 00:51:41.269 +So at the end Y is equal to 1 over the + +00:51:41.270 --> 00:51:44.408 +sum over all the features of 1 / sqrt. + +00:51:44.408 --> 00:51:46.690 +I mean one over Sigma I ^2. + +00:51:47.420 --> 00:51:50.580 +Plus one over Sigma y ^2 which is the + +00:51:50.580 --> 00:51:52.160 +standard deviation of the Prior of Y. + +00:51:52.160 --> 00:51:53.906 +Or if I just assumed uniform likelihood + +00:51:53.906 --> 00:51:55.520 +of Yi wouldn't need that term. + +00:51:56.610 --> 00:51:59.400 +And then that's times the sum over all + +00:51:59.400 --> 00:52:02.700 +the features of that feature value. + +00:52:02.700 --> 00:52:03.930 +This should be subscript I. + +00:52:04.940 --> 00:52:10.430 +Plus MU I divided by Sigma I ^2 plus mu + +00:52:10.430 --> 00:52:13.811 +Y, the Prior mean of Y divided by Sigma + +00:52:13.811 --> 00:52:14.539 +y ^2. + +00:52:16.150 --> 00:52:18.940 +And so this is just a, it's actually + +00:52:18.940 --> 00:52:19.849 +just a weighted. + +00:52:19.850 --> 00:52:22.823 +If you say that one over Sigma I + +00:52:22.823 --> 00:52:26.035 +squared is Wei, it's like a weight for + +00:52:26.035 --> 00:52:27.565 +that prediction of the ith feature. + +00:52:27.565 --> 00:52:29.830 +This is just a weighted average of the + +00:52:29.830 --> 00:52:31.720 +predictions from all the Features + +00:52:31.720 --> 00:52:33.250 +that's weighted by one over the + +00:52:33.250 --> 00:52:35.573 +steering deviation squared or one over + +00:52:35.573 --> 00:52:36.190 +the variance. + +00:52:37.590 --> 00:52:40.421 +And so I have one over the sum over I + +00:52:40.421 --> 00:52:45.683 +of WI plus WY times, the sum X plus mu + +00:52:45.683 --> 00:52:49.722 +I XI plus MU I times, Wei plus mu Y + +00:52:49.722 --> 00:52:50.100 +times. + +00:52:50.100 --> 00:52:50.670 +Why? + +00:52:51.630 --> 00:52:53.240 +Amy sounds similar, unfortunately. + +00:52:54.780 --> 00:52:56.430 +So it's just the weighted average of + +00:52:56.430 --> 00:52:57.910 +all the predictions of the individual + +00:52:57.910 --> 00:52:58.174 +features. + +00:52:58.174 --> 00:53:00.093 +And it makes sense that it kind of + +00:53:00.093 --> 00:53:01.624 +makes sense intuitively that the weight + +00:53:01.624 --> 00:53:02.650 +is 1 over the variance. + +00:53:02.650 --> 00:53:04.490 +So if you have really high variance, + +00:53:04.490 --> 00:53:05.790 +then the weight is small. + +00:53:05.790 --> 00:53:08.155 +So if, for example, maybe the + +00:53:08.155 --> 00:53:09.839 +temperature in Sacramento is a really + +00:53:09.840 --> 00:53:11.513 +bad predictor for the temperature in + +00:53:11.513 --> 00:53:12.984 +Cleveland, so it will have high + +00:53:12.984 --> 00:53:14.840 +variance and it gets a little weight, + +00:53:14.840 --> 00:53:16.460 +while the temperature in Cleveland the + +00:53:16.460 --> 00:53:19.130 +previous day is much more highly + +00:53:19.130 --> 00:53:20.849 +predictive, has lower variance, so + +00:53:20.850 --> 00:53:21.639 +it'll get more weight. + +00:53:32.280 --> 00:53:35.380 +So let me pause here. + +00:53:35.380 --> 00:53:38.690 +So any questions about? + +00:53:39.670 --> 00:53:43.255 +Estimating the likelihoods P of X given + +00:53:43.255 --> 00:53:47.970 +Y, or solving for the Y that makes. + +00:53:47.970 --> 00:53:49.880 +That's most likely given your + +00:53:49.880 --> 00:53:50.500 +likelihoods. + +00:53:52.460 --> 00:53:54.470 +And obviously if I'm happy to work + +00:53:54.470 --> 00:53:56.610 +through this in office hours as well in + +00:53:56.610 --> 00:53:59.940 +the TAS should also if you want to like + +00:53:59.940 --> 00:54:01.100 +spend more time working through the + +00:54:01.100 --> 00:54:01.530 +equations. + +00:54:03.920 --> 00:54:04.930 +I just want to pause. + +00:54:04.930 --> 00:54:07.830 +I know it's a lot of math to soak up. + +00:54:09.870 --> 00:54:13.260 +And really, it's not that memorizing + +00:54:13.260 --> 00:54:14.370 +these things isn't important. + +00:54:14.370 --> 00:54:15.860 +It's really the process that you just + +00:54:15.860 --> 00:54:17.385 +set the partial derivative with respect + +00:54:17.385 --> 00:54:20.140 +to Y, set it to zero, and then you do + +00:54:20.140 --> 00:54:20.540 +the. + +00:54:21.250 --> 00:54:23.120 +Do the partial derivative and solve the + +00:54:23.120 --> 00:54:23.510 +algebra. + +00:54:26.700 --> 00:54:28.050 +All right, I'll go on then. + +00:54:28.050 --> 00:54:31.990 +So far, this is pure maximum likelihood + +00:54:31.990 --> 00:54:32.530 +estimation. + +00:54:32.530 --> 00:54:34.920 +I'm not, I'm not imposing any kinds of + +00:54:34.920 --> 00:54:36.470 +Priors over my parameters. + +00:54:37.570 --> 00:54:39.600 +In practice, you do want to impose a + +00:54:39.600 --> 00:54:41.010 +Prior in your parameters to make sure + +00:54:41.010 --> 00:54:42.220 +you don't have any zeros. + +00:54:43.750 --> 00:54:46.380 +Otherwise, like if some in the digits + +00:54:46.380 --> 00:54:48.809 +case for example the test sample had a + +00:54:48.810 --> 00:54:50.470 +dot in an unlikely place. + +00:54:50.470 --> 00:54:52.662 +If I had just had like a one and some + +00:54:52.662 --> 00:54:54.030 +unlikely pixel, all the probabilities + +00:54:54.030 --> 00:54:55.630 +would be 0 and you wouldn't know what + +00:54:55.630 --> 00:54:57.620 +the label is because of that one stupid + +00:54:57.620 --> 00:54:57.970 +pixel. + +00:54:58.730 --> 00:55:01.040 +So you want to have some kind of Prior? + +00:55:01.730 --> 00:55:03.425 +To avoid these zero probabilities. + +00:55:03.425 --> 00:55:06.260 +So the most common case if you're + +00:55:06.260 --> 00:55:08.760 +estimating a distribution of discrete + +00:55:08.760 --> 00:55:10.430 +variables like a multinomial or + +00:55:10.430 --> 00:55:13.010 +Binomial, is to just initialize with + +00:55:13.010 --> 00:55:13.645 +some count. + +00:55:13.645 --> 00:55:16.180 +So you just say for example alpha + +00:55:16.180 --> 00:55:16.880 +equals one. + +00:55:17.610 --> 00:55:20.110 +And now I say the probability of X I = + +00:55:20.110 --> 00:55:21.620 +V given y = K. + +00:55:22.400 --> 00:55:24.950 +Is Alpha plus the count of how many + +00:55:24.950 --> 00:55:27.740 +times XI equals V and y = K. + +00:55:28.690 --> 00:55:31.865 +Divided by the all the different values + +00:55:31.865 --> 00:55:35.300 +of alpha plus account of XI equals that + +00:55:35.300 --> 00:55:37.610 +value in y = K probably for clarity I + +00:55:37.610 --> 00:55:39.700 +should have used something other than B + +00:55:39.700 --> 00:55:41.630 +in the denominator, but hopefully + +00:55:41.630 --> 00:55:42.230 +that's clear enough. + +00:55:43.060 --> 00:55:46.170 +Here's the and then here's the Python + +00:55:46.170 --> 00:55:47.070 +for that, so it's just. + +00:55:47.880 --> 00:55:50.350 +Sum of all the values where XI equals V + +00:55:50.350 --> 00:55:52.470 +and y = K Plus some alpha. + +00:55:53.300 --> 00:55:54.980 +So if alpha equals zero, then I don't + +00:55:54.980 --> 00:55:55.710 +have any Prior. + +00:55:56.840 --> 00:56:00.450 +And then I'm just dividing by the sum + +00:56:00.450 --> 00:56:04.270 +of times at y = K and there will be. + +00:56:04.850 --> 00:56:06.540 +The number of alphas will be equal to + +00:56:06.540 --> 00:56:08.150 +the number of different values, so this + +00:56:08.150 --> 00:56:10.510 +is like a little bit of a shortcut, but + +00:56:10.510 --> 00:56:11.330 +it's the same thing. + +00:56:12.860 --> 00:56:14.760 +If I have a continuous variable and + +00:56:14.760 --> 00:56:15.060 +I've. + +00:56:15.730 --> 00:56:17.010 +Modeled it with the Gaussian. + +00:56:17.010 --> 00:56:18.470 +Then the usual thing to do is just to + +00:56:18.470 --> 00:56:20.180 +add a small value to your steering + +00:56:20.180 --> 00:56:21.420 +deviation or your variance. + +00:56:22.110 --> 00:56:24.320 +And you might want to make that value + +00:56:24.320 --> 00:56:27.650 +if N is unknown, then make it dependent + +00:56:27.650 --> 00:56:29.300 +on north so that if you have a huge + +00:56:29.300 --> 00:56:31.395 +number of samples then the effect of + +00:56:31.395 --> 00:56:33.880 +the Prior will go down, which is what + +00:56:33.880 --> 00:56:34.170 +you want. + +00:56:36.140 --> 00:56:39.513 +So for example, you can say that the + +00:56:39.513 --> 00:56:41.990 +stern deviation is whatever this + +00:56:41.990 --> 00:56:44.770 +whatever the MLE estimate of the stern + +00:56:44.770 --> 00:56:47.340 +deviation is, plus some small value + +00:56:47.340 --> 00:56:49.730 +sqrt 1 over the length of north. + +00:56:50.420 --> 00:56:51.350 +Of X, sorry. + +00:57:00.440 --> 00:57:02.670 +So what the Prior does is it. + +00:57:02.810 --> 00:57:05.995 +In the case of the discrete variables, + +00:57:05.995 --> 00:57:09.110 +the Prior is trying to push your + +00:57:09.110 --> 00:57:11.152 +Estimate towards a uniform likelihood. + +00:57:11.152 --> 00:57:13.000 +In fact, in both cases it's pushing it + +00:57:13.000 --> 00:57:14.280 +towards a uniform likelihood. + +00:57:15.400 --> 00:57:18.670 +So if you had a really large alpha, + +00:57:18.670 --> 00:57:20.550 +then let's say. + +00:57:22.090 --> 00:57:23.440 +Let's say that. + +00:57:24.620 --> 00:57:25.850 +Or I don't know if I can think of + +00:57:25.850 --> 00:57:26.170 +something. + +00:57:28.140 --> 00:57:29.550 +Let's say you have a population of + +00:57:29.550 --> 00:57:30.900 +students and you're trying to estimate + +00:57:30.900 --> 00:57:32.510 +the probability that a student is male. + +00:57:33.520 --> 00:57:36.570 +If I say alpha equals 1000, then I'm + +00:57:36.570 --> 00:57:37.860 +going to need like an awful lot of + +00:57:37.860 --> 00:57:40.156 +students before I budge very far from a + +00:57:40.156 --> 00:57:42.070 +5050 chance that a student is male or + +00:57:42.070 --> 00:57:42.620 +female. + +00:57:42.620 --> 00:57:44.057 +Because I'll start with saying there's + +00:57:44.057 --> 00:57:46.213 +1000 males and 1000 females, and then + +00:57:46.213 --> 00:57:48.676 +I'll count all the males and add them + +00:57:48.676 --> 00:57:50.832 +to 1000, count all the females, add + +00:57:50.832 --> 00:57:53.370 +them to 1000, and then I would take the + +00:57:53.370 --> 00:57:55.210 +male plus 1000 count and divide it by + +00:57:55.210 --> 00:57:57.660 +2000 plus the total population. + +00:57:59.130 --> 00:58:00.860 +If Alpha is 0, then I'm going to get + +00:58:00.860 --> 00:58:03.410 +just my raw empirical Estimate. + +00:58:03.410 --> 00:58:06.810 +So if I had like 3 students and I say + +00:58:06.810 --> 00:58:09.090 +alpha equals zero, and I have two males + +00:58:09.090 --> 00:58:11.140 +and a female, then I'll say 2/3 of them + +00:58:11.140 --> 00:58:11.550 +are male. + +00:58:12.410 --> 00:58:14.670 +If I say alpha is 1 and I have two + +00:58:14.670 --> 00:58:17.110 +males and a female, then I would say + +00:58:17.110 --> 00:58:20.490 +that my probability of male is 3 / 5 + +00:58:20.490 --> 00:58:24.100 +because it's 2 + 1 / 3 + 2. + +00:58:27.060 --> 00:58:28.330 +Their deviation it's the same. + +00:58:28.330 --> 00:58:30.240 +It's like trying to just broaden your + +00:58:30.240 --> 00:58:32.600 +variance from what you would Estimate + +00:58:32.600 --> 00:58:33.580 +directly from the data. + +00:58:36.500 --> 00:58:39.260 +So I think I will not ask you all these + +00:58:39.260 --> 00:58:41.210 +probabilities because they're kind of + +00:58:41.210 --> 00:58:43.220 +you've shown the ability to count + +00:58:43.220 --> 00:58:44.810 +before mostly. + +00:58:46.550 --> 00:58:47.640 +And. + +00:58:47.850 --> 00:58:50.060 +So here's for example, the probability + +00:58:50.060 --> 00:58:54.509 +of X 1 = 0 and y = 0 is 2 out of four. + +00:58:54.510 --> 00:58:56.050 +I can get that just by looking down + +00:58:56.050 --> 00:58:56.670 +these rows. + +00:58:56.670 --> 00:58:58.870 +It takes a little bit of time, but + +00:58:58.870 --> 00:59:02.786 +there's four times that y = 0 and out + +00:59:02.786 --> 00:59:06.660 +of those two times X 1 = 0 and so this + +00:59:06.660 --> 00:59:07.440 +is 2 out of four. + +00:59:08.090 --> 00:59:08.930 +And the same. + +00:59:08.930 --> 00:59:11.260 +I can use the same counting method to + +00:59:11.260 --> 00:59:13.120 +get all of these other probabilities + +00:59:13.120 --> 00:59:13.410 +here. + +00:59:15.770 --> 00:59:19.450 +So just to check that everyone's awake, + +00:59:19.450 --> 00:59:22.970 +if I, what is the probability of Y? + +00:59:23.840 --> 00:59:27.370 +And X 1 = 1 and X 2 = 1. + +00:59:28.500 --> 00:59:30.019 +So can you get it from? + +00:59:30.019 --> 00:59:32.560 +Can you get it from this guy under an + +00:59:32.560 --> 00:59:33.450 +independence? + +00:59:33.450 --> 00:59:35.670 +So get it from this under an under an I + +00:59:35.670 --> 00:59:36.540 +Bayes assumption. + +00:59:41.350 --> 00:59:43.240 +Let's say I should say probability of Y + +00:59:43.240 --> 00:59:43.860 +equal 1. + +00:59:45.380 --> 00:59:47.910 +Probability of y = 1 given X 1 = 1 and + +00:59:47.910 --> 00:59:48.930 +X 2 = 1. + +00:59:57.500 --> 01:00:00.560 +And you don't worry about simplifying + +01:00:00.560 --> 01:00:02.610 +your numerator and denominator. + +01:00:03.530 --> 01:00:05.110 +What are the things that get multiplied + +01:00:05.110 --> 01:00:05.610 +together? + +01:00:10.460 --> 01:00:14.350 +Not sort of, partly that's in there. + +01:00:15.220 --> 01:00:17.880 +Raise your hand if you think the + +01:00:17.880 --> 01:00:18.560 +answer. + +01:00:19.550 --> 01:00:21.130 +I just want to give everyone time. + +01:00:24.650 --> 01:00:27.962 +But I mean probability of y = 1 given X + +01:00:27.962 --> 01:00:29.960 +1 = 1 and X 2 = 1. + +01:00:39.830 --> 01:00:41.220 +A Naive Bayes assumption. + +01:01:24.310 --> 01:01:25.800 +The raise your hand if you. + +01:01:26.490 --> 01:01:27.030 +Finished. + +01:01:56.450 --> 01:01:57.740 +But don't tell me the answer yet. + +01:02:18.470 --> 01:02:19.260 +Equals one. + +01:02:23.210 --> 01:02:23.420 +Alright. + +01:02:23.420 --> 01:02:24.830 +Did anybody get it yet? + +01:02:24.830 --> 01:02:25.950 +Raise your hand if you did. + +01:02:25.950 --> 01:02:26.910 +I just don't want to. + +01:02:28.170 --> 01:02:29.110 +Give it too early. + +01:03:46.370 --> 01:03:46.960 +Alright. + +01:03:48.170 --> 01:03:52.029 +Example, some people have gotten it, so + +01:03:52.030 --> 01:03:53.950 +let me I'll start going through it. + +01:03:53.950 --> 01:03:55.480 +All right, so the Naive Bayes + +01:03:55.480 --> 01:03:56.005 +assumption. + +01:03:56.005 --> 01:03:57.760 +So this would be. + +01:03:58.060 --> 01:03:58.250 +OK. + +01:04:00.690 --> 01:04:02.960 +OK, probability it's actually my touch + +01:04:02.960 --> 01:04:03.230 +screen. + +01:04:03.230 --> 01:04:04.400 +I think is kind of broken. + +01:04:05.250 --> 01:04:09.560 +Probability of X1 given Y times + +01:04:09.560 --> 01:04:14.815 +probability X2 given Y sorry equals + +01:04:14.815 --> 01:04:15.200 +one. + +01:04:16.630 --> 01:04:19.050 +Times probability of Y equal 1. + +01:04:19.910 --> 01:04:21.950 +Right, so it's the product of the + +01:04:21.950 --> 01:04:23.180 +probabilities of the Features. + +01:04:23.180 --> 01:04:24.730 +Give them label times the probability + +01:04:24.730 --> 01:04:25.240 +of the label. + +01:04:26.500 --> 01:04:29.990 +And so that will be probability of XYX. + +01:04:31.030 --> 01:04:32.819 +1 = 1. + +01:04:33.850 --> 01:04:37.317 +Given probability of Yi mean given y = + +01:04:37.317 --> 01:04:38.260 +1 is 3/4. + +01:04:42.110 --> 01:04:46.010 +And probably the X 2 = 1 given y = 1 is + +01:04:46.010 --> 01:04:46.750 +3/4. + +01:04:49.250 --> 01:04:52.550 +And the probability that y = 1 is two + +01:04:52.550 --> 01:04:53.940 +quarters or 1/2. + +01:04:58.570 --> 01:05:00.180 +So it's 930 seconds. + +01:05:01.120 --> 01:05:01.390 +Right. + +01:05:02.580 --> 01:05:05.846 +And the probability that y = 0 given X + +01:05:05.846 --> 01:05:08.059 +1 = 1 and Y 1 = 1. + +01:05:09.800 --> 01:05:11.840 +I mean sorry, the probability of y = 0 + +01:05:11.840 --> 01:05:14.480 +given the X is equal equal 1. + +01:05:15.620 --> 01:05:16.770 +Is. + +01:05:18.600 --> 01:05:19.190 +Let's see. + +01:05:20.250 --> 01:05:23.780 +So that would be 2 fourths times 2 + +01:05:23.780 --> 01:05:24.300 +fourths. + +01:05:25.180 --> 01:05:26.320 +Times 2 fourths. + +01:05:27.260 --> 01:05:31.300 +So if X 1 = 1 and X2 equal 1, then it's + +01:05:31.300 --> 01:05:33.540 +more likely that Y is equal to 1 than + +01:05:33.540 --> 01:05:35.070 +that Y is equal to 0. + +01:05:41.720 --> 01:05:46.750 +If I had if I use my Prior, this is how + +01:05:46.750 --> 01:05:48.055 +the probabilities would change. + +01:05:48.055 --> 01:05:51.060 +So if I say alpha equals one, you can + +01:05:51.060 --> 01:05:52.900 +see that the probabilities get less + +01:05:52.900 --> 01:05:53.510 +Peaky. + +01:05:53.510 --> 01:05:56.422 +So I went from 1/4 to 261 quarter and + +01:05:56.422 --> 01:05:58.951 +3/4 to 2/6 and four six for example. + +01:05:58.951 --> 01:06:02.316 +So 1/3 and 2/3 is more uniform than 1/4 + +01:06:02.316 --> 01:06:03.129 +and 3/4. + +01:06:05.050 --> 01:06:07.040 +And then if the initial estimate was + +01:06:07.040 --> 01:06:09.020 +1/2, the final Estimate will still be + +01:06:09.020 --> 01:06:11.620 +1/2 because it's because this Prior is + +01:06:11.620 --> 01:06:13.650 +just trying to push things towards 1/2. + +01:06:20.780 --> 01:06:24.220 +So I want to give one example of a use + +01:06:24.220 --> 01:06:24.550 +case. + +01:06:24.550 --> 01:06:25.685 +So I've actually. + +01:06:25.685 --> 01:06:28.360 +I mean I want to say like I used Naive + +01:06:28.360 --> 01:06:30.630 +Bayes, but I use that assumption pretty + +01:06:30.630 --> 01:06:31.440 +often. + +01:06:31.440 --> 01:06:33.480 +For example if I wanted to Estimate a + +01:06:33.480 --> 01:06:35.210 +distribution of RGB colors. + +01:06:36.740 --> 01:06:38.410 +I would first convert it to a different + +01:06:38.410 --> 01:06:39.860 +color space, but let's just say I want + +01:06:39.860 --> 01:06:41.780 +to Estimate distribution of LGBT RGB + +01:06:41.780 --> 01:06:42.390 +colors. + +01:06:42.390 --> 01:06:45.055 +Then even though it's 3 dimensions, is + +01:06:45.055 --> 01:06:45.690 +a pretty. + +01:06:45.690 --> 01:06:47.920 +You need like a lot of data to estimate + +01:06:47.920 --> 01:06:48.610 +that distribution. + +01:06:48.610 --> 01:06:50.700 +And So what I might do is I'll say, + +01:06:50.700 --> 01:06:52.820 +well, I'm going to assume that RG and B + +01:06:52.820 --> 01:06:54.645 +are independent and so the probability + +01:06:54.645 --> 01:06:57.350 +of RGB is just the probability of R + +01:06:57.350 --> 01:06:58.808 +times probability of G times + +01:06:58.808 --> 01:06:59.524 +probability B. + +01:06:59.524 --> 01:07:01.600 +And I compute a histogram for each of + +01:07:01.600 --> 01:07:04.940 +those, and I use that to get my as my + +01:07:04.940 --> 01:07:06.230 +likelihood Estimate. + +01:07:06.560 --> 01:07:08.520 +So it's like really commonly used in + +01:07:08.520 --> 01:07:10.120 +that kind of setting where you want to + +01:07:10.120 --> 01:07:11.770 +Estimate the distribution of multiple + +01:07:11.770 --> 01:07:13.380 +variables and there's just no way to + +01:07:13.380 --> 01:07:13.810 +get a Joint. + +01:07:13.810 --> 01:07:17.100 +The only options you really have are to + +01:07:17.100 --> 01:07:18.410 +make something the Naive Bayes + +01:07:18.410 --> 01:07:21.330 +assumption or to do a mixture of + +01:07:21.330 --> 01:07:23.416 +Gaussians, which we'll talk about later + +01:07:23.416 --> 01:07:24.320 +in the semester. + +01:07:26.380 --> 01:07:27.940 +Right, But here's the case where it's + +01:07:27.940 --> 01:07:29.450 +used for object detection. + +01:07:29.450 --> 01:07:32.280 +So this was by Schneiderman Kanadi and + +01:07:32.280 --> 01:07:35.500 +it was the most accurate face and car + +01:07:35.500 --> 01:07:36.520 +detector for a while. + +01:07:37.450 --> 01:07:39.720 +They detector is based on wavelet + +01:07:39.720 --> 01:07:41.420 +coefficients which are just like local + +01:07:41.420 --> 01:07:42.610 +intensity differences. + +01:07:43.320 --> 01:07:46.010 +And the. + +01:07:46.090 --> 01:07:48.880 +The It's a Probabilistic framework, so + +01:07:48.880 --> 01:07:51.070 +they're trying to say whether if you + +01:07:51.070 --> 01:07:54.107 +Extract a window of Features from the + +01:07:54.107 --> 01:07:56.386 +image, some Features over some part of + +01:07:56.386 --> 01:07:56.839 +the image. + +01:07:57.450 --> 01:07:59.020 +And Extract all the wavelet + +01:07:59.020 --> 01:08:00.330 +coefficients. + +01:08:00.330 --> 01:08:02.390 +Then you want to say that it's a face + +01:08:02.390 --> 01:08:03.950 +if the probability of those + +01:08:03.950 --> 01:08:05.853 +coefficients is greater given that it's + +01:08:05.853 --> 01:08:08.390 +a face, than given that's not a face + +01:08:08.390 --> 01:08:10.330 +times the probability that's a face + +01:08:10.330 --> 01:08:11.730 +over the probability that's not a face. + +01:08:12.430 --> 01:08:14.680 +So it's this basic Probabilistic Model. + +01:08:14.680 --> 01:08:16.740 +And again, the probability modeling. + +01:08:16.740 --> 01:08:17.920 +The probability of all those + +01:08:17.920 --> 01:08:19.370 +coefficients is way too hard. + +01:08:20.330 --> 01:08:23.290 +On the other hand, modeling all the + +01:08:23.290 --> 01:08:25.560 +Features as independent given the label + +01:08:25.560 --> 01:08:26.950 +is a little bit too much of a + +01:08:26.950 --> 01:08:28.410 +simplifying assumption. + +01:08:28.410 --> 01:08:30.270 +So they use this algorithm that they + +01:08:30.270 --> 01:08:33.220 +call semi Naive Bayes which is proposed + +01:08:33.220 --> 01:08:34.040 +earlier. + +01:08:35.220 --> 01:08:37.946 +Where you just you Model the + +01:08:37.946 --> 01:08:39.803 +probabilities of little groups of + +01:08:39.803 --> 01:08:41.380 +features and then you say that the + +01:08:41.380 --> 01:08:43.166 +total probability is the probability + +01:08:43.166 --> 01:08:44.830 +the product or the probabilities of + +01:08:44.830 --> 01:08:45.849 +these groups of Features. + +01:08:46.710 --> 01:08:47.845 +So they call these patterns. + +01:08:47.845 --> 01:08:50.160 +So first you do some look at the mutual + +01:08:50.160 --> 01:08:51.870 +information, you have ways of measuring + +01:08:51.870 --> 01:08:54.050 +the dependence of different variables, + +01:08:54.050 --> 01:08:56.470 +and you cluster the Features together + +01:08:56.470 --> 01:08:58.280 +based on their dependencies. + +01:08:58.920 --> 01:09:00.430 +And then for little clusters of + +01:09:00.430 --> 01:09:02.149 +Features, 3 Features. + +01:09:03.060 --> 01:09:05.800 +You Estimate the probability of the + +01:09:05.800 --> 01:09:08.500 +Joint combination of these features and + +01:09:08.500 --> 01:09:11.230 +then the total probability of all the + +01:09:11.230 --> 01:09:11.620 +Features. + +01:09:11.620 --> 01:09:12.920 +I'm glad this isn't worker. + +01:09:12.920 --> 01:09:14.788 +The total probability of all the + +01:09:14.788 --> 01:09:16.660 +features is the product of the + +01:09:16.660 --> 01:09:18.270 +probabilities of each of these groups + +01:09:18.270 --> 01:09:18.840 +of Features. + +01:09:19.890 --> 01:09:21.140 +And so you Model. + +01:09:21.140 --> 01:09:23.616 +Likely a set of features are given that + +01:09:23.616 --> 01:09:25.270 +it's a face, and how likely they are + +01:09:25.270 --> 01:09:27.790 +given that it's not a face or given a + +01:09:27.790 --> 01:09:29.280 +random patch from an image. + +01:09:29.930 --> 01:09:32.260 +And then that can be used to classify + +01:09:32.260 --> 01:09:33.060 +images as face. + +01:09:33.060 --> 01:09:33.896 +You're not face. + +01:09:33.896 --> 01:09:35.560 +And you would Estimate this separately + +01:09:35.560 --> 01:09:37.120 +for cars and for each orientation of + +01:09:37.120 --> 01:09:38.110 +car question. + +01:09:43.310 --> 01:09:45.399 +So the question was what beat the 2005 + +01:09:45.400 --> 01:09:45.840 +model? + +01:09:45.840 --> 01:09:47.750 +I'm not really sure that there was + +01:09:47.750 --> 01:09:50.180 +something that beat it in 2006, but + +01:09:50.180 --> 01:09:53.820 +that when Dalal Triggs SVM based + +01:09:53.820 --> 01:09:55.570 +detector came out. + +01:09:56.200 --> 01:09:57.680 +And I think it might have been, I + +01:09:57.680 --> 01:10:00.617 +didn't look it up so I'm not sure, but + +01:10:00.617 --> 01:10:02.930 +I was, I'm pretty confident it was the + +01:10:02.930 --> 01:10:04.947 +most accurate up to 2005, but not + +01:10:04.947 --> 01:10:06.070 +confident after that. + +01:10:07.250 --> 01:10:10.430 +And now it took a while for face + +01:10:10.430 --> 01:10:12.650 +detection to get more accurate than + +01:10:12.650 --> 01:10:15.630 +most famous face detector was actually + +01:10:15.630 --> 01:10:18.330 +the Viola joins detector, which was + +01:10:18.330 --> 01:10:20.515 +popular because it was really fast. + +01:10:20.515 --> 01:10:24.046 +This thing man at a couple frames per + +01:10:24.046 --> 01:10:26.414 +second, but Viola Jones ran at 15 + +01:10:26.414 --> 01:10:28.560 +frames per second in 2001. + +01:10:30.310 --> 01:10:31.960 +But Viola Jones wasn't quite as + +01:10:31.960 --> 01:10:32.460 +accurate. + +01:10:35.210 --> 01:10:37.840 +Alright, so Summary of Naive bees. + +01:10:38.180 --> 01:10:38.790 +And. + +01:10:39.940 --> 01:10:41.740 +So the key assumption is that the + +01:10:41.740 --> 01:10:43.460 +Features are independent given the + +01:10:43.460 --> 01:10:43.870 +labels. + +01:10:46.730 --> 01:10:48.110 +The parameters are just the + +01:10:48.110 --> 01:10:50.173 +probabilities, are the parameters of + +01:10:50.173 --> 01:10:51.990 +each of these probability functions, + +01:10:51.990 --> 01:10:53.908 +the probability of each feature given Y + +01:10:53.908 --> 01:10:55.750 +and probability of Y and justice. + +01:10:55.750 --> 01:10:57.250 +Like in the Simple fruit example I + +01:10:57.250 --> 01:10:59.405 +gave, you can use different models for + +01:10:59.405 --> 01:10:59.976 +different features. + +01:10:59.976 --> 01:11:02.340 +Some of the features could be discrete + +01:11:02.340 --> 01:11:04.120 +values and some could be continuous + +01:11:04.120 --> 01:11:04.560 +values. + +01:11:04.560 --> 01:11:05.520 +That's not a problem. + +01:11:08.520 --> 01:11:10.150 +You have to choose which probability + +01:11:10.150 --> 01:11:11.510 +function you're going to use for each + +01:11:11.510 --> 01:11:11.940 +feature. + +01:11:14.450 --> 01:11:16.250 +Nine days can be useful if you have + +01:11:16.250 --> 01:11:18.080 +limited training data, because you only + +01:11:18.080 --> 01:11:19.560 +have to Estimate these one-dimensional + +01:11:19.560 --> 01:11:21.150 +distributions, which you can do from + +01:11:21.150 --> 01:11:22.370 +relatively few Samples. + +01:11:23.000 --> 01:11:24.420 +And if the features are not highly + +01:11:24.420 --> 01:11:26.540 +interdependent, and it can also be + +01:11:26.540 --> 01:11:27.970 +useful as a baseline if you want + +01:11:27.970 --> 01:11:29.766 +something that's fast to code, train + +01:11:29.766 --> 01:11:30.579 +and test. + +01:11:30.580 --> 01:11:32.900 +So as you do your homework, I think out + +01:11:32.900 --> 01:11:34.860 +of the methods, Naive Bayes has the + +01:11:34.860 --> 01:11:37.140 +lowest training plus test time. + +01:11:37.140 --> 01:11:40.139 +Logistic regression is going to be + +01:11:40.140 --> 01:11:42.618 +roughly tied for test time, but it + +01:11:42.618 --> 01:11:43.680 +takes an awful lot. + +01:11:43.680 --> 01:11:45.980 +Well, it takes longer to train. + +01:11:45.980 --> 01:11:48.379 +KNN takes no time to train, but takes a + +01:11:48.380 --> 01:11:49.570 +whole lot longer to test. + +01:11:54.630 --> 01:11:56.830 +So when not to use? + +01:11:56.830 --> 01:11:58.760 +Usually Logistic or linear regression + +01:11:58.760 --> 01:12:01.070 +will work better if you have enough + +01:12:01.070 --> 01:12:01.440 +data. + +01:12:02.230 --> 01:12:05.510 +And the reason is that under most + +01:12:05.510 --> 01:12:07.860 +probability the exponential + +01:12:07.860 --> 01:12:09.790 +distribution of probability models + +01:12:09.790 --> 01:12:11.940 +which include Binomial, multinomial and + +01:12:11.940 --> 01:12:12.530 +Gaussian. + +01:12:13.640 --> 01:12:15.657 +You can rewrite Naive Bayes as a linear + +01:12:15.657 --> 01:12:18.993 +function of the input features, but the + +01:12:18.993 --> 01:12:21.740 +linear function is highly constrained + +01:12:21.740 --> 01:12:23.750 +based on this, estimating likelihoods + +01:12:23.750 --> 01:12:25.650 +for each feature separately. + +01:12:25.650 --> 01:12:27.500 +Where linear and logistic regression, + +01:12:27.500 --> 01:12:28.970 +which we'll talk about next Thursday, + +01:12:28.970 --> 01:12:30.815 +are not constrained, you can solve for + +01:12:30.815 --> 01:12:32.300 +the full range of coefficients. + +01:12:33.440 --> 01:12:35.050 +The other issue is that it doesn't + +01:12:35.050 --> 01:12:37.890 +provide a very good confidence Estimate + +01:12:37.890 --> 01:12:39.720 +because it over counts the influence of + +01:12:39.720 --> 01:12:40.880 +dependent variables. + +01:12:40.880 --> 01:12:42.860 +If you repeat a feature of many times, + +01:12:42.860 --> 01:12:44.680 +it's going to count it every time, and + +01:12:44.680 --> 01:12:47.215 +so it will tend to have too much weight + +01:12:47.215 --> 01:12:48.930 +and give you bad confidence estimates. + +01:12:51.010 --> 01:12:55.100 +9 Bayes is easy and fast to train, Fast + +01:12:55.100 --> 01:12:56.130 +for inference. + +01:12:56.130 --> 01:12:57.400 +You can use it with different kinds of + +01:12:57.400 --> 01:12:58.040 +variables. + +01:12:58.040 --> 01:12:59.220 +It doesn't account for feature + +01:12:59.220 --> 01:13:00.730 +interaction, doesn't provide good + +01:13:00.730 --> 01:13:01.670 +confidence estimates. + +01:13:02.390 --> 01:13:04.210 +And it's best when used with discrete + +01:13:04.210 --> 01:13:06.270 +variables, those that can be fit well + +01:13:06.270 --> 01:13:08.830 +by a Gaussian, or if you use kernel + +01:13:08.830 --> 01:13:10.690 +density estimation, which is something + +01:13:10.690 --> 01:13:11.840 +that we'll talk about later in this + +01:13:11.840 --> 01:13:13.580 +semester, a more general like + +01:13:13.580 --> 01:13:15.080 +continuous distribution function. + +01:13:17.210 --> 01:13:19.560 +And justice, as a reminder, don't pack + +01:13:19.560 --> 01:13:21.730 +up until I'm done, but this will be the + +01:13:21.730 --> 01:13:22.570 +second to last slide. + +01:13:24.220 --> 01:13:25.890 +So things remember. + +01:13:27.140 --> 01:13:28.950 +So Probabilistic models are really + +01:13:28.950 --> 01:13:30.837 +large class of machine learning + +01:13:30.837 --> 01:13:31.160 +methods. + +01:13:31.160 --> 01:13:32.590 +There are many different kinds of + +01:13:32.590 --> 01:13:34.690 +machine learning methods that are based + +01:13:34.690 --> 01:13:36.480 +on estimating the likelihoods of the + +01:13:36.480 --> 01:13:38.170 +label given the data or the data given + +01:13:38.170 --> 01:13:38.730 +the label. + +01:13:39.580 --> 01:13:41.630 +Naive Bayes assumes that Features are + +01:13:41.630 --> 01:13:45.430 +independent given the label, and it's + +01:13:45.430 --> 01:13:46.860 +easy and fast to estimate the + +01:13:46.860 --> 01:13:48.920 +parameters and reduces the risk of + +01:13:48.920 --> 01:13:50.480 +overfitting when you have limited data. + +01:13:52.270 --> 01:13:52.590 +It's. + +01:13:52.590 --> 01:13:55.190 +You don't usually have to derive how to + +01:13:55.190 --> 01:13:57.910 +solve for the likelihood parameters, + +01:13:57.910 --> 01:13:59.660 +but you can do it if you want to by + +01:13:59.660 --> 01:14:00.954 +taking the partial derivative. + +01:14:00.954 --> 01:14:03.540 +Usually it's usually you would be using + +01:14:03.540 --> 01:14:06.140 +a common a common kind of Model and you + +01:14:06.140 --> 01:14:07.290 +can just look up the Emily. + +01:14:09.490 --> 01:14:11.160 +The Prediction involves finding the way + +01:14:11.160 --> 01:14:13.190 +that maximizes the probability of the + +01:14:13.190 --> 01:14:15.150 +data and the label, either by trying + +01:14:15.150 --> 01:14:17.250 +all the possible values of Y or solving + +01:14:17.250 --> 01:14:18.230 +the partial derivative. + +01:14:19.270 --> 01:14:21.535 +And finally, Maximizing log probability + +01:14:21.535 --> 01:14:24.060 +of I is equivalent to Maximizing + +01:14:24.060 --> 01:14:25.360 +probability of. + +01:14:25.520 --> 01:14:27.310 +Sorry, Maximizing log probability of + +01:14:27.310 --> 01:14:30.270 +X&Y is equivalent to maximizing the + +01:14:30.270 --> 01:14:32.250 +probability of X&Y, and it's usually + +01:14:32.250 --> 01:14:34.000 +much easier, so it's important to + +01:14:34.000 --> 01:14:34.390 +remember that. + +01:14:35.970 --> 01:14:36.180 +Right. + +01:14:36.180 --> 01:14:37.840 +And then next class I'm going to talk + +01:14:37.840 --> 01:14:40.030 +about logistic regression and linear + +01:14:40.030 --> 01:14:40.700 +regression. + +01:14:41.530 --> 01:14:44.870 +And one more thing is I posted a review + +01:14:44.870 --> 01:14:49.310 +questions and answers to the 1st 2 + +01:14:49.310 --> 01:14:51.440 +cannon and this lecture on the web + +01:14:51.440 --> 01:14:52.050 +page. + +01:14:52.050 --> 01:14:53.690 +You don't have to do them but they're + +01:14:53.690 --> 01:14:55.410 +good review for the exam or just the + +01:14:55.410 --> 01:14:56.820 +check your knowledge after each + +01:14:56.820 --> 01:14:57.200 +lecture. + +01:14:57.890 --> 01:14:58.750 +Thank you. + +01:15:11.030 --> 01:15:11.320 +I. +