WEBVTT Kind: captions; Language: en-US NOTE Created on 2024-02-07T20:53:13.8059397Z by ClassTranscribe 00:01:57.930 --> 00:01:59.190 It seems like there's like. 00:02:01.950 --> 00:02:02.550 Yes, it's OK. 00:02:03.590 --> 00:02:04.800 Alright, good morning everybody. 00:02:08.160 --> 00:02:10.626 So I thought it I was trying to figure 00:02:10.626 --> 00:02:12.030 out why this seems like there's a lot 00:02:12.030 --> 00:02:13.520 of light on the screen, but I can't 00:02:13.520 --> 00:02:14.192 figure it out. 00:02:14.192 --> 00:02:16.300 I thought it was interesting that this 00:02:16.300 --> 00:02:17.490 for this picture. 00:02:17.580 --> 00:02:18.110 And. 00:02:19.140 --> 00:02:20.760 So I'm generating all of these with 00:02:20.760 --> 00:02:21.510 Dolly. 00:02:21.510 --> 00:02:22.880 This one was a dirt Rd. 00:02:22.880 --> 00:02:24.500 splits around a large gnarly tree 00:02:24.500 --> 00:02:25.640 fractal art. 00:02:25.640 --> 00:02:27.830 But I thought it was really funny how 00:02:27.830 --> 00:02:30.780 it without my bidding it put like some 00:02:30.780 --> 00:02:32.530 kind of superhero or something behind 00:02:32.530 --> 00:02:34.756 the tree there's like some looks like 00:02:34.756 --> 00:02:36.360 there's like some superhero that's like 00:02:36.360 --> 00:02:37.230 flying in and. 00:02:38.130 --> 00:02:39.420 I don't know where that came from. 00:02:41.170 --> 00:02:42.430 Can you guys see the screen OK? 00:02:43.750 --> 00:02:44.830 Seems a little faded. 00:03:05.390 --> 00:03:05.650 But. 00:03:12.610 --> 00:03:13.580 OK I. 00:03:16.440 --> 00:03:18.730 Yeah, I put the lights are on all off. 00:03:21.400 --> 00:03:22.580 But those are still on. 00:03:25.090 --> 00:03:27.020 Alright, let me just take one second. 00:03:48.110 --> 00:03:48.600 All right. 00:03:48.600 --> 00:03:50.860 Anyway, I'll move with it. 00:03:51.500 --> 00:03:53.410 Alright, so. 00:03:53.540 --> 00:03:55.570 And so for some Logistics, I wanted to 00:03:55.570 --> 00:03:57.160 I never got to introduce some of the 00:03:57.160 --> 00:03:58.740 TAS because the couple couldn't be here 00:03:58.740 --> 00:04:00.110 in the first day and I kept forgetting. 00:04:01.130 --> 00:04:03.890 So, Josh, are you here? 00:04:04.950 --> 00:04:08.900 OK, cool if you want to just actually. 00:04:10.130 --> 00:04:11.160 I can give my mic. 00:04:11.160 --> 00:04:12.780 If you want to just introduce yourself 00:04:12.780 --> 00:04:14.540 a little bit, you can say like what 00:04:14.540 --> 00:04:14.940 kind of. 00:04:19.890 --> 00:04:20.450 Yeah. 00:04:20.450 --> 00:04:20.870 Hi, everyone. 00:04:20.870 --> 00:04:21.290 I'm Josh. 00:04:21.290 --> 00:04:23.110 I've been applying machine learning to 00:04:23.110 --> 00:04:25.980 autonomous cars and airplanes. 00:04:27.150 --> 00:04:27.480 Cool. 00:04:27.480 --> 00:04:27.900 Thank you. 00:04:28.830 --> 00:04:31.020 And cassette, cassette. 00:04:37.760 --> 00:04:38.320 Yeah. 00:04:41.120 --> 00:04:46.150 OK hey everyone, I'm a TA for CS441 and 00:04:46.150 --> 00:04:49.230 I have experience with NLP majorly. 00:04:49.230 --> 00:04:49.720 Thank you. 00:04:50.960 --> 00:04:51.190 Great. 00:04:51.190 --> 00:04:52.080 Thank you. 00:04:52.080 --> 00:04:54.230 And I don't think Peter's here, but 00:04:54.230 --> 00:04:55.820 Peter, are you here, OK. 00:04:56.520 --> 00:04:58.240 Usually has a conflict on Tuesday, so 00:04:58.240 --> 00:05:00.780 also we have Pedro is a. 00:05:01.510 --> 00:05:04.170 A pro stock course Assistant. 00:05:04.170 --> 00:05:07.020 So it's not like a regular TA, but 00:05:07.020 --> 00:05:07.650 he's. 00:05:08.730 --> 00:05:10.510 Doing a postdoc with Nancy Amato. 00:05:11.170 --> 00:05:12.880 And here is helping out with the online 00:05:12.880 --> 00:05:15.190 course for a couple semesters. 00:05:15.830 --> 00:05:17.060 And so he's helping out with this 00:05:17.060 --> 00:05:19.140 course and he's. 00:05:20.920 --> 00:05:23.700 One of the things he's doing is holding 00:05:23.700 --> 00:05:25.640 office hours, and so especially if you 00:05:25.640 --> 00:05:27.710 have, if you want help with your 00:05:27.710 --> 00:05:31.940 projects or homeworks, they're like 00:05:31.940 --> 00:05:34.000 higher level advice, then he can be a 00:05:34.000 --> 00:05:35.676 really good resource for that. 00:05:35.676 --> 00:05:37.400 So I know a lot of people want to meet 00:05:37.400 --> 00:05:39.250 with me about their side projects, 00:05:39.250 --> 00:05:40.720 which is also fine, you're welcome to 00:05:40.720 --> 00:05:42.100 do that. 00:05:42.100 --> 00:05:44.090 But he's also a good person for that. 00:05:46.480 --> 00:05:49.550 Alright, so just as a reminder for 00:05:49.550 --> 00:05:51.210 anybody who wasn't here, the first 00:05:51.210 --> 00:05:53.630 lecture, all the notes and everything 00:05:53.630 --> 00:05:55.635 are on this web page. 00:05:55.635 --> 00:05:57.890 So make sure that you go there and sign 00:05:57.890 --> 00:06:00.290 up for CampusWire where announcements 00:06:00.290 --> 00:06:01.600 will be made. 00:06:01.600 --> 00:06:06.120 Also, I sent a survey by e-mail and I 00:06:06.120 --> 00:06:07.760 got a little bit of responses last 00:06:07.760 --> 00:06:08.240 night. 00:06:08.240 --> 00:06:10.390 Do you take some time to respond to it 00:06:10.390 --> 00:06:11.300 please? 00:06:11.300 --> 00:06:12.290 There's two parts. 00:06:12.290 --> 00:06:14.340 One is just asking for feedback about 00:06:14.340 --> 00:06:15.300 like piece of the course. 00:06:15.380 --> 00:06:16.340 And stuff like that. 00:06:16.420 --> 00:06:16.900 And. 00:06:17.720 --> 00:06:20.640 One part is asking about your interests 00:06:20.640 --> 00:06:23.060 for some of the possible. 00:06:24.070 --> 00:06:26.332 Challenges that I'll pick for final 00:06:26.332 --> 00:06:29.810 project and so basically for the final 00:06:29.810 --> 00:06:32.149 project there will be 3 challenges that 00:06:32.150 --> 00:06:33.600 are like pre selected. 00:06:34.230 --> 00:06:35.720 But if you don't want to do those, you 00:06:35.720 --> 00:06:38.370 can also just do some benchmark that's 00:06:38.370 --> 00:06:40.070 online or you can even do a custom 00:06:40.070 --> 00:06:40.860 task. 00:06:40.860 --> 00:06:43.960 And I'll post the specifications for 00:06:43.960 --> 00:06:46.640 final project soon as homework 2. 00:06:47.960 --> 00:06:50.140 Also, just based on the feedback I've 00:06:50.140 --> 00:06:52.810 seen so far, I think nobody thinks it's 00:06:52.810 --> 00:06:54.570 way too easy or too slow. 00:06:54.570 --> 00:06:57.150 Some people think it's much too fast 00:06:57.150 --> 00:06:57.930 and too hard. 00:06:57.930 --> 00:06:59.710 So I'm going to take some time on 00:06:59.710 --> 00:07:03.450 Thursday to Reconsolidate and present. 00:07:03.450 --> 00:07:07.280 Kind of go over what we've done so far, 00:07:07.280 --> 00:07:09.750 talk in more depth or maybe not more 00:07:09.750 --> 00:07:11.439 depth, but at least go over the 00:07:11.440 --> 00:07:12.150 concepts. 00:07:13.270 --> 00:07:16.080 And the algorithms and a little bit of 00:07:16.080 --> 00:07:18.380 code now that you've had a first pass 00:07:18.380 --> 00:07:18.640 edit. 00:07:20.460 --> 00:07:22.830 So I'll tap the brakes a little bit to 00:07:22.830 --> 00:07:24.500 do that because I think it's really 00:07:24.500 --> 00:07:27.215 important that these that everyone is 00:07:27.215 --> 00:07:28.790 really solid on these fundamentals. 00:07:28.790 --> 00:07:31.260 And I know that there's a pretty big 00:07:31.260 --> 00:07:33.090 range of backgrounds of people taking 00:07:33.090 --> 00:07:35.060 the course, many people from other 00:07:35.060 --> 00:07:35.710 departments. 00:07:37.290 --> 00:07:39.900 As well as other different kinds of. 00:07:41.230 --> 00:07:43.280 Of like academic foundations. 00:07:44.270 --> 00:07:44.610 Alright. 00:07:45.910 --> 00:07:47.890 So just to recap what we talked about 00:07:47.890 --> 00:07:49.640 in the last few lectures, very briefly, 00:07:49.640 --> 00:07:51.040 we talked about Nearest neighbor. 00:07:51.780 --> 00:07:53.210 And the superpower is the nearest 00:07:53.210 --> 00:07:55.170 neighbor are that it can instantly 00:07:55.170 --> 00:07:56.230 learn new classes. 00:07:56.230 --> 00:07:58.020 You can just add a new example to your 00:07:58.020 --> 00:07:58.790 training set. 00:07:58.790 --> 00:08:00.780 And since there's no model that has to 00:08:00.780 --> 00:08:04.110 be like tuned, you can just learn super 00:08:04.110 --> 00:08:04.720 quickly. 00:08:04.720 --> 00:08:07.450 And it's also a pretty good predictor 00:08:07.450 --> 00:08:08.980 from either one or many examples. 00:08:08.980 --> 00:08:10.430 So it's a really good. 00:08:10.530 --> 00:08:13.690 It's a really good algorithm to have in 00:08:13.690 --> 00:08:15.330 your tool belt as a baseline and 00:08:15.330 --> 00:08:16.760 sometimes as a best performer. 00:08:18.500 --> 00:08:20.160 We also talked about Naive bees. 00:08:21.050 --> 00:08:24.140 Night Bayes is not a great performer as 00:08:24.140 --> 00:08:26.984 like a full algorithm, but it's often 00:08:26.984 --> 00:08:27.426 a. 00:08:27.426 --> 00:08:30.075 It's an important concept because it's 00:08:30.075 --> 00:08:31.760 often part of an assumption that you 00:08:31.760 --> 00:08:32.920 make when you're trying to model 00:08:32.920 --> 00:08:35.560 probabilities that you'll assume that 00:08:35.560 --> 00:08:37.630 the different features are independent 00:08:37.630 --> 00:08:39.010 given the thing that you're trying to 00:08:39.010 --> 00:08:39.330 predict. 00:08:41.780 --> 00:08:44.290 It does have its pros, so the pros are 00:08:44.290 --> 00:08:46.560 that it's really fast to estimate even 00:08:46.560 --> 00:08:48.113 if you've got a lot of data. 00:08:48.113 --> 00:08:49.909 And if you don't have a lot of data and 00:08:49.910 --> 00:08:51.130 you're trying to get a probabilistic 00:08:51.130 --> 00:08:53.300 classifier, then it might be your best 00:08:53.300 --> 00:08:53.750 choice. 00:08:53.750 --> 00:08:56.700 Because of its strong assumptions, you 00:08:56.700 --> 00:08:59.880 can get decent estimates on those 00:08:59.880 --> 00:09:02.160 single variable functions from even 00:09:02.160 --> 00:09:02.800 limited data. 00:09:05.460 --> 00:09:07.964 We talked about logistic regression. 00:09:07.964 --> 00:09:10.830 Logistic regression is another super 00:09:10.830 --> 00:09:12.230 widely used classifier. 00:09:13.580 --> 00:09:16.400 I think the AML book says that SVM is 00:09:16.400 --> 00:09:18.776 should be or like go to as a first as 00:09:18.776 --> 00:09:20.660 like a first thing you try, but in my 00:09:20.660 --> 00:09:22.130 opinion Logistic Regression is. 00:09:23.810 --> 00:09:25.085 It's very effective. 00:09:25.085 --> 00:09:26.810 It's a very effective predictor if you 00:09:26.810 --> 00:09:28.720 have high dimensional features and it 00:09:28.720 --> 00:09:30.250 also provides good confidence 00:09:30.250 --> 00:09:31.460 estimates, meaning that. 00:09:32.150 --> 00:09:35.320 You get not only most likely class, but 00:09:35.320 --> 00:09:37.470 the probability that prediction is 00:09:37.470 --> 00:09:40.800 correct and those probabilities fairly 00:09:40.800 --> 00:09:41.400 trustworthy. 00:09:43.320 --> 00:09:44.970 We also talked about Linear Regression, 00:09:44.970 --> 00:09:46.560 where you're fitting a line to a set of 00:09:46.560 --> 00:09:47.050 points. 00:09:47.670 --> 00:09:50.610 And you can extrapolate to predict like 00:09:50.610 --> 00:09:52.620 new values that are outside of your 00:09:52.620 --> 00:09:53.790 Training range. 00:09:54.530 --> 00:09:55.840 And so. 00:09:56.730 --> 00:09:58.450 Linear regression is also useful for 00:09:58.450 --> 00:10:00.270 explaining relationships you're very 00:10:00.270 --> 00:10:02.100 commonly see, like trend lines. 00:10:02.100 --> 00:10:03.390 That's just Linear Regression. 00:10:04.130 --> 00:10:05.850 And you can predict continuous values 00:10:05.850 --> 00:10:07.600 from many variables in linear 00:10:07.600 --> 00:10:10.130 regression is also like probably the 00:10:10.130 --> 00:10:12.760 most common tool for. 00:10:12.830 --> 00:10:15.590 For things like, I don't know, like 00:10:15.590 --> 00:10:17.790 economics or analyzing. 00:10:18.770 --> 00:10:23.100 Yeah, time series analyzing like fMRI 00:10:23.100 --> 00:10:25.930 data or all kinds of scientific and 00:10:25.930 --> 00:10:27.180 economic analysis. 00:10:30.420 --> 00:10:33.810 So almost all algorithms involve these 00:10:33.810 --> 00:10:35.760 Nearest neighbor, logistic regression 00:10:35.760 --> 00:10:36.850 or linear regression. 00:10:37.540 --> 00:10:41.040 And the reason that there's thousand 00:10:41.040 --> 00:10:43.330 papers published in the last 10 years 00:10:43.330 --> 00:10:45.060 or so, probably a lot more than that 00:10:45.060 --> 00:10:47.030 actually, is that. 00:10:47.850 --> 00:10:50.120 Is really the feature learning, so it's 00:10:50.120 --> 00:10:52.090 getting the right representation so 00:10:52.090 --> 00:10:54.490 that when you feed that representation 00:10:54.490 --> 00:10:56.610 into these like Linear models or 00:10:56.610 --> 00:10:59.080 Nearest neighbor, you get good results. 00:11:00.080 --> 00:11:00.660 And so. 00:11:01.510 --> 00:11:03.020 Pretty much the rest of what we're 00:11:03.020 --> 00:11:05.160 going to learn in the supervised 00:11:05.160 --> 00:11:07.520 learning section of the course is how 00:11:07.520 --> 00:11:08.460 to learn features. 00:11:11.930 --> 00:11:14.150 So I did want to just briefly go over 00:11:14.150 --> 00:11:15.640 the homework and remind you that it's 00:11:15.640 --> 00:11:18.180 due on February 6th on Monday. 00:11:19.060 --> 00:11:21.400 And I'll be going over some related 00:11:21.400 --> 00:11:22.830 things again in more detail on 00:11:22.830 --> 00:11:23.350 Thursday. 00:11:24.300 --> 00:11:27.200 But there's two parts to the main 00:11:27.200 --> 00:11:27.590 homework. 00:11:27.590 --> 00:11:29.770 There's Digit Classification where 00:11:29.770 --> 00:11:31.186 you're trying to predict a label zero 00:11:31.186 --> 00:11:33.530 to 9 based on a 28 by 28 image. 00:11:34.410 --> 00:11:36.409 These images get reshaped into like a 00:11:36.410 --> 00:11:38.840 single vector, so you have a feature 00:11:38.840 --> 00:11:41.020 vector that corresponds to the pixel 00:11:41.020 --> 00:11:42.280 intensities of the image. 00:11:44.350 --> 00:11:46.510 And then you have to do K and Naive 00:11:46.510 --> 00:11:49.200 Bayes, linear logistic regression. 00:11:50.060 --> 00:11:52.510 And plot the Error versus. 00:11:52.670 --> 00:11:55.420 A plot Error versus Training size to 00:11:55.420 --> 00:11:57.310 get a sense for like how performance 00:11:57.310 --> 00:11:58.745 changes as you vary the number of 00:11:58.745 --> 00:11:59.530 training examples. 00:12:00.380 --> 00:12:02.820 And then to select the best parameter 00:12:02.820 --> 00:12:06.490 using validation set which is a really 00:12:06.490 --> 00:12:07.240 hyper parameter. 00:12:07.240 --> 00:12:09.020 Tuning is like something that you do 00:12:09.020 --> 00:12:10.100 all the time in machine learning. 00:12:13.150 --> 00:12:14.839 The second problem is Temperature 00:12:14.840 --> 00:12:15.700 Regression. 00:12:15.700 --> 00:12:18.182 So I got this Temperature. 00:12:18.182 --> 00:12:20.178 This data set of like the temperature 00:12:20.178 --> 00:12:22.890 is a big cities in the US and then 00:12:22.890 --> 00:12:24.111 made-up a problem from it. 00:12:24.111 --> 00:12:26.550 So the problem is to try to predict the 00:12:26.550 --> 00:12:28.220 next day's temperature in Cleveland 00:12:28.220 --> 00:12:30.000 which stays zero given all the previous 00:12:30.000 --> 00:12:30.550 temperatures. 00:12:31.490 --> 00:12:34.350 And these features have meanings. 00:12:34.350 --> 00:12:37.417 Every feature is some previous is, like 00:12:37.417 --> 00:12:39.530 the temperature of 1 in the big cities 00:12:39.530 --> 00:12:40.990 from one in the past five days. 00:12:42.570 --> 00:12:44.753 But you can kind of. 00:12:44.753 --> 00:12:46.110 You don't really need to know those 00:12:46.110 --> 00:12:48.110 meanings in order to solve the problem 00:12:48.110 --> 00:12:48.480 again. 00:12:48.480 --> 00:12:50.780 You essentially just have a feature 00:12:50.780 --> 00:12:53.730 vector of a bunch of continuous values 00:12:53.730 --> 00:12:55.570 in this case, and you're trying to 00:12:55.570 --> 00:12:57.290 predict a new continuous value, which 00:12:57.290 --> 00:13:00.460 is Cleveland Cleveland's temperature in 00:13:00.460 --> 00:13:01.240 the next day. 00:13:02.010 --> 00:13:04.545 And again you can use KNN and a Bayes 00:13:04.545 --> 00:13:06.000 and now Linear Regression. 00:13:07.020 --> 00:13:08.935 KNN implementation will be essentially 00:13:08.935 --> 00:13:11.440 the same for these A2 line change of 00:13:11.440 --> 00:13:13.820 code because now instead of predicting 00:13:13.820 --> 00:13:16.230 a categorical variable, you're 00:13:16.230 --> 00:13:18.440 predicting a continuous variable. 00:13:18.440 --> 00:13:20.580 So if K is greater than one, you 00:13:20.580 --> 00:13:23.770 average the predictions for Regression 00:13:23.770 --> 00:13:26.280 where for the Classification you choose 00:13:26.280 --> 00:13:27.430 the most common prediction. 00:13:28.580 --> 00:13:29.840 That's the only change. 00:13:29.840 --> 00:13:31.590 Naive Bayes does change quite a bit 00:13:31.590 --> 00:13:32.710 because you're using a different 00:13:32.710 --> 00:13:33.590 probabilistic model. 00:13:34.360 --> 00:13:36.710 And remember that there's one lecture 00:13:36.710 --> 00:13:38.670 slide that has the derivation for how 00:13:38.670 --> 00:13:40.545 you do the inference for nibbies under 00:13:40.545 --> 00:13:41.050 the setting. 00:13:42.330 --> 00:13:44.760 And then for linear and logistic 00:13:44.760 --> 00:13:46.820 regression you're able to use the 00:13:46.820 --> 00:13:48.350 modules in sklearn. 00:13:49.890 --> 00:13:51.680 And then the final part is to identify 00:13:51.680 --> 00:13:53.550 the most important features using L1 00:13:53.550 --> 00:13:54.320 Linear Regression. 00:13:55.030 --> 00:13:57.160 So the reason that we use. 00:13:58.020 --> 00:13:59.810 And when we do like. 00:14:01.000 --> 00:14:03.170 Linear and logistic regression, we're 00:14:03.170 --> 00:14:03.580 trying. 00:14:03.580 --> 00:14:05.228 We're mainly trying to fit the data. 00:14:05.228 --> 00:14:06.600 We're trying to come up with a model 00:14:06.600 --> 00:14:08.340 that fits the data or fits our 00:14:08.340 --> 00:14:09.920 predictions given the features. 00:14:10.630 --> 00:14:13.720 But also we often express some 00:14:13.720 --> 00:14:14.490 preference. 00:14:15.190 --> 00:14:19.892 Over the model, in particular that the 00:14:19.892 --> 00:14:21.669 weights don't get too large, and the 00:14:21.670 --> 00:14:25.170 reason for that is to avoid like 00:14:25.170 --> 00:14:27.070 overfitting or over relying on 00:14:27.070 --> 00:14:30.410 particular features, as well as to 00:14:30.410 --> 00:14:34.795 improve the generalization to new data 00:14:34.795 --> 00:14:36.209 and generalization. 00:14:36.210 --> 00:14:37.810 Research shows that if you can fit 00:14:37.810 --> 00:14:39.220 something with smaller weights, then 00:14:39.220 --> 00:14:42.013 you're more likely to generalize to new 00:14:42.013 --> 00:14:42.209 data. 00:14:44.640 --> 00:14:46.550 And here we're going to use it for 00:14:46.550 --> 00:14:47.520 feature selection, yeah? 00:14:58.910 --> 00:15:02.370 The so the parameters are. 00:15:02.370 --> 00:15:04.510 You're talking about 1/3. 00:15:04.510 --> 00:15:07.825 OK, so for Naive Bayes the parameter is 00:15:07.825 --> 00:15:10.430 the prior, so that's like the alpha of 00:15:10.430 --> 00:15:11.010 like your. 00:15:11.670 --> 00:15:14.903 In the, it's the initial count, so you 00:15:14.903 --> 00:15:15.882 have a Naive Bayes. 00:15:15.882 --> 00:15:17.360 You have a prior that's essentially 00:15:17.360 --> 00:15:19.200 that you pretend like you've seen all 00:15:19.200 --> 00:15:20.230 combinations of. 00:15:20.950 --> 00:15:23.930 Of things that you're counting, you 00:15:23.930 --> 00:15:26.210 pretend that you see alpha times, and 00:15:26.210 --> 00:15:28.510 so that kind of gives you a bias 00:15:28.510 --> 00:15:30.200 towards estimating that everything's 00:15:30.200 --> 00:15:33.170 equally likely, and that alpha is a 00:15:33.170 --> 00:15:34.190 parameter that you can use. 00:15:34.810 --> 00:15:36.270 You can learn using validation. 00:15:37.010 --> 00:15:39.920 For Logistic Regression, it's the 00:15:39.920 --> 00:15:42.650 Lambda which is your weight on the 00:15:42.650 --> 00:15:43.760 regularization term. 00:15:45.650 --> 00:15:48.180 And for K&N, it's your K, which is the 00:15:48.180 --> 00:15:49.320 number of nearest neighbors you 00:15:49.320 --> 00:15:49.710 consider. 00:15:57.960 --> 00:15:58.220 Yeah. 00:16:00.180 --> 00:16:03.284 So the K&N is. 00:16:03.284 --> 00:16:05.686 It's almost the same whether you're 00:16:05.686 --> 00:16:08.260 doing Regression or Classification. 00:16:08.260 --> 00:16:09.980 When you find the K nearest neighbors, 00:16:09.980 --> 00:16:11.790 it's the exact same code. 00:16:11.790 --> 00:16:14.016 The difference is that if you're doing 00:16:14.016 --> 00:16:15.270 Regression, you're trying to predict 00:16:15.270 --> 00:16:16.200 continuous values. 00:16:16.200 --> 00:16:19.340 So if K is greater than one, then you 00:16:19.340 --> 00:16:21.532 want to average those continuous values 00:16:21.532 --> 00:16:23.150 to get your final prediction. 00:16:23.850 --> 00:16:26.060 And if you're doing Classification, you 00:16:26.060 --> 00:16:28.490 find the most common label instead of 00:16:28.490 --> 00:16:29.803 averaging because you don't want to 00:16:29.803 --> 00:16:31.470 say, well it could be a four, it could 00:16:31.470 --> 00:16:31.980 be a 9. 00:16:31.980 --> 00:16:33.110 So I'm going to like split the 00:16:33.110 --> 00:16:34.270 difference and say it's a 6. 00:16:42.030 --> 00:16:45.420 That are averaging just that you so 00:16:45.420 --> 00:16:47.600 like if K&N returns like the 00:16:47.600 --> 00:16:54.870 temperatures of 1012 and 13 then you 00:16:54.870 --> 00:16:57.190 would say that the average temperature 00:16:57.190 --> 00:16:59.530 is like 11.3 or whatever that works out 00:16:59.530 --> 00:16:59.730 to. 00:17:04.600 --> 00:17:06.440 Yeah, at the end, if K is greater than 00:17:06.440 --> 00:17:09.333 one, then you take the arithmetic mean 00:17:09.333 --> 00:17:11.210 of the average of the. 00:17:11.940 --> 00:17:14.560 Predictions of your K nearest 00:17:14.560 --> 00:17:14.970 neighbors. 00:17:16.590 --> 00:17:16.790 Yeah. 00:17:18.610 --> 00:17:20.560 And so you could also get a variance 00:17:20.560 --> 00:17:22.370 from that, which you don't need to do 00:17:22.370 --> 00:17:24.500 for the homework, but so as a result 00:17:24.500 --> 00:17:26.550 you can have some like confidence bound 00:17:26.550 --> 00:17:27.840 on your estimate as well. 00:17:30.050 --> 00:17:31.780 Alright then you have stretch goals, 00:17:31.780 --> 00:17:32.170 so. 00:17:32.850 --> 00:17:34.100 Stretch goals are. 00:17:35.130 --> 00:17:37.000 Mainly intended for people taking the 00:17:37.000 --> 00:17:39.020 four credit version, but you can anyone 00:17:39.020 --> 00:17:39.510 can try them. 00:17:40.240 --> 00:17:42.570 So there's just improving the MNIST 00:17:42.570 --> 00:17:44.370 classification, like some ideas. 00:17:44.370 --> 00:17:47.360 Or you could try to crop around the 00:17:47.360 --> 00:17:49.000 Digit, or you could make sure that 00:17:49.000 --> 00:17:51.840 they're all centered, or do some 00:17:51.840 --> 00:17:53.410 whitening or other kinds of feature 00:17:53.410 --> 00:17:54.340 transformations. 00:17:55.430 --> 00:17:56.770 Improving Temperature Regression. 00:17:56.770 --> 00:18:00.070 To be honest, I'm not sure exactly how 00:18:00.070 --> 00:18:01.829 much this can be improved or how to 00:18:01.830 --> 00:18:02.280 improve it. 00:18:03.030 --> 00:18:04.720 Again, there's. 00:18:04.720 --> 00:18:07.370 What I would do is try like subtracting 00:18:07.370 --> 00:18:08.110 off the mean. 00:18:08.110 --> 00:18:09.220 For example, you can. 00:18:10.380 --> 00:18:12.370 You can normalize your features before 00:18:12.370 --> 00:18:15.540 you do the fitting by subtracting off 00:18:15.540 --> 00:18:16.750 means and dividing by steering 00:18:16.750 --> 00:18:17.410 deviations. 00:18:17.410 --> 00:18:18.140 That's one idea. 00:18:19.060 --> 00:18:22.320 But we'll look at it after submissions 00:18:22.320 --> 00:18:24.095 if it turns out that. 00:18:24.095 --> 00:18:27.020 So the targets I Choose are because I 00:18:27.020 --> 00:18:29.273 was able to do like some simple things 00:18:29.273 --> 00:18:32.383 to bring down the Error by a few tenths 00:18:32.383 --> 00:18:33.510 of a percent. 00:18:33.510 --> 00:18:35.000 So I kind of figured that if you do 00:18:35.000 --> 00:18:36.346 more things, you'll be able to bring it 00:18:36.346 --> 00:18:38.420 down further, but it's hard to tell so. 00:18:39.240 --> 00:18:40.960 If you do this and you put a lot of 00:18:40.960 --> 00:18:42.600 effort into it, describe your effort 00:18:42.600 --> 00:18:45.594 and we'll assign points even if you 00:18:45.594 --> 00:18:47.680 even if it turns out that there's not 00:18:47.680 --> 00:18:48.640 like a big improvement. 00:18:48.640 --> 00:18:50.676 So don't stress out if you can't get 00:18:50.676 --> 00:18:51.609 like 119. 00:18:52.450 --> 00:18:54.200 RMS a year or something like that. 00:18:55.130 --> 00:18:55.335 Right. 00:18:55.335 --> 00:18:57.306 The last one is to generate a train 00:18:57.306 --> 00:18:58.806 set, train Test, Classification set. 00:18:58.806 --> 00:19:00.380 So this actually means don't like 00:19:00.380 --> 00:19:02.020 generate it out of MNIST to create 00:19:02.020 --> 00:19:02.804 synthetic data. 00:19:02.804 --> 00:19:05.020 So you can Naive Bayes make certain 00:19:05.020 --> 00:19:05.405 assumptions. 00:19:05.405 --> 00:19:07.180 So if you generate your data according 00:19:07.180 --> 00:19:09.390 to those Assumptions, you should be 00:19:09.390 --> 00:19:11.900 able to create a problem that we're 00:19:11.900 --> 00:19:13.520 Naive bees can outperform the other 00:19:13.520 --> 00:19:13.980 methods. 00:19:18.970 --> 00:19:22.130 So for these homeworks, make sure that 00:19:22.130 --> 00:19:24.020 you of course read the assignment. 00:19:24.020 --> 00:19:25.040 Read the tips. 00:19:25.530 --> 00:19:26.210 00:19:27.060 --> 00:19:29.190 And then you should be adding code to 00:19:29.190 --> 00:19:30.045 the starter code. 00:19:30.045 --> 00:19:31.610 The starter code doesn't really solve 00:19:31.610 --> 00:19:33.030 the problems for you, but it loads the 00:19:33.030 --> 00:19:34.570 data and gives you some examples. 00:19:34.570 --> 00:19:38.160 So for example, for example, there's a. 00:19:38.810 --> 00:19:41.340 In the Regression, I think it includes 00:19:41.340 --> 00:19:43.710 like a baseline where it computes RMSE 00:19:43.710 --> 00:19:46.450 and median absolute error, so that 00:19:46.450 --> 00:19:48.760 function can essentially be reused 00:19:48.760 --> 00:19:50.060 later to compute the errors. 00:19:51.120 --> 00:19:53.073 And that baseline gives you some idea 00:19:53.073 --> 00:19:55.390 of like what kind of performance you 00:19:55.390 --> 00:19:55.870 might get. 00:19:55.870 --> 00:19:57.320 Like you should beat that baseline 00:19:57.320 --> 00:19:58.620 because that's just based on a single 00:19:58.620 --> 00:19:58.940 feature. 00:20:00.300 --> 00:20:02.980 And then you complete the report and 00:20:02.980 --> 00:20:04.940 make sure to include expected points. 00:20:04.940 --> 00:20:07.040 So when the grader is graded they will 00:20:07.040 --> 00:20:09.140 essentially just say if they disagree 00:20:09.140 --> 00:20:09.846 with you. 00:20:09.846 --> 00:20:12.470 So you if you claim like 10 points but 00:20:12.470 --> 00:20:14.345 something was wrong then they might say 00:20:14.345 --> 00:20:16.310 you lose like 3 points for this reason 00:20:16.310 --> 00:20:20.340 and so that streamlines their grading. 00:20:21.930 --> 00:20:23.580 The assignment, the report Submit your 00:20:23.580 --> 00:20:26.070 notebook and either if you just have 00:20:26.070 --> 00:20:28.800 one file, submitting the IPYNB is fine 00:20:28.800 --> 00:20:29.890 or otherwise you can zip it. 00:20:30.860 --> 00:20:32.220 And that's it. 00:20:33.960 --> 00:20:34.620 Yeah, question. 00:20:41.730 --> 00:20:47.160 So you need in three Credit was at 450, 00:20:47.160 --> 00:20:47.640 is that right? 00:20:48.650 --> 00:20:50.810 So think I think in the three credit 00:20:50.810 --> 00:20:52.300 you need 450 points. 00:20:53.660 --> 00:20:55.430 Each assignment without doing any 00:20:55.430 --> 00:20:56.230 stretch goals. 00:20:56.230 --> 00:20:58.620 Each assignment is worth 100 points and 00:20:58.620 --> 00:21:01.240 the final project is worth 50 points. 00:21:01.240 --> 00:21:02.673 I mean sorry, the final projects worth 00:21:02.673 --> 00:21:03.460 100 points also. 00:21:04.150 --> 00:21:06.310 So if you're in the three Credit 00:21:06.310 --> 00:21:08.210 version and you don't do any stretch 00:21:08.210 --> 00:21:10.960 goals, and you do all the assignments 00:21:10.960 --> 00:21:12.500 and you do the final project, you will 00:21:12.500 --> 00:21:13.570 have more points than you need. 00:21:14.190 --> 00:21:17.740 So the so you can kind of pick 00:21:17.740 --> 00:21:19.270 something that you don't want to do and 00:21:19.270 --> 00:21:20.910 skip it if you're in the three credit 00:21:20.910 --> 00:21:24.100 course and or like if you just are 00:21:24.100 --> 00:21:26.330 already a machine learning guru, you 00:21:26.330 --> 00:21:29.290 can do like 3 assignments with all the 00:21:29.290 --> 00:21:31.630 extra parts and then take a vacation. 00:21:32.920 --> 00:21:34.720 If you're in the four credit version, 00:21:34.720 --> 00:21:37.490 then you will have to do some of the. 00:21:37.670 --> 00:21:39.520 Some of the stretch goals in order to 00:21:39.520 --> 00:21:41.470 get your full points, which are 550. 00:21:49.580 --> 00:21:52.715 Alright, so now I'm going to move on to 00:21:52.715 --> 00:21:54.180 the main topic. 00:21:54.180 --> 00:21:57.340 So we've seen so far, we've seen 2 main 00:21:57.340 --> 00:21:59.116 choices for how to use the features. 00:21:59.116 --> 00:22:01.025 We could do Nearest neighbor when we 00:22:01.025 --> 00:22:03.200 use all the features jointly in order 00:22:03.200 --> 00:22:05.280 to find similar examples, and then we 00:22:05.280 --> 00:22:06.970 predict the most similar label. 00:22:07.910 --> 00:22:10.160 Or we can use a linear model where 00:22:10.160 --> 00:22:11.980 essentially you're making a prediction 00:22:11.980 --> 00:22:14.530 out of a of all the feature values. 00:22:16.070 --> 00:22:18.490 But there's some other things that are 00:22:18.490 --> 00:22:20.270 kind of intuitive, so. 00:22:21.220 --> 00:22:24.010 For example, if you consider this where 00:22:24.010 --> 00:22:26.260 you're trying to split the red X's from 00:22:26.260 --> 00:22:27.710 the Green O's. 00:22:28.370 --> 00:22:30.820 What's like another way that you might 00:22:30.820 --> 00:22:33.180 try to define what that Decision 00:22:33.180 --> 00:22:35.130 boundary is if you wanted to, say, tell 00:22:35.130 --> 00:22:35.730 somebody else? 00:22:35.730 --> 00:22:37.110 Like how do you identify whether 00:22:37.110 --> 00:22:38.770 something is a no? 00:22:52.240 --> 00:22:55.600 Yeah, I mean you so your jaw some kind 00:22:55.600 --> 00:22:56.200 of boundary. 00:22:57.150 --> 00:22:57.690 And. 00:22:58.620 --> 00:23:00.315 And one way that you might think about 00:23:00.315 --> 00:23:03.440 that is creating a kind of like simple 00:23:03.440 --> 00:23:04.220 rule like this. 00:23:04.220 --> 00:23:05.890 Like you might say that if. 00:23:06.600 --> 00:23:09.040 You basically draw a boundary, but if 00:23:09.040 --> 00:23:11.252 you want to specify you might say if X2 00:23:11.252 --> 00:23:15.820 is less than .6 and X2 is greater than 00:23:15.820 --> 00:23:16.500 two. 00:23:17.460 --> 00:23:21.480 And tX2, oops, that's just say X1 and 00:23:21.480 --> 00:23:22.082 the last one. 00:23:22.082 --> 00:23:24.630 And if X1 is less than seven then it's 00:23:24.630 --> 00:23:26.672 an O and otherwise it's an X. 00:23:26.672 --> 00:23:28.110 So basically you could create like a 00:23:28.110 --> 00:23:29.502 set of rules like that, right? 00:23:29.502 --> 00:23:32.161 So say if it meets these criteria then 00:23:32.161 --> 00:23:34.819 it's one class and if it meets these 00:23:34.820 --> 00:23:37.070 other criteria it's another class. 00:23:40.160 --> 00:23:42.930 And So what we're going to learn today 00:23:42.930 --> 00:23:45.280 is how we can try to learn these rules 00:23:45.280 --> 00:23:48.220 automatically, even if we have a lot of 00:23:48.220 --> 00:23:50.520 features in more complicated kinds of 00:23:50.520 --> 00:23:51.250 predictions. 00:23:52.920 --> 00:23:55.108 So this is basically the idea of 00:23:55.108 --> 00:23:55.744 Decision trees. 00:23:55.744 --> 00:23:58.490 So we all use Decision trees in our own 00:23:58.490 --> 00:24:00.264 life, even if we don't think about it 00:24:00.264 --> 00:24:00.812 that way. 00:24:00.812 --> 00:24:02.710 Like you often say, if this happens, 00:24:02.710 --> 00:24:04.121 I'll do that, and if it doesn't, then 00:24:04.121 --> 00:24:05.029 I'll do this other thing. 00:24:05.030 --> 00:24:06.685 That's like a Decision tree, right? 00:24:06.685 --> 00:24:10.400 You had some kind of criteria, and 00:24:10.400 --> 00:24:12.306 depending on the outcome of that 00:24:12.306 --> 00:24:13.886 criteria, you do one thing. 00:24:13.886 --> 00:24:16.680 And if it's the other way, if you get 00:24:16.680 --> 00:24:17.900 the other outcome, then you would be 00:24:17.900 --> 00:24:18.920 doing the other thing. 00:24:18.920 --> 00:24:20.310 And maybe you have a whole chain of 00:24:20.310 --> 00:24:22.090 them if I. 00:24:22.250 --> 00:24:23.700 If I have time today, I'm going to go 00:24:23.700 --> 00:24:25.990 to the grocery store, but if the car is 00:24:25.990 --> 00:24:27.330 not there then I'm going to do this 00:24:27.330 --> 00:24:28.480 instead and so on. 00:24:29.850 --> 00:24:32.370 All right, so in Decision trees, the 00:24:32.370 --> 00:24:34.500 Training is essentially to iteratively 00:24:34.500 --> 00:24:37.340 Choose the attribute and split in a 00:24:37.340 --> 00:24:40.080 split value that will best separate 00:24:40.080 --> 00:24:41.530 your classes from each other. 00:24:42.920 --> 00:24:44.610 Or if you're doing continuous values 00:24:44.610 --> 00:24:47.010 that kind of group things into similar 00:24:47.010 --> 00:24:48.240 prediction values. 00:24:49.480 --> 00:24:52.440 So for example you might say if these 00:24:52.440 --> 00:24:56.600 red circles are oranges and these 00:24:56.600 --> 00:24:59.264 triangles are lemons, where there 00:24:59.264 --> 00:25:01.090 oranges and lemons are plotted 00:25:01.090 --> 00:25:02.250 according to their width and their 00:25:02.250 --> 00:25:02.750 height. 00:25:02.750 --> 00:25:07.726 You might decide well if it's less than 00:25:07.726 --> 00:25:10.170 6.5 centimeters then. 00:25:10.170 --> 00:25:12.690 Or I'll use greater since it's there if 00:25:12.690 --> 00:25:14.190 it's greater than 6.5 centimeters. 00:25:15.450 --> 00:25:17.267 Then I'm going to split it into this 00:25:17.267 --> 00:25:19.410 section where it's like mostly oranges 00:25:19.410 --> 00:25:22.110 and if it's less than 6.5 centimeters 00:25:22.110 --> 00:25:24.395 width, then I'll split it into this 00:25:24.395 --> 00:25:26.220 section where it's mostly lemons. 00:25:27.250 --> 00:25:30.560 Neither of these a perfect split still. 00:25:30.560 --> 00:25:32.910 So then I go further and say if it was 00:25:32.910 --> 00:25:35.309 on this side of the split, if it's 00:25:35.310 --> 00:25:37.915 greater than 95 centimeter height then 00:25:37.915 --> 00:25:40.350 it's a lemon, and if it's less than 00:25:40.350 --> 00:25:42.130 that then it's a. 00:25:42.820 --> 00:25:43.760 Then it's an orange. 00:25:44.900 --> 00:25:46.660 And now that's like a pretty confident 00:25:46.660 --> 00:25:47.170 prediction. 00:25:47.930 --> 00:25:49.610 And then if I'm on this side then I can 00:25:49.610 --> 00:25:51.560 split it by height and say if it's less 00:25:51.560 --> 00:25:51.990 than. 00:25:53.690 --> 00:25:55.530 If it's greater than 6 centimeters then 00:25:55.530 --> 00:25:57.714 it's a lemon, and if it's less than 6 00:25:57.714 --> 00:25:59.450 centimeters then it's an orange. 00:25:59.450 --> 00:26:01.130 So you can like iteratively Choose a 00:26:01.130 --> 00:26:03.180 test and then keep splitting the data. 00:26:03.780 --> 00:26:06.510 And every time you choose a test, test 00:26:06.510 --> 00:26:09.510 another test that splits the data 00:26:09.510 --> 00:26:10.910 further according to what you're trying 00:26:10.910 --> 00:26:11.320 to predict. 00:26:12.270 --> 00:26:14.890 Essentially, this method Combines a 00:26:14.890 --> 00:26:16.760 feature selection and modeling with 00:26:16.760 --> 00:26:17.410 prediction. 00:26:18.670 --> 00:26:20.420 So at the end of this, you transform 00:26:20.420 --> 00:26:22.940 what we're two continuous values into 00:26:22.940 --> 00:26:24.770 these four discrete values. 00:26:25.450 --> 00:26:27.360 Of different chunks, different 00:26:27.360 --> 00:26:30.130 partitions of the feature space and for 00:26:30.130 --> 00:26:31.350 each of those. 00:26:32.420 --> 00:26:34.850 Each of those parts of the partition. 00:26:35.810 --> 00:26:38.360 You make a prediction. 00:26:39.240 --> 00:26:41.620 A partitioning is just when you take a 00:26:41.620 --> 00:26:44.390 continuous space and divide it up into 00:26:44.390 --> 00:26:46.850 different cells that cover the entire 00:26:46.850 --> 00:26:47.400 space. 00:26:47.400 --> 00:26:49.859 That's a partition where the cells 00:26:49.860 --> 00:26:51.040 don't overlap with each other. 00:26:54.340 --> 00:26:56.460 And then if you want to classify, once 00:26:56.460 --> 00:26:57.940 you've trained your tree, you get some 00:26:57.940 --> 00:26:59.450 new test sample and you want to know is 00:26:59.450 --> 00:27:01.450 that a lemon or an orange kind of looks 00:27:01.450 --> 00:27:01.920 in between. 00:27:02.610 --> 00:27:05.295 So you is it greater than 6.5 00:27:05.295 --> 00:27:05.740 centimeters? 00:27:05.740 --> 00:27:06.185 No. 00:27:06.185 --> 00:27:08.355 Is a tight greater than 6 centimeters? 00:27:08.355 --> 00:27:08.690 No. 00:27:08.690 --> 00:27:10.110 And so therefore it's an orange 00:27:10.110 --> 00:27:10.970 according to your rule. 00:27:13.260 --> 00:27:15.053 And you could take this tree and could 00:27:15.053 --> 00:27:17.456 you could rewrite it as a set of rules, 00:27:17.456 --> 00:27:20.560 like one rule is greater than 6.5, 00:27:20.560 --> 00:27:23.478 height greater than 9.5, another rule 00:27:23.478 --> 00:27:26.020 is greater than 65, height less than 00:27:26.020 --> 00:27:27.330 9.5, and so on. 00:27:27.330 --> 00:27:28.640 There's like 4 different rules 00:27:28.640 --> 00:27:31.180 represented by this tree, and each rule 00:27:31.180 --> 00:27:33.950 corresponds to some section of the 00:27:33.950 --> 00:27:36.440 feature space, and each rule yields 00:27:36.440 --> 00:27:37.140 some prediction. 00:27:40.950 --> 00:27:44.020 So here's another example with some 00:27:44.020 --> 00:27:45.580 discrete inputs. 00:27:45.580 --> 00:27:48.030 So here the prediction problem is to 00:27:48.030 --> 00:27:49.955 tell whether or not somebody's going to 00:27:49.955 --> 00:27:50.350 wait. 00:27:50.350 --> 00:27:52.440 If they go to a restaurant and they're 00:27:52.440 --> 00:27:54.173 told they have to wait, so do they wait 00:27:54.173 --> 00:27:55.160 or do they leave? 00:27:56.290 --> 00:27:58.170 And the features are things like 00:27:58.170 --> 00:28:00.160 whether there's an alternative nearby, 00:28:00.160 --> 00:28:02.240 whether there's a bar they can wait at, 00:28:02.240 --> 00:28:03.900 whether it's Friday or Saturday, 00:28:03.900 --> 00:28:05.289 whether they're Hungry, whether the 00:28:05.290 --> 00:28:07.106 restaurants full, what the price is, 00:28:07.106 --> 00:28:08.740 whether it's raining, whether they had 00:28:08.740 --> 00:28:10.560 a Reservation, what type of restaurant 00:28:10.560 --> 00:28:12.900 is, and they would wait time. 00:28:12.900 --> 00:28:14.747 And these are all categorical, so the 00:28:14.747 --> 00:28:16.100 wait time is split into different 00:28:16.100 --> 00:28:16.540 chunks. 00:28:20.660 --> 00:28:22.670 And so you could. 00:28:24.110 --> 00:28:27.810 You could train a tree from these 00:28:27.810 --> 00:28:29.820 categorical variables, and of course I 00:28:29.820 --> 00:28:31.590 will tell you more about like how you 00:28:31.590 --> 00:28:32.390 would learn this tree. 00:28:33.960 --> 00:28:35.670 But you might have a tree like this 00:28:35.670 --> 00:28:36.500 where you say. 00:28:37.730 --> 00:28:39.770 First, are there are there people in 00:28:39.770 --> 00:28:40.370 the restaurant? 00:28:40.370 --> 00:28:41.800 Patrons means like it's a restaurant 00:28:41.800 --> 00:28:42.684 full or not. 00:28:42.684 --> 00:28:46.310 If it's not full, then you leave right 00:28:46.310 --> 00:28:47.790 away because they're just being rude. 00:28:47.790 --> 00:28:49.330 If they tell, you have to wait I guess. 00:28:49.930 --> 00:28:52.140 If it's partly full then you'll wait, 00:28:52.140 --> 00:28:54.080 and if it's full then you then you have 00:28:54.080 --> 00:28:55.680 like consider further things. 00:28:55.680 --> 00:28:58.360 If it's a WaitEstimate, really short, 00:28:58.360 --> 00:28:58.990 then you wait. 00:28:58.990 --> 00:28:59.660 Is it really long? 00:28:59.660 --> 00:29:00.170 Then you don't. 00:29:00.960 --> 00:29:03.290 Otherwise, are you hungry? 00:29:03.290 --> 00:29:04.693 If you're not, then you'll wait. 00:29:04.693 --> 00:29:06.600 If you are, then you keep thinking. 00:29:06.600 --> 00:29:08.320 So you have like, all this series of 00:29:08.320 --> 00:29:08.820 choices. 00:29:10.350 --> 00:29:12.790 That these trees and practice like if 00:29:12.790 --> 00:29:14.230 you were to use a Decision tree on 00:29:14.230 --> 00:29:14.680 MNIST. 00:29:15.810 --> 00:29:17.600 Where the features are pretty weak 00:29:17.600 --> 00:29:19.510 individually, they're just like pixel 00:29:19.510 --> 00:29:20.140 values. 00:29:20.140 --> 00:29:21.610 You can imagine that this tree could 00:29:21.610 --> 00:29:23.160 get really complicated and long. 00:29:27.970 --> 00:29:28.390 Right. 00:29:28.390 --> 00:29:31.840 So just to mostly be state. 00:29:32.450 --> 00:29:34.080 And then Decision tree. 00:29:34.080 --> 00:29:36.410 The internal nodes are Test Attributes, 00:29:36.410 --> 00:29:38.150 so it's some kind of like feature. 00:29:38.150 --> 00:29:40.050 Attribute and feature are synonymous, 00:29:40.050 --> 00:29:41.110 they're the same thing. 00:29:41.830 --> 00:29:45.880 Some kind of feature attribute and. 00:29:45.960 --> 00:29:47.420 And if it's a continuous attribute then 00:29:47.420 --> 00:29:48.420 you have to have some kind of 00:29:48.420 --> 00:29:53.420 threshold, so width greater than 6.5 or 00:29:53.420 --> 00:29:54.650 is it raining or not? 00:29:54.650 --> 00:29:56.050 Those are two examples of. 00:29:56.740 --> 00:29:57.440 Of tests. 00:29:58.370 --> 00:29:59.984 Then depending on the outcome of that 00:29:59.984 --> 00:30:02.310 test, you split in different ways, and 00:30:02.310 --> 00:30:03.860 when you're Training, you split all 00:30:03.860 --> 00:30:05.935 your data according to that test, and 00:30:05.935 --> 00:30:07.960 then you're going to solve again within 00:30:07.960 --> 00:30:09.390 each of those nodes separately. 00:30:10.480 --> 00:30:11.570 For the next Test. 00:30:12.260 --> 00:30:14.110 Until you get to a leaf node, and at 00:30:14.110 --> 00:30:16.125 the leaf node you provide an output or 00:30:16.125 --> 00:30:18.532 a prediction, which could be, which in 00:30:18.532 --> 00:30:20.480 this case is a class, in this 00:30:20.480 --> 00:30:21.780 particular example whether it's a 00:30:21.780 --> 00:30:22.540 Linear orange. 00:30:25.060 --> 00:30:25.260 Yep. 00:30:29.360 --> 00:30:31.850 So the question is how does it Decision 00:30:31.850 --> 00:30:34.700 tree account for anomalies as in late 00:30:34.700 --> 00:30:36.480 mislabeled data or really weird 00:30:36.480 --> 00:30:37.260 examples or? 00:30:50.100 --> 00:30:52.370 So the so the question is like how does 00:30:52.370 --> 00:30:54.400 it Decision tree deal with weird or 00:30:54.400 --> 00:30:55.860 unlikely examples? 00:30:55.860 --> 00:30:58.020 And that's a good question because one 00:30:58.020 --> 00:31:00.200 of the things about a Decision tree is 00:31:00.200 --> 00:31:01.510 that if you train it. 00:31:02.350 --> 00:31:04.460 If you train it, if you train the full 00:31:04.460 --> 00:31:06.560 tree, then you can always. 00:31:06.560 --> 00:31:09.970 As long as the feature vectors for each 00:31:09.970 --> 00:31:11.560 sample are unique, you can always get 00:31:11.560 --> 00:31:13.470 perfect Classification Error. 00:31:13.470 --> 00:31:14.900 A tree has no bias. 00:31:14.900 --> 00:31:16.580 You can always like fit your training 00:31:16.580 --> 00:31:18.980 data perfectly because you just keep on 00:31:18.980 --> 00:31:20.530 chopping it into smaller and smaller 00:31:20.530 --> 00:31:22.070 bits until finally the answer. 00:31:22.800 --> 00:31:24.960 So as a result, that can be dangerous 00:31:24.960 --> 00:31:26.767 because if you do have some unusual 00:31:26.767 --> 00:31:29.100 examples, you can end up creating rules 00:31:29.100 --> 00:31:31.410 based on those examples that don't 00:31:31.410 --> 00:31:32.920 generalize well tuning data. 00:31:33.640 --> 00:31:36.191 And so some things that you can do are 00:31:36.191 --> 00:31:38.119 you can stop Training, stop Training 00:31:38.120 --> 00:31:38.590 early. 00:31:38.590 --> 00:31:40.440 So you can say I'm not going to split 00:31:40.440 --> 00:31:42.530 once I only have 5 examples of my leaf 00:31:42.530 --> 00:31:43.990 node, I'm going to quit splitting and 00:31:43.990 --> 00:31:45.460 I'll just output my best guess. 00:31:46.520 --> 00:31:47.070 00:31:47.990 --> 00:31:49.240 There's also like. 00:31:52.250 --> 00:31:53.770 Probably on Tuesday. 00:31:53.770 --> 00:31:54.810 Actually, I'm going to talk about 00:31:54.810 --> 00:31:56.790 ensembles, which is ways of combining 00:31:56.790 --> 00:31:58.770 money trees, which is another way of 00:31:58.770 --> 00:31:59.840 getting rid of this problem. 00:32:01.360 --> 00:32:01.750 Question. 00:32:09.850 --> 00:32:11.190 That's a good question too. 00:32:11.190 --> 00:32:12.890 So the question is whether Decision 00:32:12.890 --> 00:32:14.500 trees are always binary. 00:32:14.500 --> 00:32:17.880 So like in this example, it's not 00:32:17.880 --> 00:32:21.640 binary, they're splitting like the 00:32:21.640 --> 00:32:23.250 Patrons is splitting based on three 00:32:23.250 --> 00:32:23.780 values. 00:32:24.510 --> 00:32:28.260 But typically they are binary. 00:32:28.260 --> 00:32:29.030 So if you're. 00:32:29.810 --> 00:32:32.277 If you're using continuous values, it 00:32:32.277 --> 00:32:32.535 will. 00:32:32.535 --> 00:32:34.410 It will almost always be binary, 00:32:34.410 --> 00:32:35.440 because you could. 00:32:35.440 --> 00:32:37.330 Even if you wanted to split continuous 00:32:37.330 --> 00:32:40.260 variables into many different chunks, 00:32:40.260 --> 00:32:42.350 you can do that through a sequence of 00:32:42.350 --> 00:32:43.380 binary decisions. 00:32:44.780 --> 00:32:47.620 In SK learn as well, their Decision 00:32:47.620 --> 00:32:50.160 trees cannot deal with like multi 00:32:50.160 --> 00:32:53.040 valued attributes and so you need to 00:32:53.040 --> 00:32:55.000 convert them into binary attributes in 00:32:55.000 --> 00:32:56.470 order to use sklearn. 00:32:57.350 --> 00:32:59.470 And I think often that's done as a 00:32:59.470 --> 00:33:01.590 design Decision, because otherwise like 00:33:01.590 --> 00:33:03.050 some features will be like 00:33:03.050 --> 00:33:04.780 intrinsically more powerful than other 00:33:04.780 --> 00:33:06.690 features if they create like more 00:33:06.690 --> 00:33:07.340 splits. 00:33:07.340 --> 00:33:09.160 So it can cause like a bias in your 00:33:09.160 --> 00:33:09.990 feature selection. 00:33:10.740 --> 00:33:12.990 So they don't have to be binary, but 00:33:12.990 --> 00:33:15.160 it's a common common setting. 00:33:21.880 --> 00:33:24.480 Alright, so the Training 4 Decision 00:33:24.480 --> 00:33:26.760 tree again without yet getting into the 00:33:26.760 --> 00:33:27.000 math. 00:33:27.710 --> 00:33:30.935 Is Recursively for each node in the 00:33:30.935 --> 00:33:32.590 tree, if the labels and the node are 00:33:32.590 --> 00:33:33.120 mixed. 00:33:33.120 --> 00:33:35.256 So to start with, we're at the of the 00:33:35.256 --> 00:33:38.030 tree and we have all this data, and so 00:33:38.030 --> 00:33:39.676 essentially there's just right now some 00:33:39.676 --> 00:33:41.025 probability that's a no, some 00:33:41.025 --> 00:33:41.950 probability that's an 784x1. 00:33:41.950 --> 00:33:43.900 Those probabilities are close to 5050. 00:33:45.530 --> 00:33:47.410 Then I'm going to choose some attribute 00:33:47.410 --> 00:33:50.020 and split the values based on the data 00:33:50.020 --> 00:33:51.200 that reaches that node. 00:33:52.310 --> 00:33:54.310 So here I Choose this attribute the 00:33:54.310 --> 00:33:55.430 tree I'm creating up there. 00:33:56.110 --> 00:33:58.060 X2 is less than .6. 00:34:00.630 --> 00:34:05.310 If it's less than .6 then I go down one 00:34:05.310 --> 00:34:07.067 branch and if it's greater than I go 00:34:07.067 --> 00:34:08.210 down the other branch. 00:34:08.210 --> 00:34:10.870 So now then I can now start making 00:34:10.870 --> 00:34:13.440 decisions separately about this region 00:34:13.440 --> 00:34:14.260 in this region. 00:34:15.910 --> 00:34:19.200 So then I Choose another node and I say 00:34:19.200 --> 00:34:21.660 if X1 is less than 7. 00:34:22.630 --> 00:34:24.360 So I create this split and this only 00:34:24.360 --> 00:34:25.490 pertains to the data. 00:34:25.490 --> 00:34:27.570 Now that came down the first node so 00:34:27.570 --> 00:34:28.989 it's this side of the data. 00:34:29.710 --> 00:34:31.292 So if it's over here, then it's a no, 00:34:31.292 --> 00:34:33.220 if it's over here, then it's an X and 00:34:33.220 --> 00:34:34.620 Now I don't need to create anymore 00:34:34.620 --> 00:34:36.690 Decision nodes for this whole region of 00:34:36.690 --> 00:34:38.870 space because I have perfect 00:34:38.870 --> 00:34:39.630 Classification. 00:34:40.760 --> 00:34:43.010 Then I go to my top side. 00:34:43.730 --> 00:34:45.390 And I can make another split. 00:34:45.390 --> 00:34:47.760 So here there's actually more than one 00:34:47.760 --> 00:34:48.015 choice. 00:34:48.015 --> 00:34:49.960 I think that's like kind of equally 00:34:49.960 --> 00:34:51.460 good, but. 00:34:51.570 --> 00:34:56.230 Again, say if X2 is less than .8, then 00:34:56.230 --> 00:34:57.960 it goes down here where I'm still 00:34:57.960 --> 00:34:58.250 unsure. 00:34:58.250 --> 00:35:00.080 If it's greater than eight, then it's 00:35:00.080 --> 00:35:00.980 definitely a red X. 00:35:03.260 --> 00:35:05.120 And then I can keep doing that until I 00:35:05.120 --> 00:35:07.190 finally have a perfect Classification 00:35:07.190 --> 00:35:08.030 in the training data. 00:35:08.810 --> 00:35:10.000 So that's the full tree. 00:35:11.070 --> 00:35:13.830 And if you could stop early, you could 00:35:13.830 --> 00:35:15.739 say I'm not going to go past like 3 00:35:15.740 --> 00:35:18.310 levels, or that I'm going to stop 00:35:18.310 --> 00:35:20.910 splitting once my leaf node doesn't 00:35:20.910 --> 00:35:23.210 have more than five examples. 00:35:39.470 --> 00:35:41.560 Well, the question was does the first 00:35:41.560 --> 00:35:42.320 split matter? 00:35:42.320 --> 00:35:43.929 So I guess there's two parts to that. 00:35:43.930 --> 00:35:45.880 One is that I will tell you how we do 00:35:45.880 --> 00:35:46.980 this computationally. 00:35:46.980 --> 00:35:48.780 So you try to greedily find like the 00:35:48.780 --> 00:35:49.900 best split every time. 00:35:50.990 --> 00:35:53.530 And the other thing is that finding the 00:35:53.530 --> 00:35:57.290 minimum size tree is like a 00:35:57.290 --> 00:35:59.540 computationally hard problem. 00:36:00.540 --> 00:36:01.766 So it's infeasible. 00:36:01.766 --> 00:36:04.530 So you end up with a greedy solution 00:36:04.530 --> 00:36:06.020 where for every node you're choosing 00:36:06.020 --> 00:36:08.200 the best split for that node. 00:36:08.200 --> 00:36:10.045 But that doesn't necessarily give you 00:36:10.045 --> 00:36:11.680 the shortest tree overall, because you 00:36:11.680 --> 00:36:13.020 don't know like what kinds of splits 00:36:13.020 --> 00:36:14.250 will be available to you later. 00:36:16.710 --> 00:36:19.050 So it does matter, but you have like 00:36:19.050 --> 00:36:20.630 there's an algorithm for doing it in a 00:36:20.630 --> 00:36:21.550 decent way, yeah. 00:36:55.320 --> 00:36:55.860 00:36:57.660 --> 00:36:59.080 There have well. 00:37:01.160 --> 00:37:02.650 How will you know that it will work for 00:37:02.650 --> 00:37:03.320 like new data? 00:37:05.000 --> 00:37:09.209 So basically if you want to know, you 00:37:09.210 --> 00:37:10.740 do always want to know, you always want 00:37:10.740 --> 00:37:12.420 to know, right, if you if the model 00:37:12.420 --> 00:37:13.620 that you learned is going to work for 00:37:13.620 --> 00:37:14.540 new data. 00:37:14.540 --> 00:37:16.370 And so that's why I typically you would 00:37:16.370 --> 00:37:18.030 carve off, if you have some Training 00:37:18.030 --> 00:37:19.800 set, you'd carve off a validation set. 00:37:20.450 --> 00:37:22.380 And you would train it say with like 00:37:22.380 --> 00:37:25.040 70% of the Training examples and test 00:37:25.040 --> 00:37:27.850 it on the 30% of the held out Samples? 00:37:28.530 --> 00:37:30.040 And then there was held out Samples 00:37:30.040 --> 00:37:32.170 will give you an estimate of how well 00:37:32.170 --> 00:37:33.260 your method works. 00:37:33.260 --> 00:37:35.250 And so then like if you find for 00:37:35.250 --> 00:37:37.626 example that I trained a full tree and 00:37:37.626 --> 00:37:39.850 of course I got like 0% Training error, 00:37:39.850 --> 00:37:41.850 but my Test error is like 40%. 00:37:42.590 --> 00:37:44.990 Then you would probably say maybe I 00:37:44.990 --> 00:37:46.803 should try Training a shorter tree and 00:37:46.803 --> 00:37:48.930 then you can like retrain it with some 00:37:48.930 --> 00:37:51.120 constraints and then test it again on 00:37:51.120 --> 00:37:52.755 your validation set and Choose like 00:37:52.755 --> 00:37:53.830 your Parameters that way. 00:37:54.860 --> 00:37:57.581 There's also I'll talk about most 00:37:57.581 --> 00:37:59.140 likely, most likely this. 00:37:59.140 --> 00:38:01.020 I was planning to do it Thursday, but 00:38:01.020 --> 00:38:02.187 I'll probably do it next Tuesday. 00:38:02.187 --> 00:38:04.580 I'll talk about ensembles, including 00:38:04.580 --> 00:38:06.622 random forests, and those are like kind 00:38:06.622 --> 00:38:09.150 of like brain dead always work methods 00:38:09.150 --> 00:38:11.080 that combine a lot of trees and are 00:38:11.080 --> 00:38:13.439 really reliable whether you have a lot 00:38:13.439 --> 00:38:16.087 of data or well, you kind of need data. 00:38:16.087 --> 00:38:17.665 But whether you have a lot of features 00:38:17.665 --> 00:38:19.850 or a little features, they always work. 00:38:20.480 --> 00:38:21.480 They always work pretty well. 00:38:23.820 --> 00:38:25.950 Right, so in prediction then you just 00:38:25.950 --> 00:38:27.560 basically descend the tree, so you 00:38:27.560 --> 00:38:29.920 check the conditions is tX2 greater 00:38:29.920 --> 00:38:32.370 than .6 blah blah blah blah blah until 00:38:32.370 --> 00:38:33.750 you find yourself in a leaf node. 00:38:34.380 --> 00:38:36.630 So for example, if I have this data 00:38:36.630 --> 00:38:38.500 point and I'm trying to classify it, I 00:38:38.500 --> 00:38:40.902 would end up following these rules down 00:38:40.902 --> 00:38:44.290 to down to the leaf node of. 00:38:45.960 --> 00:38:47.418 Yeah, like right over here, right? 00:38:47.418 --> 00:38:50.158 X2 is less than .6 and X1 is less than 00:38:50.158 --> 00:38:50.460 .7. 00:38:51.260 --> 00:38:52.740 And so that's going to be no. 00:38:53.860 --> 00:38:56.500 And if I am over here then I end up 00:38:56.500 --> 00:38:59.420 following going down to here to here. 00:39:00.480 --> 00:39:03.020 To here to here and I end up in this 00:39:03.020 --> 00:39:07.299 leaf node and so it's an X and it 00:39:07.300 --> 00:39:09.395 doesn't matter like where it falls in 00:39:09.395 --> 00:39:10.580 this part of the space. 00:39:10.580 --> 00:39:11.700 Usually this isn't like. 00:39:12.390 --> 00:39:13.770 Even something you necessarily 00:39:13.770 --> 00:39:15.520 visualize, but. 00:39:16.060 --> 00:39:18.025 But it's worth noting that even parts 00:39:18.025 --> 00:39:20.020 of your feature space that are kind of 00:39:20.020 --> 00:39:22.640 far away from any Example can still get 00:39:22.640 --> 00:39:24.360 classified by this Decision tree. 00:39:25.070 --> 00:39:27.670 And it's not necessarily the Nearest 00:39:27.670 --> 00:39:28.450 neighbor Decision. 00:39:28.450 --> 00:39:31.186 Like this star here is actually closer 00:39:31.186 --> 00:39:33.390 to the 784x1 than it is to the O's, but 00:39:33.390 --> 00:39:35.010 it would still be a no because it's on 00:39:35.010 --> 00:39:36.050 that side of the boundary. 00:39:40.650 --> 00:39:42.350 So the key question is, how do you 00:39:42.350 --> 00:39:45.810 choose what attribute to split and 00:39:45.810 --> 00:39:46.384 where to split? 00:39:46.384 --> 00:39:48.390 So how do you decide what test you're 00:39:48.390 --> 00:39:50.000 going to use for a given node? 00:39:50.920 --> 00:39:53.615 And so let's take this example. 00:39:53.615 --> 00:39:56.290 So here I've got some table of features 00:39:56.290 --> 00:39:57.180 and predictions. 00:39:58.020 --> 00:39:59.010 And if. 00:40:00.410 --> 00:40:02.280 And if I were to split, these are 00:40:02.280 --> 00:40:04.570 binary features so they just have two 00:40:04.570 --> 00:40:06.430 values T2 false I guess. 00:40:07.400 --> 00:40:07.940 If. 00:40:09.620 --> 00:40:12.440 If I split based on X1 and I go in One 00:40:12.440 --> 00:40:14.570 Direction, then it's all true. 00:40:15.200 --> 00:40:17.585 The prediction is true and if I go in 00:40:17.585 --> 00:40:19.920 the other direction then 3/4 of the 00:40:19.920 --> 00:40:21.080 time the prediction is false. 00:40:22.810 --> 00:40:26.343 If I split based on X2, then 3/4 of the 00:40:26.343 --> 00:40:27.948 time the prediction is true. 00:40:27.948 --> 00:40:31.096 If it's true and 50% of the time the 00:40:31.096 --> 00:40:32.819 prediction is false, X2 is false. 00:40:33.530 --> 00:40:36.300 So which of these features is a better 00:40:36.300 --> 00:40:37.400 Test? 00:40:39.550 --> 00:40:41.530 So how many people think that the left 00:40:41.530 --> 00:40:42.530 is a better Test? 00:40:43.790 --> 00:40:45.070 How many people think they're right is 00:40:45.070 --> 00:40:45.730 a better Test. 00:40:46.840 --> 00:40:48.380 Right the left is a better Test 00:40:48.380 --> 00:40:48.950 because. 00:40:50.620 --> 00:40:53.380 Because my uncertainty is greatly 00:40:53.380 --> 00:40:54.990 reduced on the left side. 00:40:54.990 --> 00:40:58.750 So initially, initially I had like a 00:40:58.750 --> 00:41:01.280 5/8 chance of getting it right if I 00:41:01.280 --> 00:41:02.280 just guessed true. 00:41:02.910 --> 00:41:06.706 But if I know X1, then I've got a 100% 00:41:06.706 --> 00:41:08.600 chance of getting it right, at least in 00:41:08.600 --> 00:41:09.494 the training data. 00:41:09.494 --> 00:41:13.338 If I know that X1 is true, and I've got 00:41:13.338 --> 00:41:15.132 a 3/4 chance of getting it right if I 00:41:15.132 --> 00:41:16.280 know that X1 is false. 00:41:16.280 --> 00:41:19.135 So X1 tells me a lot about the 00:41:19.135 --> 00:41:19.572 prediction. 00:41:19.572 --> 00:41:22.035 It greatly reduces my uncertainty about 00:41:22.035 --> 00:41:22.890 the prediction. 00:41:24.510 --> 00:41:26.412 And to quantify this, we need to 00:41:26.412 --> 00:41:28.560 quantify uncertainty and then be able 00:41:28.560 --> 00:41:32.350 to measure how much a certain feature 00:41:32.350 --> 00:41:33.950 reduces our uncertainty in the 00:41:33.950 --> 00:41:34.720 prediction. 00:41:34.720 --> 00:41:36.800 And that's called the information gain. 00:41:40.470 --> 00:41:44.540 So to quantify the uncertainty, I'll 00:41:44.540 --> 00:41:45.595 use these two examples. 00:41:45.595 --> 00:41:47.790 So imagine that you're flipping a coin. 00:41:47.790 --> 00:41:50.150 These are like heads and tails, or 00:41:50.150 --> 00:41:51.510 present them as zeros and ones. 00:41:52.180 --> 00:41:54.820 And so one time I've got two different 00:41:54.820 --> 00:41:56.186 sequences, let's say two different 00:41:56.186 --> 00:41:57.740 coins and one in the coins. 00:41:57.740 --> 00:42:00.120 It's a biased coin, so I end up with 00:42:00.120 --> 00:42:03.330 zeros or heads like 16 out of 18 times. 00:42:04.250 --> 00:42:06.520 And the other for the other Coin I get 00:42:06.520 --> 00:42:09.400 closer to 5058 out of. 00:42:10.050 --> 00:42:12.390 18 times I get heads so. 00:42:13.530 --> 00:42:17.520 Which of these has higher uncertainty? 00:42:18.540 --> 00:42:19.730 The left or the right? 00:42:21.330 --> 00:42:22.580 Right, correct. 00:42:22.580 --> 00:42:23.070 They're right. 00:42:23.070 --> 00:42:24.900 Has a lot higher uncertainty. 00:42:24.900 --> 00:42:27.370 So if I with that Coin, I really don't 00:42:27.370 --> 00:42:28.470 know if it's going to be heads or 00:42:28.470 --> 00:42:30.860 tails, but on the left side, I'm pretty 00:42:30.860 --> 00:42:31.820 sure it's going to be heads. 00:42:32.590 --> 00:42:33.360 Or zeros. 00:42:34.720 --> 00:42:36.770 So we can measure that with this 00:42:36.770 --> 00:42:38.645 function called Entropy. 00:42:38.645 --> 00:42:41.350 So the entropy is a measure of 00:42:41.350 --> 00:42:42.030 uncertainty. 00:42:42.960 --> 00:42:45.740 And it's defined as the negative sum 00:42:45.740 --> 00:42:48.070 over all the values of some variable of 00:42:48.070 --> 00:42:50.220 the probability of that value. 00:42:51.020 --> 00:42:53.490 Times the log probability that value, 00:42:53.490 --> 00:42:56.470 and people usually sometimes use like 00:42:56.470 --> 00:42:57.520 log base 2. 00:42:58.630 --> 00:43:00.700 Just because that way the Entropy 00:43:00.700 --> 00:43:02.550 ranges from zero to 1 if you have 00:43:02.550 --> 00:43:03.490 binary variables. 00:43:07.600 --> 00:43:10.820 So for this case here, the Entropy 00:43:10.820 --> 00:43:13.280 would be -, 8 ninths, because eight out 00:43:13.280 --> 00:43:14.600 of nine times it's zero. 00:43:15.270 --> 00:43:17.300 Times log two of eight ninths. 00:43:18.230 --> 00:43:21.210 Minus one ninth times, log 2 of 1 ninth 00:43:21.210 --> 00:43:22.790 and that works out to about 1/2. 00:43:24.370 --> 00:43:28.480 And over here the Entropy is -, 4 00:43:28.480 --> 00:43:30.270 ninths because four out of nine times, 00:43:30.270 --> 00:43:32.900 or 8 out of 18 times, it's a 0. 00:43:34.410 --> 00:43:37.104 Times log 24 ninths, minus five ninths, 00:43:37.104 --> 00:43:39.010 times log two of five ninths, and 00:43:39.010 --> 00:43:41.490 that's about 99. 00:43:43.430 --> 00:43:45.280 The Entropy measure is how surprised 00:43:45.280 --> 00:43:47.595 are we by some new value of this 00:43:47.595 --> 00:43:47.830 Sequence? 00:43:47.830 --> 00:43:50.123 How surprised are we likely to be in, 00:43:50.123 --> 00:43:52.460 or how much information does it convey 00:43:52.460 --> 00:43:54.895 that we know that we're in this 00:43:54.895 --> 00:43:56.974 Sequence, or more generally, that we 00:43:56.974 --> 00:43:57.940 know some feature? 00:44:01.100 --> 00:44:03.425 So this is just showing the Entropy if 00:44:03.425 --> 00:44:05.450 the probability if you have a binary 00:44:05.450 --> 00:44:06.340 variable X. 00:44:07.110 --> 00:44:09.720 And the probability of X is 0, then 00:44:09.720 --> 00:44:12.180 your Entropy is 0 because you always 00:44:12.180 --> 00:44:15.127 know that if probability of X = 2 is 00:44:15.127 --> 00:44:16.818 zero, that means that probability of X 00:44:16.818 --> 00:44:18.210 equals false is 1. 00:44:18.860 --> 00:44:20.530 And so therefore you have complete 00:44:20.530 --> 00:44:22.470 confidence that the value will be 00:44:22.470 --> 00:44:22.810 false. 00:44:24.070 --> 00:44:27.740 If probability of X is true is 1, then 00:44:27.740 --> 00:44:29.590 you have complete confidence that the 00:44:29.590 --> 00:44:30.650 value will be true. 00:44:31.440 --> 00:44:35.570 But if it's .5, then you have no 00:44:35.570 --> 00:44:37.120 information about whether it's true or 00:44:37.120 --> 00:44:39.520 false, and so you have maximum entropy, 00:44:39.520 --> 00:44:40.190 which is 1. 00:44:45.770 --> 00:44:47.280 So here's another example. 00:44:47.280 --> 00:44:49.340 So suppose that we've got two 00:44:49.340 --> 00:44:51.070 variables, whether it's raining or not, 00:44:51.070 --> 00:44:52.220 and whether it's cloudy or not. 00:44:52.820 --> 00:44:55.700 And we've observed 100 days and marked 00:44:55.700 --> 00:44:57.260 down whether it's rainy or cloudy. 00:44:58.870 --> 00:45:00.150 Many and or Cloudy. 00:45:00.930 --> 00:45:01.500 00:45:02.600 --> 00:45:06.300 So 24 days it was raining and cloudy. 00:45:06.300 --> 00:45:08.210 One day it was raining and not Cloudy. 00:45:09.320 --> 00:45:11.244 25 days it was not raining and cloudy 00:45:11.244 --> 00:45:13.409 and 50 days it was not raining and not 00:45:13.409 --> 00:45:13.649 Cloudy. 00:45:15.620 --> 00:45:17.980 The probabilities are just dividing by 00:45:17.980 --> 00:45:18.766 the total there. 00:45:18.766 --> 00:45:20.850 So the probability of Cloudy and not 00:45:20.850 --> 00:45:22.630 raining is 25 out of 100. 00:45:24.040 --> 00:45:26.660 And so I can also compute an Entropy of 00:45:26.660 --> 00:45:27.855 this whole joint distribution. 00:45:27.855 --> 00:45:31.150 So I can say that the entropy of X&Y 00:45:31.150 --> 00:45:33.446 together is the sum all the different 00:45:33.446 --> 00:45:35.428 values of X and the over all the 00:45:35.428 --> 00:45:36.419 different values of Y. 00:45:37.060 --> 00:45:39.770 Of probability of X&Y times log 2, 00:45:39.770 --> 00:45:41.920 probability of X&Y, and then that's all 00:45:41.920 --> 00:45:42.880 just like written out here. 00:45:43.650 --> 00:45:45.115 And then I get some Entropy value. 00:45:45.115 --> 00:45:47.940 And sometimes people call those units 00:45:47.940 --> 00:45:51.490 bits, so 156 bits because that's the 00:45:51.490 --> 00:45:53.008 amount of, that's the number of bits 00:45:53.008 --> 00:45:54.680 that I would need that I would expect 00:45:54.680 --> 00:45:55.040 to. 00:45:55.790 --> 00:45:57.780 Be able to like represent this. 00:45:58.630 --> 00:45:59.700 This information. 00:46:00.430 --> 00:46:04.395 If you if it were always not Cloudy and 00:46:04.395 --> 00:46:04.990 not raining. 00:46:05.850 --> 00:46:08.020 If it were 100% of the time not Cloudy 00:46:08.020 --> 00:46:10.280 and not raining, then you'd have 0 bits 00:46:10.280 --> 00:46:11.830 because you don't need any data to 00:46:11.830 --> 00:46:12.810 represent the. 00:46:13.710 --> 00:46:15.770 That uncertainty, it's just always 00:46:15.770 --> 00:46:16.300 true. 00:46:16.300 --> 00:46:18.300 I mean it's always like one value. 00:46:18.300 --> 00:46:20.790 So 15 bits means that you have pretty 00:46:20.790 --> 00:46:21.490 high uncertainty. 00:46:25.250 --> 00:46:27.680 There's also a concept called specific 00:46:27.680 --> 00:46:28.510 Entropy. 00:46:28.510 --> 00:46:29.780 So that is. 00:46:29.780 --> 00:46:33.560 That means that if one thing, then how 00:46:33.560 --> 00:46:34.516 much does that? 00:46:34.516 --> 00:46:36.610 How much uncertainty do you have left? 00:46:37.460 --> 00:46:41.170 So, for example, what is the entropy of 00:46:41.170 --> 00:46:43.610 cloudiness given that I know that it's 00:46:43.610 --> 00:46:44.000 raining? 00:46:45.420 --> 00:46:48.940 And the Conditional Entropy is very 00:46:48.940 --> 00:46:51.280 similar form, it's just negative sum 00:46:51.280 --> 00:46:52.720 over the values of the. 00:46:53.710 --> 00:46:54.970 The thing that you're measuring the 00:46:54.970 --> 00:46:55.780 Entropy over. 00:46:56.800 --> 00:46:58.880 The probability of that given the thing 00:46:58.880 --> 00:46:59.500 that. 00:47:00.150 --> 00:47:03.610 Times the log probability of Y given X, 00:47:03.610 --> 00:47:04.760 where Y is the thing you're measuring 00:47:04.760 --> 00:47:06.400 the uncertainty of, and X is a thing 00:47:06.400 --> 00:47:06.850 that you know. 00:47:09.200 --> 00:47:12.660 So if I know that it's Cloudy, then 00:47:12.660 --> 00:47:15.690 there's a 24 out of 25 chance that 00:47:15.690 --> 00:47:16.150 it's. 00:47:17.340 --> 00:47:17.950 Wait, no. 00:47:17.950 --> 00:47:19.910 If I know that it's raining, sorry. 00:47:19.910 --> 00:47:21.599 If I know that it's raining, then 00:47:21.600 --> 00:47:23.931 there's a 24 out of 25 chance that it's 00:47:23.931 --> 00:47:24.430 Cloudy, right? 00:47:24.430 --> 00:47:26.190 And then one out of 25 chance that it's 00:47:26.190 --> 00:47:26.760 not Cloudy. 00:47:27.600 --> 00:47:30.340 So I get 24 to 25 there and one out of 00:47:30.340 --> 00:47:33.070 25 there, and now my Entropy is greatly 00:47:33.070 --> 00:47:33.660 reduced. 00:47:39.810 --> 00:47:41.280 And then you can also measure. 00:47:41.930 --> 00:47:44.250 In expected Conditional Entropy. 00:47:46.020 --> 00:47:50.570 So that's just the probability of. 00:47:50.570 --> 00:47:53.870 That's just taking the specific 00:47:53.870 --> 00:47:54.940 Conditional Entropy. 00:47:55.780 --> 00:47:58.660 At times the probability of each of the 00:47:58.660 --> 00:48:00.360 values that I might know. 00:48:01.260 --> 00:48:03.710 Summed up over the different values, 00:48:03.710 --> 00:48:04.180 so. 00:48:04.900 --> 00:48:06.130 The. 00:48:06.820 --> 00:48:09.920 The expected Conditional value Entropy 00:48:09.920 --> 00:48:11.950 for knowing whether or not it's raining 00:48:11.950 --> 00:48:15.179 would be the Conditional Entropy. 00:48:16.040 --> 00:48:19.010 Of it raining if I know it's raining. 00:48:19.920 --> 00:48:21.280 Times the probability that it's 00:48:21.280 --> 00:48:21.720 raining. 00:48:22.460 --> 00:48:24.460 Plus the. 00:48:25.190 --> 00:48:28.270 Entropy of cloudiness given that it's 00:48:28.270 --> 00:48:30.210 not raining, times the probability 00:48:30.210 --> 00:48:30.900 that's not raining. 00:48:33.530 --> 00:48:35.550 And that's also equal to this thing. 00:48:42.960 --> 00:48:43.400 Right. 00:48:43.400 --> 00:48:46.168 So if I want to know what is the 00:48:46.168 --> 00:48:47.790 entropy of cloudiness, I guess I said 00:48:47.790 --> 00:48:48.730 it a little early. 00:48:48.730 --> 00:48:50.890 What is the entropy of cloudiness given 00:48:50.890 --> 00:48:52.720 whether that we know whether or not 00:48:52.720 --> 00:48:53.340 it's raining? 00:48:54.310 --> 00:48:56.240 Then that is. 00:48:56.850 --> 00:48:59.540 Going to be like 1/4, which is the 00:48:59.540 --> 00:49:02.009 probability that it's raining, is that 00:49:02.010 --> 00:49:02.320 right? 00:49:02.320 --> 00:49:04.790 25 out of 100 times it's raining. 00:49:05.490 --> 00:49:08.225 So 1/4 is the probability that it's 00:49:08.225 --> 00:49:11.240 raining times the Entropy of cloudiness 00:49:11.240 --> 00:49:13.840 given that it's raining plus three 00:49:13.840 --> 00:49:15.710 quarter times it's not raining times 00:49:15.710 --> 00:49:17.570 the entropy of the cloudiness given 00:49:17.570 --> 00:49:18.810 that it's not raining. 00:49:20.470 --> 00:49:23.420 So that's a measure of how much does 00:49:23.420 --> 00:49:25.930 knowing whether or not it's rainy, or 00:49:25.930 --> 00:49:28.470 how much uncertainty do I have left if 00:49:28.470 --> 00:49:29.880 I know whether or not it's raining. 00:49:32.430 --> 00:49:34.030 How much do I expect to have left? 00:49:37.700 --> 00:49:39.800 So some useful things to know is that 00:49:39.800 --> 00:49:41.585 the Entropy is always nonnegative. 00:49:41.585 --> 00:49:43.580 You can never have negative Entropy, 00:49:43.580 --> 00:49:45.410 but do make sure you remember. 00:49:46.480 --> 00:49:47.310 00:49:48.750 --> 00:49:50.380 So do make sure you remember these 00:49:50.380 --> 00:49:53.390 negative signs in this like 00:49:53.390 --> 00:49:54.910 probability, otherwise if you end up 00:49:54.910 --> 00:49:56.780 with a negative Entropy that you left 00:49:56.780 --> 00:49:57.490 something out. 00:49:59.760 --> 00:50:02.815 You also have this chain rule, so the 00:50:02.815 --> 00:50:06.320 entropy X&Y is the entropy of X given Y 00:50:06.320 --> 00:50:08.580 plus the entropy of Y, which kind of 00:50:08.580 --> 00:50:10.260 makes sense because the Entropy of 00:50:10.260 --> 00:50:11.280 knowing two things. 00:50:12.310 --> 00:50:14.540 Of the values of two things, is the 00:50:14.540 --> 00:50:15.914 value of knowing one. 00:50:15.914 --> 00:50:18.785 Is the OR sorry, the Entropy or the 00:50:18.785 --> 00:50:20.199 uncertainty of knowing two things? 00:50:20.199 --> 00:50:22.179 Is the uncertainty of knowing one of 00:50:22.180 --> 00:50:22.580 them? 00:50:23.280 --> 00:50:24.940 Plus the uncertainty of knowing the 00:50:24.940 --> 00:50:26.515 other one, given that you already know 00:50:26.515 --> 00:50:27.350 One South. 00:50:27.350 --> 00:50:30.169 It's either Entropy of X given Y plus 00:50:30.169 --> 00:50:32.323 Entropy of Y, or Entropy of Y given X 00:50:32.323 --> 00:50:33.250 plus Entropy of 784x1. 00:50:34.640 --> 00:50:38.739 X&Y are independent, then Entropy of Y 00:50:38.740 --> 00:50:40.659 given X is equal the entropy of Y. 00:50:42.870 --> 00:50:44.520 Meaning that 784X1 doesn't reduce our 00:50:44.520 --> 00:50:45.240 uncertainty at all. 00:50:46.530 --> 00:50:48.845 And Entropy of anything with itself is 00:50:48.845 --> 00:50:50.330 0, because once you know it, then 00:50:50.330 --> 00:50:51.480 there's no uncertainty anymore. 00:50:52.880 --> 00:50:53.390 And then? 00:50:54.110 --> 00:50:57.970 If you do know something, Entropy of Y 00:50:57.970 --> 00:50:59.780 given X or at least has to be less than 00:50:59.780 --> 00:51:01.430 or equal the entropy of Y. 00:51:01.430 --> 00:51:04.020 So knowing something can never increase 00:51:04.020 --> 00:51:04.690 your uncertainty. 00:51:07.660 --> 00:51:09.520 So then finally we can get to this 00:51:09.520 --> 00:51:11.132 information gain. 00:51:11.132 --> 00:51:14.730 So information gain is the change in 00:51:14.730 --> 00:51:17.530 the Entropy due to learning something 00:51:17.530 --> 00:51:17.810 new. 00:51:20.100 --> 00:51:23.310 So I can say, for example, what is? 00:51:23.310 --> 00:51:26.160 How much does knowing whether or not 00:51:26.160 --> 00:51:27.010 it's rainy? 00:51:27.960 --> 00:51:30.610 Reduce my uncertainty of cloudiness. 00:51:31.620 --> 00:51:34.542 So that would be the Entropy of 00:51:34.542 --> 00:51:37.242 cloudiness minus the entropy of 00:51:37.242 --> 00:51:39.120 cloudiness given whether or not it's 00:51:39.120 --> 00:51:39.450 raining. 00:51:41.710 --> 00:51:43.990 So that's the Entropy of cloudiness 00:51:43.990 --> 00:51:46.500 minus the entropy of cloudiness given 00:51:46.500 --> 00:51:47.640 whether it's raining. 00:51:47.640 --> 00:51:49.660 And that's 25 bits. 00:51:49.660 --> 00:51:50.993 So that's like the value. 00:51:50.993 --> 00:51:52.860 It's essentially the value of knowing 00:51:52.860 --> 00:51:54.100 whether or not it's meaning. 00:51:59.210 --> 00:52:01.140 And then finally we can use this in our 00:52:01.140 --> 00:52:02.140 Decision tree. 00:52:02.140 --> 00:52:03.660 So if we recall. 00:52:04.300 --> 00:52:07.310 The Decision tree algorithm is that. 00:52:08.410 --> 00:52:10.940 If I'm trying to I go through like 00:52:10.940 --> 00:52:12.700 splitting my data. 00:52:13.550 --> 00:52:15.050 Choose some Test. 00:52:15.050 --> 00:52:17.280 According to the test, I split the data 00:52:17.280 --> 00:52:18.970 into different nodes and then I choose 00:52:18.970 --> 00:52:20.440 a new test for each of those nodes. 00:52:21.440 --> 00:52:22.840 So the key thing we're trying to figure 00:52:22.840 --> 00:52:24.007 out is how do we do that Test? 00:52:24.007 --> 00:52:25.800 How do we choose the features or 00:52:25.800 --> 00:52:27.480 attributes and the splitting value? 00:52:28.370 --> 00:52:30.100 To try to split things into different 00:52:30.100 --> 00:52:32.030 classes, or in other words, to try to 00:52:32.030 --> 00:52:33.640 reduce the uncertainty of our 00:52:33.640 --> 00:52:34.150 prediction. 00:52:36.190 --> 00:52:39.790 And the solution is to choose the 00:52:39.790 --> 00:52:42.450 attribute to choose the Test that 00:52:42.450 --> 00:52:44.780 maximizes the information gain. 00:52:44.780 --> 00:52:46.770 In other words, that reduces the 00:52:46.770 --> 00:52:49.600 entropy of the most for the current 00:52:49.600 --> 00:52:50.370 data in that node. 00:52:52.000 --> 00:52:52.530 So. 00:52:53.260 --> 00:52:56.478 What you would do is for each for each 00:52:56.478 --> 00:52:58.700 discrete attribute or discrete feature. 00:52:59.630 --> 00:53:02.063 You can compute the information gain of 00:53:02.063 --> 00:53:04.140 using that using that feature. 00:53:04.140 --> 00:53:06.620 So in the case of. 00:53:07.360 --> 00:53:08.280 Go back a bit. 00:53:09.010 --> 00:53:11.670 To this simple true false all right, so 00:53:11.670 --> 00:53:12.650 for example. 00:53:13.650 --> 00:53:15.520 Here I started out with a pretty high 00:53:15.520 --> 00:53:17.550 Entropy, close to one because 5/8 of 00:53:17.550 --> 00:53:18.050 the time. 00:53:18.690 --> 00:53:20.850 The value of Y is true and three it's 00:53:20.850 --> 00:53:21.180 false. 00:53:22.030 --> 00:53:26.620 And so I can say for X1, what's my 00:53:26.620 --> 00:53:28.970 Entropy after X1? 00:53:28.970 --> 00:53:31.020 It's a 5050 chance that it goes either 00:53:31.020 --> 00:53:31.313 way. 00:53:31.313 --> 00:53:34.020 So this will be 50 * 0 because the 00:53:34.020 --> 00:53:36.541 Entropy here is 0 and this will be 50 00:53:36.541 --> 00:53:36.815 times. 00:53:36.815 --> 00:53:38.630 I don't know, one or something, 00:53:38.630 --> 00:53:40.659 whatever that Entropy is, and so this 00:53:40.659 --> 00:53:42.100 Entropy will be really low. 00:53:43.000 --> 00:53:45.700 And this Entropy is just about as high 00:53:45.700 --> 00:53:46.590 as I started with. 00:53:46.590 --> 00:53:48.330 It's only a little bit lower maybe 00:53:48.330 --> 00:53:50.510 because if I go this way, I have 00:53:50.510 --> 00:53:52.691 Entropy of 1, there's a 50% chance of 00:53:52.691 --> 00:53:55.188 that, and if I go this way, then I have 00:53:55.188 --> 00:53:56.721 lower Entropy and there's a 50% chance 00:53:56.721 --> 00:53:57.159 of that. 00:53:57.870 --> 00:54:00.010 And so my information gain is my 00:54:00.010 --> 00:54:01.600 initial entropy of Y. 00:54:02.980 --> 00:54:06.550 Minus the entropy of each of these, and 00:54:06.550 --> 00:54:08.005 here the Entropy gain. 00:54:08.005 --> 00:54:10.210 The information gain of X1 is much 00:54:10.210 --> 00:54:12.940 lower than X2 and so I Choose X1. 00:54:18.810 --> 00:54:20.420 So if I have discrete values, I just 00:54:20.420 --> 00:54:22.449 compute the information gain for the 00:54:22.450 --> 00:54:24.290 current node for each of those discrete 00:54:24.290 --> 00:54:25.725 values, and then I choose the one with 00:54:25.725 --> 00:54:26.860 the highest information gain. 00:54:27.780 --> 00:54:29.650 If I have continuous values, it's 00:54:29.650 --> 00:54:31.576 slightly more complicated because then 00:54:31.576 --> 00:54:34.395 I have to also choose a threshold in 00:54:34.395 --> 00:54:36.230 the lemons and. 00:54:36.920 --> 00:54:40.150 And oranges we were choosing saying if 00:54:40.150 --> 00:54:42.010 the height is greater than six then we 00:54:42.010 --> 00:54:42.620 go one way. 00:54:44.560 --> 00:54:46.640 So we have to choose which feature and 00:54:46.640 --> 00:54:47.420 which threshold. 00:54:48.430 --> 00:54:49.580 So typically. 00:54:51.060 --> 00:54:53.512 Something this I don't know. 00:54:53.512 --> 00:54:56.295 Like who thought putting a projector in 00:54:56.295 --> 00:54:57.930 a jewel would be like a nice way to? 00:54:58.590 --> 00:55:00.260 Right and stuff, but anyway. 00:55:04.700 --> 00:55:06.340 But at least it's something, all right? 00:55:06.340 --> 00:55:08.400 So let's say that I have some feature. 00:55:09.420 --> 00:55:11.910 And I've got like some different 00:55:11.910 --> 00:55:13.240 classes and that feature. 00:55:16.190 --> 00:55:18.560 So what I would do is I would usually 00:55:18.560 --> 00:55:19.950 you would sort the values. 00:55:20.890 --> 00:55:22.440 And you're never going to want to split 00:55:22.440 --> 00:55:24.010 between two of the same class, so I 00:55:24.010 --> 00:55:26.469 would never split between the two X's, 00:55:26.470 --> 00:55:29.250 because that's always going to be worse 00:55:29.250 --> 00:55:31.070 than some split that's between 00:55:31.070 --> 00:55:31.930 different classes. 00:55:32.630 --> 00:55:35.450 So I can consider the thresholds that 00:55:35.450 --> 00:55:35.810 are. 00:55:36.460 --> 00:55:37.730 Between different classes. 00:55:42.380 --> 00:55:44.000 Really. 00:55:44.000 --> 00:55:44.530 No. 00:55:46.130 --> 00:55:48.380 Yeah, I can, but I'm not going to draw 00:55:48.380 --> 00:55:50.220 that long, so it's not worth it to me 00:55:50.220 --> 00:55:50.860 to move on here. 00:55:50.860 --> 00:55:52.160 Then I have to move my laptop and. 00:55:53.030 --> 00:55:56.030 So I'm fine. 00:55:56.750 --> 00:55:59.310 So I would choose these two thresholds. 00:55:59.310 --> 00:56:01.680 If it's this threshold, then it's 00:56:01.680 --> 00:56:04.152 basically two and zero. 00:56:04.152 --> 00:56:07.470 So it's a very low Entropy here. 00:56:07.470 --> 00:56:10.420 And the probability of that is 2 out of 00:56:10.420 --> 00:56:11.505 five, right? 00:56:11.505 --> 00:56:16.820 So it would be 0.4 * 0 is the. 00:56:17.460 --> 00:56:18.580 Entropy on this side. 00:56:19.570 --> 00:56:20.820 And if I go this way? 00:56:21.670 --> 00:56:23.500 Then it's going to be. 00:56:24.440 --> 00:56:25.290 Then I've got. 00:56:26.660 --> 00:56:27.840 Sorry, two out of seven. 00:56:29.750 --> 00:56:31.320 Out of seven times. 00:56:32.470 --> 00:56:33.930 Times Entropy of 0 this way. 00:56:34.650 --> 00:56:37.630 And if I go this way, then it's five 00:56:37.630 --> 00:56:38.020 out of. 00:56:38.980 --> 00:56:39.760 7. 00:56:41.040 --> 00:56:41.770 Times. 00:56:44.510 --> 00:56:47.560 Two out of five times log. 00:56:52.980 --> 00:56:53.690 Thank you. 00:56:53.690 --> 00:56:55.330 I always forget the minus sign. 00:56:56.140 --> 00:56:58.270 OK, so minus 5 to 7, which is a 00:56:58.270 --> 00:56:59.880 probability that I go in this direction 00:56:59.880 --> 00:57:03.805 times one out of five times log one out 00:57:03.805 --> 00:57:04.700 of five. 00:57:05.550 --> 00:57:07.760 Plus four out of five. 00:57:09.170 --> 00:57:10.710 Four to five times log. 00:57:13.360 --> 00:57:14.100 Right. 00:57:14.100 --> 00:57:15.750 So there's a one fifth chance that it's 00:57:15.750 --> 00:57:16.270 an X. 00:57:17.350 --> 00:57:19.180 I do 1/5 times log 1/5. 00:57:19.820 --> 00:57:22.200 Minus 4/5 chance that it's a no, so 00:57:22.200 --> 00:57:23.790 minus 4/5 times log four fifth. 00:57:24.510 --> 00:57:26.210 And this whole thing is the Entropy 00:57:26.210 --> 00:57:27.140 after that split. 00:57:28.590 --> 00:57:30.650 And then likewise I can evaluate this 00:57:30.650 --> 00:57:32.850 split as well and so. 00:57:33.620 --> 00:57:35.650 Out of these two splits, which one do 00:57:35.650 --> 00:57:37.190 you think will have the most 00:57:37.190 --> 00:57:38.040 information gain? 00:57:41.220 --> 00:57:43.320 Yeah, the left split, the first one has 00:57:43.320 --> 00:57:45.050 the most information gain because then 00:57:45.050 --> 00:57:47.168 I get a confident Decision about two 00:57:47.168 --> 00:57:49.943 X's and like 4 out of five chance of 00:57:49.943 --> 00:57:51.739 getting it right on the other side, 00:57:51.740 --> 00:57:53.520 where if I choose the right split, I 00:57:53.520 --> 00:57:56.791 only get a perfect confidence about 1X 00:57:56.791 --> 00:57:59.529 and A2 out of three chance of getting 00:57:59.529 --> 00:58:00.529 it right on the other side. 00:58:15.920 --> 00:58:19.580 OK, so if I continuous features I would 00:58:19.580 --> 00:58:21.490 just try all the different like 00:58:21.490 --> 00:58:23.110 candidate thresholds for all those 00:58:23.110 --> 00:58:24.690 features and then choose the best one. 00:58:26.430 --> 00:58:28.360 And. 00:58:28.460 --> 00:58:29.720 She's the best one, all right. 00:58:29.720 --> 00:58:30.090 That's it. 00:58:30.090 --> 00:58:31.430 And then I do that for all the nodes, 00:58:31.430 --> 00:58:32.590 then I do it Recursively. 00:58:33.670 --> 00:58:35.660 So if you have a lot of features and a 00:58:35.660 --> 00:58:37.050 lot of data, this can kind of take a 00:58:37.050 --> 00:58:37.600 long time. 00:58:38.250 --> 00:58:40.610 But I mean these operations are super 00:58:40.610 --> 00:58:41.710 fast so. 00:58:42.980 --> 00:58:45.919 In practice, when you run it so in 00:58:45.920 --> 00:58:48.380 homework two, I'll have you train tree 00:58:48.380 --> 00:58:50.930 train forests of Decision trees, where 00:58:50.930 --> 00:58:54.165 you train 100 of them for example, and 00:58:54.165 --> 00:58:56.090 it takes like a few seconds, so it's 00:58:56.090 --> 00:58:57.204 like pretty fast. 00:58:57.204 --> 00:58:59.070 These are these are actually not that 00:58:59.070 --> 00:59:01.030 computationally expensive, even though 00:59:01.030 --> 00:59:02.610 doing it manually would take forever. 00:59:05.590 --> 00:59:06.980 So. 00:59:08.860 --> 00:59:10.970 We're close to the we're close to the 00:59:10.970 --> 00:59:11.690 end of the lecture. 00:59:12.320 --> 00:59:14.320 But I will give you just a second to 00:59:14.320 --> 00:59:15.230 catch your breath. 00:59:15.230 --> 00:59:17.030 And while you're doing that, think 00:59:17.030 --> 00:59:17.690 about. 00:59:19.060 --> 00:59:22.640 If I were to try and in this case I'm 00:59:22.640 --> 00:59:23.760 showing like all the different 00:59:23.760 --> 00:59:25.210 examples, the numbers are different 00:59:25.210 --> 00:59:27.530 examples there and the color is whether 00:59:27.530 --> 00:59:28.270 they wait or not. 00:59:28.850 --> 00:59:30.570 And I'm trying to decide whether I'm 00:59:30.570 --> 00:59:33.090 going to make a decision based on the 00:59:33.090 --> 00:59:35.096 type of restaurant or based on whether 00:59:35.096 --> 00:59:35.860 the restaurant's full. 00:59:36.490 --> 00:59:40.840 So take a moment to stretch or zone 00:59:40.840 --> 00:59:42.760 out, and then I'll ask you what the 00:59:42.760 --> 00:59:43.200 answer is. 01:00:05.270 --> 01:00:06.606 Part of it, yeah. 01:00:06.606 --> 01:00:08.755 So this is all Training one tree. 01:00:08.755 --> 01:00:10.840 And for a random forest you just 01:00:10.840 --> 01:00:14.246 randomly sample features and randomly 01:00:14.246 --> 01:00:16.760 sample data, and then you train a tree 01:00:16.760 --> 01:00:19.250 and then you do that like N times and 01:00:19.250 --> 01:00:20.600 then you average the predictions. 01:00:27.420 --> 01:00:27.810 Yeah. 01:00:30.860 --> 01:00:33.610 And so essentially, since the previous 01:00:33.610 --> 01:00:35.440 Entropy is fixed when you're trying to 01:00:35.440 --> 01:00:36.140 make a decision. 01:00:36.910 --> 01:00:38.739 You're just essentially choosing the 01:00:38.740 --> 01:00:41.810 Decision, choosing the attribute that 01:00:41.810 --> 01:00:45.320 will minimize your expected Entropy 01:00:45.320 --> 01:00:47.160 after, like given that attribute. 01:00:57.950 --> 01:01:00.790 Alright, so how many people think that 01:01:00.790 --> 01:01:02.610 we should split? 01:01:03.300 --> 01:01:04.710 How many people think we should split 01:01:04.710 --> 01:01:05.580 based on type? 01:01:08.180 --> 01:01:09.580 How many people think we should split 01:01:09.580 --> 01:01:10.520 based on Patrons? 01:01:12.730 --> 01:01:13.680 Yeah, OK. 01:01:14.430 --> 01:01:17.380 So I would say the answer is Patrons 01:01:17.380 --> 01:01:19.870 and because splitting based on type. 01:01:20.590 --> 01:01:22.310 I end up no matter what type of 01:01:22.310 --> 01:01:24.200 restaurant is, I end up with an equal 01:01:24.200 --> 01:01:26.120 number of greens and Reds. 01:01:26.120 --> 01:01:30.140 So green green means I didn't like say 01:01:30.140 --> 01:01:32.672 it very clearly, but green means that 01:01:32.672 --> 01:01:35.842 you think that you go, that you wait, 01:01:35.842 --> 01:01:37.540 and red means that you don't wait. 01:01:38.460 --> 01:01:40.820 So type tells me nothing, right? 01:01:40.820 --> 01:01:42.310 It doesn't help me split anything at 01:01:42.310 --> 01:01:42.455 all. 01:01:42.455 --> 01:01:44.898 I knew initially I had complete Entropy 01:01:44.898 --> 01:01:47.780 Entropy of 1 and after knowing type I 01:01:47.780 --> 01:01:48.880 still have Entropy of 1. 01:01:49.900 --> 01:01:52.140 Where if I know Patrons, then a lot of 01:01:52.140 --> 01:01:55.720 the time I have my Decision, and only 01:01:55.720 --> 01:01:57.230 some fraction of the time I still have 01:01:57.230 --> 01:01:57.590 to. 01:01:57.590 --> 01:01:59.040 I need more information. 01:02:00.990 --> 01:02:02.700 So here's like all the math. 01:02:04.250 --> 01:02:05.230 To go through that but. 01:02:08.910 --> 01:02:11.790 All right, So what if I? 01:02:12.780 --> 01:02:14.730 So sometimes a lot of times trees are 01:02:14.730 --> 01:02:16.930 used for continuous values and then 01:02:16.930 --> 01:02:18.320 it's called a Regression tree. 01:02:20.760 --> 01:02:22.960 The Regression tree is learned in the 01:02:22.960 --> 01:02:23.510 same way. 01:02:24.570 --> 01:02:29.490 Except that you would use the instead 01:02:29.490 --> 01:02:30.840 of, sorry. 01:02:32.260 --> 01:02:34.170 In the Regression tree, it's the same 01:02:34.170 --> 01:02:36.530 way, but you're typically trying to 01:02:36.530 --> 01:02:38.703 minimize the sum of squared error of 01:02:38.703 --> 01:02:41.862 the node instead of minimizing the 01:02:41.862 --> 01:02:42.427 cross entropy. 01:02:42.427 --> 01:02:44.050 You could still do it actually based on 01:02:44.050 --> 01:02:45.500 cross entropy if you're seeing like 01:02:45.500 --> 01:02:47.770 Gaussian distributions, but here let me 01:02:47.770 --> 01:02:48.900 show you an example. 01:02:54.540 --> 01:02:55.230 So. 01:02:57.600 --> 01:02:59.170 Let's just say I'm doing like one 01:02:59.170 --> 01:03:00.700 feature, let's say like. 01:03:01.480 --> 01:03:04.610 This is my feature X and my prediction 01:03:04.610 --> 01:03:05.240 value. 01:03:06.000 --> 01:03:07.830 Is the number that I'm putting here. 01:03:18.340 --> 01:03:18.690 OK. 01:03:19.430 --> 01:03:20.160 So. 01:03:21.350 --> 01:03:22.980 I'm trying to predict what this number 01:03:22.980 --> 01:03:25.460 is given like where I fell on this X 01:03:25.460 --> 01:03:25.950 axis. 01:03:27.030 --> 01:03:28.940 So the best split I could do is 01:03:28.940 --> 01:03:30.550 probably like here, right? 01:03:31.330 --> 01:03:33.800 And if I take this split, then I would 01:03:33.800 --> 01:03:37.313 say that if I'm in this side of the 01:03:37.313 --> 01:03:37.879 split. 01:03:38.640 --> 01:03:42.450 Then my prediction is 4 out of three, 01:03:42.450 --> 01:03:44.560 which is the average of the values that 01:03:44.560 --> 01:03:45.690 are on this side of the split. 01:03:46.510 --> 01:03:48.710 And if I'm on this side of the split, 01:03:48.710 --> 01:03:51.030 then my prediction is 6. 01:03:51.800 --> 01:03:53.900 Which is 18 / 3, right? 01:03:53.900 --> 01:03:55.815 So it's the average of these values. 01:03:55.815 --> 01:03:58.270 So if I'm doing Regression, I'm still 01:03:58.270 --> 01:04:00.520 like I'm choosing a split that's going 01:04:00.520 --> 01:04:03.580 to give me the best prediction in each 01:04:03.580 --> 01:04:04.390 side of the split. 01:04:04.980 --> 01:04:06.745 And then my estimate on each side of 01:04:06.745 --> 01:04:08.170 the split is just the average of the 01:04:08.170 --> 01:04:10.100 values after that split. 01:04:11.000 --> 01:04:13.950 And the scoring, the scoring that I can 01:04:13.950 --> 01:04:16.120 use is the squared error. 01:04:16.120 --> 01:04:20.464 So if the squared error would be 1 -, 4 01:04:20.464 --> 01:04:22.536 thirds squared, plus 2 -, 4 thirds 01:04:22.536 --> 01:04:24.905 squared plus 1 -, 4 thirds squared plus 01:04:24.905 --> 01:04:29.049 5 -, 6 ^2 + 8 -, 6 ^2 + 5 -, 6 ^2. 01:04:29.890 --> 01:04:31.665 And so I could try like every 01:04:31.665 --> 01:04:33.549 threshold, compute my squared error 01:04:33.550 --> 01:04:35.635 given every threshold and then choose 01:04:35.635 --> 01:04:37.060 the one that gives me the lowest 01:04:37.060 --> 01:04:37.740 squared error. 01:04:41.040 --> 01:04:43.055 So it's the same algorithm, except that 01:04:43.055 --> 01:04:44.245 you have a different Criterion. 01:04:44.245 --> 01:04:46.370 You might use squared error. 01:04:47.820 --> 01:04:49.530 Because it's continuous values that I'm 01:04:49.530 --> 01:04:50.100 predicting. 01:04:50.840 --> 01:04:53.030 And then the output of the node will be 01:04:53.030 --> 01:04:54.390 the average of the Samples that fall 01:04:54.390 --> 01:04:55.069 into that node. 01:04:56.480 --> 01:04:58.300 And for Regression trees, that's 01:04:58.300 --> 01:05:00.020 especially important to. 01:05:01.330 --> 01:05:03.490 Stop growing your tree early, because 01:05:03.490 --> 01:05:05.420 obviously otherwise you're going to 01:05:05.420 --> 01:05:10.090 always separate your data into one leaf 01:05:10.090 --> 01:05:12.030 node per data point, since you have 01:05:12.030 --> 01:05:13.995 like continuous values, unless there's 01:05:13.995 --> 01:05:15.410 like many of the same value. 01:05:16.020 --> 01:05:17.410 And so you're going to tend to like 01:05:17.410 --> 01:05:17.870 overfit. 01:05:23.330 --> 01:05:25.920 Overfitting, by the way, that's a term 01:05:25.920 --> 01:05:27.060 that comes up a lot in machine 01:05:27.060 --> 01:05:27.865 learning. 01:05:27.865 --> 01:05:30.870 Overfitting means that your model you 01:05:30.870 --> 01:05:32.910 have a very complex model so that you 01:05:32.910 --> 01:05:34.940 achieve like really low Training Error. 01:05:35.660 --> 01:05:37.550 But due to the complexity you're Test 01:05:37.550 --> 01:05:38.740 error has gone up. 01:05:38.740 --> 01:05:41.570 So if you plot your. 01:05:42.250 --> 01:05:44.210 If you plot your Test error as, you 01:05:44.210 --> 01:05:45.510 increase complexity. 01:05:46.200 --> 01:05:48.000 You're Test error will go down for some 01:05:48.000 --> 01:05:50.030 time, but then at some point as your 01:05:50.030 --> 01:05:51.880 complexity keeps rising, you're Test 01:05:51.880 --> 01:05:53.590 Error will start to increase. 01:05:53.590 --> 01:05:55.040 So the point at which you're. 01:05:55.740 --> 01:05:57.650 You're Test Error increases due to 01:05:57.650 --> 01:05:59.260 increasing complexity is where you 01:05:59.260 --> 01:06:00.040 start overfitting. 01:06:00.870 --> 01:06:02.300 We'll talk about that more at the start 01:06:02.300 --> 01:06:03.000 of the ensembles. 01:06:04.840 --> 01:06:06.610 Right, so there's a few variants. 01:06:06.610 --> 01:06:08.620 You can use different splitting 01:06:08.620 --> 01:06:09.490 criteria. 01:06:09.490 --> 01:06:12.010 For example, the genie like impurity or 01:06:12.010 --> 01:06:14.580 Genie Diversity index is just one minus 01:06:14.580 --> 01:06:17.460 the sum over all the values of X 01:06:17.460 --> 01:06:18.700 probability of X ^2. 01:06:19.480 --> 01:06:22.140 This actually is like almost the same 01:06:22.140 --> 01:06:25.800 thing as the Entropy. 01:06:26.410 --> 01:06:27.800 But it's a little bit faster to 01:06:27.800 --> 01:06:29.840 compute, so it's actually more often 01:06:29.840 --> 01:06:30.740 used as the default. 01:06:33.830 --> 01:06:35.890 Most times you split on one attribute 01:06:35.890 --> 01:06:38.460 at a time, but you can also. 01:06:39.190 --> 01:06:40.820 They're in some algorithms. 01:06:40.820 --> 01:06:42.790 You can solve for slices through the 01:06:42.790 --> 01:06:44.600 feature space you can. 01:06:45.280 --> 01:06:47.490 Do like linear discriminant analysis or 01:06:47.490 --> 01:06:49.200 something like that to try to find like 01:06:49.200 --> 01:06:51.970 a multivariable split that separates 01:06:51.970 --> 01:06:53.870 the data, but usually it's just single 01:06:53.870 --> 01:06:54.310 attribute. 01:06:56.180 --> 01:06:57.970 And as I mentioned a couple of times, 01:06:57.970 --> 01:07:00.010 you can stop early so you don't need to 01:07:00.010 --> 01:07:02.010 grow like the full tree until you get 01:07:02.010 --> 01:07:03.025 perfect Training accuracy. 01:07:03.025 --> 01:07:06.110 You can stop after you reach a Max 01:07:06.110 --> 01:07:09.475 depth or stop after you have a certain 01:07:09.475 --> 01:07:11.540 number of nodes per certain number of 01:07:11.540 --> 01:07:12.620 data points per node. 01:07:13.710 --> 01:07:15.470 And the reason that you had stopped 01:07:15.470 --> 01:07:16.920 early is because you the tree to 01:07:16.920 --> 01:07:18.990 generalized new data and if you grow 01:07:18.990 --> 01:07:20.450 like a really big tree, you're going to 01:07:20.450 --> 01:07:23.000 end up with these like little like 01:07:23.000 --> 01:07:26.240 micro applicable rules that might not 01:07:26.240 --> 01:07:27.750 work well when you get new Test 01:07:27.750 --> 01:07:28.270 Samples. 01:07:29.220 --> 01:07:31.190 Where if you have a shorter tree that 01:07:31.190 --> 01:07:34.260 then you might have some uncertainty 01:07:34.260 --> 01:07:36.147 left in your leaf nodes, but you can 01:07:36.147 --> 01:07:38.300 have more confidence that will reflect 01:07:38.300 --> 01:07:39.240 the true distribution. 01:07:42.350 --> 01:07:45.630 So if we look at Decision trees versus 01:07:45.630 --> 01:07:46.280 one and north. 01:07:46.980 --> 01:07:49.500 They're actually kind of similar in a 01:07:49.500 --> 01:07:49.950 way. 01:07:49.950 --> 01:07:51.620 They both have piecewise linear 01:07:51.620 --> 01:07:52.120 decisions. 01:07:52.750 --> 01:07:54.620 So here's the boundary that I get with 01:07:54.620 --> 01:07:56.420 one and N in this example. 01:07:57.110 --> 01:08:00.550 It's going to be based on like if you 01:08:00.550 --> 01:08:03.380 chop things up into cells where each 01:08:03.380 --> 01:08:05.770 sample is like everything within the 01:08:05.770 --> 01:08:07.550 cell is closest to a particular sample. 01:08:08.260 --> 01:08:09.460 I would get this boundary. 01:08:11.100 --> 01:08:12.915 And with the Decision tree you tend to 01:08:12.915 --> 01:08:14.440 get, if you're doing 1 attribute at a 01:08:14.440 --> 01:08:15.980 time, you get this access to line 01:08:15.980 --> 01:08:16.630 boundary. 01:08:16.630 --> 01:08:18.832 So it ends up being like going straight 01:08:18.832 --> 01:08:20.453 over and then up and then straight over 01:08:20.453 --> 01:08:22.160 and then down and then a little bit 01:08:22.160 --> 01:08:23.320 over and then down. 01:08:23.320 --> 01:08:25.226 But they're kind of similar. 01:08:25.226 --> 01:08:28.220 So they're the overlap of those spaces 01:08:28.220 --> 01:08:28.690 is similar. 01:08:31.900 --> 01:08:34.170 The Decision tree also has the ability 01:08:34.170 --> 01:08:36.042 for over stopping to improve 01:08:36.042 --> 01:08:36.520 generalization. 01:08:36.520 --> 01:08:38.530 While they can and doesn't the K&N you 01:08:38.530 --> 01:08:40.700 can increase K to try to improve 01:08:40.700 --> 01:08:42.110 generalization to make it like a 01:08:42.110 --> 01:08:44.506 smoother boundary, but it doesn't have 01:08:44.506 --> 01:08:46.540 like as doesn't have very many like 01:08:46.540 --> 01:08:47.930 controls or knobs to tune. 01:08:50.390 --> 01:08:53.010 And the true power that Decision trees 01:08:53.010 --> 01:08:54.580 arise with ensembles. 01:08:54.580 --> 01:08:56.920 So if you combine lots of these trees 01:08:56.920 --> 01:08:59.250 together to make a prediction, then 01:08:59.250 --> 01:09:01.050 suddenly it becomes very effective. 01:09:01.750 --> 01:09:04.430 In practice, people don't usually use 01:09:04.430 --> 01:09:06.620 this one Decision tree in machine 01:09:06.620 --> 01:09:07.998 learning to make an automated 01:09:07.998 --> 01:09:08.396 prediction. 01:09:08.396 --> 01:09:10.710 They usually use a whole bunch of them 01:09:10.710 --> 01:09:12.397 and then average the results or train 01:09:12.397 --> 01:09:14.870 them in a way that they that they 01:09:14.870 --> 01:09:17.126 incrementally build up your prediction. 01:09:17.126 --> 01:09:18.850 And that's what I'll talk about when I 01:09:18.850 --> 01:09:19.730 talk about ensembles. 01:09:22.360 --> 01:09:23.750 So Decision trees are really a 01:09:23.750 --> 01:09:26.740 component in two of the most successful 01:09:26.740 --> 01:09:28.970 algorithms of all time, but they're not 01:09:28.970 --> 01:09:29.630 the whole thing. 01:09:30.940 --> 01:09:33.160 Here's an example of a Regression tree 01:09:33.160 --> 01:09:34.470 for Temperature prediction. 01:09:35.560 --> 01:09:37.200 Just so that I can make the tree simple 01:09:37.200 --> 01:09:39.370 enough to put on a Slide, I set the Min 01:09:39.370 --> 01:09:41.840 leaf size to 200 so there. 01:09:41.840 --> 01:09:44.000 So I stopped splitting once the node 01:09:44.000 --> 01:09:44.990 has 200 points. 01:09:46.120 --> 01:09:49.080 And then I computed the root mean 01:09:49.080 --> 01:09:50.450 squared error and the R2. 01:09:51.680 --> 01:09:53.280 And so you can see for example like. 01:09:54.430 --> 01:09:55.990 One thing that is interesting to me 01:09:55.990 --> 01:09:58.278 about this is that I would have thought 01:09:58.278 --> 01:09:59.872 that the temperature in Cleveland 01:09:59.872 --> 01:10:01.510 yesterday would be the best predictor 01:10:01.510 --> 01:10:03.150 of the temperature in Cleveland today, 01:10:03.150 --> 01:10:05.056 but it's actually not the best 01:10:05.056 --> 01:10:05.469 predictor. 01:10:05.470 --> 01:10:09.090 So the best single like criteria is the 01:10:09.090 --> 01:10:11.090 temperature in Chicago yesterday, 01:10:11.090 --> 01:10:13.590 because I guess the weather like moves 01:10:13.590 --> 01:10:15.460 from West to east a bit. 01:10:16.850 --> 01:10:19.665 And I guess downward, so knowing the 01:10:19.665 --> 01:10:21.477 weather in Chicago yesterday, whether 01:10:21.477 --> 01:10:23.420 the weather was less than whether the 01:10:23.420 --> 01:10:25.530 Temperature was less than 8.4 Celsius 01:10:25.530 --> 01:10:26.950 or greater than 8.4 Celsius. 01:10:27.590 --> 01:10:29.040 Is the best single thing that I can 01:10:29.040 --> 01:10:29.290 know. 01:10:30.480 --> 01:10:31.160 And then? 01:10:32.290 --> 01:10:34.920 That reduces my initial squared error 01:10:34.920 --> 01:10:36.400 was 112. 01:10:38.170 --> 01:10:39.680 And then if you divide it by number of 01:10:39.680 --> 01:10:41.560 Samples, then or. 01:10:42.810 --> 01:10:44.720 Yeah, take divided by number of samples 01:10:44.720 --> 01:10:45.960 and take square root or something to 01:10:45.960 --> 01:10:47.390 get the per sample. 01:10:48.300 --> 01:10:51.010 Then depending on that answer, then I 01:10:51.010 --> 01:10:53.209 check to see what is the temperature in 01:10:53.210 --> 01:10:55.458 Milwaukee yesterday or what is the 01:10:55.458 --> 01:10:57.140 temperature in Grand Rapids yesterday. 01:10:58.060 --> 01:10:59.600 And then depending on those answers, I 01:10:59.600 --> 01:11:02.040 check Chicago again, a different value 01:11:02.040 --> 01:11:04.170 of Chicago, and then I get my final 01:11:04.170 --> 01:11:04.840 decision here. 01:11:11.120 --> 01:11:13.720 Yeah, it's like my sister lives in 01:11:13.720 --> 01:11:16.140 Harrisburg, so I always know that 01:11:16.140 --> 01:11:17.750 they're going to get our weather like a 01:11:17.750 --> 01:11:18.240 day later. 01:11:19.020 --> 01:11:20.680 So it's like, it's really warm here. 01:11:20.680 --> 01:11:22.190 They're like, it's cold, it's warm 01:11:22.190 --> 01:11:22.450 here. 01:11:22.450 --> 01:11:23.600 Well, I guess it will be warm for you 01:11:23.600 --> 01:11:24.930 tomorrow or in two days. 01:11:26.130 --> 01:11:26.620 Yeah. 01:11:27.540 --> 01:11:29.772 But part of the reason that I share 01:11:29.772 --> 01:11:31.300 this is that the one thing that's 01:11:31.300 --> 01:11:32.890 really cool about Decision trees is 01:11:32.890 --> 01:11:34.910 that you get some explanation, like you 01:11:34.910 --> 01:11:37.450 can understand the data better by 01:11:37.450 --> 01:11:39.600 looking at the tree like this kind of 01:11:39.600 --> 01:11:41.860 violated my initial assumption that the 01:11:41.860 --> 01:11:43.340 best thing to know for the Temperature 01:11:43.340 --> 01:11:44.830 is your Temperature the previous day. 01:11:45.460 --> 01:11:46.600 It's actually the temperature of 01:11:46.600 --> 01:11:48.965 another city the previous day and you 01:11:48.965 --> 01:11:51.100 can get you can create these rules that 01:11:51.100 --> 01:11:53.030 help you understand, like how to make 01:11:53.030 --> 01:11:53.740 predictions. 01:11:56.130 --> 01:11:56.710 This is. 01:11:56.710 --> 01:11:58.320 I'm not expecting you to read this now, 01:11:58.320 --> 01:12:00.370 but this is the code to generate this 01:12:00.370 --> 01:12:00.820 tree. 01:12:06.080 --> 01:12:07.600 Right on Summary. 01:12:08.580 --> 01:12:10.800 The key assumptions of this of the 01:12:10.800 --> 01:12:12.570 Classification or Regression trees are 01:12:12.570 --> 01:12:15.255 that Samples with similar features have 01:12:15.255 --> 01:12:16.070 similar predictions. 01:12:16.070 --> 01:12:17.590 So it's a similar assumption in Nearest 01:12:17.590 --> 01:12:19.580 neighbor, except this time we're trying 01:12:19.580 --> 01:12:21.680 to figure out how to like split up the 01:12:21.680 --> 01:12:23.420 feature space to define that 01:12:23.420 --> 01:12:25.159 similarity, rather than using like a 01:12:25.160 --> 01:12:29.090 preset distance function like Euclidean 01:12:29.090 --> 01:12:29.560 distance. 01:12:30.970 --> 01:12:32.610 The model parameters are the split 01:12:32.610 --> 01:12:34.560 criteria, each internal node, and then 01:12:34.560 --> 01:12:36.630 the final prediction at each leaf node. 01:12:38.200 --> 01:12:40.020 The designs are putting limits on the 01:12:40.020 --> 01:12:42.080 tree growth and what kinds of splits 01:12:42.080 --> 01:12:43.545 you can consider, like whether to split 01:12:43.545 --> 01:12:45.260 on one attribute or whole groups of 01:12:45.260 --> 01:12:48.930 attributes and then choosing their 01:12:48.930 --> 01:12:50.030 criteria for this split. 01:12:51.520 --> 01:12:52.120 01:12:53.300 --> 01:12:56.060 You Decision trees by themselves are 01:12:56.060 --> 01:12:57.645 useful if you want some explainable 01:12:57.645 --> 01:12:58.710 Decision function. 01:12:58.710 --> 01:13:00.270 So they could be used for like medical 01:13:00.270 --> 01:13:02.090 diagnosis for example, because you want 01:13:02.090 --> 01:13:03.750 to be able to tell people like why. 01:13:04.710 --> 01:13:07.324 Like why I know you have cancer, like 01:13:07.324 --> 01:13:08.840 you don't want to just be like I use 01:13:08.840 --> 01:13:10.270 this machine learning algorithm and it 01:13:10.270 --> 01:13:12.070 says you have like a 93% chance of 01:13:12.070 --> 01:13:13.750 having cancer and so sorry. 01:13:15.000 --> 01:13:16.785 You want to be able to say like because 01:13:16.785 --> 01:13:19.086 of like this thing and because of this 01:13:19.086 --> 01:13:21.099 thing and because of this thing like 01:13:21.100 --> 01:13:24.919 out of all these 1500 cases like 90% of 01:13:24.920 --> 01:13:26.750 them ended up having cancer. 01:13:26.750 --> 01:13:28.180 So we need to do, we need to do a 01:13:28.180 --> 01:13:29.110 biopsy, right. 01:13:29.110 --> 01:13:30.305 So you want some explanation. 01:13:30.305 --> 01:13:31.900 A lot of times it's not always good 01:13:31.900 --> 01:13:33.600 enough to have like a good prediction. 01:13:35.590 --> 01:13:37.240 And they're also like really effective 01:13:37.240 --> 01:13:38.520 as part of a ensemble. 01:13:38.520 --> 01:13:39.650 And again, I think we might see a 01:13:39.650 --> 01:13:40.650 Tuesday instead of Thursday. 01:13:43.150 --> 01:13:44.960 It's not like a really good predictor 01:13:44.960 --> 01:13:47.320 by itself, but it is really good as 01:13:47.320 --> 01:13:47.770 part of an. 01:13:48.500 --> 01:13:48.790 Alright. 01:13:49.670 --> 01:13:51.250 So things you remember, Decision 01:13:51.250 --> 01:13:52.690 Regression trees learn to split up the 01:13:52.690 --> 01:13:54.590 feature space into partitions into 01:13:54.590 --> 01:13:56.110 different cells with similar values. 01:13:57.150 --> 01:13:59.070 And then Entropy is a really important 01:13:59.070 --> 01:13:59.600 concept. 01:13:59.600 --> 01:14:01.030 It's a measure of uncertainty. 01:14:02.730 --> 01:14:05.170 Information gain measures how much 01:14:05.170 --> 01:14:07.090 particular knowledge reduces the 01:14:07.090 --> 01:14:08.710 prediction uncertainty, and that's the 01:14:08.710 --> 01:14:10.260 basis for forming our tree. 01:14:11.630 --> 01:14:13.650 So on Thursday I'm going to do a bit of 01:14:13.650 --> 01:14:15.680 review of our concepts and then I think 01:14:15.680 --> 01:14:17.200 most likely next Tuesday I'll talk 01:14:17.200 --> 01:14:19.730 about ensembles and random forests and 01:14:19.730 --> 01:14:21.540 give you an extensive example of how 01:14:21.540 --> 01:14:23.560 it's used in the Kinect algorithm. 01:14:24.800 --> 01:14:25.770 Alright, thanks everyone. 01:14:25.770 --> 01:14:26.660 See you Thursday. 01:19:07.020 --> 01:19:08.790 Hello. 01:19:10.510 --> 01:19:11.430 Training an assault.