Docsity
Docsity

Prepare for your exams
Prepare for your exams

Study with the several resources on Docsity


Earn points to download
Earn points to download

Earn points by helping other students or get them with a premium plan


Guidelines and tips
Guidelines and tips

Cross-Validation & Feature Selection in ML: CS229 Lecture Notes by Andrew Ng, Study notes of Machine Learning

A part of the lecture notes from stanford university's cs229 machine learning course, covering cross-validation and feature selection methods for model selection. The notes discuss the importance of selecting the best model for a learning problem, especially when dealing with large numbers of features. Cross-validation techniques, such as k-fold cross-validation, and their application to model selection. Additionally, the notes cover feature selection algorithms, including wrapper methods and filter methods, which help reduce the number of features and improve learning performance.

Typology: Study notes

2010/2011

Uploaded on 10/30/2011

ilyastrab
ilyastrab 🇺🇸

4.4

(50)

131 documents

1 / 8

Toggle sidebar

Related documents


Partial preview of the text

Download Cross-Validation & Feature Selection in ML: CS229 Lecture Notes by Andrew Ng and more Study notes Machine Learning in PDF only on Docsity! CS229 Lecture notes Andrew Ng Part VI Regularization and model selection Suppose we are trying select among several different models for a learning problem. For instance, we might be using a polynomial regression model hθ(x) = g(θ0 + θ1x + θ2x 2 + · · · + θkx k), and wish to decide if k should be 0, 1, . . . , or 10. How can we automatically select a model that represents a good tradeoff between the twin evils of bias and variance1? Alternatively, suppose we want to automatically choose the bandwidth parameter τ for locally weighted regression, or the parameter C for our `1-regularized SVM. How can we do that? For the sake of concreteness, in these notes we assume we have some finite set of models M = {M1, . . . ,Md} that we’re trying to select among. For instance, in our first example above, the model Mi would be an i-th order polynomial regression model. (The generalization to infinite M is not hard.2) Alternatively, if we are trying to decide between using an SVM, a neural network or logistic regression, then M may contain these models. 1Given that we said in the previous set of notes that bias and variance are two very different beasts, some readers may be wondering if we should be calling them “twin” evils here. Perhaps it’d be better to think of them as non-identical twins. The phrase “the fraternal twin evils of bias and variance” doesn’t have the same ring to it, though. 2If we are trying to choose from an infinite set of models, say corresponding to the possible values of the bandwidth τ ∈ R+, we may discretize τ and consider only a finite number of possible values for it. More generally, most of the algorithms described here can all be viewed as performing optimization search in the space of models, and we can perform this search over infinite model classes as well. 1 2 1 Cross validation Lets suppose we are, as usual, given a training set S. Given what we know about empirical risk minimization, here’s what might initially seem like a algorithm, resulting from using empirical risk minimization for model selec- tion: 1. Train each model Mi on S, to get some hypothesis hi. 2. Pick the hypotheses with the smallest training error. This algorithm does not work. Consider choosing the order of a poly- nomial. The higher the order of the polynomial, the better it will fit the training set S, and thus the lower the training error. Hence, this method will always select a high-variance, high-degree polynomial model, which we saw previously is often poor choice. Here’s an algorithm that works better. In hold-out cross validation (also called simple cross validation), we do the following: 1. Randomly split S into Strain (say, 70% of the data) and Scv (the remain- ing 30%). Here, Scv is called the hold-out cross validation set. 2. Train each model Mi on Strain only, to get some hypothesis hi. 3. Select and output the hypothesis hi that had the smallest error ε̂Scv(hi) on the hold out cross validation set. (Recall, ε̂Scv(h) denotes the empir- ical error of h on the set of examples in Scv.) By testing on a set of examples Scv that the models were not trained on, we obtain a better estimate of each hypothesis hi’s true generalization error, and can then pick the one with the smallest estimated generalization error. Usually, somewhere between 1/4 − 1/3 of the data is used in the hold out cross validation set, and 30% is a typical choice. Optionally, step 3 in the algorithm may also be replaced with selecting the model Mi according to arg mini ε̂Scv(hi), and then retraining Mi on the entire training set S. (This is often a good idea, with one exception being learning algorithms that are be very sensitive to perturbations of the initial conditions and/or data. For these methods, Mi doing well on Strain does not necessarily mean it will also do well on Scv, and it might be better to forgo this retraining step.) The disadvantage of using hold out cross validation is that it “wastes” about 30% of the data. Even if we were to take the optional step of retraining 5 The outer loop of the algorithm can be terminated either when F = {1, . . . , n} is the set of all features, or when |F| exceeds some pre-set thresh- old (corresponding to the maximum number of features that you want the algorithm to consider using). This algorithm described above one instantiation of wrapper model feature selection, since it is a procedure that “wraps” around your learning algorithm, and repeatedly makes calls to the learning algorithm to evaluate how well it does using different feature subsets. Aside from forward search, other search procedures can also be used. For example, backward search starts off with F = {1, . . . , n} as the set of all features, and repeatedly deletes features one at a time (evaluating single-feature deletions in a similar manner to how forward search evaluates single-feature additions) until F = ∅. Wrapper feature selection algorithms often work quite well, but can be computationally expensive given how that they need to make many calls to the learning algorithm. Indeed, complete forward search (terminating when F = {1, . . . , n}) would take about O(n2) calls to the learning algorithm. Filter feature selection methods give heuristic, but computationally much cheaper, ways of choosing a feature subset. The idea here is to compute some simple score S(i) that measures how informative each feature xi is about the class labels y. Then, we simply pick the k features with the largest scores S(i). One possible choice of the score would be define S(i) to be (the absolute value of) the correlation between xi and y, as measured on the training data. This would result in our choosing the features that are the most strongly correlated with the class labels. In practice, it is more common (particularly for discrete-valued features xi) to choose S(i) to be the mutual information MI(xi, y) between xi and y: MI(xi, y) = ∑ xi∈{0,1} ∑ y∈{0,1} p(xi, y) log p(xi, y) p(xi)p(y) . (The equation above assumes that xi and y are binary-valued; more generally the summations would be over the domains of the variables.) The probabil- ities above p(xi, y), p(xi) and p(y) can all be estimated according to their empirical distributions on the training set. To gain intuition about what this score does, note that the mutual infor- mation can also be expressed as a Kullback-Leibler (KL) divergence: MI(xi, y) = KL (p(xi, y)||p(xi)p(y)) You’ll get to play more with KL-divergence in Problem set #3, but infor- mally, this gives a measure of how different the probability distributions 6 p(xi, y) and p(xi)p(y) are. If xi and y are independent random variables, then we would have p(xi, y) = p(xi)p(y), and the KL-divergence between the two distributions will be zero. This is consistent with the idea if xi and y are independent, then xi is clearly very “non-informative” about y, and thus the score S(i) should be small. Conversely, if xi is very “informative” about y, then their mutual information MI(xi, y) would be large. One final detail: Now that you’ve ranked the features according to their scores S(i), how do you decide how many features k to choose? Well, one standard way to do so is to use cross validation to select among the possible values of k. For example, when applying naive Bayes to text classification— a problem where n, the vocabulary size, is usually very large—using this method to select a feature subset often results in increased classifier accuracy. 3 Bayesian statistics and regularization In this section, we will talk about one more tool in our arsenal for our battle against overfitting. At the beginning of the quarter, we talked about parameter fitting using maximum likelihood (ML), and chose our parameters according to θML = arg max θ m ∏ i=1 p(y(i)|x(i); θ). Throughout our subsequent discussions, we viewed θ as an unknown param- eter of the world. This view of the θ as being constant-valued but unknown is taken in frequentist statistics. In the frequentist this view of the world, θ is not random—it just happens to be unknown—and it’s our job to come up with statistical procedures (such as maximum likelihood) to try to estimate this parameter. An alternative way to approach our parameter estimation problems is to take the Bayesian view of the world, and think of θ as being a random variable whose value is unknown. In this approach, we would specify a prior distribution p(θ) on θ that expresses our “prior beliefs” about the parameters. Given a training set S = {(x(i), y(i))}mi=1, when we are asked to make a prediction on a new value of x, we can then compute the posterior 7 distribution on the parameters p(θ|S) = p(S|θ)p(θ) p(S) = ( ∏m i=1 p(y (i)|x(i), θ) ) p(θ) ∫ θ ( ∏m i=1 p(y (i)|x(i), θ)p(θ)) dθ (1) In the equation above, p(y(i)|x(i), θ) comes from whatever model you’re using for your learning problem. For example, if you are using Bayesian logistic re- gression, then you might choose p(y(i)|x(i), θ) = hθ(x (i))y (i) (1−hθ(x (i)))(1−y (i)), where hθ(x (i)) = 1/(1 + exp(−θT x(i))).3 When we are given a new test example x and asked to make it prediction on it, we can compute our posterior distribution on the class label using the posterior distribution on θ: p(y|x, S) = ∫ θ p(y|x, θ)p(θ|S)dθ (2) In the equation above, p(θ|S) comes from Equation (1). Thus, for example, if the goal is to the predict the expected value of y given x, then we would output4 E[y|x, S] = ∫ y yp(y|x, S)dy The procedure that we’ve outlined here can be thought of as doing “fully Bayesian” prediction, where our prediction is computed by taking an average with respect to the posterior p(θ|S) over θ. Unfortunately, in general it is computationally very difficult to compute this posterior distribution. This is because it requires taking integrals over the (usually high-dimensional) θ as in Equation (1), and this typically cannot be done in closed-form. Thus, in practice we will instead approximate the posterior distribution for θ. One common approximation is to replace our posterior distribution for θ (as in Equation 2) with a single point estimate. The MAP (maximum a posteriori) estimate for θ is given by θMAP = arg max θ m ∏ i=1 p(y(i)|x(i), θ)p(θ). (3) 3Since we are now viewing θ as a random variable, it is okay to condition on it value, and write “p(y|x, θ)” instead of “p(y|x; θ).” 4The integral below would be replaced by a summation if y is discrete-valued.
Docsity logo



Copyright © 2024 Ladybird Srl - Via Leonardo da Vinci 16, 10126, Torino, Italy - VAT 10816460017 - All rights reserved