hw-6

Homework 6 in DSCI445: Statistical Machine Learning @ CSU

Assignment

Be sure to set.seed(445).

  1. We will explore the maximal margin classifier on a toy data set.

    1. We are given \(n = 7\) observations in \(p = 2\) dimensions. For each observation, there is an associated class label.
    Obs X_1 X_2 Y
    1 3 4 Red
    2 2 2 Red
    3 4 4 Red
    4 1 4 Red
    5 2 1 Blue
    6 4 3 Blue
    7 4 1 Blue

    Sketch the observations.

    1. Sketch the optimal separating hyperplane and provide the equation for this hyperplane.

    2. Describe the classification rule for the maximal marginal classifier. It should be along the lines of "Classify to Red if \(\beta_0 + \beta_1 X_1 + \beta_2 X_2 > 0\), and classify as Blue otherwise. Provide the values of \(\beta_0, \beta_1, \beta_2\).

    3. On your sketch, indicate the margin for the maximal margin classifier.

    4. Indicate the support vectors for the maximal margin classifier.

    5. Argue that a slight movement of the seventh observation would not affect the maximal margin hyperplane.

    6. Draw an additional observation on the plot so that the two classes are no longer separable by a hyperplane.

  2. We have seen that we can fit an SVM with a non-linear kernel in order to perform classification using a non-linear decision boundary. We will now see that we can also obtain a non-linear decision boundary by performing logistic regression using non-linear transformations of the features.

    1. Generate a data set with \(n = 500\) and \(p = 2\), such that the observations belong to two classes with a quadratic decision boundary. E.g.,
    n <- 500
    x1 <- runif(n) - 0.5
    x2 <- runif(n) - 0.5
    y <- as.numeric(x1^2 - x2^2 > 0)
    1. Plot the observations, colored according to their class labels.

    2. Fit a logistic regression model to the data using \(X_1\) and \(X_2\) as predictors.

    3. Apply this model to the training data in order to obtain a predicted class label for each training observation. Plot the observations, colored according to the predicted class labels. What shape is the decision boundary?

    4. Now fit a logistic regression model to the data using non-linear functions of \(X_1\) and \(X_2\) as predictors (e.g., \(X_1^2, X_1 \times X_2, \log(X_2)\), etc.)

    5. Apply this model to the training data in order to obtain a predicted class label for each training observation. Plot the observations, colored according to the predicted class labels. What shape is the decision boundary? Repear a)- e) until you come up with an example in which the predicted class labels are obviously non-linear.

    6. Fit a support vector classifier with \(X_1\) and \(X_2\) as predictors. Obtain a class predictor for each training observation. Plot the observations, colored according to the predicted class labels.

    7. Fit an SVM using a non-linear kernel to the data with \(X_1\) and \(X_2\) as predictors. Obtain a class predictor for each training observation. Plot the observations, colored according to the predicted class labels.

    8. Comment on your results.

  3. In this problem, you will use support vector approaches to predict whether a given car gets high or low gas mileage based on the Auto data set in the ISLR package.

    1. Create a binary variable that takes value 1 for gas mileage above the median and 0 for cars below the median.

    2. Fit a support vector classifier to the data with various values of cost, in order to predict whether a car gets high or low gas mileage (be sure not to include the original gass mileage variable – no cheating!). Report the cross-validation errors associated with different values of this parameter, comment on your results.

    3. Now repeat (b) using SVMs with radial and polynomial basis kernels, with different values of gamma, degree, and cost. Report on your results.

    4. Make some plots to back up your assertions in b) and c).

  4. This problem involves the OJ data set in the ISLR package.

    1. Create a training set containing a random sample of 900 observations and a test set containg the remaining observations.

    2. Fit a support vecotr classifier to the training set using cost = 0.01 with Purchase as the response and the other variables as predictors. Use the summary() function to produce summary statistics and describe the results obtained.

    3. What are the training and test error rates?

    4. Use the tune() function to select an optimal cost. Consider values between \(0.01\) and \(10\).

    5. Compute the training and test error rates using this new value for cost.

    6. Repeat b) through e) using a support vector machine with a radial kernal and default value for gamma.

    7. Repeat b) through e) using a support vector machine with a polynomial kernal and degree = 2.

    8. Which approach gives the best results on this data?