Home » How to Create a Confusion Matrix in R (Step-by-Step)

How to Create a Confusion Matrix in R (Step-by-Step)

by Tutor Aspire

Logistic regression is a type of regression we can use when the response variable is binary.

One common way to evaluate the quality of a logistic regression model is to create a confusion matrix, which is a 2×2 table that shows the predicted values from the model vs. the actual values from the test dataset.

The following step-by-step example shows how to create a confusion matrix in R.

Step 1: Fit the Logistic Regression Model

For this example we’ll use the Default dataset from the ISLR package. We’ll use student status, bank balance, and annual income to predict the probability that a given individual defaults on their loan.

The following code shows how to fit a logistic regression model to this dataset:

#load necessary packages

#load dataset
data #split dataset into training and testing set
sample TRUE, prob=c(0.7,0.3))
train #fit logistic regression model
model binomial", data=train)

Step 2: Create the Confusion Matrix

Next, we’ll use the confusionMatrix() function from the caret package to create a confusion matrix:

#use model to predict probability of default
#convert defaults from "Yes" and "No" to 1's and 0's
test$default Yes", 1, 0)

#find optimal cutoff probability to use to maximize accuracy
#create confusion matrix
confusionMatrix(test$default, predicted)

     0  1
0 2912 64
1   21 39

Step 3: Evaluate the Confusion Matrix

We can also calculate the following metrics using the confusion matrix:

  • Sensitivity: The “true positive rate” – the percentage of individuals the model correctly predicted would default.
  • Specificity: The “true negative rate” – the percentage of individuals the model correctly predicted would not default.
  • Total misclassification rate: The percentage of total incorrect classifications made by the model.

The following code shows how to calculate these metrics:

#calculate sensitivity
sensitivity(test$default, predicted)

[1] 0.3786408

#calculate specificity
specificity(test$default, predicted)

[1] 0.9928401

#calculate total misclassification error rate
misClassError(test$default, predicted, threshold=optimal)

[1] 0.027

The total misclassification error rate is 2.7% for this model.

In general, the lower this rate the better the model is able to predict outcomes, so this particular model turns out to be very good at predicting whether an individual will default or not. 

You may also like