*53*

Logistic Regression is a statistical method that we use to fit a regression model when the response variable is binary. To assess how well a logistic regression model fits a dataset, we can look at the following two metrics:

**Sensitivity:**The probability that the model predicts a positive outcome for an observation when indeed the outcome is positive. This is also called the “true positive rate.”**Specificity:**The probability that the model predicts a negative outcome for an observation when indeed the outcome is negative. This is also called the “true negative rate.”

One way to visualize these two metrics is by creating a **ROC curve**, which stands for “receiver operating characteristic” curve.

This is a plot that displays the sensitivity along the y-axis and (1 – specificity) along the x-axis. One way to quantify how well the logistic regression model does at classifying data is to calculate **AUC**, which stands for “area under curve.”

The closer the AUC is to 1, the better the model.

The following step-by-step example shows how to calculate AUC for a logistic regression model in R.

**Step 1: Load the Data**

First, we’ll load the **Default** dataset from the **ISLR** package, which contains information about whether or not various individuals defaulted on a loan.

#load dataset data #view first six rows of dataset head(data) default student balance income 1 No No 729.5265 44361.625 2 No Yes 817.1804 12106.135 3 No No 1073.5492 31767.139 4 No No 529.2506 35704.494 5 No No 785.6559 38463.496 6 No Yes 919.5885 7491.559

**Step 2: Fit the Logistic Regression Model**

Next, we’ll fit a logistic regression model to predict the probability that an individual defaults:

#make this example reproducible set.seed(1) #Use 70% of dataset as training set and remaining 30% as testing set sample TRUE, prob=c(0.7,0.3)) train #fit logistic regression model model binomial", data=train)

**Step 3: Calculate the AUC of the Model**

Next, we’ll use the **auc()** function from the **pROC** package to calculate the AUC of the model. This function uses the following syntax:

**auc(response, predicted)**

Here’s how to use this function in our example:

#calculate probability of default for each individual in test dataset predicted response") #calculate AUC library(pROC) auc(test$default, predicted) Setting levels: control = No, case = Yes Setting direction: controls

The AUC of the model turns out to be **0.9437**.

Since this value is close to 1, this indicates that the model does a very good job of predicting whether or not an individual will default on their loan.