# How to apply and interpret linear regression in R

By Data Tricks, 28 May 2020

For this tutorial we’re going to use some House Price data from Kaggle. You’ll need to download both the train.csv and test.csv files from here.

rm(list=ls())
test <- read.csv("C:/Users/Me/Desktop/test.csv")

Remember to change the file path to where you’ve saved the files. You can type str(train) to see a list of the variables and their types, and you should see that there are a total of 81 variables.

## Step 2 – select dependent and independent variables

A linear regression models the relationship between a dependent variable (that is the variable you are trying to predict) and one or more independent variables (the variables you are using to make the prediction). It therefore follows that one of the first steps in the process of linear regression should be to check the correlation between the dependent variable and potential independent variables. The process of variable selection can be an iterative one – once a linear regression has been carried out we can go back and tweak our variable selection.

Check the correlation between all numeric variables in the dataset.

cor(train[,unlist(lapply(train, is.numeric))])

Tip! The command unlist(lapply(train. is.numeric)) returns a list of TRUE and FALSE values for each column in the dataframe according to whether that column is numeric. If we wanted to do a more thorough linear regression, we would first look into transforming all non-numeric variables into numeric variables in order to include them in the linear regression, but for the purposes of this tutorial we will ignore these variables for now.

You should now have a matrix of correlation values for each column of the dataframe. At this stage we are only interested in the correlation of each variable with the dependent variable, which in this case is the very last column SalePrice.

The five highest correlation values (remember that you should be looking for the highest absolute values, so include negative values as well) are for the following variables:

OverallQual: the overall material and finish of the house (correlation = 0.79)

GrLivArea: the above-ground living area in feet (correlation = 0.71)

GarageCars: the size of the garage in car capacity (correlation = 0.64)

GarageArea: the size of the garage in square feet (correlation = 0.62)

TotalBsmtSF: the below-ground living area in fee (correlation = 0.61)

One of the requirements of linear regression is that independent variable must not be highly correlated with one another (the fancy word is multicollinearity). From the previous correlation matrix we ran, we can see that GarageCars and GarageArea are strongly correlated, with a value of 0.88, so one of these two variables can probably be removed. In addition, GrLivArea and TotalBsmtSF are also moderately correlated at 0.45 and it might be reasonable to remove one of these variables as well.

For the purposes of this tutorial, we are going to remove GarageArea and keep GarageCars, given the latter had a slightly stronger correlation with SalePrice. But we’ll keep both GrLivArea and TotalBsmtSF – we can always remove these later once we’ve applied linear regression.

## Step 3 – apply a linear regression

Helpfully, the method of applying machine learning regression algorithms in R usually follows a similar syntax, which is simply:

function(formula, data)

where function is the function for the algorithm you’ve chosen (in the case of linear regression, this is lm), formula is where you specify your dependent and independent variables separated by a ~ (ie. dependent.variable ~ independent.variable.1 + independent.variable.2….etc,), and data is where you specify the dataframe on which to train the algorithm.

To run a linear regression, type the following:

library(dplyr)
train.sub <- train %>%
select(SalePrice, OverallQual, GarageCars, GrLivArea, TotalBsmtSF)
myModel <- lm(SalePrice ~., data = train.sub)

Tip! If you put a full-stop after the ~ in the formula, this tells R to use all the variables (except the dependent variable already specified) in your dataframe as independent variables. Notice that we’ve used dplyr first to select only the columns that contain our dependent and independent variables.

## Step 4 – interpret the linear regression

Type summary(myModel) into the console and you should get the following:

> summary(myModel)
Call:
lm(formula = SalePrice ~ ., data = train.sub)

Residuals:
Min      1Q  Median     3Q     Max
-469856  -19956   -1360  17200  286304

Coefficients:
Estimate Std. Error t value Pr(>|t|)
(Intercept) -99248.853   4639.866  -21.39   <2e-16 ***
OverallQual  23572.236   1072.465   21.98   <2e-16 ***
GarageCars   18582.209   1747.412   10.63   <2e-16 ***
GrLivArea       45.643      2.484   18.38   <2e-16 ***
TotalBsmtSF     32.520      2.838   11.46   <2e-16 ***
---
Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ‘ ’ 1

Residual standard error: 38920 on 1455 degrees of freedom
Multiple R-squared: 0.7607, Adjusted R-squared: 0.76
F-statistic: 1156 on 4 and 1455 DF, p-value: < 2.2e-16

Some of the key statistics that are helpful in interpreting a linear regression are as follows:

The R-squared value, or R2, is a measure of goodness-of-fit. It represents the percentage of the variance of the dependent variable (in this case the SalePrice) that is explained collectively by the independent variables. It is therefore a common measure of accuracy of a regression model.

It is common to use the Adjusted R2 value because this takes into account the number of independent variables used in the regression and adjusts teh R2 such that it increases only if any new independent variables improve the model more than would be expected by chance. Using the Adjusted R2 value therefore reduces the risk of getting incorrectly inflated values purely based on having a large number of independent variables.

Our model has an Adjusted R2 value of 0.76, meaning that 76% of the variation in SalePrice can be explained by the collective effect of OverallQual, GarageCars, GrLivArea and TotalBsmtSF. Not bad for a first effort.

#### Residuals

Another useful set of statistics is the residuals, as these provide an indication of how far out your predictive model is from the real-world values. A residual is the difference between the actual values of your dependent variables and the predicted values. In our case, we have a median residual of -1,360, which means that on ‘average’ our model undershot the SalePrice. That doesn’t seem too bad, however there are clearly some outliers as the largest negative residual is -470k and the largest positive is 286k, indicating that our model under- and overshot some values quite dramatically.

#### Intercepts and coefficients

A linear regression model can be used to predict the dependent variable from the independent variables, and the intercepts and coefficients can help us determine how this is done. The regression can be expressed as a formula as follows:

y = β0 + β1x1 + β2x2

where y is the dependent variable, β0 is the intercept, x1, x2 etc. are the independent variables along with their coefficients β1, β2 etc.

Thus, our linear regression can be expressed as:

SalePrice = -99248.853 + 23572.236(OverallQual) + 18582.209(GarageCars) + 45.643(GrLivArea) + 32.520(TotalBsmtSF)

Using this formula, we can predict the value of SalePrice given any new values of independent variables.

However, before using this regression formula, it is also important to look at the other statistics available alongside the coefficients. You will see that we are also given the Standard Error, a T value and a “Pr” value.

The Standard Error of a coefficient is an estimate of the standard deviation of the error in measuring it.

The T value is the coefficient divided by its standard error. Thus in general, the larger the t value the better, because a large value indicates that the size of the error is small in comparison to the coefficient itself.

The “Pr” value, or simply the P value tests the null hypothesis that there is no relationship between the independent variable and dependent variable. It is common to take any value below 0.05 as being grounds to reject the null hypothesis – that is to say, the coefficient is statistically significant.

## Step 5 – test and evaluate

You may remember that we had a train.csv and test.csv dataset. We can now use our linear regression model to predict the SalePrice values of the test.csv dataset and submit it to Kaggle for scoring:

test.sub <- test %>%select(Id, OverallQual, GarageCars, GrLivArea, TotalBsmtSF)test.sub$GarageCars[is.na(test.sub$GarageCars)] <- mean(test.sub$GarageCars, na.rm = TRUE)test.sub$TotalBsmtSF[is.na(test.sub$TotalBsmtSF)] <- mean(test.sub$TotalBsmtSF, na.rm = TRUE)predictions <- predict(myModel, test.sub)test.results <- cbind(test.sub, predictions)test.results <- test.results[,c(1,6)]colnames(test.results)[2] <- "SalePrice"write.csv(test.results, "submission.csv", row.names = FALSE)

You will notice that we have had to impute some missing values for GarageCars and TotalBsmtSF as there was one missing value in each of these variables in test.csv. For speed we have simply replaced these missing values with the mean for the entire dataset, however there are better approaches to imputing.

Exporting our predictions to CSV and uploading them to Kaggle yields a score of 0.62 and a lowly 5000th place on the leaderboard. However, as a very quick 2-minute effort there’s lots of room for improvement.

To improve the score, here are some ideas of what to look at:

1. Investigate the correlation of other variables with a view to adding more independent variables.

2. Converting categorical or character variables into numeric variables using one-hot encoding or other methods.

3. Standardisation or normalisation of variables.

4. Feature engineering.