# Intermediate Regression in R

Learn to perform linear and logistic regression with multiple explanatory variables.

4 Hours14 Videos50 Exercises12,354 Learners4150 XPData Scientist TrackMachine Learning Scientist TrackStatistician TrackStatistics Fundamentals TrackSupervised Machine Learning Track

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA.

## Course Description

Linear regression and logistic regression are the two most widely used statistical models and act like master keys, unlocking the secrets hidden in datasets. This course builds on the skills you gained in "Introduction to Regression in R", covering linear and logistic regression with multiple explanatory variables. Through hands-on exercises, you’ll explore the relationships between variables in real-world datasets, Taiwan house prices and customer churn modeling, and more. By the end of this course, you’ll know how to include multiple explanatory variables in a model, understand how interactions between variables affect predictions, and understand how linear and logistic regression work.
1. 1

### Parallel Slopes

Free

Extend your linear regression skills to "parallel slopes" regression, with one numeric and one categorical explanatory variable. This is the first step towards conquering multiple linear regression.

Parallel slopes linear regression
50 xp
Fitting a parallel slopes linear regression
100 xp
Interpreting parallel slopes coefficients
100 xp
Visualizing each explanatory variable
100 xp
Visualizing parallel slopes
100 xp
Predicting parallel slopes
50 xp
Predicting with a parallel slopes model
100 xp
Manually calculating predictions
100 xp
Assessing model performance
50 xp
Comparing coefficients of determination
100 xp
Comparing residual standard error
100 xp
2. 2

### Interactions

Explore the effect of interactions between explanatory variables. Considering interactions allows for more realistic models that can have better predictive power. You'll also deal with Simpson's Paradox: a non-intuitive result that arises when you have multiple explanatory variables.

3. 3

### Multiple Linear Regression

See how modeling, and linear regression in particular, makes it easy to work with more than two explanatory variables. Once you've mastered fitting linear regression models, you'll get to implement your own linear regression algorithm.

4. 4

### Multiple Logistic Regression

Extend your logistic regression skills to multiple explanatory variables. Understand the logistic distribution, which underpins this form of regression. Finally, implement your own logistic regression algorithm.

In the following tracks

Data ScientistMachine Learning ScientistStatisticianStatistics FundamentalsSupervised Machine Learning

Collaborators #### Richie Cotton

Data Evangelist at DataCamp

Richie is a Data Evangelist at DataCamp. He has been using R since 2004, in the fields of proteomics, debt collection, and chemical health and safety. He has released almost 30 R packages on CRAN and Bioconductor – most famously the assertive suite of packages – as well as creating and contributing to many others. He also has written two books on R programming, Learning R and Testing R Code.