# Intermediate Statistical Modeling in R

In this follow-up course, you will expand your stat modeling skills from the introduction and dive into more advanced concepts.

Start Course for Free4 Hours10 Videos41 Exercises7,808 Learners3300 XP

## Create Your Free Account

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA. You confirm you are at least 16 years old (13 if you are an authorized Classrooms user).## Loved by learners at thousands of companies

## Course Description

*Statistical Modeling in R* is a multi-part course designed to get you up to speed with the most important and powerful methodologies in statistics. In *this intermediate course 2*, we'll take a look at effect size and interaction, the concepts of total and partial change, sampling variability and mathematical transforms, and the implications of something called collinearity. This course has been written from scratch, specifically for DataCamp users. As you'll see, by using computing and concepts from machine learning, we'll be able to leapfrog many of the marginal and esoteric topics encountered in traditional 'regression' courses.

- 1
### Effect size and interaction

**Free**Effect sizes were introduced in

*Part 1*of this course series as a way to quantify how each explanatory variable is connected to the response. In this chapter, you'll meet some high-level tools that make it easier to calculate and visualize effect sizes. You'll see how to extend the notion of effect size to models with a categorical response variable. And you'll start to use*interactions*in constructing models to reflect the way that one explanatory variable can influence the effect size of another explanatory variable on the response.Multiple explanatory variables50 xpGraphing a model of house prices100 xpBody-mass index (BMI)100 xpCategorical response variables50 xpEager runners50 xpWho are the mellow runners?100 xpSmoking and survival100 xpInteractions among explanatory variables50 xpWith and without an interaction term100 xpWorking together50 xpMileage and age interacting100 xpInteractions and effect size50 xpOptimal temperature50 xp - 2
### Total and partial change

In many circumstances, an effect size tells you exactly what you need to know: how much the model output will change when one, and only one, explanatory variable changes. This is called

*partial change*. In other situations, you will want to look at*total change*, which combines the effects of two or more explanatory variables. You'll also see an additional, but limited way of quantifying the extent to which the explanatory variables influence the response: R-squared. Finally, we'll describe the notion of*degrees of freedom*, a way of describing the complexity of a model.Total and partial change50 xpAnother bedroom?100 xpCalculating total change100 xpCar prices100 xpR-squared50 xpCalculating R-squared100 xpWarming in Minneapolis?100 xpR-squared goes up100 xpDegrees of freedom50 xpRules for counting50 xpIs bigger R-squared better? (1)100 xpIs bigger R-squared better? (2)100 xpAccidental "perfection"100 xp - 3
### Sampling variability and mathematical transforms

This chapter examines the precision with which a model can estimate an effect size. The lack of precision comes from

*sampling variability*, which can be quantified using resampling and bootstrapping. You'll also see some ways to improve precision using mathematical transformations of variables.Bootstrapping and precision50 xpA bootstrap trial100 xpFrom a bootstrap ensemble to the standard error100 xpExample: fireplaces100 xpScales and transformations50 xpTypical values of data100 xpExponential growth100 xpPrediction with log transforms100 xpConfidence intervals on log-transformed models100 xp - 4
### Variables working together

In this final chapter, you'll learn about why you'd want to avoid collinearity, a common phenomenon in statistical modeling. You'll wrap up the course by discussing some of the ways models can be improved by involving the modeler in the design of the data collecting process.

Datasets

Used FordsPrerequisites

Introduction to Statistical Modeling in R#### Daniel Kaplan

DeWitt Wallace Professor at Macalester College

Danny is the DeWitt Wallace Professor of Mathematics, Statistics, and Computer Science at Macalester College in Saint Paul, Minnesota. At Macalester, he has developed the introductory sequence in calculus and statistics as well as an introduction to computing for scientists. He’s co-authored the mosaic R package and written several textbooks: Understanding Nonlinear Dynamics,

*Introduction to Scientific Computation and Programming*, and*Statistical Modeling: A Fresh Approach*.## What do other learners have to say?

I've used other sites—Coursera, Udacity, things like that—but DataCamp's been the one that I've stuck with.

Devon Edwards Joseph

Lloyds Banking Group

DataCamp is the top resource I recommend for learning data science.

Louis Maiden

Harvard Business School

DataCamp is by far my favorite website to learn from.

Ronald Bowers

Decision Science Analytics, USAA

## Join over 9 million learners and start Intermediate Statistical Modeling in R today!

### Create Your Free Account

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA. You confirm you are at least 16 years old (13 if you are an authorized Classrooms user).