Skip to main content
HomeR

Course

Dimensionality Reduction in R

BasicSkill Level
4.8+
86 reviews
Updated 12/2024
Learn dimensionality reduction techniques in R and master feature selection and extraction for your own data and models.
Start Course for Free
RMachine Learning4 hr16 videos56 Exercises4,600 XP2,696Statement of Accomplishment

Create Your Free Account

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA.

Loved by learners at thousands of companies

Group

Training 2 or more people?

Try DataCamp for Business

Course Description

Do you ever work with datasets with an overwhelming number of features? Do you need all those features? Which ones are the most important? In this course, you will learn dimensionality reduction techniques that will help you simplify your data and the models that you build with your data while maintaining the information in the original data and good predictive performance.

Why learn dimensionality reduction?



We live in the information age—an era of information overload. The art of extracting essential information from data is a marketable skill. Models train faster on reduced data. In production, smaller models mean faster response time. Perhaps most important, smaller data and models are often easier to understand. Dimensionality reduction is your Occam’s razor in data science.

What will you learn in this course?



The difference between feature selection and feature extraction! Using R, you will learn how to identify and remove features with low or redundant information, keeping the features with the most information. That’s feature selection. You will also learn how to extract combinations of features as condensed components that contain maximal information. That’s feature extraction!

But most importantly, using R’s new tidymodel package, you will use real-world data to build models with fewer features without sacrificing significant performance.

Prerequisites

Modeling with tidymodels in R
1

Foundations of Dimensionality Reduction

Prepare to simplify large data sets! You will learn about information, how to assess feature importance, and practice identifying low-information features. By the end of the chapter, you will understand the difference between feature selection and feature extraction—the two approaches to dimensionality reduction.
Start Chapter
2

Feature Selection for Feature Importance

3

Feature Selection for Model Performance

4

Feature Extraction and Model Performance

In this final chapter, you'll gain a strong intuition of feature extraction by understanding how principal components extract and combine the most important information from different features. Then learn about and apply three types of feature extraction — principal component analysis (PCA), t-SNE, and UMAP. Discover how you can use these feature extraction methods as a preprocessing step in the tidymodels model-building process.
Start Chapter
Dimensionality Reduction in R
Course
Complete

Earn Statement of Accomplishment

Add this credential to your LinkedIn profile, resume, or CV
Share it on social media and in your performance review
Enroll Now

Don’t just take our word for it

*4.8
from 86 reviews
84%
16%
0%
0%
0%
  • Noah
    4 days ago

  • EDWIN SALIM
    2 weeks ago

  • Karina
    2 weeks ago

  • Matthieu
    2 weeks ago

  • Luis Enrique
    2 weeks ago

    amazing

  • Ryan
    4 weeks ago

Noah

EDWIN SALIM

Karina

FAQs

Is this course appropriate for someone new to machine learning in R?

Yes. It is listed as beginner level and assumes familiarity with tidyverse basics and tidymodels rather than advanced machine learning experience.

What dimensionality reduction techniques are covered?

You learn feature selection methods like variance and correlation filtering, lasso and random forest selection, and feature extraction techniques including PCA, t-SNE, and UMAP.

Does the course use the tidymodels framework?

Yes. You build tidymodels recipes and workflows throughout the course to select and extract features as part of a reproducible model-building pipeline.

What is the difference between feature selection and feature extraction?

Feature selection removes low-information variables, while feature extraction creates new condensed components like principal components that combine information from multiple original features.

Will reducing features hurt my model performance?

The course teaches you how to simplify data while maintaining strong predictive performance, showing that removing redundant features often improves model speed and interpretability.

Join over 19 million learners and start Dimensionality Reduction in R today!

Create Your Free Account

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA.

Grow your data skills with DataCamp for Mobile

Make progress on the go with our mobile courses and daily 5-minute coding challenges.