Kurs
Supervised Learning in R: Regression
Orta SeviyeBeceri Seviyesi
Güncel 01.2025Kursa Ücretsiz Başlayın
Şuna dahil:Premium or Takımlar
RMachine Learning4 sa19 video65 Egzersiz5,300 XP45,988Başarı Belgesi
Ücretsiz Hesabınızı Oluşturun
veya
Devam ederek Kullanım Şartlarımızı, Gizlilik Politikamızı ve verilerinizin ABD’de saklandığını kabul etmiş olursunuz.Binlerce şirketten öğrencinin sevgisini kazandı
2 veya daha fazla kişiyi mi eğitiyorsunuz?
DataCamp for Business ürününü deneyinKurs Açıklaması
Önkoşullar
Introduction to Regression in R1
What is Regression?
In this chapter we introduce the concept of regression from a machine learning point of view. We will present the fundamental regression method: linear regression. We will show how to fit a linear regression model and to make predictions from the model.
2
Training and Evaluating Regression Models
Now that we have learned how to fit basic linear regression models, we will learn how to evaluate how well our models perform. We will review evaluating a model graphically, and look at two basic metrics for regression models. We will also learn how to train a model that will perform well in the wild, not just on training data. Although we will demonstrate these techniques using linear regression, all these concepts apply to models fit with any regression algorithm.
3
Issues to Consider
Before moving on to more sophisticated regression techniques, we will look at some other modeling issues: modeling with categorical inputs, interactions between variables, and when you might consider transforming inputs and outputs before modeling. While more sophisticated regression techniques manage some of these issues automatically, it's important to be aware of them, in order to understand which methods best handle various issues -- and which issues you must still manage yourself.
4
Dealing with Non-Linear Responses
Now that we have mastered linear models, we will begin to look at techniques for modeling situations that don't meet the assumptions of linearity. This includes predicting probabilities and frequencies (values bounded between 0 and 1); predicting counts (nonnegative integer values, and associated rates); and responses that have a non-linear but additive relationship to the inputs. These algorithms are variations on the standard linear model.
5
Tree-Based Methods
In this chapter we will look at modeling algorithms that do not assume linearity or additivity, and that can learn limited types of interactions among input variables. These algorithms are tree-based methods that work by combining ensembles of decision trees that are learned from the training data.
Supervised Learning in R: Regression
Kurs Tamamlandı
Başarı Belgesi Kazanın
Bu kimlik bilgisini LinkedIn profilinize, özgeçmişinize veya CV'nize ekleyinSosyal medyada ve performans incelemenizde paylaşın
Şuna dahil:Premium or Takımlar
Şimdi KaydolunBugün 19 milyondan fazla öğrenciye katılın ve Supervised Learning in R: Regression eğitimine başlayın!
Ücretsiz Hesabınızı Oluşturun
veya
Devam ederek Kullanım Şartlarımızı, Gizlilik Politikamızı ve verilerinizin ABD’de saklandığını kabul etmiş olursunuz.