Accéder au contenu principal
AccueilPython

Machine Learning with Tree-Based Models in Python

In this course, you'll learn how to use tree-based models and ensembles for regression and classification using scikit-learn.

Commencer Le Cours Gratuitement
5 heures15 vidéos57 exercices95 211 apprenantsTrophyDéclaration de réalisation

Créez votre compte gratuit

GoogleLinkedInFacebook

ou

En continuant, vous acceptez nos Conditions d'utilisation, notre Politique de confidentialité et le fait que vos données sont stockées aux États-Unis.
Group

Formation de 2 personnes ou plus ?

Essayer DataCamp for Business

Apprécié par les apprenants de milliers d'entreprises


Description du cours

Decision trees are supervised learning models used for problems involving classification and regression. Tree models present a high flexibility that comes at a price: on one hand, trees are able to capture complex non-linear relationships; on the other hand, they are prone to memorizing the noise present in a dataset. By aggregating the predictions of trees that are trained differently, ensemble methods take advantage of the flexibility of trees while reducing their tendency to memorize noise. Ensemble methods are used across a variety of fields and have a proven track record of winning many machine learning competitions. In this course, you'll learn how to use Python to train decision trees and tree-based models with the user-friendly scikit-learn machine learning library. You'll understand the advantages and shortcomings of trees and demonstrate how ensembling can alleviate these shortcomings, all while practicing on real-world datasets. Finally, you'll also understand how to tune the most influential hyperparameters in order to get the most out of your models.
Pour les entreprises

Formation de 2 personnes ou plus ?

Donnez à votre équipe l’accès à la plateforme DataCamp complète, y compris toutes les fonctionnalités.
DataCamp Pour Les EntreprisesPour une solution sur mesure , réservez une démo.

Dans les titres suivants

Certification disponible

Scientifique de données associé en Python

Aller à la piste

Scientifique en apprentissage automatique en Python

Aller à la piste

Apprentissage automatique supervisé en Python

Aller à la piste
  1. 1

    Classification and Regression Trees

    Gratuit

    Classification and Regression Trees (CART) are a set of supervised learning models used for problems involving classification and regression. In this chapter, you'll be introduced to the CART algorithm.

    Jouez Au Chapitre Maintenant
    Decision tree for classification
    50 xp
    Train your first classification tree
    100 xp
    Evaluate the classification tree
    100 xp
    Logistic regression vs classification tree
    100 xp
    Classification tree Learning
    50 xp
    Growing a classification tree
    50 xp
    Using entropy as a criterion
    100 xp
    Entropy vs Gini index
    100 xp
    Decision tree for regression
    50 xp
    Train your first regression tree
    100 xp
    Evaluate the regression tree
    100 xp
    Linear regression vs regression tree
    100 xp
  2. 2

    The Bias-Variance Tradeoff

    The bias-variance tradeoff is one of the fundamental concepts in supervised machine learning. In this chapter, you'll understand how to diagnose the problems of overfitting and underfitting. You'll also be introduced to the concept of ensembling where the predictions of several models are aggregated to produce predictions that are more robust.

    Jouez Au Chapitre Maintenant
  3. 3

    Bagging and Random Forests

    Bagging is an ensemble method involving training the same algorithm many times using different subsets sampled from the training data. In this chapter, you'll understand how bagging can be used to create a tree ensemble. You'll also learn how the random forests algorithm can lead to further ensemble diversity through randomization at the level of each split in the trees forming the ensemble.

    Jouez Au Chapitre Maintenant
  4. 4

    Boosting

    Boosting refers to an ensemble method in which several models are trained sequentially with each model learning from the errors of its predecessors. In this chapter, you'll be introduced to the two boosting methods of AdaBoost and Gradient Boosting.

    Jouez Au Chapitre Maintenant
  5. 5

    Model Tuning

    The hyperparameters of a machine learning model are parameters that are not learned from data. They should be set prior to fitting the model to the training set. In this chapter, you'll learn how to tune the hyperparameters of a tree-based model using grid search cross validation.

    Jouez Au Chapitre Maintenant
Pour les entreprises

Formation de 2 personnes ou plus ?

Donnez à votre équipe l’accès à la plateforme DataCamp complète, y compris toutes les fonctionnalités.

Dans les titres suivants

Certification disponible

Scientifique de données associé en Python

Aller à la piste

Scientifique en apprentissage automatique en Python

Aller à la piste

Apprentissage automatique supervisé en Python

Aller à la piste

ensembles de données

Auto-mpgBike Sharing DemandWisconsin Breast CancerIndian Liver Patient

collaborateurs

Collaborator's avatar
Kara Woo
Collaborator's avatar
Eunkyung Park
Collaborator's avatar
Sumedh Panchadhar
Elie Kawerk HeadshotElie Kawerk

Senior Data Scientist

Voir Plus

Qu’est-ce que les autres apprenants ont à dire ?

Inscrivez-vous 15 millions d’apprenants et commencer Machine Learning with Tree-Based Models in Python Aujourd’hui!

Créez votre compte gratuit

GoogleLinkedInFacebook

ou

En continuant, vous acceptez nos Conditions d'utilisation, notre Politique de confidentialité et le fait que vos données sont stockées aux États-Unis.