Extreme Gradient Boosting with XGBoost
Learn the fundamentals of gradient boosting and build state-of-the-art machine learning models using XGBoost to solve classification and regression problems.
Kurs Kostenlos Starten4 Stunden16 Videos49 Übungen52.557 LernendeLeistungsnachweis
Kostenloses Konto erstellen
oder
Durch Klick auf die Schaltfläche akzeptierst du unsere Nutzungsbedingungen, unsere Datenschutzrichtlinie und die Speicherung deiner Daten in den USA.Trainierst du 2 oder mehr?
Versuchen DataCamp for BusinessBeliebt bei Lernenden in Tausenden Unternehmen
Kursbeschreibung
Do you know the basics of supervised learning and want to use state-of-the-art models on real-world datasets? Gradient boosting is currently one of the most popular techniques for efficient modeling of tabular datasets of all sizes. XGboost is a very fast, scalable implementation of gradient boosting, with models using XGBoost regularly winning online data science competitions and being used at scale across different industries. In this course, you'll learn how to use this powerful library alongside pandas and scikit-learn to build and tune supervised learning models. You'll work with real-world datasets to solve classification and regression problems.
Trainierst du 2 oder mehr?
Verschaffen Sie Ihrem Team Zugriff auf die vollständige DataCamp-Plattform, einschließlich aller Funktionen.In den folgenden Tracks
Machine Learning Scientist mit Python
Gehe zu TrackÜberwachtes Machine Learning in Python
Gehe zu Track- 1
Classification with XGBoost
KostenlosThis chapter will introduce you to the fundamental idea behind XGBoost—boosted learners. Once you understand how XGBoost works, you'll apply it to solve a common classification problem found in industry: predicting whether a customer will stop being a customer at some point in the future.
Welcome to the course!50 xpWhich of these is a classification problem?50 xpWhich of these is a binary classification problem?50 xpIntroducing XGBoost50 xpXGBoost: Fit/Predict100 xpWhat is a decision tree?50 xpDecision trees100 xpWhat is Boosting?50 xpMeasuring accuracy100 xpMeasuring AUC100 xpWhen should I use XGBoost?50 xpUsing XGBoost50 xp - 2
Regression with XGBoost
After a brief review of supervised regression, you'll apply XGBoost to the regression task of predicting house prices in Ames, Iowa. You'll learn about the two kinds of base learners that XGboost can use as its weak learners, and review how to evaluate the quality of your regression models.
Regression review50 xpWhich of these is a regression problem?50 xpObjective (loss) functions and base learners50 xpDecision trees as base learners100 xpLinear base learners100 xpEvaluating model quality100 xpRegularization and base learners in XGBoost50 xpUsing regularization in XGBoost100 xpVisualizing individual XGBoost trees100 xpVisualizing feature importances: What features are most important in my dataset100 xp - 3
Fine-tuning your XGBoost model
This chapter will teach you how to make your XGBoost models as performant as possible. You'll learn about the variety of parameters that can be adjusted to alter the behavior of XGBoost and how to tune them efficiently so that you can supercharge the performance of your models.
Why tune your model?50 xpWhen is tuning your model a bad idea?50 xpTuning the number of boosting rounds100 xpAutomated boosting round selection using early_stopping100 xpOverview of XGBoost's hyperparameters50 xpTuning eta100 xpTuning max_depth100 xpTuning colsample_bytree100 xpReview of grid search and random search50 xpGrid search with XGBoost100 xpRandom search with XGBoost100 xpLimits of grid search and random search50 xpWhen should you use grid search and random search?50 xp - 4
Using XGBoost in pipelines
Take your XGBoost skills to the next level by incorporating your models into two end-to-end machine learning pipelines. You'll learn how to tune the most important XGBoost hyperparameters efficiently within a pipeline, and get an introduction to some more advanced preprocessing techniques.
Review of pipelines using sklearn50 xpExploratory data analysis50 xpEncoding categorical columns I: LabelEncoder100 xpEncoding categorical columns II: OneHotEncoder100 xpEncoding categorical columns III: DictVectorizer100 xpPreprocessing within a pipeline100 xpIncorporating XGBoost into pipelines50 xpCross-validating your XGBoost model100 xpKidney disease case study I: Categorical Imputer100 xpKidney disease case study II: Feature Union100 xpKidney disease case study III: Full pipeline100 xpTuning XGBoost hyperparameters50 xpBringing it all together100 xpFinal Thoughts50 xp
Trainierst du 2 oder mehr?
Verschaffen Sie Ihrem Team Zugriff auf die vollständige DataCamp-Plattform, einschließlich aller Funktionen.In den folgenden Tracks
Machine Learning Scientist mit Python
Gehe zu TrackÜberwachtes Machine Learning in Python
Gehe zu TrackMitwirkende
Voraussetzungen
Supervised Learning with scikit-learnSergey Fogelson
Mehr AnzeigenHead of Data Science, TelevisaUnivision
Was sagen andere Lernende?
Melden Sie sich an 15 Millionen Lernende und starten Sie Extreme Gradient Boosting with XGBoost Heute!
Kostenloses Konto erstellen
oder
Durch Klick auf die Schaltfläche akzeptierst du unsere Nutzungsbedingungen, unsere Datenschutzrichtlinie und die Speicherung deiner Daten in den USA.