Skip to main content
HomeRFeature Engineering in R

Feature Engineering in R

Learn the principles of feature engineering for machine learning models and how to implement them using the R tidymodels framework.

Start Course for Free
4 hours14 videos58 exercises

Create Your Free Account

GoogleLinkedInFacebook

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA.
Group

Training 2 or more people?

Try DataCamp For Business

Loved by learners at thousands of companies


Course Description

Discover Feature Engineering for Machine Learning

In this course, you’ll learn about feature engineering, which is at the heart of many times of machine learning models. As the performance of any model is a direct consequence of the features it’s fed, feature engineering places domain knowledge at the center of the process. You’ll become acquainted with principles of sound feature engineering, helping to reduce the number of variables where possible, making learning algorithms run faster, improving interpretability, and preventing overfitting.

Implement Feature Engineering Techniques in R

You will learn how to implement feature engineering techniques using the R tidymodels framework, emphasizing the recipe package that will allow you to create, extract, transform, and select the best features for your model.

Engineer Features and Build Better ML Models

When faced with a new dataset, you will be able to identify and select relevant features and disregard non-informative ones to make your model run faster without sacrificing accuracy. You will also become comfortable applying transformations and creating new features to make your models more efficient, interpretable, and accurate!
For Business

GroupTraining 2 or more people?

Get your team access to the full DataCamp library, with centralized reporting, assignments, projects and more
Try DataCamp for BusinessFor a bespoke solution book a demo.

In the following Tracks

Certification Available

Data Scientist in R

Go To Track

Machine Learning Scientist in R

Go To Track
  1. 1

    Introducing Feature Engineering

    Free

    Raw data does not always come in its best shape for analysis. In this opening chapter, you will get a first look at how to transform and create features that enhance your model's performance and interpretability.

    Play Chapter Now
    What is feature engineering?
    50 xp
    A tentative model
    100 xp
    Manually engineering a feature
    100 xp
    Creating new features using domain knowledge
    50 xp
    Setting up your data for analysis
    100 xp
    Building a workflow
    100 xp
    Increasing the information content of raw data
    50 xp
    Identifying missing values
    100 xp
    Imputing missing values and creating dummy variables
    100 xp
    Fitting and assessing the model
    100 xp
    Predicting hotel bookings
    100 xp
  2. 2

    Transforming Features

    In this chapter, you’ll learn that, beyond manually transforming features, you can leverage tools from the tidyverse to engineer new variables programmatically. You’ll explore how this approach improves your models' reproducibility and is especially useful when handling datasets with many features.

    Play Chapter Now
  3. 3

    Extracting Features

    You’ll now learn how models often benefit from reducing dimensionality and extracting features from high-dimensional data, including converting text data into numeric values, encoding categorical data, and ranking the predictive power of variables. You’ll explore methods including principal component analysis, kernel principal component analysis, numerical extraction from text, categorical encodings, and variable importance scores.

    Play Chapter Now
  4. 4

    Selecting Features

    You’ll wrap up the course by learning about feature engineering and machine learning techniques. You’ll begin by focusing on the problems associated with using all available features in a model and the importance of identifying irrelevant and redundant features and learning to remove these features using embedded methods such as lasso and elastic-net. Next, you’ll explore shrinkage methods such as lasso, ridge, and elastic-net, which can be used to regularize feature weights or select features by setting coefficients to zero. Finally, you’ll finish by focusing on creating an end-to-end feature engineering workflow and reviewing and practicing the previously learned concepts and functions in a small project.

    Play Chapter Now
For Business

GroupTraining 2 or more people?

Get your team access to the full DataCamp library, with centralized reporting, assignments, projects and more

In the following Tracks

Certification Available

Data Scientist in R

Go To Track

Machine Learning Scientist in R

Go To Track

collaborators

Collaborator's avatar
Maham Khan
Collaborator's avatar
Arne Warnke

prerequisites

Supervised Learning in R: ClassificationSupervised Learning in R: Regression
Jorge Zazueta HeadshotJorge Zazueta

Research Professor

Jorge Zazueta is the Head of the Modeling Group at the School of Economics, UASLP.
See More

What do other learners have to say?

Join over 14 million learners and start Feature Engineering in R today!

Create Your Free Account

GoogleLinkedInFacebook

or

By continuing, you accept our Terms of Use, our Privacy Policy and that your data is stored in the USA.