Introduction to Feature Engineering in R
Learn a variety of feature engineering techniques to develop meaningful features that will uncover useful insights about your machine learning models.
Course Description
Feature engineering helps you uncover useful insights from your machine learning models. The model building process is iterative and requires creating new features using existing variables that make your model more efficient. In this course, you will explore different data sets and apply a variety of feature engineering techniques to both continuous and discrete variables.
What You’ll Learn
Creating Features from Categorical Data
In this chapter, you will learn how to change categorical features into numerical representations that models can interpret. You’ll learn about one-hot encoding and using binning for categorical features.
Transforming Numerical Features
In this chapter, you will learn about using transformation techniques, like Box-Cox and Yeo-Johnson, to address issues with non-normally distributed features. You’ll also learn about methods to scale features, including mean centering and z-score standardization.
Creating Features from Numeric Data
In this chapter, you will learn how to manipulate numerical features to create meaningful features that can give better insights into your model. You will also learn how to work with dates in the context of feature engineering.
Advanced Methods
In the final chapter, we will use feature crossing to create features from two or more variables. We will also discuss principal component analysis, and methods to explore and visualize those results.
User Reviews
Be the first to review “Introduction to Feature Engineering in R”
You must be logged in to post a review.



There are no reviews yet.