Book contents
- Frontmatter
- Dedication
- Contents
- Preface
- Part One Fundamentals
- 1 Introduction
- 2 Features, Combined: Normalization, Discretization and Outliers
- 3 Features, Expanded: Computable Features, Imputation and Kernels
- 4 Features, Reduced: Feature Selection, Dimensionality Reduction and Embeddings
- 5 Advanced Topics: Variable-Length Data and Automated Feature Engineering
- Part II Case Studies
- Bibliography
- Index
3 - Features, Expanded: Computable Features, Imputation and Kernels
from Part One - Fundamentals
Published online by Cambridge University Press: 29 May 2020
- Frontmatter
- Dedication
- Contents
- Preface
- Part One Fundamentals
- 1 Introduction
- 2 Features, Combined: Normalization, Discretization and Outliers
- 3 Features, Expanded: Computable Features, Imputation and Kernels
- 4 Features, Reduced: Feature Selection, Dimensionality Reduction and Embeddings
- 5 Advanced Topics: Variable-Length Data and Automated Feature Engineering
- Part II Case Studies
- Bibliography
- Index
Summary
This chapter deals with the topic of feature expansion and imputation with a particular emphasis on computable features. While poor domain modelling may result in too many features being added to the model, there are times when plenty of value can be gained by looking into generating features from existing ones. The excess features can then be removed using feature selection techniques (discussed in the next chapter). Computable Features will be particularly useful if we know the underlining ML model is unable to do certain operations over the features, like multiplying them (e.g., if the ML involves a simple, linear modelling). Another type of feature expansion involves calculating a best effort approximation of values missing in the data (Feature Imputation). The most straightforward expansion for features happens when the raw data contains multiple items of information under a single column (Decomposing Complex Features). The chapter concludes by borrowing ideas from a technique used in SVMs called the kernel trick. The type of projections that practitioners have found useful can lend themselves to be applied directly without the use of kernels.
Keywords
- Type
- Chapter
- Information
- The Art of Feature EngineeringEssentials for Machine Learning, pp. 59 - 78Publisher: Cambridge University PressPrint publication year: 2020