Before we examine how autoencoders learn new representations from data, it's essential to first understand what we mean by "features" within a dataset. Think of features as the individual, measurable properties or characteristics of the data you're working with. They are the building blocks that describe each piece of information, and machine learning models use these features to make predictions or find patterns.
Perhaps the most straightforward way to understand features is by looking at structured data, often organized in tables, like a spreadsheet. In such a table:
For instance, if you have a dataset about customers, the features might include:
Age
(e.g., 34 years)Annual Income
(e.g., $50,000)City of Residence
(e.g., "New York")Has Subscribed
(e.g., True/False)Each of these columns provides a distinct piece of information (a feature) about each customer. A machine learning model could use these features to, for example, predict whether a new customer is likely to subscribe to a service.
Consider the following simple representation of a dataset:
In this housing dataset example, "Square Footage," "No. of Bedrooms," and "Distance to City Center" are features that describe each house. "Price" might be what we want to predict.
Features aren't limited to neat columns in a table.
The common thread is that features are the numerical or categorical inputs derived from raw data that a machine learning algorithm uses to perform its task.
The selection and quality of features are fundamental to the success of any machine learning project.
In essence, features are the lens through which a machine learning model views the data.
As we proceed through this chapter, we'll see that autoencoders don't just work with predefined features; their strength lies in their ability to learn new, often more compact and informative, features from the initial raw input features. The bottleneck layer in an autoencoder, which we've discussed previously, is where these learned features reside. Having a clear understanding of what "features" are in their basic sense provides the foundation for appreciating how autoencoders transform them.
Was this section helpful?
© 2025 ApX Machine Learning