In previous chapters, our work centered on supervised learning, where algorithms learn from data that includes known outcomes or labels. This chapter shifts focus to unsupervised learning. This area of machine learning deals with datasets that lack predefined labels. Here, the objective is for algorithms to independently identify patterns, structures, or relationships within the data.
We will also address dimensionality reduction. Datasets with many features can present challenges for modeling and computation. Dimensionality reduction techniques aim to reduce the number of input variables while retaining meaningful properties of the data. This can lead to simpler models, faster computations, and can assist in visualizing data by projecting it into a lower-dimensional space, for instance, from Rn to R2 or R3.
Throughout this chapter, you will learn to implement several widely used unsupervised learning algorithms in Julia. We will cover clustering methods such as K-Means, which segments data into a specified number k of clusters, and DBSCAN, a density-based approach. Additionally, you will work with Principal Component Analysis (PCA), a common technique for reducing data dimensions. Finally, we'll examine approaches to evaluate the performance of your clustering results.
4.1 Clustering with K-Means in Julia
4.2 Density-Based Clustering: DBSCAN
4.3 Principal Component Analysis (PCA) for Dimensionality Reduction
4.4 Other Dimensionality Reduction Techniques
4.5 Evaluating Clustering Performance
4.6 Hands-on practical: Applying Clustering and PCA
© 2025 ApX Machine Learning