Last Updated on October 20, 2021
Principal component analysis (PCA) is an unsupervised machine learning technique. Perhaps the most popular use of principal component analysis is dimensionality reduction. Besides using PCA as a data preparation technique, we can also use it to help visualize data. A picture is worth a thousand words. With the data visualized, it is easier for us to get some insight and decide on the next step in our machine learning models.
In this tutorial, you will discover how to visualize data using PCA, as well as using visualization to help determining the parameter for dimensionality reduction.
After completing this tutorial, you will know:
How to use visualize a high dimensional data
What is explained variance in PCA
Visually observe the explained variance from the result of PCA of high dimensional data
Let’s get started.
Tutorial Overview
This tutorial is divided into two parts; they are:
Scatter plot of high dimensional data
Visualizing the explained variance
Prerequisites
For this tutorial, we assume that you are already familiar with:
How to Calculate Principal Component Analysis (PCA) from Scratch in Python
Principal Component Analysis for Dimensionality Reduction in Python
Scatter plot of high dimensional data
Visualization is a crucial step to get insight from data. We can learn from the visualization that whether a pattern can be observed and hence estimate which machine learning model is suitable.
It is easy to depict things in two dimension. Normally a scatter plot with x- and y-axis are in two dimensional. Depicting things in three dimensional is a bit challenging but not impossible. In matplotlib, for example, can plot in 3D. The only problem is on paper or on screen, we need can only look at a 3D plot at one viewport or projection at a time. In matplotlib, this is controlled by the degree of elevation and azimuth. Depicting things in four or five dimensions is impossible because we live in a three-dimensional world and have no idea of how things in such a high dimension would look like.
This is where a dimensionality reduction technique such as PCA comes into play. We can reduce the dimension to two or three so we can visualize it. Let’s start with an example.
We start with the wine dataset, which is a classification dataset with 13 features and 3 classes. There are 178 samples:
from sklearn.datasets import load_wine
winedata = load_wine()
X, y = winedata[‘data’], winedata[‘target’]
print(X.shape)
print(y.shape)
(178, 13)
(178,)
Among the 13 features, we can pick any two and plot with matplotlib (we color-coded the different classes using the c argument):
…
import matplotlib.pyplot as plt
plt.scatter(X[:,1], X[:,2], c=y)
plt.show()
or we can also pick any three and show in 3D:
…
ax = fig.add_subplot(projection=’3d’)
ax.scatter(X[:,1], X[:,2], X[:,3], c=y)
plt.show()
But these doesn’t reveal much of how the data looks like, because majority of the features are not shown. We now resort to principal component analysis:
…
from sklearn.decomposition import PCA
pca = PCA()
Xt = pca.fit_transform(X)
plot = plt.scatter(Xt[:,0], Xt[:,1], c=y)
plt.legend(handles=plot.legend_elements()[0], labels=list(winedata[‘target_names’]))
plt.show()
Here we transform the input data X by PCA into Xt. We consider only the first two columns, which contains the most information, and plot it in two dimensional. We can see that the purple class is quite distinctive, but there is still some overlap. But if we scale the data before PCA, the result would be different:
…
from sklearn.preprocessing import StandardScaler
from sklearn.pipeline import Pipeline
pca = PCA()
pipe = Pipeline([(‘scaler’, StandardScaler()), (‘pca’, pca)])
Xt = pipe.fit_transform(X)
plot = plt.scatter(Xt[:,0], Xt[:,1], c=y)
plt.legend(handles=plot.legend_elements()[0], labels=list(winedata[‘target_names’]))
plt.show()
Because PCA is sensitive to the scale, if we normalized each feature by StandardScaler we can see a better result. Here the different classes are more distinctive. By looking at this plot, we are confident that a simple model such as SVM can classify this dataset in high accuracy.
Putting these together, the following is the complete code to generate the visualizations:
from sklearn.datasets import load_wine
from sklearn.decomposition import PCA
from sklearn.preprocessing import StandardScaler
from sklearn.pipeline import Pipeline
import matplotlib.pyplot as plt
# Load dataset
winedata = load_wine()
X, y = winedata[‘data’], winedata[‘target’]
print(“X shape:”, X.shape)
print(“y shape:”, y.shape)
# Show any two features
plt.figure(figsize=(8,6))
plt.scatter(X[:,1], X[:,2], c=y)
plt.xlabel(winedata[“feature_names”][1])
plt.ylabel(winedata[“feature_names”][2])
plt.title(“Two particular features of the wine dataset”)
plt.show()
# Show any three features
fig = plt.figure(figsize=(10,8))
ax = fig.add_subplot(projection=’3d’)
ax.scatter(X[:,1], X[:,2], X[:,3], c=y)
ax.set_xlabel(winedata[“feature_names”][1])
ax.set_ylabel(winedata[“feature_names”][2])
ax.set_zlabel(winedata[“feature_names”][3])
ax.set_title(“Three particular features of the wine dataset”)
plt.show()
# Show first two principal components without scaler
pca = PCA()
plt.figure(figsize=(8,6))
Xt = pca.fit_transform(X)
plot = plt.scatter(Xt[:,0], Xt[:,1], c=y)
plt.legend(handles=plot.legend_elements()[0], labels=list(winedata[‘target_names’]))
plt.xlabel(“PC1”)
plt.ylabel(“PC2”)
plt.title(“First two principal components”)
plt.show()
# Show first two principal components with scaler
pca = PCA()
pipe = Pipeline([(‘scaler’, StandardScaler()), (‘pca’, pca)])
plt.figure(figsize=(8,6))
Xt = pipe.fit_transform(X)
plot = plt.scatter(Xt[:,0], Xt[:,1], c=y)
plt.legend(handles=plot.legend_elements()[0], labels=list(winedata[‘target_names’]))
plt.xlabel(“PC1”)
plt.ylabel(“PC2”)
plt.title(“First two principal components after scaling”)
plt.show()
If we apply the same method on a different dataset, such as MINST handwritten digits, the scatterplot is not showing distinctive boundary and therefore it needs a more complicated model such as neural network to classify:
from sklearn.datasets import load_digits
from sklearn.decomposition import PCA
from sklearn.preprocessing import StandardScaler
from sklearn.pipeline import Pipeline
import matplotlib.pyplot as plt
digitsdata = load_digits()
X, y = digitsdata[‘data’], digitsdata[‘target’]
pca = PCA()
pipe = Pipeline([(‘scaler’, StandardScaler()), (‘pca’, pca)])
plt.figure(figsize=(8,6))
Xt = pipe.fit_transform(X)
plot = plt.scatter(Xt[:,0], Xt[:,1], c=y)
plt.legend(handles=plot.legend_elements()[0], labels=list(digitsdata[‘target_names’]))
plt.show()
Visualizing the explained variance
PCA in essence is to rearrange the features by their linear combinations. Hence it is called a feature extraction technique. One characteristic of PCA is that the first principal component holds the most information about the dataset. The second principal component is more informative than the third, and so on.
To illustrate this idea, we can remove the principal components from the original dataset in steps and see how the dataset looks like. Let’s consider a dataset with fewer features, and show two features in a plot:
from sklearn.datasets import load_iris
irisdata = load_iris()
X, y = irisdata[‘data’], irisdata[‘target’]
plt.figure(figsize=(8,6))
plt.scatter(X[:,0], X[:,1], c=y)
plt.show()
This is the iris dataset which has only four features. The features are in comparable scales and hence we can skip the scaler. With a 4-features data, the PCA can produce at most 4 principal components:
…
pca = PCA().fit(X)
print(pca.components_)
[[ 0.36138659 -0.08452251 0.85667061 0.3582892 ]
[ 0.65658877 0.73016143 -0.17337266 -0.07548102]
[-0.58202985 0.59791083 0.07623608 0.54583143]
[-0.31548719 0.3197231 0.47983899 -0.75365743]]
For example, the first row is the first principal axis on which the first principal component is created. For any data point $p$ with features $p=(a,b,c,d)$, since the principal axis is denoted by the vector $v=(0.36,-0.08,0.86,0.36)$, the first principal component of this data point has the value $0.36 times a – 0.08 times b + 0.86 times c + 0.36times d$ on the principal axis. Using vector dot product, this value can be denoted by
$$
p cdot v
$$
Therefore, with the dataset $X$ as a 150 $times$ 4 matrix (150 data points, each has 4 features), we can map each data point into to the value on this principal axis by matrix-vector multiplication:
$$
X times v
$$
and the result is a vector of length 150. Now if we remove from each data point corresponding value along the principal axis vector, that would be
$$
X – (X times v) times v^T
$$
where the transposed vector $v^T$ is a row and $Xtimes v$ is a column. The product $(X times v) times v^T$ follows matrix-matrix multiplication and the result is a $150times 4$ matrix, same dimension as $X$.
If we plot the first two feature of $(X times v) times v^T$, it looks like this:
…
# Remove PC1
Xmean = X – X.mean(axis=0)
value = Xmean @ pca.components_[0]
pc1 = value.reshape(-1,1) @ pca.components_[0].reshape(1,-1)
Xremove = X – pc1
plt.scatter(Xremove[:,0], Xremove[:,1], c=y)
plt.show()
The numpy array Xmean is to shift the features of X to centered at zero. This is required for PCA. Then the array value is computed by matrix-vector multiplication.
The array value is the magnitude of each data point mapped on the principal axis. So if we multiply this value to the principal axis vector we get back an array pc1. Removing this from the original dataset X, we get a new array Xremove. In the plot we observed that the points on the scatter plot crumbled together and the cluster of each class is less distinctive than before. This means we removed a lot of information by removing the first principal component. If we repeat the same process again, the points are further crumbled:
…
# Remove PC2
value = Xmean @ pca.components_[1]
pc2 = value.reshape(-1,1) @ pca.components_[1].reshape(1,-1)
Xremove = Xremove – pc2
plt.scatter(Xremove[:,0], Xremove[:,1], c=y)
plt.show()
This looks like a straight line but actually not. If we repeat once more, all points collapse into a straight line:
…
# Remove PC3
value = Xmean @ pca.components_[2]
pc3 = value.reshape(-1,1) @ pca.components_[2].reshape(1,-1)
Xremove = Xremove – pc3
plt.scatter(Xremove[:,0], Xremove[:,1], c=y)
plt.show()
The points all fall on a straight line because we removed three principal components from the data where there are only four features. Hence our data matrix becomes rank 1. You can try repeat once more this process and the result would be all points collapse into a single point. The amount of information removed in each step as we removed the principal components can be found by the corresponding explained variance ratio from the PCA:
…
print(pca.explained_variance_ratio_)
[0.92461872 0.05306648 0.01710261 0.00521218]
Here we can see, the first component explained 92.5% variance and the second component explained 5.3% variance. If we removed the first two principal components, the remaining variance is only 2.2%, hence visually the plot after removing two components looks like a straight line. In fact, when we check with the plots above, not only we see the points are crumbled, but the range in the x- and y-axes are also smaller as we removed the components.
In terms of machine learning, we can consider using only one single feature for classification in this dataset, namely the first principal component. We should expect to achieve no less than 90% of the original accuracy as using the full set of features:
…
from sklearn.model_selection import train_test_split
from sklearn.metrics import f1_score
from collections import Counter
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.33)
from sklearn.svm import SVC
clf = SVC(kernel=”linear”, gamma=’auto’).fit(X_train, y_train)
print(“Using all features, accuracy: “, clf.score(X_test, y_test))
print(“Using all features, F1: “, f1_score(y_test, clf.predict(X_test), average=”macro”))
mean = X_train.mean(axis=0)
X_train2 = X_train – mean
X_train2 = (X_train2 @ pca.components_[0]).reshape(-1,1)
clf = SVC(kernel=”linear”, gamma=’auto’).fit(X_train2, y_train)
X_test2 = X_test – mean
X_test2 = (X_test2 @ pca.components_[0]).reshape(-1,1)
print(“Using PC1, accuracy: “, clf.score(X_test2, y_test))
print(“Using PC1, F1: “, f1_score(y_test, clf.predict(X_test2), average=”macro”))
Using all features, accuracy: 1.0
Using all features, F1: 1.0
Using PC1, accuracy: 0.96
Using PC1, F1: 0.9645191409897292
The other use of understanding the explained variance is on compression. Given the explained variance of the first principal component is large, if we need to store the dataset, we can store only the the projected values on the first principal axis ($Xtimes v$), as well as the vector $v$ of the principal axis. Then we can approximately reproduce the original dataset by multiplying them:
$$
X approx (Xtimes v) times v^T
$$
In this way, we need storage for only one value per data point instead of four values for four features. The approximation is more accurate if we store the projected values on multiple principal axes and add up multiple principal components.
Putting these together, the following is the complete code to generate the visualizations:
from sklearn.datasets import load_iris
from sklearn.model_selection import train_test_split
from sklearn.decomposition import PCA
from sklearn.metrics import f1_score
from sklearn.svm import SVC
import matplotlib.pyplot as plt
# Load iris dataset
irisdata = load_iris()
X, y = irisdata[‘data’], irisdata[‘target’]
plt.figure(figsize=(8,6))
plt.scatter(X[:,0], X[:,1], c=y)
plt.xlabel(irisdata[“feature_names”][0])
plt.ylabel(irisdata[“feature_names”][1])
plt.title(“Two features from the iris dataset”)
plt.show()
# Show the principal components
pca = PCA().fit(X)
print(“Principal components:”)
print(pca.components_)
# Remove PC1
Xmean = X – X.mean(axis=0)
value = Xmean @ pca.components_[0]
pc1 = value.reshape(-1,1) @ pca.components_[0].reshape(1,-1)
Xremove = X – pc1
plt.figure(figsize=(8,6))
plt.scatter(Xremove[:,0], Xremove[:,1], c=y)
plt.xlabel(irisdata[“feature_names”][0])
plt.ylabel(irisdata[“feature_names”][1])
plt.title(“Two features from the iris dataset after removing PC1”)
plt.show()
# Remove PC2
Xmean = X – X.mean(axis=0)
value = Xmean @ pca.components_[1]
pc2 = value.reshape(-1,1) @ pca.components_[1].reshape(1,-1)
Xremove = Xremove – pc2
plt.figure(figsize=(8,6))
plt.scatter(Xremove[:,0], Xremove[:,1], c=y)
plt.xlabel(irisdata[“feature_names”][0])
plt.ylabel(irisdata[“feature_names”][1])
plt.title(“Two features from the iris dataset after removing PC1 and PC2”)
plt.show()
# Remove PC3
Xmean = X – X.mean(axis=0)
value = Xmean @ pca.components_[2]
pc3 = value.reshape(-1,1) @ pca.components_[2].reshape(1,-1)
Xremove = Xremove – pc3
plt.figure(figsize=(8,6))
plt.scatter(Xremove[:,0], Xremove[:,1], c=y)
plt.xlabel(irisdata[“feature_names”][0])
plt.ylabel(irisdata[“feature_names”][1])
plt.title(“Two features from the iris dataset after removing PC1 to PC3”)
plt.show()
# Print the explained variance ratio
print(“Explainedd variance ratios:”)
print(pca.explained_variance_ratio_)
# Split data
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.33)
# Run classifer on all features
clf = SVC(kernel=”linear”, gamma=’auto’).fit(X_train, y_train)
print(“Using all features, accuracy: “, clf.score(X_test, y_test))
print(“Using all features, F1: “, f1_score(y_test, clf.predict(X_test), average=”macro”))
# Run classifier on PC1
mean = X_train.mean(axis=0)
X_train2 = X_train – mean
X_train2 = (X_train2 @ pca.components_[0]).reshape(-1,1)
clf = SVC(kernel=”linear”, gamma=’auto’).fit(X_train2, y_train)
X_test2 = X_test – mean
X_test2 = (X_test2 @ pca.components_[0]).reshape(-1,1)
print(“Using PC1, accuracy: “, clf.score(X_test2, y_test))
print(“Using PC1, F1: “, f1_score(y_test, clf.predict(X_test2), average=”macro”))
Further reading
This section provides more resources on the topic if you are looking to go deeper.
Books
Tutorials
How to Calculate Principal Component Analysis (PCA) from Scratch in Python
Principal Component Analysis for Dimensionality Reduction in Python
APIs
scikit-learn toy datasets
scikit-learn iris dataset
scikit-learn wine dataset
matplotlib scatter API
The mplot3d toolkit
Summary
In this tutorial, you discovered how to visualize data using principal component analysis.
Specifically, you learned:
Visualize a high dimensional dataset in 2D using PCA
How to use the plot in PCA dimensions to help choosing an appropriate machine learning model
How to observe the explained variance ratio of PCA
What the explained variance ratio means for machine learning
Â
The post Principal Component Analysis for Visualization appeared first on Machine Learning Mastery.
Read MoreMachine Learning Mastery