Making developers awesome at machine learning
Making developers awesome at machine learning
Dimensionality reduction is an unsupervised learning technique.
Nevertheless, it can be used as a data transform pre-processing step for machine learning algorithms on classification and regression predictive modeling datasets with supervised learning algorithms.
There are many dimensionality reduction algorithms to choose from and no single best algorithm for all cases. Instead, it is a good idea to explore a range of dimensionality reduction algorithms and different configurations for each algorithm.
In this tutorial, you will discover how to fit and evaluate top dimensionality reduction algorithms in Python.
After completing this tutorial, you will know:
Kick-start your project with my new bookData Preparation for Machine Learning, includingstep-by-step tutorials and thePython source code files for all examples.
Let’s get started.

Dimensionality Reduction Algorithms With Python
Photo byBernard Spragg. NZ, some rights reserved.
This tutorial is divided into three parts; they are:
Dimensionality reduction refers to techniques for reducing the number of input variables in training data.
When dealing with high dimensional data, it is often useful to reduce the dimensionality by projecting the data to a lower dimensional subspace which captures the “essence” of the data. This is called dimensionality reduction.
— Page 11,Machine Learning: A Probabilistic Perspective, 2012.
High-dimensionality might mean hundreds, thousands, or even millions of input variables.
Fewer input dimensions often means correspondingly fewer parameters or a simpler structure in the machine learning model, referred to asdegrees of freedom. A model with too many degrees of freedom is likely to overfit the training dataset and may not perform well on new data.
It is desirable to have simple models that generalize well, and in turn, input data with few input variables. This is particularly true for linear models where the number of inputs and the degrees of freedom of the model are often closely related.
Dimensionality reduction is a data preparation technique performed on data prior to modeling. It might be performed after data cleaning and data scaling and before training a predictive model.
… dimensionality reduction yields a more compact, more easily interpretable representation of the target concept, focusing the user’s attention on the most relevant variables.
— Page 289,Data Mining: Practical Machine Learning Tools and Techniques, 4th edition, 2016.
As such, any dimensionality reduction performed on training data must also be performed on new data, such as a test dataset, validation dataset, and data when making a prediction with thefinal model.
Take my free 7-day email crash course now (with sample code).
Click to sign-up and also get a free PDF Ebook version of the course.
There are many algorithms that can be used for dimensionality reduction.
Two main classes of methods are those drawn from linear algebra and those drawn from manifold learning.
Matrix factorization methods drawn from the field of linear algebra can be used for dimensionality.
For more on matrix factorization, see the tutorial:
Some of the more popular methods include:
Manifold learning methods seek a lower-dimensional projection of high dimensional input that captures the salient properties of the input data.
Some of the more popular methods include:
Each algorithm offers a different approach to the challenge of discovering natural relationships in data at lower dimensions.
There is no best dimensionality reduction algorithm, and no easy way to find the best algorithm for your data without using controlled experiments.
In this tutorial, we will review how to use each subset of these popular dimensionality reduction algorithms from the scikit-learn library.
The examples will provide the basis for you to copy-paste the examples and test the methods on your own data.
We will not dive into the theory behind how the algorithms work or compare them directly. For a good starting point on this topic, see:
Let’s dive in.
In this section, we will review how to use popular dimensionality reduction algorithms in scikit-learn.
This includes an example of using the dimensionality reduction technique as a data transform in a modeling pipeline and evaluating a model fit on the data.
The examples are designed for you to copy-paste into your own project and apply the methods to your own data. There are some algorithms available in the scikit-learn library that are not demonstrated because they cannot be used as a data transform directly given the nature of the algorithm.
As such, we will use a synthetic classification dataset in each example.
First, let’s install the library.
Don’t skip this step as you will need to ensure you have the latest version installed.
You can install the scikit-learn library using the pip Python installer, as follows:
1 | sudo pip install scikit-learn |
For additional installation instructions specific to your platform, see:
Next, let’s confirm that the library is installed and you are using a modern version.
Run the following script to print the library version number.
1 2 3 | # check scikit-learn version importsklearn print(sklearn.__version__) |
Running the example, you should see the following version number or higher.
1 | 0.23.0 |
We will use themake_classification() function to create a test binary classification dataset.
The dataset will have 1,000 examples with 20 input features, 10 of which are informative and 10 of which are redundant. This provides an opportunity for each technique to identify and remove redundant input features.
The fixed random seed for the pseudorandom number generator ensures we generate the same synthetic dataset each time the code runs.
An example of creating and summarizing the synthetic classification dataset is listed below.
1 2 3 4 5 6 | # synthetic classification dataset fromsklearn.datasetsimportmake_classification # define dataset X,y=make_classification(n_samples=1000,n_features=20,n_informative=10,n_redundant=10,random_state=7) # summarize the dataset print(X.shape,y.shape) |
Running the example creates the dataset and reports the number of rows and columns matching our expectations.
1 | (1000, 20) (1000,) |
It is a binary classification task and we will evaluate aLogisticRegression model after each dimensionality reduction transform.
The model will be evaluated using the gold standard ofrepeated stratified 10-fold cross-validation. The mean and standard deviation classification accuracy across all folds and repeats will be reported.
The example below evaluates the model on the raw dataset as a point of comparison.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 | # evaluate logistic regression model on raw data fromnumpyimportmean fromnumpyimportstd fromsklearn.datasetsimportmake_classification fromsklearn.model_selectionimportcross_val_score fromsklearn.model_selectionimportRepeatedStratifiedKFold fromsklearn.linear_modelimportLogisticRegression # define dataset X,y=make_classification(n_samples=1000,n_features=20,n_informative=10,n_redundant=10,random_state=7) # define the model model=LogisticRegression() # evaluate model cv=RepeatedStratifiedKFold(n_splits=10,n_repeats=3,random_state=1) n_scores=cross_val_score(model,X,y,scoring='accuracy',cv=cv,n_jobs=-1) # report performance print('Accuracy: %.3f (%.3f)'%(mean(n_scores),std(n_scores))) |
Running the example evaluates the logistic regression on the raw dataset with all 20 columns, achieving a classification accuracy of about 82.4 percent.
A successful dimensionality reduction transform on this data should result in a model that has better accuracy than this baseline, although this may not be possible with all techniques.
Note: we are not trying to “solve” this dataset, just provide working examples that you can use as a starting point.
1 | Accuracy: 0.824 (0.034) |
Next, we can start looking at examples of dimensionality reduction algorithms applied to this dataset.
I have made some minimal attempts to tune each method to the dataset. Each dimensionality reduction method will be configured to reduce the 20 input columns to 10 where possible.
We will use aPipeline to combine the data transform and model into an atomic unit that can be evaluated using the cross-validation procedure; for example:
1 2 3 4 | ... # define the pipeline steps=[('pca',PCA(n_components=10)),('m',LogisticRegression())] model=Pipeline(steps=steps) |
Let’s get started.
Can you get a better result for one of the algorithms?
Let me know in the comments below.
Principal Component Analysis, or PCA, might be the most popular technique for dimensionality reduction with dense data (few zero values).
For more on how PCA works, see the tutorial:
The scikit-learn library provides thePCA class implementation of Principal Component Analysis that can be used as a dimensionality reduction data transform. The “n_components” argument can be set to configure the number of desired dimensions in the output of the transform.
The complete example of evaluating a model with PCA dimensionality reduction is listed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 | # evaluate pca with logistic regression algorithm for classification fromnumpyimportmean fromnumpyimportstd fromsklearn.datasetsimportmake_classification fromsklearn.model_selectionimportcross_val_score fromsklearn.model_selectionimportRepeatedStratifiedKFold fromsklearn.pipelineimportPipeline fromsklearn.decompositionimportPCA fromsklearn.linear_modelimportLogisticRegression # define dataset X,y=make_classification(n_samples=1000,n_features=20,n_informative=10,n_redundant=10,random_state=7) # define the pipeline steps=[('pca',PCA(n_components=10)),('m',LogisticRegression())] model=Pipeline(steps=steps) # evaluate model cv=RepeatedStratifiedKFold(n_splits=10,n_repeats=3,random_state=1) n_scores=cross_val_score(model,X,y,scoring='accuracy',cv=cv,n_jobs=-1) # report performance print('Accuracy: %.3f (%.3f)'%(mean(n_scores),std(n_scores))) |
Running the example evaluates the modeling pipeline with dimensionality reduction and a logistic regression predictive model.
Note: Yourresults may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.
In this case, we don’t see any lift in model performance in using the PCA transform.
1 | Accuracy: 0.824 (0.034) |
Singular Value Decomposition, or SVD, is one of the most popular techniques for dimensionality reduction forsparse data (data with many zero values).
For more on how SVD works, see the tutorial:
The scikit-learn library provides theTruncatedSVD class implementation of Singular Value Decomposition that can be used as a dimensionality reduction data transform. The “n_components” argument can be set to configure the number of desired dimensions in the output of the transform.
The complete example of evaluating a model with SVD dimensionality reduction is listed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 | # evaluate svd with logistic regression algorithm for classification fromnumpyimportmean fromnumpyimportstd fromsklearn.datasetsimportmake_classification fromsklearn.model_selectionimportcross_val_score fromsklearn.model_selectionimportRepeatedStratifiedKFold fromsklearn.pipelineimportPipeline fromsklearn.decompositionimportTruncatedSVD fromsklearn.linear_modelimportLogisticRegression # define dataset X,y=make_classification(n_samples=1000,n_features=20,n_informative=10,n_redundant=10,random_state=7) # define the pipeline steps=[('svd',TruncatedSVD(n_components=10)),('m',LogisticRegression())] model=Pipeline(steps=steps) # evaluate model cv=RepeatedStratifiedKFold(n_splits=10,n_repeats=3,random_state=1) n_scores=cross_val_score(model,X,y,scoring='accuracy',cv=cv,n_jobs=-1) # report performance print('Accuracy: %.3f (%.3f)'%(mean(n_scores),std(n_scores))) |
Running the example evaluates the modeling pipeline with dimensionality reduction and a logistic regression predictive model.
Note: Yourresults may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.
In this case, we don’t see any lift in model performance in using the SVD transform.
1 | Accuracy: 0.824 (0.034) |
Linear Discriminant Analysis, or LDA, is a multi-class classification algorithm that can be used for dimensionality reduction.
The number of dimensions for the projection is limited to 1 and C-1, where C is the number of classes. In this case, our dataset is a binary classification problem (two classes), limiting the number of dimensions to 1.
For more on LDA for dimensionality reduction, see the tutorial:
The scikit-learn library provides theLinearDiscriminantAnalysis class implementation of Linear Discriminant Analysis that can be used as a dimensionality reduction data transform. The “n_components” argument can be set to configure the number of desired dimensions in the output of the transform.
The complete example of evaluating a model with LDA dimensionality reduction is listed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 | # evaluate lda with logistic regression algorithm for classification fromnumpyimportmean fromnumpyimportstd fromsklearn.datasetsimportmake_classification fromsklearn.model_selectionimportcross_val_score fromsklearn.model_selectionimportRepeatedStratifiedKFold fromsklearn.pipelineimportPipeline fromsklearn.discriminant_analysisimportLinearDiscriminantAnalysis fromsklearn.linear_modelimportLogisticRegression # define dataset X,y=make_classification(n_samples=1000,n_features=20,n_informative=10,n_redundant=10,random_state=7) # define the pipeline steps=[('lda',LinearDiscriminantAnalysis(n_components=1)),('m',LogisticRegression())] model=Pipeline(steps=steps) # evaluate model cv=RepeatedStratifiedKFold(n_splits=10,n_repeats=3,random_state=1) n_scores=cross_val_score(model,X,y,scoring='accuracy',cv=cv,n_jobs=-1) # report performance print('Accuracy: %.3f (%.3f)'%(mean(n_scores),std(n_scores))) |
Running the example evaluates the modeling pipeline with dimensionality reduction and a logistic regression predictive model.
Note: Yourresults may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.
In this case, we can see a slight lift in performance as compared to the baseline fit on the raw data.
1 | Accuracy: 0.825 (0.034) |
Isomap Embedding, or Isomap, creates an embedding of the dataset and attempts to preserve the relationships in the dataset.
The scikit-learn library provides theIsomap class implementation of Isomap Embedding that can be used as a dimensionality reduction data transform. The “n_components” argument can be set to configure the number of desired dimensions in the output of the transform.
The complete example of evaluating a model with SVD dimensionality reduction is listed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 | # evaluate isomap with logistic regression algorithm for classification fromnumpyimportmean fromnumpyimportstd fromsklearn.datasetsimportmake_classification fromsklearn.model_selectionimportcross_val_score fromsklearn.model_selectionimportRepeatedStratifiedKFold fromsklearn.pipelineimportPipeline fromsklearn.manifoldimportIsomap fromsklearn.linear_modelimportLogisticRegression # define dataset X,y=make_classification(n_samples=1000,n_features=20,n_informative=10,n_redundant=10,random_state=7) # define the pipeline steps=[('iso',Isomap(n_components=10)),('m',LogisticRegression())] model=Pipeline(steps=steps) # evaluate model cv=RepeatedStratifiedKFold(n_splits=10,n_repeats=3,random_state=1) n_scores=cross_val_score(model,X,y,scoring='accuracy',cv=cv,n_jobs=-1) # report performance print('Accuracy: %.3f (%.3f)'%(mean(n_scores),std(n_scores))) |
Running the example evaluates the modeling pipeline with dimensionality reduction and a logistic regression predictive model.
Note: Yourresults may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.
In this case, we can see a lift in performance with the Isomap data transform as compared to the baseline fit on the raw data.
1 | Accuracy: 0.888 (0.029) |
Locally Linear Embedding, or LLE, creates an embedding of the dataset and attempts to preserve the relationships between neighborhoods in the dataset.
The scikit-learn library provides theLocallyLinearEmbedding class implementation of Locally Linear Embedding that can be used as a dimensionality reduction data transform. The “n_components” argument can be set to configure the number of desired dimensions in the output of the transform
The complete example of evaluating a model with LLE dimensionality reduction is listed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 | # evaluate lle and logistic regression for classification fromnumpyimportmean fromnumpyimportstd fromsklearn.datasetsimportmake_classification fromsklearn.model_selectionimportcross_val_score fromsklearn.model_selectionimportRepeatedStratifiedKFold fromsklearn.pipelineimportPipeline fromsklearn.manifoldimportLocallyLinearEmbedding fromsklearn.linear_modelimportLogisticRegression # define dataset X,y=make_classification(n_samples=1000,n_features=20,n_informative=10,n_redundant=10,random_state=7) # define the pipeline steps=[('lle',LocallyLinearEmbedding(n_components=10)),('m',LogisticRegression())] model=Pipeline(steps=steps) # evaluate model cv=RepeatedStratifiedKFold(n_splits=10,n_repeats=3,random_state=1) n_scores=cross_val_score(model,X,y,scoring='accuracy',cv=cv,n_jobs=-1) # report performance print('Accuracy: %.3f (%.3f)'%(mean(n_scores),std(n_scores))) |
Running the example evaluates the modeling pipeline with dimensionality reduction and a logistic regression predictive model.
Note: Yourresults may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.
In this case, we can see a lift in performance with the LLE data transform as compared to the baseline fit on the raw data.
1 | Accuracy: 0.886 (0.028) |
Modified Locally Linear Embedding, or Modified LLE, is an extension of Locally Linear Embedding that creates multiple weighting vectors for each neighborhood.
The scikit-learn library provides theLocallyLinearEmbedding class implementation of Modified Locally Linear Embedding that can be used as a dimensionality reduction data transform. The “method” argument must be set to ‘modified’ and the “n_components” argument can be set to configure the number of desired dimensions in the output of the transform which must be less than the “n_neighbors” argument.
The complete example of evaluating a model with Modified LLE dimensionality reduction is listed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 | # evaluate modified lle and logistic regression for classification fromnumpyimportmean fromnumpyimportstd fromsklearn.datasetsimportmake_classification fromsklearn.model_selectionimportcross_val_score fromsklearn.model_selectionimportRepeatedStratifiedKFold fromsklearn.pipelineimportPipeline fromsklearn.manifoldimportLocallyLinearEmbedding fromsklearn.linear_modelimportLogisticRegression # define dataset X,y=make_classification(n_samples=1000,n_features=20,n_informative=10,n_redundant=10,random_state=7) # define the pipeline steps=[('lle',LocallyLinearEmbedding(n_components=5,method='modified',n_neighbors=10)),('m',LogisticRegression())] model=Pipeline(steps=steps) # evaluate model cv=RepeatedStratifiedKFold(n_splits=10,n_repeats=3,random_state=1) n_scores=cross_val_score(model,X,y,scoring='accuracy',cv=cv,n_jobs=-1) # report performance print('Accuracy: %.3f (%.3f)'%(mean(n_scores),std(n_scores))) |
Running the example evaluates the modeling pipeline with dimensionality reduction and a logistic regression predictive model.
Note: Yourresults may vary given the stochastic nature of the algorithm or evaluation procedure, or differences in numerical precision. Consider running the example a few times and compare the average outcome.
In this case, we can see a lift in performance with the modified LLE data transform as compared to the baseline fit on the raw data.
1 | Accuracy: 0.846 (0.036) |
This section provides more resources on the topic if you are looking to go deeper.
In this tutorial, you discovered how to fit and evaluate top dimensionality reduction algorithms in Python.
Specifically, you learned:
Do you have any questions?
Ask your questions in the comments below and I will do my best to answer.

...with just a few lines of python code
Discover how in my new Ebook:
Data Preparation for Machine Learning
It providesself-study tutorials withfull working code on:
Feature Selection,RFE,Data Cleaning,Data Transforms,Scaling,Dimensionality Reduction, and much more...
Some methods provide insight about the number of components/dimensions are useful, e.g. PCA.
Generally, treat it as a hyperparameter for your modeling pipeline.
It is not mentioned which inputs were dropped and which may be considered “strong”. In a real scenario, it is necessary know these informations to justify the dimensionality reduction.
I use results to guide me. E.g. use the number of dimensions that result in the best performing model.
Hello Jason. Can any of the manifold methods be used for time series data?
Perhaps, I am not up to speed on clustering for time series sorry.
Hi Jason,
Is there any thing wrong if I perform Linear Discriminant Analysis dimensionality reduction technique to reduce the number of dimensions before KMeans clustering.
I performed PCA + KMeans and LDA + KMeans. I got good results for LDA + KMeans. But I couldn’t find any work that uses LDA + KMeans. As I am working on a labelled dataset, I can do LDA technique, but in the real situations, there maynot be labelled dataset. So my question is, if I use LDA + KMeans does that make sense. I didnt find any works that use this technique for clustering.
Thankyou

I think that makes sense. Also, I can see some scholarly work on LDA+KMeans (https://journals.sagepub.com/doi/full/10.1155/2015/491910) too!
Jason am lost here ,with dimensionality reduction, do we used eg PCA +logistic regression to find the accuracy of the reduction before using another model to train and test the data?

PCA is a technique of dimensionality reduction. But any reduction will lost some information, only a matter of whether it is material. Applying logistic regression, for example, can help you find a way to measure whether the reduction can preserve the information you needed for a specific problem.
Hi all,
Is there any techniques related to PLA (Piecewise Linear Approximation) and which can be used for dimensionalilty reduction. I have the following problem: I have a dataset as a matrix consisting of time series and each row corresponds to a time series. I need an algorithm that uses PLA to reduce the number of columns.
Thanks and best regards
Med
Hi Med…The following discussion is a great resource for this topic:
https://stackoverflow.com/questions/29382903/how-to-apply-piecewise-linear-fit-in-python
Hello, I tried to use spectral embedding to see the cross val score, but it returned nan values. Can you help where is the mistake ? Thank you, Sarka
# Spectral embedding
from sklearn import metrics
from sklearn.datasets import load_digits
from sklearn.manifold import SpectralEmbedding
from sklearn.metrics import accuracy_score
# define dataset
X, y = make_classification(n_samples=1000, n_features=20, n_informative=10, n_redundant=10, random_state=7)
# define the pipeline
model = SpectralEmbedding(n_components=2)
X_transformed = model.fit_transform(X)
# evaluate model
cv = RepeatedStratifiedKFold(n_splits=10, n_repeats=3, random_state=1)
n_scores = cross_val_score(model, X, y, scoring=’accuracy’, cv=cv, n_jobs=-1)
# report performance
print(‘Accuracy: %.3f (%.3f)’ % (mean(n_scores), std(n_scores)))
I have a query, if I have categorical and numeric variables in my data set, what technique do you recommend I use or what should I do in those cases. Thank you.
Hi Bryam…The following discussion may be of interest to you.
Welcome!
I'mJason Brownlee PhD
and Ihelp developers get results withmachine learning.
Read more
TheData Preparation EBook is
where you'll find theReally Good stuff.