0% found this document useful (0 votes)
2 views

Linear_algebra_and_some_of_it_application_to_machine_learning__1_ (2)

Uploaded by

albertjessica919
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
2 views

Linear_algebra_and_some_of_it_application_to_machine_learning__1_ (2)

Uploaded by

albertjessica919
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 17

Contents

1 CHAPTER 1 4
1.1 Background of the Study . . . . . . . . . . . . . . . . . . . . . . 4
1.2 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.3 Problem Statement . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.3.1 1: Solving a System of Linear Equations . . . . . . . . . . 5
1.3.2 2: Eigenvalue Decomposition . . . . . . . . . . . . . . . . 5
1.3.3 3: Singular Value Decomposition (SVD) . . . . . . . . . 5
1.3.4 4: Least Squares Method . . . . . . . . . . . . . . . . . . 5
1.4 Objectives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.5 Definition of Terms . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5.1 Linear Algebra: . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5.2 Vector: . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5.3 Matrix: . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5.4 Eigenvalue: . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5.5 Eigenvector: . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5.6 SVD: . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5.7 PCA: . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.5.8 Least Squares Method: . . . . . . . . . . . . . . . . . . . . 6

2 CHAPTER 2 7
2.1 Literature Review . . . . . . . . . . . . . . . . . . . . . . . . . . 7
2.1.1 Vectors and Matrices in Machine Learning . . . . . . . . . 7
2.1.2 Eigenvalues and Eigenvectors . . . . . . . . . . . . . . . . 7
2.1.3 Singular Value Decomposition (SVD) . . . . . . . . . . . 7
2.1.4 Principal Component Analysis (PCA) . . . . . . . . . . . 8
2.1.5 Least Squares and Linear Regression . . . . . . . . . . . . 8
2.1.6 Applications in Machine Learning . . . . . . . . . . . . . 8

3 CHAPTER 3 9
3.1 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.1.1 Theoretical Framework . . . . . . . . . . . . . . . . . . . 9
3.1.2 Practical Implementation . . . . . . . . . . . . . . . . . . 9
3.1.3 Performance Evaluation . . . . . . . . . . . . . . . . . . . 10
3.1.4 Tools and Environment . . . . . . . . . . . . . . . . . . . 10

4 CHAPTER 4 12
4.1 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
4.2 Challenges and Insights . . . . . . . . . . . . . . . . . . . . . . . 12
4.3 Evaluating Model Performance . . . . . . . . . . . . . . . . . . . 13
4.4 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
4.4.1 Solving a System of Linear Equations . . . . . . . . . . . 13
4.4.2 Eigenvalue Decomposition . . . . . . . . . . . . . . . . . 14
4.4.3 Singular Value Decomposition (SVD) . . . . . . . . . . . 15
4.4.4 Least Squares Method . . . . . . . . . . . . . . . . . . . 15

1
5 CHAPTER 5 16
5.1 Conclusion and Recommendation . . . . . . . . . . . . . . . . . . 16
5.2 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
5.3 Recommendations . . . . . . . . . . . . . . . . . . . . . . . . . . 16

2
Linear algebra and some of it application to
machine learning
estherosadare
August 2024

[12pt,a4paper]article graphicx amsmath

3
1 CHAPTER 1
1.1 Background of the Study
Linear algebra is a fundamental area of mathematics that plays a critical role in
understanding and implementing various algorithms in machine learning. It in-
volves the study of vectors, vector spaces, linear transformations, and matrices.
These concepts are essential not only for theoretical insights but also for prac-
tical applications in numerous scientific fields, particularly in computer science
and machine learning. In the context of machine learning, linear algebra forms
the foundation for algorithms related to regression, classification, and dimen-
sionality reduction.Linear algebra is a foundational discipline in mathematics,
focusing on vector spaces and linear mappings between these spaces. It provides
the necessary tools for handling and manipulating high-dimensional data, which
is central to various fields of science and engineering. In the realm of machine
learning, linear algebra serves as the backbone for many algorithms, especially
those involving large datasets and complex computations.
Matrix operations, vector spaces, eigenvalues, and singular value decompo-
sition (SVD) are just a few concepts from linear algebra that are integral to
understanding and implementing machine learning models. For instance, linear
regression, one of the simplest machine learning algorithms, relies heavily on
solving systems of linear equations—a direct application of linear algebra. Sim-
ilarly, algorithms like Principal Component Analysis (PCA) and Support Vector
Machines (SVM) depend on matrix factorization techniques and vector space
transformations to reduce dimensionality and classify data points, respectively.
As the data available to us grows in volume and complexity, the need for
efficient algorithms to process and extract meaningful insights from this data
becomes more pressing. Linear algebra provides the mathematical framework
that underpins these algorithms, making it an essential area of study for anyone
looking to delve into machine learning.
Understanding linear algebra allows data scientists and engineers to design,
implement, and optimize machine learning algorithms with greater precision and
efficiency. As such, it is crucial to explore how the principles of linear algebra can
be leveraged to enhance the capabilities and performance of machine learning
models.

1.2 Motivation
My motivation for studying linear algebra, especially in the context of machine
learning, stems from the necessity to efficiently manage and process large vol-
umes of data.

1.3 Problem Statement


Despite the widespread adoption of machine learning, there is often a gap in un-
derstanding the underlying mathematical principles that drive these algorithms.

4
This study aims to explore the intersection of linear algebra and machine learn-
ing by providing solvable examples that illustrate how fundamental concepts in
linear algebra can be effectively utilized to enhance machine learning applica-
tions.

1.3.1 1: Solving a System of Linear Equations


Consider a machine learning problem where we need to solve a system of linear
equations to determine the optimal weights in a linear regression model:

Ax = b

Where A is a matrix of input features, x is the vector of weights, and b is


the output vector. Let A = ( 1 ) 2
34 and b = ( 5 )
11. Solve for x.

1.3.2 2: Eigenvalue Decomposition


In Principal Component Analysis (PCA), we often need to compute the eigen-
values and eigenvectors of the covariance matrix of the data. Given a covariance
matrix C = ( 2 ) 1
12, find its eigenvalues and eigenvectors.

1.3.3 3: Singular Value Decomposition (SVD)


Singular Value Decomposition is used in recommender systems to reduce the
dimensionality of the user-item interaction matrix. Given a matrix M = ( 1 ) 00
010
001, perform SVD.

1.3.4 4: Least Squares Method


In linear regression, the least squares method is used to find the best-fitting line
to a set of data points by minimizing the sum of the squares of the residuals.
Consider the data points (1, 2), (2, 3), and (3, 5). Find the best-fit line y =
mx + c.

1.4 Objectives
The primary objectives of this study are:
1. To demonstrate the fundamental concepts of linear algebra and how they
underpin many machine learning algorithms.
2. To solve practical problems in machine learning using linear algebra tech-
niques such as matrix operations, eigenvalue decomposition, and singular
value decomposition.

5
3. To illustrate how linear algebra can be applied to improve data prepro-
cessing, model optimization, and feature extraction in machine learning
applications.

1.5 Definition of Terms


1.5.1 Linear Algebra:
A branch of mathematics concerned with vector spaces and linear mappings
between such spaces, involving matrices and vectors.

1.5.2 Vector:
A data point in high-dimensional space, often representing features in machine
learning.

1.5.3 Matrix:
A rectangular array of numbers used to represent linear transformations.

1.5.4 Eigenvalue:
Scalar showing how a corresponding eigenvector is scaled during a transforma-
tion.

1.5.5 Eigenvector:
Non-zero vector that changes only by a scalar factor during a transformation.

1.5.6 SVD:
Factorization of a matrix into singular values, often used in dimensionality re-
duction.

1.5.7 PCA:
A technique to reduce the dimensionality of data while retaining most of its
variance.

1.5.8 Least Squares Method:


Minimizes the sum of the squared residuals between observed and predicted
values.

6
cite

2 CHAPTER 2
2.1 Literature Review
The application of linear algebra in machine learning has been extensively ex-
plored in various studies, given its fundamental role in understanding and imple-
menting algorithms. This section reviews key literature that highlights the use
of linear algebraic techniques in machine learning, focusing on concepts such
as vectors, matrices, eigenvalues, singular value decomposition, and principal
component analysis.

2.1.1 Vectors and Matrices in Machine Learning


Vectors and matrices are the basic building blocks in linear algebra and are
widely used in machine learning. According to [7], vectors are utilized to rep-
resent data points in high-dimensional spaces, which is essential for tasks such
as classification, clustering, and regression. Matrices, on the other hand, are
used to perform linear transformations, which are fundamental in operations
like rotation, scaling, and projection of data. The manipulation of matrices is
particularly important in neural networks, where the weights and inputs are
represented as matrices, and matrix multiplication is used to compute the out-
puts.

2.1.2 Eigenvalues and Eigenvectors


Eigenvalues and eigenvectors play a crucial role in understanding the behav-
ior of linear transformations. [8] emphasizes the importance of eigenvalues and
eigenvectors in Principal Component Analysis (PCA), where they are used to
identify the directions of maximum variance in the data. PCA is a dimensional-
ity reduction technique that transforms the data into a new coordinate system,
where the first coordinate (principal component) has the largest variance, and
each succeeding component has the highest variance possible under the con-
straint that it is orthogonal to the preceding components. The eigenvalues give
the magnitude of the variance in the direction of the corresponding eigenvectors,
which represent the principal components.

2.1.3 Singular Value Decomposition (SVD)


Singular Value Decomposition (SVD) is a powerful matrix factorization tech-
nique widely used in machine learning, particularly in dimensionality reduction,
noise reduction, and collaborative filtering. [9] discusses how SVD decomposes
a matrix M into three matrices U , Σ, and V T , where U and V are orthogo-
nal matrices, and Σ is a diagonal matrix containing the singular values. The
application of SVD in machine learning is particularly evident in recommender

7
systems, where it helps in reducing the dimensionality of the user-item matrix,
thus enabling efficient predictions and recommendations.

2.1.4 Principal Component Analysis (PCA)


Principal Component Analysis (PCA) is another fundamental technique that
relies on linear algebra. As discussed by [10], PCA is used for dimensionality re-
duction in high-dimensional datasets, which is crucial for visualization, noise re-
duction, and improving model performance by removing multicollinearity. PCA
works by computing the eigenvalues and eigenvectors of the covariance matrix
of the data, and then projecting the data onto the eigenvectors corresponding
to the largest eigenvalues. This results in a lower-dimensional representation of
the data that retains most of the variance.

2.1.5 Least Squares and Linear Regression


The least squares method is a standard approach for solving overdetermined
systems of linear equations, which is the basis for linear regression in machine
learning. According to [11], the least squares method minimizes the sum of the
squares of the residuals, which are the differences between the observed and
predicted values. This method is essential for finding the best-fitting line in
regression analysis, and it is also used in more complex models such as ridge
regression and lasso regression, where regularization terms are added to prevent
overfitting.

2.1.6 Applications in Machine Learning


Linear algebra is not only theoretical but also practical in various machine
learning applications. For instance, [12] explores the use of kernel methods,
which involve nonlinear transformations of data, in Support Vector Machines
(SVMs). These methods extend the concept of linear separation by mapping
data to a higher-dimensional space, where it becomes linearly separable. This
is made possible by the kernel trick, which implicitly computes the dot product
in the transformed space without explicitly mapping the data.
In the context of deep learning, [13] discusses the use of t-Distributed Stochas-
tic Neighbor Embedding (t-SNE), a technique that builds on PCA for visualizing
high-dimensional data by reducing its dimensions to two or three while preserv-
ing the structure. t-SNE uses a probabilistic approach to map the similarities
between data points in the high-dimensional space to the low-dimensional space,
making it easier to visualize complex patterns in the data.
Finally, [14] introduces Sparse Principal Component Analysis (Sparse PCA),
which extends the traditional PCA by enforcing sparsity on the principal com-
ponents. Sparse PCA is particularly useful in situations where the data is high-
dimensional but sparse, as it helps in identifying the most relevant features while
ignoring the rest, leading to more interpretable models.

8
3 CHAPTER 3
3.1 Methodology
This section outlines the methodology used to explore and demonstrate the
applications of linear algebra in machine learning. The approach is divided
into three main parts: theoretical framework, practical implementation, and
performance evaluation. Each part builds upon the previous to provide a com-
prehensive understanding of how linear algebra techniques can be utilized in
machine learning tasks.

3.1.1 Theoretical Framework


The first step involves a detailed exploration of the mathematical foundations
of linear algebra relevant to machine learning. This includes:

1. Matrix Operations: Understanding matrix multiplication, inversion,


and decomposition techniques such as LU decomposition, QR decomposi-
tion, and Cholesky decomposition. These operations are fundamental in
various machine learning algorithms.
2. Vector Spaces and Transformations: Analyzing how vector spaces
and linear transformations apply to machine learning. This involves study-
ing concepts like basis vectors, orthogonality, and projections, which are
crucial for understanding data transformations and feature extraction.
3. Eigenvalues and Eigenvectors: Investigating how eigenvalues and eigen-
vectors are derived and their significance in techniques like PCA and spec-
tral clustering. The focus is on understanding the role of these concepts in
reducing dimensionality and capturing the essence of the data’s structure.
4. Singular Value Decomposition (SVD): Examining the SVD tech-
nique and its applications in machine learning, particularly in recom-
mender systems and data compression. The mathematical derivation of
SVD is explored along with its interpretation in the context of reducing
the rank of matrices.

3.1.2 Practical Implementation


After establishing the theoretical groundwork, the next step is to implement
the discussed linear algebra techniques in practical machine learning scenarios.
This involves:

1. Data Preprocessing: Utilizing linear algebra for data normalization,


feature scaling, and dimensionality reduction. Techniques like PCA are
applied to real-world datasets to demonstrate how they reduce the dimen-
sionality of data while retaining significant information.

9
2. Model Training: Implementing linear regression, support vector ma-
chines, and other machine learning models that heavily rely on linear
algebra. For instance, the normal equation in linear regression is solved
using matrix operations, and the SVM decision boundary is computed
using inner products.

3. Dimensionality Reduction: Applying PCA and SVD on datasets to re-


duce their dimensionality and compare the performance of models trained
on reduced data versus the original data. The focus is on understanding
the trade-offs between model complexity and computational efficiency.
4. Matrix Factorization in Recommender Systems: Using SVD to
decompose user-item interaction matrices and predict missing values. This
technique is implemented in a collaborative filtering recommender system,
demonstrating how linear algebra can be used to improve recommendation
accuracy.

3.1.3 Performance Evaluation


The final part of the methodology involves evaluating the performance of the
machine learning models and techniques that utilize linear algebra. This in-
cludes:

1. Computational Efficiency: Assessing the computational time and re-


sources required to implement linear algebra techniques in machine learn-
ing models. Comparisons are made between different approaches to iden-
tify the most efficient methods.
2. Model Accuracy: Evaluating the accuracy of machine learning models
trained with and without the application of linear algebra techniques. This
includes comparing the results of models using dimensionality reduction
methods such as PCA with those using all features.
3. Scalability: Analyzing how well the implemented techniques scale with
increasing data size and complexity. The methodology includes exper-
iments with varying dataset sizes to determine the scalability of linear
algebra-based methods.
4. Robustness: Testing the robustness of the machine learning models
against noisy and incomplete data. The effectiveness of linear algebra
techniques like SVD in handling such scenarios is particularly emphasized.

3.1.4 Tools and Environment


1. Programming Languages: The practical implementation is carried out
using Python, a widely-used programming language in data science and
machine learning.

10
2. Libraries: Key libraries such as NumPy and SciPy for numerical compu-
tations, scikit-learn for machine learning algorithms, and matplotlib for
data visualization are utilized.
3. Datasets: Publicly available datasets such as the UCI Machine Learning
Repository and Kaggle datasets are used for experiments.

11
4 CHAPTER 4
4.1 Discussion
The application of linear algebra in machine learning provides significant ad-
vantages in data manipulation, model training, and dimensionality reduction.
This discussion elaborates on the practical implications of using linear algebra
techniques, the challenges encountered, and the insights gained from solving the
provided examples.
*Application and Implications
1. Matrix Operations and Model Training: Matrix operations, such
as multiplication and inversion, are essential in many machine learning
algorithms. For example, in linear regression, matrix inversion allows for
the efficient computation of model parameters using the normal equation.
This operation is crucial for determining the coefficients that minimize the
residual sum of squares.
2. Principal Component Analysis (PCA): PCA is a powerful technique
for dimensionality reduction, helping to visualize high-dimensional data
and reduce computational complexity. By projecting data onto principal
components that capture the most variance, PCA simplifies models and
improves performance.
3. Singular Value Decomposition (SVD): SVD’s role in recommender
systems and data compression illustrates its practical utility. By decom-
posing matrices into singular values and vectors, SVD helps identify latent
features and improve recommendation accuracy.
4. Eigenvalue Decomposition: Eigenvalue decomposition provides insights
into the principal directions of data variance, which are used in techniques
like PCA and spectral clustering.

4.2 Challenges and Insights


1. Computational Complexity: One challenge in applying linear alge-
bra techniques is the computational complexity associated with matrix
operations, especially with large datasets.
2. Numerical Stability: Numerical stability is another concern, particu-
larly when dealing with ill-conditioned matrices. Techniques like regular-
ization and matrix decomposition help mitigate issues related to numerical
instability.
3. Dimensionality Reduction Trade-offs: While dimensionality reduc-
tion techniques like PCA are valuable, they involve trade-offs between
data representation and computational efficiency. Reducing dimensions
may lead to loss of information, which needs to be carefully managed to
maintain model performance.

12
4. Robustness to Noisy Data: Linear algebra techniques, such as SVD,
can be sensitive to noisy or incomplete data. Incorporating regularization
techniques and robust optimization methods helps improve the resilience
of models against such challenges.

article amsmath

4.3 Evaluating Model Performance


Evaluating model performance with various levels of noise provides insights into
the robustness and reliability of the techniques used.

4.4 Results
4.4.1 Solving a System of Linear Equations
Given the system of equations:

Ax = b

where:

A = ( 1 ) 234 and b = ( 5 ) 11

To solve for x, we find the inverse of matrix A and multiply it by b:

x = A−1 b

First, compute the inverse of A:


1
A−1 = adj(A)
det(A)
where:

det(A) = (1 · 4) − (2 · 3) = 4 − 6 = −2

adj(A) = ( 4 ) − 2 − 31

Thus:
1
A−1 = ( 4 ) − 2 − 31 = ( − ) 211.5 − 0.5
−2
Multiply A−1 by b:

x = ( − ) 211.5−0.5 ( 5 ) 11 = ( − ) 2·5+1·111.5·5−0.5·11 = ( − ) 10+117.5−5.5 = ( 1 ) 2

The solution is x = ( 1 )
2.

13
4.4.2 Eigenvalue Decomposition
Given the covariance matrix:

C = ( 2 ) 112

To find eigenvalues λ and eigenvectors v, solve the characteristic equation:

det(C − λI) = 0

where:

C − λI = ( 2 ) − λ112 − λ

det(C − λI) = (2 − λ)(2 − λ) − 1 · 1 = λ2 − 4λ + 3

Setting this equal to zero:

λ2 − 4λ + 3 = 0

(λ − 3)(λ − 1) = 0

So, the eigenvalues are λ1 = 3 and λ2 = 1.


To find the eigenvectors:
For λ1 = 3:

C − 3I = ( − ) 111 − 1

Solving (C − 3I)v = 0:

( − ) 111 − 1 ( x ) y = ( 0 ) 0

The eigenvector is v1 = ( 1 )
1.
For λ2 = 1:

C − I = ( 1 ) 111

Solving (C − I)v = 0:

( 1 ) 111 ( x ) y = ( 0 ) 0

The eigenvector is v2 = ( − ) 1
1.

14
4.4.3 Singular Value Decomposition (SVD)
Given the matrix:

M = ( 1 ) 00010001

Perform SVD to decompose M :

M = U ΣV T

where U and V are orthogonal matrices, and Σ is a diagonal matrix with


singular values.
Since M is an identity matrix, U , Σ, and V are:

U = ( 1 ) 00010001

Σ = ( 1 ) 00010001

V T = ( 1 ) 00010001

Thus, M = U ΣV T with U , Σ, and V T as identity matrices.

4.4.4 Least Squares Method


Given data points (1, 2), (2, 3), and (3, 5), find the best-fit line y = mx + c using
the least squares method.
1. Set up matrices X and y:

X = ( 1 ) 12131 and y = ( 2 ) 35

2. Compute (X T X)−1 X T y:

X T X = ( 1 ) 4663

1
(X T X)−1 = ( 3 ) − 6 − 614
24

1
(X T X)−1 X T y = ( 3 ) − 6 − 614 ( 1 ) 46 = ( 1 ) 1
24
The coefficients are m = 1 and c = 1, giving the best-fit line y = x + 1.

15
5 CHAPTER 5
5.1 Conclusion and Recommendation
5.2 Conclusion
In this study, we explored the fundamental principles of linear algebra and
their critical applications in machine learning. Through various examples, we
demonstrated how linear algebra techniques such as matrix inversion, eigenvalue
decomposition, and singular value decomposition (SVD) can be applied to solve
complex problems in data analysis and model training. These techniques not
only provide a mathematical foundation for understanding machine learning
algorithms but also enhance their computational efficiency and accuracy. The
successful application of these methods in solving systems of linear equations,
reducing dimensionality, and fitting models to data underscores the importance
of linear algebra in modern machine learning practices. The integration of
these concepts into machine learning workflows can significantly improve the
performance and interpretability of models, leading to more robust and reliable
outcomes.

5.3 Recommendations
To further advance the application of linear algebra in machine learning, it is
recommended that practitioners focus on optimizing computational efficiency,
particularly when dealing with large-scale data. Incorporating advanced lin-
ear algebra techniques, such as regularized matrix decompositions and iterative
methods for matrix inversion, can help address challenges related to compu-
tational complexity and numerical stability. Additionally, it is advisable to
continue exploring the balance between dimensionality reduction and data rep-
resentation, ensuring that important information is preserved while reducing
computational costs. Finally, integrating linear algebra more deeply into ma-
chine learning education will equip future practitioners with the necessary tools
to develop more sophisticated and efficient algorithms, ultimately pushing the
boundaries of what is possible in data-driven decision-making.

16
References
[1] Strang, G. (2016). Introduction to Linear Algebra. 5th ed. Wellesley-
Cambridge Press.
[2] Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep Learning. MIT
Press.
[3] Golub, G. H., & Van Loan, C. F. (2012). Matrix Computations. 4th ed.
Johns Hopkins University Press.
[4] Horn, R. A., & Johnson, C. R. (2012). Matrix Analysis. Cambridge Uni-
versity Press.
[5] Seber, G. A. F., & Lee, A. J. (2003). Linear Regression Analysis. Wiley-
Interscience.
[6] Draper, N. R., & Smith, H. (1998). Applied Regression Analysis. Wiley.
[7] Bishop, C. M. (2006). Pattern Recognition and Machine Learning. Springer.
[8] Jolliffe, I. T. (2016). Principal Component Analysis. 2nd ed. Springer.

[9] McLachlan, G. J., & Peel, D. (2002). Finite Mixture Models. Wiley.
[10] Price, B., & Kershaw, R. (2006). Principal Component Analysis. Wiley.
[11] Ross, S. M. (2008). Incremental Linear Regression. Springer.

[12] Schölkopf, B., & Smola, A. J. (1998). Nonlinear Component Analysis as a


Kernel Eigenvalue Problem. Neural Computation.
[13] van der Maaten, L., & Hinton, G. (2008). Visualizing Data using t-SNE.
Journal of Machine Learning Research.
[14] Zou, H., & Hastie, T. (2006). Sparse Principal Component Analysis. Jour-
nal of Computational and Graphical Statistics.

17

You might also like