A Geometric Interpretation of The Covariance Matrix
A Geometric Interpretation of The Covariance Matrix
Go to...
Home » Math basics » Linear algebra » A geometric interpretation of the covariance matrix
1 Introduction
2 Eigendecomposition of a covariance matrix
3 Covariance matrix as a linear transformation
4 Conclusion
Introduction
In this article, we provide an intuitive, geometric interpretation of the covariance matrix, by exploring the
relation between linear transformations and the resulting data covariance. Most textbooks explain the shape
of data based on the concept of covariance matrices. Instead, we take a backwards approach and explain
the concept of covariance matrices based on the shape of data.
In a previous article, we discussed the concept of variance, and provided a derivation and proof of the well
known formula to estimate the sample variance. Figure 1 was used in this article to show that the standard
deviation, as the square root of the variance, provides a measure of how much the data is spread across the
feature space.
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 1/32
20/11/2019 A geometric interpretation of the covariance matrix
Figure 1. Gaussian density function. For normally distributed data, 68% of the samples fall
within the interval defined by the mean plus and minus the standard deviation.
We showed that an unbiased estimator of the sample variance can be obtained by:
(1)
However, variance can only be used to explain the spread of the data in the directions parallel to the axes of
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 2/32
20/11/2019 A geometric interpretation of the covariance matrix
covariance.
For this data, we could calculate the variance in the x-direction and the variance in
the y-direction. However, the horizontal spread and the vertical spread of the data does not explain the clear
diagonal correlation. Figure 2 clearly shows that on average, if the x-value of a data point increases, then
also the y-value increases, resulting in a positive correlation. This correlation can be captured by extending
(2)
(3)
If x is positively correlated with y, y is also positively correlated with x. In other words, we can state that
. Therefore, the covariance matrix is always a symmetric matrix with the variances
on its diagonal and the covariances off-diagonal. Two-dimensional normally distributed data is explained
completely by its mean and its covariance matrix. Similarly, a covariance matrix is used to
capture the spread of three-dimensional data, and a covariance matrix captures the spread of N-
dimensional data.
Figure 3 illustrates how the overall shape of the data defines the covariance matrix:
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 3/32
20/11/2019 A geometric interpretation of the covariance matrix
Figure 3. The covariance matrix defines the shape of the data. Diagonal spread is
transforms white data into the data we observed. However, before diving into the technical details, it is
important to gain an intuitive understanding of how eigenvectors and eigenvalues uniquely define the
As we saw in figure 3, the covariance matrix defines both the spread (variance), and the orientation
(covariance) of our data. So, if we would like to represent the covariance matrix with a vector and its
magnitude, we should simply try to find the vector that points into the direction of the largest spread of the
data, and whose magnitude equals the spread (variance) in this direction.
If we define this vector as , then the projection of our data onto this vector is obtained as , and
the variance of the projected data is . Since we are looking for the vector that points into the
direction of the largest variance, we should choose its components such that the covariance matrix
of the projected data is as large as possible. Maximizing any function of the form with respect to ,
where is a normalized unit vector, can be formulated as a so called Rayleigh Quotient. The maximum of
such a Rayleigh Quotient is obtained by setting equal to the largest eigenvector of matrix .
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 4/32
20/11/2019 A geometric interpretation of the covariance matrix
Check
In other words, the largest out my top-4
eigenvector of theof must-read
covariance matrix always
machine pointsbooks
learning into the direction of the
largest variance of the data, and the magnitude of this vector equals the corresponding eigenvalue. The
second largest eigenvector is always orthogonal to the largest eigenvector, and points into the direction of
Now let’s have a look at some examples. In an earlier article we saw that a linear transformation matrix
is completely defined by its eigenvectors and eigenvalues. Applied to the covariance matrix, this means
that:
(4)
If the covariance matrix of our data is a diagonal matrix, such that the covariances are zero, then this means
that the variances must be equal to the eigenvalues . This is illustrated by figure 4, where the eigenvectors
are shown in green and magenta, and where the eigenvalues clearly equal the variance components of the
covariance matrix.
However, if the covariance matrix is not diagonal, such that the covariances are not zero, then the situation
is a little more complicated. The eigenvalues still represent the variance magnitude in the direction of the
largest spread of the data, and the variance components of the covariance matrix still represent the
variance magnitude in the direction of the x-axis and y-axis. But since the data is not axis aligned, these
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 5/32
20/11/2019 A geometric interpretation of the covariance matrix
By comparing figure 5 with figure 4, it becomes clear that the eigenvalues represent the variance of the data
along the eigenvector directions, whereas the variance components of the covariance matrix represent the
spread along the axes. If there are no covariances, then both values are equal.
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 6/32
20/11/2019 A geometric interpretation of the covariance matrix
Check
Let the data shown by figure out my
6 be top-4
, then of must-read
each of the examples shownlearning
machine by figure 3 can be obtained by
books
linearly transforming :
(5)
(6)
(7)
(8)
where and are the scaling factors in the x direction and the y direction respectively.
In the following paragraphs, we will discuss the relation between the covariance matrix , and the linear
transformation matrix .
Let’s start with unscaled (scale equals 1) and unrotated data. In statistics this is often refered to as ‘white
data’ because its samples are drawn from a standard normal distribution and therefore correspond to white
(uncorrelated) noise:
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 7/32
20/11/2019 A geometric interpretation of the covariance matrix
The covariance matrix of this ‘white’ data equals the identity matrix, such that the variances and standard
(9)
(10)
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 8/32
20/11/2019 A geometric interpretation of the covariance matrix
(11)
Thus, the covariance matrix of the resulting data is related to the linear transformation that is
(12)
However, although equation (12) holds when the data is scaled in the x and y direction, the question rises if
it also holds when a rotation is applied. To investigate the relation between the linear transformation matrix
and the covariance matrix in the general case, we will therefore try to decompose the covariance
As we saw earlier, we can represent the covariance matrix by its eigenvectors and eigenvalues:
(13)
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 9/32
20/11/2019 A geometric interpretation of the covariance matrix
eigenvectors and two eigenvalues. The system of two equations defined by equation (13) can be represented
(14)
where is the matrix whose columns are the eigenvectors of and is the diagonal matrix whose non-
This means that we can represent the covariance matrix as a function of its eigenvectors and eigenvalues:
(15)
Equation (15) is called the eigendecomposition of the covariance matrix and can be obtained using a
Singular Value Decomposition algorithm. Whereas the eigenvectors represent the directions of the largest
variance of the data, the eigenvalues represent the magnitude of this variance in those directions. In other
words, represents a rotation matrix, while represents a scaling matrix. The covariance matrix can
(16)
(17)
In other words, if we apply the linear transformation defined by to the original white data
shown by figure 7, we obtain the rotated and scaled data with covariance matrix
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 10/32
20/11/2019 A geometric interpretation of the covariance matrix
Figure 10. The covariance matrix represents a linear transformation of the original data.
The colored arrows in figure 10 represent the eigenvectors. The largest eigenvector, i.e. the eigenvector with
the largest corresponding eigenvalue, always points in the direction of the largest variance of the data and
thereby defines its orientation. Subsequent eigenvectors are always orthogonal to the largest eigenvector
due to the orthogonality of rotation matrices.
Conclusion
In this article we showed that the covariance matrix of observed data is directly related to a linear
transformation of white, uncorrelated data. This linear transformation is completely defined by the
eigenvectors and eigenvalues of the data. While the eigenvectors represent the rotation matrix, the
eigenvalues correspond to the square of the scaling factor in each dimension.
Summary
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 11/32
20/11/2019 A geometric interpretation of the covariance matrix
Share
April 24, 2014 Vincent Spruyt Linear algebra 47 Comments covariance matrix,
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 12/32
20/11/2019 A geometric interpretation of the covariance matrix
Chris says:
May 14, 2014 at 3:42 pm
Reply
Alex says:
May 14, 2014 at 8:09 pm
V^T$.
From computational point of view it is much simpler to
find $V^T$ than $V^{-1}$.
Reply
Reply
Brian says:
May 24, 2014 at 1:03 am
Reply
Hi Brian:
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 14/32
20/11/2019 A geometric interpretation of the covariance matrix
(D^t*R)^t
Reply
Konstantin says:
August 18, 2014 at 1:47 am
Thank you for this great post! But let me please correct
one fundamental mistake that you made. The square
wrong).
Reply
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 15/32
20/11/2019 A geometric interpretation of the covariance matrix
Reply
Reply
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 16/32
20/11/2019 A geometric interpretation of the covariance matrix
Reply
Reply
Paul says:
April 25, 2015 at 5:40 pm
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 17/32
20/11/2019 A geometric interpretation of the covariance matrix
Paul says:
April 25, 2015 at 5:53 pm
Reply
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 18/32
20/11/2019 A geometric interpretation of the covariance matrix
Reply
harmyder says:
January 17, 2017 at 9:30 am
Reply
Lukas says:
June 2, 2015 at 5:09 pm
Hello Vincent,
Thank you very much for this blog post. I have one
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 19/32
20/11/2019 A geometric interpretation of the covariance matrix
Lukas
Reply
Gordon
Reply
intuitive topics!!!
Reply
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 20/32
20/11/2019 A geometric interpretation of the covariance matrix
Reply
seravee says:
June 17, 2015 at 9:09 am
Reply
Pradeep says:
June 21, 2015 at 10:31 am
Reply
haining says:
September 11, 2015 at 6:36 pm
Reply
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 21/32
20/11/2019 A geometric interpretation of the covariance matrix
Priyamvad says:
September 13, 2015 at 11:52 pm
Reply
Joe says:
October 30, 2015 at 2:21 pm
Reply
Superb!
Reply
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 22/32
20/11/2019 A geometric interpretation of the covariance matrix
dk sunil says:
January 18, 2016 at 9:59 pm
Reply
Gennaro says:
January 22, 2016 at 11:53 am
me to be as meticulous as possible
Thank you in advance!
Reply
simon says:
March 10, 2017 at 6:11 am
Reply
Nariman says:
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 23/32
20/11/2019 A geometric interpretation of the covariance matrix
Reply
beena says:
March 20, 2016 at 6:22 pm
Reply
Marcell says:
March 29, 2016 at 6:56 pm
Reply
Treble says:
April 15, 2016 at 2:16 am
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 24/32
20/11/2019 A geometric interpretation of the covariance matrix
I found.
Reply
Artur RC says:
April 18, 2016 at 6:23 pm
Reply
Reply
ninjajack says:
April 28, 2016 at 7:20 am
Reply
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 25/32
20/11/2019 A geometric interpretation of the covariance matrix
with zero…
Where I am going geometrically worng.
Second
Why don’t we a complex eigen vector conjugate
when we rotate the white data by rotational
matrix….
Reply
Thomas says:
July 13, 2016 at 10:39 am
Summer says:
July 28, 2016 at 7:51 pm
Reply
0]
if I missunderstand something im sorry
Reply
Reply
J. tipton says:
October 1, 2016 at 6:28 pm
me question. –ekwaysan
Reply
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 28/32
20/11/2019 A geometric interpretation of the covariance matrix
Thanks in advance!
Reply
humansofportsmouth says:
November 13, 2016 at 4:29 pm
Reply
Dave says:
November 25, 2016 at 8:42 pm
Reply
renzocoppola says:
April 2, 2017 at 8:09 am
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 29/32
20/11/2019 A geometric interpretation of the covariance matrix
Reply
Kevin says:
April 24, 2017 at 12:27 am
Reply
Kamesh says:
June 8, 2017 at 10:32 pm
Reply
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 30/32
20/11/2019 A geometric interpretation of the covariance matrix
NA says:
June 12, 2017 at 4:04 am
Reply
Follow On Twitter
Article Topics
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 31/32
20/11/2019 A geometric interpretation of the covariance matrix
Dimensionality reduction
Check (2)out my top-4 of must-read machine learning books
Statistics (2)
Other (2)
Recent Posts
Hybrid deep learning for modeling driving behavior from sensor data
Rss Feed
RSS - Posts
RSS - Comments
https://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix/ 32/32