August 5, 2023
August 5, 2023
August 5, 2023
The aim of this project is to explore and analyze matrix iterative meth-
ods for computing eigenvalues and eigenvectors. Eigenvalue problems play a
fundamental role in various mathematical and scientific disciplines, includ-
ing linear algebra, physics, and computer science. Matrix iterative methods
provide efficient and practical approaches for approximating eigenvalues and
eigenvectors, especially for large and sparse matrices. This project will in-
vestigate popular iterative methods such as the power method, inverse power
method, and the Rayleigh Quotient Iteration method, and compare their per-
formance in terms of accuracy for matrices with different sizes and properties,
convergence rate, and analyzing their computational complexity.
1 Introduction
1.1 Background
The term ”eigenvalue” comes from the German word ”eigen,” meaning ”own”
or ”characteristic.” The development of the theory dates back to the early
19th century, with contributions from mathematicians such as Augustin-
Louis Cauchy, William Rowan Hamilton, and Arthur Cayley. The system-
atic study of eigenvalues and eigenvectors began in the mid-19th century,
pioneered by the German mathematicians Carl Friedrich Gauss and Philipp
Ludwig von Seidel. Eigenvalue and eigenvector computations play a fun-
damental role in various areas of mathematics, physics, and engineering.
They are essential for understanding the behavior of linear systems, analyz-
ing structural vibrations, solving differential equations, and numerous other
applications. The eigenvalues represent the scalar values associated with a
matrix, while the eigenvectors are the corresponding non-zero vectors that
1
remain in the same direction after applying the matrix transformation. Solv-
ing eigenvalue problems analytically is often challenging or even impossible,
especially for large and complex matrices. As a result, numerical methods
are employed to approximate eigenvalues and eigenvectors. Matrix iterative
methods have emerged as powerful techniques for computing these approxi-
mations, particularly in situations where the matrix is large or sparse.
1.3 Objectives
The objectives of this project are as follows:
2
ues and eigenvectors.
3
multiplied by its corresponding eigenvector x, the result is a scaled version
of x: Ax = λ ×x
4
indication of the efficiency and scalability of a method for solving a problem.
2 Literature Review
2.1 Review of Eigenvalues and Eigenvectors
Eigenvalues are often introduced in the context of linear algebra or matrix
theory and numerical analysis. Historically, however, they arose in the study
of quadratic forms and differential equations.In the 18th century, Euler stud-
ied the rotational motion of a rigid body and discovered the importance of
the principal axes. Lagrange realized that the principal axes are the eigenvec-
tors of the inertia matrix.(Hawkins,1975). In the early 19th century,Cauchy
saw how their work could be used to classify the quadratic surfaces, and
generalized it to arbitrary dimension.( Hawkins,1975)Cauchy also coined the
term racine caracteristique (characteristic root) for what is now called eigen-
value; his term survives in characteristic equation. Fourier used the work
of Laplace and Lagrange to solve the heat equation by separation of vari-
5
ables in his famous book ” Theorie analytique de la Chaleur” published in
1822. (Kline , 1972) . Sturm development Fourier ideas further and brought
them to the attention of Cauchy, who combined them with his own ideas
and arrived at the fact that real symmetric matrices have real eigenvalues.
(Hawkins,1975), This was extended by Hermit in 1855 to what are now called
Hermitian matrices(Kline,1972). Around the same time, Brioschi proved that
the eigenvalues of orthogonal matrices lie on the unit circle,(Hawkins,1975)
and Clebseh found the corresponding result for skew - symmetric matrices.
(Kline, 1972) talked on finding eigenvalues and eigenvectors.
Finally, weierstrass clarified an important aspect in the stability theory of
eigenvalue problems by Laplace by realizing that defective matrices can cause
instability of eigenvalues(Hawkins,1975). In the mean time, Liouville studied
eigenvalue problems similar to those of Sturm; the discipline that grew out of
their work is now Sturm-Liouville theory.(kline,1972). Schwarz studied the
first eigenvalue of Laplace’s equation on general domains toward the end of
the 19th century, while Poincare studied Poison’s equation a few years later.
(Kline, 1973). At the start of the 20th century , David Hibert studied the
eigenvalues of integral operators by viewing the operations as infinite matri-
ces matrices (Kline,19720). He was the first to use the German word eigen,
which means ”own”, to denote eigenvalues and eigenvectors in 1904, though
he may have been following a related usage by Hermann von Helmholtz . For
some time, the standard term in English was ”proper value”, but the more
distinctive term ”eigenvalue” is the standard today.(Aldrich,2006),
The first numerical algorithm for computing eigenvalues and eigenvectors ap-
peared in 1929, when Richard Von Mises published the POWER METHOD.
One of the most popular methods today, the QR algorithm, was proposed
independently by John G.F. Francis (Francis ,1961)and Vera Kublanovskaya
(Kublanovskaya,1961) in 1961. (Golub $ Van Loan ,1996), (meyer,2000) An-
other important eigenvalue algorithm developed recently is the shift - invert
residual Arnold method for finding eigenvalues, this method has a few sim-
ilarities to Jacobi method; both aim at only one eigen-pair at a time and
expand the subspace with the solution to a correction equation where the
right-hand side is the current eigenvalue residual vector (the coefficient ma-
trices are different for the two methods, though in the initial steps, the inner
solves can be done with only a moderate accuracy, but the tolerance for the
inner decreases as the outer iteration proceeds.
To further study the Arnold and other methods, extends the strategy and ac-
curacy of matrix vector products with shift - invert transformation for stan-
6
dard eigenvalue problems. Numerical experiments and analysis show that
these techniques lead to significant savings in computational cost without af-
fecting the convergence of outer iterations to the desired eigen pairs, knowing
fully well that the eigenvalue algorithm are outer iteration Eigenvalues and
eigenvectors are fundamental concepts in linear algebra, playing a central
role in various mathematical, scientific, and engineering applications. They
arise in the study of linear transformations, systems of differential equations,
and eigenvalue problems. This literature review aims to provide an in-depth
analysis of the theoretical foundations, numerical methods for computation,
properties, and diverse applications of eigenvalues and eigenvectors.
7
*Characteristic Equation: In mathematics,the characteristic equation
(or auxiliary equation)(Edwards $Penney,2008)is an algebraic equation of
degree n upon which depends the solution of a given nth-order differential
equation or difference equation (William,1970) (Chiang,2004) The character-
istic equation can only be formed when the differential or difference equation
is linear and homogeneous, and has constant coefficients (Edwards $ Pen-
ney,2008).
The eigenvalues of a matrix A are solutions to the characteristic equation:
det(A - λI) = 0 .where I is the n x n identity matrix. The characteristic
equation is derived by finding the values of λ that make the matrix (A -
λI) singular, which means its determinant is zero. Solving the characteristic
equation yields the eigenvalues of the matrix A.
8
have no defective eigenvalues (i.e., eigenvalues whose geometric multiplicity
is less than their algebraic multiplicity). A square matrix A is said to be
diagonalizable if it has n linearly independent eigenvectors. Diagonalization
refers to the process of expressing A in the form of PDP∧(-1), where P is
a matrix formed by the eigenvectors of A, and D is a diagonal matrix with
the eigenvalues on its main diagonal. In this form, A is transformed into a
simpler diagonal matrix, making computations involving A more straightfor-
ward (Taboga,2021) .
9
terminants.(Sheldon,1997),(Hall,2013).The spectral theorem as generalized
by John von Neumann is today perhaps the most important result of opera-
tor theory.
10
ent trade-offs in terms of computational complexity, convergence speed, and
applicability to different types of matrices.(Trefethen $ Bau ,1997).
11
systems. This literature review aims to provide a comprehensive analysis of
the Inverse Power Method, including its theoretical foundations, numerical
implementation, convergence properties, and practical applications.
The Inverse Power Method is based on the observation that if λ is an eigen-
value of a matrix A, then (A - $lambdaI)∧-1, where I is the identity matrix,
has an eigenvalue of 1/
λ
. Thus, by applying inverse iteration to (A - λI)-1, we can compute an
eigenvalue close to 1/λ. The algorithm starts with an initial vector and
successively applies the linear system solver to update the vector until con-
vergence. The convergence rate of the Inverse Power Method is determined
by the gap between the target eigenvalue and the other eigenvalues of the
matrix.(Golub $ Van Loan,2012).The numerical implementation of the In-
verse Power Method involves solving the shifted linear system (A - λI)X =
b at each iteration, where b is the current approximation of the eigenvector
and X is the updated eigenvector. Several numerical methods, such as LU
decomposition, QR factorization, or iterative solvers like the Conjugate Gra-
dient Method, can be employed to solve the linear system efficiently.
Preconditioning techniques can also be applied to improve the convergence
speed of the Inverse Power Method for ill-conditioned matrices.(Kelley,1995)
The convergence behavior of the Inverse Power Method is influenced by vari-
ous factors, such as the initial guess, the shift value (λ), the condition number
of the matrix, and the eigenvector corresponding to the target eigenvalue. For
matrices with distinct eigenvalues, the method typically converges quickly
to the desired eigenvalue. However, if the matrix has multiple eigenvalues
close to the target value, the convergence rate may slow down. Researchers
have investigated different strategies, such as using deflation techniques and
adaptive shifts, to accelerate convergence and handle clustered eigenvalues
effectively(Gupta,2020).
The Inverse Power Method finds various applications in real-world problems,
such as quantum mechanics for computing energy levels of molecules, struc-
tural analysis for determining the lowest natural frequencies of mechanical
systems, and control systems for stability analysis. It is often used to find the
fundamental mode shape and frequency of structures, which are crucial for
predicting their dynamic behavior and response to external forces(Sudarshan
$ Mukunda ,1974).
The Inverse Power Method has a powerful and versatile iterative algorithm
12
for computing eigenvalues and eigenvectors, particularly for finding the small-
est eigenvalue and corresponding eigenvector of a matrix. Its theoretical
foundation, numerical implementation, convergence properties, and practi-
cal applications make it a valuable tool in various scientific and engineering
domains. By understanding its strengths and limitations, researchers and
practitioners can effectively utilize the Inverse Power Method to tackle com-
plex eigenvalue problems in their respective fields and enhance their under-
standing of system behavior and properties. Further research may focus on
developing hybrid approaches that combine the Inverse Power Method with
other iterative techniques to address challenging eigenvalue problems and
extend its applicability to broader contexts.
13
to make them zero, while preserving the symmetry of the matrix. The Ja-
cobi Method is highly parallelizable, making it well-suited for modern parallel
computing architectures. However, the Jacobi Method is relatively slow com-
pared to other iterative techniques, such as the Power Method and the QR
Algorithm, particularly for large dense matrices(Kelley,1995).
The Jacobi Method guarantees convergence to the eigenvalues and eigenvec-
tors of a symmetric matrix. However, its convergence rate is relatively slow,
especially for matrices with clustered eigenvalues. The number of iterations
required for convergence is proportional to the condition number of the ma-
trix and can be large for ill-conditioned matrices. (Horn $ Johnson,2013).
Researchers have proposed various strategies to accelerate convergence, such
as adaptive strategies that focus on off-diagonal elements with larger magni-
tudes and variations of the Jacobi Method combined with other techniques.
The Jacobi Method finds numerous applications in various fields, including
structural mechanics, quantum mechanics, computational physics, and sig-
nal processing. It is often used to compute the eigen modes and natural
frequencies of mechanical structures, to diagonalize Hamiltonian matrices in
quantum mechanics, and for solving linear systems of equations. The Jacobi
Method is particularly valuable for symmetric matrices, where it exhibits
robustness and accuracy(Hjelmervik $ Hesthaven,2014).
Researchers and practitioners can effectively apply the Jacobi Method to
solve eigenvalue problems in their respective fields, leveraging its advantages
for real-world applications. Further research may focus on developing hybrid
methods that combine the Jacobi Method with other iterative techniques
to improve convergence speed and extend its applicability to more general
classes of matrices.
14
vectors of certain matrices, particularly those that are diagonally dominant
or symmetric. The Gauss-Seidel method converges to the solution of the
linear system if the matrix A is diagonally dominant or symmetric positive
definite. Diagonal dominance means that the magnitude of the diagonal el-
ement of A is greater than the sum of the magnitudes of the other elements
in the same row. One key advantage of the Gauss-Seidel method is that it
updates the solution vector elements one at a time, using the latest available
values. This feature allows for efficient implementations in cases where the
matrix A is sparse, as it avoids the need to store the entire matrix explic-
itly.However, the Gauss-Seidel method has some limitations:
15
scenarios. Research by (Kim et al.,2019). Comparative analysis of iterative
methods for computing eigenvalues and eigenvectors.(Martinez et al.,2020).
Performance comparison of iterative methods on diverse matrix types. (Bar-
rett et al.,1994) has assessed the strengths and limitations of each method
on diverse matrix types, providing valuable insights for selecting the most
suitable method for specific applications.
16
is a fundamental aspect of quantum mechanics and is described mathemat-
ically using eigenvectors and eigenvalues. The concept of superposition and
the probabilistic nature of measurements are central to quantum mechanics
and have profound implications for understanding the behavior of particles
at the atomic and subatomic levels (Ball,2005).
17
*Data Analysis: Eigenvectors and eigenvalues are of paramount impor-
tance in data analysis, particularly in techniques like Principal Component
Analysis (PCA) and Singular Value Decomposition (SVD). They enable data
scientists and analysts to uncover patterns, reduce dimensionality, and ex-
tract essential information from high-dimensional datasets. One of the key
applications of eigenvectors and eigenvalues in data analysis is in PCA, a
widely used method for dimensionality reduction and extraction of essential
features In PCA, the eigenvectors of the covariance matrix of the data rep-
resent the principal components, which are new orthogonal axes in the data
space. The corresponding eigenvalues indicate the variance of the data along
these principal components. The principal components are ordered based
on the magnitude of their eigenvalues, allowing data analysts to select the
most significant components that capture the most variability in the data.
By projecting the data onto these principal components, high-dimensional
data can be represented in a lower-dimensional space while preserving the
most critical information.This dimensionality reduction helps in visualiza-
tion, simplifies data analysis, and can speed up computation in machine
learning algorithms. Additionally, PCA can be used for data compression,
noise reduction, and feature selection in various applications.(Hastie $ Fried-
man,2009).
18
as pole placement, allows engineers to control the system’s stability and per-
formance. Moreover, eigenvalues and eigenvectors also help in understanding
the dynamic response of the system to different inputs. The eigenvectors of
the system matrix represent the natural modes of oscillation or behavior of
the system. These modes describe how the system responds when subjected
to different initial conditions or input signals. Engineers can use this infor-
mation to optimize the control system’s design and ensure robustness against
disturbances(Ogata,2010).
19