(Fas) Multigrid Methods for Nonlinear Problems
(Fas) Multigrid Methods for Nonlinear Problems
Abstract. Since their early application to elliptic partial differential equations, multigrid
methods have been applied successfully to a large and growing class of problems, from elas-
ticity and computational fluid dynamics to geodetics and molecular structures. Classical
multigrid begins with a two-grid process. First, iterative relaxation is applied, whose effect is
to smooth the error. Then a coarse-grid correction is applied, in which the smooth error is
determined on a coarser grid. This error is interpolated to the fine grid and used to correct the
fine-grid approximation. Applying this method recursively to solve the coarse-grid problem
leads to multigrid.
The coarse-grid correction works because the residual equation is linear. But this is not the
case for nonlinear problems, and different strategies must be employed. In this presentation
we describe how to apply multigrid to nonlinear problems. There are two basic approaches.
The first is to apply a linearization scheme, such as the Newton’s method, and to employ
multigrid for the solution of the Jacobian system in each iteration.
The second is to apply multigrid directly to the nonlinear problem by employing the so-
called Full Approximation Scheme (FAS). In FAS a nonlinear iteration is applied to smooth
the error. The full equation is solved on the coarse grid, after which the coarse-grid error is
extracted from the solution. This correction is then interpolated and applied to the fine grid
approximation.
We describe these methods in detail, and present numerical experiments that indicate the
efficacy of them.
1. Introduction
Multigrid methods have become a common approach for solving many types of linear prob-
lems of the form Au = f . We denote by v an approximation to the exact solution u and by e
the error, e = u−v. Defining the residual to be r = f −Av, we observe the critical relationship
known as the residual equation, namely Ae = r. In brief, multigrid is the recursive application
of a two-grid process. First, an iterative method, such as Gauss-Seidel or Jacobi relaxation,
is applied to the fine-grid problem. These iterations have the property that after relaxation
the error will be smooth. This in turn means that the error can be accurately represented
on a coarser grid (e.g., with every second gridpoint removed in each coordinate direction).
Since the coarse grid is much smaller than the fine grid, it is much less expensive to work on
the coarse grid. These facts permit the second part of the process, known as the coarse-grid
correction. The fine-grid residual rh is computed and restricted to the coarse grid r2h = Ih2h rh ,
where it is used as the right-hand side of the coarse-grid residual equation A2h e2h = r2h . This
equation is solved, and the error thus determined is then interpolated back to the fine grid
where it is used to correct the fine-grid approximation, vh ← vh + I2h h 2h
e . By recursively
solving the coarse-grid equation with this two-grid process, a multigrid algorithm is defined.
Nonlinear problems necessarily must be solved using iterative methods, and for this reason
it is natural to expect that multigrid ideas should be effective on these problems. Consider a
system of nonlinear algebraic equations, A(u) = f , where u, f ∈ Rn . Once again the error is
given by e = u − e while now r = f − A(v) is the residual. Subtracting the original equation
from the residual, we obtain
Since A is nonlinear, in general A(e) 6= r, implying that for the nonlinear problem we can
not determine the error by solving a simple linear equation on the coarse grid, as in standard
multigrid. Instead, we must use (1.1) as the residual equation.
There are two basic approaches to using multigrid in the solution of (1.1). One is to use
multigrid as the linear solver in a standard linearization, such as in Newton’s method or
Picard iteration. We briefly describe one such approach, Newton-multigrid, in the following
section. The second approach, known as the Full Approximation Scheme (FAS), is to apply
multigrid methodology directly to the original equation A(u) = f and to base the coarse-grid
correction on the nonlinear residual equation (1.1). We describe FAS in section 3. Finally, we
demonstrate the usage of both FAS and Newton-multigrid on some simple model problems in
section 5.
2. Newton-Multigrid
Undoubtedly most important method for solving nonlinear equations is Newton’s Method.
Suppose we wish to solve the scalar equation F (x) = 0. By expanding F in a Taylor series
about an initial guess x0 we obtain
s2 00
F (x0 + s) = F (x0 ) + sF 0 (x0 ) + F (ξ)
2
for some ξ between x0 and x0 + s. If x0 + s were the solution, then neglecting the second-
order term and solving for s we obtain an improved approximate solution given by x0 ←
x0 − F (x0 )/F 0 (x0 ). This is easily turned into the Newton iteration:
F (xj )
xj ← xj − .
F 0 (xj )
We can use (1.1) as a basis for a multigrid solver by applying Newton’s method to the system
of n nonlinear equations
A1 (x1 , x2 , . . . xn ) 0
A2 (x1 , x2 , . . . xn ) 0
A(x) ≡ .. = . .
..
.
An (x1 , x2 , . . . xn ) 0
Letting J(v) be the Jacobian matrix
∂A1 ∂A1 ∂A1
∂x1 ∂x2
··· ∂xn
∂A2 ∂A2
··· ∂A2
∂x1 ∂x2 ∂xn
J(v) = .. .. .. .. ,
. . . .
∂An ∂An ∂An
∂x1 ∂x2
··· ∂xn x=v
and letting u = v + e be the exact solution, the Taylor series for this system is
A(v + e) = A(v) + J(v)e + higher order terms.
Neglecting the higher order terms and subtracting A(v) from both sides, we see that
(2.1) J(v)e, = r,
which represents an approximation to the nonlinear system (1.1). It can be solved for e and
the current approximation v can be updated by v ← v + e. Iteration of this step is
(2.2) vj ← vj + J −1 (vj ) (fj − A(vj )) ,
a form of Newton’s method. By using multigrid to solve linear system (2.1) at each step we
obtain a combination of Newton’s method for the outer iteration and multigrid for the (linear)
inner iteration. This is called Newton-multigrid. We will present examples of this method in
section 5.
The right side of this nonlinear system is known, and the equation is of the same form as
the original fine-grid equation (3.1). Assume we can find a solution to this system, which we
denote u2h . The coarse-grid error can be extracted from the solution by e2h = u2h −Ih2h vh , and
can then be interpolated up to the fine grid and used to correct the fine-grid approximation
vh :
h 2h
vh ← vh + I2h e or vh ← vh + I2hh
(u2h − Ih2h vh ).
This scheme is the most commonly used nonlinear version of multigrid. Because the coarse-
grid problem is solved for the full approximation, rather than the error e2h , the method is
called the full approximation scheme (FAS). A two-grid version of this scheme is described as
follows:
It is easy to see that if A is a linear operator, then FAS reduces directly to the linear coarse-
grid correction scheme. Hence, FAS is in fact a generalization of the coarse-grid correction to
nonlinear problems. It is also useful to observe that vh is a fixed point of the FAS iteration if
and only if it is an exact solution of the fine-grid problem.
We may also view FAS as a way to alter the coarse-grid equations so that their approxima-
tion properties are substantially enhanced. This may be seen if the FAS coarse-grid equation
is written as
A2h (u2h ) = f 2h + τ 2h
h where τ 2h 2h 2h h 2h h h
h = A (Ih v ) − Ih A (v )
Typically, this is a nonlinear scalar equation in the scalar s, and the scalar Newton’s method
can be used to solve it efficiently [5]. The jth component is then updated by v ← v + s²j .
Updating all n components sequentially constitutes one iteration sweep of nonlinear Gauss-
Seidel.
The two-stage process of a) nonlinear relaxation followed by b) FAS correction is exactly
analogous to the relaxation/coarse-grid correction that comprises the two-grid scheme under-
lying conventional multigrid. Hence it is natural to ask at this point, “How do we solve the
coarse-grid problem A2h (u2h ) = f 2h + τ 2hh ?” Nonlinear multigrid results by recursively apply-
ing the same nonlinear relaxation/FAS correction algorithm to solve this problem. Thus, FAS,
like its linear counterparts, is usually implemented as a V-cycle or W-cycle scheme (Figure 1).
Another important technique is the use of full multigrid (FMG) for obtaining a good initial
guess the fine-grid problem [3], in which for any given grid level, the problem is first solved
on the next coarser grid to provide and interpolated to the given grid to provide a good
initial guess. The solution process begins on the coarsest grid (Figure 1(c)). Convergence of
nonlinear iteration depends even more critically on a good initial guess, since the closer the
fine-grid initial guess is to the solution, the more linear the fine-grid problem appears, and the
more effective the solver will be. Using FMG for nonlinear problems typically ensures that the
h 2h
interpolant I2h u lies in the basin of attraction of the fine-grid solver. Thus, we can expect
one FMG cycle to provide good accuracy whether we use Newton-multigrid or FAS V-cycles
on each new level. If greater accuracy is required, the FMG cycle is followed by successive
V-cycles.
(5.1) h−2 (4ui,j − ui−1,j − ui+1,j − ui,j−1 − ui,j+1 ) + γ ui,j eui,j = fi,j ,
for 1 < i, j < n. To enforce the boundary condition, we set
u0,j = uN,j = ui,0 = ui,N = 0
wherever these terms appear in the equations.
We consider first the implementation of an FAS solver for this problem. We must select
intergrid transfer operators; the easiest and most natural for this problem are to use a lin-
ear interpolation operator and a full-weighting restriction operator. The linear interpolation
operator can be defined by vh = I2h v , with components of vh given, for 0 ≤ i, j ≤ n2 − 1, by
h 2h
h
v2i,2j = vij2h ,
h 1 ¡ 2h 2h
¢
v2i+1,2j = vij + vi+1,j ,
2
h 1 ¡ 2h 2h
¢
v2i,2j+1 = vij + vi,j+1 ,
2
h 1 ¡ 2h 2h 2h 2h
¢
v2i+1,2j+1 = vij + vi+1,j + vi,j+1 + vi+1,j+1 .
4
The full-weighting restriction operator produces at a coarse-grid point a value that is just an
average of the values at the corresponding fine-grid point and its eight nearest neighbors:
1£ h
vij2h = v h
+ v2i−1,2j+1 h
+ v2i+1,2j−1 h
+ v2i+1,2j+1
16 ¡ 2i−1,2j−1 ¢
h h h h
+ 2 v2i,2j−1 + v2i,2j+1 + v2i−1,2j + v2i+1,2j
h
¤
+ 4v2i,2j ,
n
for 1 ≤ i, j ≤ 2
− 1.
Coarse-grid versions of the nonlinear operator are obtained in the same manner as the fine-
grid operator. That is, we employ discretization (5.1) with the appropriate grid spacing (2h,
4h, . . .) and index limits. Hence we may examine the individual component equations of the
FAS correction (3.3). Once a fine-grid approximation vh has been obtained and v2h = Ih2h vh
and r2h = Ih2h (f h −Ah (uh )) are computed, the individual component equations of (3.3) appear
as
2h
2h ui,j
4h−2 (4u2h 2h 2h 2h 2h
i,j − ui−1,j − ui+1,j − ui,j−1 − ui,j+1 ) + γ ui,j e
(5.2) 2h
2h vi,j
= 4h−2 (4vi,j
2h 2h
− vi−1,j 2h
− vi+1,j 2h
− vi,j−1 2h
− vi,j+1 ) + γ vi,j 2h
e + ri,j .
for 1 ≤ i, j ≤ n2 − 1. Note that the left side of this equation is exactly the same form as (5.1)
and represents the nonlinear operator; the right side carries the effect of the nonlinearity in
the full approximation mode.
We employ nonlinear Gauss-Seidel as the relaxation method. Because the individual com-
ponent equations of the system are nonlinear, the relaxation uses scalar Newton’s method to
solve the (i, j)th equation for ui,j :
γ
0 1 10 100 1000 10000
Convergence Factor .136 .135 .124 .098 .072 .039
Number of FAS cycles 12 12 11 11 10 8
This particular problem becomes easier to solve as the nonlinear term becomes more dom-
inant, which is, in general, an uncommon effect in a nonlinear problem. Since the nonlinear
term involves only ui,j and none of its neighboring values, as γ becomes larger the problem
becomes more local in nature. As this occurs, the relaxation sweeps become more effective,
being local in nature. This phenomenon is analogous to increasing diagonal dominance in the
linear case. This type of behavior would not be expected for other types of nonlinearity.
We can compare the performance of the FAS solver Newton’s method for this problem. The
Jacobian matrix for this nonlinear problem is a block tridiagonal system
J1 B
B J2 B
B J3 B
J(u) = ... ... ... ,
B J B
N −2
B JN −1
where each of the block matrices is (n − 1) × (n − 1). The off-diagonal blocks B are all
−1/h2 times the identity matrix. The diagonal blocks are tridiagonal, with the constant value
−1/h2 on the super- and sub-diagonals. The diagonal entries of Jj , corresponding to the grid
locations (xi , yj ) = (ih, jh) for fixed j and 1 ≤ i ≤ n − 1, are given by
4
(Jj )i,i =
2
+ γ ui,j eui,j .
h
For this experiment, we use a banded LU decomposition to Jacobian matrix at each Newton
step. Because J(u) is sparse and narrow-banded, so are the factors L and U and doth the
decomposition and the solve can be computed quite efficiently. Table 2 gives the results of
applying Newton’s method to the problem for various choices of γ.
γ
0 1 10 100 1000 10000
Convergence Factor 2.6e-13 3.9e-5 7.4e-5 3.2e-4 1.9e-4 1.2e-4
Newton Iterations 1 3 3 4 4 4
Tables 1 and 2 reveal that Newton’s method converges much faster than does FAS. This
should not be surprising, as Newton’s method converges quadratically, while multigrid is gen-
erally linear in its convergence. However, this comparison fails to consider the computational
costs of the methods, which we will discuss shortly.
Before doing so, however, we consider the use of multigrid V-cycles as a an method for
inverting the Jacobian system in the Newton iteration. If we solve the Jacobian system to the
same level of accuracy by V-cycles as was done with the LU decomposition, the results should
be essentially the same as in Table 2. However, it may be effective to solve the Jacobian system
only approximately. This technique is known as the inexact Newton’s method[4]. Performing
inexact Newton by using only a few V-cycles is a method we denote Newton-MG.
The following experiment is designed to compare the performance of FAS, Newton’s meth-
ods, and Newton-MG. Consider the same operator as before, with γ = 10, and let the exact
solution be u(x, y) = (x2 − x3 ) sin(3πy). Unlike in the previous experiments, this solution
results in a nontrivial discretization error. The resulting source term is
³³ ´ ´
u 2 (x2 −x3 ) sin(3πy) 2 3
(5.4) −∆u + γue = 9π + γe (x − x ) + 6x − 2 sin(3πy).
The problem is discretized with n = 128 so that the interior grid is 127 × 127.
Table 3 shows the performance on and the computational costs of FAS, Newton (with a
direct LU -based solve), and Newton-MG applied to (5.4), resulting in the nonlinear system
(5.1). The iteration is stopped when the residual norm is less than 10−10 . The column labeled
outer iterations lists how many V-cycles (for FAS) or Newton steps (for Newton and Newton-
MG) are required to achieve the desired tolerance. For Newton-MG, we vary the number of
V-cycles used to approximately solve the Jacobian system. The column labeled inner iterations
gives the number of V-cycles used in the approximate inversion of the Jacobian system. The
last column, labeled Megaflops, is the number of millions of floating-point operations expended
by the methods. For this example, using these performance measurements, it appears that
both of the multigrid-based methods are much more efficient than Newton’s method using a
direct solver. Furthermore, Newton-MG compares well to FAS when the number of inner MG
iterations is properly tuned.
While the convergence properties may be representative for a fairly broad class of nonlinear
problems, we must emphasize that several factors can affect the operation counts dramatically.
For example, they may depend heavily on the character of the nonlinearity, the details of
implementation, the computer architecture, and the programming language. It should also
be remembered that we have not accounted for the cost of evaluating the nonlinear function,
which is highly problem-dependent, and, for many problems, may be the most costly part of
the process. It does seem fairly clear, however, that for large problems the multigrid-based
methods will outperform Newton’s method using direct solvers. Further, it is apparent that,
if Newton-MG is to be competitive with FAS, only a few V-cycles should be used in solving
the Jacobian system. In the end, there may be very little difference in the performance
between the two carefully designed multigrid-based schemes. The choice may depend largely
on convenience and other available algorithm features (for example, τ -extrapolation in FAS
[2]).
We should consider one further option: combining the Newton and FAS methods with a full
multigrid (FMG) scheme [2, 3]. In FMG, for each new fine grid, the initial guess is obtained
by first solving the nonlinear problem on the next coarser grid. For linear multigrid problems,
a single FMG cycle suffices to reduce the error to the level of discretization. One indication
that this is the case is that after a single FMG cycle continued V-cycles have the effect of
further decreasing the residual (as the linear systems are solved with increasing accuracy) but
the norm of the error no longer decreases.
Table 4 shows the results of applying the FMG-FAS combination to (5.4). Norms are shown,
for the residual and the error, after one FMG-FAS cycle and after each of eight subsequent
FAS V-cycles. Both the FMG cycle and the FAS V-cycles are performed using nonlinear
Gauss-Seidel. The norm of the error is reduced to 2.0 × 10−5 by the FMG-FAS cycle alone.
Further FAS V-cycling does not reduce the error, indicating that it has reached the level of
discretization. However, subsequent FAS V-cycles do reduce the residual norm further until
it reaches the prescribed tolerance of 10−10 . The column labeled Mflops in the table gives the
cumulative number of floating point operations after each stage of the computation.
Analogous results for FMG with a Newton solver applied to (5.4) are shown in Table 5.
Again we use FMG, applying one step of Newton-MG on each level in the FMG process.
Each Newton-MG step starts with an initial guess from the next coarser grid and uses one V-
cycle. Norms are shown for the residual and error after one FMG-Newton-MG cycle followed
by subsequent Newton-MG cycles on the fine grid. The results are very similar to those for
Cycle ||r|| ratio ||e|| Mflops
FMG-FAS 1.07e-2 2.00e-5 3.1
FAS V 1 6.81e-4 0.064 2.44e-5 5.4
FAS V 2 5.03e-5 0.074 2.49e-5 7.6
FAS V 3 3.89e-6 0.077 2.49e-5 9.9
FAS V 4 3.25e-7 0.083 2.49e-5 12.2
FAS V 5 2.98e-8 0.092 2.49e-5 14.4
FAS V 6 2.94e-9 0.099 2.49e-5 16.7
FAS V 7 3.01e-10 0.102 2.49e-5 18.9
FAS V 8 3.16e-11 0.105 2.49e-5 21.2
FMG-FAS in Table 4. Observe that the norm of the actual error is reduced to the level
of discretization by one FMG-Newton cycle. Subsequent Newton-MG cycles do, however,
continue to reduce the norm of the residual effectively to below the prescribed tolerance.
Both of these methods reduce the error to the level of discretization in one FMG cycle.
The flop count indicates that the methods are similar in cost, with the FMG-Newton-MG
cycle somewhat less expensive (2.4 Mflops) than the FMG-FAS (3.1 Mflops). However, the
individual Newton-MG steps on the fine grid, although cheaper, are not quite as effective as
FAS V-cycles for reducing the residual norm. Indeed, if the goal is to reduce the residual
norm to 10−10 , it is somewhat less expensive (21.2 vs. 26.2 Mflops) to use FAS than Newton-
MG. Again, however, we note that comparative measures of efficiency depend critically on
specific implementation details, computing environment, and problem characteristics. The
major conclusion to be reached is that both methods are very efficient and robust.
6. Conclusions
We have presented this material very much in an overview or tutorial style, and have not
discussed particular applications. Both Newton-MG and FAS are in widespread use for many
application areas, including computational fluid dynamics, radiation transport, image recon-
struction, and porous media flow, among others. Still other techniques, including hybrids of
various linear and nonlinear methods, are known or are under development. Current research
also includes applying FAS to unconventional multigrid, such as algebraic multigrid (AMG).
Two major factors determine what sort of nonlinear multigrid solver is applied. First,
one must determine how the nonlinear operator can be represented on the coarse grid. The
most common approach is to apply the same discretization method with a coarse spacing
as was used to discretize the fine grid problem. If this is unfeasible of overly expensive,
however, other choices must be considered if FAS is to be used. Sometimes it is possible to
use the so-called Galerkin method, in which the action of the coarse-grid operator is given
by A2h (u2h ) = Ih2h Ah (I2h
h 2h
u ). The second factor is the relative expense of evaluating the
nonlinear function versus the cost of forming the Jacobian matrix. In many applications
evaluating the nonlinear function is a very expensive computational procedure; indeed, it can
be the most costly part of the FAS algorithm. If the Jacobian matrix can be evaluated with
relative ease, Newton-MG may be preferable. These determinations are highly dependent on
the nature of the specific problem at hand.
For any nonlinear method, it is well-known that the iteration is much more reliable and
much faster if the initial guess is good. Hence, it is well worth the effort to obtain the highest
quality initial guess possible. As a result, nested iteration, the method that obtains a good
initial guess by first solving the problem on a coarser grid, should be employed. Doing this
on all levels is the heart of the FMG approach. Whether FAS, Newton-MG, or some other
nonlinear solver is obtained, using the FMG scheme will improve both speed of convergence
and robustness.
7. Acknowledgements
This discussion follows closely that of Chapter 6 of the book A Multigrid Tutorial, Second
Edition, published by the Society for Industrial and Applied Mathematics and co-authored
by Bill Briggs, Steve McCormick,and the author of this paper [3]. The numerical examples
are taken directly from that work. This has been done with the gracious permission of the
authors and the publisher.
This work was performed under the auspices of the U. S. Department of Energy by Univer-
sity of California Lawrence Livermore National Laboratory under contract W-7405-Eng-48.
The author’s contact information is: Van Emden Henson, Center for Applied Scientific
Computing, Lawrence Livermore National Laboratory, L-560, PO Box 808, Livermore, CA,
94551, email:[email protected].
References
[1] A. Brandt, Generally highly accurate algebraic coarsening, Elec. Trans. Num. Anal., 10 (2000), pp. 1–20.
[2] A. Brandt, Multigrid techniques: 1984 guide with applications to fluid dynamics, GMD-studien Nr. 85,
Gesellschaft für Mathematik und Datenverarbeitung, St. Agustin, 1984.
[3] W. L. Briggs, V. E. Henson, and S. F. McCormick,A Multigrid Tutorial, Second Edition, SIAM
Books, Philadelphia, 2000.
[4] J.E. Dennis and R.B. Schnabel, Numerical Methods for Unconstrained Optimization and Nonlinear
Equations, SIAM Books, Philadelphia, 1996.
[5] J.M. Ortega and W.C. Rheinboldt, Iterative Solution of Nonlinear Equations in Several Variables,
Academic Press, San Diego, 1970.