Ad Joint Course
Ad Joint Course
Department of Mathematics
Colorado State University
Fort Collins, CO 80523
Definition
A sequence {xn } in X is a Cauchy sequence if we can make
the distance between elements in the sequence arbitrarily small
by restricting the indices to be large.
Example
Consider the sequence {1/n}∞
n=1 in [0, 1]. It is a Cauchy
sequence since
1 1 m−n max{m, n} 2
− = ≤2 =
n m mn mn min{m, n}
Example
The sequence {1/n}∞ n=1 is a Cauchy sequence in (0, 1) but does
not converge in (0, 1).
Duality, Adjoints, Green’s Functions – p. 7/304
Background in linear algebra
Definition
A Banach space is a vector space with a norm such that every
Cauchy sequence converges to a limit in the space.
Example
Rn with the norms defined for x = (x1 , · · · , xn )> ,
Spaces of functions:
Definition
For an interval [a, b], the space of continuous functions is
denoted C([a, b]), where we take the maximum norm
kf k = max |f (x)|.
a≤x≤b
Spaces of functions:
Definition
For 1 ≤ p ≤ ∞,
L2 is particularly important.
Theorem
The Lp spaces and C([a, b]) are Banach spaces.
Definition
A map or operator L from X to Y is a rule or association that
assigns to each x in X a unique element y in Y .
Definition
A map L : X → Y is linear if L(αx1 + βx2 ) = αL(x1 ) + βL(x2 )
for all numbers α, β and x1 , x2 in X .
Example
Every linear map from Rm to Rn is obtained by multiplying
vectors in Rm by a n × m matrix, i.e., they have the form Ax,
where A is a n × m matrix.
Example
Differentiation is a linear map from C 1 ([a, b]) to C([a, b]).
Definition
A map L : X → Y is continuous if for every sequence {xn } in
X that converges to a limit x in X , i.e., xn → x, we have
L(xn ) → L(x).
Example
Linear maps from Rm to Rn are continuous.
Definition
A linear map L : X → Y is bounded if there is a constant C > 0
such that kLxkY ≤ CkxkX for all x in X .
Theorem
A linear map between normed vector spaces is continuous if
and only if it is bounded.
Definition
If X and Y are normed vector spaces, we use L(X, Y ) to denote
the vector space of all bounded linear maps from X to Y .
L(X, Y ) is a normed vector space under the operator norm
kLxkY
kLk = sup kLxkY = sup . (1)
kxkX =1 x6=0 kxkX
Example
If the linear transformation L is given by the n × n matrix A, then
n
X
kLk1 = kAk1 = max |aij |
1≤j≤n
i=1
q
kLk2 = kAk2 = σ(AT A)
n
X
kLk∞ = kAk∞ = max |aij |
1≤i≤n
j=1
Theorem
If X and Y are normed vector spaces and Y is complete, then
L(X, Y ) is complete.
Definition
A linear functional on a vector space X is a linear map from X
to R.
Example
Let v in Rn be fixed. The map F (x) = v · x = (x, v) is a linear
functional on Rn .
Example
Rb
Consider C([a, b]). Both I(f ) = a f (x) dx and F (f ) = f (y) for
a ≤ y ≤ b are linear functionals.
Example
In the example above with F (x) = (x, v), consider v = ei , the ith
standard basis function. Then F (x) = xi where x = (x1 , · · · , xn ).
Example
δy gives a linear functional on sufficiently smooth, real valued
functions via
Z
F (u) = u(y) = δy (x)u(x) dx.
Ω
Definition
If X is a normed vector space, the space L(X, R) of bounded
linear functionals on X is called the dual space of or on or to
X , and is denoted by X ∗ .
The dual space is a normed vector space under the dual norm
defined for y ∈ X ∗ as
|y(x)|
kykX ∗ = sup |y(x)| = sup .
x∈X x∈X kxk
kxkX =1 x6=0
Example
Consider X = Rn . Every vector v in Rn is associated with a
linear functional Fv (·) = (·, v). This functional is clear bounded
since |(x, v)| ≤ kvk kxk (The “C ” in the definition is kvk).
Example
Rb
For C([a, b]), consider I(f ) = a f (x) dx. It is easy to compute
Z b
kIkC([a,b])∗ = sup f (x) dx
f ∈C([a,b]) a
max |f |=1
by looking at a picture.
possible functions
b
a
-1
Example
Recall Hölder’s inequality for f ∈ Lp (Ω) and g ∈ Lq (Ω) with
1 1
+ = 1 for 1 ≤ p, q ≤ ∞ is
p q
Example
By considering the set of n functionals corresponding to taking
the inner product with {e1 , · · · , en }, we can “reconstruct” any
given vector in Rn by looking at the functional values.
Issues:
• What does it mean to extend?
• Does the norm increase upon extension?
• What if there is an infinite number of extensions involved?
Theorem
Hahn-Banach Let X be a Banach space and X0 a subspace of
X . Suppose that F0 (x) is a bounded linear functional defined on
X0 . There is a linear functional F defined on X such that
F (x) = F0 (x) for x in X0 and kF k = kF0 k.
The dual space can be better behaved than the original normed
vector space.
Theorem
If X is a normed vector space over R, then X ∗ is a Banach
space (whether or not X is a Banach space).
Useful notation:
Definition
If x is in X and y is in X ∗ , we denote the value
The norms on X and its dual X ∗ are closely related. Recall that
if y ∈ X ∗ , then
|y(x)|
kyk X∗ = sup |y(x)| = sup .
x∈X x∈X kxk
kxkX =1 x6=0
Definition
The generalized Cauchy inequality is
Theorem
If X is a Banach space, then
|y(x)|
kxkX = sup = sup |y(x)|
y∈X ∗ kykX ∗ y∈X ∗
y6=0 kykX ∗ =1
for all x in X .
Example
Rn with the standard Euclidean norm k k = k k2 can be identified
with its dual space.
Theorem
If X has an inner product (x, y), then it is a normed vector space
with norm kxk = (x, x)1/2 for x in X .
Definition
A vector space with an inner product that is a Banach space
with respect to the associated norm is called a Hilbert space.
Example
Rn with k k = k k2 and L2 are both Hilbert spaces.
Theorem
Riesz Representation For every bounded linear functional F
on a Hilbert space X , there is a unique element y in X such that
|F (x)|
F (x) = (x, y), all x ∈ X and kykX ∗ = sup .
x∈X kxk
x6=0
Definition
Two normed vector spaces X and Y are isometric if there is a
linear 1-1 and onto map L : X → Y such that kL(x)kY = kxkX
for all x in X .
Example
For k = 1, 2, 3, · · · , we define H k (Ω) to be the distribution
functions in L2 (Ω) whose partial derivatives of order k and less
are also distributions in L2 (Ω).
Definition
index notation: For α = (α1 , · · · , αn ) with integer coefficients,
we define
∂ |α|
∂ ∂ α
D= ,··· , , D =
∂x1 ∂xn ∂xα1 1 · · · ∂xαnn
with |α| = α1 + · · · + αn .
For each y ∗ ∈ Y ∗ ,
x∗ is unique.
y ∗ (L(x)) = L∗ y ∗ (x)
Example
Let X = Rm and Y = Rn , where we take the standard inner
product and norm.
and
m
X
yi = aij xj , 1 ≤ i ≤ n.
j=1
The matrix A∗ of L∗ is
∗ ∗
a11 · · · a1n a11 a21 · · · an1
∗ .. .
. .. .. = A> .
A = . . = . .
a∗m1 · · · a∗mn a1m a2m · · · anm
proof
Therefore,
|L∗ y ∗ (x)|
kL∗ y ∗ kX ∗ = sup ≤ ky ∗ kY ∗ kLk,
x6=0 kxk
and
|y ∗ (Lx)| ∗
sup ∗
≤ kL k kxkX , x ∈ X.
y ∗ 6=0 ky kY ∗
Theorem
Let X , Y , and Z be normed linear spaces. Then,
0∗ = 0
(L1 + L2 )∗ = L∗1 + L∗2 , all L1 , L2 ∈ L(X, Y )
(αL)∗ = αL∗ , all α ∈ R, L ∈ L(X, Y )
We are often deal with linear operators that are not defined on
the entire space.
Example
Consider D = d/dx on X = C([0, 1]). This linear map is only
defined on the subspace C 1 ([0, 1]).
Definition
Let X and Y be normed vector spaces. A map L that assigns to
each x in a subset D(L) of X a unique element y in Y is called a
map or operator with domain D(L).
We want
L∗ y ∗ (x) = y ∗ (Lx) all x ∈ D(L).
Definition
A subspace A of a normed linear space X is dense if every
point in X is either in A or the limit of a sequence of points in A.
Example
The rational numbers are dense in the real numbers and the
polynomials are dense in C([a, b]).
Lx = b.
Definition
The set of b for which there is a solution is called the range,
R(L), of L.
The set of x for which L(x) = 0 is called the null space, N (L),
of L.
0 is always in N (L).
Theorem
N (L) is a subspace of X and R(L) is a subspace of Y .
For y ∗ ∈ Y ∗ ,
y ∗ (Lx) = y ∗ (y).
By the definition of the adjoint,
L∗ y ∗ (x) = y ∗ (y).
Is this sufficient?
Definition
A subset A of a normed vector space X is closed if every
sequence {xn } in A that has a limit in X has its limit in A.
Theorem
Let X and Y be normed linear spaces and L ∈ L(X, Y ). A
necessary condition that y ∈ R(L) is y ∗ (y) = 0 for all
y ∗ ∈ N (L∗ ). This is a sufficient condition if R is closed in Y .
Example
Suppose that L ∈ L(X, Y ) is associated with the n × m matrix
A, i.e., L(x) = Ax.
Example
In the case X is a Hilbert space and L ∈ L(X, Y ), then
necessarily R(L∗ ) ⊂ N (L)⊥ , where S ⊥ is the subspace of
vectors that are orthogonal to a subspace S .
Example
The setting of general partial differential equations:
Theorem
Holmgren Uniqueness The generalized initial value problem
consisting of the equation
X
L(u) = Aα (x)Dα u = f (x), x ∈ Rn ,
|α|≤m
where {Aα } and f are analytic functions, together with the data
Example
y 00 = d2 y/dx2 requires two boundary conditions to define a
problem for the associated differential equation that has a
unique solution.
Example
Divergence,
∂u1 ∂un
div u = + ··· + ,
∂x1 ∂xn
associates a scalar with a given vector function, and the
associated differential equation is very “under-determined”.
The gradient,
∂u1 ∂un
grad u = ,··· , ,
∂x1 ∂xn
associates a vector field with a given scalar function, and the
associated equations are very “over-determined”.
Sz = d,
with ! ! !
y b 0 A
z= , d= , S= .
x c A> 0
Consequences:
• The theorem on the adjoint condition for solvability falls out
right away.
• It yields a “natural” definition of a solution or gives
conditions for a solution to exist in the over-determined and
under-determined cases.
• It also gives a way of determining the condition of the
solution process.
Example
Consider 2x!1 + x2 = 4, where L : R2 → R. L∗ : R → R2 is given
2
by L∗ = . The extended system is
1
0 2 1 y1 4
2 0 0 x1 = c1 ,
1 0 0 x2 c2
Example
On the other hand, if the problem is
2x1 + x2 = 4
x2 = 3,
1 1 0 0 x2 3
where we can specify any values for c1 , c2 .
Example
Consider the under-determined problem
div F = ρ.
Example
Consider
Ax = b
A> φ = ei
Example
A standard difference approximation of
(
−u00 (x) = f (x), 0 < x < 1,
u(0) = 0, u(1) = 0,
Example
If we change the boundary conditions to
(
−u00 (x) = f (x), 0 < x < 1,
u(0) = 0, u0 (0) = 0,
Example
( (
u0 = f (x), 0 < x < 1, −v 0 = g(x), 1 > x > 0,
⇒
u(0) = 0 v(1) = 0,
Example
(
−u00 = f (x), 0 < x < 1,
no boundary conditions
(
−v 00 = g(x), 0 < x < 1,
⇒
v(0) = v 0 (0) = v(1) = v 0 (1) = 0.
which is
Definition
We say that we are evaluating the bilinear identity when we
compute
Definition
A function on a domain Ω has compact support in Ω if it
vanishes identically outside a bounded set contained inside Ω.
Definition
Let L be a differential operator. The formal adjoint L∗ is the
differential operator that satisfies
Z Z
(Lu, v) = (u, L∗ v) Lu · v dx = u · L∗ v dx
Ω Ω
Example
Consider
d d d
Lu(x) = − a(x) u(x) + (b(x)u(x))
dx dx dx
1 1 1
d d
Z Z
(b(x)u(x))v(x) dx = − u(x)b(x) v(x) dx+b(x)u(x)v(x) ,
0 dx 0 dx 0
Therefore,
d d d
L∗ v = − a(x) v(x) − b(x) (v(x)).
dx dx dx
Example
A general linear second order differential operator L in Rn can
be written
n X
n n
X ∂2u X ∂u
L(u) = aij + bi + cu,
∂xi ∂xj ∂xi
i=1 j=1 i=1
where
n
X ∂u ∂(aij v)
pi = aij v −u + bi uv.
∂xj ∂xj
j=1
A typical term,
∂2u
∂ ∂u ∂ ∂u ∂(a11 v) ∂u
va11 = va11 = va11 −
∂x21 ∂x1 ∂x1 ∂x1 ∂x1 ∂x1 ∂x1
∂ 2 (a11 v)
∂ ∂u ∂ ∂(a11 v)
= va11 − u +u
∂x1 ∂x1 ∂x1 ∂x1 ∂x21
yielding
∂2u ∂ 2 (a11 v)
∂ ∂u ∂(a11 v)
va11 − u = a11 v −u .
∂x21 ∂x21 ∂x1 ∂x1 ∂x1
Example
Let L be a differential operator of order 2p of the form
X
|α| α β
Lu = (−1) D aαβ (x)D u ,
|α|,|β|≤p
then X
∗ |α| α β
L v= (−1) D aβα (x)D v ,
|α|,|β|≤p
then X
∗
L v= (−1)|α| Dα (aα (x)v(x)).
|α|≤p
Example
If
Lu = ρutt − ∇ · (a∇u) + bu
then L∗ = L.
If
Lu = ut − ∇ · (a∇u) + bu
then
L∗ v = −vt − ∇ · (a∇v) + bv
where time runs “backwards”.
Example
Let
L(~u) = A~ux + B~uy + C~u,
where A, B , and C are n × n matrices, then
so that
~v > L~u − ~u> L∗~v = ∇ · (~v > A~u, ~v > B~u).
Definition
The adjoint boundary conditions posed on the formal adjoint
operator are the minimal conditions required to make the
boundary terms that appear when evaluating the bilinear identity
for general smooth functions vanish.
Example
Consider Newton’s equation of motion s00 (x) = f (x) with x =
“time”, normalized with mass 1. First, suppose we assume
s(0) = s0 (0) = 0, and 0 < x < 1. We have
00 d
00
s v − sv = (vs0 − sv 0 )
dx
and
Z 1
1
(s00 v − sv 00 ) dx = (vs0 − sv 0 ) 0 .
0
Example
Suppose instead we impose conditions such that at the mass
returns to the origin with zero speed at x = 1.
Example
Based on the discussion above, we require the data f to be
orthogonal to the solution of the adjoint problem v 00 = 0, which is
v = a + bx.
Example
Since
∂v ∂u
Z Z
(u∆v − v∆u) dx = u −v ds,
Ω ∂Ω ∂n ∂n
Example
Let Ω ⊂ R2 be bounded with a smooth boundary and let s =
arclength along the boundary. Consider
(
−∆u = f, x ∈ Ω,
∂u ∂u
∂n + ∂s = 0, x ∈ ∂Ω.
Since
∂v ∂v ∂u ∂u
Z Z
(u∆v − v∆u) dx = u − −v + ds,
Ω ∂Ω ∂n ∂s ∂n ∂s
Definition
The Green’s function satisfies
(
L∗ φ(y, x) = δy (x), x ∈ Ω,
adjoint b.c. and i.v., x ∈ ∂Ω,
Example
For (
−∆u = f, x ∈ Ω,
u = 0, x ∈ ∂Ω,
φ solves
(
−∆φ(y; x) = δy (x), x ∈ Ω,
φ(y; x) = 0, x ∈ ∂Ω,
Example
For
ut − ∆u = f,
x ∈ Ω, 0 < t ≤ T,
u(t, x) = 0, x∂Ω, 0 < t ≤ T,
u(0, x) = 0, x ∈ Ω,
φ solves
−φt − ∆φ = δ(s,y) (t, x),
x ∈ Ω, T > t ≥ 0,
φ(t, x) = 0, x∂Ω, T > t ≥ 0,
φ(T, x) = 0, x ∈ Ω,
Motivations:
• The Green’s function is defined without reference to the
particular data f that determines a particular solution. The
Green’s function depends on the operator and its properties.
• The Green’s function generally has special properties
arising from the properties of the delta function, such as
localized support and symmetry.
• The Green’s function gives a “low-dimensional” snapshot of
the solution.
Example
The Green’s function for the Dirichlet problem for the Laplacian
L = −∆ on the ball Ω of radius r centered at the origin in R3 is
−1
(
−1 −1 r2 y
1 |y − x| − r|y| |y|2 − x , y 6= 0,
φ(y; x) = ×
4π |x|−1 − r−1 , y = 0,
When the original problem has a unique solution and the original
operator is the adjoint to the adjoint operator, the Green’s
function φ for the original problem and the Green’s function for
the adjoint problem φ∗ are related via
This depends on the dual space of the dual space of the original
space being identifiable with the original space.
We carry out the analysis using the Green’s function for the
homogeneous problem and some additional integrals involving
data and the Green’s function will appear.
Example
Suppose the problem is
(
−∆u = f, x ∈ Ω,
u = g, x ∈ ∂Ω.
This yields
∂φ(y; s)
Z Z
u(y) = f (x)φ(y; x) dx − g(s) ds.
Ω ∂Ω ∂n
Example
For (
−∆u = f, x ∈ Ω,
u = 0, x ∈ ∂Ω,
∂u
Z
(y) = − φ(y; x)f (x) dx.
∂n Ω
Theorem
Sobolev If s > k + n/2, then there is a constant C such that for
f ∈ H s (Rn ),
max sup |Dα (x)| ≤ Ckf ks .
|α|≤k x∈Rn
We consider
(
Lu = f, x ∈ Ω,
suitable b.c. and i.v., x ∈ ∂Ω,
Definition
The generalized Green’s function corresponding to the
quantity of interest (u, ψ) satisfies
(
L∗ φ(y, x) = ψ(x), x ∈ Ω,
adjoint b.c. and i.v., x ∈ ∂Ω,
We concentrate on
(
Lu = f, x ∈ Ω,
u = 0, x ∈ ∂Ω,
where
We assume
• Ω ⊂ Rn , n = 2, 3, is a smooth or polygonal domain
• a = (aij ), where ai,j are continuous in Ω for 1 ≤ i, j ≤ n and
there is a a0 > 0 such that v > av ≥ a0 for all v ∈ Rn \ {0} and
x∈Ω
• b = (bi ) where bi is continuous in Ω
• c and f are continuous in Ω
Th
hK
K
U=0
Definition
The finite element method is:
Ax = b
R = AX − b
Ae = −R.
A> φ = ψ,
Definition
kφk is called the stability factor
e kRk
,ψ ≤ cond ψ (A) ,
kx k kbk
where the “weak” condition number is
Definition
The generalized Green’s function φ solves the weak adjoint
problem : Find φ ∈ H01 (Ω) such that
A∗ (v, φ) = (∇v, a∇φ)−(v, div (bφ))+(v, cφ) = (v, ψ) for all v ∈ H01 (Ω),
Definition
πh φ denotes an approximation of φ in Vh
Theorem
The error in the quantity of interest computed from the finite
element solution satisfies the error representation,
(e, ψ) = (f, φ−πh φ)−(a∇U, ∇(φ−πh φ))−(b·∇U, φ−πh φ)−(cU, φ−πh φ),
Definition
The approximate generalized Green’s function satisfies:
Compute Φ ∈ Vh2 such that
A∗ (v, Φ) = (∇v, a∇Φ)−(v, div (bΦ))+(v, cΦ) = (v, ψ) for all v ∈ Vh2 .
Example
An elliptic problem on the unit square Ω = (0, 1) × (0, 1)
(
−∆u = 200 sin(10πx) sin(10πy), (x, y) ∈ Ω,
u(x, y) = 0, (x, y) ∈ ∂Ω
u0
-1
x
y
1.2 3
error/estimate
error/estimate
0.8 2
0.4 1
0.0 0
102 103 104 0.0 0.1 1.0 10.0 100.0
Number of elements percent error
Definition
The mesh acceptance criterion is
X Z
(f − b · ∇U − cU )(Φ − πh Φ) − a∇U · ∇(Φ − πh Φ) dx ≤ TOL
K∈Th K
Definition
Two element acceptance criteria for the element indicators
are
TOL
max (f − b · ∇U − cU )(Φ − πh Φ) − a∇U · ∇(Φ − πh Φ) . ,
K |Ω|
or
Z
TOL
(f − b · ∇U − cU )(Φ − πh Φ) − a∇U · ∇(Φ − πh Φ) dx . ,
K M
Example
−∆u =
4800
2 2
−400((x−.5)2 +(y−.5)2 )
π 1 − 400((x − .5) + (y − .5) ) e ,
(x, y) ∈ Ω,
u(x, y) = 0, (x, y) ∈ ∂Ω,
350 0.08
0 0
-10
-20
-30
Mesh Element Contributions
140 0.08
0 0
-10
-20
-30
Mesh Element Contributions
1.10
1.05
error/estimate 1.00
0.95
0.90
0 50 100 150 200
percent error
where the last term is a line integral and n∂K denotes the
outward normal to ∂K .
X Z
(e, ψ) = − (∆U + f )(φ − πh φ) dx
K∈Th K
1
Z
− [∇U ] · n∂K (φ − πh φ) ds .
2 ∂K
Definition
The residual and corresponding adjoint or dual weights are
! !
k∆U + f kK kφ − πh φkK
RK = −1/2
, WK = 1/2
.
kh [∇U ]k∂K /2 kh (φ − πh φ)k∂K
Theorem
The error of the finite element approximation is bounded by
X
|(e, ψ)| ≤ RK · WK .
K∈Th
RK ≤ C|K|1/2 ,
k[∇U ]kσ = k∇U |K2 −∇U |K1 kσ ≤ k∇U |K2 −∇u|σ kσ +k∇u|σ −∇U |K1 kσ .
1 −1/2
kh [∇U ]k∂K ≤ C max |f | × |K|1/2 .
2 Ω
Duality, Adjoints, Green’s Functions – p. 166/304
Further analysis on the a posteriori error estimate
Elliptic problems also have the property that the strength of the
effect of a localized perturbation on a solution decays
significantly with the distance from the support of the
perturbation, at least in some directions.
Example
Consider the Green’s function for the Dirichlet problem for the
Laplacian −∆u = f on the ball Ω of radius r centered at the
origin in R3 .
r2 y
|y − x| ≤ 2
−x , x ∈ supp(δf ), y ∈ Ω \ supp(δf ).
|y|
We conclude that
r
max |δf | × volume of supp(δf ) × 1 + |y|
|δu(y)| ≤ .
4π × the distance from y to supp(δf )
Recall,
r2 x
1 ln |x|
−y
|x|2
, x 6= 0,
G(x; y) = × r|x−y|
2π ln r
|y| , x = 0.
∂Ω
∂ωδ
∂ω
We conclude,
Theorem
For any δ > 0 small enough that ωδ ⊂ Ω but large enough that
N (K) ∩ ω = ∅ for K ⊂ Ω \ ωδ , there is a constant C such that
X X X
2 α 1/2
kek1,ω ≤ Ckh D φkK |K| + C|hφ|1,N (K) |K|1/2 .
K⊂Ω\ωδ |α|=2 K∩ωδ 6=∅
Ch2K 1/2
kφ − πh φkK ≤ 2
|K| .
dist (K, ω)
To handle the second sum on the right, we use the basic stability
estimate,
kφk1,Ω ≤ kψk−1,Ω = kψk−1,ω = 1.
If we assume a uniform (small) size hK = h for elements such
that K ∩ ωδ 6= ∅, we obtain
X C X
ChK |φ|1,N (K) ≤ Chkφk1,Ω = Ch ≤ h|K|.
|ωδ |
K∩ωδ 6=∅ K∩ωδ 6=∅
We conclude
Theorem
For any δ > 0 small enough that ωδ ⊂ Ω but large enough that
N (K) ∩ ω = ∅ for K ⊂ Ω \ ωδ , there is a constant C such that
X Ch2K X
kek1,ω ≤ 2
|K| + Ch|K|.
dist (K, ω)
K⊂Ω\ωδ K∩ωδ 6=∅
The element indicators are Ch2K /dist (K, ω)2 respectively Ch,
and in an optimal mesh,
h2K
2
≈h or hK ≈ h1/2 × dist (K, ω), K ⊂ Ω \ ωδ .
dist (K, ω)
Definition
An effective domain of influence corresponding to the data ψ
is the region ωψ in which the corresponding elements must be
significantly smaller in size than the elements used in the
complement Ω \ ωψ in order to satisfy the adaptive goal.
If the data ψ has the property that the adjoint weight φ − πh φ has
a more-or-less uniform size throughout Ω, then the degree of
non-uniformity in an adapted mesh depends largely on the
spatial variation of the residual.
Definition
Suppose that {Ωi }N i=1 is a finite open cover of Ω. A Lipschitz
partition of unity subordinate to {Ωi } is a collection of functions
{pi }N
i=1 with the properties
N
X
supp (pi ) ⊂ Ωi , 1 ≤ i ≤ N, pi (x) = 1, x ∈ Ω,
i=1
pi is continuous on Ω and differentiable on Ωi , 1 ≤ i ≤ N,
kpi kL∞ (Ω) ≤ C and k∇pi kL∞ (Ωi ) ≤ C/diam (Ωi ), 1 ≤ i ≤ N,
Definition
The quantities {(U, ψpi )} corresponding to the data {ψi = ψpi }
are called the localized information corresponding to the
partition of unity.
Compute Ûi ∈ V̂i such that A(Ûi , v) = (f, v) for all v ∈ V̂i ,
We hope that this will require a locally refined mesh because the
corresponding data has localized support.
Definition
The partition of unity approximation is defined by
PN
Up = i=1 Ui pi , which is in the partition of unity finite element
space
N
(N )
X X
Vp = Vi p i = vi pi : vi ∈ Vi .
i=1 i=1
Note
N
X N
X N
X
Up = Ui pi = χi Ûi pi ≡ Ûi pi .
i=1 i=1 i=1
Find φi ∈ H01 (Ω) such that A∗ (v, φi ) = (v, ψi ) for all v ∈ H01 (Ω).
N
X
(u − Up , ψ) = (u − Ui )pi , ψ .
i=1
Theorem
The error of the partition of unity finite element solution Up
satisfies the error representation,
N
X
(u − Up , ψ) = (f, φi − πi φi ) − (a∇Ûi , ∇(φi − πi φi ))
i=1
− (b · ∇Ûi , φi − πi φi ) − (cÛi , φi − πi φi ) .
Definition
The approximate generalized Green’s functions solve:
Compute Φi ∈ Vi2 such that
Two approaches:
Approach 1: A Global Computation
Find one triangulation such that the corresponding finite
element solution satisfies |(e, ψi )| ≤ TOLi , for 1 ≤ i ≤ N .
Approach 2: A Decomposed Computation
Find N independent triangulations and finite element
solutions Ui so that the errors satisfy |(ei , ψi )| ≤ TOLi , for
1 ≤ i ≤ N.
or
Z
Ei |K = (f − b · ∇Ûi − cÛi )(Φi − πi Φi ) − a∇Ûi · ∇(Φi − πi Φi ) dx,
K
Definition
We let Ei (x) denote the piecewise constant element error
indicator function associated to data ψi with Ei (x) ≡ Ei |K for
K ∈ T0 .
Definition
The (cross-)correlation of two functions f ∈ Lp (Ω) and
g ∈ Lq (Ω), defined as:
Z
(f ◦ g)(y) = f (x)g(y + x) dx,
Ω
Definition
The correlation indicator c(Ei , Ej ) is
Definition
We say that the effective domain of influence associated to ψi is
significantly correlated to the domain of influence associated
to ψj if two conditions hold.
Definition
We say that the effective domain of influence associated to ψi is
significantly correlated to the domain of influence associated
to ψj if two conditions hold.
Condition 1
The correlation of Ei and Ej is larger than a fixed fraction of the
norm of Ej , or
c(Ei , Ej )
Correlation Ratio 1 = 2
≥ γ1 ,
kEj k
Element Indicator
Element Indicator
Element Indicator
Ei
Ej Ei
Ej
Ei
Ej
Element Element Element
Definition
We say that the effective domain of influence associated to ψi is
significantly correlated to the domain of influence associated
to ψj if two conditions hold.
Condition 2
The component of Ej orthogonal to Ei is smaller than a fixed
fraction of the norm of Ej , or
c(Ej ,Ei )
Ej − kEi k2 Ei
Correlation Ratio 2 = ≤ γ2 ,
kEj k
This corrects for the difficulties that arise when Ei is much larger
than Ej .
Duality, Adjoints, Green’s Functions – p. 212/304
Identifying significant correlations
Element Indicator
Ej
Ei
Element
E1 E6
3
E2 E7
E3 E8
E4 E9
2
E5
0
1 9 18 27 36
Element Number
Definition
The Final Element Ratio is the number of elements in the final
mesh refinement level required to achieve accuracy in the
specified quantities of interest in the Global Computation to the
maximum number of elements in the final mesh refinement
levels for the individual computations in the Decomposed
Computation.
Example 1
We solve
(
1
− 10π 2 ∆u(x) = sin(πx) sin(πy), (x, y) ∈ Ω,
u(x, y) = 0, (x, y) ∈ ∂Ω,
Numerical solutions.
6 7 10 11
1 3
5 8 9 12
2 3 14 15
2 4
1 4 13 16
Example 2
We solve
−∇ · (1.1 + sin(πx) sin(πy))∇u(x, y)
= −3 cos2 (πx) + 4 cos2 (πx) cos2 (πx)
+2.2 sin(πx) sin(πy) + 2 − 3 cos 2 (πy), (x, y) ∈ Ω,
u(x, y) = 0, (x, y) ∈ ∂Ω,
We use
400 −400((x−cx )2 +(y−cy )2 )
δ̂(cx ,cy ) = e
π
to approximate the delta function δ(cx ,cy ) .
ψ1 ψ2
Lev. Elt’s Est. Err. Rat. Est. Err. Rat.
1 64 .035 .035 1.0 .090 .29 3.3
2 201 .0088 .0089 1.0 .042 .082 1.9
3 763 .0027 .0027 1.0 .020 .020 .99
4 2917 .00044 .00044 1.0 .0050 .00504 1.0
ψ3
Lev. Elt’s Est. Err. Rat.
1 64 .24 .022 .091
2 201 .0024 .014 6.0
3 763 .0020 .0020 1.0
4 2917 .0049 .00504 1.0
We vary the initial meshes; using 7 × 7 for Û1 ; 9 × 9 for Û2 and
Û4 ; and 12 × 12 for Û3 and Û5 .
^ ^
Final Mesh for U ^
Final Mesh for U
Final Mesh for U1 2 3
Example 3
We solve
(
−∆u = 16(y − y 2 + x − x2 ) (x, y) ∈ Ω,
u(x, y) = 0, (x, y) ∈ ∂Ω,
Since the domain is small and the solution and the generalized
Green’s function are very smooth, the gain from decomposing
the solution is greatly reduced.
Final meshes.
We consider the error in the average value and the point values
at (.25, .25) and (.5, .5).
1.0 1.0
Corr. Rat. 1: E1 on E2
Corr. Rat. 1: E1 on E3
Corr. Rat. 2: E1 on E2
Corr. Rat. 2: E1 on E3
Corr. Rat. 1: E2 on E1
Corr. Rat. 1: E3 on E1
Corr. Rat. 2: E2 on E1
0.5 0.5 Corr. Rat. 2: E3 on E1
0.0 0.0
0 50 100 150 0 50 100 150
Number of Elements Number of Elements
10 4
Corr. Rat. 1: E1 on E5
Corr. Rat. 2: E1 on E5 Corr. Rat. 1: E1 on E6
Corr. Rat. 1: E5 on E1 Corr. Rat. 2: E1 on E6
5 2
Corr. Rat. 2: E5 on E1 Corr. Rat. 1: E6 on E1
Corr. Rat. 2: E6 on E1
0 0
0 100 200 300 400 0 100 200 300 400
Number of Elements Number of Elements
1.2 3
Corr. Rat. 1: E3 on E5 Corr. Rat. 1: E5 on E3
Corr. Rat. 2: E3 on E5 Corr. Rat. 2: E5 on E3
2
Corr. Rat. 1: E5 on E6
Corr. Rat. 2: E5 on E6
0.6
Corr. Rat. 1: E6 on E5
Corr. Rat. 2: E6 on E5 1
0.0 0
0 50 100 150 0 100 200 300 400
Number of Elements Number of Elements
Example 4
We solve
2 2
−∇ · .05 + tanh 10(x − 5) + 10(y − 1) ! ∇u
−100
+ · ∇u = 1, (x, y) ∈ Ω,
0
u(x, y) = 0, (x, y) ∈ ∂Ω,
Diffusion 1 10
x
2
y 0
Final Mesh.
Pe=1000
Pe=.1
Plots of the meshes for the original problem with P e = 1000 and
a problem with P e = .1.
11 12 13 14 15 16 17 18 19 20
1 2 3 4 5 6 7 8 9 10
Significant Correlations:
^
Final Mesh for U1
^
Final Mesh for U5
Plots of the final meshes for the localized solutions Û1 and Û5 .
Duality, Adjoints, Green’s Functions – p. 256/304
Examples
^
Final Mesh for U9
x 10 -4 x 10 -4
6 6
4 4
2 2
0 10 0 10
8 8
2 6 2 6
4 x 4 x
1 2 1 2
y y
0 0 0 0
Example 5
We solve
(
1 −5((x−.5) 2
+(y−2.5) 2
) , (x, y) ∈ Ω,
− π2 ∆u = 2 + 4e
u(x, y) = 0, (x, y) ∈ ∂Ω,
The initial mesh has 48 elements and the error tolerance is 1%.
6 1.6
5 1.2
4
3 0.8
2 0.4
1 3 3
2 2
0 0
1 y 1 1 y
1
2 2
3 0
x 3 0 x
7 6 5
8 4
1 2 3
^ ^
Final Mesh for U3 Final Mesh for U4
Final meshes.
^ ^
Final Mesh for U6 Final Mesh for U7
Final meshes.
1.4
1
0.6
3 0.2 3
2 2
0 0
1 1 y 1 1 y
2 2
x 3 0 x 3 0
f (x) = b
where the data b, nonlinearity f , and the solution x all have the
same dimension.
We get
Ãe = −R
with
Z 1
à = f 0 (sx + (1 − s)X) ds.
0
We obtain
Definition
A subset A of a vector space is convex if for any a, b ∈ A, the
set of points on the “line segment” joining a and b, i.e.,
{sa + (1 − s)b| 0 ≤ s ≤ 1} is contained in A.
F (e) = f (u + e) − f (u),
The domain of F is
Definition
An operator A∗ (e) with domain D(A∗ ) ⊂ Y ∗ and range in X ∗ is
an adjoint operator corresponding to F if
Example
Suppose that F can be represented as F (e) = A(e)e, where
A(e) is a linear operator with D(F ) ⊂ D(A).
Example
Let (t, x) ∈ Ω = (0, 1) × (0, 1), with X = X ∗ = Y = Y ∗ = L2
denoting the space of periodic functions in t and x, with period
equal to 1.
∂e ∂e
F (e) = +e + ae = f
∂t ∂x
where a > 0 is a constant and the domain of F is the set of
continuously differentiable functions.
∂v ∂v
A1 (e)v = +e + av
∂t ∂x
∂v ∂e
A2 (e)v = + a+ v
∂t ∂x
∂v 1 ∂(ev)
A3 (e)v = + + av.
∂t 2 ∂x
∂w ∂(ew)
A∗1 (e)w=− − + aw
∂t ∂x
∗ ∂w ∂e
A2 (e)w = − + a+ w
∂t ∂x
∗ ∂w e ∂w
A3 (e)w = − − + aw.
∂t 2 ∂x
We rewrite this as
with
Z 1
A(e) = f 0 (u + se) ds.
0
Definition
For a fixed e, the adjoint operator A∗ (e), defined in the usual way
for the linear operator A(e), is said to be an adjoint for F .
Example
Continuing the previous example,
0 ∂v ∂v ∂e
F (e)v = +e + a+ v.
∂t ∂x ∂x
Assumptions:
• Ω is a convex polygonal domain in R2 with boundary ∂Ω
• = (i ) and f = (fi ) have smooth second derivatives
uo = u − up
Time
Ω Space
A space-time discretization.
We define
q
X
Wnq = w(x, t) : w(x, t) = tj vj (x), vj ∈ Vn , (x, t) ∈ Sn .
j=0
[w]n = wn+ − wn− , where wn± = lims→tn ± w(s), denotes the jump
across tn
Definition
The continuous Galerkin cG(q) approximation U ∈ W q
satisfies U0− = P0 u0 and for n ≥ 1, the Galerkin orthogonality
relation
Z t Z tn
n
(U̇i , vi ) + (i (U )∇Ui , ∇vi ) dt = (fi (U ), vi ) dt
tn−1 tn−1
for all v ∈ Wnq−1 , 1 ≤ i ≤ D,
U + = P U − .
n−1 n n−1
Definition
The discontinuous Galerkin dG(q) approximation U ∈ W q
satisfies U0− = P0 u0 and for n ≥ 1,
Z tn
[Ui ]n−1 , vi+
(U̇i , vi ) + (i (U )∇Ui , ∇vi ) dt +
tn−1
Z tn
= (fi (U ), vi ) dt
tn−1
for all v ∈ Wnq , 1 ≤ i ≤ D.
Example
We discretize the scalar problem
u̇ − ∆u = f (u), (x, t) ∈ Ω × R +,
u(x, t) = 0, (x, t) ∈ ∂Ω × R+ ,
u(x, 0) = u0 (x), x ∈ Ω,
Un satisfies
~ n − F~ (U − )kn = Bn,n−1 U
~ n−1 ,
Bn + kn An U n n ≥ 1,
Example
If Mn is constant and the lumped mass quadrature is used to
evaluate the coefficients of Bn and Bn,n−1 = Bn , then the
resulting set of equations for the dG(0) approximation is the
same as the equations for the nodal values of the backward
Euler difference scheme.
Example
In the case of the scalar problem with constant diffusion, the
adjoint problem is
¯
−φ̇ − ∆φ − f φ = ψ, (x, t) ∈ Ω × (tn , 0],
φ(x, t) = 0, (x, t) ∈ ∂Ω × (tn , 0],
φ(x, tn ) = 0, x ∈ Ω.
Example
In the case of one parabolic equation with nonlinear diffusion
coupled to one ordinary differential equation, the dual problem is
¯11 φ1 − f¯12 φ2 = ψ1 ,
−φ̇ 1 − ∇ ·
¯1 ∇φ 1 + β̄11 ∇φ1 − f
−φ̇ + β̄ ∇φ − f¯ φ − f¯ φ = ψ ,
2 12 1 21 1 22 2 2
φ1 (x, t) = 0,
φ(x, tn ) = 0.