Functional Analysis and Applied Optimization in Banach Spaces
Functional Analysis and Applied Optimization in Banach Spaces
Functional Analysis and Applied Optimization in Banach Spaces
Functional Analysis
and Applied
Optimization in
Banach Spaces
Applications to Non-Convex Variational
Models
Functional Analysis and Applied Optimization
in Banach Spaces
Fabio Botelho
Functional Analysis
and Applied Optimization
in Banach Spaces
Applications to Non-Convex Variational
Models
123
Fabio Botelho
Department of Mathematics and Statistics
Federal University of Pelotas
Pelotas, RS-Brazil
The first objective of this work is to present, to some extent, a deep introduction
to the basic concepts on real and functional analysis.
In principle, the text is written for applied mathematicians and postgraduate stu-
dents in applied mathematics, with interest in applications of functional analysis,
calculus of variations, and optimization to problems in physics and engineering.
However, engineers, physicists, and other professionals in related areas may
find the text very interesting by the possibility of background development towards
graduate-level mathematics applicable in their respective work fields.
We have proven almost all results presented. The proofs are rigorous, but we
believe are almost all very clear and relatively easy to read, even at the most complex
text parts.
The material presented in Parts I and II concerns standard real and functional
analysis. Hence in these two parts the results in general are not new, with the excep-
tion of some sections on domains of class Ĉ1 and relating Sobolev spaces and some
sections about Lagrange multiplier results and the basic theorem about relaxation
for the scalar case, where we show a different proof concerning the original one
in the book Convex Analysis and Variational Problems (indeed such a book is the
theoretical base of the present work) by Ekeland and Témam’s.
About the basic part, specifically Chaps. 1–3 correspond to standard functional
analysis. In Chaps. 4–6 we present basic and advanced concepts in measure and in-
tegration which will be relevant in subsequent results (in fact perhaps a little more
than the minimum necessary). Moreover, Chaps. 7 and 8 correspond to a basic expo-
sition on Sobolev spaces and again, the fundamental results presented are relevant
for subsequent developments. In Chaps. 9–11 we introduce some basic and more
advanced concepts on calculus of variations, convex analysis, and optimization.
Finally, the applications presented in Chaps. 12–23 correspond to the work of
the present author along the last years, and almost all results including the applica-
tions of duality for micro-magnetism, composites in elasticity, and conductivity and
phase transitions are extensions and natural developments of prior ones presented
in the author’s Ph.D. thesis at Virginia Tech, USA, and the previous book Topics
on Functional Analysis, Calculus of Variations and Duality published by Academic
v
vi Preface
Publications. The present book overlaps to some extent with the previous one just
on a part concerning standard mathematics. The applications in the present one are
almost all new developments.
Anyway, a key feature of the present work is that while all problems studied
here are nonlinear with corresponding non-convex variational formulation, it has
been almost always possible to develop convex (in fact concave) dual variational
formulations, which in general are more amenable to numerical computations.
The section on relaxation for the vectorial case, as its title suggests, presents du-
ality principles that are valid even for vectorial problems. It is worth noting that
such results were used in this text to develop concave dual variational formulations
in situations such as for conductivity in composites and vectorial examples in phase
transitions. In Chap. 15 we present the generalized method of lines, a numerical pro-
cedure in which the solution of the partial differential equation in question is written
on lines as functions of boundary conditions and boundary shape. In Chap. 22 we
develop some examples concerning the Navier–Stokes system.
Chapter 12 develops duality for a model in finite elasticity. The dual formulations
obtained allow the matrix of stresses to be nonpositive or nonnegative definite. This
is, in some sense, an extension of earlier results (which establish the complementary
energy as a perfect global optimization duality principle only if the stress tensor is
positive definite at the equilibrium point). The results are based on standard tools of
convex analysis and the concept of the Legendre transform.
Chapters 14–16 are concerned with existence theory and the development of dual
variational formulations for Ginzburg–Landau-type equations. Since the primal for-
mulations are non-convex, we use specific results for distance between two convex
functions to obtain the dual approaches. Note that we obtain a convex dual formula-
tion for the simpler real case. For such a formulation optimality conditions are also
established.
The main focus of Chaps. 17 and 18 is the development of dual variational formu-
lations for multi-well optimization problems in phase transitions, conductivity, and
elasticity. The primal formulation may not have minimizers in the classical sense.
In this case, the solution through the dual formulation is a weak limit of minimizing
sequences for the original problem.
The first part of Chap. 20 develops a dual variational formulation for the optimal
design of a plate of variable thickness. The design variable, namely the plate thick-
ness, is supposed to minimize the plate deformation work due to a given external
load. The second part is concerned with the optimal design for a two-phase problem
in elasticity. In this case, we are looking for the mixture of two constituents that
minimizes the structural internal work. In both applications the dual formulations
were obtained through basic tools of convex analysis. Finally, we highlight that this
chapter is coauthored by myself and my colleague Professor Alexandre Molter.
viii Preface
Chapter 23 develops duality for the optimal control and design of a beam model.
We emphasize the dual formulation is useful to obtain numerical results. Finally,
numerical examples of optimal design are provided, concerning the maximization
of buckling load and fundamental frequency, respectively.
Acknowledgments
xi
xii Contents
7 Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167
7.1 Basic Definitions and Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 167
7.2 Differentiation of Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 171
7.3 Examples of Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172
7.3.1 First Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172
7.3.2 Second Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172
7.3.3 Third Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172
Contents xiii
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 553
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 557
Acronyms
xvii
xviii Acronyms
∃ Exists symbol
Cm (Ω ) Space of real functions on Ω that are continuously differ-
entiable up to order m, 0 ≤ m ≤ ∞
Ccm (Ω ) Set of functions in Cm (Ω ) with compact support in Ω
D(Ω ),Cc∞ (Ω ) Set of functions in C∞ (Ω ) with compact support in Ω
L p (Ω ) Space of measurable functions whose p-power of their
absolute values is finite integrable
W m,p (Ω ) Sobolev space {u ∈ L p (Ω ), Dα u ∈ L p (Ω ), |α | ≤ m, 1 ≤
p ≤ ∞}, where derivatives are in the distributional sense
L p (Ω ; Rn ) n-Dimensional L p
W m,p (Ω ; Rn ) n-Dimensional W m,p
Part I
Basic Functional Analysis
Chapter 1
Topological Vector Spaces
1.1 Introduction
The main objective of this chapter is to present an outline of the basic tools of
analysis necessary to develop the subsequent chapters. We assume the reader has
a background in linear algebra and elementary real analysis at an undergraduate
level. The main references for this chapter are the excellent books on functional
analysis: Rudin [58], Bachman and Narici [6], and Reed and Simon [52]. All proofs
are developed in details.
Definition 1.2.1 (Vector Spaces). A vector space over F is a set which we will de-
note by U whose elements are called vectors, for which are defined two operations,
namely, addition denoted by (+) : U ×U → U and scalar multiplication denoted by
(·) : F × U → U, so that the following relations are valid:
1. u + v = v + u, ∀u, v ∈ U,
2. u + (v + w) = (u + v) + w, ∀u, v, w ∈ U,
3. there exists a vector denoted by θ such that u + θ = u, ∀u ∈ U,
4. for each u ∈ U, there exists a unique vector denoted by
−u such that u + (−u) = θ ,
5. α · (β · u) = (α · β ) · u, ∀α , β ∈ F, u ∈ U,
6. α · (u + v) = α · u + α · v, ∀α ∈ F, u, v ∈ U,
7. (α + β ) · u = α · u + β · u, ∀α , β ∈ F, u ∈ U,
8. 1 · u = u, ∀u ∈ U.
Remark 1.2.2. From now on we may drop the dot (·) in scalar multiplications and
denote α · u simply as α u.
Remark 1.2.6. When necessary, to clarify the notation, we shall denote the vector
space U endowed with the topology σ by (U, σ ).
A \ B = {u ∈ A | u ∈ B}.
Proof.
1. Since 0/ is open and U = 0/ c , by Definition 1.2.7, U is closed. Similarly, since U
is open and 0/ = U \ U = U c , 0/ is closed.
2. A, B closed implies that Ac and Bc are open, and by Definition 1.2.5, Ac ∪ Bc is
open, so that A ∩ B = (Ac ∪ Bc )c is closed.
1.2 Vector Spaces 5
Remark 1.2.11. From Proposition 1.2.9 item 3 we have that Ā is the smallest closed
set that contains A, in the sense that if C is closed and A ⊂ C, then Ā ⊂ C.
Proposition 1.2.15. If A ⊂ U is a set such that for each u ∈ A there exists a neigh-
borhood Vu u such that Vu ⊂ A, then A is open.
Proof. This follows from the fact that A = ∪u∈A Vu and any arbitrary union of open
sets is open.
Definition 1.2.16 (Function). Let U and V be two topological spaces. We say that
f : U → V is a function if f is a collection of pairs (u, v) ∈ U ×V such that for each
u ∈ U there exists only one v ∈ V such that (u, v) ∈ f .
f −1 (V ) = {u ∈ U | f (u) ∈ V }. (1.2)
Definition 1.2.20. We say that (U, σ ) is a Hausdorff topological space if, given u1 ,
u2 ∈ U, u1 = u2 , there exists V1 , V2 ∈ σ such that
u1 ∈ V1 , u2 ∈ V2 and V1 ∩ V2 = 0.
/ (1.3)
and
V1 + V2 ⊂ V (1.7)
u0 + U + U ⊂ u0 + V ⊂ U \ {u1} (1.8)
so that
u0 + v1 + v2 = u1 , ∀v1 , v2 ∈ U , (1.9)
or
u0 + v1 = u1 − v2 , ∀v1 , v2 ∈ U , (1.10)
and since U = −U
(u0 + U ) ∩ (u1 + U ) = 0.
/ (1.11)
is said to be convex.
Definition 1.3.4 (Locally Convex Spaces). A topological vector space U is said to
be locally convex if there is a local base Ω whose elements are convex.
8 1 Topological Vector Spaces
Proof.
1. Suppose U is a neighborhood of zero. From the continuity of scalar multiplica-
tion, there exist V (neighborhood of zero) and δ > 0, such that α V ⊂ U when-
ever |α | < δ . Define W = ∪|α |<δ α V ; thus W ⊂ U is a balanced neighborhood
of zero.
2. Suppose U is a convex neighborhood of zero in U. Define
so that
βu ∈ u+V , (1.16)
or
and thus
un ∈ V , ∀n ≥ N.
∩α ∈L Kα = 0.
/
That is,
Kα0 ∩ [∩αα = α0
∈L Kα ] = 0.
/
Thus,
∩αα = α0
∈L Kα ⊂ Kα0 ,
c
so that
Kα0 ⊂ [∩αα = α0
∈L Kα ] ,
c
α =α
Kα0 ⊂ [∪α ∈L 0 Kαc ].
However, Kα0 is compact and Kαc is open, ∀α ∈ L.
Hence, there exist α1 , . . . , αn ∈ L such that
The idea here is to prepare a route for the study of Banach spaces defined below.
We start with the definition of norm.
Definition 1.5.1 (Norm). A vector space U is said to be a normed space, if it is pos-
sible to define a function · U : U → R+ = [0, +∞), called a norm, which satisfies
the following properties:
1. u U > 0, if u = θ and u U = 0 ⇔ u = θ ,
2. u + v U ≤ u U + v U , ∀ u, v ∈ U,
3. α u U = |α | u U , ∀u ∈ U, α ∈ F.
Now we present the definition of metric.
Definition 1.5.2 (Metric Space). A vector space U is said to be a metric space if it
is possible to define a function d : U × U → R+ , called a metric on U, such that
1. 0 ≤ d(u, v), ∀u, v ∈ U,
2. d(u, v) = 0 ⇔ u = v,
3. d(u, v) = d(v, u), ∀u, v ∈ U,
4. d(u, w) ≤ d(u, v) + d(v, w), ∀u, v, w ∈ U.
A metric can be defined through a norm, that is,
d(u, v) = u − v U . (1.19)
The following are some basic definitions concerning metric and normed spaces:
Definition 1.5.3 (Convergent Sequences). Given a metric space U, we say that
{un } ⊂ U converges to u0 ∈ U as n → ∞, if for each ε > 0, there exists n0 ∈ N,
such that if n ≥ n0 , then d(un , u0 ) < ε . In this case we write un → u0 as n → +∞.
Definition 1.5.6 (Limit Point). Let (U, d) be a metric space and let E ⊂ U. We say
that v ∈ U is a limit point of E if for each r > 0 there exists w ∈ Br (v) ∩ E such that
w = v.
12 1 Topological Vector Spaces
Definition 1.5.7 (Interior Point, Topology for (U, d)). Let (U, d) be a metric space
and let E ⊂ U. We say that u ∈ E is interior point if there exists r > 0 such that
Br (u) ⊂ E. We may define a topology for a metric space (U, d) by declaring as
open all set E ⊂ U such that all its points are interior. Such a topology is said to be
induced by the metric d.
Definition 1.5.8. Let (U, d) be a metric space. The set σ of all open sets, defined
through the last definition, is indeed a topology for (U, d).
Proof.
1. Obviously 0/ and U are open sets.
2. Assume A and B are open sets and define C = A ∩ B. Let u ∈ C = A ∩ B; thus,
from u ∈ A, there exists r1 > 0 such that Br1 (u) ⊂ A. Similarly from u ∈ B there
exists r2 > 0 such that Br2 (u) ⊂ B.
Define r = min{r1 , r2 }. Thus, Br (u) ⊂ A ∩ B = C, so that u is an interior point of
C. Since u ∈ C is arbitrary, we may conclude that C is open.
3. Suppose {Aα , α ∈ L} is a collection of open sets. Define E = ∪α ∈L Aα , and we
shall show that E is open.
Choose u ∈ E = ∪α ∈L Aα . Thus there exists α0 ∈ L such that u ∈ Aα0 . Since Aα0
is open there exists r > 0 such that Br (u) ⊂ Aα0 ⊂ ∪α ∈L Aα = E. Hence u is an
interior point of E, since u ∈ E is arbitrary, we may conclude that E = ∪α ∈L Aα
is open.
The proof is complete.
Definition 1.5.9. Let (U, d) be a metric space and let E ⊂ U. We define E as the set
of all the limit points of E.
Theorem 1.5.10. Let (U, d) be a metric space and let E ⊂ U. Then E is closed if
and only if E ⊂ E.
Proof. Suppose E ⊂ E. Let u ∈ E c ; thus u ∈ E and u ∈ E . Therefore there exists
r > 0 such that Br (u) ∩ E = 0, / so that Br (u) ⊂ E c . Therefore u is an interior point
of E . Since u ∈ E is arbitrary, we may infer that E c is open, so that E = (E c )c is
c c
closed.
Conversely, suppose that E is closed, that is, E c is open.
If E = 0,
/ we are done.
Thus assume E = 0/ and choose u ∈ E . Thus, for each r > 0, there exists v ∈
Br (u) ∩ E such that v = u. Thus Br (u) E c , ∀r > 0 so that u is not a interior point
of E c . Since E c is open, we have that u ∈ E c so that u ∈ E. We have thus obtained,
u ∈ E, ∀u ∈ E , so that E ⊂ E.
The proof is complete.
Remark 1.5.11. From this last result, we may conclude that in a metric space, E ⊂ U
is closed if and only if E ⊂ E.
Definition 1.5.12 (Banach Spaces). A normed vector space U is said to be a Banach
space if each Cauchy sequence related to the metric induced by the norm converges
to an element of U.
1.6 Compactness in Metric Spaces 13
Definition 1.6.2. Let (U, d) be a metric space. We say that {Fk } ⊂ U is a nested
sequence of sets if
F1 ⊃ F2 ⊃ F3 ⊃ . . . .
Theorem 1.6.3. If (U, d) is a complete metric space, then every nested sequence of
nonempty closed sets {Fk } such that
lim diam(Fk ) = 0
k→+∞
∩∞
k=1 Fk = 0.
/
Proof. Suppose {Fk } is a nested sequence and lim diam(Fk ) = 0. For each n ∈ N,
k→∞
select un ∈ Fn . Suppose given ε > 0. Since
lim diam(Fn ) = 0,
n→∞
diam(Fn ) < ε .
d(un , um ) < ε .
Hence {un } is a Cauchy sequence. Being U complete, there exists u ∈ U such that
un → u as n → ∞.
14 1 Topological Vector Spaces
u ∈ F̄m = Fm .
u ∈ ∩∞
m=1 Fm .
d(u, v) > K.
Observe that
A ⊂ ∪u∈A B1 (u).
Since A is compact there exists u1 , u2 , . . . , un ∈ A such that
A =⊂ ∪nk=1 B1 (uk ).
Define
R = max{d(ui , u j ) | i, j ∈ {1, . . . , n}}.
Choose u, v ∈ A such that
u ∈ B1 (ui ), v ∈ B1 (u j ).
Thus
d(u, v) < ε .
1.6 Compactness in Metric Spaces 15
Thus
Theorem 1.6.9. Let (U, d) be a metric space. If from each sequence {un } ⊂ A we
can select a convergent subsequence {unk }, then A is totally bounded.
Proof. Suppose, to obtain contradiction, that A is not totally bounded. Thus there
exists ε0 > 0 such that there exists no ε0 -net with respect to A. Choose u1 ∈ A; hence
{u1 } is not a ε0 -net, that is, there exists u2 ∈ A such that
d(u1 , u2 ) > ε0 .
Again {u1 , u2 } is not a ε0 -net for A, so that there exists u3 ∈ A such that
Definition 1.6.10 (Sequentially Compact Sets). Let (U, d) be a metric space. A set
A ⊂ U is said to be sequentially compact if for each sequence {un } ⊂ A, there exist
a subsequence {unk } and u ∈ A such that
unk → u, as k → ∞.
δ (un ) → δ0 as n → ∞.
δ (u0 )
d(unk , u0 ) < . (1.25)
4
We claim that
δ (u0 )
δ (unk ) ≥ , if k > K0 .
4
To prove the claim, suppose
Since
δ (u0 )
< δ (u0 ),
2
there exists some α1 ∈ L such that
However, since
δ (u0 )
d(unk , u0 ) < , if k > K0 ,
4
we obtain
B δ (u0 ) (u0 ) ⊃ B δ (u0 ) (unk ), if k > K0 ,
2 4
so that
δ (u0 )
δ (unk ) ≥ , ∀k > K0 .
4
Therefore
δ (u0 )
lim δ (unk ) = δ0 ≥ .
k→∞ 4
Choose ε > 0 such that
δ0 > ε > 0.
From the last theorem since A is sequentially compact, it is totally bounded. For the
ε > 0 chosen above, consider an ε -net contained in A (the fact that the ε -net may be
chosen contained in A is also a consequence of the last theorem) and denote it by N
that is,
N = {v1 , . . . , vn } ∈ A.
Since δ0 > ε , there exists
α1 , . . . , αn ∈ L
such that
Bε (vi ) ⊂ Gαi , ∀i ∈ {1, . . . , n},
considering that
δ (vi ) ≥ δ0 > ε > 0, ∀i ∈ {1, . . . , n}.
For u ∈ A, since N is an ε -net we have
A ⊂ ∪ni=1 Gαi .
18 1 Topological Vector Spaces
Thus
{Gα1 , . . . , Gαn }
is a finite subcover for A of
{Gα , α ∈ L}.
Hence, A is compact.
The proof is complete.
Proof. Suppose A is relatively compact. Thus A is compact so that from the last
theorem, A is sequentially compact.
Thus from each sequence in A we may select a subsequence which converges
to some element of A. In particular, for each sequence in A ⊂ A, we may select a
subsequence that converges to some element of A.
Conversely, suppose that for each sequence in A, we may select a convergent sub-
sequence. It suffices to prove that A is sequentially compact. Let {vn } be a sequence
in A. Since A is dense in A, there exists a sequence {un } ⊂ A such that
1
d(un , vn ) < .
n
From the hypothesis we may obtain a subsequence {unk } and u0 ∈ A such that
unk → u0 , as k → ∞.
Thus,
vnk → u0 ∈ A, as k → ∞.
Therefore A is sequentially compact so that it is compact.
Proof.
1. Suppose A ⊂ U is relatively compact. From the last theorem, from each sequence
in A, we can extract a convergent subsequence. From Theorem 1.6.9, A is totally
bounded.
2. Let (U, d) be a metric space and let A be a totally bounded subset of U.
Let {un } be a sequence in A. Since A is totally bounded for each k ∈ N we find a
εk -net where εk = 1/k, denoted by Nk where
(k) (k) (k)
Nk = {v1 , v2 , . . . , vnk }.
1.7 The Arzela–Ascoli Theorem 19
In particular for k = 1 {un } is contained in the 1-net N1 . Thus at least one ball
of radius 1 of N1 contains infinitely many points of {un }. Let us select a subse-
(1)
quence {unk }k∈N of this infinite set (which is contained in a ball of radius 1).
(2)
Similarly, we may select a subsequence here just partially relabeled {unl }l∈N
(1)
of {unk } which is contained in one of the balls of the 12 -net. Proceeding in this
(k)
fashion for each k ∈ N we may find a subsequence denoted by {unm }m∈N of the
original sequence contained in a ball of radius 1/k.
(k)
Now consider the diagonal sequence denoted by {unk }k∈N = {zk }. Thus
2
d(zn , zm ) < , if m, n > k,
k
that is, {zk } is a Cauchy sequence, and since (U, d) is complete, there exists u ∈ U
such that
zk → u as k → ∞.
From Theorem 1.6.12, A is relatively compact.
The proof is complete.
Proof. Let {un } be a countable dense set in (U, d). By hypothesis, { fn (u1 )} is a
bounded sequence; therefore, it has a convergent subsequence, which is denoted by
{ fnk (u1 )}. Let us denote
fnk (u1 ) = f˜1,k (u1 ), ∀k ∈ N.
20 1 Topological Vector Spaces
f˜1,k (u1 ) → g1 , as k → ∞.
Observe that { fnk (u2 )} is also bounded and also it has a convergent subsequence,
which similarly as above we will denote by { f˜2,k (u2 )}. Again there exists g2 ∈ C
such that
f˜2,k (u1 ) → g1 , as k → ∞.
f˜2,k (u2 ) → g2 , as k → ∞.
Proceeding in this fashion for each m ∈ N we may obtain { f˜m,k } such that
where the set {g1 , g2 , . . . , gm } is obtained as above. Consider the diagonal sequence
{ f˜k,k },
is such that
f˜k,k (um ) → gm ∈ C, as k → ∞, ∀m ∈ N.
Therefore we may conclude that from { fn } we may extract a subsequence also de-
noted by
{ fnk } = { f˜k,k }
which is convergent in
E = {un }n∈N .
Now suppose K ⊂ U, being K compact. Suppose given ε > 0. From the equiconti-
nuity hypothesis there exists δ > 0 such that if u, v ∈ U and d(u, v) < δ we have
ε
| fnk (u) − fnk (v)| < , ∀k ∈ N.
3
Observe that
K ⊂ ∪u∈K B δ (u),
2
K ⊂ ∪M
j=1 B δ (ũ j ).
2
lim fnk (v j )
k→∞
ε
| fnk (v j ) − fnl (v j )| < .
3
Pick u ∈ K; thus
u ∈ B δ (ũ jˆ)
2
Therefore if
k, l > max{K01 , . . . , K0M },
then
| fnk (u) − fnl (u)| ≤ | fnk (u) − fnk (v jˆ)| + | fnk (v jˆ) − fnl (v jˆ)|
+| fnl (v jˆ) − fnl (u)|
ε ε ε
≤ + + = ε. (1.26)
3 3 3
Since u ∈ K is arbitrary, we conclude that { fnk } is uniformly Cauchy on K.
The proof is complete.
Definition 1.8.2 (Null Space and Range). Given f : U → V , we define the null
space and the range of f, denoted by N( f ) and R( f ), respectively, as
N( f ) = {u ∈ U | f (u) = θ } (1.30)
and
f (U ) ⊂ V . (1.32)
Thus
or
1. f is continuous,
2. f is bounded,
3. if un → θ , then { f (un )} is bounded,
4. if un → θ , then f (un ) → θ .
Then,
• 1 implies 2,
• 2 implies 3,
• if U is metrizable, then 3 implies 4, which implies 1.
Proof.
1. 1 implies 2: Suppose f is continuous, for W ⊂ V neighborhood of zero, there
exists a neighborhood of zero in U, denoted by V , such that
f (V ) ⊂ W . (1.35)
If E is bounded, there exists t0 ∈ R+ such that E ⊂ tV , ∀t ≥ t0 , so that
f (E) ⊂ f (tV ) = t f (V ) ⊂ tW , ∀t ≥ t0 , (1.36)
and thus f is bounded.
2. 2 implies 3: Suppose un → θ and let W be a neighborhood of zero. Then, there
exists N ∈ N such that if n ≥ N, then un ∈ V ⊂ W where V is a balanced
neighborhood of zero. On the other hand, for n < N, there exists Kn such that
un ∈ Kn V . Define K = max{1, K1 , . . . , Kn }. Then, un ∈ KV , ∀n ∈ N and hence
{un} is bounded. Finally from 2, we have that { f (un )} is bounded.
3. 3 implies 4: Suppose U is metrizable and let un → θ . Given K ∈ N, there exists
nK ∈ N such that if n > nK , then d(un , θ ) < K12 . Define γn = 1 if n < n1 and
γn = K, if nK ≤ n < nK+1 so that
d(γn un , θ ) = d(Kun , θ ) ≤ Kd(un , θ ) < K −1 . (1.37)
Thus since 2 implies 3 we have that { f (γn un )} is bounded so that, by
Proposition 1.9.2, f (un ) = γn−1 f (γn un ) → θ as n → ∞.
4. 4 implies 1: suppose 1 fails. Thus there exists a neighborhood of zero W ⊂ V
such that f −1 (W ) contains no neighborhood of zero in U. Particularly, we can
select {un} such that un ∈ B1/n (θ ) and f (un ) not in W so that { f (un )} does not
converge to zero. Thus 4 fails.
A(u) V < K u U , ∀u ∈ U.
24 1 Topological Vector Spaces
We claim that for each ε > 0 there exists δ > 0 such that if u U < δ , then
A(u) V < ε .
Suppose, to obtain contradiction, that the claim is false.
Thus there exists ε0 > 0 such that for each n ∈ N there exists un ∈ U such that
un U ≤ 1n and A(un ) V ≥ ε0 .
Therefore un → θ and A(un ) does not converge to θ , which contradicts (1.38).
Thus the claim holds.
In particular, for ε = 1, there exists δ > 0 such that if u U < δ , then
A(u) V < 1. Thus given an arbitrary not relabeled u ∈ U, u = θ , for
δu
w=
2 u U
we have
δ A(u) V
A(w) V = < 1,
2 u U
that is
2 u U
A(u) V < , ∀u ∈ U.
δ
Defining
2
K=
δ
the first part of the proof is complete. Reciprocally, suppose there exists K > 0 such
that
A(u) V < K u U , ∀u ∈ U.
Hence un → θ implies A(un ) V → θ , so that from Proposition 1.9.3, A is continu-
ous.
The proof is complete.
In this section we present some important results in Banach spaces. We start with
the following theorem.
Theorem 1.11.1. Let U and V be Banach spaces and let A : U → V be a linear
operator. Then A is bounded if and only if the set C ⊂ U has at least one interior
point, where
C = A−1 [{v ∈ V | v V ≤ 1}].
1.11 Some Classical Results on Banach Spaces 25
Proof. Suppose there exists u0 ∈ U in the interior of C. Thus, there exists r > 0 such
that
Br (u0 ) = {u ∈ U | u − u0 U < r} ⊂ C.
Fix u ∈ U such that u U < r. Thus, we have
A(u + u0) V ≤ 1
and hence
A(u) V ≤ 1 + A(u0) V , (1.39)
A(u) V ≤ K u U , ∀u ∈ U,
U = ∪∞
n=1 An ,
that is,
B2 ∩ A2 = 0.
/
Proceeding inductively in this fashion, for each n ∈ N, we may obtain un ∈ Bn−1 \ Ān
such that we may choose an open ball Bn = Brn (un ) such that
B̄n ⊂ Bn−1 ,
Bn ∩ An = 0,
/
and
0 < rn < 21−n.
Observe that {un } is a Cauchy sequence, considering that if m, n > N, then un , um ∈
BN , so that
d(un , um ) < 2(21−N ).
Define
u = lim un .
n→∞
Since
un ∈ BN , ∀n > N,
we get
u ∈ B̄N ⊂ BN−1 .
Therefore u is not in AN−1 , ∀N > 1, which means u is not in ∪∞
n=1 An = U, a
contradiction.
The proof is complete.
T (u) V < Ku , ∀T ∈ F .
T < K, ∀T ∈ F .
Proof. Define
Bn = {u ∈ U | T (u) V ≤ n, ∀T ∈ F }.
By the hypotheses, given u ∈ U, u ∈ Bn for all n is sufficiently big. Thus,
U = ∪∞
n=1 Bn .
Moreover each Bn is closed. By the Baire category theorem there exists n0 ∈ N such
that Bn0 has nonempty interior. That is, there exists u0 ∈ U and r > 0 such that
Br (u0 ) ⊂ Bn0 .
T (u) V ≤ n0 , ∀u ∈ Br (u0 ).
T (u + u0) V ≤ n0 ,
that is,
T (u) V − T (u0 ) V ≤ n0 .
Thus,
T (u) V ≤ 2n0 , if u U < r. (1.40)
For u ∈ U arbitrary, u = θ , define
ru
w= ,
2 u U
from (1.40) we obtain
r T (u) V
T (w) V = ≤ 2n0 ,
2 u U
so that
4n0 u U
T (u) V ≤ , ∀u ∈ U.
r
Hence
4n0
T ≤ , ∀T ∈ F .
r
The proof is complete.
28 1 Topological Vector Spaces
Theorem 1.11.5 (The Open Mapping Theorem). Let U and V be Banach spaces
and let A : U → V be a bounded onto linear operator. Thus, if O ⊂ U is open, then
A(O) is open in V .
Proof. First we will prove that given r > 0, there exists r > 0 such that
V = ∪∞
n=1 A(nB1 (θ )).
By the Baire category theorem, there exists n0 ∈ N such that the closure of
A(n0 B1 (θ )) has nonempty interior, so that A(B1 (θ )) has nonempty interior. We
will show that there exists r > 0 such that
A(Br2 (θ )) ⊃ BVr1 (θ ),
y ∈ A(B1 (θ ))
u − u0 U ≤ u U + u0 U
< 1 + u0 U
= r2 , (1.43)
so that
A(u − u0) ∈ A(Br2 (θ )).
This means
y = A(u) ∈ A(u0 ) + A(Br2 (θ )),
and hence
A(B1 (θ )) ⊂ A(u0 ) + A(Br2 (θ )).
1.11 Some Classical Results on Banach Spaces 29
and therefore
A(Br2 (θ )) ⊃ BVr1 (θ ).
Since
A(Br2 (θ )) = r2 A(B1 (θ )),
A(B1 (θ )) ⊃ BVr1 (θ ).
Define
∞
u= ∑ un ,
n=1
Therefore
A(B1 (θ )) ⊂ A(B2 (θ )).
The proof of (1.41) is complete.
30 1 Topological Vector Spaces
To finish the proof of this theorem, assume O ⊂ U is open. Let v0 ∈ A(O). Let
u0 ∈ O be such that A(u0 ) = v0 . Thus there exists r > 0 such that
Br (u0 ) ⊂ O.
From (1.41),
A(Br (θ )) ⊃ BVr (θ ),
for some r > 0. Thus
Theorem 1.11.8 (The Closed Graph Theorem). Let U and V be Banach spaces
and let A : U → V be a linear operator. Then A is bounded if and only if its graph is
closed.
Π1 (u, A(u)) = u
and
Π2 (u, A(u)) = A(u).
Observe that Π1 is a bijection, so that by the inverse mapping theorem, Π1−1 is
continuous. As
A = Π2 ◦ Π1−1 ,
it follows that A is continuous. The converse is trivial.
1.12 Hilbert Spaces 31
0 ≤ (u − α v, u − α v)H
= (u, u)H + α 2 (v, v)H − 2α (u, v)H
= u 2H + α 2 v 2H − 2α (u, v)H . (1.44)
(u, v)2H
0 ≤ u 2H − ,
v 2H
that is,
|(u, v)H | ≤ u H v H .
The remaining conclusions are left to the reader.
· H : H → R
u + v 2H = (u + v, u + v)H
= (u, u)H + (v, v)H + 2(u, v)H
≤ (u, u)H + (v, v)H + 2|(u, v)H |
≤ u 2H + v 2H + 2 u H v H
= ( u H + v H )2 . (1.45)
Therefore
u + v H ≤ u H + v H , ∀u, v ∈ H.
The proof is complete.
Theorem 1.12.7. Let H be a Hilbert space and M a closed subspace of H and sup-
pose u ∈ H. Under such hypotheses there exists a unique m0 ∈ M such that
u − m0 H = min { u − m H }.
m∈M
Moreover n0 = u − m0 ∈ M ⊥ so that
u = m0 + n 0 ,
Proof. Define d by
d = inf { u − m H }.
m∈M
u − mi H → d, as i → ∞.
mi − m j 2H = mi − u − (m j − u) 2H
= 2 mi − u 2H + 2 m j − u 2H
−2 − 2u + mi + m j 2H
= 2 mi − u 2H + 2 m j − u 2H
1.12 Hilbert Spaces 33
−4 − u + (mi + m j )/2 2H
→ 2d 2 + 2d 2 − 4d 2 = 0, as i, j → +∞. (1.46)
d 2 ≤ u − (m0 − tm) 2H
= n0 + tm 2H
= n0 2H + 2(n0, m)H t + m 2H t 2 . (1.47)
Since
n0 2H = u − m0 2H = d 2 ,
we obtain
2(n0 , m)H t + m 2H t 2 ≥ 0, ∀t ∈ R
so that
(n0 , m)H = 0.
Being m ∈ M arbitrary, we obtain
n0 ∈ M ⊥ .
u − m 2H = u − m0 + m0 − m 2H
= u − m0 2H + m − m0 2H , (1.48)
since
(u − m0, m − m0 )H = (n0 , m − m0)H = 0.
From (1.48) we obtain
u − m 2H > u − m0 2H = d 2 ,
if m = m0 .
Therefore m0 is unique.
Now suppose
u = m1 + n 1 ,
34 1 Topological Vector Spaces
u − m 2H = u − m1 + m1 − m 2H
= u − m1 2H + m − m1 2H ,
≥ u − m1 H (1.49)
f (u) = (u, u0 )H , ∀u ∈ H.
Moreover
f H ∗ = u0 H .
Proof. Define N by
N = {u ∈ H | f (u) = 0}.
Thus, as f is a continuous and linear, N is a closed subspace of H. If N = H, then
f (u) = 0 = (u, θ )H , ∀u ∈ H and the proof would be complete. Thus, assume N = H.
By the last theorem there exists v = θ such that v ∈ N ⊥ .
Define
f (v)
u0 = v.
v 2H
Thus,if u ∈ N we have
f (u) = 0 = (u, u0 )H = 0.
On the other hand, if u = α v for some α ∈ R, we have
f (u) = α f (v)
f (v)(α v, v)H
=
v 2H
f (v)v
= α v,
v 2H H
= (α v, u0 )H . (1.50)
Therefore f (u) equals (u, u0 )H in the space spanned by N and v. Now we show that
this last space (then span of N and v) is in fact H. Just observe that given u ∈ H we
1.12 Hilbert Spaces 35
may write
f (u)v f (u)v
u = u− + . (1.51)
f (v) f (v)
Since
f (u)v
u− ∈N
f (v)
we have finished the first part of the proof, that is, we have proven that
f (u) = (u, u0 )H , ∀u ∈ H.
f (u) = (u, u1 )H , ∀u ∈ H.
Thus,
u0 − u1 2H = (u0 − u1 , u0 − u1)H
= (u0 − u1 , u0 )H − (u0 − u1, u1 )H
= f (u0 − u1) − f (u0 − u1) = 0. (1.52)
Hence u1 = u0 .
Let us now prove that
f H ∗ = u0 H .
First observe that
f H ∗ = sup{ f (u) | u ∈ H, u H ≤ 1}
= sup{|(u, u0 )H | | u ∈ H, u H ≤ 1}
≤ sup{ u H u0 H | u ∈ H, u H ≤ 1}
≤ u0 H . (1.53)
f H ∗ = sup{ f (u) | u ∈ H, u H ≤ 1}
u0
≥ f
u0 H
(u0 , u0 )H
=
u0 H
= u0 H . (1.54)
Remark 1.12.9. Similarly as above we may define a Hilbert space H over C, that is,
a complex one. In this case the complex inner product (·, ·)H : H × H → C is defined
through the following properties:
1. (u, v)H = (v, u)H , ∀u, v ∈ H,
2. (u + v, w)H = (u, w)H + (v, w)H , ∀u, v, w ∈ H,
3. (α u, v)H = α (u, v)H , ∀u, v ∈ H, α ∈ C,
4. (u, u)H ≥ 0, ∀u ∈ H, and (u, u) = 0, if and only if u = θ .
Observe that in this case we have
In this section we study separable Hilbert spaces and the related orthonormal
bases.
Definition 1.13.1. Let H be a Hilbert space. A set S ⊂ H is said to be orthonormal if
u H = 1,
and
(u, v)H = 0, ∀u, v ∈ S, such that u = v.
If S is not properly contained in any other orthonormal set, it is said to be an or-
thonormal basis for H.
Theorem 1.13.2. Let H be a Hilbert space and let {un }Nn=1 be an orthonormal set.
Then, for all u ∈ H, we have
2
N N
u 2H = ∑ |(u, un )H | + u − ∑ (u, un )H un .
2
n=1
n=1
H
u 2H = (u, u)H
2 2
N N
= ∑ |(u, un )H un + u − ∑ (u, un )H un
n=1 n=1
H H
2
N N
= ∑ |(u, un )H |2 + u − ∑ (u, un )H un . (1.55)
n=1
n=1
H
Corollary 1.13.3 (Bessel Inequality). Let H be a Hilbert space and let {un }Nn=1 be
an orthonormal set. Then, for all u ∈ H, we have
N
u 2H ≥ ∑ |(u, un)H |2 .
n=1
v= ∑ (uα , v)H uα ,
α ∈L
and
v 2H = ∑ |(uα , v)H |2 .
α ∈L
∑ |(uα , v)H | ≤ v 2H .
α ∈L
From this, we may infer that the set An = {α ∈ L | |(uα , v)H | > 1/n} is finite, so that
is at most countable.
Thus (uα , v)H = 0 for at most countably many α s ∈ L, which we order by
{αn }n∈N . Since the sequence
N
sN = ∑ |(uαi , v)H |2 ,
i=1
Also, if α = αl , ∀l ∈ N, then
∞
(v − v , uα )H = lim (v − ∑ (uαi , v)H uαi , uα )H = 0.
N→∞
i=1
Hence
v − v⊥uα , ∀α ∈ L.
If
v − v = θ ,
then we could obtain an orthonormal set
v − v
uα , α ∈ L,
v − v H
1.13 Orthonormal Basis 39
{uα , α ∈ L},
a contradiction.
Therefore, v − v = θ , that is,
N
v = lim
N→∞
∑ (uαi , v)H uαi .
i=1
Observe that clearly {vn } is an orthonormal set and for each m ∈ N, {vk }m
k=1 and
{uk }m
k=1 span the same vector subspace of H.
Such a process of obtaining the orthonormal set {vn } is known as the Gram–
Schmidt orthonormalization.
We finish this section with the following theorem.
Proof. Suppose H is separable and let {un } be a countable dense set in H. To ob-
tain an orthonormal basis it suffices to apply the Gram–Schmidt orthonormalization
procedure to the greatest linearly independent subset of {un }.
Conversely, if B = {vn } is an orthonormal basis for H, the set of all finite linear
combinations of elements of B with rational coefficients are dense in H, so that H is
separable.
40 1 Topological Vector Spaces
2.1 Introduction
and
and therefore
1 1
[−p(v1 − α z)+g(v1 )] ≤ [p(v2 + β z)−g(v2 )], ∀v1 , v2 ∈ V, α , β > 0. (2.9)
α β
Thus, there exists a ∈ R such that
1 1
sup [ (−p(v − α z) + g(v))] ≤ a ≤ inf [ (p(v + α z) − g(v))]. (2.10)
v∈V,α >0 α v∈V,α >0 α
Definition 2.2.2 (Topological Dual Space). For a Banach space U, we define its
topological dual space as the set of all linear continuous functionals defined on U.
We suppose that such dual space of U may be identified with a space denoted by U ∗
through a bilinear form ·, ·U : U × U ∗ → R (here we are referring to the standard
2.2 The Hahn–Banach Theorem 43
representations of dual spaces concerning Lebesgue and Sobolev spaces). That is,
given f : U → R linear continuous functional, there exists u∗ ∈ U ∗ such that
and
u∗ U ∗ = g V ∗ . (2.15)
|u, u∗ U | ≤ u U u∗ U ∗ , ∀u ∈ U, u∗ ∈ U ∗
we have
However, from last corollary, we have that there exists u∗0 ∈ U ∗ such that u∗0 U ∗ =
u U and u, u∗0 U = u U2 . Define u∗1 = u U−1 u∗0 . Then u∗1 U = 1 and u, u∗1 U =
u U .
44 2 The Hahn–Banach Theorems and Weak Topologies
H = {u ∈ U | u, u∗ U = α } (2.19)
We say that H separates A and B strictly if there exists ε > 0 such that
u U
p(u) ≤ , ∀u ∈ U (2.25)
r
which proves (2.23). Now suppose u ∈ C. Since C is open, (1 + ε )u ∈ C for ε is
sufficiently small. Therefore p(u) ≤ 1+1 ε < 1. Conversely, if p(u) < 1, there exists
0 < α < 1 such that α −1 u ∈ C and therefore, since C is convex, u = α (α −1 u) + (1 −
α )θ ∈ C.
2.2 The Hahn–Banach Theorem 45
Lemma 2.2.11. Consider C ⊂ U a convex open set and let u0 ∈ U be a vector not
in C. Then there exists u∗ ∈ U ∗ such that u, u∗ U < u0 , u∗ U , ∀u ∈ C
g(tu0 ) = t, t ∈ R. (2.26)
We have that g(u) ≤ p(u), ∀u ∈ V . From the Hahn–Banach theorem, there exists a
linear functional f on U which extends g such that
Here we have used Lemma 2.2.10. In particular, f (u0 ) = 1 and (also from the last
lemma) f (u) < 1, ∀u ∈ C. The existence of u∗ satisfying the theorem follows from
the continuity of f indicated in (2.27).
Proof. There exists ε > 0 sufficiently small such that Aε = A + B(0, ε ) and Bε =
B + B(0, ε ) are convex disjoint sets. From Theorem 2.2.9, there exists u∗ ∈ U ∗ such
that u∗ = θ and
u, u∗ U = 0, ∀u ∈ V. (2.32)
Definition 2.3.1 (Weak Neighborhoods and Weak Topologies). For the topologi-
cal space U and u0 ∈ U, we define a weak neighborhood of u0 , denoted by Vw as
for some m ∈ N, ε > 0, and u∗i ∈ U ∗ , ∀i ∈ {1, . . . , m}. Also, we define the weak
topology for U, denoted by σ (U,U ∗ ), as the set of arbitrary unions and finite inter-
sections of weak neighborhoods in U.
Proof. It is clear that if ψ is continuous with U endowed with the weak topology,
then u∗ ◦ ψ is continuous for all u∗ ∈ U ∗ . Conversely, consider U a weakly open set
in U. We have to show that ψ −1 (U ) is open in Z. But observe that U = ∪λ ∈L Vλ ,
where each Vλ is a weak neighborhood. Thus ψ −1 (U ) = ∪λ ∈L ψ −1 (Vλ ). The result
follows considering that u∗ ◦ ψ is continuous for all u∗ ∈ U ∗ , so that ψ −1 (Vλ ) is
open, for all λ ∈ L.
Proposition 2.3.3. A Banach space U is Hausdorff as endowed with the weak topol-
ogy σ (U,U ∗ ).
Defining
and
Proof.
1. The result follows directly from the definition of topology σ (U,U ∗ ).
2. This follows from the inequality
3. Since for every u∗ ∈ U ∗ the sequence {un , u∗ U } is bounded, from the uni-
form boundedness principle, we have that there exists M > 0 such that un U ≤
M, ∀n ∈ N. Furthermore, for u∗ ∈ U ∗ , we have
|un , u∗ U | ≤ u∗ U ∗ un U , (2.39)
Thus
Theorem 2.3.6. Consider A ⊂ U a convex set. Thus A is weakly closed if and only
if it is strongly closed.
Proof. Suppose A is strongly closed. Consider u0 ∈ A. By the Hahn–Banach the-
orem there exists a closed hyperplane which separates u0 and A strictly. Therefore
there exists α ∈ R and u∗ ∈ U ∗ such that
Define
is onto.
The weak topology for U ∗ is denoted by σ (U ∗ ,U ∗∗ ). By analogy, we can define
the topology σ (U ∗ ,U), which is called the weak-star topology. A standard neighbor-
hood of u∗0 ∈ U ∗ for the weak-star topology, which we denoted by Vw∗ , is given by
for some ε > 0, m ∈ N, ui ∈ U, ∀i ∈ {1, . . . , m}. It is clear that the weak topology for
U ∗ and the weak-star topology coincide if U is reflexive.
Proposition 2.4.2. Let U be a Banach space. U ∗ as endowed with the weak-star
topology is a Hausdorff space.
Proof. The proof is similar to that of Proposition 2.3.3.
Proof. For each u ∈ U, we will associate a real number ωu and denote ω = ∏u∈U ωu .
We have that ω ∈ RU and let us consider the projections Pu : RU → R, where
Pu (ω ) = ωu . Consider the weakest topology σ for which the functions Pu (u ∈ U)
are continuous. For U ∗ , with the topology σ (U ∗ ,U), define φ : U ∗ → RU by
Since for each fixed u the mapping u∗ → u, u∗ U is weakly star continuous, we
see that φ is σ continuous, since weak-star convergence and convergence in σ are
equivalent in U ∗ . To prove that φ −1 is continuous, from Proposition 2.3.2, it suf-
fices to show that the function ω → u, φ −1 (ω )U is continuous on φ (U ∗ ). This is
true because u, φ −1 (ω )U = ωu on φ (U ∗ ). On the other hand, it is also clear that
φ (BU ∗ ) = K, where
K = {ω ∈ RU | |ωu | ≤ u U ,
ωu+v = ωu + ωv , ωλ u = λ ωu , ∀u, v ∈ U, λ ∈ R}. (2.48)
and
and
Bλ ,u = {ω ∈ RU |ωλ u − λ ωu = 0} (2.52)
we may write
We recall that the K2 is closed because arbitrary intersections of closed sets are
closed. Finally, we have that K1 ∩ K2 is compact, which completes the proof.
BU = {u ∈ U | u U ≤ 1} (2.54)
and therefore
n n
∑ βi αi − ∑ βi uε , fi U < ε S (2.57)
i=1 i=1
or
n n n
∑ βi αi < ∑ βi fi uε U + ε S ≤ ∑ βi fi + εS (2.58)
i=1 i=1 ∗
U i=1 U∗
so that
n n
∑ βi αi ≤ ∑ βi fi (2.59)
i=1 i=1
U∗
to the closure of ϕ (BU ). Let us suppose that α does not belong to the closure of
ϕ (BU ) and obtain a contradiction. Thus we can separate α and the closure of ϕ (BU )
strictly, that is, there exists β = (β1 , . . . , βn ) ∈ Rn and γ ∈ R such that
or,
and hence
Now we will complete the proof of Kakutani theorem. Suppose BU is weakly com-
pact (i.e., compact for the topology σ (U,U ∗ )). Observe that J : U → U ∗∗ is weakly
continuous, that is, it is continuous with U endowed with the topology σ (U,U ∗ )
and U ∗∗ endowed with the topology σ (U ∗∗ ,U ∗ ). Thus as BU is weakly com-
pact, we have that J(BU ) is compact for the topology σ (U ∗∗ ,U ∗ ). From the last
lemma, J(BU ) is dense BU ∗∗ for the topology σ (U ∗∗ ,U ∗ ). Hence J(BU ) = BU ∗∗ , or
J(U) = U ∗∗ , which completes the proof.
Proposition 2.5.5. Let U be a reflexive Banach space. Let K ⊂ U be a convex closed
bounded set. Then K is weakly compact.
Proof. From Theorem 2.3.6, K is weakly closed (closed for the topology σ (U,U ∗ )).
Since K is bounded, there exists α ∈ R+ such that K ⊂ α BU . Since K is weakly
closed and K = K ∩ α BU , we have that it is weakly compact.
52 2 The Hahn–Banach Theorems and Weak Topologies
It can be easily verified that these two topologies coincide (through restrictions and
extensions of linear forms). From Theorem 2.5.2, it suffices to show that BM is
compact for the topology σ (M, M ∗ ). But BU is compact for σ (U,U ∗ ) and M ⊂
U is closed (strongly) and convex so that it is weakly closed; thus, from the last
proposition, BM is compact for the topology σ (U,U ∗ ), and therefore it is compact
for σ (M, M ∗ ).
Define U0 as the vector space on Q spanned by {un } and U1 as the vector space
on R spanned by {un }. It is clear that U0 is dense in U1 and we will show that U1
is dense in U, so that U0 is a dense set in U. Suppose u∗ is such that u, u∗ U =
0, ∀u ∈ U1 . Since {u∗n } is dense in U ∗ , given ε > 0, there exists n ∈ N such that
u∗n − u∗ U ∗ < ε , so that
1 ∗
u U ∗ ≤ un , u∗n U = un , u∗n − u∗ U + un, u∗ U
2 n
≤ u∗n − u∗ U ∗ un U + 0 < ε (2.68)
or
u∗ w ≤ u∗ U .
d(u∗ , v∗ ) = u∗ − v∗ w .
Now we shall prove that the topology induced by d coincides with σ (U ∗ ,U) in U ∗ .
Let u∗0 ∈ BU ∗ and let V be neighborhood of u∗0 in σ (U ∗ ,U).
We need to prove that there exists r > 0 such that
vi U ≤ 1, ∀i ∈ {1, . . . , k}.
54 2 The Hahn–Banach Theorems and Weak Topologies
Since {un } is dense in U, for each i ∈ {1, . . . , k}, there exists ni ∈ N such that
ε
uni − vi U < .
4
Choose r > 0 small enough such that
ε
2ni r < , ∀i ∈ {1, . . . , k}.
2
We are going to show that Vw ⊂ V , where
For u∗ ∈ Vw we have
k
1
d(u∗ , u∗0 ) = ∑ 2n |un, u∗ − u∗0U |
n=1
∞
1
+ ∑ n
|un , u∗ − u∗0 U |
n=k+1 2
∞
1
< ε +2 ∑ 2 n
n=k+1
1
= ε+ . (2.71)
2k−1
2.7 Uniformly Convex Spaces 55
1
< r/2.
2k−1
The first part of the proof is finished.
Conversely, assume BU ∗ is metrizable in σ (U ∗ ,U). We are going to show that U
is separable.
Define,
1
Ṽn = u∗ ∈ BU ∗ | d(u∗ , θ ) < .
n
From the first part, we may find Vn a neighborhood of zero in σ (U ∗ ,U) such that
Vn ⊂ Ṽn .
Thus D is countable and we are going to prove that such a set is dense in U.
Suppose u∗ ∈ U ∗ is such that
u, u∗ U = 0, ∀u ∈ D.
Hence,
u∗ ∈ Vn ⊂ Ṽn , ∀n ∈ N,
so that u∗ = θ .
The proof is complete.
Theorem 2.7.2 (Milman Pettis). Every uniformly convex Banach space is reflexive.
Thus, suppose given ε > 0. Let δ > 0 be the corresponding constant relating the
uniformly convex property.
Choose f ∈ U ∗ such that f U ∗ = 1 and
δ
f , η U ∗ > 1 − . (2.72)
2
Define
δ
V = ζ ∈ U ∗∗ | | f , ζ − η U ∗ | < .
2
Observe that V is neighborhood of η in σ (U ∗∗ ,U ∗ ). Since J(BU ) is dense in BU ∗∗
concerning the topology σ (U ∗∗ ,U ∗ ), we have that V ∩ J(BU ) = 0/ and thus there
exists u ∈ BU such that J(u) ∈ V. Suppose, to obtain contradiction, that
η − J(u) U ∗∗ > ε .
Therefore, defining
W = (J(u) + ε BU ∗∗ )c ,
we have that η ∈ W , where W is also a weak neighborhood of η in σ (U ∗∗ ,U ∗ ),
since BU ∗∗ is closed in σ (U ∗∗ ,U ∗ ).
Hence V ∩ W ∩ J(BU ) = 0, / so that there exists some v ∈ BU such that J(v) ∈
V ∩W. Thus, J(u) ∈ V and J(v) ∈ V , so that
δ
|u, f U − f , η U ∗ | < ,
2
and
δ
|v, f U − f , η U ∗ | < .
2
Hence,
2 f , η U ∗ < u + v, f U + δ
≤ u + v U + δ . (2.73)
u + v U
> 1−δ,
2
and thus from the definition of uniform convexity, we obtain
u − v U ≤ ε . (2.74)
On the other hand, since J(v) ∈ W , we have
J(u) − J(v) U ∗∗ = u − v U > ε ,
The main references for this chapter are Reed and Simon [52] and Bachman and
Narici [6].
First we recall that the set of all bounded linear operators, denoted by L (U,Y ),
is a Banach space with the norm
A = sup{ Au Y | u U ≤ 1}.
The topology related to the metric induced by this norm is called the uniform oper-
ator topology.
Let us introduce now the strong operator topology, which is defined as the weak-
est topology for which the functions
Eu : L (U,Y ) → Y
In the next lines we describe the weak operator topology in L (U,Y ). Such a topol-
ogy is weakest one such that the functions
Eu,v : L (U,Y ) → C
where ε > 0, u1 , . . . , un ∈ U, v1 , . . . , vm ∈ Y ∗ .
A sequence {An } ⊂ L (U,Y ) converges to A ∈ L (U,Y ) if
as n → ∞, ∀u ∈ U, v ∈ Y ∗ .
|T (u)| ≤ Au Y · v∗ ≤ A v∗ Y ∗ u U .
T (u) = u, u∗ U , ∀u ∈ U.
that is,
u, A∗ v∗ U = Au, v∗ Y , ∀u ∈ U, v∗ ∈ Y ∗ .
We call A∗ : Y ∗ → U ∗ the adjoint operator relating A : U → Y.
Theorem 3.2.2. Let U,Y be Banach spaces and let A : U → Y be a bounded linear
operator. Then
A = A∗ .
3.2 Adjoint Operators 59
A = sup{ Au | u U = 1}
u∈U
∗ ∗ ∗
= sup sup {Au, v Y | v Y = 1}, u U = 1
u∈U v∗ ∈Y ∗
= sup {Au, v∗ Y | v∗ Y ∗ = 1, u U = 1}
(u,v∗ )∈U×Y ∗
= sup {u, A∗ v∗ U | v∗ Y ∗ = 1, u U = 1}
(u,v∗ )∈U×Y ∗
∗ ∗ ∗
= sup sup{u, A v U | u U = 1}, v Y ∗ = 1
v∗ ∈Y ∗ u∈U
= sup { A v , v∗ Y ∗ = 1}
∗ ∗
v∗ ∈Y ∗
∗
= A . (3.1)
4. AA∗ = A 2 .
Proof.
1. Observe that
2. Observe that
3. We have that
I = AA−1 = A−1 A,
so that
I = I ∗ = (AA−1 )∗ = (A−1 )∗ A∗ = (A−1 A)∗ = A∗ (A−1 )∗ .
4. Observe that
A∗ A ≤ A A∗ = A 2 ,
and
A∗ A ≥ sup{(u, A∗ Au)H | u U = 1}
u∈U
60 3 Topics on Linear Operators
= sup{(Au, Au)H | u U = 1}
u∈U
= sup{ Au 2H | u U = 1} = A 2, (3.2)
u∈U
and hence
A∗ A = A 2 .
A = A∗ .
Proof. Let v∗ ∈ N(A∗ ). Choose v ∈ R(A). Thus there exists u in U such that Au = v
so that
v, v∗ Y = Au, v∗ Y = u, A∗ v∗ U = 0.
Since v ∈ R(A) is arbitrary we have obtained
N(A∗ ) ⊂ [R(A)]⊥ .
Au, v∗ Y = 0,
so that
u, A∗ v∗ U , ∀u ∈ U.
Therefore A∗ v∗ = θ , that is, v∗ ∈ N(A∗ ). Since v∗ ∈ [R(A)]⊥ is arbitrary, we get
[R(A)]⊥ ⊂ N(A∗ ).
A(u0 ) = v
and
u0 U ≤ K v Y .
3.2 Adjoint Operators 61
Proof. Define L = N(A) = {u ∈ U : A(u) = θ } (the null space of A). Consider the
space U/L, where
U/L = {u : u ∈ U},
where
u = {u + w : w ∈ L}.
Define A : U/L → R(A), by
A(u) = A(u).
Observe that A is one-to-one, linear, onto, and bounded. Moreover R(A) is closed so
that it is a Banach space. Hence by the inverse mapping theorem we have that A has
a continuous inverse. Thus, for any v ∈ R(A), there exists u ∈ U/L such that
A(u) = v
so that
−1
u=A (v),
and therefore
−1
u ≤ A v Y .
Recalling that
u = inf { u + w U },
w∈L
and so that
A(u0 ) = A(u0 ) = A(u) = v.
−1
Taking K = 2 A we have completed the proof.
Theorem 3.1. Let U,Y be Banach spaces and let A : U → Y be a bound linear
operator. Assume R(A) is closed. Under such hypotheses
R(A∗ ) = [N(A)]⊥ .
u∗ = A∗ (v∗ ).
R(A∗ ) ⊂ [N(A)]⊥ .
62 3 Topics on Linear Operators
u, u∗ U = 0, ∀u ∈ N(A).
Fix v ∈ R(A). From the Lemma 3.1, there exists K > 0 (which does not depend on v)
and uv ∈ U such that
A(uv ) = v
and
uv U ≤ K v Y .
Define f : R(A) → R by
f (v) = uv , u∗ U .
Observe that
| f (v)| ≤ uv U u∗ U ∗ ≤ K v Y u∗ U ∗ ,
u = (u − uv0 ) + uv0 ,
and
A(u − uv0 ) = A(u) − A(uv0 ) = v0 − v0 = θ .
Hence,
u, u∗ U = A(u), v∗ Y , ∀u ∈ U.
We may conclude that u∗ = A∗ (v∗ ) ∈ R(A∗ ). Since u∗ ∈ [N(A)]⊥ is arbitrary we
obtain
[N(A)]⊥ ⊂ R(A∗ ).
The proof is complete.
Definition 3.2.6. Let U be a Banach space and S ⊂ U. We define the positive con-
jugate cone of S, denoted by S⊕ by
where
(A∗ )−1 = {v∗ ∈ Y ∗ : A∗ v∗ ∈ S⊕ }.
A(u), v∗ Y ≥ 0,
so that
u, A∗ (v∗ )U ≥ 0.
Since u ∈ S is arbitrary, we get
From this
[A(S)]⊕ ⊂ (A∗ )−1 (S⊕ ).
Reciprocally, let v∗ ∈ (A∗ )−1 (S⊕ ). Hence A∗ (v∗ ) ∈ S⊕ so that for u ∈ S we obtain
and therefore
A(u), v∗ Y ≥ 0.
64 3 Topics on Linear Operators
ψn ∈ [φ1 , . . . , φn ]⊥ ,
ψn θ , weakly in H.
and
∞
∑ |b j |2 = 1.
j=n+1
Therefore
∞
|(ψn , ψ )H | = ∑ (φ j , φ j )H a j · b j
∗
j=n+1
∞
= ∑ a j · b j
j=n+1
∞ ∞
≤ ∑ |a j |2 ∑ |b j |2
j=n+1 j=n+1
√
≤ ε, (3.6)
(ψn , ψ ∗ )H → 0, as n → ∞.
ψn θ , weakly in H.
66 3 Topics on Linear Operators
Aψn → θ in norm ,
Ez = x,
where z = x + y, x ∈ M, and y ∈ M ⊥ .
Observe that
1. E is linear,
2. E is idempotent, that is, E 2 = E,
3. R(E) = M,
4. N(E) = M ⊥ .
Also observe that from
Ez = x
we have
Ez 2H = x 2H ≤ x 2H + y 2H = z 2H ,
so that
E ≤ 1.
A≥θ
if
(Au, u)H ≥ 0, ∀u ∈ H,
3.4 The Square Root of a Positive Operator 67
A≥B
if
A−B ≥ θ.
Theorem 3.4.3. Let A and B be bounded self-adjoint operators such that A ≥ θ and
B ≥ θ . If AB = BA, then
AB ≥ θ .
Proof. If A = θ , the result is obvious. Assume A = θ and define the sequence
A
A1 = , An+1 = An − A2n, ∀n ∈ N.
A
We claim that
θ ≤ An ≤ I, ∀n ∈ N.
We prove the claim by induction.
For n = 1, it is clear that A1 ≥ θ . And since A1 = 1, we get
so that
A1 ≤ I.
Thus
θ ≤ A1 ≤ I.
Now suppose θ ≤ An ≤ I. Since An is self-adjoint, we have
where v = An u. Therefore
A2n (I − An ) ≥ θ .
Similarly, we may obtain
An (I − An )2 ≥ θ ,
so that
θ ≤ A2n (I − An) + An (I − An )2 = An − A2n = An+1 .
So, also we have
θ ≤ I − An + A2n = I − An+1,
that is,
θ ≤ An+1 ≤ I,
68 3 Topics on Linear Operators
so that
θ ≤ An ≤ I, ∀n ∈ N.
Observe that
A1 = A21 + A2
= A21 + A22 + A3
... ...................
= A21 + . . . + A2n + An+1. (3.8)
Since An+1 ≥ θ , we obtain
Hence
(ABu, u)H ≥ 0, ∀u ∈ H.
The proof is complete.
Ai B = BAi , ∀i ∈ N.
Cn = B − An ≥ 0, ∀n ∈ N.
and also
Cn (Cm − Cn ) ≥ θ .
Thus,
(Cm2 u, u)H ≥ (CnCm u, u)H ≥ (Cn2 u, u)H ,
and we may conclude that
(Cn2 u, u)H
is a monotone nonincreasing sequence of real numbers, bounded below by 0, so that
there exists α ∈ R such that
as
m, n → ∞.
Therefore {Cn u} is a Cauchy sequence in norm, so that there exists the limit
Now define A by
Au = lim An u.
n→∞
sup{ An } < ∞
n∈N
An ≤ K, ∀n ∈ N.
Therefore
An u H ≤ K u H ,
so that
Au = lim { An u H } ≤ K u H , ∀u ∈ H
n→∞
(Au, v)H = lim (An u, v)H = lim (u, An v)H = (u, Av)H ,
n→∞ n→∞
so that
(Au, u) = lim (An u, u)H ≤ (Bu, u)H , ∀u ∈ H.
n→∞
Hence A ≤ B.
The proof is complete.
Theorem 3.4.6. Suppose A ∈ L (H) is positive. Then there exists B ≥ θ such that
B2 = A.
A ≤ 1,
so that if
A
C2 =
A
then
B = A 1/2C.
Let
B0 = θ ,
and consider the sequence of operators given by
1
Bn+1 = Bn + (A − B2n), ∀n ∈ N ∪ {0}.
2
Since each Bn is polynomial in A, we have that Bn is self-adjoint and commute with
any operator with commutes with A. In particular
Bi B j = B j Bi , ∀i, j ∈ N.
1 1
I − Bn+1 = I − Bn − A + B2n
2 2
1 1
= (I − Bn ) + (I − A) ≥ θ
2
(3.13)
2 2
so that
Bn+1 ≤ I.
The induction is complete, that is,
Bn ≤ I, ∀n ∈ N.
B0 ≤ B1 ,
and supposing
Bn−1 ≤ Bn ,
we have
1 1
Bn+1 − Bn = Bn + (A − B2n) − Bn−1 − (A − B2n−1)
2 2
1
= Bn − Bn−1 − (B2n − B2n−1)
2
1
= Bn − Bn−1 − (Bn + Bn−1)(Bn − Bn−1)
2
1
= (I − (Bn + Bn−1))(Bn − Bn−1)
2
1
= ((I − Bn−1) + (I − Bn))(Bn − Bn−1) ≥ θ .
2
The induction is complete, that is,
θ = B0 ≤ B1 ≤ B2 ≤ . . . ≤ Bn ≤ . . . ≤ I.
Bn → B in norm.
Fixing u ∈ H we have
1
Bn+1 u = Bn u + (A − B2n)u,
2
so that taking the limit in norm as n → ∞, we get
θ = (A − B2 )u.
3.5 About the Spectrum of a Linear Operator 73
Definition 3.5.1. Let U be a Banach space and let A ∈ L (U). A complex number
λ is said to be in the resolvent set ρ (A) of A, if
λI −A
Rλ (A) = (λ I − A)−1
the resolvent of A in λ .
If λ ∈ ρ (A), we write
λ ∈ σ (A) = C − ρ (A),
where σ (A) is said to be the spectrum of A.
Theorem 3.5.3. Let U be a Banach space and suppose that A ∈ L (U). Then ρ (A)
is an open subset of C and
F(λ ) = Rλ (A)
is an analytic function with values in L (U) on each connected component of ρ (A).
For λ , μ ∈ σ (A), Rλ (A), and Rμ (A) commute and
1 1
=
λ −A λ − λ0 + (λ0 − A)
⎡ ⎤
1 ⎣ 1
= ⎦
λ 0 − A 1 − λ0 −λ
λ0 −A
∞
1 λ0 − λ n
= 1+ ∑ . (3.14)
λ0 − A n=1 λ0 − A
Define
∞
R̂λ (A) = Rλ0 (A) I + ∑ (λ − λ0) (Rλ0 ) n n
. (3.15)
n=1
Observe that
(Rλ0 )n ≤ Rλ0 n .
Thus, the series indicated in (3.15) will converge in norm if
Hence, for λ satisfying (3.16), R̂(A) is well defined and we can easily check that
Therefore
R̂λ (A) = Rλ (A), if |λ − λ0| < Rλ0 −1 ,
so that λ0 is an interior point. Since λ0 ∈ ρ (A) is arbitrary, we have that ρ (A) is
open. Finally, observe that
Corollary 3.5.4. Let U be a Banach space and A ∈ L (U). Then the spectrum of A
is nonempty.
In particular
Rλ (A) → 0, as |λ | → ∞. (3.19)
σ (A) = 0.
/
In such a case Rλ (A) would be an entire bounded analytic function. From Liouville’s
theorem, Rλ (A) would be constant, so that from (3.19) we would have
Rλ (A) = θ , ∀λ ∈ C,
which is a contradiction.
Proposition 3.5.5. Let H be a Hilbert space and A ∈ L (H).
1. If λ ∈ Res[σ (A)], then λ ∈ Pσ (A∗ ).
2. If λ ∈ Pσ (A), then λ ∈ Pσ (A∗ ) ∪ Res[σ (A∗ )].
Proof.
1. If λ ∈ Res[σ (A)], then
R(A − λ I) = H.
Therefore there exists v ∈ (R(A − λ I))⊥, v = θ such that
(v, (A − λ I)u)H = 0, ∀u ∈ H
that is,
((A∗ − λ I)v, u)H = 0, ∀u ∈ H
so that
(A∗ − λ I)v = θ ,
which means that λ ∈ Pσ (A∗ ).
2. Suppose there exists v = θ such that
(A − λ I)v = θ ,
and
λ ∈ Pσ (A∗ ).
76 3 Topics on Linear Operators
Thus
(u, (A − λ I)v))H = 0, ∀u ∈ H,
so that
((A∗ − λ I)u, v)H , ∀u ∈ H.
Since
(A∗ − λ I)u = θ , ∀u ∈ H, u = θ ,
we get v ∈ (R(A∗ − λ I))⊥ , so that R(A∗ − λ I) = H.
Hence λ ∈ Res[σ (A∗ )].
(A − (λ + μ i))u 2 = (A − λ )u 2 + μ 2 u 2,
so that
(A − (λ + μ i))u 2 ≥ μ 2 u 2.
Therefore if μ = 0, A − (λ + μ i) has a bounded inverse on its range, which is closed.
If R(A − (λ + μ i)) = H, then by the last result (λ − μ i) would be in the point spec-
trum of A, which contradicts the last inequality. Hence, if μ = 0, then λ + μ i ∈ ρ (A).
To complete the proof, suppose
Au1 = λ1 u1 ,
and
Au2 = λ2 u2 ,
where
λ1 , λ2 ∈ R, λ1 = λ2 , and u1 , u2 = θ .
Thus
Since λ1 − λ2 = 0 we get
(u1 , u2 )H = 0.
3.6 The Spectral Theorem for Bounded Self-Adjoint Operators 77
and
M = sup{(Au, u)H | u H = 1}.
u∈H
−ε I ≤ A − B ≤ ε I,
α = max{|m|, |M|},
where
m = inf {(Au, u)H | u H = 1},
u∈H
and
M = sup{(Au, u)H | u H = 1}.
u∈H
Then
A = α .
Proof. Observe that
and
(A(u − v), u − v)H = (Au, u)H + (Av, v)H − 2(Au, v)H .
Thus,
so that
4(Au, v)H ≤ α ( u + v U2 + u − v U2 ).
Hence, replacing v by −v, we obtain
−4(Au, v)H ≤ α ( u + v U2 + u − v U2 ),
and therefore
4|(Au, v)H | ≤ α ( u + v U2 + u − v U2 ).
Replacing v by β v, we get
4|(A(u), v)H | ≤ 2α ( u U2 /β + β v U2 ).
β = u U / v U ,
we obtain
|(Au, v)H | ≤ α u U v U , ∀u, v ∈ U.
Thus
A ≤ α .
On the other hand,
|(Au, u)H | ≤ A u U2 ,
so that
|M| ≤ A
and
|m| ≤ A ,
so that
α ≤ A .
The proof is complete.
At this point we start to develop the spectral theory. Define by P the set of all real
polynomials defined in R. Define
Φ1 : P → L (H),
by
Φ1 (p(λ )) = p(A), ∀p ∈ P.
Thus we have
1. Φ1 (p1 + p2 ) = p1 (A) + p2(A),
2. Φ1 (p1 · p2 ) = p1 (A)p2 (A),
3. Φ1 (α p) = α p(A), ∀α ∈ R, p ∈ P,
4. if p(λ ) ≥ 0, on [m, M], then p(A) ≥ θ .
3.6 The Spectral Theorem for Bounded Self-Adjoint Operators 79
v1 + i μ 1 , . . . , vk + i μ k .
Observe that if μi = 0, then vi ∈ (m, M). The assumption that p(λ ) ≥ 0 on [m, M]
implies that any real root in (m, M) must be of even multiplicity.
Since complex roots must occur in conjugate pairs, we have the following repre-
sentation for p(λ ) :
n l k
p(λ ) = a ∏(λ − αi ) ∏(βi − λ ) ∏((λ − vi )2 + μi2 ),
i=1 i=1 i=1
(A − vk I)2 + μk2I ≥ θ .
Therefore,
p(A) ≥ θ .
The idea is now to extend the domain of Φ1 to the set of upper semicontinuous
functions, and such set we will denote by Cup .
Observe that if f ∈ Cup , there exists a sequence of continuous functions {gn }
such that
gn ↓ f , pointwise ,
that is,
gn (λ ) ↓ f (λ ), ∀λ ∈ R.
Considering the Weierstrass Theorem, since gn ∈ C([m, M]), we may obtain a se-
quence of polynomials {pn } such that
gn + 1 − pn < 1 ,
2 n 2n
∞
80 3 Topics on Linear Operators
pn (λ ) ↓ f (λ ), on [m, M].
Therefore
p1 (A) ≥ p2 (A) ≥ p3 (A) ≥ . . . ≥ pn (A) ≥ . . .
Since pn (A) is self-adjoint for all n ∈ N, we have
pn ↓ f , and qn ↓ f ,
First observe that being {pn } and {qn } sequences of continuous functions we have
that
ĥnk (λ ) = max{pn (λ ), qk (λ )}, ∀λ ∈ [m, M]
is also continuous, ∀n, k ∈ N. Now fix n ∈ N and define
hk (λ ) = max{pk (λ ), qn (λ )}.
Observe that
hk (λ ) ↓ qn (λ ), ∀λ ∈ R,
so that by Dini’s theorem
It follows that for each n ∈ N there exists kn ∈ N such that if k > kn then
1
hk (λ ) − qn(λ ) ≤ , ∀λ ∈ [m, M].
n
Since
pk (λ ) ≤ hk (λ ), ∀λ ∈ [m, M],
3.6 The Spectral Theorem for Bounded Self-Adjoint Operators 81
we obtain
1
pk (λ ) − qn (λ ) ≤ , ∀λ ∈ [m, M].
n
By analogy, we may show that for each n ∈ N there exists k̂n ∈ N such that if k > k̂n ,
then
1
qk (λ ) − pn(λ ) ≤ .
n
From above we obtain
1
lim pk (A) ≤ qn (A) + .
k→∞ n
Since the self-adjoint qn (A) + 1/n commutes with the
lim pk (A)
k→∞
we obtain
1
lim pk (A) ≤ lim qn (A) +
k→∞ n→∞ n
≤ lim qn (A). (3.21)
n→∞
so that
lim qn (A) = lim pn (A) = f (A).
n→∞ n→∞
Φ2 ( f ) = f (A).
f1 (A) ≥ f2 (A).
h = f1 − g1 and h = f2 − g2.
Thus
f 1 − g1 = f 2 − g2 ,
that is
f 1 + g2 = f 2 + g1 ,
so that from the definition of Φ2 we obtain
that is,
f1 (A) − g1(A) = f2 (A) − g2 (A).
Therefore Φ3 is well defined. Finally observe that for α < 0
α ( f − g) = −α g − (−α ) f ,
Φ3 (α f ) = α f (A) = αΦ3 ( f ), ∀α ∈ R.
Denote
E(μ ) = Φ3 (h μ ) = hμ (A).
Observe that
h μ (λ )h μ (λ ) = h μ (λ ), ∀λ ∈ R,
so that
[E(μ )]2 = E(μ ), ∀μ ∈ R.
Therefore
{E(μ ) | μ ∈ R}
is a family of orthogonal projections. Also observe that if ν ≥ μ , we have
3.6 The Spectral Theorem for Bounded Self-Adjoint Operators 83
hν (λ )h μ (λ ) = hμ (λ )hν (λ ) = h μ (λ ),
so that
E(ν )E(μ ) = E(μ )E(ν ) = E(μ ), ∀ν ≥ μ .
If μ < m, then h μ (λ ) = 0, on [m, M], so that
E(μ ) = 0, if μ < m.
E(μ ) = I, if μ ≥ M.
Next we show that the family {E(μ )} is strongly continuous from the right. First
we will establish a sequence of polynomials {pn } such that
pn ↓ h μ
and
pn (λ ) ≥ hμ + 1 (λ ), on [m, M].
n
Observe that for any fixed n there exists a sequence of polynomials {pnj } such that
Thus
gn (λ ) ≥ h μ + 1 (λ ), ∀λ ∈ R,
n
and we obtain
lim gn (λ ) ≥ lim hμ + 1 (λ ) = hμ (λ ).
n→∞ n→∞ n
so that
lim gn (λ ) ≤ lim prn (λ ).
n→∞ n→∞
Therefore
Thus
lim gn (λ ) = h μ (λ ).
n→∞
1
|gn (λ ) + 1/n − pn(λ )| < , ∀λ ∈ [m, M], n ∈ N,
2n
so that
pn (λ ) ≥ gn (λ ) + 1/n − 1/2n ≥ gn (λ ) ≥ h μ +1/n (λ ).
Thus
pn (A) → E(μ ),
and
pn (A) ≥ h μ + 1 (A) = E(μ + 1/n) ≥ E(μ ).
n
Thus
lim E(μ + 1/n) = E(μ ).
n→∞
From this we may easily obtain the strong continuity from the right.
For μ ≤ ν we have
μ (hν (λ ) − hμ (λ )) ≤ λ (hν (λ ) − hμ (λ ))
≤ ν (hν (λ ) − hμ (λ )). (3.24)
To verify this observe that if λ < μ or λ > ν , then all terms involved in the above
inequalities are zero. On the other hand if
μ ≤λ ≤ν
then
hν (λ ) − hμ (λ ) = 1,
so that in any case (3.24) holds. From the monotonicity property we have
a < m and b ≥ M.
3.6 The Spectral Theorem for Bounded Self-Adjoint Operators 85
P0 = {a = λ0 , λ1 , . . . , λn = b},
such that
max {|λk − λk−1 |} < ε .
k∈{1,...,n}
Hence
we obtain
n
∑ λk−1(E(λk ) − E(λk−1)) ≤ A
k=1
n
≤ ∑ λk (E(λk ) − E(λk−1)). (3.27)
k=1
Let λk0 ∈ [λk−1 , λk ]. Since (λk − λk0) ≤ (λk − λk−1) from (3.26) we obtain
n n
A − ∑ λk0 (E(λk ) − E(λk−1)) ≤ ε ∑ (E(λk ) − E(λk−1))
k=1 k=1
= ε I. (3.28)
By analogy
n
− ε I ≤ A − ∑ λk0 (E(λk ) − E(λk−1)). (3.29)
k=1
Since
n
A − ∑ λk0 (E(λk ) − E(λk−1))
k=1
is self-adjoint we obtain
n
A − ∑ λk0 (E(λk ) − E(λk−1)) < ε .
k=1
that is, M
A= λ dE(λ ).
m−
Also if
p(eiφ ) ≥ 0
3.7 The Spectral Decomposition of Unitary Transformations 87
so that
p(U) = [q(U)]∗ q(U).
Therefore
which means
p(U) ≥ 0.
Define the function hμ (φ ) by
1, if 2kπ < φ ≤ 2kπ + μ ,
h μ (φ ) = (3.30)
0, if 2kπ + μ < φ ≤ 2(k + 1)π ,
for each k ∈ {0, ±1, ±2, ±3, . . .}. Define E(μ ) = hμ (U). Observe that the family
{E(μ )} are projections and in particular
E(0) = 0,
E(2π ) = I
and if μ ≤ ν , since
h μ (φ ) ≤ hν (φ ),
we have
E(μ ) ≤ E(ν ).
Suppose given ε > 0. Let P0 be a partition of [0, 2π ], that is,
P0 = {0 = φ0 , φ1 , . . . , φn = 2π }
such that
max {|φ j − φ j−1 |} < ε .
j∈{1,...,n}
φ ∈ [φ j−1 , φ j ].
n
|eiφ − ∑ eiφk (hφk (φ ) − hφk−1 (φ ))| = |eiφ − eiφ j |
k=1
≤ |φ − φ j | < ε . (3.31)
88 3 Topics on Linear Operators
Thus,
n
0 ≤ |eiφ − ∑ eiφk (hφk (φ ) − hφk−1 (φ ))|2 ≤ ε 2
k=1
n n
0 ≤ [U − ∑ eiφk (E(φk ) − E(φk−1 ))]∗ [U − ∑ eiφk (E(φk ) − E(φk−1 ))]
k=1 k=1
≤ ε 2I
and hence
n
U − ∑ e (E(φk ) − E(φk−1 ) < ε .
iφk
k=1
Being ε > 0 arbitrary, we may infer that
2π
U= eiφ dE(φ ).
0
3.8.1 Introduction
Γ (A) = Γ (A).
3.8 Unbounded Operators 89
so that
A ⊂ C.
However C ⊂ B and since B is an arbitrary closed extension of A we have
C=A
so that
Γ (C) = Γ (A) = Γ (A).
Observe that by the Riesz lemma, φ ∈ D(A∗ ) if and only if there exists K > 0 such
that
|(Aψ , φ )H | ≤ K ψ H , ∀ψ ∈ D.
Also note that if
A ⊂ B then B∗ ⊂ A∗ .
Finally, as D is dense in H, then
η = A∗ (φ )
is uniquely defined. However the domain of A∗ may not be dense, and in some
situations we may have D(A∗ ) = {θ }.
If D(A∗ ) is dense, we define
A∗∗ = (A∗ )∗ .
90 3 Topics on Linear Operators
A = A∗∗ ,
V (φ , ψ ) = (−ψ , φ ).
V (φ , ψ ) = (−ψ , φ ) = (φ1 , ψ1 ).
Hence
ψ = −φ1 and φ = ψ1 ,
so that for (ψ1 , −φ1 ) ∈ E⊥ and (w1 , w2 ) ∈ E we have
Thus
(φ1 , −w2 )H + (ψ1 , w1 )H = 0,
and therefore
((φ1 , ψ1 ), (−w2 , w1 ))H×H = 0,
that is,
((φ1 , ψ1 ),V (w1 , w2 ))H×H = 0, ∀(w1 , w2 ) ∈ E.
This means that
(φ1 , ψ1 ) ∈ (V (E))⊥ ,
so that
V (E ⊥ ) ⊂ (V (E))⊥ .
It is easily verified that the implications from which the last inclusion results are
in fact equivalences, so that
V (E ⊥ ) = (V (E))⊥ .
(φ , Aψ )H = (η , ψ )H , ∀ψ ∈ D,
Γ (A) = [Γ (A)⊥ ]⊥
= V 2 [Γ (A)⊥ ]⊥
= [V [V (Γ (A))⊥ ]]⊥
= [V (Γ (A∗ )]⊥ (3.32)
Γ (A) = Γ [(A∗ )∗ ].
Conversely, suppose D(A∗ ) is not dense. Thus there exists ψ ∈ [D(A∗ )]⊥ such
that ψ = θ . Let (φ , A∗ φ ) ∈ Γ (A∗ ). Hence
((ψ , θ ), (φ , A∗ φ ))H×H = (ψ , φ )H = 0,
so that
(ψ , θ ) ∈ [Γ (A∗ )]⊥ .
Therefore V [Γ (A∗ )]⊥ is not the graph of a linear operator. Since Γ (A) =
V [Γ (A∗ )]⊥ A is not closable.
3. Observe that if A is closable, then
A∗ φ = Aφ , ∀φ ∈ D.
(Aφ , ψ )H = (φ , Aψ )H , ∀φ , ψ ∈ D.
92 3 Topics on Linear Operators
Proof.
• 1 implies 2:
Suppose A is self-adjoint, let φ ∈ D = D(A∗ ) be such that
Aφ = iφ
so that
A∗ φ = iφ .
Observe that
− i(φ , φ )H = (iφ , φ )H
= (Aφ , φ )H
= (φ , Aφ )H
= (φ , iφ )H
= i(φ , φ )H , (3.33)
N(A − iI) = {θ }.
A∗ φ = −iφ
and hence by above ψ = θ . Now we will prove that R(A − iI) is closed and
conclude that
R(A − iI) = H.
Given φ ∈ D we have
(A − iI)φ 2H = Aφ 2H + φ 2H . (3.34)
Let ψ0 ∈ H be a limit point of R(A − iI). Thus we may find {φn } ⊂ D such that
(A − iI)φn → ψ0 .
From (3.34)
φn − φm H ≤ (A − iI)(φn − φm ) H , ∀m, n ∈ N
R(A − iI) = H.
Similarly
R(A + iI) = H.
• 3 implies 1: Let φ ∈ D(A∗ ). Since R(A − iI) = H, there is an η ∈ D such that
(A∗ − iI)(φ − η ) = θ .
U = (A − i)(A + i)−1,
The operator U is called the Cayley transform of A. We have already proven that
2π
U= eiφ dF(φ ),
0
F(0) = F(0+ )
we obtain
F(0+ ) = 0 = F(0− ),
that is, F(φ ) is continuous at φ = 0. We claim that F is continuous at φ = 2π .
Observe that F(2π ) = F(2π + ) so that we need only to show that
F(2π − ) = F(2π ).
Suppose
F(2π ) − F(2π − ) = θ .
Thus, there exists some u, v ∈ H such that
Therefore
F(φ )v = F(φ )[(F(2π ) − F(2π − ))u],
so that
0, if φ < 2π ,
F(φ )v = (3.36)
v, if φ ≥ 2π .
3.9 Symmetric and Self-Adjoint Operators 95
Observe that
2π
U −I = (eiφ − 1)dF(φ ),
0
and
2π
U∗ − I = (e−iφ − 1)dF(φ ).
0
Let {φn } be a partition of [0, 2π ]. From the monotonicity of [0, 2π ] and pairwise
orthogonality of
{F(φn ) − F(φn−1 )}
we can show that (this is not proved in details here)
2π
(U ∗ − I)(U − I) = (e−iφ − 1)(eiφ − 1)dF(φ ),
0
The last two equalities result from e2π i − 1 = 0 and d F(φ )v = θ on [0, 2π ). Since
v = θ the last equation implies that 1 ∈ Pσ (U), which contradicts the existence of
(I − U)−1.
Thus, F is continuous at φ = 2π .
Now choose a sequence of real numbers {φn } such that φn ∈ (0, 2π ), n =
0, ±1, ±2, ±3, . . . such that
φn
− cot = n.
2
Now define Tn = F(φn ) − F(φn−1 ). Since U commutes with F(φ ), U commutes
with Tn . Since
A = i(I + U)(I − U)−1 ,
96 3 Topics on Linear Operators
this implies that the range of Tn is invariant under U and A. Observe that
∑ Tn = ∑(F(φn ) − F(φn−1 ))
n n
= lim F(φ ) − lim F(φ )
φ →2π φ →0
= I − θ = I. (3.38)
Hence
∑ R(Tn ) = H.
n
so that
2π
(I − U)Tnu = (1 − eiφ )dF(φ )Tn u
0
φn
= (1 − eiφ )dF(φ )u. (3.40)
φn−1
Therefore
φn
(1 − eiφ )−1 dF(φ )(I − U)Tn u
φn−1
φn φn
= (1 − eiφ )−1 dF(φ ) (1 − eiφ )dF(φ )u
φn−1 φn−1
φn
= (1 − eiφ )−1 (1 − eiφ )dF(φ )u
φn−1
φn
= dF(φ )u
φn−1
2π
= dF(φ )Tn u = Tn u. (3.41)
0
Hence
−1 φn
(I − U)|R(Tn ) = (1 − eiφ )−1 dF(φ ).
φn−1
we obtain
φn
ATn u = i(1 + eiφ )(1 − eiφ )−1 dF(φ )u.
φn−1
Therefore defining
φ
λ = − cot ,
2
and
E(λ ) = F(−2 cot−1 λ ),
we get
iφ iφ −1 φ
i(1 + e )(1 − e ) = − cot = λ.
2
Hence,
n
ATn u = λ dE(λ )u.
n−1
Finally, from
∞
u= ∑ Tn u,
n=−∞
we can obtain
∞
Au = A( ∑ Tn u)
n=−∞
∞
= ∑ ATn u
n=−∞
∞ n
= ∑ λ dE(λ )u. (3.42)
n=−∞ n−1
The main references for this chapter are Rudin [57], Royden [59], and Stein and
Shakarchi [62], where more details may be found. All these three books are excellent
and we strongly recommend their reading.
Remark 4.1.4.
1. Observe that 0/ = U \ U so that from 1 and 2 in Definition 4.1.1, we have that
0/ ∈ M .
2. From 1 and 3 from Definition 4.1.1, it is clear that ∪ni=1 Ai ∈ M whenever Ai ∈
M , ∀i ∈ {1, . . . , n}.
3. Since ∩∞ ∞
i=1 Ai = (∪i=1 Ai ) also from Definition 4.1.1, it is clear that M is closed
c c
Theorem 4.1.5. Let F be any collection of subsets of U. Then there exists a smallest
σ -algebra M0 in U such that F ⊂ M0 .
Proof. Let Ω be the family of all σ -algebras that contain F . Since the set of all
subsets in U is a σ -algebra, Ω is nonempty.
Let M0 = ∩Mλ ⊂Ω Mλ ; it is clear that M0 ⊃ F , and it remains to prove that in
fact M0 is a σ -algebra. Observe that:
1. U ∈ Mλ , ∀Mλ ∈ Ω , so that U ∈ M0 ,
2. A ∈ M0 implies A ∈ Mλ , ∀Mλ ∈ Ω , so that Ac ∈ Mλ , ∀Mλ ∈ Ω , which means
Ac ∈ M0 ,
3. {An } ⊂ M0 implies {An } ⊂ Mλ , ∀Mλ ∈ Ω , so that ∪∞
n=1 An ∈ Mλ , ∀Mλ ∈ Ω ,
which means ∪∞
n=1 An ∈ M0 .
From Definition 4.1.1 the proof is complete.
Definition 4.1.6 (Borel Sets). Let U be a topological space, considering the last
theorem, there exists a smallest σ -algebra in U, denoted by B, which contains the
open sets of U. The elements of B are called the Borel sets.
[−∞, α ) = ∪∞ ∞
n=1 [−∞, αn ] = ∪n=1 (αn , ∞] ,
C
(4.1)
K δn ≤ t ≤ (K + 1)δn . (4.3)
Defining
Kn (t)δn , i f 0 ≤ t < n,
ϕn (t) = (4.4)
n, i f t ≥ n,
4.3 Measures
A1 ⊂ A2 ⊂ A3 ⊂ . . . (4.6)
Proof.
1. Take An+1 = An+2 = . . . = 0/ in Definition 4.1.1 item 1.
2. Observe that B = A ∪ (B − A) and A ∩ (B − A) = 0/ so that by the above, μ (A ∪
(B − A)) = μ (A) + μ (B − A) ≥ μ (A).
3. Let B1 = A1 and let Bn = An − An−1; then Bn ∈ M , Bi ∩ B j = 0/ if i = j, An =
B1 ∪ . . . ∪ Bn , and A = ∪∞
i=1 Bi . Thus
∞ n
μ (A) = μ (∪∞
i=1 Bi ) = ∑ μ (Bi ) = lim
n→∞
∑ μ (Bi ) = lim μ (An ).
n→∞
(4.8)
n=1 i=1
Definition 4.4.1 (Integral for Simple Functions). For s : U → [0, ∞], a measurable
simple function, that is,
n
s = ∑ αi χAi , (4.10)
i=1
where
1, if u ∈ Ai ,
χAi (u) = (4.11)
0, otherwise,
we define the integral of s over E ⊂ M , denoted by E s d μ as
4.4 Integration of Simple Functions 103
n
s d μ = ∑ αi μ (Ai ∩ E). (4.12)
E i=1
where
Then,
(a) f is measurable,
(b) U fn d μ → U f d μ as n → ∞.
Proof. Since U fn d μ ≤ U f n+1 d μ , ∀n ∈ N, there exists α ∈ [0, ∞] such that
fn d μ → α , as n → ∞, (4.15)
U
Let s be any simple function such that 0 ≤ s ≤ f , and let c ∈ R such that 0 < c < 1.
For each n ∈ N we define
104 4 Basic Results on Measure and Integration
so that
α≥ sd μ , ∀s simple and measurable such that 0 ≤ s ≤ f . (4.21)
U
This implies
α≥ f dμ. (4.22)
U
We do not prove the next result (it is a direct consequence of the last theorem). For
a proof see [57].
Corollary 4.4.5. Let { fn } be a sequence of nonnegative measurable functions de-
fined on U ( fn : U → [0, ∞], ∀n ∈ N). Defining f (u) = ∑∞
n=1 f n (u), ∀u ∈ U, we have
∞
U
f dμ = ∑ fn d μ .
n=1 U
Then
gk ≤ f k (4.25)
so that
gk d μ ≤ fk d μ , ∀k ∈ N. (4.26)
U U
4.4 Integration of Simple Functions 105
Proof.
1. This inequality holds since f is measurable and | f | ≤ g.
2. Since 2g − | fn − f | ≥ 0, we may apply Fatou’s lemma and obtain
2gd μ ≤ lim inf (2g − | fn − f |)d μ , (4.32)
U n→∞ U
so that
lim sup | fn − f | d μ ≤ 0. (4.33)
n→∞ U
Hence
lim | fn − f | d μ = 0. (4.34)
n→∞ U
In this section we study signed measures. We start with the following definition.
Definition 4.5.1. Let (U, M ) be a measurable space. We say that a measure μ is
finite if μ (U) < ∞. On the other hand, we say that μ is σ -finite if there exists a
sequence {Un } ⊂ U such that U = ∪∞ n=1Un and μ (Un ) < ∞, ∀n ∈ N.
Definition 4.5.2 (Signed Measure). Let (U, M ) be a measurable space. We say that
ν : M → [−∞, +∞] is a signed measure if
• ν may assume at most one the values −∞, +∞,
• ν (0)
/ = 0,
• ν (∑∞ ∞
n=1 En ) = ∑n=1 ν (En ) for all sequence of measurable disjoint sets {En }.
Lemma 4.5.3. Considering the last definitions, we have that a countable union of
positive measurable sets is positive.
Proof. Let A = ∪∞
n=1 An where An is positive, ∀n ∈ N. Choose a measurable set
E ⊂ A. Set
En = (E ∩ An ) \ (∪i=1
n−1
Ai ).
Thus, En is a measurable subset of An so that ν (En ) ≥ 0. Observe that
E = ∪∞
n=1 En ,
Lemma 4.5.4. Considering the last definitions, let E be a measurable set such that
such that
ν (Ek ) < −1/nk .
Define
A = E \ (∪∞
k=1 Ek ) .
Then
E = A ∪ (∪∞
k=1 Ek ) .
is convergent so that nk → ∞ as k → ∞.
From ν (E) > 0 we must have ν (A) > 0.
Now, we will show that A is positive. Let ε > 0. Choose k sufficiently big such
that 1/(nk − 1) < ε .
Since
A ⊂ E \ ∪kj=1 E j ,
−1/(nk − 1) > −ε ,
Proof. Without losing generality, suppose ν does not assume the value +∞ (the
other case may be dealt similarly). Define
lim ν (An ) = λ .
n→∞
Define
A = ∪∞
i=1 Ai .
λ ≥ ν (A).
λ ≥ ν (E ∪ A)
= ν (E) + ν (A)
= ν (E) + λ , (4.35)
so that ν (E) = 0.
Thus, B contains no positive set of positive measure, so that by Lemma 4.5.4, B
contains no subsets of positive measure, that is, B is negative.
The proof is complete.
4.6 The Radon–Nikodym Theorem 109
Remark 4.5.6. Denoting the Hahn decomposition of U relating ν by {A, B}, we may
define the measures ν + and ν − by
ν + (E) = ν (E ∩ A),
and
ν − (E) = −ν (E ∩ B),
so that
ν = ν + − ν −.
We recall that two measures ν1 and ν2 are mutually singular if there are disjoint
measurable sets such that
U = A∪B
and
ν1 (A) = ν2 (B) = 0.
Observe that the measures ν+ and ν − above defined are mutually singular. The
decomposition
ν = ν+ − ν−
is called the Jordan one of ν . The measures ν + and ν − are called the positive and
negative parts of ν , respectively.
Observe that either ν + or ν − is finite since only one of the values +∞, −∞ may
be assumed by ν . We may also define
|F( f )| = |( f , 1)L2 (μ ) |
≤ f L2 (μ ) [μ (U)]1/2
≤ f L2 (λ ) [μ (U)]1/2 , (4.36)
since
f 2L2 (μ ) = | f |2 d μ ≤ | f |2 d λ = f 2L2 (λ ) .
U U
Thus,
f dμ = f g dλ ,
U U
and in particular,
f dμ = f g (d μ + d ν ).
U U
4.6 The Radon–Nikodym Theorem 111
Hence
f (1 − g) d μ = f g dν . (4.37)
U U
Assume, to obtain contradiction, that g < 0 in a set A such that μ (A) > 0.
Thus
(1 − g) d μ > 0,
A
so that from this and (4.37) with f = χA we get
g d ν > 0.
A
and hence
ν (B) = 0.
Thus, Bg d ν = 0 so that
(1 − g) d μ = 0,
B
which implies that μ (B) = 0, a contradiction.
From above we conclude that
0 ≤ g ≤ 1, a.e. [μ ] in U.
On the other hand, for a fixed E μ -measurable again from (4.37) with f = χE , we
get
(1 − g) d μ = g dν ,
E E
so that
(1 − g) d μ = g dν − d ν + ν (E),
E E E
and therefore
ν (E) = (1 − g) (d μ + d ν ),
E
that is,
ν (E) = (1 − g) d λ , ∀E ∈ M .
E
112 4 Basic Results on Measure and Integration
Define
B = {u ∈ U : g(u) = 0}.
Hence, μ (B) = B g d λ = 0.
From this, since λ μ , we obtain
g−1 g = 1, a.e. [λ ].
Thus,
ν (E) = g−1 d μ − μ (E)
E
= (g−1 − 1) d μ
E
= (1 − g)g−1 d μ , ∀E ∈ M . (4.39)
E
The proof for the finite case is complete. The proof for σ -finite is developed in the
next lines.
Since U is σ -finite, there exists a sequence {Un } such that U = ∪∞ n=1Un , and
μ (Un ) < ∞ and ν (Un ) < ∞, ∀n ∈ N.
Define
Fn = Un \ ∪n−1
j=1 U j ,
thus U = ∪∞ n=1 Fn and {Fn } is a sequence of disjoint sets, such that μ (Fn ) < ∞ and
ν (Fn ) < ∞, ∀n ∈ N.
Let E ∈ M . For each n ∈ N from above we may obtain fn such that
ν (E ∩ Fn ) = fn d μ , ∀E ∈ M .
E∩Fn
From this and the monotone convergence theorem corollary we may write
∞
ν (E) = ∑ ν (E ∩ Fn)
n=1
4.6 The Radon–Nikodym Theorem 113
∞
= ∑ fn d μ
n=1 E∩Fn
∞
= ∑ fn χFn d μ
n=1 E
∞
= ∑ fn χFn d μ
E n=1
= f dμ, (4.40)
E
where
∞
f= ∑ fn χFn .
n=1
ν = ν0 + ν1 .
λ = ν + μ.
Observe that ν and μ are absolutely continuous with respect to λ . Hence, by the
Radon–Nikodym theorem, there exist nonnegative measurable functions f and g
such that
μ (E) = f d λ , ∀E ∈ M
E
and
ν (E) = g d λ ∀E ∈ M .
E
Define
A = {u ∈ U | f (u) > 0},
and
B = {u ∈ U | f (u) = 0}.
Thus,
U = A ∪ B,
and
A ∩ B = 0.
/
114 4 Basic Results on Measure and Integration
Also define
ν0 (E) = ν (E ∩ B), ∀E ∈ M .
We have that ν0 (A) = 0 so that
ν0 ⊥ μ .
Define
ν1 (E) = ν (E ∩ A)
= g dλ . (4.41)
E∩A
Therefore,
ν = ν0 + ν1 .
To finish the proof, we have only to show that
ν1 μ .
and in particular
f d λ = 0.
(E∩A)
λ (A ∩ E) = 0.
ν (E ∩ A) = 0,
so that
ν1 (E) = ν (E ∩ A) = 0.
From this we may infer that
ν1 μ .
The proof of uniqueness is left to the reader.
1. μ ∗ (0)
/ = 0,
2. if A ⊂ B, then μ ∗ (A) ≤ μ ∗ (B), ∀A, B ⊂ U,
3. if E ⊂ ∪∞
n=1 En , then
∞
μ ∗ (E) ≤ ∑ μ ∗(En ).
n=1
μ ∗ (A) = μ ∗ (A ∩ E2 ) + μ ∗ (A ∩ E2c ),
Since
A ∩ (E1 ∪ E2 ) = (A ∩ E2 ) ∪ (A ∩ E1 ∩ E2c ),
we obtain
μ ∗ (A ∩ (E1 ∪ E2 )) ≤ μ ∗ (A ∩ E2 ) + μ ∗ (A ∩ E2c ∩ E1 ). (4.43)
From this and (4.42) we obtain
Hence E1 ∪ E2 is μ ∗ -measurable.
By induction, the union of a finite number of μ ∗ -measurable sets is μ ∗ -
measurable.
Assume E = ∪∞ ∗
i=1 Ei where {Ei } is a sequence of disjoint μ -measurable sets.
Define Gn = ∪ni=1 Ei . Then Gn is μ ∗ -measurable and for a given A ⊂ U we have
μ ∗ (A) = μ ∗ (A ∩ Gn ) + μ ∗ (A ∩ Gcn )
≥ μ ∗ (A ∩ Gn ) + μ ∗ (A ∩ E c ), (4.45)
since E c ⊂ Gcn , ∀n ∈ N.
116 4 Basic Results on Measure and Integration
Observe that
Gn ∩ E n = E n
and
Gn ∩ Enc = Gn−1 .
Thus, from the measurability of En , we may get
μ ∗ (A ∩ Gn ) = μ ∗ (A ∩ Gn ∩ En ) + μ ∗(A ∩ Gn ∩ Enc )
= μ ∗ (A ∩ En) + μ ∗ (A ∩ Gn−1). (4.46)
By induction we obtain
n
μ (A ∩ Gn ) = ∑ μ ∗ (A ∩ Ei ),
i=1
so that
n
μ ∗ (A) ≥ μ ∗ (A ∩ E c ) + ∑ μ ∗ (A ∩ Ei ), ∀n ∈ N,
i=1
we get
∞
μ ∗ (A) ≥ μ ∗ (A ∩ E c ) + ∑ μ ∗ (A ∩ Ei )
i=1
≥ μ ∗ (A ∩ E c ) + μ ∗ (A ∩ E). (4.47)
μ (E1 ∪ E2 ) = μ ∗ (E1 ∪ E2 )
= μ ∗ ((E1 ∪ E2 ) ∩ E2 ) + μ ∗((E1 ∪ E2 ) ∩ E2c )
= μ ∗ (E2 ) + μ ∗(E1 ). (4.48)
Therefore,
∞
μ (E) ≥ ∑ μ (Ei ).
i=1
and thus
∞
μ (E) = ∑ μ (Ei ).
i=1
∞
μ (E) = ∑ μ (Ei ).
i=1
where Ai ∈ A , ∀i ∈ N.
A ⊂ ∪∞
i=1 Ai .
Proof. Define
Bn = (A ∩ An ) \ (∪i=1
n−1
Ai ).
Thus
Bn ⊂ An , ∀n ∈ N,
Bn ∈ A , ∀n ∈ N, and
A = ∪∞
i=1 Bi .
118 4 Basic Results on Measure and Integration
Proof. The only not immediate property to be proven is the countably sub-
additivity.
Suppose E ⊂ ∪∞ ∗
i=1 Ei . If μ (Ei ) = +∞ for some i ∈ N, the result holds.
∗
Thus, assume μ (Ei ) < +∞, ∀i ∈ N.
Let ε > 0. Thus for each i ∈ N there exists {Ai j } ⊂ A such that Ei ⊂ ∪∞j=1 Ai j ,
and
∞
ε
∑ μ (Ai j ) ≤ μ ∗ (Ei ) + 2i .
j=1
Therefore,
∞ ∞ ∞
μ (E) ≤ ∑ ∑ μ (Ai j ) ≤ ∑ μ ∗ (Ei ) + ε .
i=1 j=1 i=1
∞
∑ μ (Ai) < μ ∗ (E) + ε .
i=1
Observe that
μ (Ai ) = μ (Ai ∩ A) + μ (Ai ∩ Ac ),
so that from the fact that
E ∩ A ⊂ ∪∞
i=1 (Ai ∩ A),
and
(E ∩ Ac ) ⊂ ∪∞
i=1 (Ai ∩ A ),
c
we obtain
∞ ∞
μ ∗ (E) + ε > ∑ (Ai ∩ A) + ∑ μ (Ai ∩ Ac )
i=1 i=1
≥ μ ∗ (E ∩ A) + μ ∗(E ∩ Ac ). (4.49)
4.7 Outer Measure and Measurability 119
μ ∗ (E) ≥ μ ∗ (E ∩ A) + μ ∗(E ∩ Ac ).
μ ∗ (E) = μ ∗ (B).
E ⊂ ∪∞
i=1 Ai
and
∞
∑ μ (Ai) ≤ μ ∗ (E) + ε .
i=1
Define A = ∪∞
i=1 Ai , then
∞
μ ∗ (A) ≤ ∑ μ ∗(Ai )
i=1
∞
= ∑ μ (Ai )
i=1
∗
≤ μ (E) + ε . (4.50)
E ⊂ An
and
μ ∗ (An ) ≤ μ (E) + 1/n.
Define B = ∩∞
n=1 An . Thus, B ∈ Aσ δ , E ⊂ B and
Hence
μ ∗ (B) = μ ∗ (E).
The proof is complete.
120 4 Basic Results on Measure and Integration
U = ∪∞
i=1Ui .
Observe that
E = ∪∞
i=1 Ei ,
where
Ei = E ∩Ui ,
is μ ∗ -measurable for each i ∈ N.
Let ε > 0. From the last proposition for each i, n ∈ N there exists Ani ∈ Aσ such
that
1
μ (Ani ) < μ ∗ (Ei ) + i .
n2
Define
An = ∪∞ i=1 Ani .
Thus,
E ⊂ An
and
An \ E ⊂ ∪∞
i=1 (Ani \ Ei ),
and therefore,
∞ ∞
1 1
μ (An \ E) ≤ ∑ μ (Ani \ Ei ) ≤ ∑ i
= .
i=1 i=1 n2 n
Since An ∈ Aσ , defining
A = ∩∞
n=1 An ,
Since B ⊂ A, we obtain
μ̃ (B) ≤ μ ∗ (B).
μ ∗ (A) ≤ μ ∗ (B) + ε
so that
μ ∗ (A) = μ ∗ (B) + μ ∗(A \ B),
from this and above
μ̃ (A \ B) ≤ μ ∗ (A \ B) ≤ ε ,
if μ ∗ (B) < ∞.
Therefore,
μ ∗ (B) ≤ μ̃ (B),
so that μ ∗ (B) = μ̃ (B). Finally, since μ is σ -finite, there exists a sequence of count-
able disjoint sets {Ui } such that μ (Ui ) < ∞, ∀i ∈ N, and U = ∪∞ i=1Ui .
122 4 Basic Results on Measure and Integration
If B ∈ B, then
B = ∪∞
i=1 (Ui ∩ B).
Remark 4.7.11. We may start the process of construction of a measure by the action
of a set function on a semi-algebra. Here, a semi-algebra C is a collection of subsets
of U such that the intersection of any two sets in C is in C and the complement of
any set in C is a finite disjoint union of sets in C .
If C is any semi-algebra of sets, then the collection consisting of the empty set
and all finite disjoint unions of sets in C is an algebra, which is said to be generated
by C . We denote such algebra by A .
If we have a set function acting on C , we may extend it to A by defining
n
μ (A) = ∑ μ (Ei ),
i=1
where A = ∪ni=1 Ei and Ei ∈ C , ∀i ∈ {1, . . ., n}, so that this last union is disjoint. We
recall that any A ∈ A admits such a representation.
Let (U, M1 , μ1 ) and (V, M2 , μ2 ) be two complete measure spaces. We recall that
the Cartesian product between U and V , denoted by U × V , is defined by
U × V = {(u, v) | u ∈ U and v ∈ V }.
(A × B) ∩ (C × D) = (A ∩C) × (B ∩ D),
and
(A × B)c = (Ac × B) ∪ (A × Bc) ∪ (Ac × Bc ).
We define λ : M1 × M2 → R+ by
λ (A × B) = μ1 (A)μ2 (B).
Hence, for the fixed u in question, from the corollary of Lebesgue monotone con-
vergence theorem, we may write
∞
V
χA×B(u, v)d μ2 (v) = ∑ χAi (u)χBi (v)d μ2 (v)
i=1
∞
= ∑ χAi (u)μ2 (Bi ) (4.53)
i=1
Observe that
d μ1 (u) χA×B (u, v)d μ2 (v) = d μ1 (u) χA (u)χB (v)d μ2 (v)
U V U V
= μ1 (A)μ2 (B).
Eu = {v | (u, v) ∈ E},
and
Ev = {u | (u, v) ∈ E}.
Observe that
χEu (v) = χE (u, v),
(E c )u = (Eu )c ,
and
(∪Eα )u = ∪(Eα )u ,
for any collection {Eα }.
We denote by Rσ as the collection of sets which are countable unions of mea-
surable rectangles. Also, Rσ δ will denote the collection of sets which are countable
intersections of elements of Rσ .
Lemma 4.8.5. Let u ∈ U and E ∈ Rσ δ . Then Eu is a measurable subset of V .
where Ei ∈ R, ∀i ∈ N. Thus,
χ(Ei )u (v)
4.8 The Fubini Theorem 125
χEu (v)
E = ∩∞
i=1 Ei ,
Thus as from above χ(Ei )u (v) is measurable for each i ∈ N, we have that χEu is also
measurable so that Eu is measurable.
Lemma 4.8.6. Let E be a set in Rσ δ with (μ1 × μ2 )(E) < ∞. Then the function g
defined by
g(u) = μ2 (Eu )
is a measurable function and
g d μ1 (u) = (μ1 × μ2 )(E).
U
gi (u) = μ2 ((Ei )u ).
U
g(u)d μ1 (u) = ∑ gi (u)d μ1 (u)
i=1 U
∞
= ∑ (μ1 × μ2)(Ei )
i=1
= (μ1 × μ2 )(E). (4.56)
126 4 Basic Results on Measure and Integration
Ei+1 ⊂ Ei
and
E = ∩∞
i=1 Ei .
we have that
g1 (u) < ∞ a.e. in E1 .
For an u ∈ E1 such that g1 (u) < ∞ we have that {(Ei )u } is a sequence of measurable
sets of finite measure whose intersection is Eu . Thus
that is,
gi → g, a.e. in E.
We may conclude that g is also measurable. Since
0 ≤ gi ≤ g, ∀i ∈ N
Lemma 4.8.7. Let E be a set such that (μ1 × μ2 )(E) = 0. Then for almost all u ∈ U
we have
μ2 (Eu ) = 0.
(μ1 × μ2)(F) = 0.
μ2 (Eu ) = 0
(μ1 × μ2 )(G) = 0.
Theorem 4.8.9 (Fubini). Let (U, M1 , μ1 ) and (V, M2 , μ2 ) be two complete measure
spaces and f an integrable function on U × V . Then
1. fu (v) = f (u, v) is measurable and integrable for almost all u,
2. fv (u) = f (u, v) is measurable and integrable for almost all v,
3. h1 (u) = V f (u, v) d μ2 (v) is integrable on U,
4. h2 (v) = U f (u, v) d μ1 (u) is integrable on V ,
5.
f d μ2 (v) d μ1 (u) = f d μ1 (u) d μ2 (v)
U V V U
= f d(μ1 × μ2 ). (4.58)
U×V
Proof. It suffices to consider the case where f is nonnegative (we can then apply the
result to f + = max( f , 0) and f − = max(− f , 0)). The last proposition asserts that
the theorem is true if f is a simple function which vanishes outside a set of finite
measure. Similarly as in Theorem 4.2.2, we may obtain a sequence of nonnegative
simple functions {φn } such that
φn ↑ f .
128 4 Basic Results on Measure and Integration
(φn )u ↑ fu , a.e. .
so that this last resulting function is integrable in U. Again by the Lebesgue mono-
tone convergence theorem, we obtain
f d μ2 (v) d μ1 (u) = lim φn d μ2 (v) d μ1 (u)
U V n→∞ U V
= lim φn d(μ1 × μ2 )
n→∞ U×V
= f d(μ1 × μ2 ). (4.59)
U×V
Chapter 5
The Lebesgue Measure in Rn
5.1 Introduction
In this chapter we will define the Lebesgue measure and the concept of Lebesgue
measurable set. We show that the set of Lebesgue measurable sets is a σ -algebra
so that the earlier results, proven for more general measure spaces, remain valid in
the present context (such as the Lebesgue monotone and dominated convergence
theorems). The main reference for this chapter is [62].
We start with the following theorems without proofs.
Theorem 5.1.1. Every open set A ⊂ R may be expressed as a countable union of
disjoint open intervals.
Remark 5.1.2. In this text Q j denotes a closed cube in Rn and |Q j | its volume, that
is, |Q j | = ∏ni=1 (bi − ai ), where Q j = ∏ni=1 [ai , bi ]. Also we assume that if two Q1 and
Q2 , closed or not, have the same interior, then |Q1 | = |Q2 | = |Q̄1 |. We recall that two
cubes Q1 , Q2 ⊂ Rn are said to be quasi-disjoint if their interiors are disjoint.
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 129
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 5,
© Springer International Publishing Switzerland 2014
130 5 The Lebesgue Measure in Rn
First observe that given ε > 0, there exists a sequence {Q j } such that
E ⊂ ∪∞j=1 Q j
and
∞
∑ |Q j | ≤ m∗ (E) + ε .
j=1
1. Monotonicity: If E1 ⊂ E2 then m∗ (E1 ) ≤ m∗ (E2 ). This follows from the fact that
if E2 ⊂ ∪∞j=1 Q j then E1 ⊂ ∪∞j=1 Q j .
2. Countable sub-additivity : If E ⊂ ∪∞j=1 E j , then m∗ (E) ≤ ∑∞j=1 m∗ (E j ).
E j ⊂ ∪∞
k=1 Qk, j
and
∞
ε
∑ |Qk, j | < m∗ (E j ) + 2 j .
k=1
Hence
E ⊂ ∪∞j,k=1 Qk, j
and therefore
∞ ∞ ∞
∗
m (E) ≤ ∑ |Qk, j | = ∑ ∑ |Qk, j |
j,k=1 j=1 k=1
∞
ε
≤ ∑ m∗ (E j ) + j
2
j=1
∞
= ∑ m∗ (E j ) + ε . (5.1)
j=1
3. If
E ⊂ Rn ,
and
α = inf{m∗ (A) | A is open and E ⊂ A},
5.2 Properties of the Outer Measure 131
then
m∗ (E) = α .
Proof. From the monotonicity, we have
Thus
m∗ (E) ≤ α .
Suppose given ε > 0. Choose a sequence {Q j } of closed cubes such that
E ⊂ ∪∞j=1 Q j
and
∞
∑ |Q j | ≤ m∗ (E) + ε .
j=1
Therefore
α ≤ m∗ (E) + 2ε .
Being ε > 0 arbitrary, we have
α ≤ m∗ (E).
E ⊂ ∪∞j=1 Q j ,
and
∞
∑ |Q j | ≤ m∗ (E) + ε .
j=1
Hence,
m∗ (E1 ) + m∗ (E2 ) ≤ ∑ |Q j | + ∑ |Q j |
j∈J1 j∈J2
∞
≤ ∑ |Q j | ≤ m∗ (E) + ε . (5.3)
j=1
E = ∪∞j=1 Q j ,
then
∞
m∗ (E) = ∑ |Q j |.
j=1
Thus, for each N ∈ N the cubes Q̃1 , . . . , Q̃N are disjoint and each pair have a finite
distance. Hence,
N N ε
m∗ (∪Nj=1 Q̃ j ) = ∑ |Q̃ j | ≥ ∑ |Q j | − j .
2
j=1 j=1
Being
∪Nj=1 Q̃ j ⊂ E,
we obtain
N N
m∗ (E) ≥ ∑ |Q̃ j | ≥ ∑ |Q j | − ε .
j=1 j=1
Therefore
∞
∑ |Q j | ≤ m∗ (E) + ε .
j=1
E ⊂A
and
m∗ (A − E) ≤ ε .
If E is measurable, we define its Lebesgue measure, denoted by m(E), as
m(E) = m∗ (E).
Proof. Let E ⊂ Rn be such that m∗ (E) = 0. Suppose given ε > 0, thus there
exists A ⊂ Rn open such that E ⊂ A and m∗ (A) < ε . Therefore
m∗ (A − E) < ε .
(A − E) ⊂ ∪∞j=1 (A j − E j ).
where Bk denotes a closed ball of radius k with center at origin. Thus F may be
expressed as a countable union of compact sets. Hence, we have only to show
that if F is compact then it is measurable. Let F be a compact set. Observe that
m∗ (F) < ∞.
m∗ (A) ≤ m∗ (F) + ε .
A − F = ∪∞j=1 Q j .
For each N ∈ N
K = ∪Nj=1 Q j
5.4 Properties of Measurable Sets 135
is compact; therefore
d(K, F) > 0.
Being K ∪ F ⊂ A, we have
N
m∗ (A) ≥ m∗ (F ∪ K) = m∗ (F) + m∗ (K) = m∗ (F) + ∑ |Q j |.
j=1
Therefore
N
∑ |Q j | ≤ m∗ (A) − m∗(F) ≤ ε .
j=1
Finally,
∞
m∗ (A − F) ≤ ∑ |Q j | < ε .
j=1
E c − S ⊂ Ak − E, ∀k ∈ N.
Therefore
1
m∗ (E c − S) ≤ , ∀k ∈ N.
k
Thus
m∗ (E c − S) = 0.
This means that E c − S is measurable, so that
E c = S ∪ (E c − S)
Proof. First assume that E j is bounded. Being E cj measurable, given ε > 0, there
exists an open H j ⊃ E cj such that
ε
m∗ (H j − E cj ) < , ∀ j ∈ N.
2j
Denoting Fj = H cj we have that Fj ⊂ E j is closed and
ε
m∗ (E j − Fj ) < , ∀ j ∈ N.
2j
For each N ∈ N the sets F1 , . . . , FN are compact and disjoint, so that
N
m(∪Nj=1 Fj ) = ∑ m(Fj ).
j=1
As
∪Nj=1 Fj ⊂ E
we have
N N
m(E) ≥ ∑ m(Fj ) ≥ ∑ m(E j ) − ε .
j=1 j=1
Hence
∞
m(E) ≥ ∑ m(E j ) − ε .
j=1
For the general case, select a sequence of cubes {Qk } such that
Rn = ∪∞
k=1 Qk
5.4 Properties of Measurable Sets 137
E j,k = E j ∩ Sk , ∀ j, k ∈ N.
Thus
E = ∪∞j=1 ∪∞ ∞
k=1 E j,k = ∪ j,k=1 E j,k ,
where such a union is disjoint and each E j,k is bounded. Through the last result,
we get
∞
m(E) = ∑ m(E j,k )
j,k=1
∞ ∞
= ∑ ∑ m(E j,k )
j=1 k=1
∞
= ∑ m(E j ). (5.4)
j=1
m(A − E) < ε .
m(E − F) < ε .
m(E \ K) < ε .
F = ∪Nj=1 Q j
such that
m(E F) ≤ ε ,
where
E F = (E \ F) ∪ (F \ E).
Proof.
1. This item follows from the definition of measurable set.
2. Being E c measurable, there exists an open B ⊂ Rn such that E c ⊂ B and
m∗ (B \ E c ) < ε .
138 5 The Lebesgue Measure in Rn
Being m(E) < ∞ the series converges and there exists N0 ∈ N such that
∞
ε
∑ |Q j | < .
2
N0 +1
N
Defining F = ∪ j=1
0
Q j , we have
Observe that:
1. If
f −1 ([−∞, a))
is measurable for each a ∈ R, then
f −1 ([−∞, a]) = ∩∞
k=1 f
−1
([−∞, a + 1/k))
f −1 ([−∞, a)) = ∪∞
k=1 f
−1
([−∞, a − 1/k])
f −1 ([−∞, a))
f −1 ((a, b))
Proof. The first implication is obvious. For the second, being φ continuous
φ −1 ([−∞, a))
is measurable, ∀a ∈ R.
140 5 The Lebesgue Measure in Rn
and
lim sup fk (x), lim inf fk (x)
k→∞ k→∞
are measurable.
Proof. We will prove only that supn∈N fn (x) is measurable. The remaining proofs
are analogous. Let
f (x) = sup fn (x).
n∈N
Thus
f −1 ((a, +∞]) = ∪∞ −1
n=1 f n ((a, +∞]).
inf fk (x)
k∈N
is measurable and
lim sup fk (x) = inf sup f j (x),
k→∞ k≥1 j≥k
and
lim inf fk (x) = sup inf f j (x)
k→∞ k≥1 j≥k
are measurable.
Then f is measurable.
The next result we do not prove it. For a proof see [62].
Proposition 5.5.5. If f and g are measurable functions, then
1. f 2 is measurable,
2. f + g and f · g are measurable if both assume finite values.
5.5 Lebesgue Measurable Functions 141
Observe that
Proof. Let N ∈ N. Let QN be the cube with center at origin and side of measure N.
Define ⎧
⎨ f (x), if x ∈ QN and f (x) ≤ N,
FN (x) = N, if x ∈ QN and f (x) > N,
⎩
0, otherwise.
Thus FN (x) → f (x) as N → ∞, ∀x ∈ Rn . Fixing M, N ∈ N define
l l+1
El,M = x ∈ QN : ≤ FN (x) ≤ ,
M M
for 0 ≤ l ≤ N · M. Defining
NM
l
FN,M = ∑ M χEl,M ,
l=0
142 5 The Lebesgue Measure in Rn
1
0 ≤ FN (x) − FN,M (x) ≤ .
M
If ϕK (x) = FK,K (x), we obtain
1
0 ≤ |FK (x) − ϕK (x)| ≤ .
K
Hence
| f (x) − ϕK (x)| ≤ | f (x) − FK (x)| + |FK (x) − ϕK (x)|.
Therefore
lim | f (x) − ϕK (x)| = 0, ∀x ∈ Rn .
K→∞
and
lim ϕk (x) = f (x), ∀x ∈ Rn .
k→∞
Proof. Write
f (x) = f + (x) − f − (x),
where
f + (x) = max{ f (x), 0}
and
f − (x) = max{− f (x), 0}.
Thus f + and f − are nonnegative measurable functions so that from the last theorem
there exist increasing sequences of nonnegative simple functions such that
(1)
ϕk (x) → f + (x), ∀x ∈ Rn ,
and
(2)
ϕk (x) → f − (x), ∀x ∈ Rn ,
as k → ∞. Defining
(1) (2)
ϕk (x) = ϕk (x) − ϕk (x),
we obtain
ϕk (x) → f (x), ∀x ∈ Rn
5.5 Lebesgue Measurable Functions 143
as k → ∞ and
(1) (2)
|ϕk (x)| = ϕk (x) + ϕk (x) " | f (x)|, ∀x ∈ Rn ,
as k → ∞.
Proof. From the last theorem, it suffices to prove that if E is measurable and
m(E) < ∞, then χE may be approximated almost everywhere in E by step func-
tions. Suppose given ε > 0. Observe that from Proposition 5.4.2, there exist cubes
Q1 , . . . , QN such that
m(E ∪Nj=1 Q j ) < ε .
We may obtain almost disjoints rectangles R̃ j such that ∪M
j=1 R̃ j = ∪ j=1 Q j and dis-
N
Thus
M
f (x) = ∑ χR j ,
j=1
possibly except in a set of measure < 2ε . Hence, for each k > 0, there exists a step
function ϕk such that m(Ek ) < 2−k where
Ek = {x ∈ Rn | f (x) = ϕk (x)}.
Defining
Fk = ∪∞j=k+1 E j
we have
∞
m(Fk ) ≤ ∑ m(E j )
j=k+1
∞
≤ ∑ 2− j
j=k+1
2−(k+1)
=
1 − 1/2
= 2−k . (5.8)
m(F) ≤ 2−k , ∀k ∈ N.
144 5 The Lebesgue Measure in Rn
fk → f , ∀x ∈ E.
EkN ⊂ Ek+1
N
and that ∪∞
k=1 Ek = E. Thus we may obtain kN such that
N
1
m(E − EkNN ) < .
2N
Observe that
1
| f j (x) − f (x)| < , ∀ j ≥ kN , x ∈ EkNN .
N
Choose M ∈ N such that
∞
ε
∑ 2−k ≤ 2 .
k=M
Define
Ãε = ∩∞
N≥M EkN .
N
Thus
∞
ε
m(E − Ãε ) ≤ ∑ m(E − EkNN ) < .
2
N=M
Suppose given δ > 0. Let N ∈ N be such that N > M and 1/N < δ . Thus if x ∈ Ãε
then x ∈ EkNN so that
| f j (x) − f (x)| < δ , ∀ j > kN .
Hence fk → f uniformly in Ãε . Observe that Ãε is measurable and thus there exists
a closed set Aε ⊂ Ãε such that
ε
m(Ãε − Aε ) < .
2
5.5 Lebesgue Measurable Functions 145
That is
ε ε
m(E − Aε ) ≤ m(E − Ãε ) + m(Ãε − Aε ) < + = ε,
2 2
and
fk → f
uniformly in Aε . The proof is complete.
Definition 5.5.12. We say that a set A ⊂ L1 (Rn ) is dense in L1 (Rn ), if for each
f ∈ L1 (Rn ) and each ε > 0 there exists g ∈ A such that
f − g L1(Rn ) = | f − g| dx < ε .
Rn
Proof.
1. From the last theorems given f ∈ L1 (Rn ) there exists a sequence of simple func-
tions such that
ϕk (x) → f (x) a.e. in Rn .
Since {ϕk } may be also such that
|ϕk | ≤ | f |, ∀k ∈ N
ϕk − f L1 (Rn ) → 0,
as k → ∞.
2. From the last item, it suffices to show that simple functions may be approximated
by step functions. As a simple function is a linear combination of characteristic
functions of sets of finite measure, it suffices to prove that given ε > 0 and a set
of finite measure, there exists ϕ a step function such that
χE − ϕ L1 (Rn ) < ε .
Thus
f − g L1 (Rn ) < 2ε .
for the general case of a rectangle in Rn , we just recall that in this case f is the
product of the characteristic functions of n intervals. Therefore we may approxi-
mate f by the product of n functions similar to g defined above.
Chapter 6
Other Topics in Measure and Integration
In this chapter we present some important results which may be found in similar
form at Chapters 2, 6, and 7 in the excellent book Real and Complex Analysis, [57]
by Rudin, where more details may be found.
then
f = 0, a.e. in E.
Proof. Define
An = {u ∈ E | f (u) > 1/n}, ∀n ∈ N.
Thus
μ (An )/n ≤ f dμ ≤ f d μ = 0.
An E
Therefore μ (An ) = 0, ∀n ∈ N.
Define
A = {u ∈ E | f (u) > 0}.
Hence,
A = ∪∞
n=1 An ,
so that μ (A) = 0.
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 147
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 6,
© Springer International Publishing Switzerland 2014
148 6 Other Topics in Measure and Integration
Thus,
f = 0, a.e. in E.
Theorem 6.1.2. Assume f ∈ L1 (μ ) and E f d μ = 0, ∀E ∈ M . Under such hypothe-
ses, f = 0, a.e. in U.
Proof. Consider first the case f : U → [−∞, +∞]. Define
Thus,
μ (An )/n ≤ f d μ = 0.
An
Hence, μ (An ) = 0, ∀n ∈ N.
Define
A = {u ∈ E | f (u) > 0}.
Therefore,
A = ∪∞
n=1 An ,
so that μ (A) = 0.
Thus,
f ≤ 0, a.e. in U.
By analogy we get
f ≥ 0, a.e. in U,
so that
f = 0, a.e. in U.
To complete the proof, just apply this last result to the real and imaginary parts of a
complex f .
Theorem 6.1.3. Suppose μ (U) < ∞ and f ∈ L1 (μ ). Moreover, assume
E |f| dμ
≤ α ∈ [0, ∞), ∀E ∈ M .
μ (E)
| f | ≤ α , a.e. in U.
Proof. Define
An = {u ∈ U | | f (u)| > α + 1/n}, ∀n ∈ N.
Thus, if μ (An ) > 0, we get
An (| f | − α ) d μ An | f |dμ
1/n ≤ = − α ≤ 0,
μ (An ) μ (An )
Define
A = {u ∈ U | | f (u)| > α }.
Therefore,
A = ∪∞
n=1 An ,
so that μ (A) = 0.
Thus,
| f (u)| ≤ α , a.e. in U.
The proof is complete.
At this point we present some preliminary results to the development of the well-
known Urysohn’s lemma.
Proof. For each u ∈ K there exist open sets Wu ,Vvu ⊂ U such that u ∈ Wu , v ∈ Wvu ,
and Wu ∩Vvu = 0.
/
Observe that K ⊂ ∪u∈K Wu so that, since K is compact, there exist u1 , u2 , . . . , un ∈
K such that
K ⊂ ∪ni=1Wui .
Finally, defining the open sets
V = ∩ni=1Vvui
and
W = ∪ni=1Wui ,
we get
V ∩W = 0,
/
v ∈ V , and K ⊂ W .
The proof is complete.
Hence
Kα0 ⊂ [∩α ∈L\{α0 } Kα ]c ,
that is,
Kα0 ⊂ ∪α ∈L\{α0 } Kαc = ∪α ∈L\{α0 }Vα .
150 6 Other Topics in Measure and Integration
Proof. Let u ∈ K. Since U is locally compact there exists an open Vu ⊂ U such that
u ∈ Vu and V u is compact.
Observe that
K ⊂ ∪u∈K Vu
and since K is compact there exist u1 , u2 , . . . , un ∈ K such that
K ⊂ ∪nj=1Vu j .
K ⊂ G,
where G is compact.
If W = U define V = G and the proof would be complete.
Otherwise, if W = U define C = U \ W . From Theorem 6.1.4, for each v ∈ C,
there exists an open Wv such that K ⊂ Wv and v ∈ W v .
Hence {C ∩ G ∩ W v : v ∈ C} is a collection of compact sets with empty
intersection.
From Theorem 6.1.5 there are points v1 , . . . , vn ∈ C such that
C ∩ G ∩W v1 ∩ . . . ∩W vn = 0.
/
Defining
V = G ∩Wv1 ∩ . . . ∩Wvn
we obtain
V ⊂ G ∩W v1 ∩ . . . ∩W vn .
Also,
K ⊂ V ⊂ V ⊂ W.
This completes the proof.
6.1 Some Preliminary Results 151
Observe that from this last definition f is continuous if and only if it is both lower
and upper semicontinuous.
Here we state and prove a very important result, namely, the Uryshon’s lemma.
and
r j = min{rk , | k ∈ {1, . . . , n} and rk > rn+1 }.
We may find again an open set Vrn+1 such that
Thus, we have obtained a sequence Vr of open sets such that for every r rational in
(0, 1), V r is compact and if s > r then V s ⊂ Vr . Define
r, if u ∈ Vr ,
fr (u) =
0, otherwise,
and
1, if u ∈ V s ,
gs (u) =
s, otherwise.
Also define
f (u) = sup fr (u), ∀u ∈ V
r∈Q∩(0,1)
and
g(u) = inf gs (u), ∀u ∈ V.
s∈Q∩(0,1)
152 6 Other Topics in Measure and Integration
h1 = g1
h2 = (1 − g1)g2
h3 = (1 − g1)(1 − g2)g3
... ...............
hn = (1 − g1)(1 − g2) . . . (1 − gn−1)gn . (6.1)
Thus,
0 ≤ hi ≤ 1 and hi ∈ Cc (Vi ), ∀i ∈ {1, . . ., n}.
Furthermore, by induction, we may obtain
h1 + h2 + . . . + hn = 1 − (1 − g1)(1 − g2) . . . (1 − gn).
Finally, if u ∈ K then u ∈ W̃i for some i ∈ {1, . . . , n}, so that gi (u) = 1 and hence
(h1 + . . . + hn )(u) = 1, ∀u ∈ K.
The set {h1 , . . . , hn } is said to be a partition of unity on K subordinate to the open
cover {V1, . . . ,Vn }.
The proof is complete.
In the next lines we introduce the main result in this section, namely, the Riesz
representation theorem.
Theorem 6.2.1 (Riesz Representation Theorem). Let U be a locally compact
Hausdorff space and let F be a positive linear functional on Cc (U). Then there
exists a σ -algebra M in U which contains all the Borel sets and there exists a
unique positive measure μ on M such that
1. F( f ) = U f d μ , ∀ f ∈ Cc (U),
2. μ (K) < ∞, for every compact K ⊂ U,
3. μ (E) = inf{μ (V ) | E ⊂ V, V open }, ∀E ∈ M ,
4. μ (E) = sup{μ (K) | K ⊂ E, K compact} holds for all open E and all E ∈ M
such that μ (E) < ∞,
5. If E ∈ M , A ⊂ E and μ (E) = 0 then A ∈ M .
154 6 Other Topics in Measure and Integration
μ1 (K) = μ2 (K)
for every compact K ⊂ U. Let ε > 0. Fix a compact K ⊂ U. By 2 and 3, there exists
an open V ⊃ K such that
μ2 (V ) < μ2 (K) + ε .
By the Urysohn’s lemma, there exists a f ∈ Cc (V ) such that
0 ≤ f (u) ≤ 1, ∀u ∈ V
and
f (u) = 1, ∀u ∈ K.
Thus,
μ1 (K) = χ K d μ1
U
≤ f d μ1
U
= F( f )
= f d μ2
U
≤ χV d μ2
U
= μ2 (V )
< μ2 (K) + ε . (6.2)
μ1 (K) ≤ μ2 (K).
μ2 (K) ≤ μ1 (K),
so that
μ1 (K) = μ2 (K).
The proof of uniqueness is complete.
Now for every open V ⊂ U, define
μ (V1 ) ≤ μ (V2 ).
Hence,
μ (E) = inf{μ (V ) | E ⊂ V, V open},
if E is an open set. Define
Finally, define by M the collection of all sets such that E ⊂ U and E ∩ K ∈ MF for
all compact K ⊂ U. Since
μ (A) ≤ μ (B),
if A ⊂ B we have that μ (E) = 0 implies E ∩ K ∈ MF for all K compact, so that
E ∈ M . Thus, 5 holds and so does 3 by definition.
Observe that if f ≥ 0, then F( f ) ≥ 0, that is, if f ≤ g then F( f ) ≤ F(g).
Now we prove that if {En } ⊂ U is a sequence, then
∞
μ (∪∞
n=1 En ) ≤ ∑ μ (En ). (6.3)
n=1
0 ≤ g ≤ 1.
By Theorem 6.1.10 there exist functions h1 and h2 such that hi ∈ Cc (Vi ) and
0 ≤ hi ≤ 1
and so that h1 + h2 = 1 on the support of g. Hence, hi ∈ Cc (Vi ), 0 ≤ hi g ≤ 1, and
g = (h1 + h2)g and thus
F(g) = F(h1 g) + F(h2 g) ≤ μ (V1 ) + μ (V2).
Since g is arbitrary, from the definition of μ , we obtain
Let a not relabeled ε > 0. Therefore for each n ∈ N there exists an open Vn ⊃ En
such that
ε
μ (Vn ) < μ (En ) + n .
2
Define
V = ∪∞ n=1Vn ,
∞
n=1 En ) ≤ μ (V ) ≤ ∑ μ (En ) + ε .
μ (∪∞
i=1
Vα = {u ∈ U | f (u) > α }.
μ (V ) < μ (K) + ε .
F( f ) ≤ μ (K),
so that
α < μ (K),
where K ⊂ U is a compact set.
Hence (6.7) holds.
Suppose that
E = ∪∞
n=1 En ,
μ (E) ≥ μ (KN )
N
= ∑ μ (Hn )
n=1
N
≥ ∑ μ (En ) − ε , ∀N ∈ N. (6.11)
n=1
Let ε0 > 2ε . If μ (E) < ∞, there exists N0 ∈ N such that μ (KN0 ) > ∑∞
n=1 μ (En ) − ε0 .
From this and (6.12) we obtain
μ (E) ≤ μ (KN0 ) + ε0 .
Therefore, since ε > 0 and ε0 > 2ε are arbitrary, we may conclude that E satisfies 4
so that E ∈ MF .
Now we prove the following.
6.2 The Riesz Representation Theorem 159
so that
μ (V \ K) < ε .
In the next lines we prove that if A, B ∈ MF then
A \ B, A ∪ B and A ∩ B ∈ MF .
By above there exist compact sets K1 , K2 and open sets V1 ,V2 such that
K1 ⊂ A ⊂ V1 , K2 ⊂ B ⊂ V2
and
μ (Vi \ Ki ) < ε , ∀i ∈ {1, 2}.
Since
(A \ B) ⊂ (V1 \ K2 ) ⊂ (V1 \ K1 ) ∪ (K1 \ V2) ∪ (V2 \ K2 ),
we get
μ (A \ B) < ε + μ (K1 \ V2) + ε ,
Since K1 \ V2 ⊂ A \ B is compact and ε > 0 is arbitrary, we get
A \ B ∈ MF .
Since
A ∪ B = (A \ B) ∪ B,
we obtain
A ∪ B ∈ MF .
Since
A ∩ B = A \ (A \ B)
we get
A ∩ B ∈ MF .
160 6 Other Topics in Measure and Integration
At this point we prove that M is a σ -algebra in U which contains all the Borel sets.
Let K ⊂ U be a compact subset. If A ∈ M then
Ac ∩ K = K \ (A ∩ K),
so that Ac ∩ K ∈ MF considering that K ∈ MF and A ∩ K ∈ MF .
Thus if A ∈ M then Ac ∈ M .
Next suppose
A = ∪∞n=1 An ,
where An ∈ M , ∀n ∈ N.
Define B1 = A1 ∩ K and
Bn = (An ∩ K) \ (B1 ∪ B2 ∪ . . . ∪ Bn−1 ),
∀n ≥ 2, n ∈ N.
Then {Bn } is disjoint sequence of sets in MF .
Thus
A ∩ K = ∪∞ n=1 Bn ∈ MF .
μ (V \ K) < ε .
μ (E ∩ K) < μ (K1 ) + ε .
Since
E ⊂ (E ∩ K) ∪ (V \ K),
it follows that
μ (E) ≤ μ (E ∩ K) + μ (V \ K) < μ (K1 ) + 2ε .
This implies E ∈ MF .
To finish the proof, we show that
F( f ) = f d μ , ∀ f ∈ Cc (U).
U
From linearity it suffices to prove the result for the case where f is real.
6.2 The Riesz Representation Theorem 161
Let f ∈ Cc (U). Let K be the support of f and let [a, b] ⊂ R be such that
R( f ) ⊂ (a, b),
Hence
n
f = ∑ hi f
i=1
and
n n
μ (K) ≤ F ∑ hi f = ∑ F(hi f ).
i=1 i=1
Observe that
ε
μ (Ei ) + > μ (Vi )
n
= sup{F( f ) | f ∈ Cc (Vi ), 0 ≤ f ≤ 1}
> F(hi ), ∀i ∈ {1, . . . , n}. (6.13)
162 6 Other Topics in Measure and Integration
Thus
n
F( f ) = ∑ F(hi f )
i=1
n
≤ ∑ F(hi (yi−1 + 2ε ))
i=1
n
= ∑ (yi−1 + 2ε )F(hi )
i=1
n ε
< ∑ (yi−1 + 2ε ) μ (Ei ) +
n
i=1
n n
ε n
< ∑ yi−1 μ (Ei ) + ∑ (yi−1 ) n + 2ε ∑ μ (Ei ) + 2ε 2
i=1 i=1 i=1
< f d μ + bε + 2ε μ (K) + 2ε . 2
(6.14)
U
From this
F(− f ) ≤ (− f ) d μ , ∀ f ∈ Cc (U),
U
that is,
F( f ) ≥ f d μ , ∀ f ∈ Cc (U).
U
Hence
F( f ) = f d μ , ∀ f ∈ Cc (U).
U
The proof is complete.
Since K is compact, there exists a finite number of such balls which covers K. By
Lemma 6.3.1, there exists a disjoint subcollection here denoted by {Br1 , . . . , BrN }
such that K ⊂ ∪Nk=1 B3rk , so that
N
m(K) ≤ 3n ∑ m(Brk )
k=1
N
≤ 3n λ −1 ∑ |μ |(Brk )
k=1
≤ 3n λ −1 μ . (6.15)
η (Br (u))
G f (u) = sup
0<r<∞ m(B r (u))
and
Aλ = {u ∈ U | G f (u) > λ },
we have
m(Aλ ) ≤ 3n λ −1 f 1 .
Finally in this section we present the main definition of Lebesgue points and
some relating results.
Definition 6.3.4. Let f ∈ L1 (Rn ). A point u ∈ L1 (Rn ) such that
1
lim | f (v) − f (u)| dm(v) = 0
r→0 m(Br (u)) Br (u)
Proof. Define
1
Hr f (u) = | f − f (u)| dm, ∀u ∈ Rn , r > 0,
m(Br (u)) Br (u)
f − g 1 < 1/k.
so that
Hh < Gh + |h|.
Since
Hr f ≤ Hrg + Hrh ,
we obtain
H f ≤ Gh + |h|.
Define
Ay = {u ∈ Rn | H f (u) > 2y},
By,k = {u ∈ Rn | Gh (u) > y},
and
Cy,k = {u ∈ Rn | |h| > y}.
Observe that h 1 < 1/k, so that from Remark 6.3.3 we obtain
3n
m(By,k ) ≤
yk
and
1
m(Cy,k ) ≤
yk
and hence
3n + 1
m(By,k ∪Cy,k ) ≤ .
yk
166 6 Other Topics in Measure and Integration
Therefore
3n + 1
m(Ay ) ≤ m(By,k ∪Cy,k ) ≤ .
yk
Since k is arbitrary, we get m(Ay ) = 0, ∀y > 0 so that m{u ∈ Rn | H f (u) > 0} = 0.
The proof is complete.
Definition 7.1.1 (Test Functions, the Space D(Ω )). Let Ω ⊂ Rn be a nonempty
open set. For each K ⊂ Ω compact, consider the space DK , the set of all C∞ (Ω )
functions with support in K. We define the space of test functions, denoted by
D(Ω ) as
2. σ̂ denotes the collection of all convex balanced sets W ∈ D(Ω ) such that W ∩
DK ⊂ σK for every compact K ⊂ Ω .
3. We define σ in D(Ω ) as the collection of all unions of sets of the form φ + W ,
for φ ∈ D(Ω ) and W ∈ σ̂ .
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 167
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 7,
© Springer International Publishing Switzerland 2014
168 7 Distributions
Proof.
1. From item 3 of Definition 7.1.2, it is clear that arbitrary unions of elements of
σ are elements of σ . Let us now show that finite intersections of elements of σ
also belong to σ . Suppose V1 ∈ σ and V2 ∈ σ ; if V1 ∩ V2 = 0,
/ we are done. Thus,
suppose φ ∈ V1 ∩ V2 . By the definition of σ there exist two sets of indices L1 and
L2 , such that
φ ∈ φi + Wi , for i = 1, 2. (7.5)
φ − φi + δi Wi ⊂ (1 − δi)Wi + δi Wi = Wi (7.7)
so that
φ + Wφ ⊂ V i , (7.9)
1 1
φ1 + V + φ2 + V = φ1 + φ2 + V . (7.12)
2 2
(c) To prove the continuity of scalar multiplication, first consider φ0 ∈ D(Ω ) and
α0 ∈ R. Then,
αφ − α0 φ0 = α (φ − φ0 ) + (α − α0 )φ0 . (7.13)
so that
1
(|α0 | + δ )(φ − φ0 ) ∈ V . (7.15)
2
This means
1 1
α (φ − φ0 ) + (α − α0 )φ0 ∈ V + V = V . (7.16)
2 2
Therefore αφ − α0 φ0 ∈ V whenever |α − α0 | < δ and φ − φ0 ∈ cV .
For the next result the proof may be found in Rudin [58].
Proposition 7.1.4. A convex balanced set V ⊂ D(Ω ) is open if and only if V ∈ σ .
Proposition 7.1.5. The topology σK of DK ⊂ D(Ω ) coincides with the topology that
DK inherits from D(Ω ).
V ∈ σ implies DK ∩ V ∈ σK . (7.17)
{ϕ ∈ DK | ϕ − φ N < δφ } ⊂ V . (7.18)
170 7 Distributions
Define
Then Uφ ∈ σ̂ and
DK ∩ (φ + Uφ ) = φ + (DK ∩ Uφ ) ⊂ V . (7.20)
lim φi − φ j N = 0, ∀N ∈ N. (7.21)
i, j→∞
|T (φ )| ≤ c φ N , ∀φ ∈ DK . (7.22)
Proof. The proof follows from the equivalence of 1 and 4 in Theorem 7.1.9.
Then T ∈ D (Ω ), and
Dα Tn → Dα T in D (Ω ). (7.26)
Proof. Let K be an arbitrary compact subset of Ω . Since (7.25) holds for every
φ ∈ DK , the principle of uniform boundedness implies that the restriction of T to
DK is continuous. It follows from Theorem 7.1.9 that T is continuous in D(Ω ), that
is, T ∈ D (Ω ). On the other hand
Thus,
|T (φ )| = |φ (1/2) + φ (1/3)| ≤ φ ∞ + φ ∞ ≤ 2 φ 1 ,
so that T is also a distribution (bounded and linear).
where f ∈ L1 (Ω ).
Observe that the derivative of T for the multi-index α = (α1 , . . . , αn ) is defined by
Dα T (φ ) = (−1)|α | T (Dα φ ) = (−1)|α | f Dα φ dx.
Ω
7.3 Examples of Distributions 173
and
T (φ ) = f φ dx,
Ω
where φ ∈ Cc∞ (Ω ) ,we have
dφ
Dx T (φ ) = − f dx
Ω dx
1
dφ
=− (1) dx
1/2 dx
= −φ (1) + φ (1/2) = φ (1/2), (7.29)
that is,
Dx T (φ ) = φ (1/2), ∀φ ∈ Cc∞ (Ω ).
Finally, defining f : Ω → R by
x, if x ∈ [0, 1/2],
f (x) =
−x + 1, if x ∈ (1/2, 1],
and
T (φ ) = f φ dx,
Ω
where φ ∈ Cc∞ (Ω ) we have
dφ
Dx T (φ ) = − f dx
Ω dx
1
dφ
=− f dx
0 dx
1
= gφ dx, (7.30)
0
where
1, if x ∈ [0, 1/2],
g(x) =
−1, if x ∈ (1/2, 1].
174 7 Distributions
Here, we emphasize that the two main references for this chapter are Adams [2]
and Evans [26]. We start with the definition of Lebesgue spaces, denoted by L p (Ω ),
where 1 ≤ p ≤ ∞ and Ω ⊂ Rn is an open set. In this chapter, integrals always refer
to the Lebesgue measure.
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 175
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 8,
© Springer International Publishing Switzerland 2014
176 8 The Lebesgue and Sobolev Spaces
Thus,
1 p 1
ab ≤ (a ) + (bq ), ∀a ≥ 0, b ≥ 0. (8.5)
p q
Therefore
1 1
|u(x)||v(x)| ≤ |u(x)| p + |v(x)|q , a.e. in Ω . (8.6)
p q
For λ = u −1
q/p
p v q we obtain the Hölder inequality.
The next step is to prove that · p is a norm.
Theorem 8.1.4. L p (Ω ) is a vector space and · p is norm ∀p such that 1 ≤ p ≤ ∞.
Proof. The only nontrivial property to be proved concerning the norm definition is
the triangle inequality. If p = 1 or p = ∞, the result is clear. Thus, suppose 1 < p < ∞.
For u, v ∈ L p (Ω ) we have
that is,
u + v p ≤ u p + v p, ∀u, v ∈ L p (Ω ). (8.12)
1
|um (x) − un (x)| < , ∀x ∈ Ω \ Ek , ∀m, n ≥ Nk . (8.14)
k
Observe that E = ∪∞ k=1 Ek is such that m(E) = 0. Thus {un (x)} is a real Cauchy
sequence at each x ∈ Ω \ E. Define u(x) = lim un (x) on Ω \ E. Letting m → ∞
n→∞
in (8.14) we obtain
1
|u(x) − un(x)| < , ∀x ∈ Ω \ E, ∀n ≥ Nk . (8.15)
k
Thus u ∈ L∞ (Ω ) and un − u ∞ → 0 as n → ∞.
Now suppose 1 ≤ p < ∞. Let {un } be a Cauchy sequence in L p (Ω ). We can
extract a subsequence {unk } such that
1
unk+1 − unk p ≤ , ∀k ∈ N. (8.16)
2k
To simplify the notation we write uk in place of unk , so that
1
uk+1 − uk p ≤ , ∀k ∈ N. (8.17)
2k
Defining
n
gn (x) = ∑ |uk+1 (x) − uk (x)|, (8.18)
k=1
we obtain
gn p ≤ 1, ∀n ∈ N. (8.19)
178 8 The Lebesgue and Sobolev Spaces
From the monotone convergence theorem and (8.19), gn (x) converges to a limit g(x)
with g ∈ L p (Ω ). On the other hand, for m ≥ n ≥ 2, we have
Hence {un (x)} is Cauchy a.e. in Ω and converges to a limit u(x) so that
which means u ∈ L p (Ω ). Finally from |un (x) − u(x)| → 0, a.e. in Ω , |un (x) −
u(x)| p ≤ |g(x)| p , and the Lebesgue dominated convergence theorem we get
un − u p → 0 as n → ∞. (8.22)
α p + β p ≤ (α 2 + β 2 ) p/2 , ∀α , β ≥ 0. (8.24)
Now taking α = a+b
2
and β = a−b in (8.24), we obtain (using the convexity
2
of t p/2 )
p/2 2 p/2
a + b p a − b p a + b 2 a − b 2 a b2
+ ≤ + = +
2 2 2 2 2 2
1 1
≤ |a| p + |b| p . (8.25)
2 2
The inequality (8.23) follows immediately.
8.1 Definition and Properties of L p Spaces 179
and therefore
u + v
2 < 1−δ, (8.28)
p
so that
Tu (Lq (Ω ))∗ ≤ u p. (8.31)
Pick u ∈ L p (Ω ) and define f0 (x) = |u(x)| p−2 u(x) ( f0 (x) = 0 if u(x) = 0). Thus,
we have that f0 ∈ Lq (Ω ), f0 q = u pp−1, and Tu, f0 Lq (Ω ) = u pp. Therefore,
Tu, f0 Lq (Ω )
Tu (Lq (Ω ))∗ ≥ = u p. (8.32)
f 0 q
Theorem 8.1.8 (Riesz Representation Theorem). Let 1 < p < ∞ and let f be a
continuous linear functional on L p (Ω ). Then there exists a unique u0 ∈ Lq such that
f (v) = vu0 dx, ∀v ∈ L p (Ω ). (8.34)
Ω
180 8 The Lebesgue and Sobolev Spaces
Furthermore
f (L p )∗ = u0 q . (8.35)
Cm,λ (Ω̄ ) is a Banach space with the norm denoted by · m,λ given by
α
|D u(x) − Dα u(y)|
u m,λ = u B,m + max sup , x = y .
0≤|α |≤m x,y∈Ω |x − y|λ
is compact and W ⊂ Ω .
Theorem 8.1.11. The space C0 (Ω ) is dense in L p (Ω ), for 1 ≤ p < ∞.
Then f = 0 a.e. in Ω .
First suppose f ∈ L1 (Ω ) and Ω bounded, so that m(Ω ) < ∞. Given ε > 0, since
C0 (Ω ) is dense in L1 (Ω ), there exists f1 ∈ C0 (Ω ) such that f − f1 1 < ε and thus,
from (8.39), we obtain
f1 u dx ≤ ε u ∞ , ∀u ∈ C0 (Ω ). (8.40)
Ω
Defining
K1 = {x ∈ Ω | f1 (x) ≥ ε } (8.41)
and
K2 = {x ∈ Ω | f1 (x) ≤ −ε }, (8.42)
182 8 The Lebesgue and Sobolev Spaces
as K1 and K2 are disjoint compact sets, by the Urysohn theorem, there exists u0 ∈
C0 (Ω ) such that
+1, if x ∈ K1 ,
u0 (x) = (8.43)
−1, if x ∈ K2
and
so that
| f1 | dx = | f1 | dx + | f1 | dx ≤ ε + ε m(Ω ). (8.47)
Ω K Ω −K
Hence
f 1 ≤ f − f1 1 + f1 1 ≤ 2ε + ε m(Ω ). (8.48)
Proof. The result follows from last theorem and from the fact that C0 (K) is sep-
arable for each K ⊂ Ω compact [from the Weierstrass theorem, polynomials with
rational coefficients are dense C0 (K)]. Observe that Ω = ∪∞
n=1 Ω n , Ω n defined as in
(8.49), where Ω̄n is compact, ∀n ∈ N.
8.2 The Sobolev Spaces 183
Definition 8.2.2. We define the norm · m,p for W m,p (Ω ), where m ∈ N and 1 ≤
p ≤ ∞, as
1/p
u m,p = ∑ Dα u pp , if 1 ≤ p < ∞, (8.51)
0≤|α |≤m
and
Remark 8.2.4. Observe that distributional and classical derivatives coincide when
the latter exist and are continuous. We define S ⊂ W m,p (Ω ) by
Moreover,
f U ∗ = v qN , (8.59)
where · qN = · Lq (Ω ,RN ) .
we have
f (uk ) = u, vk L2 (Ω ) = |vk |dx ≥ ( (vk ∞ − ε ) uk 1
A
= ( v ∞N − ε ) u 1N . (8.65)
Definition 8.2.9. Let Ω ⊂ Rn be a domain. For m a positive integer and 1 ≤ p < ∞we
define W0m,p (Ω ) as the closure in · m,p of Cc∞ (Ω ), where we recall that Cc∞ (Ω )
denotes the set of C∞ (Ω ) functions with compact support contained in Ω . Finally,
we also recall that the support of φ : Ω → R, denoted by spt(φ ), is given by
Now we present the Sobolev imbedding theorem. We recall that for normed
spaces X,Y the notation
X → Y
means that X ⊂ Y and there exists a constant K > 0 such that
u Y ≤ K u X , ∀u ∈ X.
If in addition the imbedding is compact, then for any bounded sequence {un } ⊂ X
there exists a convergent subsequence {unk }, which converges to some u in the norm
· Y . At this point, we first introduce the following definition.
Definition 8.3.1. Let Ω ⊂ Rn be an open bounded set. We say that ∂ Ω is Ĉ1 if for
each x0 ∈ ∂ Ω , denoting x̂ = (x1 , . . . , xn−1 ) for a local coordinate system, there exist
r > 0 and a function f (x1 , . . . , xn−1 ) = f (x̂) such that
Now we present a collection of results which imply the proof of the Sobolev
imbedding theorem. We start with the approximation by smooth functions.
Definition 8.4.1. Let Ω ⊂ Rn be an open bounded set. For each ε > 0 define
Ωε = {x ∈ Ω | dist(x, ∂ Ω ) > ε }.
Theorem 8.4.4 (Properties of Mollifiers). The mollifiers have the following prop-
erties:
1. fε ∈ C∞ (Ωε ),
2. fε → f a.e. as ε → 0,
3. If f ∈ C(Ω ), then fε → f uniformly on compact subsets of Ω .
Proof.
1. Fix x ∈ Ωε , i ∈ {1, . . . , n} and a h small enough such that
x + hei ∈ Ωε .
Thus
fε (x + hei) − fε (x) 1 1 x + hei − y x−y
= n η −η
h ε Ω h ε ε
× f (y) dy
1 1 x + hei − y x−y
= n η −η
ε Vh ε ε
× f (y) dy, (8.79)
8.4 The Proof of the Sobolev Imbedding Theorem 189
as h → 0, uniformly on V , we obtain
∂ fε (x) ∂ ηε (x − y)
= f (y) dy.
∂ xi Ω ∂ xi
By analogy, we may show that
Dα fε (x) = Dα ηε (x − y) f (y) dy, ∀x ∈ Ωε .
Ω
V ⊂⊂ W ⊂⊂ Ω ,
and note that f is uniformly continuous on W . Thus the limit indicated in (8.80)
holds uniformly on V , and therefore fε → f uniformly on V .
ηε ∗ u ∈ L p (Ω ),
ηε ∗ u p ≤ u p, ∀ε > 0
and
lim ηε ∗ u − u p = 0.
ε →0+
190 8 The Lebesgue and Sobolev Spaces
Proof. Suppose u ∈ L p (Ω ) and 1 < p < ∞. Defining q = p/(p − 1), from the Hölder
inequality, we have
|ηε ∗ u(x)| = ηε (x − y)u(y) dy
Rn
= [ηε (x − y)](1−1/p)[ηε (x − y)]1/pu(y) dy
Rn
1/q 1/p
≤ ηε (x − y) dy ηε (x − y)|u(y)| dy
p
Rn Rn
1/p
= ηε (x − y)|u(y)| p dy . (8.82)
Rn
u − φ p < ρ /3.
ηε ∗ φ − φ p < ρ /3
ηε ∗ u − u p = ηε ∗ u − ηε ∗ φ + ηε ∗ φ − φ + φ − u p
≤ ηε ∗ u − ηε ∗ φ p + ηε ∗ φ − φ p + φ − u p
≤ ρ /3 + ρ /3 + ρ /3 = ρ . (8.84)
2.
|u(x)| p dx < ε p . (8.86)
Ω −G
N = {v1 , . . . , vm } ⊂ L p (Ω ).
φk − u p ≤ φk − vk p + vk − u p
ε ε ε
≤ + = . (8.87)
6 6 3
Define
k=1 spt(φk ),
G = ∪m
where
spt(φk ) = {x ∈ Rn | φk (x) = 0}.
We have that
G ⊂⊂ Ω ,
192 8 The Lebesgue and Sobolev Spaces
Thus,
ε p
|φk (x + h) − φk (x)| p dx < .
Ω 3
Also observe that since
ε
u − φk p < ,
3
we have that
ε
Th u − Thφk p < ,
3
where Th u = u(x + h). Thus, if |h| < δ = min{δ̃ , 1}, we obtain
Th ũ − ũ p ≤ Th ũ − Th φk p + Thφk − φk p
+ φk − u p
ε ε ε
< + + = ε. (8.88)
3 3 3
For the converse, it suffices to consider the special case Ω = Rn , because for the
general Ω we can define K̃ = {ũ | u ∈ K}. Suppose given ε > 0 and choose G ⊂⊂ Rn
such that for all u ∈ K we have
ε
|u(x)| p dx < .
Rn −G 3
For each ρ > 0 the function ηρ ∗ u ∈ C∞ (Rn ), and in particular ηρ ∗ u ∈ C(G).
Suppose φ ∈ C0 (Rn ). Fix ρ > 0. By the Hölder inequality we have
p
|ηρ ∗ φ (x) − φ (x)| =
p ηρ (y)(φ (x − y) − φ (x)) dy
Rn
p
= (ηρ (y)) 1−1/p
(ηρ (y)) (T−y φ (x) − φ (x)) dy
1/p
Rn
≤ (ηρ (y))|T−y φ (x) − φ (x)| p dy. (8.89)
B ρ (θ )
8.4 The Proof of the Sobolev Imbedding Theorem 193
φk → u, in L p (Rn ).
Observe that
ηρ ∗ φk → ηρ ∗ u, in L p (Rn ),
as k → ∞. Also
Th φk → Th u, in L p (Rn ),
as k → ∞. Thus
Th φk − φk p → Th u − u p,
in particular
" #
lim sup sup { Th φk − φk ≤ sup Th u − u p .
k→∞ h∈Bρ (θ ) h∈Bρ (θ )
Therefore as
ηρ ∗ φk − φk p → ηρ ∗ u − u p,
as k → ∞, from (8.91) we get
ηρ ∗ u − u p ≤ sup { Th u − u p}.
h∈Bρ (θ )
ηρ ∗ u − u p → 0, uniformly in K as ρ → 0.
Observe that
|ηρ0 ∗ u(x)| = ηρ0 (x − y)u(y) dy
Rn
= [ηρ0 (x − y)](1−1/p)[ηρ0 (x − y)]1/pu(y) dy
Rn
1/q 1/p
≤ ηρ0 (x − y) dy ηρ0 (x − y)|u(y)| dy
p
Rn Rn
1/p
= ηρ0 (x − y)|u(y)| p dy . (8.92)
Rn
where K1 = K2 K3 , 1/p
K2 = sup ηρ0 (y) ,
y∈Rn
{ηρ0 ∗ u | u ∈ K}
is relatively compact in C(G), and it is totally bounded so that there exists a ε0 -net
N = {v1 , . . . , vm } where
1/p
ε
ε0 = .
3 · 2 p−1|G|
Thus for some k ∈ {1, . . . , m} we have
vk − ηρ0 ∗ u ∞ < ε0 .
8.4 The Proof of the Sobolev Imbedding Theorem 195
Hence,
|u(x) − ṽk (x)| p dx = |u(x)| p dx + |u(x) − vk (x)| p dx
Rn Rn −G G
ε
≤ + 2 p−1 (|u(x) − (ηρ0 ∗ u)(x)| p
3 G
+|ηρ0 ∗ u(x) − vk (x)| p ) dx
ε ε ε |G|
≤ + 2 p−1 +
3 3 · 2 p−1 3 · 2 p−1|G|
= ε. (8.93)
Proof. Assertion 1 has been already proved. Let us prove 2. We will show that if
|α | ≤ m, then
Dα uε = ηε ∗ Dα u, in Ωε .
For that, let x ∈ Ωε . Thus,
Dα uε (x) = Dα ηε (x − y)u(y) dy
Ω
= Dαx ηε (x − y)u(y) dy
Ω
= (−1)|α | Dαy (ηε (x − y))u(y) dy. (8.94)
Ω
φ (y) = ηε (x − y) ∈ Cc∞ (Ω ).
Therefore,
Dαy (ηε (x − y)) u(y) dy = (−1)|α | ηε (x − y)Dαy u(y) dy,
Ω Ω
196 8 The Lebesgue and Sobolev Spaces
and hence,
Dα uε (x) = (−1)|α |+|α | ηε (x − y)Dα u(y) dy
Ω
= (ηε ∗ Dα u)(x). (8.95)
Now choose any open bounded set such that V ⊂⊂ Ω . We have that
Dα uε → Dα u, in L p (V ) as ε → 0,
for each |α | ≤ m.
Thus,
p
uε − u m,p,V = ∑ Dα uε − Dα u p,V → 0,
|α |≤m
as ε → 0.
Theorem 8.4.8. Let Ω ⊂ Rn be a bounded open set and suppose u ∈ W m,p (Ω ) for
some 1 ≤ p < ∞. Then there exists a sequence {uk } ⊂ C∞ (Ω ) such that
uk → u in W m,p (Ω ).
where
Ωi = {x ∈ Ω | dist(x, ∂ Ω ) > 1/i}.
Define
Vi = Ωi+3 − Ω̄i+1,
and choose any open set V0 such that V0 ⊂⊂ Ω , so that
Ω = ∪∞
i=0Vi .
Let {ζi }∞ ∞
i=0 be a smooth partition of unit subordinate to the open sets {Vi }i=0 .
That is,
0 ≤ ζi ≤ 1, ζi ∈ Cc∞ (Vi )
∑∞ i=0 ζi = 1, on Ω .
ui = ηεi ∗ (ζi u)
satisfies
δ
ui − ζi u m,p,Ω ≤ ,
2i+1
8.4 The Proof of the Sobolev Imbedding Theorem 197
Thus such a function belongs to C∞ (Ω ), since for each open V ⊂⊂ Ω there are at
most finitely many nonzero terms in the sum. Since
∞
u = ∑ ζi u,
i=0
v − u m,p,Ω < δ .
The next result is also relevant. For a proof see Evans [26], p. 232.
Theorem 8.4.9. Let Ω ⊂ Rn be a bounded set such that ∂ Ω is C1 . Suppose u ∈
W m,p (Ω ) where 1 ≤ p < ∞. Thus there exists a sequence {un } ⊂ C∞ (Ω ) such that
un → u in W m,p (Ω ), as n → ∞.
uk − u m,p,Ω → 0, as k → ∞.
Observe that
For the fixed δ > 0, there exists Nε ∈ N such that if n > Nε we have
and
vn ∈ C∞ (W ).
Hence
∂ Ω ⊂ ∪M
i=1 Bri (xi ).
Ω ⊂ ∪M
i=0 Bi .
Let {ζi }M
i=0 be a concerned partition of unity relating {Bi }i=0 .
M
∞
Thus ζi ∈ Cc (Bi ) and 0 ≤ ζi ≤ 1, ∀i ∈ {0, . . . , M} and also
M
∑ ζi = 1 on Ω .
i=0
8.4 The Proof of the Sobolev Imbedding Theorem 199
From above, we may find vi ∈ C∞ (W i ) such that vi − u m,p,Wi < ε , ∀i ∈ {1, . . ., M}.
Define u0 = v0 = u on B0 ≡ W0 ,
ui = ζi u, ∀i ∈ {0, . . . , M}
and
M
v = ∑ ζi vi .
i=0
We emphasize
v ∈ C∞ (Ω ).
Therefore
M
v − u m,p,Ω = ∑ (ζi u − ζi vi )
i=0
m,p,Ω
M
≤ C2 ∑ u − vi m,p,(Ω ∩Bi )
i=0
M
= C2 ∑ u − vi m,p,Wi
i=0
< C2 M ε . (8.97)
8.4.3 Extensions
Theorem 8.4.12. Assume Ω ⊂ Rn is an open bounded set and that ∂ Ω is Ĉ1 . Let
V be a bounded open set such that Ω ⊂⊂ V . Then there exists a bounded linear
operator
E : W 1,p (Ω ) → W 1,p (Rn ),
200 8 The Lebesgue and Sobolev Spaces
B+ = B+
r1 (y0 ) = {y ∈ Br1 (y0 ) | yn ≥ 0},
and
B− = B−
r1 (y0 ) = {y ∈ Br1 (y0 ) | yn < 0}.
Define φ̂N : B → R by
+
φ̃N (y) if y ∈ B
φ̂N (y) =
−3φ̃N (y1 , . . . , yn−1 , −yn ) + 4φ̃N (y1 , . . . , yn−1 , −yn /2) if y ∈ B− .
It may be easily verified that φ̂N ∈ C1 (B). Also, there exists C2 > 0 such that
u 1,p,W ≤ C5 u 1,p,W + .
Now denoting W = Wx0 we have that ∂ Ω ⊂ ∪x0 ∈∂ Ω Wx0 and since ∂ Ω is compact,
there exist x1 , . . . , xM ∈ ∂ Ω , such that
∂ Ω ⊂ ∪M
i=1Wxi .
Ω ⊂ ∪M
i=0Wi .
M
∑ ζi = 1, in Ω ,
i=0
M
u 1,p,Rn ≤ ∑ ζi ui 1,p,Rn
i=0
M
≤ C5 ∑ ui 1,p,Wi
i=0
M
≤ C5 u 1,p,W0 + C5 ∑ ui 1,p,W +
i
i=1
≤ (M + 1)C5 u 1,p,Ω
= C u 1,p,Ω , (8.101)
where C = (M + 1)C5 .
We recall that the partition of unity may be chosen so that its support is on V .
Finally, we denote Eu = u.
The proof is complete.
8.4 The Proof of the Sobolev Imbedding Theorem 203
Proof. Suppose p = 1. Let u ∈ Cc1 (Rn ). From the fundamental theorem of calculus
we have
xi
∂ u(x1 , . . . , xi−1 , yi , xi+1 , . . . , xn )
u(x) = dyi ,
−∞ ∂ xi
so that
∞
|u(x)| ≤ |Du(x1 , . . . , xi−1 , yi , xi+1 , . . . , xn )| dyi .
−∞
Therefore,
n ∞
1/(n−1)
|u(x)| n/(n−1)
≤∏ |Du(x1 , . . . , xi−1 , yi , xi+1 , . . . , xn )| dyi .
i=1 −∞
Integrating in x2 we obtain
∞ ∞
|u(x)|n/(n−1) dx1 dx2
−∞ −∞
∞
∞ 1/(n−1)
≤ |Du| dy1
−∞ −∞
n ∞ ∞ 1/(n−1)
×∏ |Du| dx1 dyi dx2 ,
i=2 −∞ −∞
so that
∞ ∞
|u(x)|n/(n−1) dx1 dx2
−∞ −∞
∞ ∞ 1/(n−1)
≤ |Du| dy2 dx1
−∞ −∞
∞
1/(n−1)
∞
× |Du| dy1
−∞ −∞
n ∞ ∞ 1/(n−1)
×∏ |Du| dx1 dyi dx2 . (8.104)
i=3 −∞ −∞
v = |u|γ ,
to obtain
(n−1)/n
|u(x)|γ n/(n−1) dx ≤ |D|u|γ |; dx
Rn Rn
≤γ |u|γ −1 |Du|; dx
Rn
(p−1)/p
≤γ |u|(γ −1)p/(p−1) dx
Rn
1/p
× |Du| dx
p
. (8.107)
Rn
γn (γ − 1)p np
= = ,
n−1 p−1 n− p
we get
((n−1)/n−(p−1)/p) 1/p
|u| dx
r
≤C |Du| dx
p
.
Rn Rn
ū 1,p,Rn ≤ C u 1,p,Ω ,
where C does not depend on u. As ū has compact support, from Theorem 8.4.10,
there exists a sequence {uk } ∈ Cc∞ (Rn ) such that
uk → ū in W 1,p (Rn ),
Hence,
uk → ū in Lr (Rn ),
also from the last theorem
so that
ū r,Rn ≤ K Dū p,Rn .
Therefore, we may get
u r,Ω ≤ ū r,Rn
≤ K Dū p,Rn
≤ K1 ū 1,p,Rn
≤ K2 u 1,p,Ω . (8.108)
Theorem 8.4.16. Let Ω ⊂ Rn be a bounded open set such that ∂ Ω ∈ Ĉ1 . If mp < n,
then W m,p (Ω ) → Lq (Ω ) for p ≤ q ≤ (np)/(n − mp).
W m,p → Lq0 (Ω ).
where
r1 = np/(n − (m − 1)p) = np/(n − np + p),
8.4 The Proof of the Sobolev Imbedding Theorem 207
whenever n > (m − 1)p. If u ∈ W m,p (Ω ) where n > mp, then u and D j u are in
W m−1,p (Ω ), so that from (8.109) we have u ∈ W 1,r1 (Ω ) and
Since n > mp we have that r1 = np/((n − mp) + p) < n, from q0 = nr1 /(n − r1) =
np/(n − mp) by the last theorem, we have
u q0 ,Ω ≤ K2 u 1,r1,Ω ,
where the constant K2 does not depend on u, and therefore from this and (8.110) we
obtain
since
p/t + q0/t = q.
This completes the proof.
W m,p (Ω ) → W m,s (Ω )
with the imbedding constant depending only on |Ω |. Since ms < n, by the last the-
orem, we obtain
W m,p (Ω ) → W m,s (Ω ) → Lq (Ω ).
208 8 The Lebesgue and Sobolev Spaces
Now if p ≤ q ≤ p , from above we have W m,p (Ω ) → L p (Ω ) and the obvious imbed-
ding W m,p (Ω ) → L p (Ω ). Define s = (p − q)p/(p − p), and the result follows from
a reasoning analogous to the final chain of inequalities of last theorem, indicated
in (8.112).
About the next theorem, note that its hypotheses are satisfied if ∂ Ω is Ĉ1 (here we
do not give the details).
Theorem 8.4.18. Let Ω ⊂ Rn be an open bounded set, such that for each x ∈ Ω
there exists a convex set Cx ⊂ Ω whose shape depends on x but such that |Cx | > α ,
for some α > 0 that does not depend on x. Thus, if mp > n, then
W m,p (Ω ) → CB0 (Ω ).
so that, from the Hölder inequality and the Fubini theorem, we get
Therefore 1
|φ (x)|α ≤ φ p,Ω |Ω |1/p + |∇φ (z)|δ t −n dz dt,
0 V
where |V | = t n |Cx | and δ denote the diameter of Ω . From the Hölder inequality
again, we obtain
1 1/p
|φ (x)|α ≤ φ p,Ω |Ω |1/p + δ |∇φ (z)| p dy t −n (t n |Cx |)1/p dt,
0 V
and thus
1
1/p 1/p
|φ (x)|α ≤ φ p,Ω |Ω | + δ |Cx | ∇φ p,Ω t −n(1−1/p ) dt.
0
8.4 The Proof of the Sobolev Imbedding Theorem 209
From this, the last inequality and from the fact that |Cx | ≤ |Ω |, we have that there
exists K > 0 such that
φk → u, in W 1,p (Ω ).
φk → u, a.e. in Ω . (8.114)
Fix x ∈ Ω such that the limit indicated in (8.114) holds. Suppose given ε > 0. There-
fore, there exists k0 ∈ N such that
and
φk0 − u 1,p,Ω < ε /(2K).
Thus,
Since ε > 0 is arbitrary, the proof for m = 1 is complete, because for {φk } ∈ C∞ (Ω )
such that φk → u in W 1,p (Ω ), from (8.113), we have that {φk } is a uniformly Cauchy
sequence, so that it converges to a continuous u∗ , where u∗ = u, a.e. in Ω .
For m > 1 but p > n we still have
Now define
r̂1 = p1 = np/(n − ( j − 1)p)
and 1 ≤ p2 ≤ n such that
r̂1 = np2 /(n − p2),
so that
np/(n − ( j − 1)p) = np2 /(n − p2).
Hence p2 = np/(n − ( j − 2)p) so that by the last theorem
If p ≤ q < ∞, we have
u qq,Ω = |u(x)| p |u(x)|q−p dx
Ω
q−p
≤ |u(x)| p K u m,p,Ω dx
Ω
≤K q−p
u pp,Ω u m,p,
q−p
Ω
p q−p
≤ K q−p u m,p,Ω u m,p,Ω
q
= K q−p u m,p,Ω . (8.116)
Proof. First consider λ = 1 − n/p and u ∈ C1 (S). Let x, y ∈ S such that |x − y| < 1
and define σ = |x − y|. Consider a fixed cube denoted by Rσ ⊂ S such that |Rσ | = σ n
and x, y ∈ R̄σ . For z ∈ Rσ , we may write
212 8 The Lebesgue and Sobolev Spaces
1
du(x + t(z − x))
u(x) − u(z) = − dt,
0 dt
that is,
1
u(x)σ n = u(z) dz − ∇u(x + t(z − x)) · (z − x) dt dz.
Rσ Rσ 0
Thus, denoting in the next lines V by an appropriate set such that |V | = t n |Rσ |, we
obtain
1
√
|u(x) − u(z) dz/σ n | ≤ nσ 1−n |∇u(x + t(z − x))| dt dz
Rσ Rσ 0
1
√
≤ nσ 1−n t −n |∇u(z)| dz dt
0 V
1
√
≤ nσ 1−n t −n ∇u p,S|V |1/p dt
0
1
√
≤ nσ 1−n σ n/p ∇u p,S t −nt n/p dt
0
1
√
≤ nσ 1−n/p ∇u p,S t −n/p dt
0
≤σ 1−n/p u 1,p,SK, (8.117)
where
√ 1 −n/p √
K= n t dt = n/(1 − n/p).
0
A similar inequality holds with y in place of x, so that
Now consider 0 < λ < 1 − n/p. Observe that, as |x− y|λ ≥ |x− y|1−n/p, if |x− y| < 1,
we have
|u(x) − u(y)|
sup | x =
y, |x − y| < 1
x,y∈S |x − y|λ
|u(x) − u(y)|
≤ sup | x = y, |x − y| < 1 ≤ K u 1,p,S. (8.118)
x,y∈S |x − y|1−n/p
Also,
|u(x) − u(y)|
sup | |x − y| ≥ 1 ≤ 2 u ∞,S ≤ 2K1 u 1,p,S
x,y∈S |x − y|λ
8.4 The Proof of the Sobolev Imbedding Theorem 213
so that
|u(x) − u(y)|
sup | x = y ≤ (K + 2K1) u 1,p,S , ∀u ∈ C1 (S).
x,y∈S |x − y|λ
Theorem 8.4.21. Let Ω ⊂ Rn be an open bounded set such that ∂ Ω is Ĉ1 . Assume
n < p ≤ ∞.
Then
W 1,p (Ω ) → C0,λ (Ω ),
for all 0 < λ ≤ 1 − n/p.
Proof. Fix 0 < λ ≤ 1 − n/p and let u ∈ W 1,p (Ω ). Since ∂ Ω is Ĉ1 , from Theo-
rem 8.4.12, there exists an extension Eu = ū such that ū = u, a.e. in Ω , and
ū 1,p,Rn ≤ K u 1,p,Ω ,
where the constant K does not depend on u. From the proof of this same theorem,
we may assume that spt(ū) is on an n-dimensional sphere S ⊃ Ω with sufficiently
big radius and such sphere does not depend on u. Thus, in fact, we have
ū 1,p,S ≤ K u 1,p,Ω .
Since C∞ (S) is dense in W 1,p (S), there exists a sequence {φk } ⊂ C∞ (S) such that
uk → ū, in W 1,p (S). (8.119)
Up to a not relabeled subsequence, we have
uk → ū, a.e. in Ω .
From last theorem we have
uk − ul C0,λ (S) ≤ C uk − ul 1,p,S ,
so that {uk } is a Cauchy sequence in C0,λ (S), and thus uk → u∗ for some u∗ ∈
C0,λ (S). Hence, from this and (8.119), we have
u∗ = ū, a.e. in S.
Finally, from above and last theorem, we may write
u∗ C0,λ (Ω ) ≤ u∗ C0,λ (S) ≤ K1 ū 1,p,S ≤ K2 u 1,p,Ω .
The proof is complete.
214 8 The Lebesgue and Sobolev Spaces
T : W 1,p (Ω ) → L p (∂ Ω ),
such that
• Tu = u|∂ Ω if u ∈ W 1,p (Ω ) ∩C(Ω ),
•
Tu p,∂ Ω ≤ C u 1,p,Ω , ∀u ∈ W 1,p (Ω ),
where the constant C depends only on p and Ω .
Proof. Let u ∈ W 1,p (Ω ) ∩C(Ω ). Choose x0 ∈ ∂ Ω .
Since ∂ Ω is Ĉ1 , there exists r > 0 such that for a local coordinate system we may
write
Ω ∩ Br (x0 ) = {x ∈ Br (x0 ) | xn ≥ f (x1 , . . . , xn−1 )},
where denoting x̂ = (x1 , . . . , xn−1 ), f (x̂) is continuous and such that its partial deriva-
tives are classically defined a.e. and bounded on its domain. Furthermore
yi = xi , ∀i ∈ {1, . . . , n − 1},
and
yn = xn − f (x1 , . . . , xn−1 ).
Define û(y) by
Also define y0 = (x01 , . . . , x0n−1 , x0n − f (x01 , . . . , x0n−1 ) = (y01 , . . . , y0n−1 , 0) and
choose r1 > 0 such that
Ψ −1 (B+
r1 (y0 )) ⊂ Ω ∩ Br (x0 ).
Observe that this is possible since Ψ and Ψ −1 are continuous, where y = Ψ (x). Here
B+
r1 (y0 ) = {y ∈ Br1 (y0 ) | yn > 0}.
+
For each N ∈ N, choose, by mollification, for example, φN ∈ C∞ (Br1 (y0 )) such that
8.5 The Trace Theorem 215
1
φN − û ∞,B+ (y ) < .
r1 0 N
Denote B = Br1 /2 (y0 ), and B+ = B+ (y ). Now choose η ∈ Cc∞ (Br1 (y0 )), such that
r1 /2 0
η > 0 and η ≡ 1 on B. Also denote
Γ̃ = {y ∈ B | yn = 0},
and
Γ̃1 = {y ∈ Br1 (y0 ) | yn = 0}.
Observe that
|φN | p dΓ ≤ η |φN | p dΓ
Γ̃ Γ̃1
=− (η |φN | p )yn dy
B+
r1
≤− (ηyn |φN | p ) dy
B+
r1
+ (p|φN | p−1 |(φN )yn |η ) dy. (8.120)
B+
r1
Letting N → +∞ we obtain
|u(x)| dΓ ≤ C2
p
|û(y)| p dΓ
Γ Γ̃
≤ C3 |û| dy +
p
|Dû| dy
p
B+
r1 B+
r1
≤ C4 |u| dx +
p
|Du| dx ,
p
(8.122)
W+ W+
Observe that denoting W = Wx0 we have that ∂ Ω ⊂ ∪x∈∂ Ω Wx , and thus, since
∂ Ω is compact, we may select x1 , . . . , xM such that ∂ Ω ⊂ ∪M
i=1Wi . We emphasize to
have denoted Wxi = Wi , ∀i ∈ {1, . . . , M}. Denoting Wi+ = Wi ∩ Ω we may obtain
M
∂Ω
|u(x)| p dΓ ≤ ∑ |u(x)| p dΓ
i=1 Γi
M
≤ ∑ C4i Wi+
|u| dx +
p
Wi+
|Du| dx
p
i=1
≤ C5 M |u| p dx + |Du| p dx
Ω Ω
=C |u| p dx + |Du| p dx . (8.123)
Ω Ω
uk − u 1,p,Ω → 0, as k → ∞.
From above
Tuk − Tul p,∂ Ω ≤ C uk − ul 1,p,Ω ,
so that
{Tuk }
is a Cauchy sequence. Hence we may define
Tu = lim Tuk , in L p (∂ Ω ).
k→∞
Remark 8.5.2. Similar results are valid for W0m,p ; however, in this case the traces
relative to derivatives of order up to m − 1 are involved.
Theorem 8.6.1. Let m be a nonnegative integer and let 0 < ν < λ ≤ 1. Then the
following imbeddings exist:
Cm+1 (Ω ) → Cm (Ω ), (8.124)
Cm,λ (Ω ) → Cm (Ω ), (8.125)
8.6 Compact Imbeddings 217
φ Cm (Ω ) ≤ φ Cm+1 (Ω ) ,
φ Cm (Ω ) ≤ φ Cm,λ (Ω ) .
To establish (8.126) note that for |α | ≤ m
α
|D φ (x) − Dα φ (y)|
sup | x =
y, |x − y| < 1
x,y∈Ω |x − y|ν
α
|D φ (x) − Dα φ (y)|
≤ sup | x = y, |x − y| < 1 , (8.127)
x,y∈Ω |x − y|λ
and also,
|Dα φ (x) − Dα φ (y)|
sup | |x − y| ≥ 1 ≤ 2 sup {|Dα φ |}. (8.128)
x,y∈Ω |x − y|ν x∈Ω
φk → φ in C(Ω ).
However, {Di φk } is also bounded in C0,λ (Ω ), so that there exists a not relabeled
subsequence, also denoted by {φk } and ψi such that
Di φk → ψi , in C(Ω ).
The convergence in C(Ω̄ ) being the uniform one, we have ψi = Di φ . We can proceed
extracting (not relabeled) subsequences until obtaining
218 8 The Lebesgue and Sobolev Spaces
Dα φk → Dα φ , in C(Ω ), ∀ 0 ≤ |α | ≤ m.
This completes the proof of compactness of (8.125). For (8.126), let S be a bounded
set in Cm,λ (Ω ). Observe that
ν /λ
|Dα φ (x) − Dα φ (y)| |Dα φ (x) − Dα φ (y)|
=
|x − y|ν |x − y|λ
·|Dα φ (x) − Dα φ (y)|1−ν /λ
≤ K|Dα φ (x) − Dα φ (y)|1−ν /λ , (8.129)
W j+m,p (Ω ) → W j,q (Ω ),
if 0 < n − mp < n and 1 ≤ np/(n − mp), (8.130)
W j+m,p → CB (Ω ),
j
(8.132)
W j+m,p (Ω ) → C j,λ (Ω ),
if mp > n ≥ (m − 1)p and 0 < λ < m − n/p. (8.135)
(despite the fact we are assuming Ω bounded, the general results may be found in
Adams [1]).
Proof of Parts II and III. If mp > n > (m − 1)p and 0 < λ < (m − n)/p, then
there exists μ such that λ < μ < m − (n/p). Since Ω is bounded, the imbedding
C0,μ (Ω ) → C0,λ (Ω ) is compact by Theorem 8.6.1. Since by the Sobolev imbed-
ding theorem we have W m,p (Ω ) → C0,μ (Ω ), we have that Imbedding (8.135) is
compact.
If mp > n, let j∗ be the nonnegative integer satisfying (m − j∗ )p > n ≥
(m − j∗ − 1)p. Thus we have the chain of imbeddings
∗ ,p
W m,p (Ω ) → W m− j (Ω ) → C0,μ (Ω ) → C(Ω ), (8.137)
where 0 < μ < m − j∗ − (n/p). The last imbedding in (8.137) is compact by Theo-
rem 8.6.1, so that (8.134) is compact for j = 0. By analogy (8.132) is compact for
j = 0. Therefore from the above remarks (8.133) is also compact. For the proof of
Part I, we need the following lemma:
Lemma 8.6.6. Let Ω be a bounded domain in Rn . Let 1 ≤ q1 ≤ q0 and suppose
W m,p → Lq (Ω ) (8.140)
is compact.
Suppose A is a bounded set of functions in W m,p (Ω ), that is, suppose there exists
K1 > 0 such that
u W m,p (Ω ) < K1 , ∀u ∈ A.
Also, suppose given ε > 0, and define, for u ∈ W m,p (Ω ), ũ(x) = u(x) if x ∈ Ω ,
ũ(x) = 0, if x ∈ Rn \ Ω . Fix u ∈ A. From the Hölder inequality and considering that
W m,p (Ω ) → Lq0 (Ω ), we have
1/q0 1−1/q0
|u(x)|dx ≤ |u(x)|q0 dx 1dx
Ω −Ω k Ω −Ω k Ω −Ω k
We emphasize the main references for this chapter are [37, 38, 68].
Here we recall that a functional is a function whose co-domain is the real set.
We denote such functionals by F : U → R, where U is a Banach space. In our work
format, we consider the special cases:
1. F(u) = Ω f (x, u, ∇u) dx, where Ω ⊂ Rn is an open, bounded, and connected set.
2. F(u) = Ω f (x, u, ∇u, D2 u) dx, here
∂ ui
Du = ∇u =
∂xj
and
∂ 2 ui
D u = {D ui } =
2 2
,
∂ xk ∂ xl
for i ∈ {1, . . . , N} and j, k, l ∈ {1, . . ., n}.
Also, f : Ω × RN × RN×n → R is denoted by f (x, s, ξ ) and we assume
1.
∂ f (x, s, ξ )
∂s
and
2.
∂ f (x, s, ξ )
∂ξ
are continuous ∀(x, s, ξ ) ∈ Ω × RN × RN×n .
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 225
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 9,
© Springer International Publishing Switzerland 2014
226 9 Basic Concepts on the Calculus of Variations
F(u0 ) = min{F(u)}.
u∈U
V = {ϕ | u + ϕ ∈ U, ∀u ∈ U}.
where Ω ⊂ R3 and
U = {u ∈ W 1,2 (Ω ) | u = û on ∂ Ω }
we have
V = W01,2 (Ω ).
Observe that in this example U is a subset of a Banach space.
Definition 9.1.3 (Local Minimum). Given F : U → R, we say that u0 ∈ U is a local
minimum for F if there exists δ > 0 such that
or equivalently
F(u + εϕ ) − F(u)
δ F(u, ϕ ) = lim ,
ε →0 ε
if such a limit is well defined. Furthermore, if there exists u∗ ∈ U ∗ such that
δ F(u, ϕ ) = ϕ , u∗ U , ∀ϕ ∈ U,
∂ F(u)
u∗ = δ F(u) or u∗ = .
∂u
Hence, there exists ε0 > 0 such that for each n ∈ N there exists 0 < εn < 1/n such
that
G(x, u, ϕ , εn ) dx − G̃(x, u, ϕ ) dx ≥ ε0 . (9.1)
Ω Ω
228 9 Basic Concepts on the Calculus of Variations
Define
cn = max{|G(x, u(x), ϕ (x), εn ) − G̃(x, u(x), ϕ (x))|}.
x∈Ω
Since the function in question is continuous on the compact set Ω , {xn } is well
defined. Also from the fact that Ω is compact, there exists a subsequence {xn j } and
x0 ∈ Ω̄ such that
lim xn j = x0 .
j→+∞
Thus
lim cn j = c0
j→+∞
cn j < ε0 /|Ω |.
that is,
∂ f (x, u, ∇u) ∂ f (x, u, ∇u)
δ F(u, ϕ ) = ·ϕ + · ∇ϕ dx.
Ω ∂s ∂ξ
Then u = 0, a.e. in Ω .
Remark 9.2.2. Of course a similar result is valid for the vectorial case. A proof of
such a result was given in Chap. 8.
δ F(u, ϕ ) = 0, ∀ϕ ∈ V .
Proof. Fix ϕ ∈ V . Define φ (ε ) = F(u + εϕ ). Since by hypothesis φ is differentiable
and attains a minimum at ε = 0, from the standard necessary condition φ (0) = 0,
we obtain φ (0) = δ F(u, ϕ ) = 0.
Theorem 9.2.4. Consider the hypotheses stated in Section 9.1 on F : U → R. Sup-
pose F attains a local minimum at u ∈ C2 (Ω̄ ; RN ) and additionally assume that
f ∈ C2 (Ω , RN , RN×n ). Then the necessary conditions for a local minimum for F are
given by the Euler–Lagrange equations:
∂ f (x, u, ∇u) ∂ f (x, u, ∇u)
− div = θ , in Ω .
∂s ∂ξ
Proof. From Theorem 9.2.3, the necessary condition stands for δ F(u, ϕ ) = 0, ∀ϕ
∈ V . From the above this implies, after integration by parts
∂ f (x, u, ∇u) ∂ f (x, u, ∇u)
− div · ϕ dx = 0,
Ω ∂s ∂ξ
∀ϕ ∈ Cc∞ (Ω , RN ).
The result then follows from the fundamental lemma of calculus of variations.
U = {u ∈ W 1,2 (Ω , RN ) | u = u0 in ∂ Ω }.
Suppose
F(u) = f (x, u, ∇u) dx,
Ω
where f : Ω × RN × RN×n is such that for each K > 0 there exists K1 > 0 which does
not depend on x such that
Also assume the hypotheses of Section 9.1 except for the continuity of derivatives
of f . Under such assumptions, for each u ∈ C1 (Ω ; RN ) and ϕ ∈ Cc∞ (Ω ; RN ), we
have
∂ f (x, u, ∇u) ∂ f (x, u, ∇u)
δ F(u, ϕ ) = ·ϕ + · ∇ϕ dx.
Ω ∂s ∂ξ
230 9 Basic Concepts on the Calculus of Variations
F(u + εϕ ) − F(u)
δ F(u, ϕ ) = lim .
ε →0 ε
Observe that
that is,
∂ f (x, u, ∇u) ∂ f (x, u, ∇u)
δ F(u, ϕ ) = ·ϕ + · ∇ϕ dx.
Ω ∂s ∂ξ
9.5 Elementary Convexity 231
T (u + v) − T(u) − T (u)(v) Y
lim = 0, v = θ .
v→θ v U
Thus
f (x + λ (y − x)) − f (x)
≤ f (y) − f (x), ∀λ ∈ (0, 1].
λ
Letting λ → 0+ we obtain
then f is convex.
232 9 Basic Concepts on the Calculus of Variations
that is,
f (x) ≥ x, x∗ Rn − f ∗ (x∗ ), ∀x, x∗ ∈ Rn .
Observe that from (9.4)
and thus
f (x) = sup {x, x∗ Rn − f (x∗ )}, ∀x ∈ Rn .
x∗ ∈Rn
δ 2 F(u, ϕ , η ),
by
δ F(u + εη , ϕ ) − δ F(u, ϕ )
δ 2 F(u, ϕ , η ) = lim .
ε →0 ε
If such a limit exists ∀ϕ , η ∈ V , we say that F is twice Gâteaux differentiable at u.
Finally, if η = ϕ , we denote δ 2 F(u, ϕ , η ) = δ 2 F(u, ϕ ).
Corollary 9.5.7. Let U be a Banach space. Suppose F : U → R is a twice Gâteaux
differentiable functional and that
δ 2 F(u, ϕ ) ≥ 0, ∀u ∈ U, ϕ ∈ V .
Then, F is convex.
Proof. Pick u, v ∈ U. Define φ (ε ) = F(u + ε (v − u)). By hypothesis, φ is twice
differentiable, so that
Therefore, by hypothesis,
F(v) ≥ F(u), ∀v ∈ U,
F(v) ≥ F(u), ∀v ∈ U.
φ (ε ) = F(u0 + εϕ ).
Observe that for |ε | < min{r, 1}, for some ε̃ such that |ε̃ | ≤ |ε |, we have
Hence,
F(u0 ) ≤ F(u0 + εϕ ), ∀ ε , ϕ such that |ε | < r, ϕ U < 1.
The proof is complete.
δ 2 F(u, ϕ ) ≥ 0, ∀ϕ ∈ Cc∞ (Ω , RN ),
9.6 The Legendre–Hadamard Condition 235
then
fξαi ξ k (x, u(x), ∇u(x))ρ i ρ k ηα ηβ ≥ 0, ∀x ∈ Ω , ρ ∈ RN , η ∈ Rn .
β
Proof. Suppose
δ 2 F(u, ϕ ) ≥ 0, ∀ϕ ∈ Cc∞ (Ω ; RN ).
We denote δ 2 F(u, ϕ ) by
δ 2 F(u, ϕ ) = a(x)Dϕ (x) · Dϕ (x) dx
Ω
+ b(x)ϕ (x) · Dϕ (x) dx + c(x)ϕ (x) · ϕ (x) dx, (9.6)
Ω Ω
where
a(x) = fξ ξ (x, u(x), Du(x)),
b(x) = 2 fsξ (x, u(x), Du(x)),
and
c(x) = fss (x, u(x), Du(x)).
Now consider v ∈ Cc∞(B1 (0), RN ). Thus given x0 ∈ Ω for λ sufficiently small we
have that ϕ (x) = λ v x−x λ
0
is an admissible direction. Now we introduce the new
coordinates y = (y , . . . , y ) by setting y = λ −1 (x − x0 ) and multiply (9.6) by λ −n to
1 n
obtain
{a(x0 + λ y)Dv(y) · Dv(y) + 2λ b(x0 + λ y)v(y) · Dv(y)
B1 (0)
αβ β
where a = {ai j }, b = {b jk } and c = {c jk }. Since a, b and c are continuous, we have
where
αβ αβ
f˜jk = a jk (x0 ) = fξαi ξ k (x0 , u(x0 ), ∇u(x0 )).
β
236 9 Basic Concepts on the Calculus of Variations
ρ = (ρ 1 , . . . , ρ N ) ∈ RN
and
η = (η1 , . . . , ηn ) ∈ Rn
and ζ ∈ Cc∞ (B1 (0)). From (9.7) we obtain
αβ
0 ≤ f˜jk ρ j ρ k (ηα t(−sin((η · y)t)ζ + cos((η · y)t)Dα ζ )
B1 (0)
· ηβ t(−sin((η · y)t)ζ + cos((η · y)t)Dβ ζ dy (9.8)
By analogy for
v j = ρ j sin((η · y)t)ζ (y)
we obtain
αβ
0 ≤ f˜jk ρ j ρ k (ηα t(cos((η · y)t)ζ + sin((η · y)t)Dα ζ )
B1 (0)
· ηβ t(cos((η · y)t)ζ + sin((η · y)t)Dβ ζ dy (9.9)
Summing up these last two equations, dividing the result by t 2 , and letting t → +∞
we obtain
αβ
0 ≤ f˜jk ρ j ρ k ηα ηβ ζ 2 dy,
B1 (0)
Here we present the Weierstrass condition for the special case N ≥ 1 and n = 1.
We start with a definition.
Definition 9.7.1. We say that u ∈ Ĉ1 ([a, b]; RN ) if u : [a, b] → RN is continuous in
[a, b] and Du is continuous except on a finite set of points in [a, b].
where
U = {u ∈ Ĉ1 ([a, b]; RN ) | u(a) = α , u(b) = β }.
Suppose u ∈ U minimizes locally F on U, that is, suppose that there exists ε0 > 0
such that
F(u) ≤ F(v), ∀v ∈ U, such that u − v ∞ < ε0 .
Under such hypotheses, we have
and
E(x, u(x), u (x−), w) ≥ 0, ∀x ∈ [a, b], w ∈ RN ,
where
u (x+) = lim u (x + h),
h→0+
and
E(x, s, ξ , w) = f (x, s, w) − f (x, s, ξ ) − fξ (x, s, ξ )(w − ξ ).
Proof. Fix x0 ∈ (a, b) and w ∈ RN . Choose 0 < ε < 1 and h > 0 such that u + v ∈ U
and
v ∞ < ε0
where v(x) is given by
⎧
⎨ (x − x0 )w, if 0 ≤ x − x0 ≤ ε h,
v(x) = ε̃ (h − x + x0)w, if ε h ≤ x − x0 ≤ h,
⎩
0, otherwise,
where
ε
ε̃ = .
1−ε
From
F(u + v) − F(u) ≥ 0
we obtain
x0 +h
f (x, u(x) + v(x), u (x) + v (x)) dx
x0
x0 +h
− f (x, u(x), u (x)) dx ≥ 0. (9.10)
x0
Define
x − x0
x̃ = ,
h
238 9 Basic Concepts on the Calculus of Variations
so that
dx
d x̃ = .
h
From (9.10) we obtain
1
h f (x0 + x̃h, u(x0 + x̃h) + v(x0 + x̃h), u (x0 + x̃h) + v(x0 + x̃h) d x̃
0
1
−h f (x0 + x̃h, u(x0 + x̃h), u (x0 + x̃h)) d x̃ ≥ 0. (9.11)
0
Letting h → 0 we obtain
In this section we present a proof for the Weierstrass necessary condition for
N ≥ 1, n ≥ 1. Such a result may be found in similar form in [37].
F(u + ϕ ) ≥ F(u),
ϕ ∞ < ε .
Here
F(u) = f (x, u, Du) dx,
Ω
where we recall to have denoted
∂ ui
Du = ∇u = .
∂xj
Under such hypotheses, for all x ∈ Ω and each rank-one matrix η = {ρi β α } =
{ρ ⊗ β }, we have that
where
δ F(u; ϕ ) = 0, ∀ϕ ∈ Cc∞ (Ω ; RN ),
that is,
(ϕ · fs (x, u(x), Du(x)) + Dϕ · fξ (x, u(x), Du(x)) dx = 0,
Ω
and hence,
( f (x, u(x), Du(x) + Dϕ (x)) − f (x, u(x), Du(x)) dx
Ω
− (ϕ (x) · fs (x, u(x), Du(x)) − Dϕ (x) · fξ (x, u(x), Du(x)) dx
Ω
≥ 0, (9.15)
240 9 Basic Concepts on the Calculus of Variations
∀ϕ ∈ V , where
V = {ϕ ∈ Cc∞ (Ω ; RN ) : ϕ ∞ < ε }.
Choose a unit vector e ∈ Rn and write
x = (x · e)e + x,
where
x · e = 0.
Denote De v = Dv · e and let ρ = (ρ1 , . . . ., ρN ) ∈ RN .
Also, let x0 be any point of Ω . Without loss of generality assume x0 = 0.
Choose λ0 ∈ (0, 1) such that Cλ0 ⊂ Ω , where
Cλ0 = {x ∈ Rn : |x · e| ≤ λ0 and x ≤ λ0 }.
and such that φk converges uniformly to φλ on each compact subset of
We emphasize the choice of {φk } may be such that for some K > 0 we have φ ∞ <
K, φk ∞ < K and φk ∞ < K, ∀k ∈ N.
Observe that for any sufficiently small λ > 0 we have that ϕk defined by
Note that for such a ϕ (x), the integrand of (9.15) vanishes if x ∈ Cλ , where
Cλ = {x ∈ Rn : |x · e| ≤ λ and x ≤ λ }.
Cλ− = {x ∈ Cλ : x · e ≤ 0},
and
Cλ+ = {x ∈ Cλ : x · e > 0}.
Hence, denoting
and
Now define
y = ye e + y,
where
x·e
ye = ,
λ2
and
x
y= .
λ
The sets Cλ− and Cλ+ correspond, concerning the new variables, to the sets B−
λ and
B+
λ, where
B−λ = {y : y ≤ 1, and − λ
−1
≤ ye ≤ 0},
B+ −1
λ = {y : y ≤ 1, and 0 < y ≤ λ }.
e
where
x = (x · e)e + x = λ 2 ye + λ y ≡ x(y).
Observe that
⎧
⎨ ρφ ( y 2 ) if − 1 ≤ ye ≤ 0,
De ϕ (x) = ρφ ( y )(−λ ) if 0 ≤ ye ≤ λ −1 ,
2 (9.21)
⎩
0, otherwise.
y ∈ B+ − −
λ , or y ∈ Bλ \ B1 ,
that
|g(x(y))| ≤ o(λ ), as λ → 0.
Since the Lebesgue measures of B− +
λ and Bλ are bounded by
2n−1 /λ
lim o(1)/λ = 0,
λ →0+
x(y) → 0,
and on B−
1 (up to the limit set B)
we get
[ f (0, u(0), Du(0) + ρφ ( y 2 )e) − f (0, u(0), Du(0))
B
−ρφ ( y 2 )e fξ (0, u(0), Du(0))] dy2 . . . dyn
≥ 0, (9.23)
where B is an appropriate limit set (we do not provide more details here) such that
B = {y ∈ Rn : ye = 0 and y ≤ 1}.
9.9 The du Bois–Reymond Lemma 243
Dϕ (x) → ρφ ( y 2 )e, as λ → 0+ .
Finally, inequality (9.23) is valid for a sequence {φn } (in place of φ ) such that
∀n ∈ N.
Letting n → ∞, from (9.23), we obtain
where
V = {ϕ ∈ C1 [a, b] | ϕ (a) = ϕ (b) = 0},
then there exists c ∈ R such that
Proof. Define
b
1
c= u(t) dt,
b−a a
and x
ϕ (x) = (u(t) − c) dt.
a
Thus we have ϕ (a) = 0 and
b
ϕ (b) = u(t) dt − c(b − a) = 0.
a
Therefore
b
0≤ (u(x) − c)2 dx
a
b
= (u(x) − c)ϕ (x) dx
a
b
= u(x)ϕ (x) dx − c[ϕ (x)]ba = 0. (9.25)
a
Thus b
(u(x) − c)2 dx = 0,
a
and being u(x) − c continuous, we finally obtain
∀ϕ ∈ V , where
V = {ϕ ∈ C1 [a, b] | ϕ (a) = ϕ (b) = 0},
then
v ∈ C1 ([a, b])
and
v (x) = u(x), ∀x ∈ [a, b].
Proof. Define x
u1 (x) = u(t) dt, ∀x ∈ [a, b].
a
That is,
b
(v(x) − u1 (x))ϕ (x) dx, ∀ϕ ∈ V .
a
Hence
v = u1 + c ∈ C1 ([a, b]),
so that
v (x) = u1 (x) = u(x), ∀x ∈ [a, b].
The proof is complete.
Proof. Fix ϕ ∈ Cc∞ ([a, b]; RN ). Integration by parts of the extremal condition
δ F(u, ϕ ) = 0,
246 9 Basic Concepts on the Calculus of Variations
implies that
b
fξ (x, u(x), u (x)) · ϕ (x) dx
a
b x
− fs (t, u(t), u (t)) dt · ϕ (x) dx = 0.
a a
F(u) ≤ F(v), ∀v ∈ Cr ,
Let x0 ∈ (a, b) be a corner point of u. Denoting u0 = u(x0 ), ξ0+ = u (x0 + 0), and
ξ0− = u (x0 − 0), then the following relations are valid:
1. fξ (x0 , u0 , ξ0− ) = fξ (x0 , u0 , ξ0+ ),
2.
Remark 9.10.4. The conditions above are known as the Weierstrass–Erdmann corner
conditions.
τε (x) = x + ελ (x),
where λ ∈ Cc∞ (I). Observe that τε (a) = a and τε (b) = b, ∀ε > 0. Also τ0 (x) = x.
Choose ε0 > 0 sufficiently small such that for each ε satisfying |ε | < ε0 , we have
τε (x) > 0 and
ũε (x) = (u ◦ τε−1)(x) ∈ Cr .
9.10 The Weierstrass–Erdmann Conditions 247
Define
φ (ε ) = F(x, ũε , ũε (x)).
Thus φ has a local minimum at 0, so that φ (0) = 0, that is,
Defining
x̄ = τε−1 (x),
we obtain
1
d x̄ = dx,
1 + ελ (x̄)
that is,
dx = (1 + ελ (x̄)) d x̄.
Dropping the bar for the new variable, we may write
b
u (x)
F(ũε ) = f x + ελ (x), u(x), 1 + ελ (x) dx.
a 1 + ελ (x)
From
dF(ũε )
|ε =0 ,
dε
we obtain
b
(λ fx (x, u(x), u (x)) + λ (x)( f (x, u(x), u (x))
a
− u (x) fξ (x, u(x), u (x)))) dx = 0. (9.28)
Since λ is arbitrary, from Proposition 9.9.2, (in fact from its version for u ∈ L1 ([a, b])
and respective extension for the N dimensional case, please see [37] for details), we
obtain
x
f (x, u(x), u (x)) − u (x) fξ (x, u(x), u (x)) − fx (t, u(t), u (t)) dt = c1
a
for some c1 ∈ RN .
248 9 Basic Concepts on the Calculus of Variations
Since ax fx (t, u(t), u (t)) dt + c1 is a continuous function (in fact absolutely con-
tinuous), the proof is complete.
U = {u ∈ W 1,2 (Ω ; RN ); u = u0 on Γ0 },
and in particular
G(x) > β /2, in Br (x0 ) ∩ Γ1 .
Choose 0 < r1 < r such that Br1 (x0 ) ∩ Γ0 = 0.
/ This is possible since Γ0 is closed and
x0 ∈ Γ1 .
Choose ϕ i ∈ Cc∞ (Br1 (x0 )) such that ϕ i ≥ 0 in Br1 (x0 ) and ϕ i > 0 in Br1 /2 (x0 ).
Therefore
β
G(x)ϕ i (x) dx > ϕ i dx > 0,
Γ1 2 Γ1
and this contradicts (9.29). Thus
G(x) ≤ 0, ∀x ∈ Γ1 ,
and by analogy
G(x) ≥ 0, ∀x ∈ Γ1 ,
so that
G(x) = 0, ∀x ∈ Γ1 .
The proof is complete.
Chapter 10
Basic Concepts on Convex Analysis
For this chapter the most relevant reference is Ekeland and Temam, [25].
n
Co(S) = ∑ λi ui | n ∈ N,
i=1
n
∑ λi = 1, λi ≥ 0, ui ∈ S, ∀i ∈ {1, . . . , n} . (10.2)
i=1
Definition 10.1.2 (Convex Functional). Let S be convex subset of the vector space
U. A functional F : S → R̄ = R ∪ {+∞, −∞} is said to be convex if
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 251
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 10,
© Springer International Publishing Switzerland 2014
252 10 Basic Concepts on Convex Analysis
Thus
Vλ × (−∞, λ ) ⊂ E pi(F)c
so that (u, r) is an interior point of E pi(F)c , and hence, since such a point in E pi(F)c
is arbitrary, we may conclude that E pi(F)c is open so that E pi(F) is closed in
σ (U,U ∗ ) product with the usual topology in R.
Now assume (2). Observe that
Since from the hypothesis E pi(F) is closed, we have that HγF × {γ } is closed and
hence HγF is closed.
Now assume (3). To obtain (4) just take the complement of HγF . Suppose (4) is
valid. Let γ ∈ R such that
γ < F(u).
Since GFγ is open in σ (U,U ∗ ) there exists a weak neighborhood V (u) such that
V (u) ⊂ GFγ ,
10.1 Convex Sets and Convex Functions 253
so that
F(v) > γ , ∀v ∈ V (u),
and hence
inf F(v) ≥ γ ,
v∈V (u)
Let λ < F(u). Thus there exists a weak neighborhood V (u) such that F(v) ≥ F(u) >
λ , ∀v ∈ V (u). The proof is complete.
Similar result is valid for the strong topology of the Banach space U so that a func-
tional F : U → R ∪ {+∞} is strongly lower semicontinuous (l.s.c.) at u ∈ U, if
where
Proof. By definition, the convex envelope of F is the supremum of all affine contin-
uous minorants of F. We can consider only the maximal minorants, which functions
of the form
Thus,
F ∗ ≤ F ∗∗∗ . (10.12)
so that
F(u + λ h) − F(u)
lim = h, u∗ U , ∀h ∈ U. (10.15)
λ →0 λ
The vector u∗ is said to be the Gâteaux derivative of F : U → R at u and may be
denoted as follows:
∂ F(u)
u∗ = or u∗ = δ F(u) (10.16)
∂u
Definition 10.1.16 (Sub-gradients). Given F : U → R̄, we define the set of sub-
gradients of F at u, denoted by ∂ F(u) as
u, Λ ∗ v∗ U = Λ u, v∗ Y , ∀u ∈ U, v∗ ∈ Y ∗ . (10.18)
Proof. By translation, we may reduce the problem to the case where u = θ and
F(u) = 0. Let V be a neighborhood of origin such that F(v) ≤ a < +∞, ∀v ∈ V .
Define W = V ∩ (−V ) (which is a symmetric neighborhood of origin). Pick ε ∈
(0, 1). If v ∈ ε W , since F is convex and
v
∈V (10.19)
ε
we may infer that
Also
−v
∈V. (10.21)
ε
Thus,
F(v) ε
F(θ ) ≤ + F(−v/ε ),
1+ε 1+ε
so that
Therefore
|F(v)| ≤ ε a, ∀v ∈ ε W , (10.23)
and
v − u, u∗U ≥ 0, ∀v ∈ U, (10.27)
β
− v, u∗/α U ≤ F(v), (10.28)
α
and
β
− u, u∗/α U = F(u), (10.29)
α
or
so that
and
for almost all x ∈ S, ξ → g(x, ξ ) is a continuous function.
The proof of next results may be found in Ekeland and Temam [25].
Proposition 10.1.21. Let E and F be two Banach spaces, S a Borel subset of Rn and
g : S × E → F a Carathéodory mapping. For each measurable function u : S → E,
let G1 (u) be the measurable function x → g(x, u(x)) ∈ F.
If G1 maps L p (S, E) into Lr (S, F) for 1 ≤ p, r < ∞, then G1 is continuous in the
norm topology.
For the functional G : U → R, defined by G(u) = S g(x, u(x))dS, where U =
U ∗ = [L2 (S)]l (this is a special case of the more general hypothesis presented in
[25]) we have the following result.
Proposition 10.1.22. Considering the last proposition we can express G∗ : U ∗ →
R̄ as
G∗ (u∗ ) = g∗ (x, u∗ (x))dS, (10.32)
S
∂ g(x0 )
y∗i = , (10.34)
∂ xi
and RnL = {y∗ ∈ Rn such that (10.34) has a unique solution}.
Furthermore, considering the functional G : Y → R defined as G(v) = S g(v)dS,
we define the associated Legendre transform functional, denoted by G∗L : YL∗ → R as
G∗L (v∗ ) = g∗L (v∗ )dS, (10.35)
S
∂ g −1 ∗
x0 (y∗ ) = [ ] (y ). (10.36)
∂x
Then, ∀ i ∈ {1, . . . , n} we may write
∂ g(x0 ) ∂ g∗ (y∗ )
y∗i = ⇔ x0 i = L ∗ (10.37)
∂ xi ∂ yi
∂ g(x0 )
y∗i = , ∀ i ∈ {1, . . . , n}, (10.38)
∂ xi
thus,
∂ g∗L (y∗ )
x0 i = , ∀i ∈ {1, . . . , n}. (10.43)
∂ y∗i
∂ g(x̄0 )
y∗i = ∀i ∈ {1, . . . , n}, (10.44)
∂ xi
and
∂ g∗L (y∗ )
x̄0i = = x0i , ∀ i ∈ {1, . . . , n}, (10.48)
∂ y∗i
∂ g(x̄0 ) ∂ g(x0 )
y∗i = = ∀ i ∈ {1, . . . , n}. (10.49)
∂ xi ∂ xi
∂ G(Λ (u0 ))
where v∗0 = ∂v is supposed to be such that v∗0 (x) ∈ RnL , a.e. in S and in this
case
∂ G(Λ u0 )
Λ∗ − f =θ (10.52)
∂v
∂ G(Λ u0 )
which, as v∗0 = ∂v , implies that
Λ ∗ v∗0 − f = θ , (10.53)
and
∂ g(Λ u0 )
v∗0i = . (10.54)
∂ xi
260 10 Basic Concepts on Convex Analysis
∂ g∗L (v∗0 )
Λi (u0 ) = , for i ∈ {1, .., n} (10.55)
∂ y∗i
which means
∂ G∗L (v∗0 )
Λ u0 = . (10.56)
∂ v∗
Therefore from (10.53) and (10.56) we have
that is,
Λ ∗ v∗0 − f = θ (10.59)
and
∂ G∗L (v∗0 )
Λ u0 = . (10.60)
∂ v∗
Clearly, from (10.60), the last proposition and (10.59), we can write
∂ G(Λ (u0 ))
v∗0 = (10.61)
∂v
and
∂ G(Λ u0 )
Λ∗ − f = θ, (10.62)
∂v
which implies
δ J(u0 ) = θ . (10.63)
Finally, we have
Observe that
φ ∗ (0, p∗ ) = sup {0, uU + p, p∗ Y − φ (u, p)} ≥ −φ (u, 0), (10.70)
(u,p)∈U×Y
or
then h is convex.
If h(p) = +∞ or h(q) = +∞ we are done. Thus let us assume h(p) < +∞ and h(q) <
+∞. For each a > h(p) there exists u ∈ U such that
Thus
so that
Proposition 10.2.3. The set of solutions of the problem P ∗ (the dual problem) is
identical to ∂ h∗∗ (0).
Proof. Consider p∗0 ∈ Y ∗ a solution of problem P ∗ , that is,
−φ ∗ (0, p∗0 ) ≥ −φ ∗ (0, p∗ ), ∀p∗ ∈ Y ∗ , (10.81)
which is equivalent to
−h∗ (p∗0 ) ≥ −h∗ (p∗ ), ∀p∗ ∈ Y ∗ , (10.82)
which is equivalent to
Proof. By hypothesis h(0) ∈ R and as was shown above, h is convex. As the func-
tion p → φ (u0 , p) is convex and continuous at 0 ∈ Y , there exists a neighborhood V
of zero in Y such that
h(0) = inf {φ (u, 0)} = sup {−φ ∗ (0, p∗ )} = h∗∗ (0). (10.87)
u∈U p∗ ∈Y ∗
Proof. The existence of solutions for the primal problem follows from the direct
method of calculus of variations. That is, considering a minimizing sequence, from
above (coercivity hypothesis), such a sequence is bounded and has a weakly conver-
gent subsequence to some u0 ∈ U. Finally, from the lower semicontinuity of primal
formulation, we may conclude that u0 is a minimizer. The other conclusions follow
from Theorem 10.2.4 just observing that
so that
Thus,
and solutions u0 and p∗0 for the primal and dual problems, respectively, imply that
Our main objective in this section is to state and prove the min–max theorem.
Observe that
inf F(u) ≤ F(u0 ),
u∈A
so that
inf sup L(u, v) ≤ sup L(u0 , v). (10.95)
u∈A v∈B v∈B
Define
G(v) = inf L(u, v).
u∈A
Thus
sup G(v) ≥ G(v0 ),
v∈B
10.3 The Min–Max Theorem 265
so that
sup inf L(u, v) ≥ inf L(u, v0 ). (10.96)
v∈B u∈A u∈A
Hence
so that
sup inf L(u, v) ≤ sup L(u, v), ∀u ∈ A,
v∈B u∈A vInB
and hence
sup inf L(u, v) ≤ inf sup L(u, v). (10.99)
v∈B u∈A u∈A v∈B
Conversely suppose
max inf L(u, v) = min sup L(u, v).
v∈B u∈A u∈A v∈B
As above defined,
F(u) = sup L(u, v),
v∈B
and
G(v) = inf L(u, v).
u∈A
266 10 Basic Concepts on Convex Analysis
so that
F(u0 ) = sup L(u0 , v) = inf L(u, v0 ) = G(v0 ).
v∈B u∈U
In particular
Therefore
sup L(u0 , v) = L(u0 , v0 ) = inf L(u, v0 ).
v∈B u∈U
L(u0 , v) ≤ α , ∀v ∈ B,
and
L(u, v0 ) ≥ α , ∀u ∈ A.
Under such hypotheses (u0 , v0 ) is a saddle point of L, that is,
L(u0 , v0 ) ≤ α ,
and
L(u0 , v0 ) ≥ α ,
so that
L(u0 , v) ≤ α = L(u0 , v0 ) ≤ L(u, v0 ), ∀u ∈ A, v ∈ B.
This completes the proof.
Proof. Fix v ∈ B. Observe that Gv (u) = L(u, v) is convex and lower semicontinuous.
Therefore it is weakly lower semicontinuous on the weak compact set A. At first we
assume the additional hypothesis that Gv (u) is strictly convex, ∀v ∈ B. Hence Gv (u)
attains a unique minimum on A. We denote the optimal u ∈ A by u(v).
Define
G(v) = min Gv (u) = min L(u, v).
u∈A u∈U
Thus,
G(v) = L(u(v), v).
The function G(v) is expressed as the minimum of a family of concave
weakly upper semicontinuous functions, and hence it is also concave and upper
semicontinuous.
Moreover, G(v) is bounded above on the weakly compact set B, so that there
exists v0 ∈ B such that
Observe that
G(v0 ) = min L(u, v0 ) ≤ L(u, v0 ), ∀u ∈ U.
u∈A
Observe that from the concerned concavity, for u ∈ A, v ∈ B, and λ ∈ (0, 1), we have
Therefore,
∀u ∈ A, k ∈ N.
Recalling that λnk → 0, from this and (10.106), we obtain
so that
G(v0 ) ≥ lim inf L(unk , v) ≥ L(u0 , v), ∀v ∈ B.
k→∞
and
α = G(v0 ) = min L(u, v0 ) ≤ L(u, v0 ), ∀u ∈ A.
u∈U
10.3 The Min–Max Theorem 269
From these last two results and Proposition 10.2 we have that (u0 , v0 ) is a saddle
point for L. Now assume that
Gv (u) = L(u, v)
In such a case
(Gv )n (u) = Ln (u, v)
is strictly convex for all n ∈ N.
From above we mainly obtain (un , vn ) ∈ A × B such that
L(un , v) + un U /n ≤ L(un , vn ) + un U /n
≤ L(u, vn ) + u /n. (10.108)
un u0 , weakly in U,
vn v0 , weakly in Y,
so that
Hence,
L(u0 , v) ≤ L(u, v0 ), ∀u ∈ A, v ∈ B,
so that
L(u0 , v) ≤ L(u0 , v0 ) ≤ L(u, v0 ), ∀u ∈ A, v ∈ B.
This completes the proof.
In the next result we deal with more general situations.
Theorem 10.2. Let U,Y be reflexive Banach spaces, A ⊂ U, B ⊂ Y and let L : A ×
B → R be a functional.
Suppose that
1. A ⊂ U is convex, closed, and nonempty.
2. B ⊂ Y is convex, closed, and nonempty.
3. For each u ∈ A, Fu (v) = L(u, v) is concave and upper semicontinuous.
270 10 Basic Concepts on Convex Analysis
An = {u ∈ A : u U ≤ n}.
Fix n ∈ N. The sets An and B are closed, convex, and bounded, so that from the last
Theorem 10.1 there exists a saddle point (un , vn ) ∈ An × B for
L : An × B → R.
Hence,
L(un , v) ≤ L(un , vn ) ≤ L(u, vn ), ∀u ∈ An , v ∈ B.
For a fixed ũ ∈ A1 we have
L(un , ṽ) ≤ L(un , vn )
≤ L(ũ, vn )
≤ sup L(ũ, v) ≡ b ∈ R. (10.110)
v∈B
Hence
a ≤ L(un , ṽ) ≤ L(un , vn ) ≤ b, ∀n ∈ N.
Therefore {L(un , vn )} is bounded.
Moreover, from the coercivity hypotheses and
a ≤ L(un , ṽ) ≤ b, ∀n ∈ N,
10.4 Relaxation for the Scalar Case 271
unk u0 , weakly in U,
vnk v0 , weakly in Y,
L(unk , vnk ) → α ∈ R,
as k → ∞. Fix (u, v) ∈ A × B. Observe that if nk > n0 = u U , then
that is,
L(u0 , v) ≤ α ≤ L(u, v0 ), ∀u ∈ A, v ∈ B.
From this and Proposition 10.2 we may conclude that (u0 , v0 ) is a saddle point
for L : A × B → R.
The proof is complete.
r
u(x) − ∑ αk uk < ε , ∀x ∈ Ω , (10.115)
k=1
r
u(x) = ∑ αk uk (x), ∀x ∈ ∂ Ω . (10.116)
k=1
r
|u(x) − ∑ αk uk (x)| < ε , ∀x ∈ Ω (10.120)
k=1
r
u(x) = ∑ αk uk (x), ∀x ∈ ∂ Ω . (10.121)
k=1
Proof. It is sufficient to establish the result for functions uk affine over Ω , since
Ω can be divided into pieces on which uk are affine, and such pieces can be put
together through (10.121). Let ε > 0 be given. We know that simple functions are
10.4 Relaxation for the Scalar Case 273
Now we apply Theorem 10.4.2, to each of the open sets Oi ; therefore there exists
a locally Lipschitz function u : Oi → R and there exist r open disjoints spaces Ωki ,
1 ≤ k ≤ r, such that
r
u(x) − ∑ αk uk (x) ≤ δ , ∀x ∈ Oi (10.128)
k=1
r
u(x) = ∑ αk uk (x), ∀x ∈ ∂ Oi . (10.129)
k=1
From | f (x, ∇u(x))| ≤ c(x), m(Oi − ∪rk=1 Ωki ) ≤ δ and (10.124) we obtain
274 10 Basic Concepts on Convex Analysis
r
f (x, ∇u(x))dx − ∑ f (x, ∇uk (x)dx
Oi k=1 Ω i
k
= f (x, ∇u(x))dx ≤ ε /N1 . (10.131)
Oi −∪rk=1 Ω ki
We do not prove the next result. It is a well-known result from the finite element
theory.
un → u, in L p (Ω ) (10.135)
and
∇v − ∇u ∈ V , (10.138)
u = v on ∂ Ω ,
v − u ∞ < ε , (10.139)
and
f (x, ∇v(x))dx − f ∗∗
(x, ∇u(x))dx < ε. (10.140)
Ω Ω
Proof. Suppose given ε > 0, u ∈ W 1,p (Ω ) piecewise affine continuous, and a neigh-
borhood V of zero, which may be expressed as
V = {w ∈ L p (Ω , RN ) | hm · wdx < η ,
Ω
∀m ∈ {1, . . . , M}}, (10.141)
N+1
∑ αk ξk = ∇u, ∀x ∈ Δi , (10.142)
k=1
and
N+1
∑ αk f (x, ξk ) = f ∗∗ (x, ∇u(x)). (10.143)
k=1
We recall that ρ does not depend on ε . Furthermore, for each i ∈ {1, . . . , r}, there
exists a compact subset Ki ⊂ Δi such that
ε1
[a1 (x) + c1(x) max {|ξ | p}]dx < . (10.146)
Δ i −Ki |ξ |≤ρ r
Also, observe that the sets Ki may be obtained such that the restrictions of f and f ∗∗
to Ki × ρ B are continuous, so that from this and from the compactness of ρ B, for all
x ∈ Ki , we can find an open ball ωx with center in x and contained in Ω , such that
ε1
| f ∗∗ (y, ∇u(x)) − f ∗∗ (x, ∇u(x))| < , ∀y ∈ ωx ∩ Ki , (10.147)
m(Ω )
and
ε1
| f (y, ξ ) − f (x, ξ )| < , ∀y ∈ ωx ∩ Ki , ∀ξ ∈ ρ B. (10.148)
m(Ω )
We can cover the compact set Ki with a finite number of those open ball ωx , de-
noted by ω j , 1 ≤ j ≤ l. Consider the open sets ω j = ω j − ∪i=1 ω̄i . We have that
j−1
and
Finally we set
vi = u on Δi − ∪lj=1ω j . (10.154)
and
ε1
| f (x, ∇v(x)|dx < . (10.158)
Δ i −Ki r
so that
| f (x, ∇v(x))dx − f ∗∗ (x, ∇u(x))dx|
Ki Ki
ε1 ε1 m(Ki )
≤ + . (10.160)
r m(Ω )
v − u ∞ < ε1 , (10.162)
278 10 Basic Concepts on Convex Analysis
and thus
ĥm · (∇v(x) − ∇u(x))dx = − div(ĥm )(v(x) − u(x))dx
Ω Ω
≤ div(ĥm ) Lq (Ω ) v − u L p(S)
≤ Cε1 m(Ω )1/p
η
< . (10.163)
2
Also we have that
(ĥm − hm ) · (∇v − ∇u)dx
Ω
η
≤ ĥm − hm Lq (Ω ,RN ) ∇v − ∇u L p(Ω ,RN ) ≤ . (10.164)
2
Thus
hm · (∇v − ∇u)dx < η , ∀m ∈ {1, . . . , M}. (10.165)
Ω
∇v − ∇u ∈ V , (10.166)
and
f (x, ∇v(x))dx − f ∗∗
(x, ∇u(x))dx < ε. (10.167)
Ω Ω
so that
1
∇w − ∇u ∈ V , (10.169)
2
and
ε
f ∗∗ (x, ∇w(x))dx − f ∗∗
(x, ∇u(x))dx < . (10.170)
Ω Ω
2
10.4 Relaxation for the Scalar Case 279
1
∇v − ∇w ∈ V , (10.171)
2
and
ε
f ∗∗ (x, ∇w(x))dx − f (x, ∇v(x))dx < . (10.172)
Ω Ω 2
Finally, from (10.169), (10.171), and from the fact that weak neighborhoods are
convex, we have
∇v − ∇u ∈ V . (10.174)
To finish this chapter, we present two theorems which summarize the last results.
Theorem 10.4.7. Let f be a Carathéodory function from Ω × RN into R which sat-
isfies
The solutions of relaxed problem are weak cluster points in W01,p (Ω ) of the mini-
mizing sequences of primal problem.
Proof. The existence of solutions for the convex relaxed formulation is a conse-
quence of the reflexivity of U and coercivity hypothesis, which allows an applica-
tion of the direct method of calculus of variations. That is, considering a minimizing
sequence, from above (coercivity hypothesis), such a sequence is bounded and has
a weakly convergent subsequence to some û ∈ W 1,p (Ω ). Finally, from the lower
semicontinuity of relaxed formulation, we may conclude that û is a minimizer. The
relation (10.176) follows from last theorem.
The solutions of relaxed problem are weak cluster points in W 1,p (Ω ) of the min-
imizing sequences of primal problem.
Proof. Just apply the last theorem to the integrand g(x, ξ ) = f (x, ξ + ∇u0 ). For
details see [25].
In this section we present and prove the Ekeland variational principle. This proof
may be found in Giusti, [39], pp. 160–161.
Theorem 10.5.1 (Ekeland Variational Principle). Let (U, d) be a complete metric
space and let F : U → R be a lower semicontinuous bounded below functional
taking a finite value at some point.
Let ε > 0. Assume for some u ∈ U we have
u1 = u,
and having u1 , . . . , un , select un+1 as specified in the next lines. First, define
and hence
m
ε d(un+m , un ) ≤ ∑ d(un+i , un+i−1 ) ≤ F(un ) − F(un+m ). (10.181)
i=1
F(un ) → α as n → ∞.
In particular we have
w = v. (10.183)
Thus, from this and (10.182), we have
so that
2 lim inf{F(un+1 )} ≤ F(v) − ε d(v, w) + lim inf{F(un )}.
n→∞ n→∞
Hence,
F(v) ≤ lim inf{F(un+1 )} ≤ F(v) − ε d(v, w),
n→∞
so that
0 ≤ −ε d(v, w),
which contradicts (10.183).
Thus 3 holds.
Remark 10.5.2. We may introduce in U a new metric given by d1 = ε 1/2 d. We
highlight that the topology remains the same and also F remains lower semicon-
tinuous. Under the hypotheses of the last theorem, if there exists u ∈ U such that
F(u) < infu∈U F(u) + ε , then there exists v ∈ U such that
1. d(u, v) ≤ ε 1/2 ,
2. F(v) ≤ F(u),
3. F(v) ≤ F(w) + ε 1/2 d(u, w), ∀w ∈ U.
Remark 10.5.3. Observe that if U is a Banach space,
Similarly
Thus
We have thus obtained, from the last theorem and remarks, the following result.
Theorem 10.5.4. Let U be a Banach space. Let F : U → R be a lower semicon-
tinuous Gâteaux differentiable functional. Given ε > 0 suppose that u ∈ U is such
that
√
u − v U ≤ ε, (10.191)
and
√
δ F(v) U ∗ ≤ ε. (10.192)
where
and
√
δ J(uε ) U ∗ < ε. (10.196)
We finish this chapter with an important result for vectorial problems in the calculus
of variations.
Theorem 10.5.6. Let U be a reflexive Banach space. Consider (G ◦ Λ ) : U → R and
(F ◦ Λ1 ) : U → R l.s.c. functionals such that J : U → R defined as
J(u) = (G ◦ Λ )(u) − (F ◦ Λ1 )(u) − u, f U
is below bounded. (Here Λ : U → Y and Λ1 : U → Y1 are continuous linear oper-
ators whose adjoint operators are denoted by Λ ∗ : Y ∗ → U ∗ and Λ1∗ : Y ∗ → U ∗ ,
respectively). Also we suppose the existence of L : Y1 → Y continuous and linear
operator such that L∗ is onto and
Λ (u) = L(Λ1 (u)), ∀u ∈ U.
Under such assumptions, we have
inf {J(u)} ≥ sup { ∗inf ∗ {F ∗ (L∗ z∗ ) − G∗(v∗ + z∗ )}},
u∈U v∗ ∈A∗ z ∈Y1
where
A∗ = {v∗ ∈ Y ∗ | Λ ∗ v∗ = f }.
284 10 Basic Concepts on Convex Analysis
Λ ∗ v∗0 − f = 0.
Suppose u0 ∈ U is such that
∂ F ∗ (L∗ z∗0 )
= Λ1 u 0 ,
∂ v∗
so that
Λ u0 ∈ ∂ G∗ (v∗0 + z∗0 ).
Also we assume that there exists a sequence {un } ⊂ U such that un u0 weakly in
U and
G(Λ un ) → G∗∗ (Λ u0 ) as n → ∞.
Under these additional assumptions we have
that is,
so that
∀v∗ ∈ A∗ , u ∈ U
10.5 Duality Suitable for the Vectorial Case 285
which means
where
A∗ = {v∗ ∈ Y ∗ | Λ ∗ v∗ = f }.
Now suppose
∂ F ∗ (L∗ z∗0 )
L ∈ ∂ G∗ (v∗0 + z∗0 ),
∂ v∗
and u0 ∈ U is such that
∂ F ∗ (L∗ z∗0 )
= Λ1 u 0 .
∂ v∗
Observe that
Λ u0 = L(Λ1 u0 ) ∈ ∂ G(v∗0 + z∗0 )
implies that
For this chapter the most relevant reference is the excellent book of Luenberger
[47], where more details may be found. Other relevant references are [15, 40–42].
We start with the definition of cone.
Definition 11.1.1 (Cone). Given a Banach space U, we say that C ⊂ U is a cone
with the vertex at the origin; if given u ∈ C, we have that λ u ∈ C, ∀λ ≥ 0. By
analogy we define a cone with the vertex at p ∈ U as P = p + C, where C is any
cone with the vertex at the origin. From now on we consider only cones with vertex
at origin, unless otherwise indicated.
Definition 11.1.2. Let P be a convex cone in U. For u, v ∈ U we write u ≥ v (with
respect to P) if u − v ∈ P. In particular u ≥ θ if and only if u ∈ C. Also
If u∗ ∈ P+ , we write u∗ ≥ θ ∗ .
Proposition 11.1.3. Let U be a Banach space and P be a closed cone in U. If u ∈ U
satisfies u, u∗ U ≥ 0, ∀u∗ ≥ θ ∗ , then u ≥ θ .
Proof. We prove the contrapositive. Assume u ∈ P. Then by the separating hyper-
plane theorem there is an u∗ ∈ U ∗ such that u, u∗ U < p, u∗ U , ∀p ∈ P. Since P is
a cone we must have p, u∗ U ≥ 0; otherwise we would have u, u∗ > α p, u∗ U
for some α > 0. Thus u∗ ∈ P+ . Finally, since inf p∈P{p, u∗ U } = 0, we obtain
u, u∗ U < 0 which completes the proof.
Definition 11.1.4 (Convex Mapping). Let U, Z be vector spaces. Let P ⊂ Z be a
cone. A mapping G : U → Z is said to be convex if the domain of G is convex and
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 287
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 11,
© Springer International Publishing Switzerland 2014
288 11 Constrained Variational Optimization
Define
ω (α z1 + (1 − α )z2) = inf{F(u) | u ∈ Ω
and G(u) ≤ α z1 + (1 − α )z2}
(11.4)
≤ inf{F(u) | u = α u1 + (1 − α )u2 u1 , u2 ∈ Ω
and G(u1 ) ≤ z1 , G(u2 ) ≤ z2 }
(11.5)
≤α inf{F(u1 ) | u1 ∈ Ω , G(u1 ) ≤ z1 }
+ (1 − α ) inf{F(u2 ) | u2 ∈ Ω , G(u2 ) ≤ z2 }
(11.6)
≤αω (z1 ) + (1 − α )ω (z2). (11.7)
Now we establish the Lagrange multiplier theorem for convex global optimization.
Theorem 11.1.6. Let U be a vector space, Z a Banach space, Ω a convex subset of
U, and P a positive cone of Z. Assume that P contains an interior point. Let F be
a real convex functional on Ω and G a convex mapping from Ω into Z. Assume the
existence of u1 ∈ Ω such that G(u1 ) < θ . Defining
and
B = {(r, z) ∈ R × Z | r ≤ μ0 , z ≤ θ }, (11.11)
where μ0 = infu∈Ω {F(u) | G(u) ≤ θ }. Since F and G are convex, A and B are convex
sets. It is clear that A contains no interior point of B, and since N = −P contains
an interior point, the set B contains an interior point. Thus, from the separating
hyperplane theorem, there is a nonzero element w∗0 = (r0 , z∗0 ) ∈ W ∗ such that
From the nature of B it is clear that w∗0 ≥ θ . That is, r0 ≥ 0 and z∗0 ≥ θ . We will show
that r0 > 0. The point (μ0 , θ ) ∈ B; hence
If r0 = 0, then G(u1 ), z∗0 Z ≥ 0 and z∗0 = θ . Since G(u1 ) < θ and z∗0 ≥ θ we have
a contradiction. Therefore r0 > 0 and, without loss of generality, we may assume
r0 = 1. Since the point (μ0 , θ ) is arbitrarily close to A and B, we have
Hence
Corollary 11.1.7. Let the hypothesis of the last theorem hold. Suppose
Then
Proof. Suppose there is a u1 ∈ Ω such that F(u1 ) < F(u0 ) and G(u1 ) ≤ G(u0 ). Thus
so that
and
Suppose z∗0 and z∗1 are the Lagrange multipliers related to these problems. Then
11.2 Duality
where
and
Γ = {z ∈ Z | G(u) ≤ z f or some u ∈ Ω }.
Proof. Observe that
so that
If the infimum on the left side in (11.38) is achieved at some u0 ∈ U and the max on
the right side at z∗0 ∈ Z ∗ , then
or
ϕ (z∗ ) ≤ μ0 . (11.41)
Remark 11.3.1. This section was published in similar form by the journal Computa-
tional and Applied Mathematics, SBMAC-Springer, reference [15].
In this section we develop a new and simpler proof of the Lagrange multiplier
theorem in a Banach space context. In particular, we address the problem of min-
imizing a functional F : U → R subject to G(u) = θ , where θ denotes the zero
vector and G : U → Z is a Fréchet differentiable transformation. Here U, Z are Ba-
nach spaces. General results on Banach spaces may be found in [1, 26], for example.
For the theorem in question, among others, we would cite [13, 40, 47]. Specially the
proof given in [47] is made through the generalized inverse function theorem. We
emphasize such a proof is extensive and requires the continuous Fréchet differentia-
bility of F and G. Our approach here is different and the results are obtained through
other hypotheses.
The main result is summarized by the following theorem.
Theorem 11.3.2. Let U and Z be Banach spaces. Assume u0 is a local minimum of
F(u) subject to G(u) = θ , where F : U → R is a Gâteaux differentiable functional
and G : U → Z is a Fréchet differentiable transformation such that G (u0 ) maps U
onto Z. Finally, assume there exist α > 0 and K > 0 such that if ϕ U < α , then
G (u0 + ϕ ) − G(u0 ) ≤ K ϕ U .
that is,
ϕ , F (u0 )U + G (u0 )ϕ , z∗0 Z = 0, ∀ϕ ∈ U.
Proof. First observe that there is no loss of generality in assuming 0 < α < 1. Also
from the generalized mean value inequality and our hypothesis, if ϕ U < α , then
≤ K sup { hϕ U } ϕ U ≤ K ϕ U2 . (11.42)
h∈[0,1]
that is,
A(ū) = G (u0 ) · u,
where ū = {u + v | v ∈ L0 }.
Since G (u0 ) is onto, so is A, so that by the inverse mapping theorem A has a
continuous inverse A−1 .
Let ϕ ∈ U be such that G (u0 ) · ϕ = θ . For a given t such that 0 < |t| < 1+ αϕ U ,
let ψ0 ∈ U be such that
H(t ϕ )
G (u0 ) · ψ0 + 2 = θ ,
t
Observe that from (11.42),
H(t ϕ ) ≤ Kt 2 ϕ U2 ,
Since from the hypotheses G (u) is continuous at u0 , we may choose 0 < δ < α
such that if v U < δ then
294 11 Constrained Variational Optimization
H(t ϕ + t 2ψ )
G (u0 ) · ψ + = θ.
t2
Define
−1 H(t ϕ + t 2(ψ0 − g0))
L1 = A G (u0 ) · (ψ0 − g0 ) + ,
t2
so that
H(t ϕ + t 2(ψ0 − g0))
L1 = A−1 [A(ψ0 − g0)] + A−1
t2
= ψ0 − g0 + w1
= ψ0 + w1
= {ψ0 + w1 + v | v ∈ L0 }.
that is,
H(t ϕ + t 2 (ψ0 − g0 ))
A(w1 ) = ,
t2
so that
H(t ϕ + t 2(ψ0 − g0))
G (u0 ) · w1 = .
t2
Select g1 ∈ L1 such that
g1 − g0 U ≤ 2 L1 − L0 .
So we have that
−1 H(t ϕ ) H(t ϕ + t 2(ψ0 − g0))
L1 = A − 2 + . (11.43)
t t2
11.3 The Lagrange Multiplier Theorem 295
However
< ε t 2 ψ0 U . (11.45)
Observe that
−1 H(t ϕ + t 2(ψ0 − gn−1))
Ln = A G (u0 ) · (ψ0 − gn−1) + + Ln−1
t2
H(t ϕ + t 2(ψ0 − gn−1))
= A−1 A(ψ0 − gn−1) + A−1 + gn−1
t2
H(t ϕ + t 2 (ψ0 − gn−1))
= ψ0 − gn−1 + A−1 + gn−1
t2
296 11 Constrained Variational Optimization
H(t ϕ + t 2(ψ0 − gn−1))
−1
= ψ0 + A
t2
= {ψ0 + wn + v | v ∈ L0 }.
that is,
H(t ϕ + t 2 (ψ0 − gn−1))
A(wn ) = ,
t2
so that
H(t ϕ + t 2(ψ0 − gn−1))
G (u0 ) · wn = .
t2
Choose gn ∈ Ln such that
gn − gn−1 U ≤ 2 Ln − Ln−1 .
Thus
−1 H(t ϕ + t 2 (ψ0 − gn−1))
Ln = A G (u0 ) · (ψ0 − gn−1) + + G (u0 ) · gn−1 .
t2
By analogy
H(t ϕ + t 2(ψ0 − gn−2))
Ln−1 = A−1 G (u0 ) · (ψ0 − gn−2) + + G
(u 0 ) · g n−2 .
t2
Observe that
× t 2(−gn−1 + gn−2) U
< ε t 2 gn−1 − gn−2 U .
A−1
Ln − Ln−1 ≤ H(t ϕ + t 2 (ψ0 − gn−1)) − H(t ϕ + t 2(ψ0 − gn−2))
t2
< ε A−1 gn−1 − gn−2 U
< (r/4) gn−1 − gn−2 U
1
< gn−1 − gn−2 U . (11.48)
4
Thus,
1
gn − gn−1 U ≤ 2 Ln − Ln−1 < gn−1 − gn−2 U .
2
Finally
1
gn − gn−1 U < gn−1 − gn−2 U , ∀n ∈ N,
2
so that {gn } is a Cauchy sequence, and since U is a Banach space there exists g ∈ U
such that
gn → g, in norm, as n → ∞.
Hence
Ln → L = ḡ, in norm, as n → ∞,
so that
H(t ϕ + t 2 (ψ0 − g))
Ln − Ln−1 → L − L = θ = A−1 G (u0 ) · (ψ0 − g) + .
t2
298 11 Constrained Variational Optimization
H(t ϕ + t 2(ψ˜0 ))
G (u0 ) · ψ˜0 + =θ
t2
Clarifying the dependence on t we denote ψ˜0 = ψ̃0 (t) where as above mentioned,
t ∈ R is such that
δ
0 < |t| < .
2(1 + ϕ U + K1 )
Therefore
G(u0 + t ϕ + t 2 ψ̃0 (t)) = θ .
Observe also that t 2 ψ̃0 (t) U = t 2 (ψ0 (t) − g) U ≤ t 2 (K1 + r) ≤ t 2 (K1 + 1) so
that t 2 ψ̃0 (t) → θ as t → 0. Thus, by defining t 2 ψ̃0 (t)|t=0 = θ (observe that in prin-
ciple such a function would not be defined at t = 0), we obtain
2
d(t 2 ψ̃0 (t)) t ψ̃0 (t) − θ
|t=0 = lim = θ,
dt t→0 t
considering that
t ψ̃0 (t) U ≤ |t|(K1 + 1) → 0, as t → 0.
Finally, defining
φ (t) = F(u0 + t ϕ + t 2 ψ̃0 (t)),
from the hypotheses, we have that there exists a suitable t˜2 > 0 such that
φ (0) = δ F(u0 , ϕ ) = 0,
that is,
ϕ , F (u0 )U = 0, ∀ϕ such that G (u0 ) · ϕ = θ .
In the next lines as usual N[G (u0 )] and R[G (u0 )] denote the null space and the
range of G (u0 ), respectively. Thus F (u0 ) is orthogonal to the null space of G (u0 ),
which we denote by
F (u0 ) ⊥ N[G (u0 )].
Since R[G (u0 )] is closed, we get F (u0 ) ∈ R([G (u0 )]∗ ), that is, there exists z∗0 ∈ Z ∗
such that
F (u0 ) = [G (u0 )]∗ (−z∗0 ).
The proof is complete.
11.4 Some Examples Concerning Inequality Constraints 299
In this section we assume the hypotheses of the last theorem for F and G spec-
ified below. As an application of this same result, consider the problem of locally
minimizing F(u) subject to G1 (u) = θ and G2 (u) ≤ θ , where F : U → R, U being a
function Banach space, G1 : U → [L p (Ω )]m1 , G2 : U → [L p (Ω )]m2 where 1 < p < ∞,
and Ω is an appropriate subset of RN . We refer to the simpler case in which the par-
tial order in [L p (Ω )]m2 is defined by u = {ui } ≥ θ if and only if ui ∈ L p (Ω ) and
ui (x) ≥ 0 a.e. in Ω , ∀i ∈ {1, . . . , m2 }.
Observe that defining
F̃(u, v) = F(u),
G(u, v) = {(G1 )i (u)}m1 ×1 , {(G2 )i (u) + v2i }m2 ×1
it is clear that (locally) minimizing F̃(u, v) subject to G(u, v) = (θ , θ ) is equivalent
to the original problem. We clarify the domain of F̃ is denoted by U × Y , where
Remark 11.4.1. For the case U = Rn and Rmk replacing [L p (Ω )]mk , for k ∈ {1, 2},
the conditions (z∗2 )i vi = θ mean that for the constraints not active (e.g., vi = 0)
the corresponding coordinate (z∗2 )i of the Lagrange multiplier is 0. If vi = 0, then
(G2 )i (u0 ) = 0, so that in any case (z∗2 )i (G2 )i (u0 ) = 0.
300 11 Constrained Variational Optimization
Summarizing, for this last mentioned case, we have obtained the standard
necessary optimality conditions: (z∗2 )i ≥ 0, and (z∗2 )i (G2 )i (u0 ) = 0, ∀i ∈ {1, . . . , m2 }.
In this section we develop more rigorous results concerning the Lagrange multi-
plier theorem for the case involving equalities and inequalities.
G1 (u0 ) · ϕ0 = θ
and
G2 (u0 ) · ϕ0 < θ .
Under such hypotheses, there exists a Lagrange multiplier z∗0 = (z∗1 , z∗2 ) ∈ Z1∗ × Z2∗
such that
F (u0 ) + [G1 (u0 )]∗ (z∗1 ) + [G2(u0 )]∗ (z∗2 ) = θ ,
z∗2 ≥ θ ∗ ,
and
G2 (u0 ), z∗2 Z2 = 0.
G1 (u0 ) · ϕ = θ
and
G2 (u0 ) · ϕ = v − λ G2(u0 ),
for some v ≤ θ and λ ≥ 0.
Select α ∈ (0, 1) and define
ϕα = αϕ0 + (1 − α )ϕ .
11.5 The Lagrange Multiplier Theorem for Equality and Inequality Constraints 301
Observe that G1 (u0 ) = θ and G1 (u0 ) · ϕα = θ so that as in the proof of the La-
grange multiplier Theorem 11.3.2 we may find K1 > 0, ε > 0 and ψ0 (t) such that
and
ψ0 (t) U < K1 , ∀|t| < ε .
Observe that
G2 (u0 ) · ϕα
= α G2 (u0 ) · ϕ0 + (1 − α )G2(u0 ) · ϕ
= α G2 (u0 ) · ϕ0 + (1 − α )(v − λ G2(u0 ))
= α G2 (u0 ) · ϕ0 + (1 − α )v − (1 − α )λ G2(u0 ))
= v0 − λ0G2 (u0 ), (11.51)
where
λ0 = (1 − α )λ ,
and
v0 = α G2 (u0 ) · ϕ0 + (1 − α )v < θ .
Hence, for t > 0,
where
r(t)
lim = 0.
t→0+ t
Therefore from (11.51) we obtain
where
r1 (t)
lim = 0.
t→0+ t
Observe that there exists ε1 > 0 such that if 0 < t < ε1 < ε , then
r1 (t)
v0 + < θ,
t
and
G2 (u0 ) − t λ0G2 (u0 ) = (1 − t λ0)G2 (u0 ) ≤ θ .
Hence
G2 (u0 + t ϕα + t 2ψ0 (t)) < θ , if 0 < t < ε1 .
From this there exists 0 < ε2 < ε1 such that
302 11 Constrained Variational Optimization
where
|r2 (t)|
lim = 0.
t→0+ t
Dividing the last inequality by t > 0 we get
Letting t → 0+ we obtain
Letting α → 0+ , we get
ϕ , F (u0 )U ≥ 0,
if
G1 (u0 ) · ϕ = θ ,
and
G2 (u0 ) · ϕ = v − λ G2(u0 ),
for some v ≤ θ and λ ≥ 0. Define
ϕ , F (u0 )U ≥ 0,
so that
ϕ , F (u0 )U = 0,
and r = 0. Hence (0, θ , θ ) is on the boundary of A. Therefore, by the Hahn–Banach
theorem, geometric form, there exists
such that
(β , z∗1 , z∗2 ) = (0, θ , θ )
and
ϕ , F (u0 )U + G1 (u0 )ϕ , z∗1 Z1 + G2 (u0 ) · ϕ , z∗2 Z2 ≥ 0, ∀ϕ ∈ U,
that is, since obviously such an inequality is valid also for −ϕ , ∀ϕ ∈ U, we obtain
ϕ , F (u0 )U + ϕ , [G1 (u0 )]∗ (z∗1 )U + ϕ , [G2 (u0 )]∗ (z∗2 )U = 0, ∀ϕ ∈ U,
so that
F (u0 ) + [G1 (u0 )]∗ (z∗1 ) + [G2(u0 )]∗ (z∗2 ) = θ .
The proof is complete.
304 11 Constrained Variational Optimization
and also suppose that (G2 )i (u0 ) < θ , if i ∈ A. Moreover, suppose {G1 (u0 ), {(G2 )i
(u0 )}i∈A } is onto and that there exist α > 0, K > 0 such that if ϕ U < α , then
where
G̃(u) = {G1 (u), {(G2 )i (u)}i∈A }.
Finally, suppose there exists ϕ0 ∈ U such that
G1 (u0 ) · ϕ0 = θ
and
G2 (u0 ) · ϕ0 < θ .
Under such hypotheses, there exists a Lagrange multiplier z∗0 = (z∗1 , z∗2 ) ∈ Z1∗ × (Z2∗ )k
such that
F (u0 ) + [G1 (u0 )]∗ (z∗1 ) + [G2(u0 )]∗ (z∗2 ) = θ ,
z∗2 ≥ (θ ∗ , . . . , θ ∗ ) ≡ θk∗ ,
and
(G2 )i (u0 ), (z∗2 )i Z = 0, ∀i ∈ {1, . . . , k},
(z∗2 )i = θ ∗ , if i ∈ A,
Moreover, defining
L(u, z∗1 , z∗2 ) = F(u) + G1(u), z∗1 Z1 + G2 (u), z∗2 Z2 ,
we have that
δuu
2
L(u0 , z∗1 , z∗2 ; ϕ ) ≥ 0, ∀ϕ ∈ V0 ,
where
11.6 Second-Order Necessary Conditions 305
From the last Theorem 11.1 for such an optimization problem there exists a La-
grange multiplier (z∗1 , {(z∗2 )i∈A }) such that (z∗2 )i ≥ θ ∗ , ∀i ∈ A, and
F (u0 ) + [G1(u0 )]∗ (z∗1 ) + ∑ [(G2 )i (u0 )]∗ ((z∗2 )i ) = θ . (11.55)
i∈A
z∗2 ≥ θk∗
and
(G2 )i (u0 ), (z∗2 )i Z = 0, ∀i ∈ {1, . . . , k}.
Let ϕ ∈ V0 , that is, ϕ ∈ U,
G1 (u0 )ϕ = θ
and
(G2 )i (u0 ) · ϕ = θ , ∀i ∈ A.
Recall that G̃(u) = {G1 (u), (G2 )i∈A (u)} and therefore, similarly as in the proof
of the Lagrange multiplier Theorem 11.3.2, we may obtain ψ0 (t), K > 0 and ε > 0
such that
G̃(u0 + t ϕ + t 2 ψ0 (t)) = θ , ∀|t| < ε ,
and
ψ0 (t) ≤ K, ∀|t| < ε .
Also, if i ∈ A, we have that (G2 )i (u0 ) < θ , so that
(G2 )i (u0 + t ϕ + t 2 ψ0 (t)) = (G2 )i (u0 ) + Gi (u0 ) · (t ϕ + t 2ψ0 (t)) + r(t),
where
r(t)
lim = 0,
t→0 t
that is,
where
r1 (t)
= 0,
lim
t t→0
so that
Now we recall a classical definition, namely, the Banach fixed theorem also
known as the contraction mapping theorem.
Definition 11.7.1. Let C be a subset of a Banach space U and let T : C → C be an
operator. Thus, T is said to be a contraction mapping if there exists 0 ≤ α < 1 such
that
T (u) − T (v) U ≤ α u − v U , ∀u, v ∈ C.
we have
un → u0 , in norm, as n → +∞.
un+1 = T (un ), ∀n ∈ N.
un+p − un U
= un+p − un+p−1 + un+p−1 − un+p−2 + . . . − un+1 + un+1 − un U
≤ un+p − un+p−1 U + un+p−1 − un+p−2 U + . . . + un+1 − un U
≤ (α n+p−2 + α n+p−3 + . . . + α n−1) u2 − u1 U
≤ α n−1 (α p−1 + α p−2 + . . . + α 0 ) u2 − u1 U
308 11 Constrained Variational Optimization
∞
≤ α n−1 ∑ αk u2 − u1 U
k=0
α n−1
≤ u2 − u1 U (11.57)
1−α
Denoting n + p = m, we obtain
α n−1
um − un U ≤ u2 − u1 U , ∀m > n ∈ N.
1−α
Let ε > 0. Since 0 ≤ α < 1, there exists n0 ∈ N such that if n > n0 then
α n−1
0≤ u2 − u1 U < ε ,
1−α
so that
um − un U < ε , if m > n > n0 .
From this we may infer that {un } is a Cauchy sequence, and since U is a Banach
space, there exists u0 ∈ U such that
un → u0 , in norm, as n → ∞.
Observe that
u0 − T (u0 ) U = u0 − un + un − T (u0 ) U
≤ u0 − un U + un − T (u0 ) U
≤ u0 − un U + α un−1 − u0 U
→ 0, as n → ∞. (11.58)
Thus u0 − T (u0 ) U = 0.
Finally, we prove the uniqueness. Suppose u0 , v0 ∈ C are such that
Hence,
u0 − v0 U = T (u0 ) − T (v0 ) U
≤ α u0 − v0 U . (11.59)
11.8.1 Introduction
In this section we state and prove the implicit function theorem for Banach
spaces. A similar result may be found in Ito and Kunisch [40], page 31.
We emphasize the result found in [40] is more general; however, the proof present
here is almost the same for a simpler situation. The general result found in [40] is
originally from Robinson [53].
Theorem 11.3 (Implicit Function Theorem). Let V,U,W be Banach spaces. Given
a function F̂ : V × U → W , suppose (x0 , u0 ) ∈ V × U is such that F̂(x0 , u0 ) = θ .
Assume F̂ is Fréchet differentiable, F̂x (x0 , u0 ) is continuous, and [F̂x (x0 , u0 )]−1 is a
(single-valued) bounded linear operator so that we denote [F̂x (x0 , u0 )]−1 = ρ > 0.
Under such hypotheses, for each ε > 0, there exist a neighborhood Uε of u0 , a
neighborhood Vε of x0 , and a function x : Uε → Vε such that for each u ∈ Uε , x(u)
is the unique solution of
F̂(x, u) = θ ,
that is,
F̂(x(u), u) = θ .
Moreover, for each u, v ∈ Uε , we have
where K0 = K(ρ + ε ).
Define
and
Select a ball Uε about u0 and a closed ball Vε of radius r > 0 about x0 such that
for each u ∈ Uε and x ∈ Vε we have
so that since 0 < ρδ < 1 we may infer that φu (x) is a contractor. Observe also that
x0 = T −1 (θ ), so that
φu (x0 ) − x0 ≤ ρ h(x0, u) − θ
= ρ F̂(x0 , u) − F̂(x0 , u0 )
≤ (1 − ρδ )r. (11.63)
Therefore φu (x) ∈ Vε , ∀x ∈ Vε so that from this, (11.62) and the Banach fixed
point theorem, φu has a unique fixed point in Vε , which we denote by x(u).
Thus,
x(u) = φu (x(u))
= T −1 (h(x(u), u))
= T −1 (F̂x (x0 , u0 )(x(u) − x0 ) − F̂(x(u), u))
= [F̂x (x0 , u0 )]−1 (F̂x (x0 , u0 )(x(u) − x0 ) − F̂(x(u), u)) + x0
= x(u) − x0 + x0 − [F̂x (x0 , u0 )]−1 (F̂(x(u), u))
= x(u) − [F̂x (x0 , u0 )]−1 (F̂(x(u), u)). (11.65)
From this,
11.8 Sensitivity Analysis 311
F̂(x(u), u) = F̂x (x0 , u0 )[F̂x (x0 , u0 )]−1 (F̂(x(u), u)) = F̂x (x0 , u0 )θ = θ ,
that is,
F̂(x(u), u) = θ .
Also as a consequence of the Banach fixed point theorem, we have that
Now observe that for u, v ∈ Uε , with x = x(v) in the last inequality, we get
However, x(v) = φv (x(v)), so that from this and the last inequality, we obtain
G : V × U → [L p (Ω )]m1
and
H : V × U → [L p (Ω )]m2 ,
(the cases in which the co-domains of G and H are Rm1 and Rm2 , respectively, are
dealt similarly), we redefine the concerned optimization problem, again for a fixed
u ∈ U, by minimizing F(x, u) subject to
and
H(x, u) = θ .
312 11 Constrained Variational Optimization
At this point we assume F(x, u), G̃(x, u, v) = {Gi (x, u) + v2i } ≡ G(u) + v2 (from
now on we use this general notation) and H(x, u) satisfy the hypotheses of the La-
grange multiplier Theorem 11.3.2.
Hence, for the fixed u ∈ U, we assume there exists an optimal x ∈ V which locally
minimizes F(x, u) under the mentioned constraints.
From Theorem 11.3.2 there exist Lagrange multipliers λ1 , λ2 such that denoting
[L p (Ω )]m1 and [L p (Ω )]m2 simply by L p and defining
G(x, u) + v2 = θ , (11.68)
λ1 · v = θ , (11.69)
λ1 ≥ θ , (11.70)
H(x, u) = θ . (11.71)
Clarifying the dependence on u, we denote the solution x, λ1 , λ2 , v by x(u),
λ1 (u), λ2 (u), v(u), respectively. In particular, we assume that for a u0 ∈ U,
x(u0 ), λ1 (u0 ), λ2 (u0 ), v(u0 ) satisfy the hypotheses of the implicit function theo-
rem. Thus, for any u in an appropriate neighborhood of u0 , the corresponding
x(u), λ1 (u), λ2 (u), v(u) are uniquely defined.
We emphasize that from now on the main focus of our analysis is to evaluate vari-
ations of the optimal x(u), λ1 (u), λ2 (u), v(u) with variations of u in a neighborhood
of u0 .
For such an analysis, the main tool is the implicit function theorem and its main
hypothesis is satisfied through the invertibility of the matrix of Fréchet second
derivatives.
Hence, denoting x0 = x(u0 ), (λ1 )0 = λ1 (u0 ), (λ2 )0 = λ2 (u0 ), v0 = v(u0 ), and
A2 = G(x0 , u0 ) + v20
A3 = H(x0 , u0 ),
A4 = (λ1 )0 · v0 ,
we reiterate to assume that
A1 = θ , A2 = θ , A3 = θ , A4 = θ ,
where
A = Fxx (x0 , u0 ) + (λ1 )0 · Gxx (x0 , u0 ) + (λ2 )0 · Hxx (x0 , u0 ).
Moreover, also from the implicit function theorem,
F̃xx
so that
≤ K1 K 2 ϕ 24 ε
→ 0, as ε → 0.
and
F̃x (x(u0 + εϕ ), u0 + εϕ , λ (u0 + εϕ ), v(u0 + εϕ )) = θ .
On the other hand,
F̃(x(u0 ), u0 + εϕ , λ (u0 + εϕ ), v(u0 + εϕ ))
ε
F̃(x(u0 ), u0 , λ (u0 ), v(u0 ))
−
ε
F̃(x(u0 ), u0 + εϕ , λ (u0 + εϕ ), v(u0 + εϕ ))
=
ε
F̃(x(u0 ), u0 + εϕ , λ (u0 ), v(u0 ))
−
ε
λ1 (u0 )v(u0 ) = θ ,
λ1 (u0 )v(u0 + εϕ ) → θ , as ε → 0,
316 11 Constrained Variational Optimization
and
λ1 (u0 )(v(u0 + εϕ )2 − v(u0)2 )
ε
λ1 (u0 )(v(u0 + εϕ ) + v(u0))(v(u0 + εϕ ) − v(u0))
=
ε
λ1 (u0 )(v(u0 + εϕ ) + v(u0)) K ϕ ε
≤
ε
→ 0, as ε → 0. (11.77)
Finally,
λ2 (u0 + εϕ ) − λ2(u0 ), H(x(u0 ), u0 + εϕ )L p
ε
K ε ϕ
≤ H(x(u0 ), u0 + εϕ )
ε
→ 0, as ε → 0.
and
H(x(u0 ), u0 + εϕ ) → θ , as ε → 0.
From these last results, we get
In the last lines we have proven the following corollary of the implicit function
theorem.
Corollary 11.1. Suppose (x0 , u0 , (λ1 )0 , (λ2 )0 , v0 ) is a solution of the system (11.67),
(11.68),(11.69), (11.71), and assume the corresponding hypotheses of the implicit
function theorem are satisfied. Also assume F̃(x, u, λ1 , λ2 , v) is such that the Fréchet
second derivative F̃xx (x, u, λ1 , λ2 ) is continuous in a neighborhood of
we have
12.1 Introduction
The first part of the present work develops a new duality principle applicable to
nonlinear elasticity. The proof of existence of solutions for the model in question
has been obtained in Ciarlet [21]. In earlier results (see [65] for details) the concept
of complementary energy is equivalently developed under the hypothesis of positive
definiteness of the stress tensor at a critical point. In more recent works, Gao [33,
34, 36] applied his triality theory to similar models obtaining duality principles for
more general situations, including the case of negative definite optimal stress tensor.
We emphasize our main objective is to establish a new and different duality
principle which allows the local optimal stress tensor to not be either positive or
negative definite. Such a result is a kind of extension of a more basic one obtained
in Toland [67]. Despite the fact we do not apply it directly, we follow a similar idea.
The optimality conditions are also new. We highlight the basic tools on convex anal-
ysis here used may be found in [25, 54, 67] for example. For related results about
the plate model presented in Ciarlet [22], see Botelho [11, 13].
In a second step, we present other two duality principles which qualitatively agree
with the triality theory proposed by Gao (see again [33, 34], for details).
However, our proofs again are obtained through more traditional tools of convex
analysis. Finally, in the last section, we provide a numerical example in which the
optimal stress field is neither positive nor negative definite.
At this point we start to describe the primal formulation.
Consider Ω ⊂ R3 an open, bounded, connected set, which represents the
reference volume of an elastic solid under the loads f ∈ L2 (Ω ; R3 ) and the boundary
loads fˆ ∈ L2 (Γ ; R3 ), where Γ denotes the boundary of Ω . The field of displace-
ments resulting from the actions of f and fˆ is denoted by u ≡ (u1 , u2 , u3 ) ∈ U,
where u1 , u2 , and u3 denote the displacements relating the directions x, y, and z,
respectively, in the Cartesian system (x, y, z).
Here U is defined by
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 321
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 12,
© Springer International Publishing Switzerland 2014
322 12 Duality Applied to Elasticity
where
u, f L2 (Ω ;R3 ) = fi ui dx.
Ω
(G ◦ Λ )(u)
1 ui, j + u j,i um,i um, j uk,l + ul,k um,k um,l
= Hi jkl + + dx,
2 Ω 2 2 2 2
12.2 The Main Duality Principle 323
where Λ : U → Y × Y is given by
Λ u = {Λ1 u, Λ2 u},
ui, j + u j,i
Λ1 u =
2
and
Λ2 u = {um,i }.
Here
U = {u ∈ W 1,4 (Ω ; R3 ) | u = (u1 , u2 , u3 ) = θ on Γ0 }.
Define (F ◦ Λ2 ) : U → R, (GK ◦ Λ ) : U → R, and (G1 ◦ Λ2) : U → R by
K
(F ◦ Λ2 )(u) = um,i , um,i L2 (Ω ) ,
2
K
GK (Λ u) = GK (Λ1 u, Λ2 u) = G(Λ u) + um,i , um,i L2 (Ω ) ,
4
and
K
(G1 ◦ Λ2 )(u) = um,i , um,i L2 (Ω ) ,
4
respectively.
Also define
C = {u ∈ U | G∗∗
K (Λ u) = GK (Λ u)},
inf {J(u)}
u∈C1
" ∗ ∗ ∗ ∗ ∗ ∗
#
≥ sup inf ∗ F (z ) − G̃K (σ , z , v) − G̃1(σ̃ , σ , z , v) ,
∗
(σ̃ ,σ ,v)∈Ỹ z ∈Y
where Ỹ = A∗ × Y ∗ × Ŷ ∗ , Y = Y ∗ = L2 (Ω ; R3×3 ) ≡ L2 ,
and
3 (z∗ v )2
z∗mi z∗mi
− H i jkl z∗i j z∗kl − ∑
ij mj
W ∗ (z∗ ) = . (12.7)
K m,i=1 K/2
324 12 Duality Applied to Elasticity
C2 = {u ∈ U | {ui, j } ∈ Ŷ ∗ }.
Furthermore,
Also
we define
G̃∗K (σ , z∗ , v) = G∗K (σ + z∗ , Q)
= sup {v1 , σ + z∗ Y + v2 , QY − GK (v1 , v2 )}, (12.9)
(v1 ,v2 )∈Y ×Y
so that in particular,
G̃∗K (σ , z∗ , v) = G∗K (σ + z∗ , Q)
1
= H i jkl (σi j + z∗i j )(σkl + z∗kl ) dx
2 Ω
1 K
+ (σi j + z∗i j )vmi vm j dx + vmi , vmi L2 (Ω )
2 Ω 4
if (σ̃ , σ , v, z∗ ) ∈ B∗ . We emphasize to denote
⎧ ⎫
⎨ σ11 + z∗11 + K/2 σ12 + z∗12 σ13 + z∗13 ⎬
σK (σ , z∗ ) = σ21 + z∗21 σ22 + z∗22 + K/2 σ23 + z∗23 , (12.10)
⎩ ⎭
σ31 + z∗31 σ32 + z∗32 σ33 + z∗33 + K/2
and
{H i jkl } = {Hi jkl }−1 .
12.2 The Main Duality Principle 325
Moreover,
1 3
= ∑ σ̃mi − σmi − Qmi 22
K m,i=1
1 3
= ∑ σ̃mi − σmi − (σi j + z∗i j )vm j − (K/2)vmi 22 .
K m,i=1
Finally, if there exists a point (u0 , σ̃0 , σ0 , v0 , z∗0 ) ∈ C1 × ((Ỹ × Y ∗ ) ∩ B∗ ), such that
δ u0i , −σ̃0i j, j − fi L2 (Ω ) − u0i ( fˆi − σ̃0i j n j ) dΓ
Γ1
∗ ∗ ∗ ∗ ∗ ∗
+F (z0 ) − G̃K (σ0 , z0 , v0 ) − G̃1(σ̃0 , σ0 , z0 , v0 ) = θ , (12.11)
we have
∂ gK (y)
y∗ = . (12.15)
∂y
Thus
1
y∗1i j = σi j = Hi jkl y1kl + y2mk y2ml (12.16)
2
and
1
y∗2mi = Qmi = Hi jkl y1kl + y2ok y2ol y2m j + (K/2)y2mi (12.17)
2
so that
and also
1
y1i j = H i jkl σkl − y2mi y2m j . (12.21)
2
Finally
1 1
g∗KL (σ , Q) = H i jkl σi j σkl + σ̄iKj Qmi Qm j . (12.22)
2 2
Now we will prove that g∗KL (y∗ ) = g∗K (y∗ ) if σK (y∗1 ) = σK (σ ) is positive definite.
First observe that
g∗K (y∗ ) = sup {y1 , σ R9 + y2 , QR9 − gK (y)}
y∈R18
= sup y1 , σ R9 + y2 , QR9
y∈R18
1 1 1
− Hi jkl y1i j + y2mi y2m j y1kl + y2mk y2ml
2 4 2
K
− y2mi y2mi
4
12.2 The Main Duality Principle 327
1
= sup ȳ1i j − y2mi y2m j , σi j R + y2 , QR9
(ȳ1 ,y2 )∈R9 ×R9 2
1 K
− Hi jkl [ȳ1i j ][ȳ1kl ] − y2mi y2mi .
2 4
and
1 K
sup − y2mi y2m j , σi j R + y2 , QR9 − y2mi y2mi
y2 ∈R9 2 4
1
= σ Kij Qmi Qm j (12.24)
2
if σK (y∗1 ) = σK (σ ) is positive definite.
Now observe that using the relation
we have
G̃∗K (σ , z∗ , v) = G∗K (σ + z∗ , Q)
= g∗KL (σ + z∗ , Q) dx, (12.25)
Ω
if σK (σ + z∗ ) is positive definite.
Also, considering the concerned symmetries, we may write
∀u ∈ U, z∗ ∈ Y ∗ , (σ̃ , σ , v) ∈ Ỹ , so that
G̃∗K (σ , z∗ , v) + G̃∗1(σ̃ , σ , z∗ , v)
≥ Λ2 u, z∗ L2 + Λ1 u, σ̃ L2
−GK (Λ u) − G1(Λ2 u)
= Λ2 u, z∗ L2 + u, f L2 (Ω ;R3 )
+ fˆi ui dΓ − GK (Λ u) − G1 (Λ2 u), (12.27)
Γ1
328 12 Duality Applied to Elasticity
∀u ∈ C1 , z∗ ∈ Y ∗ , (σ̃ , σ , v) ∈ Ỹ . Hence
∀u ∈ C1 , (σ̃ , σ , v) ∈ Ỹ .
Therefore,
∀u ∈ C1 , (σ̃ , σ , v) ∈ Ỹ . Finally,
Now suppose there exists a point (u0 , σ̃0 , σ0 , z∗0 , v0 ) ∈ C1 × ((Ỹ ×Y ∗ )∩B∗ ), such that
δ u0i , −σ̃0i j, j − fi L2 (Ω ) − u0i ( fˆi − σ̃0i j n j ) dΓ
Γ1
+F ∗ (z∗0 ) − G̃∗K (σ0 , z∗0 , v0 ) − G̃∗1(σ̃0 , σ0 , z∗0 , v0 ) = θ , (12.33)
12.2 The Main Duality Principle 329
that is,
δ u0i , −σ̃0i j, j − fi L2 (Ω ) − u0i ( fˆi − σ̃0i j n j ) dΓ
Γ1
1
+F ∗ (z∗0 ) − H̄i jkl (σ0i j + z∗0i j )(σ0kl + z∗0kl ) dx
2 Ω
1 K
− (σ0i j + z∗0i j )v0mi v0m j dx − v mi , v0mi L2 (Ω )
2 Ω 4 0
3
1
− ∑ σ̃0mi − σ0mi − (σ0i j + z∗0i j )v0m j − K/2v0mi 22 = θ.
m,i=1 K
From this and recalling that σ̃i j = σ̃ ji , so that we may use the replacement
σ̃i j + σ̃ ji
σ̃i j = = σ̃ ji
2
(observe that a similar remark is valid for σ0i j + z∗0i j ), the variation in σ gives us
so that
{v0i j } = {u0i, j }, in Ω . (12.37)
From this and (12.35) we get
u0k,l + u0l,k u0m,k u0m,l
σ0i j + z∗0i j = Hi jkl + . (12.38)
2 2
σ̃0i j, j + fi = 0, in Ω , (12.40)
u0 = θ on Γ0 ,
330 12 Duality Applied to Elasticity
and
σ̃0i j n j = fˆi on Γ1 ,
where from (12.34), (12.37), and (12.39), we have
u0k,l + u0l,k u0m,k u0m,l
σ̃0i j = Hi jkl +
2 2
u0k,l + u0l,k u0 p,k u0 p,l
+Hm jkl + u0i,m . (12.41)
2 2
From the hypothesis indicated in (12.6), the extremal relation through which z∗0
is obtained is in fact a global one.
From this, (12.2) and (12.42), the proof is complete.
Remark 12.2.2. About the last theorem, there is no duality gap between the primal
and dual problems, if K is big enough so that for the optimal dual point, σK (σ0 , z∗0 )
is positive definite in Ω , where
⎧ ⎫
⎨ σ11 + z∗11 + K/2 σ12 + z∗12 σ13 + z∗13 ⎬
σK (σ , z∗ ) = σ21 + z∗21 σ22 + z∗22 + K/2 σ23 + z∗23 , (12.43)
⎩ ⎭
σ31 + z∗31 σ32 + z∗32 σ33 + z∗33 + K/2
and
u0k,l + u0l,k u0m,k u0m,l
σ0i j + z∗0i j = Hi jkl + , (12.44)
2 2
and, at the same time, K is small enough so that for the fixed point {v0m j } = {u0m, j }
the quadratic form (in z∗ ) W ∗ (z∗ ) is also positive definite in Ω , where
∗
3 (z v m j )
z∗mi z∗mi
2
W ∗ (z∗ ) = − H̄i jkl z∗i j z∗kl − ∑
ij 0
. (12.45)
K m,i=1 K/2
C = {u ∈ U | G∗∗
K (Λ u) = GK (Λ u)}
= {u ∈ U | σK (σ (u), θ ) is positive definite in Ω }, (12.46)
12.3 Other Duality Principles 331
At this point we present another main result, which is summarized by the follow-
ing theorem.
(G ◦ Λ )(u)
1 ui, j + u j,i um,i um, j uk,l + ul,k um,k um,l
= Hi jkl + + dx,
2 Ω 2 2 2 2
where
Define J : U → R by
J(u) = G(Λ u) − u, f L2 (Ω ;R3 ) − fˆi ui dΓ . (12.49)
Γ1
Also define
JK : U × Y → R
by
K
JK (u, p) = G(Λ u + p) + Kp, pL2 − u, f L2 (Ω ;R3 ) − fˆi ui dΓ − p, pL2 ,
Γ1 2
and assume that K > 0 is sufficiently big so that JK (u, p) is bounded below.
Also define
2
∂ G∗ (σ )
JK∗ (σ , u) = ∗
Ff (σ ) − G (σ ) + K Λ u − + 1 σ , σ 2 , (12.50)
∂ σ L2 2K L
332 12 Duality Applied to Elasticity
where
Proof. Define
G1 (u, p) = G(Λ u + p) + Kp, pL2 ,
and
K
G2 (u, p) = u, f L2 (Ω ;R3 ) + fˆi ui dΓ + p, pL2 .
Γ1 2
12.3 Other Duality Principles 333
Thus,
−G2 (u, p) ≥ −G1 (u, p) + αK , ∀u ∈ U, p ∈ Y,
so that
Hence,
∂ G(Λ u + p)
σ= ,
∂v
we get
∂ G∗ (σ )
p +Λu = ,
∂σ
that is,
∂ G∗ (σ )
p= − Λ u,
∂σ
and
G∗ (σ ) = Λ u + p, σ L2 − G(Λ u + p).
Hence
∗ 2
∂ G (σ )
Λ u + p, σ L2 − G1 (u, p) = G∗ (σ ) − K
∂σ − Λ u .
2
L
∀σ ∈ Y ∗ , u ∈ U.
Thus,
Defining
∂ G(Λ u0 + p0 )
σ0 = , (12.57)
∂v
since for the extremal point, we have
δu G(Λ u + p0) − u, f L2 (Ω ;R3 ) − fi ui dΓ |u=u0 = θ ,
ˆ
Γ1
that is,
At this point we develop a reasoning similarly to the lines above but now for the
specific case of a neighborhood around the local optimal point. We repeat some
analogous details for the sake of clarity.
12.3 Other Duality Principles 335
From above,
G1 (u, p) = G(Λ u + p) + Kp, pL2 ,
and
K
G2 (u, p) = u, f L2 (Ω ;R3 ) + fˆi ui dΓ + p, pL2 .
Γ1 2
Observe that α = inf(u,p)∈V0 {JK (u, p)} ∈ R is such that
Thus,
−G2 (u, p) ≥ −G1 (u, p) + α , ∀(u, p) ∈ V0 ,
so that
Hence,
∂ G(Λ u + p)
σ= ,
∂v
that is,
∂ G∗ (σ )
p +Λu = ,
∂σ
we get
∂ G∗ (σ )
p= − Λ u,
∂σ
and
G∗ (σ ) = Λ u + p, σ L2 − G(Λ u + p).
Hence
∗ 2
∂ G (σ )
Λ u + p, σ L2 − G1 (u, p) = G∗ (σ ) − K
∂σ − Λ u .
2 (12.63)
L
336 12 Duality Applied to Elasticity
Finally, since
1 ∂ G(Λ u0 + p0 )
p0 = − , (12.67)
K ∂p
we get
1
p0 Y ≈ O ,
K
so that from this, (12.61), and (12.65), we may finally write
Our final result is summarized by the next theorem. It refers to a duality principle
for the case of a local maximum for the primal formulation.
Theorem 12.3.3. Let Ω ⊂ R3 be an open, bounded, connected set with a regu-
lar boundary denoted by Γ = Γ0 ∪ Γ1 , where Γ0 ∩ Γ1 = 0.
/ Consider the functional
(G ◦ Λ ) : U → R expressed by
(G ◦ Λ )(u)
1 ui, j + u j,i um,i um, j uk,l + ul,k um,k um,l
= Hi jkl + + dx,
2 Ω 2 2 2 2
where
Define J : U → R by
J(u) = G(Λ u) − u, f L2 (Ω ;R3 ) − fˆi ui dΓ . (12.70)
Γ1
∂ G(Λ u0 )
σ0 = (12.71)
∂v
is negative definite.
Also define
J ∗ (σ ) = Ff (σ ) − G∗ (σ ), (12.72)
where
∀u ∈ V0 .
Hence,
−u, f L2 (Ω ;R3 ) − fˆi ui dΓ ≤ −G(Λ u) + α , ∀u ∈ V0 ,
Γ1
so that
Λ u, σ L2 − u, f L2 (Ω ;R3 ) − fˆi ui dΓ
Γ1
≤ Λ u, σ L2 − G(Λ u) + α , ∀u ∈ V0 , σ ∈ Y ∗ . (12.75)
Therefore,
sup Λ u, σ L2 − u, f L2 (Ω ;R3 ) − fˆi ui dΓ
u∈V0 Γ1
We highlight that σ0 ∈ Ṽ0 , and from the generalized inverse function theorem,
any σ in an appropriate neighborhood of σ0 also belongs to Ṽ0 (we do not provide
the details here).
From this and (12.76), we get
where
U = {u ∈ W 1,4 ([0, 1]) | u(0) = u(1) = 0} = W01,4 ([0, 1]),
H = 105
P = −1000
where the units refer to the international system. The condition indicated in (12.45)
here stands for W ∗ (z∗ ) to be positive definite in a critical point u0 ∈ U, where
x 10−3
0
−0.2
−0.4
−0.6
−0.8
−1
−1.2
−1.4
0 0.2 0.4 0.6 0.8 1
x 10−3
5
4
3
2
1
0
−1
−2
−3
−4
−5
0 0.2 0.4 0.6 0.8 1
Fig. 12.2 The solution u0 (x) through the primal formulation
C2 = {u ∈ U | ux ∈ Ŷ ∗ }, where
1 1
H K
GK (ux ) = (ux + u2x /2)2 dx + u2x dx,
2 0 4 0
and
Ŷ ∗ = {v ∈ L2 ([0, 1]) | W ∗ (z∗ ) is positive definite in [0, 1]}.
In fact, plotting the function F(x) = H(x + x2 /2)2 /2, we may observe that inside
the set [−0.5, 0.5] there is a local minimum, that is, in a close set, the Legendre
necessary condition for a local minimum is satisfied. Please see Fig. 12.3.
We emphasize on the concerned sets there is no duality gap between the primal
and dual formulations. Also, from the graphic of u0 (x), it is clear that the stress
H(u0 + 1/2(u0)2 )
is not exclusively positive or negative in [0, 1].
12.5 Conclusion 341
x 104
5
4.5
4
3.5
3
2.5
2
1.5
1
0.5
0
−0.8 −0.6 −0.4 −0.2 0 0.2 0.4 0.6
12.5 Conclusion
13.1 Introduction
In the present work we develop dual variational formulations for the Kirchhoff–
Love thin plate model. Earlier results establish the complementary energy under
the hypothesis of positive definiteness of the membrane force tensor at a critical
point (please see [30–32, 36, 65] for details). In more recent works Gao has applied
his triality theory to models in elasticity (see [33–35] for details) obtaining duality
principles which allow the optimal stress tensor to be negative definite. Here for the
present case we have obtained a dual variational formulation which allows the global
optimal point in question to be not only positive definite (for related results see
Botelho [11, 13]) but also not necessarily negative definite. The approach developed
also includes sufficient conditions of optimality for the primal problem. Moreover,
a numerical example concerning the main duality principle application is presented
in the last section.
It is worth mentioning that the standard tools of convex analysis used in this text
may be found in [13, 25, 54], for example. Another relating result may be found
in [14].
At this point we start to describe the primal formulation.
Let Ω ⊂ R2 be an open, bounded, connected set which represents the middle
surface of a plate of thickness h. The boundary of Ω , which is assumed to be regular,
is denoted by ∂ Ω . The vectorial basis related to the Cartesian system {x1 , x2 , x3 } is
denoted by (aα , a3 ), where α = 1, 2 (in general Greek indices stand for 1 or 2)
and where a3 is the vector normal to Ω , whereas a1 and a2 are orthogonal vectors
parallel to Ω . Also, n is the outward normal to the plate surface.
The displacements will be denoted by
û = {ûα , û3 } = ûα aα + û3 a3 .
The Kirchhoff–Love relations are
ûα (x1 , x2 , x3 ) = uα (x1 , x2 ) − x3w(x1 , x2 ),α and û3 (x1 , x2 , x3 ) = w(x1 , x2 ).
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 343
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 13,
© Springer International Publishing Switzerland 2014
344 13 Duality Applied to a Plate Model
Finally, we also emphasize from now on, as their meaning is clear, we may denote
L2 (Ω ) and L2 (Ω ; R2×2 ) simply by L2 and the respective norms by · 2. Moreover
derivatives are always understood in the distributional sense, θ denotes the zero vec-
tor in appropriate Banach spaces, and the following and relating notations are used:
∂ 2w
w,αβ = ,
∂ xα ∂ xβ
∂ uα
uα ,β = ,
∂ xβ
∂ Nαβ
Nαβ ,1 = ,
∂ x1
and
∂ Nαβ
Nαβ ,2 = .
∂ x2
where
1
γαβ (u) = Λ1αβ (u) + Λ2α (u)Λ2β (u),
2
uα ,β + uβ ,α
{Λ1αβ (u)} = ,
2
{Λ2α (u)} = {w,α },
{καβ (u)} = {−Λ3αβ (u)} = {−w,αβ },
346 13 Duality Applied to a Plate Model
Also,
we have that
J(u0 ) = min{J(u)}
u∈U
" #
= sup ∗inf ∗ F̃ ∗ (z∗ ) − G∗ (v∗ , z∗ )
v∗ ∈A∗ z ∈Y0
= F̃ ∗ (z∗0 ) − G∗ (v∗0 , z∗0 ). (13.11)
∀u ∈ U, v∗ ∈ A∗ , z∗ ∈ Y0∗ .
Thus,
∀u ∈ U, v∗ ∈ A∗ . Hence,
∗ ∗ ∗ ∗ ∗
inf {J(u)} ≥ sup inf
∗ ∗
{ F̃ (z ) − G (v , z )} . (13.15)
u∈U v∗ ∈A∗ z ∈Y0
∂ G∗ (v∗0 , z∗0 )
− ((z0 )∗11 )xx /K − = θ, (13.18)
∂ z∗11
∂ G∗ (v∗0 , z∗0 )
− ((z0 )∗22 )yy /K − = θ, (13.19)
∂ z∗22
that is, from this and (13.16) we get
F̃ ∗ (z∗0 ) = (z0 )∗11 , (w0 )xx L2 (Ω ) + (z0 )∗22 , (w0 )yy L2 (Ω ) − F(Λ2 u0 ). (13.20)
13.3 Another Duality Principle 349
Λ ∗ v∗0 − P̃ = 0, (13.21)
To complete the proof just observe that from the condition indicated in (13.10), the
extremal relations (13.18) and (13.19) refer to a global optimization (in z∗ , for a
fixed v∗0 ), so that the infimum indicated in the dual formulation is attained for the z∗0
in question.
From this and (13.22), the proof is complete.
where
Λ (u) = {γ (u), κ (u)},
uα ,β + uβ ,α w,α w,β
γαβ (u) = + ,
2 2
καβ (u) = −w,αβ ,
and J : U → R by
J(u) = (G ◦ Λ )(u) − F(u).
350 13 Duality Applied to a Plate Model
where
1
G1 (γ (u)) = H γ (u)γλ μ (u) dx,
2 Ω αβ λ μ αβ
1
G2 (u) = h κ (u)κλ μ (u) dx − F(u),
2 Ω αβ λ μ αβ
so that
J(u) = G1 (γ (u)) + G2 (u).
Moreover,
and
if
N = {Nαβ } ∈ A∗ = A1 ∩ A2 ,
Also,
A1 = {N ∈ Y ∗ | J(w)
˜ > 0, ∀w ∈ W02,2 (Ω ) such that w = θ },
1 1
˜
J(w) = hαβ λ μ w,αβ w,λ μ dx + Nαβ w,α w,β dx,
2 Ω 2 Ω
and
A2 = {N ∈ Y ∗ | Nαβ ,β + Pα = 0 in Ω }.
Finally, if there exists u0 ∈ U such that δ J(u0 ) = θ and N0 = {Nαβ (u0 )} ∈ A1 ,
where Nαβ (u0 ) = Hαβ λ μ γλ μ (u0 ), then
J(u0 ) = min{J(u)}
u∈U
= max∗ {−G∗1 (N) − G̃∗2 (−N)}
N∈A
= −G∗1 (N0 ) − G̃∗2 (−N0 )}. (13.26)
13.3 Another Duality Principle 351
Proof. Clearly
Hence,
inf {J(u)} ≥ sup {−G∗1 (N) − G̃∗2 (−N)}. (13.28)
u∈U N∈A∗
Now suppose there exists u0 ∈ U such that δ J(u0 ) = θ and N0 = {Nαβ (u0 )} ∈ A1 .
First, note that from δ J(u0 ) = θ , the following extremal equation is satisfied:
(N0 )αβ ,β + Pα = 0 in Ω ,
Hence, also from the equation δ J(u0 ) = θ and (13.30), we may get ŵ = w0 , so
that from this and (13.29), we obtain
we get
G∗1 (N0 ) = γ (u0 ), N0 L2 − G1 (γ (u0 )),
352 13 Duality Applied to a Plate Model
so that
From this and (13.28) and also from the fact that N0 ∈ A∗ , the proof is complete.
where
 = A1 ∩ A2 ∩ A3 ∩ A4 ,
A1 = {N ∈ Y ∗ | J(w)
˜ > 0, ∀w ∈ W02,2 (Ω ) such that w = θ },
where
1 1
˜
J(w) = hαβ λ μ w,αβ w,λ μ dx + Nαβ w,α w,β dx,
2 Ω 2 Ω
and
A2 = {(M, N, u) ∈ Y ∗ × Y ∗ × U | Nαβ ,β + Pα = 0 in Ω }.
Finally, we could suggest as a possible approximate dual formulation the problem
of maximizing −JK∗ (M, N, u) on A1 ∩ A2 ∩ A3 , where K > 0 and
1 1
JK∗ (M, N, u) = H αβ λ μ Nαβ Nλ μ dx + hαβ λ μ Mαβ Mλ μ dx
2 Ω 2 Ω
1
+ N w,α w,β dx
2 Ω αβ
K 2
2 α ,∑
+ Mαβ − hαβ λ μ (−w,λ μ ) 22 . (13.34)
β =1
13.4 An Algorithm for Obtaining Numerical Results 353
A study about the system behavior as K → +∞ is planned for a future work. Any-
way, big values for K > 0 allow the gap function 12 Ω Nαβ w,α w,β dx to be nonposi-
tive at a possible optimal point inside the region of convexity of JK∗ .
In this section we develop an algorithm which we prove, under certain mild hy-
potheses; it is convergent up to a subsequence (the result stated in the next lines
must be seen as an existence one and, of course, it is not the full proof of conver-
gence from a numerical analysis point of view). Such a result is summarized by the
following theorem.
Theorem 13.4.1. Consider the system of equations relating the boundary value
form of the Kirchhoff–Love plate model, namely
⎧
⎪
⎪ Mαβ ,αβ + (Nαβ w,α ),β + P = 0, in Ω
⎪
⎪
⎨
Nαβ ,β + Pα = 0 in Ω (13.35)
⎪
⎪
⎪
⎪
⎩
uα = w = ∂∂ wn = 0 on ∂ Ω
where
Nαβ (u) = Hαβ λ μ γλ μ (u), (13.36)
Mαβ (u) = hαβ λ μ κλ μ (u). (13.37)
Define, as above,
1
(G ◦ Λ )(u) = Nαβ (u)γαβ (u) dx
2 Ω
1
+ Mαβ (u)καβ (u) dx, (13.38)
2 Ω
Assume { Pα 2 } are small enough so that (from [22] pages 285–287) if either
uα W 1,2 (Ω ) → ∞
or
w W 2,2 (Ω ) → ∞,
354 13 Duality Applied to a Plate Model
then
J(u) → +∞.
Let {un = ((un )α , wn )} ⊂ U be the sequence obtained through the following al-
gorithm:
1. Set n = 1.
2. Choose (z∗1 )1 , (z∗2 )1 ∈ L2 (Ω ).
3. Compute un by
K K
un = argminu∈U G(Λ u) + (wx )2 dx + (wy )2 dx
2 Ω 2 Ω
−wx , (z∗1 )n L2 − wy , (z∗2 )n L2
1 ∗ 2 1 ∗ 2
+ (z ) dx + (z ) dx − F(u) ,
2K Ω 1 n 2K Ω 2 n
that is,
(z∗1 )n+1 = K(wn )x ,
and
(z∗2 )n+1 = K(wn )y .
5. Set n → n + 1 and go to step 3 till the satisfaction of a suitable approximate
convergence criterion.
Assume {un = ((un )α , wn )} ⊂ U is such that for a sufficiently big K > 0 we have
we have
K K
J(u) = G(Λ u) + (wx )2 dx + (wy )2 dx
2 Ω 2 Ω
−wx , z∗1 L2 − wy , z∗2 L2
K K
− (wx )2 dx − (wy )2 dx
2 Ω 2 Ω
+wx , z∗1 L2 + wy , z∗2 L2 − F(u)
K K
≤ G(Λ u) + (wx )2 dx (wy )2 dx
2 Ω 2 Ω
−wx , z∗1 L2 − wy , z∗2 L2
K K
+ sup − v1 dx −
2
v22 dx
v∈L ×L
2 2 2 Ω 2 Ω
∀u ∈ U, z∗ ∈ L2 (Ω ) × L2 (Ω ).
From the hypotheses, {un} is inside the region of strict convexity of the func-
tional in U (for z∗ fixed) in question, so that it is uniquely defined for each z∗n
356 13 Duality Applied to a Plate Model
(through the general results in [11] we may infer the region of convexity of the
functional
K K
J(u) = G(Λ u) + (wx )2 dx + (wy )2 dx +
2 Ω 2 Ω
−wx , (z∗1 )n L2 − wy , (z∗2 )n L2
1 1
+ (z∗1 )2n dx + (z∗ )2 dx − F(u), (13.45)
2K Ω 2K Ω 2 n
corresponds to the satisfaction of constraints
Denoting
K K
αn = G(Λ un ) + (wn )2x dx + (wn )2y dx +
2 Ω 2 Ω
−(wn )x , (z∗1 )n L2 − (wn )y , (z∗2 )n L2
1 1
+ (z∗1 )2n dx + (z∗ )2 dx − F(un ), (13.47)
2K Ω 2K Ω 2 n
we may easily verify that {αn } is a real nonincreasing sequence bounded below by
infu∈U {J(u)}; therefore, there exists α ∈ R such that
lim αn = α . (13.48)
n→∞
wn W 2,2 (Ω ) < C, ∀n ∈ N,
0
and
(uα )n W 1,2 (Ω ) < C, ∀n ∈ N.
0
and
wn → w0 , strongly in W01,2 (Ω ),
so that, considering the algorithm in question,
zn → z∗0 strongly in L2 (Ω ; R2 ),
where
(z∗0 )α = K(w0 ),α .
From these last results, the Sobolev imbedding theorem and relating results (more
specifically, Korn’s inequality and its consequences; details may be found in [21]),
we have that there exist K1 , K2 > 0 such that
and
(wn ),α 4 < K2 , ∀n ∈ N, α ∈ {1, 2}.
On the other hand, un ∈ U such that
K K
un = argminu∈U G(Λ u) + (wx )2 dx + (wy )2 dx
2 Ω 2 Ω
−wx , (z∗1 )n L2 − wy , (z∗2 )n L2
1 1
+ (z∗1 )2n dx + (z∗2 )2n dx − F(u) (13.49)
2K Ω 2K Ω
and
(un )λ ,μ + (un )λ ,μ (wn ),λ (wn ),μ
Hαβ λ μ +
2 2 ,β
+Pα = 0 in Ω , (13.51)
in the sense of distributions (theoretical details about similar results may be found
in [25]).
358 13 Duality Applied to a Plate Model
→ 0, as n → ∞.
(u0 )λ ,μ + (u0)λ ,μ (w0 ),λ (w0 ),μ
− Hαβ λ μ + (w0 ),β − P, φ
2 2 ,α L2
./
= lim hαβ λ μ (wn ),λ μ ,αβ
n→∞
(un )λ ,μ + (un)λ ,μ (wn ),λ (wn ),μ
− Hαβ λ μ + (wn ),β
2 2 ,α
−K(wn ),αα + (z∗n )α ,α − P, φ
L2
= lim 0 = 0.
n→∞
∂ w0
(u0 )α = w0 = = 0, on ∂ Ω .
∂n
The proof is complete.
360 13 Duality Applied to a Plate Model
Remark 13.4.2. We emphasize that for each n ∈ N, from the condition indicated in
(13.42), {un } is obtained through the minimization of a convex functional. There-
fore the numerical procedure translates into the solution of a sequence of convex
optimization problems.
In this section we present some numerical results. Let Ω = [0, 1] × [0, 1] and
consider the problem of minimizing J : U → R where
Here
Γ0 = {[0, y] ∪ [x, 0], 0 ≤ x, y ≤ 1}, and P, P1 , P2 ∈ L2 denote the external loads in the
directions a3 , a1 , and a2 , respectively.
We consider the particular case where all entries of {Hαβ λ μ } and {hαβ λ μ } are
zero, except for H1111 = H2222 = H1212 = 105 and h1111 = h2222 = h1212 = 104 .
Moreover P = 1000, P1 = −100, and P2 = −100 (units refer to the international
system). In a first moment, define the trial functions w : Ω → R, u1 : Ω → R, and
u2 : Ω → R by
(a0 )1 = 0.000832
We have that
1 1
W (a1 , a2 ) = h w w dx + (N(u0 ))αβ w,α w,β dx
2 Ω αβ λ μ ,αβ ,λ μ 2 Ω
= 360319.a21 + 191.511a1a2 + 5.7668 ∗ 106a22 . (13.58)
x 10−4
6
5
4
3
2
1
0
1.5
1 1.5
1
0.5 0.5
0 0
13.6 Conclusion
In this chapter, we develop duality principles for the Kirchhoff–Love plate model.
The results are obtained through the basic tools of convex analysis and include suf-
ficient conditions of optimality. It is worth mentioning that earlier results require
the membrane force tensor to be positive definite in a critical point in order to
guarantee global optimality, whereas from the new results here presented, we are
able to guarantee global optimality for a critical point such that N11 (u0 ) < 0 and
N22 (u0 ) < 0, in Ω . Finally, the methods developed here may be applied to many
other nonlinear models of plates and shells. Applications to related areas (specially
to the shell models found in [23]) are planned for future works.
Chapter 14
About Ginzburg–Landau-Type Equations:
The Simpler Real Case
14.1 Introduction
In this chapter, our first objectives are to show existence and develop dual for-
mulations concerning the real semi-linear Ginzburg–Landau equation. We start by
describing the primal formulation.
By S ⊂ R3 we denote an open connected bounded set with a sufficiently regular
boundary Γ = ∂ S (regular enough so that the Sobolev imbedding theorem holds).
The Ginzburg–Landau equation is given by
2
−∇2 u + α ( u2 − β )u − f = 0 in S, (14.1)
u = 0 on Γ ,
where u : S → R denotes the primal field and f ∈ L2 (S). Moreover, α , β are real
positive constants.
Remark 14.1.1. The complex Ginzburg–Landau equation plays a fundamental role
in the theory of superconductivity (see [4], for details). In the present work we deal
with the simpler real form; however, the results obtained may be easily extended to
the complex case.
The corresponding variational formulation is given by the functional J : U → R,
where
1 α u2
J(u) = |∇u|2 dx + ( − β )2 dx − f u dx (14.2)
2 S 2 S 2 S
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 363
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 14,
© Springer International Publishing Switzerland 2014
364 14 About Ginzburg–Landau-Type Equations: The Simpler Real Case
Furthermore, from the above and the Poincaré inequality, it is clear that for J given
by (14.2), we have
J(u) → +∞ as u W 1,2 (S) → +∞,
that is, J is coercive.
Now we establish the existence of a minimizer for J : U → R. It is a well-known
procedure (the direct method of calculus of variations). We present it here for the
sake of completeness.
Theorem 14.1.3. For α , β ∈ R+ , f ∈ L2 (S) there exists at least one u0 ∈ U such that
J(u0 ) = min{J(u)}
u∈U
where
1 α u2
J(u) = |∇u| dx +
2
( − β )2 dx − f u dx
2 S 2 S 2 S
J(u) → +∞ as u U → +∞.
Also from the Poincaré inequality, there exists α1 ∈ R such that α1 =
infu∈U {J(u)} so that for {un } minimizing sequence, in the sense that
J(un ) → α1 as n → +∞ (14.3)
we have that un U is bounded, and thus, as W01,2 (S) is reflexive, there exists u0 ∈
W01,2 (S) and a subsequence {un j } ⊂ {un } such that
Furthermore, defining J1 : U → R as
1 α
J1 (u) = |∇u|2 dx + u4 dx − f u dx
2 S 8 S S
We start this section by enunciating the following theorem which has been proven
in [11].
Our next result refers to a convex dual variational formulation, through which we
obtain sufficient conditions for optimality.
366 14 About Ginzburg–Landau-Type Equations: The Simpler Real Case
and U = W01,2 (S). For K = 1/K0 , where K0 stands for the constant related to the
Poincaré inequality, we have the following duality principle:
where
1 1 1 (v∗1 )2
G∗L (v∗ , z∗ ) = |∇z∗ |2 dx − (z∗ )2 dx + ∗ dx
2K 2 S 2K S 2 S v0 + K
1
+ (v∗0 )2 dx + β v∗0 dx, (14.8)
2α S S
and
where
α 2
v̄∗0 = u −β,
2 0
v̄∗1 = (v̄∗0 + K)u0
and
z̄∗ = Ku0 .
where
1 α u2 K
G(Λ u) = |∇u|2 dx + ( − β + 0)2 dx + u2 dx,
S2 S 2 2 2 S
K
F(Λ1 u) = u2 dx,
2 S
14.2 A Concave Dual Variational Formulation 367
where
Λ u = {Λ0 u, Λ1 u, Λ2 u},
and
Λ0 u = 0, Λ1 u = u, Λ2 u = ∇u.
From Theorem 14.2.1 (here this is an auxiliary theorem through which we obtain
A∗ , below indicated), we have
where
1
F ∗ (z∗ ) = (z∗ )2 dx,
2K S
and
∗ ∗ 1 1 (v∗1 )2 1
G (v ) = |v∗2 |2 dx + ∗ + K dx + 2α (v∗0 )2 dx + β v∗0 dx,
2 S 2 v
S 0 S S
or
Observe that
G∗ (v∗ ) ≥ Λ u, v∗ Y − G(Λ u), ∀u ∈ U, v∗ ∈ A∗ ,
and thus
−F ∗ (z∗ ) + G∗ (v∗ ) ≥ −F ∗ (z∗ ) + Λ1u, z∗ L2 (S) + u, f U − G(Λ u), (14.9)
and hence, making z∗ an independent variable through A∗ , from (14.9), we may write
∗ ∗ ∗ ∗ ∗ ∗ ∗ ∗
sup {−F (z ) + G (v2 (v1 , z ), v1 , v0 )} ≥ sup −F ∗ (z∗ )
z∗ ∈L2 (S) z∗ ∈L2 (S)
+Λ1 u, z∗ L2 (S) + f u dx − G(Λ u) , (14.10)
S
368 14 About Ginzburg–Landau-Type Equations: The Simpler Real Case
so that
1 1 1 (v∗1 )2
sup − (z∗ )2 dx + (v∗2 (z∗ , v∗1 ))2 dx + ∗ dx
z∗ ∈L2 (S) 2K S 2 S 2 S v0 + K
1
+ (v∗0 )2 dx + β v∗0 dx
2α S S
≥ F(Λ1 u) + f udx − G(Λ u). (14.11)
S
∇z∗
v∗2 = and z∗ = 0 on Γ ,
K
so that the final format of our duality principle is given by
1 1
inf {J(u)} ≥ sup − 2 |∇z∗ |2 dx + (z∗ )2 dx
u∈U ∗ ∗ ∗
(z ,v ,v )∈B ∗ 2K S 2K S
1 0
1 (v∗1 )2 1
− ∗ + K dx − 2α (v∗0 )2 dx − β v∗0 dx , (14.12)
2 v
S 0 S S
where
and 1
1
F(u) = (u − f )2 dx
2 0
we may write
J(u) = G(Λ u) + F(u)
where, for convenience, we define Λ : U → Y ≡ L4 (S) × L2(S) as
Λ u = {u,x , 0}.
Furthermore, we have
G(Λ u) + F(u) ≥ inf {−p0 , v∗0 L2 (S) − p1, v∗1 L2 (S) + G(Λ u + p) + F(u)}
p∈Y
or
G(Λ u) + F(u) ≥ inf {−q0 , v∗0 L2 (S) − q1, v∗1 L2 (S) + G(q)
p∈Y
Here q = Λ u + p so that
G(Λ u) + F(u) ≥ −G∗ (v∗ ) + 0, v∗0 L2 (S) + u,x , v∗1 L2 (S) + F(u).
370 14 About Ginzburg–Landau-Type Equations: The Simpler Real Case
That is,
G(Λ u) + F(u) ≥ −G∗ (v∗ ) + inf {0, v∗0 L2 (S) + u,x , v∗1 L2 (S) + F(u)},
u∈U
or
inf {G(Λ u) + F(u)} ≥ sup {−G∗ (v∗ ) − F ∗ (−Λ ∗ v∗ )}
u∈U v∗ ∈A∗
where
1 (v∗1 )2 1
G∗ (v∗ ) = ∗ dx + (v∗0 )2 dx,
2 S v0 2 S
and
A∗ = {v∗ ∈ Y ∗ | v∗0 > 0, a.e. in S}.
Remark 14.3.1. Through the extremal condition v∗0 = ((u,x )2 − 1) and Weierstrass
condition (u,x )2 − 1.0 ≥ 0 we can see that the dual formulation is convex for v∗0 > 0;
however, it is possible that the primal formulation has no minimizers, and we could
expect a microstructure formation through v∗0 = 0 (i.e., u,x = ±1, depending on f(x)).
To allow v∗0 = 0 we will redefine the primal functional as indicated below.
Define G1 : U → R and F1 : U → R by
K
G1 (u) = G(Λ u) + F(u) + (u,x )2 dx
2 S
and
K
F1 (u) = (u,x )2 dx.
2 S
Also defining Ĝ(Λ u) = G(Λ u) + K2 S (u,x ) dx,
2 from Theorem 14.2.1, we can write
inf {J(u)} ≤ ∗inf ∗ sup {F1∗ (z∗ ) − Ĝ∗(v∗0 , v∗2 ) − F ∗ (v∗1 )} (14.13)
u∈U z ∈Y v∗ ∈B∗ (z∗ )
where
1
F1∗ (z∗ ) = (z∗ )2 dx,
2K S
1 (v∗2 )2 1
Ĝ∗ (v∗0 , v∗2 ) = ∗ dx + (v∗0 )2 dx,
2 S v0 + K 2 S
1
F ∗ (v∗1 ) = (v∗1 )2 dx + f , v∗1 L2 (S) − v∗2 (1)u(1)
2 S
14.3 A Numerical Example 371
and
0.6
0.5
0.4
0.3
0.2
0.1
−0.1
0 0.2 0.4 0.6 0.8 1
Fig. 14.1 Vertical axis: u0 (x)-weak limit of minimizing sequences for f(x)=0
We present numerical results for f (x) = 0 (see Fig. 14.1), f (x) = 0.3 ∗ Sin
(π ∗ x) (Fig. 14.2), and f (x) = 0.3 ∗ Cos(π ∗ x) (Fig. 14.3). The solutions indicated
as optimal through the dual formulations (denoted by u0 ) are in fact weak cluster
points of minimizing sequences for the primal formulations.
372 14 About Ginzburg–Landau-Type Equations: The Simpler Real Case
0.5
0.45
0.4
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
0 0.2 0.4 0.6 0.8 1
Fig. 14.2 Vertical axis: u0 (x)-weak limit of minimizing sequences for f (x) = 0.3 ∗ Sin(π ∗ x)
0.5
0.45
0.4
0.35
0.3
0.25
0.2
0.15
0.1
0.05
0
0 0.2 0.4 0.6 0.8 1
Fig. 14.3 Vertical axis: u0 (x)-weak limit of minimizing sequences for f (x) = 0.3 ∗Cos(π ∗ x)
Chapter 15
The Full Complex Ginzburg–Landau System
15.1 Introduction
Remark 15.1.1. This chapter was published in an article form by Applied Mathe-
matics and Computation-Elsevier, reference [12].
We recall that about the year 1950 Ginzburg and Landau introduced a theory to
model the superconducting behavior of some types of materials below a critical
temperature Tc , which depends on the material in question. They postulated that the
free-energy density may be written close to Tc as
h̄ α (T ) β (T )
Fs (T ) = Fn (T ) + |∇ψ |22 dx + |ψ |4 dx − |ψ |2 dx,
4m Ω 4 Ω 2 Ω
where ψ is a complex parameter and Fn (T ) and Fs (T ) are the normal and super-
conducting free-energy densities, respectively. (see [4, 9, 45, 46] for details). Here
Ω ⊂ R3 denotes the superconducting sample with a boundary denoted by ∂ Ω = Γ .
The complex function ψ ∈ W 1,2 (Ω ; C) is intended to minimize Fs (T ) for a fixed
temperature T .
Denoting α (T ) and β (T ) simply by α and β , the corresponding Euler–Lagrange
equations are given by
⎧ h̄ 2
⎨ − 2m ∇ ψ + α |ψ |2 ψ − β ψ = 0, in Ω
(15.1)
⎩ ∂ψ
∂ n = 0, on ∂ Ω .
This last system of equations is well known as the Ginzburg–Landau (G-L) one.
In the physics literature, it is also well known the G-L energy in which a magnetic
potential here denoted by A is included. The functional in question is given by
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 373
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 15,
© Springer International Publishing Switzerland 2014
374 15 The Full Complex Ginzburg–Landau System
2
1 h̄2 ∇ψ − 2ie Aψ dx
J(ψ , A) = | curl A − B0 |22 dx +
8π R3 4m Ω h̄c 2
α β
+ |ψ |4 dx − |ψ |2 dx (15.2)
4 Ω 2 Ω
Considering its minimization on the space U, where
and
⎧
⎨ curl (curl A) = curl B0 + 4cπ J,˜ in Ω
(15.4)
⎩
curl (curl A) = curl B0 , in R3 − Ω ,
where
ieh̄ ∗ 2e 2
J˜ = − (ψ ∇ψ − ψ ∇ψ ∗ ) − |ψ |2 A.
2m mc
and
B0 ∈ L2 (R3 ; R3 )
is a known applied magnetic field.
Sobolev imbedding theorems, the trace theorem, and the Gauss–Green formulas of
integration by parts hold. Details about such results may be found in [1, 26].
An ∞ < K, ∀n ∈ N
for some finite K > 0. Under such a hypothesis, there exists (ψ0 , A0 ) ∈ U such that
such that
An ∞ < K, ∀n ∈ N
for some finite K > 0.
Observe that
1
J(ψn , An ) ≥ | curl An − B0 |22 dx
8π R3
h̄2 h̄2 2ie
+ |∇ψn |2 dx − K
2
∇ψn 2 ψn 2
4m Ω 2m h̄c
2
+
h̄2 2ie An ψn dx + α |ψn |4 dx
4m Ω h̄c
2 4 Ω
β
− |ψn | dx, ∀n ∈ N.
2
(15.7)
2 Ω
Suppose, to obtain contradiction, that there exists a subsequence {ψnk } such that
either ψnk 4 → ∞ or ∇ψnk 2 → ∞, as k → ∞. In such a case from (15.7) we
would obtain J(ψnk , Ank ) → +∞, as k → ∞, which contradicts (15.6).
Therefore, there exists K1 > 0 such that
ψn 4 < K1 and ∇ψn 2 < K1 ,
376 15 The Full Complex Ginzburg–Landau System
∀n ∈ N. From this we may conclude that there exists K2 > 0 such that
ψn 2 < K2 , ∀n ∈ N. Hence by the Rellich–Kondrachov theorem, there exists
ψ0 ∈ W 1,2 (Ω ; C) and a subsequence not relabeled such that
∇ψn ∇ψ0 , weakly in L2 (Ω ; C3 )
and
ψn → ψ0 strongly in L2 (Ω ; C). (15.8)
On the other hand, since { An ∞ } is uniformly bounded, there exists A0 ∈
L∞ (Ω ; R3 ) such that up to a subsequence not relabeled we have
An A0 , weakly star in L∞ (Ω ; R3 ).
Fix v ∈ L2 (Ω , C3 ), since
|vψ0 | dx ≤ v 2 ψ0 2
Ω
for some K3 > 0, so that there exists v0 ∈ L2 (R3 ; R3 ) such that up to a subsequence
not relabeled, we have
A0 , curl∗ φ L2 (R3 ;R3 ) = lim An , curl∗ φ L2 (R3 ;R3 )
n→∞
= lim curl An , φ L2 (R3 ;R3 )
n→∞
= v0 , φ L2 (R3 ;R3 ) . (15.11)
v0 = curl A0 ,
In this section we study the existence of solutions for a closely related optimal
control problem. In particular the state equation is of Ginzburg–Landau type. It is
worth mentioning that the present case refers to the simpler real one. In the next
lines we describe such a problem.
Let Ω ⊂ R3 be an open, bounded, and connected set with a regular boundary
denoted by ∂ Ω = Γ . Let ψd : Ω → R be a function such that ψd ∈ L2 (Ω ). Consider
the problem P, that is, the problem of minimizing J : U → R given by
1 1
J(ψ , u) = |ψ − ψd |2 dx + |u|2 dΓ
2 Ω 2 ∂Ω
378 15 The Full Complex Ginzburg–Landau System
subject to
−∇2 ψ + αψ 3 − β ψ = f , in Ω
∂ψ (15.13)
∂ n = u, on ∂ Ω .
Here U = W 1,2 (Ω ) × L2 (∂ Ω ) and f ∈ L2 (Ω ).
We say that the set of admissible fields for problem P is nonempty if there exists
(ψ , u) ∈ U satisfying (15.13).
A similar problem is studied in [40] through a different approach. We will prove
that such a problem has a solution. We start with the following proposition:
Proposition 15.3.1. The set of admissible fields for problem P is nonempty.
Proof. From reference [13], Chap. 13, there exists ψ̃ ∈ W01,2 (Ω ) which minimizes
J˜ on W01,2 (Ω ), where
˜ ψ) = 1 α β
J( |∇ψ |22 dx + |ψ |4 dx − |ψ |2 dx − ψ , f L2 (Ω ) ,
2 Ω 4 Ω 2 Ω
so that
−∇2 ψ̃ + α ψ̃ 3 − β ψ̃ = f , in Ω
(15.14)
ψ̃ = 0, on ∂ Ω .
so that
|∇ψnk |22 dx + α |ψnk |4 dx − β |ψnk |2 dx
Ω Ω Ω
) *
∂ ψn k
−ψnk , f L2 (Ω ) − , ψn k = 0, (15.17)
∂n L2 (∂ Ω )
Hence
β |ψnk |2 dx ≥ |∇ψnk |22 dx + α |ψnk |4 dx
Ω Ω Ω
− ψnk 2 f 2 − unk L2 (∂ Ω ) ψnk L2 (∂ Ω )
≥ |∇ψnk |22 dx + α |ψnk |4 dx
Ω Ω
−K f L2 (Ω ) − K ψnk L2 (∂ Ω ) (15.18)
and thus, from the trace theorem, there exists C1 > 0 such that
β |ψnk |2 dx + ψnk 22 ≥ |∇ψnk |22 dx + α |ψnk |4 dx
Ω Ω Ω
+ ψnk 22 − K f 2
−KC1 ψnk W 1,2 (Ω ) . (15.19)
ψn ψ̃ , weakly in L4 (Ω ),
so that from
ψn ψ0 , weakly in L2 (Ω ),
we get
ψ̃ = ψ0 .
From the last results and (15.17) we may obtain
|∇ψn |22 dx ≤ α K14 + β K 2 + K f 2 + K C1 ψn W 1,2 (Ω )
Ω
!
≤ α K14 + β K 2 + K f 2 + K C1 ψn 22 + ∇ψn 22
!
≤ α K14 + β K 2 + K f 2 + K C1 K 2 + ∇ψn 22 ,
380 15 The Full Complex Ginzburg–Landau System
∇ψn 2 < K2 , ∀n ∈ N.
ψn W 1,2 (Ω ) < K3 , ∀n ∈ N.
and
ψn → ψ̂ , strongly in L2 (Ω )
so that as
ψn ψ0 , weakly in L2 (Ω )
we can get
ψ̂ = ψ0 ,
that is,
∇ψn ∇ψ0 , weakly in L2 (Ω )
and
ψn → ψ0 , strongly in L2 (Ω ).
Choose φ ∈ Cc∞ (Ω ). Clearly we have
and
ψn , φ L2 (Ω ) → ψ0 , φ L2 (Ω ) , (15.21)
and in the next lines, we will prove that
as n → ∞. Observe that
(ψ 3 − ψ 3 )φ dx ≤ (ψ 3 − ψ 2ψ0 + ψ 2 ψ0 − ψ 3 )φ dx
Ω n 0 Ω n n n 0
≤ ψn2 (ψn − ψ0 )φ dx
Ω
+ ψ0 (ψ 2 − ψ 2)φ dx. (15.23)
n 0
Ω
15.3 A Related Optimal Control Problem 381
as n → ∞. Therefore
−∇2 ψ0 + αψ03 − β ψ0 − f = 0, in Ω ,
∂ ψn
= un , on ∂ Ω , ∀n ∈ N,
∂n
we may also obtain
∂ ψ0
= u0 , on ∂ Ω .
∂n
Finally, from the convexity of the functional in question,
−r (θ )
f1 (θ ) = ,
r(θ )
we have
f1 (θ )
f2 (θ ) = 1 + ,
1 + f1(θ )2
2 f1 (θ )
f3 (θ ) = ,
1 + f1 (θ )2
and
1
f4 (θ ) = .
1 + f1 (θ )2
Observe that t ∈ [1, 2] in Ω . Discretizing in t (N equal pieces which will generate
N lines) we obtain the equation
The next result is the base of our generalized method of lines. For a proof
see [47].
384 15 The Full Complex Ginzburg–Landau System
u1 = T (0, u1 , u2 ).
We may use the contraction mapping theorem to calculate u1 as a function of u2 .
The procedure would be
1. set x0 = u2 ,
2. obtain x1 = T (0, x0 , u2 ),
3. obtain recursively
xk+1 = T (0, xk , u2 ), and
4. finally get
u1 = lim xk = g1 (u2 ).
k→∞
1
u1 − u2/2 ≤ d T̃ u2 /2 ,
1−α
Having such an estimate, we may similarly obtain
u2 ≈ u3 + O(d),
and generically
un ≈ un+1 + O(d), ∀n ∈ {1, . . . , N − 1}.
This last calculation is just to clarify that the procedure of obtaining the relation
between consecutive lines through the contraction mapping theorem is well defined.
386 15 The Full Complex Ginzburg–Landau System
Consider again the equation in finite differences for the example in question:
The approximation error in (15.35) is of order O(d 3 ) plus the error concerning the
application of the contraction mapping theorem, which is well known and, if d is
small enough, may be made arbitrarily small in a reasonable number of iterations.
Also we may infer that the approximation error in (15.36) is also of order O(d 3 ).
The discretization error in this case is known to be of order O(d) (see [63] for
details).
15.4 The Generalized Method of Lines 387
u = u0 on Γ0 and u = u f , on Γ1 .
We assume u0 and u f are smooth functions. As above Γ0 denotes the internal bound-
ary of Ω and Γ1 the external one. We consider the simpler case where
Γ1 = 2Γ0 .
Suppose there exists r(θ ), a smooth function such that
Γ0 = {(θ , r(θ )) | 0 ≤ θ ≤ 2π },
being r(0) = r(2π ).
Also assume (0, 0) ∈ Ω and
Line 1
u1 (x) = 0.1u f (x) + 0.9u0(x) − 0.034u0(x) f2 (x) + 0.034 f2(x)u f (x)
−0.034 f3(x)u0 (x) + 0.034 f3(x)uf (x)
+0.018 f4(x)u0 (x) + 0.008 f4(x)uf (x)
Line 2
u2 (x) = 0.2u f (x) + 0.8u0(x) − 0.058u0(x) f2 (x) + 0.058 f2(x)u f (x)
−0.058 f3(x)u0 (x) + 0.058 f3(x)uf (x)
+0.029 f4(x)u0 (x) + 0.015 f4(x)uf (x)
Line 3
u3 (x) = 0.3u f (x) + 0.7u0(x) − 0.075u0(x) f2 (x) + 0.075 f2(x)u f (x)
−0.075 f3(x)u0 (x) + 0.075 f3(x)uf (x)
+0.034 f4(x)u0 (x) + 0.020 f4(x)uf (x)
388 15 The Full Complex Ginzburg–Landau System
Line 4
u4 (x) = 0.4u f (x) + 0.6u0(x) − 0.083u0(x) f2 (x) + 0.083 f2(x)u f (x)
−0.083 f3(x)u0 (x) + 0.083 f3(x)uf (x)
+0.035 f4(x)u0 (x) + 0.024 f4(x)uf (x)
Line 5
u5 (x) = 0.5u f (x) + 0.5u0(x) − 0.085u0(x) f2 (x) + 0.085 f2(x)u f (x)
−0.085 f3(x)u0 (x) + 0.085 f3(x)uf (x)
+0.033 f4(x)u0 (x) + 0.026 f4(x)uf (x)
Line 6
u6 (x) = 0.6u f (x) + 0.4u0(x) − 0.080u0(x) f2 (x) + 0.080 f2(x)u f (x)
−0.080 f3(x)u0 (x) + 0.080 f3(x)uf (x)
+0.028 f4(x)u0 (x) + 0.026 f4(x)uf (x)
Line 7
u7 (x) = 0.7u f (x) + 0.3u0(x) − 0.068u0(x) f2 (x) + 0.068 f2(x)u f (x)
−0.068 f3(x)u0 (x) + 0.068 f3(x)uf (x)
+0.023 f4(x)u0 (x) + 0.023 f4(x)uf (x)
Line 8
u8 (x) = 0.8u f (x) + 0.2u0(x) − 0.051u0(x) f2 (x) + 0.051 f2(x)u f (x)
−0.051 f3(x)u0 (x) + 0.051 f3(x)uf (x)
+0.015 f4(x)u0 (x) + 0.018 f4(x)uf (x)
Line 9
u9 (x) = 0.9u f (x) + 0.1u0(x) − 0.028u0(x) f2 (x) + 0.028 f2(x)u f (x)
−0.028 f3(x)u0 (x) + 0.028 f3(x)uf (x)
+0.008 f4(x)u0 (x) + 0.010 f4(x)uf (x)
Remark 15.4.5. Here a word of caution is necessary. Consider for example the
equation
ε ∇2 u + G(u) = 0, in Ω ⊂ R2 , (15.38)
with the boundary conditions
u = u0 on Γ0 and u = u f , on Γ1 .
We assume G, u0 , and u f are smooth functions.
15.5 A First Numerical Example 389
If ε is too small, for example, about 0.001 or 0.0001, the error just truncating
the series up the order d 2 is big. It seems that higher-order approximations or even
discretizing more does not solve the problem. However, for example, for G(u) = u,
by solving the equation with ε = 1, we can infer that the solution at each line has
the general format
un (x) ≈ a1 [n]u f (x) + a2 [n]u0 (x) + a3[n]u0 (x) f2 (x) + a4[n] f2 (x)u f (x)
a5 [n] f3 (x)u0 (x) + a6[n] f3 (x)uf (x)
+a7 [n] f4 (x)u0 (x) + a8 [n] f4 (x)uf (x)
+a9 [n] f5 (x)u0 (x) + a10[n] f5 (x)u f (x),
ū = r2 cos(x).
390 15 The Full Complex Ginzburg–Landau System
See below the approximate values for the 9 lines (N = 10) obtained by the
generalized method of lines (un (x)) and the exact values ( ūn (x) for the same lines):
Line 1
u1 (x) = 1.21683 cos(x), ū1 (x) = 1.21 cos(x)
Line 2
u2 (x) = 1.44713 cos(x), ū2 (x) = 1.44 cos(x)
Line 3
u3 (x) = 1.69354 cos(x), ū3 (x) = 1.69 cos(x)
Line 4
u4 (x) = 1.95811 cos(x) ū4 (x) = 1.96 cos(x)
Line 5
u5 (x) = 2.24248 cos(x), ū5 (x) = 2.25 cos(x)
Line 6
u6 (x) = 2.54796 cos(x), ū6 (x) = 2.56 cos(x)
Line 7
u7 (x) = 2.87563 cos(x), ū7 (x) = 2.89 cos(x)
Line 8
u8 (x) = 3.22638 cos(x), ū8 (x) = 3.24 cos(x)
Line 9
u9 (x) = 3.60096 cos(x), ū9 (x) = 3.61 cos(x)
subject to
⎧
⎨ −∇ ψ + αψ − β ψ = 0, in Ω
2 3
ψ = 0, on Γ0 , (15.39)
⎩ ∂ψ
∂ n = u, on Γ1 .
Also U = W 1,2 (Ω ) × L2 (Γ1 ). In this example we consider in polar coordinates
Ω = {(r, θ ) | 1 ≤ r ≤ 2, 0 ≤ θ ≤ 2π },
Γ0 = {(1, θ ) | 0 ≤ θ ≤ 2π },
Γ1 = {(2, θ ) | 0 ≤ θ ≤ 2π },
and
ψd (r, θ ) = (r − 1)2 sin θ .
We discretize the domain in lines (in fact curves). We divide the interval [1, 2] into
10 pieces (corresponding to the discretization in r) obtaining the following system
of equations:
Line 1
ψ1 (x) = 0.150254u f (x) − 0.004917u f (x)3 + 0.0078404uf (x)
Line 2
ψ2 (x) = 0.290418u f (x) − 0.009824u f (x)3 + 0.0148543uf (x)
Line 3
ψ3 (x) = 0.420248u f (x) − 0.014651u f (x)3 + 0.0204794uf (x)
Line 4
ψ4 (x) = 0.539385u f (x) − 0.019208u f (x)3 + 0.0243293uf (x)
Line 5
ψ5 (x) = 0.64738u f (x) − 0.023125u f (x)3 + 0.0261384uf (x)
392 15 The Full Complex Ginzburg–Landau System
Line 6
ψ6 (x) = 0.743709u f (x) − 0.025792u f (x)3 + 0.0257253uf (x)
Line 7
ψ7 (x) = 0.827787u f (x) − 0.026299u f (x)3 + 0.0229684uf (x)
Line 8
ψ8 (x) = 0.898983u f (x) − 0.023376u f (x)3 + 0.0177894uf (x)
Line 9
ψ9 (x) = 0.956623u f (x) − 0.015333u f (x)3 + 0.0101412uf (x)
The second step is to replace the field ψ obtained in J and then to compute
through a numerical minimization of J the optimal u f . For the candidate to optimal
u f (x) see Fig. 15.1. Finally, we have computed a critical point, but we cannot guar-
antee it is the global optimal solution.
0.8
0.6
0.4
0.2
0
−0.2
−0.4
−0.6
−0.8
0 20 40 60 80 100
15.7 Conclusion
16.1 Introduction
We recall again here (more details may be found in the introduction at Chap. 15)
that close to a critical temperature Tc , the Ginzburg–Landau energy would be
expressed by
h̄ α (T ) β (T )
Fs (T ) = Fn (T ) + |∇ψ |22 dx + |ψ |4 dx − |ψ |2 dx,
4m Ω 4 Ω 2 Ω
where ψ is a complex parameter and Fn (T ) and Fs (T ) are the normal and super-
conducting free-energy densities, respectively (see [4, 9, 45, 46] for details). Here
Ω ⊂ R3 denotes the superconducting sample with a boundary denoted by ∂ Ω = Γ .
The complex function ψ ∈ W 1,2 (Ω ; C) is intended to minimize Fs (T ) for a fixed
temperature T .
Denoting α (T ) and β (T ) simply by α and β , the corresponding Euler–Lagrange
equations are given by
⎧ h̄ 2
⎨ − 2m ∇ ψ + α |ψ |2 ψ − β ψ = 0, in Ω
(16.1)
⎩ ∂ψ
∂n = 0, on ∂ Ω .
This last system of equations is well known as the Ginzburg–Landau (G-L) one.
In the physics literature, it is also well known the G-L energy in which a magnetic
potential here denoted by A is included. The functional in question is given by
2
1 h̄2 ∇ψ − 2ie Aψ dx
J(ψ , A) = | curl A − B0 |22 dx +
8π R3 4m Ω h̄c 2
α β
+ |ψ |4 dx − |ψ |2 dx (16.2)
4 Ω 2 Ω
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 393
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 16,
© Springer International Publishing Switzerland 2014
394 16 More on Duality and Computation for the Ginzburg–Landau System
and ⎧
⎨ curl (curl A) = curl B0 + 4cπ J,˜ in Ω
(16.4)
⎩
curl (curl A) = curl B0 , in R3 \ Ω ,
where
ieh̄ ∗ 2e 2
J˜ = − (ψ ∇ψ − ψ ∇ψ ∗ ) − |ψ |2 A.
2m mc
and
B0 ∈ L2 (R3 ; R3 )
is a known applied magnetic field.
Also,
1
G∗0 (v∗2 ) = sup 2 ∗
Ak , v2k L2 (Ω ) − |curl A − B0 |2 dx ,
2
(ψ ,A)∈U 8π R3
where u1 , v1 are the real parts and u2 , v2 are the imaginary ones of u, v, respectively.
A similar remark is valid for L2 (Ω ; C3 ).
Moreover,
G∗2 (σ ) = sup {v3 , σ L2 − G2 (v3 )}
v3 ∈L2 (Ω )
1
= (σ + β /2)2 dx (16.9)
α Ω
and
G∗3 (v∗1 , v∗2 , σ ) = sup {−|ψ |2 , σ L2 (Ω ) − A2k , v∗2k L2 (Ω ) − G3 (ψ , A)}
(ψ ,A)∈U
.
= sup −|ψ |2 , σ L2 (Ω ) − A2k , v∗2k L2 (Ω )
(ψ ,A)∈U
) *
2ie
+ψ , div(v∗1 )L2 (Ω ;C) + Aψ , v∗1
h̄c L2 (Ω ;C3 )
396 16 More on Duality and Computation for the Ginzburg–Landau System
so that if v∗ ∈ A∗ , we have
G∗3 (v∗1 , v∗2 , σ ) = −|ψ̃ |2 , σ L2 (Ω )
−Ãk , v∗2k L2 (Ω ) − G3 (ψ̃ , Ã),
2
(16.10)
where (ψ̃ , Ã) ∈ U is the only critical point of the quadratic functional indicated
in (16.10).
Here,
Y ∗ = W 1,2 (Ω ; C3 ) × L2 (Ω ; R3 ) × L2 (Ω ),
and defining
) *
2ie
G̃(ψ , A) = |ψ | 2
, σ L2 (Ω ) + A2k , v∗2k L2 (Ω ) − Aψ , v∗1 ,
h̄c L2 (Ω ;C3 )
we also define
A∗ = A1 ∩ A2 ,
Furthermore,
A2 = {v∗ ∈ Y ∗ | J(A)
˜ > 0, ∀A ∈ W01,2 (R3 ; R3 ) such that A = θ },
and
1
˜
J(A) = |curl A|22 dx − A2k , v∗2k L2 (Ω ) .
8π R3
Finally, define Ind0 : U → R ∪ {+∞} by
0, if ih̄∇ψ + 2e
c Aψ · n = 0, on ∂ Ω ,
Ind0 (ψ , A) = (16.11)
+∞, otherwise.
Assume (ψ0 , A0 ) ∈ U is such that δ J(ψ0 , A0 ) = θ and Ind0 (ψ0 , A0 ) = 0 and also
such that
v∗0 = (v∗01 , v∗02 , σ0 ) ∈ A∗ ,
where such a point is the solution of the following relations:
∂ G1 (∇ψ0 − 2ie
h̄c A0 ψ0 )
v∗01 =
∂ v1
h̄2 2ie
= ∇ψ0 − A0 ψ0 , in Ω , (16.12)
2m h̄c
3
curl curl A0 − curl B0
− ∑ 2(v∗02 )k A0k ek + = θ in Ω , (16.13)
k=1 4π
16.2 The Duality Principle 397
∂ G2 (|ψ0 |2 )
σ0 =
∂ v3
α β
= |ψ0 |2 − , in Ω . (16.14)
2 2
Under such hypotheses, we have
∀(ψ , A) ∈ U, v∗ ∈ A∗ .
Hence,
1
J(ψ , A) = −A2k , v∗2k L2 (Ω ) + | curl A − B0 |22 dx
8π R3
) * 2
2ie h̄2 2ie
− ∇ψ − ∗
Aψ , v 1 + ∇ψ − Aψ dx
h̄c L2 (Ω ;C3 ) 4m Ω h̄c 2
α β
−|ψ |2 , σ L2 (Ω ) + |ψ |4 dx − |ψ |2 dx
4Ω 2 Ω
) *
2 ∗ 2ie ∗
+Ai , v2i L2 (Ω ) + |ψ | , σ L2 (Ω ) + ∇ψ −
2
Aψ , v 1 (16.16)
h̄c L2 (Ω ;C3 )
∀(ψ , A) ∈ U, v∗ ∈ A∗ , so that
1
J(ψ , A) ≥ inf −A2k , v∗2k L2 (Ω ) + | curl A − B0 |22 dx
A∈U 8π R3
∗ h̄2
+ inf −v1 , v1 L2 (Ω ;C3 ) + |v1 |2 dx
2
v1 ∈L2 (Ω ;C3 ) 4m Ω
398 16 More on Duality and Computation for the Ginzburg–Landau System
α β
+ inf −v3 , σ L2 (Ω ) + (v3 ) dx −
2
v3 dx
v3 ∈L2 (Ω ) 4 Ω 2 Ω
) *
∗ 2ie ∗
+ inf −ψ , div(v1 )L2 (Ω ;C) − Aψ , v 1
(ψ ,A)∈U h̄c L2 (Ω ;C3 )
0
+|ψ |2 , σ L2 (Ω ) + A2k , v∗2k L2 (Ω ) ,
∀(ψ , A) ∈ U, v∗ ∈ A∗ .
Therefore,
∀(ψ , A) ∈ U, v∗ ∈ A∗ .
Now observe that since
δ J(ψ0 , A0 ) = θ ,
from the variation in ψ , we get
2
h̄ 2ie h̄2 2ie −2ieA0
−div ∇ψ0 − A0 ψ0 + ∇ψ0 − A0 ψ0 ·
2m h̄c 2m h̄c h̄c
+α |ψ0 |2 ψ0 − β ψ0 = θ , in Ω .
Finally, by (16.22), (16.23), (16.24), (16.25), and from the fact that v∗0 ∈ A∗ , we
obtain
that is,
J(ψ0 , A0 ) = −J ∗ (v∗0 ).
From this and (16.17), the proof is complete.
We first apply Newton’s method. The solution here is obtained similarly as for
the generalized method of lines procedure. See the next sections for details on such
a method for PDEs.
Consider again the equation.
⎧
⎨ u + f (u) + g = 0, in [0, 1]
(16.26)
⎩
u(0) = u0 , u(1) = u f ,
where
An = f (ũn ),
and
Bn = f (ũn ) − f (ũn )ũn + gn .
In particular for n = 1 we get
u2 − 2u1 + u0 + A1 u1 d 2 + B1 d 2 = 0.
u 1 = a 1 u 2 + b 1 u 0 + c1 ,
where
a1 = (2 − A1d 2 )−1 , b1 = a1 , c1 = a1 B1 .
16.3 On the Numerical Procedures for Ginzburg–Landau-Type Equations 401
and
un+1 − 2un + un−1 + Anun d 2 + Bn d 2 = 0,
we get
un+1 − 2un + an−1un + bn−1u0 + cn−1 + An un d 2 + Bn d 2 = 0,
so that
un = an un+1 + bnu0 + cn ,
where
an = (2 − an−1 − An d 2 )−1 ,
bn = an bn−1 ,
and
cn = an (cn−1 + Bnd 2 ),
∀n ∈ 1, . . . , N − 1.
We have thus obtained
and in particular
uN−1 = HN−1 (u f ),
so that we may calculate
The results are obtained for ε = 1.0, ε = 0.1, ε = 0.01, and ε = 0.001. Please
see Figs. 16.1, 16.2, 16.3, and 16.4, respectively. For the other two solutions for
ε = 0.01 see Figs. 16.5 and 16.6.
0.14
0.12
0.1
0.08
0.06
0.04
0.02
0
0 0.2 0.4 0.6 0.8 1
1.4
1.2
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
1.4
1.2
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
1.4
1.2
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
−1
−2
−3
−4
0 0.2 0.4 0.6 0.8 1
18
16
14
12
10
8
6
4
2
0
0 0.2 0.4 0.6 0.8 1
The generalized method of lines was originally developed in [13]. In this work
we address its matrix version. Consider the simpler case where Ω = [0, 1] × [0, 1].
We discretize the domain in x, that is, in N + 1 vertical lines obtaining the following
equation in finite differences (see [63] for details about finite differences schemes).
16.4 Numerical Results for Related PDEs 405
3α d 2
un+1 − 2un + un−1 + M̃2 un − (u0 )2n un
ε
2α β d2 d2
+ (u0 )3n d 2 + un − fn = 0, (16.30)
ε ε ε
2
where M̃2 = M2 dd 2 and
1
⎡ ⎤
−2 1 0 0 ... 0
⎢ 1 −2 1 0 ... 0 ⎥
⎢ ⎥
⎢ 0 1 −2 1 0.. 0 ⎥
M2 = ⎢
⎢... ... ... ...
⎥, (16.31)
⎢ ... ... ⎥
⎥
⎣ 0 0 ... 1 −2 1 ⎦
0 0 ... ... 1 −2
3α d 2
u2 − 2u1 + M̃2 u1 − (u0 )21 u1
ε
2α β d2 d2
+ (u0 )31 d 2 + u1 − f1 = 0. (16.32)
ε ε ε
Denoting
α d2 β d2
M12 [1] = 2Id − M̃2 + 3 (u0 )21 Id − Id ,
ε ε
where Id denotes the (N1 − 1) × (N1 − 1) identity matrix,
2α d 2 d2
Y0 [1] = (u0 )31 − f1 ,
ε ε
and M50 [1] = M12 [1]−1 , we obtain
where
z[1] = M50 [1] ·Y0 [1].
406 16 More on Duality and Computation for the Ginzburg–Landau System
3α d 2
u3 − 2u2 + u1 + M̃2 u2 − (u0 )22 u2
ε
2α β d2 d2
+ (u0 )32 d 2 + u2 − f2 = 0, (16.33)
ε ε ε
that is,
3α d 2
u3 − 2u2 + M50 [1]u2 + z[1] + M̃2 u2 − (u0 )22 u2
ε
2α β d2 d2
+ (u0 )32 d 2 + u2 − f2 = 0, (16.34)
ε ε ε
so that denoting
α d2 β d2
M12 [2] = 2Id − M̃2 − M50 [1] + 3 (u0 )22 Id − Id ,
ε ε
2α d 2 d2
Y0 [2] = (u0 )32 − f2 ,
ε ε
and M50 [2] = M12 [2]−1 , we obtain
where
z[2] = M50 [2] · (Y0 [2] + z[1]).
Proceeding in this fashion, for the line n, we obtain
3α d 2
un+1 − 2un + M50 [n − 1]un + z[n − 1] + M̃2un − (u0 )2n un
ε
2α β d2 d2
+ (u0 )3n d 2 + un − fn = 0, (16.35)
ε ε ε
so that denoting
α d2 β d2
M12 [n] = 2Id − M̃2 − M50[n − 1] + 3 (u0 )2n Id − Id ,
ε ε
and also denoting
2α d 2 d2
Y0 [n] = (u0 )3n − fn ,
ε ε
and M50 [n] = M12 [n]−1 , we obtain
un = M50 [n]un+1 + z[n],
16.4 Numerical Results for Related PDEs 407
where
z[n] = M50 [n] · (Y0[n] + z[n − 1]).
Observe that we have
uN = θ ,
where θ denotes the zero matrix (N1 − 1) × 1, so that we may calculate
and
uN−2 = M50 [N − 2] · uN−1 + z[N − 2],
and so on, up to obtaining
The next step is to replace {(u0 )n } by {un } and thus to repeat the process until
convergence is achieved.
This is Newton’s method; what seems to be relevant is the way we inverted the
big matrix ((N1 −1)·(N −1))×((N1 −1)·(N −1)), and in fact instead of inverting it
directly we have inverted N − 1 matrices (N1 − 1) × (N1 − 1) through an application
of the generalized method of lines.
So far we cannot guarantee convergence; however, through the next theorem,
we describe a procedure that always leads to a solution. Anyway, we highlight the
next result is not a formal proof of convergence in a numerical analysis context. In
fact, such a result must be seen as an existence of one of the critical points for the
equation in question.
that is,
z∗n+1 = β un .
5. Set n → n + 1 and go to step 3 (up to the satisfaction of an appropriate conver-
gence criterion).
The sequence {un } is such that up to a subsequence not relabeled
un → u0 , strongly in L2 (Ω ),
where
u0 ∈ W01,2 (Ω )
is a solution of equation (16.36).
ε α
J(u) = ∇u · ∇u dx + u4 dx
2 Ω 4 Ω
β
− u2 dx + f u dx , (16.40)
2 Ω Ω
we have
ε α β
J(u) = ∇u · ∇u dx + u dx −
4
u2 dx
2 Ω 4 Ω
2 Ω
Denoting
ε α
αn = ∇un · ∇un dx + u4n dx
2 Ω 4 Ω
1
−un , z∗n L2 + (z∗n )2 dx + f un dx, (16.42)
2β Ω Ω
we may easily verify that {αn } is a real nonincreasing sequence bounded below by
infu∈U {J(u)}; therefore there exists α ∈ R such that
lim αn = α . (16.43)
n→∞
From the Poincaré inequality (see [1] for details) we have that J(u) → +∞ if
u W 1,2 (Ω ) → ∞. From this, (16.41), (16.42), and (16.43), we may infer that
un W 1,2 (Ω ) < C, ∀n ∈ N
un → u0 strongly in L2 (Ω ),
so that considering the algorithm in question
zn → z∗0 strongly in L2 (Ω ),
where
z∗0 = β u0 .
Observe that the unique un ∈ U such that
ε α
un = argminu∈U ∇u · ∇u dx + u4 dx
2 Ω 4 Ω
1
−u, z∗n L2 + (z∗n )2 dx + f u dx , (16.44)
2β Ω Ω
as n → ∞. First observe that from (16.41), (16.42), and (16.43), it is clear that there
exists K1 > 0 such that
un 4 < K1 , ∀n ∈ N.
Observe also that
(u3n − u3 )φ dx ≤ (u3n − u2n u0 + u2n u0 − u3)φ dx
Ω 0 Ω 0
≤ u2n (un − u0)φ dx
Ω
+ u0 (u2n − u20)φ dx. (16.46)
Ω
Furthermore,
u2 (un − u0 )φ dx ≤ un 2 (un − u0 )φ 2
n 4
Ω
≤ K12 un − u0 2 φ ∞
→ 0, as n → ∞. (16.47)
On the other hand, from the generalized Hölder inequality, we get
u0 (u2n − u20 )φ dx = |u0 (un + u0 )(un − u0)φ | dx
Ω Ω
≤ u0 4 un + u0 4 (un − u0)φ 2
≤ u0 4 ( un 4 + u0 4 ) (un − u0) 2 φ ∞
≤ K1 (K1 + K1 ) φ ∞ un − u0 2
→ 0, as n → ∞. (16.48)
Summarizing the last results we get
u3n , φ L2 (Ω ) → u30 , φ L2 (Ω ) , (16.49)
as n → ∞. So, we may write
0 = lim {un , ε ∇2 φ L2 + −α u3n + z∗n − f , φ L2 }
n→∞
= u0 , ε ∇2 φ L2 + −α u30 + z∗0 − f , φ L2
= u0 , ε ∇2 φ L2 + −α u30 + β u0 − f , φ L2
(16.50)
that is,
ε ∇2 u0 − α u30 + β u0 = f in Ω ,
in the sense of distributions. From un = 0 on ∂ Ω , ∀n ∈ N we also obtain in a weak
sense
u0 = 0, on ∂ Ω .
The proof is complete.
16.5 Numerical Results 411
Remark 16.4.2. Observe that for each n, the procedure of evaluating un stands for
the solution of a convex optimization problem with unique solution, given by the
one of equation
ε ∇2 un − α u3n + z∗n + f = 0 in Ω ,
which may be easily obtained, due to convexity, through the generalized method of
lines (matrix version) associated with Newton’s method as above described.
0.08
0.06
0.04
0.02
0
1
1
0.5 0.8
0.6
0.4
0.2
0 0
1.6
1.4
1.2
0.8
1
1
0.5 0.8
0.6
0.4
0.2
0 0
2
where M̃2 = M2 dd 2 and
1
⎡ ⎤
−2 1 0 0 ... 0
⎢ 1 −2 1 0 . . . 0 ⎥
⎢ ⎥
⎢ ⎥
⎢ 0 1 −2 1 0.. 0 ⎥
M2 = ⎢
⎢...
⎥, (16.55)
⎢ ... ... ... ... ... ⎥
⎥
⎢ ⎥
⎣ 0 0 . . . 1 −2 1 ⎦
0 0 . . . . . . 1 −2
where
M12 [1] = 2Id − M̃2 ,
Id denotes the (N1 − 1) × (N1 − 1) identity matrix, M50 [1] = M12 [1]−1 ,
M60 [1] = M50 [1],
so that denoting
M12 [2] = 2Id − M̃2 − M50 [1],
and also denoting M50 [2] = M12 [2]−1 , we obtain
u2 = M50 [2]u3 + M60[2]G(u3 )d 2 + z[2] + Er[2],
where
M60 [2] = M50 [2] · (M60[1] + Id ),
z[2] = M50 [2] · (z[1] − f2d 2 ).
and
Er[2] = M50 [2](Er[1]) + M60[2](G(u2 ) − G(u3 ))d 2 .
Proceeding in this fashion, for the line n, we obtain
so that denoting
M12 [n] = 2Id − M̃2 − M50[n − 1],
and M50 [n] = M12 [n]−1 , we obtain
Remark 16.6.2. We may use, as a first approximation for the solution, the relations
un ≈ M50 [n]un+1 + M60[n]G(un+1 )d 2 + z[n].
Observe that we have
uN = θ ,
where θ denotes the zero matrix (N1 − 1) × 1, so that we may calculate
uN−1 ≈ M50 [N − 1] · uN + M60 [N − 1] · G(uN )d 2 + z[N − 1],
and
uN−2 ≈ M50 [N − 2] · uN−1 + M60 [N − 2] · G(uN−1)d 2 + z[N − 2],
and so on, up to obtaining
The next step is to use the {un } obtained as the initial solution for Newton’s
method.
What is relevant is that in general, the first approximation is a good one for the
exact solution.
We have computed the first approximation using such a method, for
−u3 u
G(u) = + ,
ε ε
1
f (x, y) = − , ∀(x, y) ∈ Ω
ε
and ε = 0.01. Please see Fig. 16.9.
This first approximation is close to the solution obtained through Newton’s
method. For the solution through the earlier approach, see Fig. 16.10.
1.5
0.5
0
1
1
0.5 0.8
0.6
0.4
0.2
0 0
1.5
0.5
0
1
1
0.5 0.8
0.6
0.4
0.2
0 0
Linearizing the equations about the first solutions ũ, and ṽ, we obtain
∂ f1 (ũ, ṽ)
u + f1 (ũ, ṽ) + (u − ũ)
∂u
∂ f1 (ũ, ṽ)
+ (v − ṽ) + g1 = 0, (16.61)
∂v
∂ f2 (ũ, ṽ)
v + f2 (ũ, ṽ) + (u − ũ)
∂u
∂ f2 (ũ, ṽ)
+ (v − ṽ) + g2 = 0. (16.62)
∂v
In finite differences, we could write
∂ f1 (ũn−1 , ṽn−1 )
un − un−1 + f1 (ũn−1 , ṽn−1 )d + (un−1 − ũn−1)d
∂u
∂ f1 (ũn−1 , ṽn−1 )
+ (vn−1 − ṽn−1)d + (g1 )n−1 d = 0, (16.63)
∂v
∂ f2 (ũn−1 , ṽn−1 )
vn − vn−1 + f2 (ũn−1 , ṽn−1 )d + (un−1 − ũn−1)d
∂u
∂ f2 (ũn−1 , ṽn−1 )
+ (vn−1 − ṽn−1)d + (g2)n−1 d = 0. (16.64)
∂v
Hence, we may write
un = an un−1 + bnvn−1 + cn ,
vn = dn un−1 + en vn−1 + fn ,
where
∂ f1 (ũn−1 , ṽn−1 )
an = − d + 1,
∂u
∂ f1 (ũn−1 , ṽn−1 )
bn = − d,
∂v
16.7 Final Results, Newton’s Method for a First-Order System 417
∂ f1 (ũn−1 , ṽn−1 )
cn = − f1 (ũn−1 , ṽn−1 )d + ũn−1 d
∂u
∂ f1 (ũn−1 , ṽn−1 )
+ ṽn−1 d − (g1 )n−1 d, (16.65)
∂v
and
∂ f2 (ũn−1 , ṽn−1 )
dn = − d,
∂u
∂ f2 (ũn−1 , ṽn−1 )
en = − d + 1,
∂v
∂ f2 (ũn−1 , ṽn−1 )
fn = − f2 (ũn−1 , ṽn−1 )d + ũn−1d
∂u
∂ f2 (ũn−1 , ṽn−1 )
+ ṽn−1 d − (g2)n−1 d. (16.66)
∂v
In particular, for n = 1, we get
u 1 = a 1 u 0 + b 1 v0 + c1 , (16.67)
and
v1 = d 1 u 0 + e1 v0 + f 1 . (16.68)
From this last equation,
v0 = (v1 − d1 u0 − f1 )/e1 ,
so that from this and Eq. (16.67), we get
where
F1 = b1 /e1 , G1 = a1 u0 − b1 (d1 u0 + f1 )/e1 + c1.
Reasoning inductively, having
where
Hn = 1/(dn Fn−1 + en ),
Ln = −Hn (dn Gn−1 + fn ).
Hence
un = an (Fn−1 vn−1 + Gn−1) + bnvn−1 + cn−1,
so that
un = an (Fn−1 (Hn vn + Ln ) + Gn−1) + bn(Hn vn + Ln ) + cn−1,
and hence
Fn = an Fn−1 Hn + bn Hn ,
and
Gn = an (Fn−1 Ln + Gn−1) + bnLn + cn−1.
Thus,
un = Fn vn + Gn ,
so that, in particular,
uN = FN v f + GN ,
vN−1 = HN v f + LN ,
and hence
uN−1 = FN−1 vN−1 + GN−1 ,
vN−2 = HN−1 vN−1 + LN−1 ,
and so on, up to finding,
u1 = F1 v1 + G1 ,
and
v0 = H0 v1 + L0 ,
where H0 = 1/e1 and L0 = −(d1 u0 + f1 )/e1 .
The next step is to replace {ũn} and {ṽn } by {un } and {vn }, respectively, and then
to repeat the process up to the satisfaction of an appropriate convergence criterion.
where n(t) is the neutron population, C(t) is the concentration of delayed neutrons,
T (t) is the core temperature, ρ (T ) is the reactivity (which depends on the tempera-
ture T ), β is the delayed neutron fraction, L is the prompt reactors generation time,
λ is the average decay constant of the precursors, and H is the inverse of the reactor
thermal capacity.
For our numerical examples we consider T (0s) = 300 K and T (100 s) = T f =
350 K. Moreover we assume the relation
1 (β − ρ (0))
C(0) = n(0),
λ L
where n(0) is unknown (to be numerically calculated by our method such that we
have T (100 s) = T f ).
Also we consider
ρ (T ) = ρ (0) − α (T − T (0)).
The remaining values are β = 0.0065, L = 0.0001 s, λ = 0.00741 s−1 , H =
0.05 K/(MWs), α = 5 · 10−5 K−1 , and ρ (0) = 0.2β .
First we linearize the system in question about (ñ, T̃ ) obtaining (in fact it is a first
approximation)
ρ (T̃ ) − β ρ (T ) − β
n (t) = n(t) + ñ(t)
L L
ρ (T̃ ) − β
− ñ(t) + λ C(t), (16.70)
L
β
C (t) = n(t) − λ C(t),
L
T (t) = Hn(t),
where ρ (T ) = ρ (0) − α (T − T (0)).
Discretizing such a system in finite differences, we get
ρ (T̃i ) − β ρ (Ti ) − β
(ni+1 − ni )/d = ni + ñi
L L
ρ (T̃i ) − β
− ñi + λ Ci , (16.71)
L
β
(Ci+1 − Ci )/d = ni − λ Ci ,
L
(Ti+1 − Ti )/d = Hni ,
where d = 100 s/N, where N is the number of nodes.
Hence, we may write
ni+1 = ai ni + bi Ti + diCi + ei , (16.72)
where
ρ (T̃i ) − β
ai = 1 + d,
L
−α
bi = ñi d,
L
di = λ d,
(ρ (0) + α T (0) − β ) (ρ (T̃i ) − β )
ei = ñi d − ñi d,
L L
β
f = d,
L
g = 1 − λ d,
h = 1,
m = Hd.
Observe that
C0 = α̃ n0 ,
where
β − ρ (0)
α̃ = .
Lλ
For i = 0 from (16.74) we obtain
T1 − hT0
n0 = = α1 T1 + β1, (16.75)
m
where α1 = 1/m and β1 = −(h/m)T0 .
Therefore,
C0 = α̃ n0 = α̃ (α1 T1 + β1).
Still for i = 0, replacing this last relation and (16.75) into (16.72), we get
n1 = a0 (α1 T1 + β1 ) + b0T0 + d0α̃ (α1 T1 + β1) + e0 ,
so that
n1 = α̃1 T1 + β̃1, (16.76)
where
α̃1 = a0 α0 + d0α̃α1 ,
and
β̃1 = a0 β1 + b0 T0 + d0α̃β1 + e0 .
Finally, from (16.73),
C1 = f (α1 T1 + β1) + gα̃ (α1 T1 + β1)
= α̂1 T1 + β̂1 , (16.77)
16.7 Final Results, Newton’s Method for a First-Order System 421
where
α̂1 = f α1 + gα̃α1 ,
and
β̂1 = f β1 + gα̃β1 .
Reasoning inductively, having
ni−1 = αi Ti + βi , (16.79)
Ci = α̂i Ti + β̂i , (16.80)
we are going to obtain the corresponding relations for i + 1, i ≥ 1. From (16.74)
and (16.78) we obtain
Ti+1 = hTi + m(α̃i Ti + β̃i),
so that
Ti = ηi Ti+1 + ξi , (16.81)
where
ηi = (h + mα̃i )−1 ,
and
ξi = −(mβ̃i )ηi .
On the other hand, from (16.72), (16.78), and (16.80), we have
where
αi+1 = α̃i ηi ,
and
βi+1 = α̃i ξi + β̃i .
422 16 More on Duality and Computation for the Ginzburg–Landau System
Moreover,
Ci+1 = f ni + gCi
= f (αi+1 Ti+1 + βi+1)
+g(α̂i Ti + β̂i )
= f (αi+1 Ti+1 + βi+1)
+g(α̂i (ηi Ti + ξi ) + β̂i )
= α̂i+1 Ti+1 + β̂i+1, (16.82)
where
α̂i+1 = f αi+1 + gα̂i ηi ,
and
β̂i = f βi+1 + gα̂i ξi + gβ̂i.
Summarizing, we have obtained linear functions (F0 )i , (F1 )i , and (F2 )i such that
Ti = (F0 )i (Ti+1 ),
ni = (F1 )i (Ti+1 ),
Ci = (F2 )i (Ti+1 ),
∀i ∈ {1, . . . , N − 1}.
Thus, considering the known value TN = T f , we obtain
T1 = (F0 )1 (T2 ),
n1 = (F1 )1 (T2 ),
C1 = (F2 )1 (T2 ),
and n0 = (F0 )1 (T1 ).
16.8 Conclusion 423
The next step is to replace (ñ, T̃ ) by the last calculated (n, T ) and then to repeat
the process until an appropriate convergence criterion is satisfied.
Concerning our numerical results through such a method, for the solution n(t)
obtained, please see Fig. 16.11. For the solution T (t), see Fig. 16.12.
12
11
10
6
0 20 40 60 80 100
We emphasize the numerical results here obtained are consistent with the current
literature (see [61] for details).
16.8 Conclusion
In this chapter, first we have presented a duality principle for the Ginzburg–
Landau system in the presence of a magnetic field. We highlight to have obtained
sufficient conditions of optimality, similarly to the canonical duality procedure int-
roduced by Gao [36]. It is worth mentioning the dual formulation is concave and
amenable to numerical computation.
In a second step, we have introduced the matrix version of the generalized
method of lines. Also we develop a convergent algorithm suitable for equations
that present strong variational formulation and, in particular, suitable for Ginzburg–
Landau-type equations. The results are rigorously proven and numerical examples
are provided. We emphasize that even as the parameter ε is very small, namely,
ε = 0.0001, the results are consistent and the convergence is very fast. Finally, in the
424 16 More on Duality and Computation for the Ginzburg–Landau System
350
345
340
335
330
325
320
315
310
305
300
0 20 40 60 80 100
last section, we develop in details Newton’s method combined with the generalized
method of lines main idea, with numerical results relating an example in nuclear
physics.
Chapter 17
On Duality Principles for Scalar and Vectorial
Multi-well Variational Problems
17.1 Introduction
In this chapter, our first objective is the establishment of a duality principle suitable
for the variational form of some nonlinear vectorial problems in physics and
engineering. The results are based on standard tools of convex analysis. As a first
example we apply them to a phase transition model, which may be found in a similar
format in Chenchiah and Bhattacharya [18]. It is relevant to observe that the study
developed in [18] is restricted to the two-well problem, whereas our new duality
principle is applicable to vectorial multi-well formulations in general, not restricted
to two or three wells.
In Sect. 17.4 we discuss how the standard tools of convex analysis can be used
to study the scalar case.
In Sect. 17.3 we present the main theorem in this chapter, namely Theorem 17.3.1,
which corresponds, as mentioned above, to a new duality principle. It is important
to emphasize that this principle stands for relaxation for a vectorial phase transition
problem. In the next lines, we describe such a result.
Consider (G ◦ Λ ) : U → R and (F ◦ Λ ) : U → R, F being a convex Gâteaux
differentiable functional such that J : U → R defined as
J(u) = (G ◦ Λ )(u) − (F ◦ Λ )(u) − u, f U
is bounded below. Here Λ : U → Y is a continuous linear injective operator whose
the respective adjoint is denoted by Λ ∗ : Y ∗ → U ∗ . Under such assumptions, we
have
inf {G∗∗ (Λ u) − F(Λ u) − u, f U }
u∈U
∗ ∗ ∗ ∗ ∗ ∗
≥ sup ∗inf ∗ {(F ◦ Λ ) (Λ z ) − G (v + z )} .
v∗ ∈A∗ z ∈Y
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 425
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 17,
© Springer International Publishing Switzerland 2014
426 17 On Duality Principles for Scalar and Vectorial Multi-well Variational Problems
where
A∗ = {v∗ ∈ Y ∗ | Λ ∗ v∗ = f }.
Furthermore, under additional assumptions to be specified, we have
inf {G∗∗ (Λ u) − F(Λ u) − u, f U }
u∈U
= sup ∗inf ∗ {(F ◦ Λ )∗ (Λ ∗ z∗ ) − G∗(v∗ + z∗ )} .
v∗ ∈A∗ z ∈Y
Now we also present a summary of our main applied result, namely, a duality
principle applied to a vectorial phase transition problem.
For an open bounded connected set S ⊂ R3 with a regular boundary denoted by
∂ S, consider the functional J : U → R, where
1
J(u) = min {gk (ε (u)) + βk } dx − u, f U ,
2 S k∈{1,...,N}
1
εi j (u) = (ui, j + u j,i ), for i, j ∈ {1, 2, 3}.
2
Furthermore {Cikjlm } are positive definite matrices and βk ∈ R for each k ∈
{1, . . . , N}, and f ∈ L2 (S; R3 ) is an external load. Here ek ∈ R3×3 for k ∈ {1, . . . , N}
represent the stress-free configurations or phases presented by a solid with field of
displacements u = (u1 , u2 , u3 ) ∈ W 1,2 (S; R3 ) (due to f ). Also
where
1 K
G(ε (u)) = min {gk (ε (u)) + βk } dx + (εi j (u))Hi jlm (εlm (u)) dx,
2 S k∈{1,...,N} 2 S
K
F(ε (u)) = (εi j (u))Hi jlm (εlm (u)) dx,
2 S
17.2 Preliminaries 427
{Hi jlm } is a positive definite matrix (the identity for example) and K > 0 is an
appropriate constant. The final duality principle is expressed by
Finally,
A∗ = {v∗ ∈ Y ∗ | v∗i j, j + fi = 0 in S},
and
17.2 Preliminaries
We denote by U and Y Banach spaces which the topological dual spaces are
identified with U ∗ and Y ∗ , respectively. Unless otherwise indicated, Y is assumed to
be reflexive. The canonical duality pairing between U and U ∗ is denoted by ·, ·U :
U × U ∗ → R, through which the linear continuous functionals defined on U are
represented.
Given F : U → R̄ = R ∪ {+∞} its polar F ∗ : U ∗ → R̄ is defined as
u, Λ ∗ v∗ U = Λ u, v∗ Y , ∀u ∈ U, v∗ ∈ Y ∗ .
Finally, the following duality principle found in [25] will be used in this text.
Theorem 17.2.2. Let G : Y → R̄ = R ∪ {+∞} and F : U → R be two convex l.s.c.
(lower semicontinuous) functionals so that J : U → R̄ defined as
and there exists at least one v∗0 ∈ Y ∗ which maximizes the dual formulation. If in
addition U is reflexive and
lim J(u) = +∞
u →+∞
then both primal and dual formulations have global extremals so that there exist
u0 ∈ U and v∗0 ∈ Y ∗ such that
Also
G(Λ u0 ) + G∗(v∗0 ) = Λ u0 , v∗0 Y ,
F(u0 ) + F ∗ (−Λ ∗ v∗0 ) = u0 , −Λ ∗ v∗0 U ,
so that
G(Λ u0 ) + F(u0 ) = −G∗ (v∗0 ) − F ∗ (−Λ ∗ v∗0 ).
Also fundamental for the construction of the main duality principle is a result found
in Toland [67] (despite we have not used it directly we have followed a similar idea)
which is as follows.
Theorem 17.2.3. Consider the functionals F, G : U → R through which we define
J : U → R as
J(u) = G(u) − F(u). (17.2)
Suppose there exists u0 ∈ U such that
J(u0 ) = inf {J(u)} (17.3)
u∈U
and ∂ F(u0 ) = 0.
/
17.3 The Main Duality Principle 429
where
A∗ = {v∗ ∈ Y ∗ | Λ ∗ v∗ = f }.
Furthermore, assuming that G∗ : Y ∗ → R is Lipschitz continuous, there exists
v∗0 ∈ Y ∗ such that
Jˆ∗ (v∗0 ) = max
∗ ∗
{Jˆ∗ (v∗ )},
v ∈Y
where
J ∗ (v∗ ) = ∗inf ∗ {(F ◦ Λ )∗ (Λ ∗ z∗ ) − G∗ (v∗ + z∗ )} ,
z ∈Y
0, if Λ ∗ v∗ + f = 0,
Ind(v∗ ) =
+∞, otherwise,
and
Jˆ∗ (v∗ ) = J ∗ (v∗ ) − Ind(v∗).
In addition we suppose that defining
we have that
J1∗ (v∗0 , z∗ ) → +∞
as
z∗ Y ∗ → +∞, or Λ ∗ z∗ L2 → +∞.
Furthermore, suppose that if {z∗n } ⊂ Y ∗ is such that Λ ∗ z∗n L2 < K, ∀n ∈ N for some
K > 0, then there exists z̃∗ ∈ Y ∗ such that for a not relabeled subsequence, we have
and
z∗n → z̃∗ , strongly in Y ∗ .
Under such additional assumptions, there exist z∗0 ∈ Y ∗ and u0 ∈ U such that
Since Y ∗ and L2 are reflexive Banach spaces, there exist v∗0 ∈ Y ∗ and ṽ∗0 ∈ L2 , such
that up to a not relabeled subsequence we have
v∗n v∗0 , weakly in Y ∗ ,
and
Λ ∗ v∗n ṽ∗0 , weakly in L2 .
Observe that given ϕ ∈ Cc∞ we have
Thus
ṽ∗0 = Λ ∗ v∗0 ,
in distributional sense, so that
Λ ∗ v∗n Λ ∗ v∗0 , weakly in L2 .
Hence, as Jˆ∗ (v∗ ) is concave and strongly continuous, it is also upper semicontinu-
ous, so that
lim sup{Jˆ∗ (v∗n )} ≤ Jˆ∗ (v∗0 ).
n→∞
where
J1∗ (v∗0 , z∗ ) = (F ◦ Λ )∗ (Λ ∗ z∗ ) − G∗ (v∗0 + z∗ ).
432 17 On Duality Principles for Scalar and Vectorial Multi-well Variational Problems
From the coercivity hypothesis, if {z∗n } is a minimizing sequence, there exists K1 > 0
such that
z∗n Y ∗ < K1 and Λ ∗ z∗n L2 < K1 , ∀n ∈ N.
Also from the hypothesis, up to a not relabeled subsequence, there exist z∗0 ∈ Y ∗
such that
Λ z∗n Λ ∗ z∗0 , weakly in L2 ,
and
z∗n → z∗0 , strongly in Y ∗ .
As G∗ (v∗ ) is strongly continuous, we obtain
Therefore
and
G∗ (v∗0 + z∗0 ) = Λ u0 , v∗0 Y + Λ u0 , z∗0 Y − G∗∗(Λ u0 ). (17.12)
From the fact that v∗0 ∈ A∗ , we have
Λ ∗ v∗0 = f .
Note that from the last three equations, we obtain
(F ◦ Λ )∗ (Λ ∗ z∗0 ) − G∗ (v∗0 + z∗0 )
= G∗∗ (Λ u0 ) − F(Λ u0 ) − u0, f U . (17.13)
Therefore, we may conclude that
inf {G∗∗ (Λ u) − F(Λ u) − u, f U }
u∈U
= G∗∗ (Λ u0 ) − F(Λ u0 ) − u0 , f U
= (F ◦ Λ )∗ (Λ ∗ z∗0 ) − G∗(v∗0 + z∗0 )
∗ ∗ ∗ ∗ ∗ ∗
= sup ∗inf ∗ {(F ◦ Λ ) (Λ z ) − G (v + z )} .
v̂∗ ∈A∗ z ∈Y
This section is dedicated to the analysis of the scalar multi-well problem via
duality.
and
G(∇u) = min {gi (∇u)}dS = g(∇u)dS. (17.15)
S i∈{1,...,N} S
We also assume
G(∇u)
→ +∞ as u U → ∞, (17.16)
u U
434 17 On Duality Principles for Scalar and Vectorial Multi-well Variational Problems
and
U = {u ∈ W 1,2 (S) | u = u0 on Γ }. (17.17)
As a preliminary result, we present Corollary 3.8, at p. 339 of Ekeland and Temam
[25] (here Ω stands for S).
Theorem 17.4.1. Let f be a Carathéodory function from Ω × (R × Rn ) into R which
satisfies
a2 (x) + c2 |ξ |α ≤ f (x, s, ξ ) ≤ a1 (x) + b|s|α + c1 |ξ |α (17.18)
where a1 , a2 ∈ L1 (Ω ), 1 < α < +∞, b ≥ 0 and c1 ≥ c2 > 0. Let u0 ∈ W 1,α (Ω ).
Under such assumptions, defining Û = {u | u − u0 ∈ W01,2 (Ω )}, we have
∗∗
inf f (x, u, ∇u)dx = min f (x, u; ∇u) dx (17.19)
u∈Û Ω u∈Û Ω
The solutions of relaxed problem are weak cluster points in W 1,α (Ω ) of the min-
imizing sequences of primal problem.
Now we can enunciate the following result.
Theorem 17.4.2. Consider the definition and assumptions about (G ◦ ∇) : U → R
indicated in (17.14), (17.15), and (17.16). Also assuming the hypothesis of
Theorem 17.4.1, we have
inf {G(∇u) − u, f L2 (S) } = inf {G∗∗ (∇u) − u, f L2 (S) } (17.20)
u∈U u∈U
The proof follows directly from Theorem 17.4.1. Our next proposition is very im-
portant to establish the subsequent results. It is simple so that we do not prove it.
Now we present the main duality principle for the scalar case.
we have
where
G∗ (v∗ ) = max {g∗i (v∗ )}dS (17.27)
S i∈{1,...,N}
and
C∗ = {v∗ ∈ Y ∗ | div(v∗ ) + f (x) = 0, in S}. (17.28)
that is,
or
inf {G∗∗ (∇u) − u, f L2 (S) } ≥ sup {−G∗ (v∗ ) + u0, v∗ · nL2 (Γ ) }. (17.31)
u∈U v∗ ∈C∗
The equality in (17.31) follows from the hypothesis indicated in (17.16) and
Theorem 17.2.2.
Observe that the dual formulation is convex but non-smooth. It is through the
points of non-smoothness that the microstructure is formed, specially when the orig-
inal primal formulation has no minimizers in the classical sense.
To start this section, we present duality for the solution of a standard scalar
multi-well problem. Consider an open bounded connected set S ⊂ R3 , with a regular
boundary Γ , and the function (W ◦ ∇) defined as
436 17 On Duality Principles for Scalar and Vectorial Multi-well Variational Problems
1
W (∇u) = min |∇u − ai|2 (17.32)
i∈{1,...,N} 2
where
U = {u ∈ W 1,2 (S) | u = u0 on Γ } (17.33)
ai are known matrices, for all i ∈ {1, . . . , N}, so that the energy of the system is
modeled by J : U → R, where
J(u) = W (∇u) dx − u, f L2 (S) (17.34)
S
or
" #
1
J(u) = min |∇u − ai|2 dx − u, f L2 (S) . (17.35)
2 S i∈{1,...,N}
inf {J(u)}
u∈U
1 ∗2 N
= sup inf − |v | + ∑ λi v ai dx + u0 , v · nL2 (Γ )
∗T ∗
v∗ ∈C∗ λ ∈B S 2 i=1
where
B = {λ = (λ1 , . . . , λN ) measurable |
λi (x) ∈ [0, 1], ∀i ∈ {1, . . . , N}
N
and ∑ λi (x) = 1}, (17.37)
i=1
and
C∗ = {v∗ ∈ Y ∗ | div(v∗ ) + f = 0, in S} (17.38)
It is important to emphasize that, in general, this kind of problem does not present
minimizers in the classical sense. The solution of the dual problem (which is well
posed and convex) reflects the average behavior of minimizing sequences as weak
cluster points (of such sequences).
17.5 Duality for a Vectorial Multi-well Model Applicable to Phase Transition Problems 437
In this section we consider duality for another class of multi-well problems simi-
lar as those found in [18]. However, here the format of our problem is more general,
not restricted to two-well formulations. Observe that the relaxation for the case of
three or more wells was so far an open question in the current literature. Our main
result is summarized by the next theorem.
1
εi j (u) = (ui, j + u j,i ), for i, j ∈ {1, 2, 3}.
2
Furthermore {Cikjlm } are positive definite matrices and βk ∈ R for each k ∈
{1, . . . , N}, and f ∈ L2 (S; R3 ) is a external load. Here ek ∈ R3×3 for k ∈ {1, . . . , N}
represent the stress-free configurations or phases presented by a solid with field of
displacements u = (u1 , u2 , u3 ) ∈ W 1,2 (S; R3 ) (due to f ). Also
We may write
J(u) = G(ε (u)) − F(ε (u)) − u, f U
where
1 K
G(ε (u)) = min {gk (ε (u)) + βk } dx + (εi j (u))Hi jlm (εlm (u)) dx,
2 S k∈{1,...,N} 2 S
K
F(ε (u)) = (εi j (u))Hi jlm (εlm (u)) dx,
2 S
{Hi jlm } is a positive definite matrix and K > 0. Under such assumptions, observing
that
∗ ∗ ∗ 1 ∗
G (v + z ) = max (vi j + z∗i j )Dkijlm (v∗lm + z∗lm )
S k∈{1,...,N} 2
0
+(v∗i j + z∗i j )Ĉikjlm eklm − βk dx, (17.39)
438 17 On Duality Principles for Scalar and Vectorial Multi-well Variational Problems
and
1
F ∗ (z∗ ) = z∗i j Ĥi jlm z∗lm dx,
2K S
we have
∗∗ ∗ ∗ ∗ ∗ ∗
inf {G (ε (u)) − F(ε (u)) − u, f U } ≥ sup inf {F (z ) − G (z + v )} .
u∈U v∗ ∈A∗ z∗ ∈Y ∗
where
{Ĥi jlm } = {Hi jlm }−1 ,
{Dkijlm } = {Cikjlm + KHi jlm }−1 ,
and
{Ĉikjlm } = {Cikjlm + KHi jlm }−1 {Cikjlm }.
Moreover, z∗ (v∗ ,t) is obtained through Eq. (17.44), that is,
N N
1
Ĥi jlm z∗lm − ∑ {tk Dkijlm (v∗lm + z∗lm )} − ∑ tkĈikjlm eklm = 0, in S.
K k=1 k=1
Also,
A∗ = {v∗ ∈ Y ∗ | v∗i j, j + fi = 0 in S},
and
Finally, assuming the hypotheses of the main duality principle, there exist u0 ∈ U
and (v∗0 , z∗0 ) ∈ Ŷ ∗ such that
∀u ∈ U, v∗ ∈ A∗ . Therefore
∗∗ ∗ ∗ ∗ ∗ ∗
inf {G (ε (u)) − F(ε (u)) − u, f U } ≥ sup inf {F (z ) − G (z + v )} ,
u∈U v∗ ∈A∗ z∗ ∈Y ∗
where
1 ∗
G∗ (v∗ + z∗ ) = max (v + z∗i j )Dkijlm (v∗lm + z∗lm )
S k∈{1,...,N} 2 ij
0
+(v∗i j + z∗i j )Ĉikjlm eklm − βk dx, (17.42)
so that
where
where z∗ ∈ Ỹ if z∗ ∈ Y ∗ and
∂ F(ε (u))
z∗ = ,
∂v
17.5 Duality for a Vectorial Multi-well Model Applicable to Phase Transition Problems 441
for some u ∈ U. We also recall that Ỹ may be defined through linear equations
of compatibility analogously to those of linear elasticity, considering that in the
present case F is a quadratic functional. We do not work these elementary details
here, postponing a more extensive analysis for a future work.
where
1
G(u ) = min{g1 (u ), g2 (u )} dx,
0
A A
g1 (u ) = (u − c1 )2 and g2 (u ) = (u − c2)2 ,
2 2
and
U = W01,2 ([0, 1]).
From Theorem 17.4.4, denoting Y ∗ = L2 ([0, 1]), we obtain
where
1
G∗ (v∗ ) = max{g∗1 (v∗ ), g∗2 (v∗ )} dx,
0
1 ∗ 2
g∗1 (v∗ ) = (v ) + c1 v∗
2A
and
1 ∗ 2
g∗2 (v∗ ) = (v ) + c2 v∗ .
2A
Also
C∗ = {v∗ ∈ Y ∗ | (v∗ ) + f = 0 in [0, 1]}.
In this case there is no duality gap between the primal and dual problems. Anyway,
let us analyze the dual problem obtained as we redefine the primal formulation as
indicated in the next lines.
Define
K 1 2
Ĝ(u ) = G(u ) + (u ) dx,
2 0
and
K 1 2
F(u ) = (u ) dx.
2 0
442 17 On Duality Principles for Scalar and Vectorial Multi-well Variational Problems
where
1
∗ ∗
Ĝ (v ) = max{ĝ∗1 (v∗ ), ĝ∗2 (v∗ )} dx,
0
1 Ac1 ∗ AKc21
ĝ∗1 (v∗ ) = (v∗ )2 + v −
2(A + K) A+K 2(A + K)
and
1 Ac2 ∗ AKc22
ĝ∗2 (v∗ ) = (v∗ )2 + v − .
2(A + K) A+K 2(A + K)
Also
1
1
F ∗ (z∗ ) = (z∗ )2 dx.
2K 0
Now defining
JK∗ (v∗ ) = ∗inf ∗ {F ∗ (z∗ ) − Ĝ∗(z∗ + v∗ )},
z ∈Y
we may write
inf {J(u)} ≥ sup {JK∗ (v∗ )},
u∈U v∗ ∈C∗
and
1
JK∗ (v∗ ) = inf{ ∗inf ∗ {F ∗ (z∗ ) − (t ĝ∗1 (v∗ + z∗ ) + (1 − t)ĝ∗2(v∗ + z∗ )) dx}},
t∈B z ∈Y 0
that is,
1 1
1
JK∗ (v∗ ) = inf{ ∗inf ∗ { (z∗ )2 dx − (t ĝ∗1 (v∗ + z∗ ) + (1 − t)ĝ∗2(v∗ + z∗ )) dx}},
t∈B z ∈Y 2K 0 0
where
B = {t measurable | t ∈ [0, 1] a.e. in [0, 1]}.
Evaluating the infimum in z∗ we obtain the final expression for JK∗ , namely,
1
JK∗ (v∗ ) = inf{ (a · (c1 − c2)2t − a · (c1 − c2)2 · t 2
t∈B 0
1
−tc1 v∗ − (1 − t)c2v∗ − (v∗ )2 ) dx}. (17.46)
2 · 10
For different values of K we have different values of a, for example:
1. for K = 5, we have a = 1.66667,
2. for K = 50, we have a = 4.166667,
17.5 Duality for a Vectorial Multi-well Model Applicable to Phase Transition Problems 443
Considering the vectorial case in question, for the analogous final value of a
(in this case a = 5), the difference observed through (17.47) will result in no duality
gap between the primal and dual problems. The difference is noted for the intermedi-
ate values of t, that is, for 0 < t < 1, and it is particularly relevant in a microstructural
context.
Finally, denoting α = a · (c1 − c2 )2 , through a Lagrange multiplier λ , we may
write
1
∗ ∗ (α − λ ) c1 + c2 ∗
JK (v ) = sup { − v } dx
λ ∈C 0 4 2
1 1
(c1 − c2 )2 (v∗ )2 1
− dx − (v∗ )2 dx ,
0 4(λ − α ) 2 · 10 0
where
C = {λ ∈ Y ∗ | λ − α > 0, a.e. in [0, 1]}.
Thus the final expression of the duality principle would be
1
(α − λ ) c1 + c2 ∗
inf {J(u)} ≥ sup { − v } dx
u∈U (v∗ ,λ )∈C∗ ×C 0 4 2
1 1
(c1 − c2 )2 (v∗ )2 1 ∗ 2
− dx − (v ) dx .
0 4(λ − α ) 2 · 10 0
Our final result is concerned with the evaluation of duality gap between the
primal and dual problems.
444 17 On Duality Principles for Scalar and Vectorial Multi-well Variational Problems
From the last theorem we may select a sequence {u0K } ⊂ U such that
∗∗ ∗ ∗ ∗ ∗ ∗ ∗
JK (u0K ) = sup ∗inf ∗ {(FK ◦ ε ) (ε z ) − GK (v + z )} (17.47)
v∗ ∈A∗ z ∈Y
∀K ∈ N.
Suppose there exists K̃ such that
and also
J(u0 ) − sup inf {(F ◦ ε )∗ ∗ ∗
(ε z ) − G ∗ ∗
(v + z∗
) < ε.
K ∗ ∗
K
z ∈Y
K
v∗ ∈A∗
we have
GK (ε (u0K )) = (G̃K )(ε (u0K )), ∀K ∈ N, (17.48)
17.5 Duality for a Vectorial Multi-well Model Applicable to Phase Transition Problems 445
and
G∗∗ ∗∗
K (ε (u0K )) = (G̃K ) (ε (u0K )), ∀K ∈ N, (17.49)
(this is possible since g is the minimum of N quadratic positive definite functions
and ε (u0K ) ∞ is uniformly bounded), where
1 K
(G̃K )(ε (u)) = g̃(ε (u)) dx + (εi j (u))Hi jlm (εlm (u)) dx,
2 S 2 S
Define
gδ (y), if |y| < K1 ,
g̃δ (y) =
+∞, otherwise.
Observe that from (17.50)
ε
|G̃K (ε (u0K )) − (G̃K )δ (ε (u0K ))| < ,
2
and
ε
|G̃∗∗ ∗∗
K (ε (u0K )) − (G̃K )δ (ε (u0K ))| < , ∀K ∈ N,
2
where
1 K
(G̃K )δ (ε (u)) = g̃δ (ε (u)) dx + (εi j (u))Hi jlm (εlm (u)) dx.
2 S 2 S
if K > Kε .
The proof is complete.
Remark 17.5.5. Through this final result we have shown that for the problem in
question, the duality gap between the primal and dual formulations becomes arbi-
trarily small as K goes to ∞.
17.6 Conclusion
18.1 Introduction
In this chapter we develop dual variational formulations for a more general class
of non-convex multi-well variational models. Such models appear in similar form
in phase transition and related problems. Please see [5, 7, 18–20, 28, 48, 64] for
details. We also address problems of conductivity in composites and optimal design
and control in elasticity.
In this section we state and prove the main result in this chapter, which is sum-
marized by the next theorem. From now on, by a regular boundary ∂ Ω of Ω ⊂ R3 ,
we mean regularity enough so that the standard Gauss–Green formulas of integra-
tions by parts, the Sobolev imbedding theorem, and the trace theorem to hold. Also,
n denotes the outward normal to ∂ Ω and derivatives must be understood always in
the distributional sense.
Remark 18.2.1. At some points of our analysis we refer to the problems in question
after discretization. In such a case, we are referring to their approximations in a
finite element or finite differences context. Finally, to simplify the notation in some
parts of the text, we may denote L2 (Ω ), L2 (Ω ; R3 ) and L2 (Ω ; R3×3 ) simply by L2 .
where
1
G(ε (u),t) = tk gk (ε (u)) dx,
2 Ω
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 447
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 18,
© Springer International Publishing Switzerland 2014
448 18 More on Duality Principles for Multi-well Problems
gk (ε (u)) = Hikjlm (εi j (u) − ekij )(εlm (u) − eklm ) + βk , ∀k ∈ {1, . . . , N}.
Here
1
ε (u) = {εi j (u)} = (ui, j + u j,i ) ,
2
and
U = {u ∈ W 1,2 (Ω ; R3 ) | u = u0 on ∂ Ω }.
Also, βk ∈ R, {ekij } ∈ L2 (Ω ; R3×3 ), f ∈ L2 (Ω ; R3 ) and
{Hikjlm }
are fourth-order positive definite constant tensors ∀k ∈ {1, . . . , N}.
Under such hypotheses,
where
J ∗ (v∗ ) = sup{G∗ (v∗ ,t)} − v∗i j n j , (u0 )i L2 (Γ ) ,
t∈B
and
A∗ = {v∗ ∈ Y ∗ | v∗i j, j + fi = 0 in Ω }.
Furthermore, there exists v∗0 ∈ A∗ ⊂ Y ∗ = Y = L2 (Ω ; R3×3 ) such that
−J ∗ (v∗0 ) = max
∗ ∗
{−J ∗ (v∗ )}.
v ∈A
Finally, for the discretized version of the problem, assume t0 ∈ B such that
G∗ (v∗0 ,t0 ) = sup{G∗ (v∗0 ,t)},
t∈B
∀(u,t) ∈ U × B, v∗ ∈ A∗ .
Thus,
inf {J(u,t)} ≥ sup {−J ∗ (v∗ )}. (18.3)
(u,t)∈U×B v∗ ∈A∗
At this point and on, we consider the discretized version of the problem.
From the hypotheses, t0 ∈ B such that
G∗ (v∗0 ,t0 ) = sup{G∗ (v∗0 ,t)},
t∈B
stands for
∂ G∗ (v∗0 ,t0 )
ε (û) = .
∂ v∗
Hence, since for a fixed t0 ∈ B G(ε (u),t0 ) is convex in u, we get
G∗ (v∗0 ,t0 ) = ε (û), v∗0 L2 − G(ε (û),t0 )
= −ûi , (v0 )∗i j, j L2 (Ω ) + (v0 )∗i j n j , (u0 )i L2 (Γ )
−G(ε (û),t0 ). (18.4)
450 18 More on Duality Principles for Multi-well Problems
Therefore,
G∗ (v∗0 ,t0 ) − (v0 )∗i j n j , (u0 )i L2 (Γ )
= −G(ε (û),t0 ) + ûi, fi L2 (Ω ) . (18.5)
From this we get
J ∗ (v∗0 ) = −J(û,t0 ).
From this last equation and (18.3), the proof is complete.
In this section we state and prove another relevant result, which is summarized
by the next theorem.
Theorem 18.3.1. Let Ω ⊂ R3 be an open, bounded, connected set with a regular
boundary denoted by ∂ Ω = Γ . Denote J : U × B → R by
J(u,t) = G(ε (u),t),
where
1
G(ε (u),t) = tk gk (ε (u)) dx,
2 Ω
gk (ε (u)) = Hikjlm (εi j (u) − ekij )(εlm (u) − eklm ) + βk , ∀k ∈ {1, . . . , N}.
Here
1
ε (u) = {εi j (u)} = (ui, j + u j,i ) ,
2
and
U = {u ∈ W 1,2 (Ω ; R3 ) | u = u0 on ∂ Ω }.
Also, βk ∈ R, {ekij } ∈ L2 (Ω ; R3×3 ) ≡ L2 , and as above,
{Hikjlm }
are fourth-order positive definite constant tensors, ∀k ∈ {1, . . . , N}.
Under such hypotheses,
where
For the primal formulation we repeat the statements found in reference [27].
Consider a material confined into a bounded domain Ω ⊂ RN , N > 1. The medium
is obtained by mixing two constituents with different electric permittivity and
452 18 More on Duality Principles for Multi-well Problems
subject to
div[χ L1 ∇u + (1 − χ )L0∇u] + f (x) = 0 (18.14)
and
χ dx ≤ t1 |Ω |,
Ω
and
0 < t1 < 1.
Also,
" #
1
G(∇u,t) = t(∇uT Q1 ∇u) + (1 − t)(∇uT Q0 ∇u) dx,
2 Ω
0 if (u,t) ∈ A
Ind(u,t) = (18.16)
+∞ otherwise,
where
A = {(u,t) ∈ U × B |
div((tL1 + (1 − t)L0)∇u) + f = 0 in Ω }. (18.17)
Here,
and
A∗ = {v∗ ∈ Y ∗ | div(v∗ ) = 0 in Ω }.
Moreover, there exists (v∗0 , λ0 ) ∈ A∗ × U1 such that
where U1 = W01,2 (Ω ).
Assume, after discretization, that t0 ∈ B such that
and
θ ∈ ∂λ J ∗ (v∗0 , λ0 ),
stand for
∂ G∗ (v∗0 − (t0 L1 + (1 − t0)L0 )∇λ0 )
∇û = ,
∂ v∗
and
∂ G∗ (v∗0 − (t0 L1 + (1 − t0)L0 )∇λ0 )
θ = div (t0 L1 + (1 − t0)L0 ) + f.
∂ v∗
Thus,
inf {J(u,t)} ≥ sup {−J ∗(v∗ , λ )}. (18.21)
(u,t)∈U×B (v∗ ,λ )∈A∗ ×U1
18.4 Duality for a Problem on Conductivity in Composites 455
From the remaining hypotheses, after discretization, we have that t0 ∈ B such that
and
θ ∈ ∂λ J ∗ (v∗0 , λ0 ),
stand for
∂ G∗ (v∗0 − (t0 L1 + (1 − t0)L0 )∇λ0 )
∇û = ,
∂ v∗
and
∂ G∗ (v∗0 − (t0 L1 + (1 − t0)L0 )∇λ0 )
θ = div (t0 L1 + (1 − t0)L0 ) + f.
∂ v∗
Hence
div((t0 L1 + (1 − t0)L0 )∇û) + f = θ , in Ω ,
and
Therefore,
so that
J(û,t0 ) = −J ∗ (v∗0 , λ0 ).
From this and (18.21), the proof is complete.
456 18 More on Duality Principles for Multi-well Problems
In this section, we develop duality for the optimal control of a two-phase plate
model. The control variable is t1 and is related to the elastic constant distribution
K(t1 ) given by
K(t1 ) = t1 K1 + (1 − t1)K2
where K1 % K2 > 0. Moreover, the plate stiffness is given by the tensor Hαβ λ μ (t) as
specified in the next theorem. We are concerned with the calculation of the optimal
t,t1 which minimizes the plate compliance (or, equivalently, its inner work), under
appropriate constraints. The plate model in question is the Kirchhoff one, where
Ω ⊂ R2 denotes the plate mid-surface. Moreover, w ∈ W 1,2 (Ω ) denotes the field of
displacements resulting from a vertical load f ∈ L2 (Ω ) action. Please see the next
theorem for details.
where U = W02,2 (Ω ),
0 < t˜ < 1,
and
0 < t˜1 < 1.
Also, Λ :→ Y = Y ∗ = L2 (Ω ; R2×2 ) ≡ L2 is given by
Λ w = {w,αβ },
1
G1 (Λ w,t) = Hαβ λ μ (t)w,αβ w,λ μ dx,
2 Ω
18.5 Optimal Design and Control for a Plate Model 457
1
G2 (w,t1 ) = K(t1 )w2 dx,
2 Ω
Hαβ λ μ (t) = t(H0 )αβ λ μ + (1 − t)(H1)αβ λ μ ,
where {(H0 )αβ λ μ } and {(H1 )αβ λ μ } are fourth-order positive definite constant
tensors. Also,
K(t1 ) = t1 K1 + (1 − t1)K2 ,
where
K1 % K2 > 0.
Moreover,
0 if (w,t,t1 ) ∈ A
Ind(w,t,t1 ) = (18.26)
+∞ otherwise,
where
A = {(w,t,t1 ) ∈ U × B × B1 |
(Hαβ λ μ (t)w,λ μ ),αβ + K(t1 )w − f = 0 in Ω }. (18.27)
Here,
J ∗ (v∗ ,t,t1 ) = G∗1 (v∗ ,t) + G∗2 (Λ ∗ v∗ − f ,t1 ),
where
K(t) = {K(t)}−1 .
Furthermore, also under the mentioned hypotheses, we have
where
ˆ ŵ) =
J( sup {G1 (Λ ŵ,t) + G2 (ŵ,t1 ) − ŵ, f L2 }.
(t,t1 )∈B×B1
stands for
δŵ {G1 (Λ ŵ0 ,t0 ) + G2 (ŵ0 , (t1 )0 ) − ŵ0 , f L2 } = θ .
Under such hypotheses
Moreover,
= −J(
ˆ ŵ), ∀(w,t,t1 ) ∈ U × B × B1, ŵ ∈ U. (18.32)
Hence,
inf {J(w,t,t1 )} ≥ sup {−J(
ˆ ŵ)}. (18.33)
(w,t,t1 )∈U×B×B1 ŵ∈U
−J(
ˆ ŵ0 ) = max{−J(
ˆ ŵ)}.
ŵ∈U
Finally, from the hypotheses, after discretization, we have that (t0 , (t1 )0 ) ∈ B× B1
such that
ˆ ŵ) =
J( sup {G1 (Λ ŵ0 ,t) + G2(ŵ0 ,t1 )
(t,t1 )∈B×B1
−ŵ0 , f L2 }
= G1 (Λ ŵ0 ,t0 ) + G2 (ŵ0 , (t1 )0 )
−ŵ0 , f L2 . (18.34)
θ ∈ ∂ J(
ˆ ŵ0 )
stands for
δŵ {G1 (Λ ŵ0 ,t0 ) + G2 (ŵ0 , (t1 )0 ) − ŵ0 , f L2 } = θ .
Thus,
ˆ ŵ0 ) = −G1 (Λ ŵ0 ,t0 ) − G2 (ŵ0 , (t1 )0 )
− J(
+ŵ0 , f L2
= G1 (Λ ŵ0 ,t0 ) + G2(ŵ0 , (t1 )0 ) + Ind(ŵ0,t0 , (t1 )0 )
= J(ŵ0 ,t0 , (t1 )0 ). (18.35)
Theorem 18.6.1. Consider a straight beam with rectangular cross section in which
the axis corresponds to the set Ω = [0, l]. Define J˜ : U × B → R = R ∪ {+∞} by
l l
1 1
˜
J(w,t) = EIw2,xx dx + K(t)w2 dx,
2 0 2 0
subject to
EIwxxxx + K(t)w − f = 0, in [0, l], (18.36)
where
U = {w ∈ W 2,2 (Ω ) | w(0) = w(l) = 0},
1
B = {t measurable | 0 ≤ t ≤ 1, a.e. in Ω and t dx ≤ c0 Ω },
0
and 0 < c0 < 1, K1 % K2 > 0.
Now, define J : U × B → R = R ∪ {+∞} by
J(w,t) = G(Λ w) + F(w,t) + Ind(w,t)
where
Λ : U → Y = Y ∗ = L2 (Ω ) ≡ L2 ,
is expressed by
Λ w = w,xx ,
l
1
G(Λ w) = EIw2,xx dx,
2 0
1 l
F(w,t) = K(t)w2 dx,
2 0
0, if (w,t) ∈ B0
Ind(w,t) = (18.37)
+∞, otherwise,
and
B0 = {(w,t) ∈ U × B such that (18.36) is satisfied }.
Under such hypotheses, we have
where A∗ = U,
J ∗ (v∗ ,t) = G∗ (v∗ ) + F ∗ (Λ ∗ v∗ − f ,t),
where
ˆ ŵ) = G(Λ ŵ) + F̂(ŵ) − ŵ, f L2 ,
J(
and
F̂(ŵ) = sup{F(ŵ,t)}.
t∈B
−J(
ˆ ŵ0 ) = max{−J(
ˆ ŵ)}.
ŵ∈U
θ ∈ ∂ J(
ˆ ŵ0 )
stands for
δŵ {G(Λ ŵ0 ) + F(ŵ0 ,t0 ) − ŵ0 , f L2 } = θ .
Under such hypotheses
We have computed, through the dual formulations, the solution for the numerical
values EI = 105 , K1 = 990, 000, K2 = 10, and P = 1, 000, t1 = 0.5, l = 1.0 with units
462 18 More on Duality Principles for Multi-well Problems
relating the international system. We consider two cases: first, for t, the constraint
0 ≤ t ≤ 1 and, in a second step, the constraint t ∈ {0, 1}.
For the results, please see Figs. 18.1, 18.2, 18.3, and 18.4, below indicated.
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0 0.2 0.4 0.6 0.8 1
18.7 Conclusion
At some point of our analysis, for almost all results, we have considered the dis-
cretized problem version. The reason is that we may not guarantee the attainability
of a measurable t0 ∈ B such that
120
100
80
60
40
20
0
0 0.2 0.4 0.6 0.8 1
1.2
0.8
0.6
0.4
0.2
−0.2
0 0.2 0.4 0.6 0.8 1
the solution related to Figs. 18.3 and 18.4 is the global optimal one. On the other
hand, the solution related to Figs. 18.1 and 18.2 is just a critical point (anyway, a
possible candidate to global optimal solution).
464 18 More on Duality Principles for Multi-well Problems
120
100
80
60
40
20
0
0 0.2 0.4 0.6 0.8 1
19.1 Introduction
U = {ϕ ∈ W 1,2 (Ω ) | ϕ = 0 on ∂ Ω } = W01,2 (Ω ).
F. Botelho ()
Department of Mathematics and Statistics, Federal University of Pelotas, Pelotas, RS-Brazil
e-mail: [email protected]
A. Ferreira
Department of Physics, Federal University of Pelotas, Brazil
e-mail: [email protected]
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 465
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 19,
© Springer International Publishing Switzerland 2014
466 19 Duality and Computation for Quantum Mechanics Models
h̄2 2
− ∇ ϕ + 2α |ϕ |2 ϕ + V (x)ϕ − μϕ = 0 in Ω , (19.1)
2m
where μ ∈ R is a suitable Lagrange multiplier to be obtained such that the following
constraint is satisfied:
|ϕ |2 dx = c. (19.2)
Ω
Here c > 0 is appropriate constant to be specified.
Furthermore, α , h̄, m are positive constants and V ∈ L2 (Ω ).
In case ϕ ∈ U satisfies (19.1) and (19.2), the function ψ (x,t), given by
iμ t
ψ (x,t) = e− h̄ ϕ (x),
∂ ψ (x,t) h̄2
ih̄ = − ∇2 ψ (x,t) + 2α |ψ (x,t)|2 ψ (x,t) + V (x)ψ (x,t) in Ω × [0, ∞),
∂t 2m
with the boundary condition ψ = 0 on ∂ Ω × [0, ∞), so that
|ψ (x,t)|2 dx = c, ∀t ∈ [0, ∞).
Ω
Remark 19.1.1. About the references, we highlight that details on the Sobolev
spaces involved may be found in [1, 26]. Duality principles for related problems are
addressed in [13]. Also, an extensive study on Lagrange multipliers may be found in
[40, 47]. For the numerical results, details on finite difference schemes are presented
in [63]. Finally, details on related physics problems are developed in [4, 45, 46].
where
h̄2
G1 (∇ϕ ) = ∇ϕ · ∇ϕ dx,
4m Ω
α
G2 (|ϕ |2 ) = |ϕ |4 dx,
2 Ω
1
G3 (ϕ ) = V (x)|ϕ |2 dx,
2 Ω
and, as above indicated,
0, if Ω |ϕ |2 dx = c,
Ind(ϕ ) = (19.5)
+∞, otherwise.
where Y = Y ∗ = L2 (Ω ),
Moreover,
δϕ {Jμ0 (ϕ0 )} = θ ,
where
and so that
|ϕ0 |2 dx − c = 0.
Ω
Also, suppose that
J˜μ0 (ϕ ) ≥ 0, ∀ϕ ∈ U,
where
∀ϕ ∈ U, v∗ ∈ Y ∗ . Thus,
that is,
Hence,
−J ∗ (v∗0 ) = J(ϕ0 ).
From this and (19.12) the proof is complete.
so that
2 2 2
3
3
∇ f (x, y) = 4∇
x y
+ ∇ .
x +y
2 2 x +y
2 2
We may compute
1
∇ f (x, y) = . (19.15)
x + y2
2
A similar result is valid for a N > 2-dimensional vector representing the dis-
cretized version of a function.
470 19 Duality and Computation for Quantum Mechanics Models
1 d 2 ϕ (x)
− + ϕ 3 (x) + V (x)ϕ (x) − μϕ (x) = 0, in Ω = [0, 1],
λ 2 dx2
with the boundary conditions
ϕ (0) = ϕ (1) = 0.
Here μ ∈ R is such that
ϕ 2 (x) dx = 1.
Ω
Moreover, the potential V (x) is given by
0, if x ∈ (0, 1),
V (x) = (19.17)
+∞, otherwise.
We denote by I(ϕ ) the diagonal matrix which the diagonal is the vector ϕ .
Furthermore, if ϕ ∈ RN is such that
ϕ = [ϕ1 , . . . , ϕN ],
we denote
ϕ p = [ϕ1p , . . . , ϕNp ], ∀p ∈ N.
The above equation, after discretization, in finite differences may be expressed by
1
− M2 ϕ + I(ϕ 2 )ϕ − μϕ = 0,
λ 2d 2
where ⎡ ⎤
−2 1 0 0 ... 0
⎢ 1 −2 1 0 ... 0 ⎥
⎢ ⎥
⎢ 0 1 −2 1 0.. 0 ⎥
M2 = ⎢
⎢... ... ... ...
⎥. (19.18)
⎢ ... ... ⎥
⎥
⎣ 0 0 ... 1 −2 1 ⎦
0 0 ... ... 1 −2
Here M2 is a N × N matrix, where N is the number of nodes and d = 1/N.
19.3 Numerical Examples 471
so that
5 6
ϕ̃n 2 2 2
−M2 + I d λ ϕ̃n+1 ≥ d 2 λ 2 ,
ϕ̃n
that is,
5 6
1 1 ϕ̃n 2 2 2
≤ −M2 + I d λ , ∀n ∈ N. (19.19)
ϕ̃n+1 d 2 λ 2 ϕ̃n
where we have used equality (19.16), as it is indicated at Remark 19.3.1, that is, for
a vector ϕ̃n ∈ RN ,
ϕ̃n
δ = 1 .
ϕ̃n ϕ̃n
Recalling that −M2 is positive definite, at this point, we assume that for an ap-
propriate choice of d, there exist n0 ∈ N and 0 < β < 1 such that
5 6
ϕ̃n 2 2 2
−M2 − I d λ
ϕ̃n
5 −2
2 6
ϕ̃
× −M2 + I
n
d λ
2 2
ϕ̃n
5 6
ϕ̃n 2 2 2
× −M2 + I d λ ≤ β < 1, ∀n > n0 .
ϕ̃n
19.3 Numerical Examples 473
so that denoting
ϕ̃0
ϕ0 = ,
ϕ̃0
we get
ϕ0 2 2
−M2 ϕ0 + I(ϕ02 )ϕ0 d 2 λ 2 − d λ = 0.
ϕ̃0
That is,
1
− M2 ϕ0 + I(ϕ02 )ϕ0 − μ0 ϕ0 = 0,
λ 2d 2
where
1
μ0 = .
ϕ̃0
Clearly we have
|ϕ0 |2 dx = 1.
Ω
Remark 19.3.2. Observe that we have not formally proven that the algorithm con-
verges. However, from the analysis developed, we have a strong indication that such
a convergence, under mild hypotheses, holds. Indeed, all numerical examples so far
worked have converged very easily.
An analogous analysis is valid for the two-dimensional case.
474 19 Duality and Computation for Quantum Mechanics Models
About the numerical results for this one-dimensional example, the ground state
is given in Fig. 19.1. Other solution with a greater eigenvalue is plotted in Fig. 19.2.
We highlight the numerical results obtained perfectly agree with the well-known
analytic solutions for this one-dimensional model. See [17] for details.
Finally, we present an analogous two-dimensional example, that is, we develop
results for the eigenvalue problem
1.4
1.2
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
1.5
0.5
−0.5
−1
−1.5
0 0.2 0.4 0.6 0.8 1
1 2
− ∇ ϕ (x, y) + ϕ 3 (x, y) + V (x, y)ϕ (x, y) − μϕ (x, y) = 0, in Ω = [0, 1] × [0, 1],
λ2
with the boundary condition
ϕ = 0 on ∂ Ω .
Here μ ∈ R is such that
ϕ 2 (x, y) dxdy = 1.
Ω
For such a two-dimensional case, the ground state for λ = 1 is given in Fig. 19.3.
The ground state, for λ = 25, is given in Fig. 19.4.
1.5
0.5
0
1
1
0.5 0.8
0.6
0.4
0.2
0 0
1.5
0.5
0
1
1
0.5 0.8
0.6
0.4
0.2
0 0
1.5
1
0.5
0
−0.5
−1
−1.5
1
1
0.5 0.8
0.6
0.4
0.2
0 0
19.4 Conclusion
20.1 Introduction
1
ei j (u) = (ui, j + u j,i ). (20.2)
2
Denoting by Hi1jkl and Hi0jkl two symmetric positive definite fourth-order constant
tensors, first we define the optimization problem of minimizing J(u,t) where
1
J(u,t) = tHi1jkl ei j (u)ekl (u) + (1 − t)Hi0jkl ei j (u)ekl (u) dx,
2 Ω
F. Botelho ()
Department of Mathematics and Statistics, Federal University of Pelotas, Pelotas, RS-Brazil
e-mail: [email protected]
A. Molter
Department of Mathematics and Statistics, Federal University of Pelotas, Brazil
e-mail: [email protected]
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 477
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 20,
© Springer International Publishing Switzerland 2014
478 20 Duality Applied to the Optimal Design in Elasticity
subject to
tHi1jkl ekl (u) + (1 − t)Hi0jkl ekl (u) + Pi = 0, in Ω , (20.3)
,j
tHi1jkl ekl (u) + (1 − t)Hi0jkl ekl (u) n j = 0, on Γ1 , (20.4)
emphasizing the result indicated in (20.8) is well known, whereas the chain of equal-
ities indicated in (20.15) and (20.17) we believe they are both new for the relaxed
problem, that is, for the case p > 1 where the min–max theorem does not apply.
Theorem 20.2.1. Considering the above expressed function Ind(u,t), defining
Hi jkl (t p )
G(e(u),t) = ei j (u)ekl (u) dx, (20.7)
Ω 2
and
J(u,t) = G(e(u),t) + Ind(u,t), ∀(u,t) ∈ U × B,
we have that
where
Moreover,
and
A∗ = {σ ∈ Y ∗ | σi j, j + Pi = 0, in Ω , σi j n j = 0 on Γ1 }.
Also, the following representation holds:
1
G(e(û),t) = sup − Hi jkl (t )vi j vkl dx + ei j (û), Hi jkl (t )vkl L2
p p
v∈Y 2 Ω
1
= Hi jkl (t p )ei j (û)ekl (û) dx. (20.12)
2 Ω
Assume there exists (σ0 , t˜0 , ũ0 ) ∈ A∗ × B × U such that
where
Jλ∗ (σ ,t, û) = G̃∗ (σ ,t) + ûi , −σi j, j − PiL2 + ûi σi j n j dΓ
Γ1
+ λ1 (t 2 − t) dx + λ2 t dx − t1 |Ω | . (20.13)
Ω Ω
= max{−J˜∗ (û)}
û∈U
= −J˜∗ (ũ0 ), (20.15)
where
σ0 = {(σ0 )i j } = {Hi jkl (t˜0p )ekl (ũ0 )}.
Finally, considering the same problem after discretization,
for the optimal û0 ∈ U satisfying (20.10), assume t0 ∈ B defined by
is such that (û0 ,t0 ) are also such that the hypotheses of Corollary 11.1 to hold.
Under such hypotheses, we have
inf {J(u,t)}
(u,t)∈U×B
= inf{ inf {J(u,t)}}
t∈B u∈U
= inf{ inf {G(e(u),t) + Ind(u,t)}}
t∈B u∈U
20.2 On the Duality Principle 481
= inf sup inf {G(e(u),t) + ûi, (Hi jkl (t p )ekl (u)), j + Pi L2
t∈B û∈U u∈U
− ûi Hi jkl (t )ekl (u)n j dΓ }
p
,
Γ1
so that
inf {J(u,t)}
(u,t)∈U×B
" #
= inf sup inf G(e(u),t) − ei j (û), Hi jkl (t p )ekl (u)L2 + ûi , Pi L2
t∈B û∈U u∈U
= inf sup {−G(e(û),t) + ûi , Pi L2 }
t∈B û∈U
" #
= inf inf∗ G̃∗ (σ ,t)
t∈B σ ∈A
∀u, û ∈ U.
Therefore,
From the hypotheses, there exists (σ0 , t˜0 , ũ0 ) ∈ A∗ × B × U such that
where
Jλ∗ (σ ,t, û) = G̃∗ (σ ,t) + ûi , −σi j, j − PiL2 + ûi σi j n j dΓ
Γ1
+ λ1 (t 2 − t) dx + λ2 t dx − t1 |Ω | , (20.22)
Ω Ω
Furthermore, −J˜∗ (û) is concave since it is the infimum of a family of concave func-
tions. From the coerciveness above verified, if {ûn} is a maximizing sequence, there
exists K1 > 0 such that
ûn U ≤ K1 , ∀n ∈ N.
Hence, there exists û0 ∈ U, such that up to a subsequence not relabeled, we have
Thus,
max{−J˜∗ (û)} = −J˜∗ (û0 ).
û∈U
is such that (û0 ,t0 ) are also such that the hypotheses of Corollary 11.1 are satisfied.
From such a corollary, the optimal equation
δ J˜∗ (û0 ) = θ
stands for
∂ {G(e(û0 ),t0 ) − û0 , PL2 }
= θ.
∂u
Hence,
(Hi jkl (t0p )ekl (û0 )), j + Pi = 0 in Ω , (20.26)
and
(Hi jkl (t0p )ekl (û0 ))n j = 0 on Γ1 , (20.27)
∀i ∈ {1, 2, 3}.
By (20.26) and (20.27), we obtain
1
− J˜∗(û0 ) = − Hi jkl (t0p )ei j (û0 )ekl (û0 ) dx + û0, PL2
2 Ω
1
= Hi jkl (t0p )ei j (û0 )ekl (û0 ) dx + Ind(û0,t0 )
2 Ω
= G(e(û0 ),t0 ) + Ind(û0,t0 ). (20.28)
Observe that
Remark 20.2.2. After discretization, for the case p = 1, the chain of equalities in-
dicated in (20.31) may be easily obtained from the min–max theorem. However,
to improve the numerical results in practical situations, it is desirable to use p > 1
and in such a case the min–max theorem does not apply (p = 3, e.g., is a standard
choice).
and
ĥ(t) = th31 + (1 − t)h30, h1 = 0.1, h0 = 10−4, and E = 107 , with the units related to
the international system (we emphasize to denote x = (x1 , x2 )). Observe that 0 ≤
t(x) ≤ 1, a.e. in Ω . Similarly, as in the last section (see [13] for details), we may
obtain the following duality principle (here in a slightly different version):
1
inf {J(w,t)} = inf H αβ λ μ (t)Mαβ Mλ μ dx ,
(w,t)∈U×B (t,{Mαβ })∈B×D∗ 2 Ω
where
{H αβ λ μ (t)} = {Hαβ λ μ (t)}−1 ,
and
We have computed the dual problem for Ω = [0, 1] × [0, 1] and a vertical load acting
on the plate given by P(x) = 10, 000, obtaining the results indicated in the respective
Figs. 20.1 and 20.2, for t0 (x) and w0 (x). We emphasize they are critical points, that
is, just candidates to optimal points. Observe that for the concerned critical point
Hαβ λ μ (t0 )
J(w0 ,t0 ) = (w0 ),αβ (w0 ),λ μ dx + Ind(w0,t0 )
Ω 2
Hαβ λ μ (t0 )
= (w0 ),αβ (w0 ),λ μ dx
Ω 2
1
= H (t0 )(M0 )αβ (M0 )λ μ dx
2 Ω αβ λ μ
= G̃∗ (σ0 ,t0 ), (20.36)
0.8
0.6
0.4
0.2
0
1
1
0.8
0.5 0.6
0.4
0.2
0 0
x 10−3
2
1.5
0.5
0
1
1
0.5 0.8
0.6
0.4
0.2
0 0
In the paper [60] Sigmund presents an algorithm for shape optimization. Such
an algorithm refers to find critical points of the Lagrangian functional Lλ (slightly
changed considering the development established in this text) given by
Lλ (u, û,t) = u, PL2 − Hi jkl (t p )ei j (u)ekl (û) dx + û, PL2
Ω
+ λ1+ (t(x) − 1) dx + λ1− (tmin − t(x)) dx
Ω
Ω
+λ2 t(x) dx − t1|Ω | . (20.37)
Ω
20.4 Another Numerical Example 487
Dual
0.5
0.4
0.3
m
0.2
0.1
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
m
Fig. 20.3 The cantilever beam density through the dual formulation, t1 = 0.50
Primal
0.5
0.4
m
0.3
F
0.2
0.1
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
m
Fig. 20.4 The cantilever beam density through the primal formulation, t1 = 0.50
488 20 Duality Applied to the Optimal Design in Elasticity
Dual F
0.2
0.18
0.16
0.14
0.12
0.1
m
0.08
0.06
0.04
0.02
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
m
Fig. 20.5 Clamped beam density through the dual formulation, t1 = 0.50
Primal F
0.2
0.18
0.16
0.14
0.12
0.1
m
0.08
0.06
0.04
0.02
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
m
Fig. 20.6 Clamped beam density through the primal formulation, t1 = 0.50
20.4 Another Numerical Example 489
Dual F
0.2
0.18
0.16
0.14
0.12
0.1
m
0.08
0.06
0.04
0.02
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
m
Fig. 20.7 Simply supported beam density through the dual formulation, t1 = 0.50
Primal F
0.2
0.18
0.16
0.14
0.12
0.1
m
0.08
0.06
0.04
0.02
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
m
Fig. 20.8 Simply supported beam density through the primal formulation, t1 = 0.50
490 20 Duality Applied to the Optimal Design in Elasticity
We made a change in the 99-line Sigmund algorithm, now updating the density t
through a critical point of the dual functional. The optimal equation would be
δt H i jkl (t p )σi j σkl dx
Ω
+ λ1+ (t(x) − 1) dx + λ1− (tmin − t(x)) dx
Ω Ω
+λ2 t(x) dx − t1 |Ω | = θ. (20.38)
Ω
For a fixed σ we update t through this last equation and update σ , for a fixed t,
through the equilibrium equation.
In a more practical fashion, concerning the 99-line finite element algorithm and
respective notation, the heuristic equation for each element through which te is up-
dated up to the OC function, that is,
η
p(te ) p−1 uTe K0 ue
te = te
new
, (20.39)
λ2 ∂∂ Vte
is replaced by
∂ (tep K0 ue )T (tenew )−p K0−1 (tep K0 ue )
+ λ2 = 0, (20.40)
∂ tenew
so that
1
p(te2p )uTe K0 ue
1+p
tenew = .
λ2
Hence,
1
p(tep−1)uTe K0 ue
1+p
tenew = te , (20.41)
λ2
It is worth emphasizing in such a procedure we must have 0 < te ,tenew < 1. To in-
clude the general case we have used an analogous OC function as found in the
original Sigmund algorithm. We may observe that Eqs. (20.41) and (20.39) are very
similar, so that in fact this last heuristic equation (20.39) almost corresponds, for
different values of η , to the Euler–Lagrange equations for the dual problem, which,
to some extent, justify the algorithm convergence. Based on the 99-line O. Sigmund
software with inclusion of the mentioned change (which corresponds, as above in-
dicated, to optimization through the dual formulation), the second author of this
chapter (Alexandre Molter) designed a code suitable for the different situations here
addressed, through which we have developed numerical examples.
20.5 Conclusion 491
20.5 Conclusion
21.1 Introduction
where
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 493
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 21,
© Springer International Publishing Switzerland 2014
494 21 Duality Applied to Micro-Magnetism
1, if x ∈ Ω ,
χΩ (x) = (21.4)
0, otherwise.
The term
α
|∇m|22 dx
2 Ω
is called the exchange energy, where
2
3 3
3
|m|2 = 4 ∑ m2k
k=1
and
3
|∇m|22 = ∑ |∇mk |22 .
k=1
We consider first the case of a uniaxial material with no exchange energy. That
is, α = 0 and ϕ (m) = β (1 − |m · e|).
Observe that
ϕ (m) = min{β (1 + m · e), β (1 − m · e)}
where β > 0 and e ∈ R3 is a unit vector.
In the next lines we present the primal formulation and related duality principle.
Define J : Y1 × Y2 × B → R by
G1 (m, f ,t) = (tg1 (m) + (1 − t)g2(m)) dx
Ω
+Ind0(m) + Ind1( f ) + Ind2(m, f )
− H(x) · m dx, (21.5)
Ω
21.2 Summary of Results for the Hard Uniaxial Case 495
and
1
G2 ( f ) = | f (x)|22 dx.
2 R3
Also,
g1 (m) = β (1 + m · e),
g2 (m) = β (1 − m · e),
0, if |m(x)|2 = 1 a.e. in Ω ,
Ind0 (m) =
+∞, otherwise,
0, if div(− f + mχΩ ) = 0 a.e. in R3 ,
Ind1 (m, f ) =
+∞, otherwise,
and
0, if Curl( f ) = θ , a.e. in R3 ,
Ind2( f ) =
+∞, otherwise.
Observe that as abovementioned,
ϕ (m) dx = β (1 − |m · e|) dx
Ω Ω
= min (tg1 (m) + (1 − t)g2(m)) dx . (21.6)
t∈B Ω
where
J ∗ (λ ) = G̃∗1 (λ ) + G∗2 (λ ) − β dx,
Ω
⎧ ⎫
⎨ 2 1/2 ⎬
3
∂ λ2
G̃∗1 (λ ) = sup
t∈B ⎩ Ω
∑ ∂ xi
+ Hi + β (1 − 2t)ei dx
⎭
i=1
where
2 1/2
3
∂ λ2
Ĝ∗1 (λ ,t) =
Ω
∑ ∂ xi
+ Hi + β (1 − 2t)ei dx,
i=1
and
1
G∗2 (λ ) = |Curl ∗ λ1 − ∇λ2 |22 dx.
2 R3
Finally,
B = {t measurable | t(x) ∈ [0, 1], a.e. in Ω },
496 21 Duality Applied to Micro-Magnetism
and
Ŷ ∗ = {λ = (λ1 , λ2 ) ∈ W 1,2 (R3 ; R3 ) × W 1,2 (R3 ) | λ2 = 0 on ∂ Ω }.
In the next lines we present one of our main results, which is summarized by the
following theorem.
Theorem 21.3.1. Define J : Y1 × Y2 × B → R by
G1 (m, f ,t) = (tg1 (m) + (1 − t)g2(m)) dx
Ω
+Ind0(m) + Ind1( f ) + Ind2(m, f )
− H(x) · m dx, (21.9)
Ω
and
1
G2 ( f ) = | f (x)|22 dx.
2 R3
Also,
g1 (m) = β (1 + m · e),
g2 (m) = β (1 − m · e),
0, if |m(x)|2 = 1 a.e. in Ω ,
Ind0 (m) =
+∞, otherwise,
0, if div(− f + mχΩ ) = 0 a.e. in R3 ,
Ind1 (m, f ) =
+∞, otherwise,
and
0, if Curl( f ) = θ , a.e. in R3 ,
Ind2 ( f ) =
+∞, otherwise.
This case refers to a uniaxial material with no exchange energy, that is, α = 0.
Observe that
ϕ (m) dx = β (1 − |m · e|) dx
Ω Ω
= min (tg1 (m) + (1 − t)g2(m)) dx . (21.10)
t∈B Ω
where
J ∗ (λ ) = G̃∗1 (λ ) + G∗2 (λ ) − β dx,
Ω
⎧ ⎫
⎨ 2 1/2 ⎬
3
∂ λ2
G̃∗1 (λ ) = sup
t∈B ⎩ Ω
∑ ∂ xi
+ Hi + β (1 − 2t)ei dx
⎭
i=1
where
2 1/2
3
∂ λ2
Ĝ∗1 (λ ,t) =
Ω
∑ ∂ xi
+ Hi + β (1 − 2t)ei dx,
i=1
1
G∗2 (λ ) = |Curl ∗ λ1 − ∇λ2 |22 dx,
2 R3
and
f0 = Curl ∗ (λ0 )1 − ∇(λ0 )2 ,
498 21 Duality Applied to Micro-Magnetism
we have that
we have that
This last infimum indicated is attained for functions satisfying the equations
∂ λ2
Hi + β (1 − 2t)ei − λ3 mi + = 0,
∂ xi
if λ2 = 0 on ∂ Ω .
That is,
Hi + β (1 − 2t)ei + ∂∂λx2i
mi =
λ3
and thus from the constraint
21.3 The Duality Principle for the Hard Case 499
3
∑ m2i − 1 = 0
i=1
we obtain 1/2
3
∂ λ2 2
λ3 = ∑ Hi + β (1 − 2t)ei + ∂ xi .
i=1
− f + Curl ∗ λ1 − ∇λ2 = θ .
∀(m, f ,t) ∈ Ỹ = Y1 × Y2 × B, λ ∈ Ŷ ∗ .
Therefore,
Also after discretization, from the hypotheses and Corollary 11.1, we have
δ {−J ∗(λ0 )} = θ ,
stands for
∂ Ĝ∗1 (λ0 ,t0 ) ∂ G∗2 (λ0 )
− − = θ,
∂ λ2 ∂ λ2
and
∂ G∗2 (λ0 )
− = θ,
∂ λ1
that is,
⎛ ⎞
∂ (λ 0 )2
3
∂ ⎜ ∂ xi + Hi + β (1 − 2t0)ei ⎟
∑ ∂ xi ⎜
⎝7
⎟
2 χΩ ⎠
∂ (λ 0 )2
i=1
∑3i=1 ∂ xi + Hi + β (1 − 2t0)ei
−div(Curl ∗ (λ0 )1 − ∇(λ0 )2 ) = 0, (21.20)
a.e. in R3 , and
Curl(Curl ∗ (λ0 )1 − ∇(λ0 )2 ) = θ , a.e. in R3 .
Hence
div(m0 χΩ − f0 ) = 0, a.e. in R3 ,
and
Curl( f0 ) = θ , a.e. in R3 .
Now observe that from the definition of m0 we get
∂ G∗2 (λ0 )
f0 = ,
∂ v1
where
v1 = Curl ∗ λ1 − ∇λ2 ,
so that
G∗2 (λ0 ) = f0 ,Curl ∗ (λ0 )1 − ∇(λ0 )2 L2 (Ω ;R3 ) − | f0 (x)|2 dx
R3
= f0 ,Curl ∗ (λ0 )1 − ∇(λ0 )2 L2 (Ω ;R3 ) − G2 ( f0 ). (21.22)
Remark 21.3.2. The reason we consider the problem just after discretization at some
point refers to the fact that we cannot guarantee the equation supt∈B {Ĝ∗1 (λ0 ,t)} =
Ĝ∗1 (λ0 ,t0 ) is satisfied, for the infinite dimensional original problem, by a measur-
able t0 ∈ B. In fact, the global optimal point for the primal formulation may not be
attained for the infinite dimensional problem, but surely it is attained for its finite
dimensional approximation. If the last theorem hypotheses are satisfied, the solution
for primal finite dimensional problem may be obtained by the corresponding dual
one with no duality gap.
G1 (m, f ,t) = (tg1 (m) + (1 − t)g2(m)) dx
Ω
+Ind0(m) + Ind1( f ) + Ind2(m, f )
− H(x) · m dx, (21.25)
Ω
and
1
G2 ( f ) = | f (x)|22 dx.
2 R3
Also,
g1 (m) = β (1 + m · e),
g2 (m) = β (1 − m · e),
0, if |m(x)|2 = 1 a.e. in Ω ,
Ind0 (m) =
+∞, otherwise,
0, if div(− f + mχΩ ) = 0 a.e. in R3 ,
Ind1 (m, f ) =
+∞, otherwise,
and
0, if Curl( f ) = θ , a.e. in R3 ,
Ind2 ( f ) =
+∞, otherwise.
We recall the present case refers to a uniaxial material with exchange energy. That
is, α > 0 and ϕ (m) = β (1 − |m · e|).
21.4 The Semi-Linear Case 503
where
1
Ĝ∗0 (m∗ , λ ) = sup {−G0 (m) + m2i , m∗i L2 (Ω ) − mi ni , λ2 L2 (∂ Ω ) },
m∈Y1 2
⎧ ⎛ 2 ⎞ ⎫
⎪
⎨1 ∂ λ2 ⎪
⎬
∂ xi + Hi + β (1 − 2t)ei
3
⎜ ⎟
G̃∗1 (m∗ , λ ) = sup ⎝∑ ⎠ dx
t∈B ⎪
⎩ 2 Ω i=1 m∗i + λ3 ⎪
⎭
where ⎛ 2 ⎞
∂ λ2
1 ⎜
3
∂ xi + Hi + β (1 − 2t)ei ⎟
Ĝ∗1 (m∗ , λ ,t) = ⎝∑ ⎠ dx,
2 Ω i=1 m∗i + λ3
1
G∗2 (λ ) = |Curl ∗ λ1 − ∇λ2 |22 dx,
2 R3
and
B = {t measurable | t(x) ∈ [0, 1], a.e. in Ω }.
Also,
A∗ = A1 ∩ A2 ∩ A3 ∩ A4 ,
where
Y3 = L2 (Ω ; R3 ) × W 1,2 (R3 ; R3 ) × W 1,2 (R3 ) × L2 (Ω ),
A1 = {(m∗ , λ ) ∈ Y3 | m∗i + λ3 > 0 in Ω , ∀i ∈ {1, 2, 3}}.
∂ m̃i
A2 = {(m∗ , λ , m̃) ∈ Y4 = Y3 × W 1,2 (Ω ; R3 ) | λ2 ni + =0
∂n
on ∂ Ω , ∀i ∈ {1, 2, 3}}, (21.29)
A3 = {(m∗ , λ ) ∈ Y3 | J(m)
˜ > 0, ∀m ∈ Y1 such that m = θ }.
504 21 Duality Applied to Micro-Magnetism
Here,
λ = (λ1 , λ2 , λ3 )
and
1, -
˜
J(m) = G0 (m) − m2i , m∗i L2 (Ω )
2
α 1
= |∇m|22 dx − m2i , m∗i L2 (Ω ) . (21.30)
2 Ω 2
And also,
is also such that Ĝ∗1 (m∗ , λ ,t) is locally Lipschitz continuous in a neighborhood of
(m∗0 , λ0 ,t0 ).
Also assume (m∗0 , λ0 ,t0 ) is such that the hypotheses of Corollary 11.1 are satis-
fied. Under such hypotheses, we have
∂ (λ 0 )2
∂ xi + Hi + β (1 − 2t0)ei
(m0 )i = , ∀i ∈ {1, 2, 3}
(m0 )∗i + (λ0 )3
and defining
f0 = Curl ∗ (λ0 )1 − ∇(λ0 )2 ,
we have that
∂ λ2
Hi + β (1 − 2t)ei − (m∗i + λ3 )mi + = 0.
∂ xi
That is,
Hi + β (1 − 2t)ei + ∂∂λx2i
mi = .
m∗i + λ3
506 21 Duality Applied to Micro-Magnetism
f = Curl ∗ λ1 − ∇λ2 .
if
∂ m̃i
λ2 ni + = 0 on ∂ Ω , ∀i ∈ {1, 2, 3}.
∂n
From this and (21.32) we obtain
stands for
, -
∂ −Ĝ∗1 (m∗0 , λ0 ,t0 ) − 12 (m0 )i , (m̃0 )i (m0 )∗i + α ∇2 (m̃0 )i L2 (Ω )
= θ, (21.38)
∂ m∗i
, -
∂ 1
(m ) , (m̃ ) (m )∗ + α ∇2 (m̃ )
2 0 i 0 i 0 i 0 i L (Ω )
2
− = θ, (21.39)
∂ m̃i
, -
∗
∂ 1
2 (m 0 )i , (m̃ 0 )i (m 0 )i + α ∇2
(m̃ 0 )i L (Ω )
2
= θ, (21.40)
∂ mi
Also,
∂ Ĝ∗0 (m∗0 , λ0 ) ∂ Ĝ∗1 (m∗0 , λ0 ,t0 ) ∂ G∗2 (λ0 )
− − − = θ, (21.43)
∂ λ2 ∂ λ2 ∂ λ2
and
∂ G∗2 (λ0 )
− = θ. (21.44)
∂ λ1
508 21 Duality Applied to Micro-Magnetism
a.e. in R3 , and
⎛ ⎞
∂ (λ 0 )2
∂ xi + Hi + β (1 − 2t0)ei
⎝ ⎠ ni − (m̃0 )i ni = 0, on ∂ Ω .
(m∗0 )i + (λ0 )3
and
3
∑ (m0 )2i = 1, a.e. in Ω .
i=1
From (21.44),
Hence
div(m0 χΩ − f0 ) = 0, a.e. in R3 ,
21.4 The Semi-Linear Case 509
and
Curl( f0 ) = θ , a.e. in R3 .
Now observe that from the expression of m0 we get
∂ G∗2 (λ0 )
f0 = ,
∂ v1
where
v1 = Curl ∗ λ1 − ∇λ2 ,
so that
G∗2 (λ0 ) = f0 ,Curl ∗ (λ0 )1 − ∇(λ0 )2 L2 (Ω ;R3 ) − | f0 (x)|2 dx
R3
= f0 ,Curl ∗ (λ0 )1 − ∇(λ0 )2 L2 (Ω ;R3 ) − G2 ( f0 ). (21.50)
H = H0 i + 0j,
where
i = (1, 0) and j = (0, 1).
√ √
Moreover, e1 = 2/2 and e2 = 2/2, where (e1 , e2 ) is the preferred direction of
magnetization. We have plotted the stream lines for the vector field m = (m1 , m2 )
for these three cases. Please see Figs. 21.1, 21.2, and 21.3. We observe that as H0
increases, the magnetization m direction approaches the magnetic field H one, which
in such an example is given by i = (1, 0).
Remark 21.5.1. It is worth mentioning that as H0 /β is smaller the magnetization m
is closer to (e1 , e2 ). However its direction approaches the H one, as H0 increases.
Such a result is consistent with the concerned problem physics.
H = H0 Ha (x, y),
where
Ha = x(0.5 − y)i − y(0.5 − x)j.
√ √
Also, again e1 = 2/2 and e2 = 2/2, where (e1 , e2 ) is the preferred direction of
magnetization. For the stream lines of Ha , please see Fig. 21.4. For the magnetiza-
tion m for these three different cases see Figs. 21.5, 21.6, and 21.7. For the parameter
t related to the case H0 /β = 0.5, see Fig. 21.8.
512 21 Duality Applied to Micro-Magnetism
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.2 0.4 0.6 0.8 1
Fig. 21.1 First example—stream lines for the magnetization m for H0 /β = 1.0
1.4
1.2
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
Fig. 21.2 First example—stream lines for the magnetization m for H0 /β = 10.0
21.5 Numerical Examples 513
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.2 0.4 0.6 0.8 1
Fig. 21.3 First example—stream lines for the magnetization m for H0 /β = 100.0
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.2 0.4 0.6 0.8 1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.2 0.4 0.6 0.8 1
Fig. 21.5 Second example—stream lines for the magnetization m for H0 /β = 0.5
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.2 0.4 0.6 0.8 1
Fig. 21.6 Second example—stream lines for the magnetization m for H0 /β = 5.0
21.6 Acknowledgments 515
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.2 0.4 0.6 0.8 1
Fig. 21.7 Second example—stream lines for the magnetization m for H0 /β = 50.0
0.8
0.6
0.4
0.2
0
1
1
0.8
0.5 0.6
0.4
0.2
0 0
21.6 Acknowledgments
The author is very grateful to Professor Robert C. Rogers for his excellent work
as Ph.D. thesis advisor at Virginia Tech-USA. We thank as well the Mathematics
Department of that institution for the constant financial support received during our
doctoral program.
516 21 Duality Applied to Micro-Magnetism
21.7 Conclusion
22.1 Introduction
In this chapter we develop solutions for the Navier–Stokes system through the
generalized method of lines. The main reference for this chapter is R. Temam [66].
At this point we describe the system in question.
Consider Ω ⊂ R2 an open, bounded, and connected set, whose internal boundary
is denoted by Γ0 and external one is denoted by Γ1 . Denoting by u : Ω → R the field
of velocity in direction x of the Cartesian system (x, y), by v : Ω → R the velocity
field in the direction y, by p : Ω → R the pressure field, so that P = p/ρ , where ρ is
the constant fluid density, ν is the viscosity coefficient, and g is the gravity constant,
the Navier–Stokes PDE system is expressed by
⎧ 2
⎪ ν ∇ u − u∂xu − v∂yu − ∂x P + gx = 0, in Ω ,
⎪
⎨
ν ∇2 v − u∂xv − v∂y v − ∂y P + gy = 0, in Ω , (22.1)
⎪
⎪
⎩
∂x u + ∂y v = 0, in Ω ,
u = v = 0, on Γ0 ,
(22.2)
u = u∞ , v = 0, P = P∞ , on Γ1
Through the next result we obtain a linear system whose solution also solves the
steady-state Euler system.
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 517
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 22,
© Springer International Publishing Switzerland 2014
518 22 The Generalized Method of Lines Applied to Fluid Mechanics
Theorem 22.2.1. A solution for the Euler system below indicated, that is,
⎧
⎪
⎪ −u∂x u − v∂yu − ∂x P + gx = 0, in Ω ,
⎪
⎪
⎨
−u∂x v − v∂yv − ∂y P + gy = 0, in Ω , (22.3)
⎪
⎪
⎪
⎪
⎩
∂x u + ∂yv = 0, in Ω ,
∇2 w0 = 0 in Ω . (22.7)
∂x F − ∂x P + gx = 0, in Ω , (22.8)
and
∂y F − ∂y P + gy = 0, in Ω , (22.9)
which may be solved in P for the particular w0 solution of (22.7) with corresponding
boundary conditions. Since
it is clear that (22.8) and (22.9) have a solution in P with the proper boundary con-
ditions above described.
22.3 The Generalized Method of Lines for the Navier–Stokes System 519
In this section we develop the solution for the Navier–Stokes system through the
generalized method of lines. About such a method see Chap. 15 for more details.
Consider
Ω = {(r, θ ) | 1 ≤ r ≤ 2, 0 ≤ θ ≤ 2π },
∂ Ω0 = {(1, θ ) | 0 ≤ θ ≤ 2π },
and
∂ Ω1 = {(2, θ ) | 0 ≤ θ ≤ 2π }.
First for the boundary conditions
u = 0, v = 0, P = P0 (x) on ∂ Ω0 ,
u = u f (x), v = v f (x), P = Pf (x) on ∂ Ω1 ,
we have obtained the following general expressions for the n − th lines (for an ap-
propriate approximate system of equations in polar coordinates):
Denoting
2
J(u, v, P) = ν ∇2 u − u∂xu − v∂y u − ∂x P dΩ
Ω
2
+ ν ∇2 v − u∂x v − v∂yv − ∂y P dΩ
Ω
+ (∂x u + ∂yv)2 d Ω , (22.10)
Ω
the coefficients {ai [n]}, {bi [n]}, {ci [n]} may be obtained through the numerical min-
imization of J(u, v, P).
u = 0, v = 0, P = P0 (x) on ∂ Ω0 ,
−r (θ )
f1 (θ ) = ,
r(θ )
we have
f0 (θ ) = 1 + f1 (θ )2 ,
f1 (θ )
f2 (θ ) = 1 + ,
1 + f1(θ )2
22.3 The Generalized Method of Lines for the Navier–Stokes System 521
2 f1 (θ )
f3 (θ ) = ,
1 + f1 (θ )2
and
1
f4 (θ ) = .
1 + f1 (θ )2
Also
∂u ∂u
d1 u/ f0 (θ ) = f5 (θ ) + ( f6 (θ )/t) ,
∂t ∂θ
∂u ∂u
d2 u/ f0 (θ ) = f7 (θ ) + ( f8 (θ )/t) ,
∂t ∂θ
where
f5 (θ ) = cos(θ )/r(θ ) + sin(θ )r (θ )/r3 (θ ),
f6 (θ ) = − sin(θ )/r(θ ),
f7 (θ ) = sin(θ )/r(θ ) − cos(θ )r (θ )/r3 (θ ),
f8 (θ ) = cos(θ )/r(θ ).
Observe that t ∈ [1, 2] in Ω .
From (22.11) and (22.12) we may write
Line 1
u1 (x) = 0.1u f (x) + 0.045 f5(x)P0 (x) − 0.045 f5(x)Pf (x)
+0.034 f2(x)u f (x) − 0.0165 f5(x)u f (x)2
−0.0165 f7(x)u f (x)v f (x) − 0.023 f6(x)P0 (x)
−0.011 f6(x)Pf (x) + 0.034 f3(x)uf (x)
−0.005 f6(x)u f (x)uf (x) − 0.005 f8(x)v f (x)uf (x)
+0.008 f4(x)uf (x)
Line 2
u2 (x) = 0.2u f (x) + 0.080 f5(x)P0 (x) − 0.080 f5(x)Pf (x)
+0.058 f2(x)u f (x) − 0.032 f5(x)u f (x)2
−0.032 f7(x)u f (x)v f (x) − 0.037 f6(x)P0 (x)
−0.022 f6(x)Pf (x) + 0.058 f3(x)uf (x)
−0.010 f6(x)u f (x)uf (x) − 0.010 f8(x)v f (x)uf (x)
+0.015 f4(x)uf (x)
Line 3
u3 (x) = 0.3u f (x) + 0.105 f5(x)P0 (x) − 0.105 f5(x)Pf (x)
+0.075 f2(x)u f (x) − 0.045 f5(x)u f (x)2
−0.045 f7(x)u f (x)v f (x) − 0.044 f6(x)P0 (x)
−0.030 f6(x)Pf (x) + 0.075 f3(x)uf (x)
−0.015 f6(x)u f (x)uf (x) − 0.015 f8(x)v f (x)uf (x)
+0.020 f4(x)uf (x)
Line 4
u4 (x) = 0.4u f (x) + 0.120 f5(x)P0 (x) − 0.120 f5(x)Pf (x)
+0.083 f2(x)u f (x) − 0.056 f5(x)u f (x)2
−0.056 f7(x)u f (x)v f (x) − 0.047 f6(x)P0 (x)
−0.037 f6(x)Pf (x) + 0.083 f3(x)uf (x)
−0.019 f6(x)u f (x)uf (x) − 0.019 f8(x)v f (x)uf (x)
+0.024 f4(x)uf (x)
Line 5
u5 (x) = 0.5u f (x) + 0.125 f5(x)P0 (x) − 0.125 f5(x)Pf (x)
+0.085 f2(x)u f (x) − 0.062 f5(x)u f (x)2
22.3 The Generalized Method of Lines for the Navier–Stokes System 523
Line 6
u6 (x) = 0.6u f (x) + 0.120 f5(x)P0 (x) − 0.120 f5(x)Pf (x)
+0.080 f2(x)u f (x) − 0.064 f5(x)u f (x)2
−0.064 f7(x)u f (x)v f (x) − 0.039 f6(x)P0 (x)
−0.040 f6(x)Pf (x) + 0.080 f3(x)uf (x)
−0.024 f6(x)u f (x)uf (x) − 0.024 f8(x)v f (x)uf (x)
+0.025 f4(x)uf (x)
Line 7
u7 (x) = 0.7u f (x) + 0.105 f5(x)P0 (x) − 0.105 f5(x)Pf (x)
+0.068 f2(x)u f (x) − 0.059 f5(x)u f (x)2
−0.059 f7(x)u f (x)v f (x) − 0.032 f6(x)P0 (x)
−0.037 f6(x)Pf (x) + 0.068 f3(x)uf (x)
−0.023 f6(x)u f (x)uf (x) − 0.023 f8(x)v f (x)uf (x)
+0.023 f4(x)uf (x)
Line 8
u8 (x) = 0.8u f (x) + 0.080 f5(x)P0 (x) − 0.080 f5(x)Pf (x)
+0.051 f2(x)u f (x) − 0.048 f5(x)u f (x)2
−0.048 f7(x)u f (x)v f (x) − 0.022 f6(x)P0 (x)
−0.029 f6(x)Pf (x) + 0.051 f3(x)uf (x)
−0.019 f6(x)u f (x)uf (x) − 0.019 f8(x)v f (x)uf (x)
+0.018 f4(x)uf (x)
Line 9
u9 (x) = 0.9u f (x) + 0.045 f5(x)P0 (x) − 0.045 f5(x)Pf (x)
+0.028 f2(x)u f (x) − 0.059 f5(x)u f (x)2
−0.028 f7(x)u f (x)v f (x) − 0.028 f6(x)P0 (x)
−0.011 f6(x)Pf (x) + 0.017 f3(x)uf (x)
524 22 The Generalized Method of Lines Applied to Fluid Mechanics
For the field of velocity v we have obtained for the following expressions for the
lines:
Line 1
v1 (x) = 0.1v f (x) + 0.045 f7(x)P0 (x) − 0.045 f7(x)Pf (x)
+0.034 f2(x)v f (x) − 0.017 f5(x)u f (x)v f (x)
−0.017 f7(x)v f (x)2 − 0.023 f8(x)P0 (x)
−0.011 f8(x)Pf (x) + 0.034 f3(x)vf (x)
−0.005 f6(x)u f (x)vf (x) − 0.005 f8(x)v f (x)vf (x)
+0.008 f4(x)vf (x)
Line 2
v2 (x) = 0.2v f (x) + 0.080 f7(x)P0 (x) − 0.080 f7(x)Pf (x)
+0.058 f2(x)v f (x) − 0.032 f5(x)u f (x)v f (x)
−0.032 f7(x)v f (x)2 − 0.037 f8(x)P0 (x)
−0.022 f8(x)Pf (x) + 0.058 f3(x)vf (x)
−0.010 f6(x)u f (x)vf (x) − 0.010 f8(x)v f (x)vf (x)
+0.015 f4(x)vf (x)
Line 3
v3 (x) = 0.3v f (x) + 0.105 f7(x)P0 (x) − 0.105 f7(x)Pf (x)
+0.075 f2(x)v f (x) − 0.045 f5(x)u f (x)v f (x)
−0.045 f7(x)v f (x)2 − 0.045 f8(x)P0 (x)
−0.030 f8(x)Pf (x) + 0.075 f3(x)vf (x)
−0.015 f6(x)u f (x)vf (x) − 0.015 f8(x)v f (x)vf (x)
+0.020 f4(x)vf (x)
Line 4
v4 (x) = 0.4v f (x) + 0.120 f7(x)P0 (x) − 0.120 f7(x)Pf (x)
+0.083 f2(x)v f (x) − 0.056 f5(x)u f (x)v f (x)
−0.056 f7(x)v f (x)2 − 0.047 f8(x)P0 (x)
−0.037 f8(x)Pf (x) + 0.083 f3(x)vf (x)
−0.019 f6(x)u f (x)vf (x) − 0.019 f8(x)v f (x)vf (x)
+0.024 f4(x)vf (x)
22.3 The Generalized Method of Lines for the Navier–Stokes System 525
Line 5
v5 (x) = 0.5v f (x) + 0.125 f7(x)P0 (x) − 0.125 f7(x)Pf (x)
+0.085 f2(x)v f (x) − 0.062 f5(x)u f (x)v f (x)
−0.062 f7(x)v f (x)2 − 0.045 f8(x)P0 (x)
−0.040 f8(x)Pf (x) + 0.085 f3(x)vf (x)
−0.022 f6(x)u f (x)vf (x) − 0.022 f8(x)v f (x)vf (x)
+0.026 f4(x)vf (x)
Line 6
v6 (x) = 0.6v f (x) + 0.120 f7(x)P0 (x) − 0.120 f7(x)Pf (x)
+0.068 f2(x)v f (x) − 0.064 f5(x)u f (x)v f (x)
−0.064 f7(x)v f (x)2 − 0.039 f8(x)P0 (x)
−0.040 f8(x)Pf (x) + 0.080 f3(x)vf (x)
−0.024 f6(x)u f (x)vf (x) − 0.024 f8(x)v f (x)vf (x)
+0.026 f4(x)vf (x)
Line 7
v7 (x) = 0.7v f (x) + 0.105 f7(x)P0 (x) − 0.105 f7(x)Pf (x)
+0.068 f2(x)v f (x) − 0.059 f5(x)u f (x)v f (x)
−0.059 f7(x)v f (x)2 − 0.032 f8(x)P0 (x)
−0.037 f8(x)Pf (x) + 0.068 f3(x)vf (x)
−0.023 f6(x)u f (x)vf (x) − 0.023 f8(x)v f (x)vf (x)
+0.023 f4(x)vf (x)
Line 8
v8 (x) = 0.8v f (x) + 0.080 f7(x)P0 (x) − 0.080 f7(x)Pf (x)
+0.051 f2(x)v f (x) − 0.048 f5(x)u f (x)v f (x)
−0.048 f7(x)v f (x)2 − 0.022 f8(x)P0 (x)
−0.029 f8(x)Pf (x) + 0.051 f3(x)vf (x)
−0.019 f6(x)u f (x)vf (x) − 0.019 f8(x)v f (x)vf (x)
+0.018 f4(x)vf (x)
526 22 The Generalized Method of Lines Applied to Fluid Mechanics
Line 9
v9 (x) = 0.9v f (x) + 0.045 f7(x)P0 (x) − 0.045 f7(x)Pf (x)
+0.028 f2(x)v f (x) − 0.028 f5(x)u f (x)v f (x)
−0.028 f7(x)v f (x)2 − 0.011 f8(x)P0 (x)
−0.017 f8(x)Pf (x) + 0.028 f3(x)vf (x)
−0.012 f6(x)u f (x)vf (x) − 0.012 f8(x)v f (x)vf (x)
+0.010 f4(x)vf (x)
Finally, for the field of pressure P, we have obtained the following lines:
Line 1
P1 (x) = 0.9P0(x) + 0.1Pf (x) − 0.034 f2(x)P1 (x)
+0.034 f2(x)Pf (x) + 0.045 f5(x)2 f0 (x)u f (x)2
+0.090 f5(x) f7 (x) f0 (x)u f (x)v f (x) + 0.045 f7(x)2 f0 (x)v f (x02
−0.034 f3(x)P0 (x) + 0.034 f3(x)Pf (x)
+0.022 f5(x) f6 (x) f0 (x)u f (x)uf (x) + 0.022 f5(x) f8 (x) f0 (x)v f (x)uf (x)
+0.003 f6(x)2 f0 (x)uf (x)2 + 0.022 f6(x) f7 (x) f0 (x)u f (x)vf (x)
+0.022 f7(x) f8 (x) f0 (x)v f (x)vf (x) + 0.007 f6(x) f8 (x) f0 (x)uf (x)vf (x)
+0.003 f8(x)2 f0 (x)vf (x)2 + 0.018 f4(x)P0 (x)
+0.008 f4(x)Pf (x)
Line 2
P2 (x) = 0.8P0(x) + 0.2Pf (x) − 0.058 f2(x)P0 (x)
+0.058 f2(x)Pf (x) + 0.080 f5(x)2 f0 (x)u f (x)2
+0.160 f5(x) f7 (x) f0 (x)u f (x)v f (x) + 0.080 f7(x)2 f0 (x)v f (x)2
−0.058 f3(x)P0 (x) + 0.058 f3(x)Pf (x)
+0.043 f5(x) f6 (x) f0 (x)u f (x)uf (x) + 0.043 f5(x) f8 (x) f0 (x)v f (x)uf (x)
+0.007 f6(x)2 f0 (x)uf (x)2 + 0.043 f6(x) f7 (x) f0 (x)u f (x)vf (x)
+0.043 f7(x) f8 (x) f0 (x)v f (x)vf (x) + 0.013 f6(x) f8 (x) f0 (x)uf (x)vf (x)
+0.007 f8(x)2 f0 (x)vf (x)2 + 0.028 f4(x)P0 (x)
+0.014 f4(x)Pf (x)
Line 3
P3 (x) = 0.7P0(x) + 0.3Pf (x) − 0.075 f2(x)P0 (x)
+0.075 f2(x)Pf (x) + 0.104 f5(x)2 f0 (x)u f (x)2
22.3 The Generalized Method of Lines for the Navier–Stokes System 527
+0.210 f5(x) f7 (x) f0 (x)u f (x)v f (x) + 0.105 f7(x)2 f0 (x)v f (x)2
−0.075 f3(x)P0 (x) + 0.075 f3(x)Pf (x)
+0.060 f5(x) f6 (x) f0 (x)u f (x)uf (x) + 0.060 f5(x) f8 (x) f0 (x)v f (x)uf (x)
+0.010 f6(x)2 f0 (x)uf (x)2 + 0.060 f6(x) f7 (x) f0 (x)u f (x)vf (x)
+0.060 f7(x) f8 (x) f0 (x)v f (x)vf (x) + 0.020 f6(x) f8 (x) f0 (x)uf (x)vf (x)
+0.010 f8(x)2 f0 (x)vf (x)2 + 0.034 f4(x)P0 (x)
+0.020 f4(x)Pf (x)
Line 4
P4 (x) = 0.6P0(x) + 0.4Pf (x) − 0.083 f2(x)P0 (x)
+0.083 f2(x)Pf (x) + 0.120 f5(x)2 f0 (x)u f (x)2
+0.240 f5(x) f7 (x) f0 (x)u f (x)v f (x) + 0.120 f7(x)2 f0 (x)v f (x)2
−0.083 f3(x)P0 (x) + 0.083 f3(x)Pf (x)
+0.073 f5(x) f6 (x) f0 (x)u f (x)uf (x) + 0.073 f5(x) f8 (x) f0 (x)v f (x)uf (x)
+0.012 f6(x)2 f0 (x)uf (x)2 + 0.073 f6(x) f7 (x) f0 (x)u f (x)vf (x)
+0.073 f7(x) f8 (x) f0 (x)v f (x)vf (x) + 0.073 f6(x) f8 (x) f0 (x)uf (x)vf (x)
+0.012 f8(x)2 f0 (x)vf (x)2 + 0.035 f4(x)P0 (x)
+0.024 f4(x)Pf (x)
Line 5
P5 (x) = 0.5P0(x) + 0.5Pf (x) − 0.085 f2(x)P0 (x)
+0.085 f2(x)Pf (x) + 0.125 f5(x)2 f0 (x)u f (x)2
+0.250 f5(x) f7 (x) f0 (x)u f (x)v f (x) + 0.125 f7(x)2 f0 (x)v f (x)2
−0.085 f3(x)P0 (x) + 0.085 f3(x)Pf (x)
+0.080 f5(x) f6 (x) f0 (x)u f (x)uf (x) + 0.080 f5(x) f8 (x) f0 (x)v f (x)uf (x)
+0.014 f6(x)2 f0 (x)uf (x)2 + 0.080 f6(x) f7 (x) f0 (x)u f (x)vf (x)
+0.080 f7(x) f8 (x) f0 (x)v f (x)vf (x) + 0.028 f6(x) f8 (x) f0 (x)uf (x)vf (x)
+0.014 f8(x)2 f0 (x)vf (x)2 + 0.033 f4(x)P0 (x)
+0.026 f4(x)Pf (x)
Line 6
P6 (x) = 0.4P0(x) + 0.6Pf (x) − 0.080 f2(x)P0 (x)
+0.080 f2(x)Pf (x) + 0.120 f5(x)2 f0 (x)u f (x)2
+0.240 f5(x) f7 (x) f0 (x)u f (x)v f (x) + 0.120 f7(x)2 f0 (x)v f (x)2
528 22 The Generalized Method of Lines Applied to Fluid Mechanics
Line 7
P7 (x) = 0.3P0(x) + 0.7Pf (x) − 0.068 f2(x)P0 (x)
+0.068 f2(x)Pf (x) + 0.105 f5(x)2 f0 (x)u f (x)2
+0.210 f5(x) f7 (x) f0 (x)u f (x)v f (x) + 0.105 f7(x)2 f0 (x)v f (x)2
−0.068 f3(x)P0 (x) + 0.068 f3(x)Pf (x)
+0.073 f5(x) f6 (x) f0 (x)u f (x)uf (x) + 0.073 f5(x) f8 (x) f0 (x)v f (x)uf (x)
+0.014 f6(x)2 f0 (x)uf (x)2 + 0.073 f6(x) f7 (x) f0 (x)u f (x)vf (x)
+0.073 f7(x) f8 (x) f0 (x)v f (x)vf (x) + 0.027 f6(x) f8 (x) f0 (x)uf (x)vf (x)
+0.014 f8(x)2 f0 (x)vf (x)2 + 0.022 f4(x)P0 (x)
+0.023 f4(x)Pf (x)
Line 8
P8 (x) = 0.2P0(x) + 0.8Pf (x) − 0.051 f2(x)P0 (x)
+0.051 f2(x)Pf (x) + 0.080 f5(x)2 f0 (x)u f (x)2
+0.160 f5(x) f7 (x) f0 (x)u f (x)v f (x) + 0.080 f7(x)2 f0 (x)v f (x)2
−0.051 f3(x)P0 (x) + 0.051 f3(x)Pf (x)
+0.058 f5(x) f6 (x) f0 (x)u f (x)uf (x) + 0.058 f5(x) f8 (x) f0 (x)v f (x)uf (x)
+0.011 f6(x)2 f0 (x)uf (x)2 + 0.058 f6(x) f7 (x) f0 (x)u f (x)vf (x)
+0.058 f7(x) f8 (x) f0 (x)v f (x)vf (x) + 0.022 f6(x) f8 (x) f0 (x)uf (x)vf (x)
+0.011 f8(x)2 f0 (x)vf (x)2 + 0.015 f4(x)P0 (x)
+0.018 f4(x)Pf (x)
Line 9
P9 (x) = 0.1P0(x) + 0.9Pf (x) − 0.028 f2(x)P0 (x)
+0.028 f2(x)Pf (x) + 0.045 f5(x)2 f0 (x)u f (x)2
+0.090 f5(x) f7 (x) f0 (x)u f (x)v f (x) + 0.045 f7(x)2 f0 (x)v f (x)2
−0.028 f3(x)P0 (x) + 0.028 f3(x)Pf (x)
22.3 The Generalized Method of Lines for the Navier–Stokes System 529
+0.034 f5(x) f6 (x) f0 (x)u f (x)uf (x) + 0.034 f5(x) f8 (x) f0 (x)v f (x)uf (x)
+0.007 f6(x)2 f0 (x)uf (x)2 + 0.034 f6(x) f7 (x) f0 (x)u f (x)vf (x)
+0.034 f7(x) f8 (x) f0 (x)v f (x)vf (x) + 0.013 f6(x) f8 (x) f0 (x)uf (x)vf (x)
+0.007 f8(x)2 f0 (x)vf (x)2 + 0.008 f4(x)P0 (x)
+0.010 f4(x)Pf (x)
Ω = {(r, θ ) | 1 ≤ r ≤ 2, 0 ≤ θ ≤ 2π },
∂ Ω0 = {(1, θ ) | 0 ≤ θ ≤ 2π },
and
∂ Ω1 = {(2, θ ) | 0 ≤ θ ≤ 2π }.
For the present example, the boundary conditions are
un (x) = a1 [n] cos(x) + a2[n] sin(x) + a3[n] cos(x)3 + a4 [n] cos(x) sin(x)2
vn (x) = b1 [n] cos(x) + b2 [n] sin(x) + b3 [n] sin(x)3 + b4 [n] cos(x)2 sin(x)
the coefficients {ai [n]}, {bi[n]}, {ci [n]} has been obtained through the numerical
minimization of J, so that for the mesh in question, we have obtained
22.4 Conclusion
−1
−2
−3
0 5 10 15 20
2.5
1.5
0.5
−0.5
−1
−1.5
−2
−2.5
0 5 10 15 20
1.5
0.5
−0.5
−1
−1.5
0 5 10 15 20
0.6
0.4
0.2
−0.2
−0.4
−0.6
−0.8
0 5 10 15 20
0.15
0.1
0.05
−0.05
−0.1
−0.15
−0.2
0 5 10 15 20
−1
−2
−3
0 5 10 15 20
0.8
0.6
0.4
0.2
−0.2
−0.4
−0.6
−0.8
0 5 10 15 20
0.6
0.4
0.2
−0.2
−0.4
−0.6
−0.8
0 5 10 15 20
0.4
0.3
0.2
0.1
−0.1
−0.2
−0.3
−0.4
0 5 10 15 20
0.08
0.06
0.04
0.02
−0.02
−0.04
−0.06
−0.08
0 5 10 15 20
23.1 Introduction
In this chapter, first we study duality for the optimal control concerning the
energy minimization of a well-known model of beams. The duality principle devel-
oped includes a concave dual variational formulation suitable to obtain numerical
results. For related results in optimization and convex analysis see [13, 14, 40, 47].
For details on the Sobolev spaces involved, see [1, 26]. We emphasize the dual
problem always has a solution through which we may verify the optimality of the
corresponding primal problem one. However in some situations the primal problem
may not have a global minimum, so that in such cases, if there is no duality gap be-
tween the dual and primal problems, the dual formulation global maximum solution
is a weak limit of minimizing sequences for the primal one. At this point we start to
describe the primal problem.
Consider a straight beam represented by the set Ω = [0, l] where l is the beam
length. Consider also the problem of minimizing the beam energy on a fixed in-
terval [0, T ], under the equilibrium equations, that is, the problem of minimizing
J : U → R, where
T l T l
1 1
J(w, u) = EI(wxx )2 dx dt + ρ A(wt )2 dx dt,
2 0 0 2 0 0
subject to
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 537
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3 23,
© Springer International Publishing Switzerland 2014
538 23 Duality Applied to the Optimal Control and Optimal Design of a Beam Model
In this section we develop a dual variational formulation for the optimal control
problem in question. Our main theoretical result is summarized by the next theorem
in which we redefine the functional J without relabeling it.
Theorem 23.2.1. Let ε > 0 be a small constant. Let J : U → R̄ = R ∪ {+∞} be
redefined by
J(w, u) = G(Λ w) + F(Λ w) + Ind(w, u),
where
Λ : U → Y = [L2 ([0, T ]; L2 (Ω ))]3
is given by
Λ w = {Λ1 w, Λ2 w, Λ3 w},
Λ1 w = wxx , Λ2 w = wx , Λ3 w = wt ,
T l T l
 B̂
G(Λ w) = (wxx )2 dx dt + (wt )2 dx dt
2 0 0 2 0 0
T l
K
− (wx )2 dx dt, (23.1)
2 0 0
T l
K
F(Λ w) = (wx )2 dx dt
2 0 0
ε T l
+ (wxx )2 dx dt
2 0 0
ε T l
+ (wt )2 dx dt (23.2)
2 0 0
where  = EI − ε , and B̂ = ρ A − ε .
23.2 The Duality Principle 539
Moreover,
0, if (w, u) ∈ B
Ind(w, u) = (23.3)
+∞, otherwise,
B = {(w, u) ∈ U | EIwxxxx + ρ Awtt + Cwt + uwx − f = 0 in Ω × [0, T ]},
U = Ũ × B̃,
Ũ = U1 ∩U2 ∩U3 ,
B̃ = B1 ∩ B2 ,
U1 = {w ∈ L2 ([0, T ]; H 2 (Ω )) |
w(0,t) = wx (0,t) = 0, in [0, T ]} (23.4)
U2 = {w ∈ L2 ([0, T ]; H 2 (Ω )) ∩ H 1 ([0, T ]; L2 (Ω )) |
w(x, 0) = w1 (x), and wt (x, 0) = w2 (x), in Ω }, (23.5)
U3 = {w ∈ L2 ([0, T ]; H 2 (Ω )) |
wxx (l,t) = wxxx (l,t) = 0, in [0, T ]} (23.6)
B1 = {u ∈ L2 ([0, T ]; L2 (Ω )) |
−M0 ≤ u(x,t) ≤ M0 , in Ω × [0, T ]}, (23.7)
and
B2 = {u ∈ L2 ([0, T ]; L2 (Ω )) |
l
|u(x,t)| dx ≤ c, in [0, T ]}. (23.8)
0
Also,
G(Λ w) ≥ 0, ∀w ∈ Ũ.
where
Also,
(G ◦ Λ )∗ (−Λ ∗ v∗ , λ )
= sup Λ w, −v∗ Y − (G ◦ Λ )(w)
w∈U
T
+ λ (0,t)EIwxxx (0,t) dt
0
T
− λx (0,t)EIwxx (0,t) dt
0
l
− λ (x, T )ρ Awt (x, T ) dx
0
l
+ λ (x, 0)ρ Aw2 (x) dx . (23.14)
0
23.2 The Duality Principle 541
∂ (G ◦ Λ )∗ (−Λ ∗ v∗0 , λ0 )
w0 = ,
∂ w∗
where w∗ = −Λ ∗ v∗ and
T l
1
F2∗ (v∗02 , λ0 ) = (v∗02 − λ0 u0 (x,t))2 dx dt
2K 0 0
where
where T l
g, hL2 = g(x,t)h(x,t) dx dt
0 0
and
F ∗∗ (Λ2 w, λ ) = sup {Λ2 w, v∗2 L2 − F2∗ (v∗2 , λ )}.
v∗2 ∈A∗
stands for
∂ F̃2∗ (v∗02 , λ0 , u0 )
Λ2 w0 = ,
∂ v∗2
where
∂ (G ◦ Λ )∗ (−Λ ∗ v∗0 , λ0 )
w0 = ,
∂ w∗
and
T l
1
F̃2∗ (v∗2 , λ , u) = (v∗2 − λ u(x,t))2 dx dt;
2K 0 0
then
∀(w, u) ∈ U, (v∗ , λ ) ∈ A∗ .
Therefore,
∀(w, u) ∈ U, (v∗ , λ ) ∈ A∗ .
Hence,
∂ (G ◦ Λ )∗ (−Λ ∗ v∗0 , λ0 )
w0 = , (23.25)
∂ w∗
we get
∂ F1∗ (v∗01 − EI λ0xx )
Λ1 w0 − = θ, (23.26)
∂ v∗1
Λ2 w0 ∈ ∂v∗2 F2∗ (v∗02 , λ0 ), (23.27)
and
∂ F3∗ (v∗03 + ρ Aλ0t − Cλ0 )
Λ3 w0 − = θ. (23.28)
∂ v∗3
Thus, from the last three relations, we obtain
From (23.25), the extremal condition concerning the variation in λ and these last
three equalities, we get
∂ F̃2∗ (v∗02 , λ0 , u0 )
Λ2 w0 = ,
∂ v∗2
where
T l
1
F̃2∗ (v∗2 , λ , u) = (v∗2 − λ u(x,t))2 dx dt,
2K 0 0
denoting
T l T l
K
F̃2 (Λ2 w, λ , u) = (λ u(x,t)wx ) dx dt + (wx )2 dx dt,
0 0 2 0 0
we have
F2∗ (v∗02 , λ0 ) = F̃2∗ (v∗02 , λ0 , u0 ) = Λ2 w0 , v∗02 L2 − F̃2(Λ2 w0 , λ0 , u0 )
= Λ2 w0 , v∗02 L2 − F2(Λ2 w0 , λ0 ). (23.32)
Moreover, the variation in λ in the dual formulation gives us the extremal inclu-
sion:
∗ ∗ ∗ ∗
∂ F1 (v01 − EI λ0xx ) ∂ F3 (v03 + ρ Aλ0t − Cλ0)
EI + ρ A
∂ v∗1 xx ∂ v∗3 t
∂ F3∗ (v∗03 + ρ Aλ0t − Cλ0)
+C − f ∈ [∂v∗2 F2∗ (v∗02 , λ0 )](−u0 ), (23.33)
∂ v∗3
so that
EIw0xxxx + ρ Aw0tt + Cwt + u0w0x − f = 0, in Ω × [0, T ].
Hence, from this last equation, (23.29) and (23.32) we have
J(w0 , u0 ) = G(Λ w0 ) + F(Λ w0 ) + Ind(w0 , u0 )
= G(Λ w0 ) + F1(Λ1 w0 )
+F̃2 (Λ2 w0 , λ0 , u0 ) + F3(Λ3 w0 )
+λ0 , EIw0xxxx + ρ Aw0tt + Cw0t + u0w0x − f L2
= −(G ◦ Λ )∗ (−Λ ∗ v∗0 , λ0 ) − F1∗ (v∗01 ) − F̃2∗ (v∗02 , λ0 , u0 )
−F3∗ (v∗03 ) − λ0 , f L2
= −(G ◦ Λ )∗ (−Λ ∗ v∗0 , λ0 ) − F1∗ (v∗01 ) − F2∗ (v∗02 , λ0 )
−F3∗ (v∗03 ) − λ0 , f L2
= −J ∗ (v∗0 , λ0 ). (23.34)
Consider a straight beam with circular cross-sectional area given by A(x), where
x ∈ [0, l], l being the beam length and [0, l] = Ω its axis. Suppose such a beam is sim-
ply supported, so that w ∈ U, where w : Ω → R is the field of vertical displacements
and
U = {w ∈ W 2,2 (Ω ) : w(0) = w(l) = 0}.
Also, the beam in question is assumed to be under a compressive axial load P
applied at x = l. We shall look for the optimal distribution A(x) which maximizes
the buckling load P, where the following designed constraints must be satisfied:
A(x) dx = V = cAmax l,
Ω
and
A(x) dx = V = cAmax l,
Ω
where 0 < c < 1.
Observe that from the concerned constraints
c0 P
A(x)2 w2,xx dx = w2,x dx,
2 Ω 2 Ω
so that through the appropriate constraint for the concerned eigenvalue problem, that
is,
w2,x dx = 1,
Ω
we get
c0 P
A(x)2 w2,xx dx = .
2 Ω 2
23.3 Some Closely Related Simpler Examples with Numerical Results 547
where
−c0
J(w, A) = A(x)2 w2,xx dx
2 Ω
and
0, if (w, P, A) ∈ A∗ ,
Ind(w, P, A) = (23.35)
+∞, otherwise,
where
A∗ = A1 ∩ A2 ∩ A3 ,
where
˜ P, A) = c0
J(w, A(x)2 w2,xx dx −
P
w2,x dx.
2 Ω 2 Ω
Finally,
A3 = w ∈ U | (w,x )2 dx = 1 .
Ω
Jλ (w, P, A) = −G(w,xx , A)
+λ , (c0 A(x)2 w,xx )xx + Pw,xx L2 (23.38)
so that
and therefore
Jλ (w, P, A) ≤ G∗ (v∗ , A) + Ind1(v∗ , λ , P, A),
where
1 (v∗ )2
G∗ (v∗ , A) = dx,
2c0 Ω A(x)2
and
0, if (v∗ , λ , P, A) ∈ B∗ ,
Ind1 (v∗ , λ , P, A) = (23.41)
+∞, otherwise,
where
B∗ = {(v∗ , λ , P, A) ∈ L2 × L2 × R+ × L2 |
v∗,xx − c0 (A(x)2 λxx ),xx + Pwxx = 0, in Ω ,
and v∗ (0) = v∗ (l) = 0}. (23.42)
Having this inequality in mind, we suggest the following algorithm to get critical
points relating the original problem. It is worth emphasizing we have not formally
proven its convergence:
1. Set k = 1 and choose w̃01 ∈ U such that
[(w̃01 ),x ]2 dx = 1.
Ω
2. Set A1 = cAmax .
3. Set n = 1 and w̃1k = w̃0k .
4. Calculate wnk ∈ U by solving the equation
5. Define
w̃kn+1 = wnk /Skn ,
where 7
Skn = [(wnk ),x ]2 dx.
Ω
6. Set n → n + 1 and go to step (4), up to the satisfaction of an appropriate conver-
gence criterion.
23.3 Some Closely Related Simpler Examples with Numerical Results 549
7. Define
w̃0k+1 = lim w̃nk ,
n→∞
1
Pk = lim .
n→∞ Skn
8. Define
v∗ = c0 Ak (x)(w̃0k+1 )xx ,
and obtain Ak+1 (x) by
where
C∗ = C1 ∩C2 ,
C1 = {A ∈ L2 | A(x) dx = V = cAmax l},
Ω
and
C2 = {A ∈ L2 | 0 < Amin ≤ A(x) ≤ Amax , in Ω }.
9. Set k → k + 1 and go to step (3), up to the satisfaction of an appropriate conver-
gence criterion.
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0 0.2 0.4 0.6 0.8 1
Fig. 23.1 Optimal distribution of area A(x)/α , intending to maximize the buckling load
where ρ denotes the beam density and w(x,t) : Ω × [0, T ] → R denotes the field
of vertical displacements. The motion results from proper initial conditions not yet
specified.
For the last equation we look for a solution of the form
so that
(c0 (A(x)2 u,xx )xx − ω 2ρ A(x)u(x) = 0, in Ω . (23.43)
At this point we consider the problem of finding A(x) which maximizes the fun-
damental frequency ω , subject to (23.43):
u2 dx = 1,
Ω
1 ω2
co A(x)2 (u,xx )2 dx − ρ A(x)u2 dx ≥ 0,
2 Ω 2 Ω
∀u ∈ U.
Moreover, the following design constraints must be satisfied:
A(x) dx = V = cAmax l,
Ω
23.3 Some Closely Related Simpler Examples with Numerical Results 551
and
0 < Amin ≤ A(x) ≤ Amax , in Ω .
This problem is mathematically similar to the previous one, related to the maxi-
mization of the buckling load.
Thus, similarly as in the last section, we define Jλ (u, A, ω ) by
Jλ (u, ω , A) = −G(u,xx , A)
+λ , (c0 A(x)2 u,xx )xx − ω 2 ρ A(x)uL2 (23.44)
where
c0
G(u,xx , A) = A(x)2 (u,xx )2 dx,
2 Ω
and λ is an appropriate Lagrange multiplier.
Observe that
so that
and therefore
Jλ (w, P, A) ≤ G∗ (v∗ , A) + Ind2(v∗ , λ , P, A),
where
1 (v∗ )2
G∗ (v∗ , A) = dx,
2c0 Ω A(x)2
and
0, if (v∗ , λ , ω , A) ∈ B∗ ,
Ind1 (v∗ , λ , ω , A) = (23.46)
+∞, otherwise,
where
B∗ = {(v∗ , λ , ω , A) ∈ L2 × L2 × R+ × L2 |
v∗,xx − c0 (A(x)2 λxx ),xx − ω 2ρ A(x)λ = 0, in Ω ,
and v∗ (0) = v∗ (l) = 0}. (23.47)
0.9
0.8
0.7
0.6
0.5
0.4
23.4 Conclusion
In this chapter we develop a concave dual variational formulation for the opti-
mal control of a well-known beam model. In practice, the results may be applied
to the energy minimization with piezoelectric actuators, through which the beam
vibration may be controlled (see [49] for related results). In a second step, we study
the optimal design for this same beam model, with the objective of maximizing its
buckling load and fundamental frequency, respectively. In both cases the numerical
results obtained are consistent with the problem physics. Finally, we emphasize the
approach here developed may be applied to many other situations, such as for plate
and shell models for example.
References
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 553
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3,
© Springer International Publishing Switzerland 2014
554 References
23. P. Ciarlet, Mathematical Elasticity, vol. III – Theory of Shells (North Holland, Elsevier, 2000)
24. B. Dacorogna, Direct Methods in the Calculus of Variations (Springer, New York, 1989)
25. I. Ekeland, R. Temam, Convex Analysis and Variational Problems (Elsevier-North Holland,
Amsterdam 1976).
26. L.C. Evans, Partial Differential Equations, Graduate Studies in Mathematics, vol. 19
(American Mathematical Society, Providence, Rhode Island, 1998)
27. U. Fidalgo and P. Pedregal: A General Lower Bound for the Relaxation of an Optimal Design
Problem with a General Quadratic Cost Functional and a General Linear State Equation,
Journal of Convex Analysis 19, number 1, 281–294 (2012)
28. N.B. Firoozye, R.V. Khon, Geometric Parameters and the Relaxation for Multiwell Ener-
gies. Microstructure and Phase Transition, the IMA volumes in mathematics and applications,
vol. 54, 85–110 (Oxford, UK, 1993)
29. I. Fonseca, G. Leoni, Modern Methods in the Calculus of Variations, L p Spaces (Springer,
New York, 2007)
30. A. Galka, J.J. Telega, Duality and the complementary energy principle for a class of geomet-
rically non-linear structures. Part I. Five parameter shell model Arch. Mech. 47, 677–698;
Part II. Anomalous dual variational principles for compressed elastic beams. Arch. Mech. 47,
699–724 (1995)
31. D.Y. Gao, G. Strang, Geometric nonlinearity: potential energy, complementary energy and the
gap function. Q. J. Appl. Math. 47, 487–504 (1989a)
32. D.Y. Gao, On the extreme variational principles for non-linear elastic plates. Q. Appl. Math.
XLVIII(2), 361–370 (1990)
33. D.Y. Gao, Pure complementary energy principle and triality theory in finite elasticity. Mech.
Res. Comm. 26(1), 31–37 (1999)
34. D.Y. Gao, General analytic solutions and complementary variational principles for large
deformation nonsmooth mechanics. Meccanica 34, 169–198 (1999)
35. D.Y. Gao, Finite deformation beam models and triality theory in dynamical post-buckling
analysis. Int. J. Non Linear Mech. 35, 103–131 (2000)
36. D.Y. Gao, Duality Principles in Nonconvex Systems, Theory, Methods and Applications
(Kluwer, Dordrecht, 2000)
37. M. Giaquinta, S. Hildebrandt, Calculus of Variations I. A Series of Comprehensive Studies
in Mathematics, vol. 310 (Springer, Berlin, 1996)
38. M. Giaquinta, S. Hildebrandt, Calculus of Variations II. A Series of Comprehensive Studies
in Mathematics, vol. 311 (Springer, Berlin, 1996)
39. E. Giusti, Direct Methods in the Calculus of Variations (World Scientific, Singapore, reprint,
2005)
40. K. Ito, K. Kunisch, Lagrange Multiplier Approach to Variational Problems and Applications.
Advances in Design and Control (SIAM, Philadelphia, 2008)
41. A. Izmailov, M. Solodov, Otimização, vol. 2 (IMPA, Rio de Janeiro, 2007)
42. A. Izmailov, M. Solodov, Otimização, vol. 1, 2nd edn. (IMPA, Rio de Janeiro, 2009)
43. R.D. James, D. Kinderlehrer, Frustration in ferromagnetic materials. Continuum Mech.
Thermodyn. 2, 215–239 (1990)
44. H. Kronmuller, M. Fahnle, Micromagnetism and the Microstructure of Ferromagnetic Solids
(Cambridge University Press, Cambridge, UK, 2003)
45. L.D. Landau, E.M. Lifschits, Course of Theoretical Physics, Vol. 5- Statistical Physics, Part 1
(Butterworth-Heinemann, Elsevier, reprint, 2008)
46. E.M. Lifschits, L.P. Pitaevskii, Course of Theoretical Physics, Vol. 9- Statistical Physics,
Part 2 (Butterworth-Heinemann, Elsevier, reprint, 2002)
47. D.G. Luenberger, Optimization by Vector Space Methods (Wiley, New York, 1969)
48. G.W. Milton, Theory of composites. Cambridge Monographs on Applied and Computational
Mathematics (Cambridge University Press, Cambridge 2002)
49. A. Molter, O.A.A. Silveira, J. Fonseca, V. Bottega, Simultaneous Piezoelectric Actuator and
Sensor Placement Optimization and Control Design of Manipulators with Flexible Links
Using SDRE Method, Mathematical Problems in Engineering, Hindawi, 2010, pp. 1–23
References 555
50. P. Pedregal, Parametrized measures and variational principles, Progress in Nonlinear Differ-
ential Equations and Their Applications, vol. 30 (Birkhauser, Basel, 1997)
51. P. Pedregal and B. Yan, On Two Dimensional Ferromagnetism. Proceedings of the Royal So-
ciety of Edinburgh, 139A, 575–594 (2009)
52. M. Reed, B. Simon, Methods of Modern Mathematical Physics, Volume I, Functional Analysis
(Reprint Elsevier, Singapore, 2003)
53. S.M. Robinson, Strongly regular generalized equations. Math. Oper. Res. 5, 43–62 (1980)
54. R.T. Rockafellar, Convex Analysis (Princeton University Press, Princeton, New Jersey, 1970)
55. R.C. Rogers, Nonlocal variational problems in nonlinear electromagneto-elastostatics, SIAM
J. Math. Anal. 19(6), 1329–1347 (1988)
56. R.C. Rogers, A nonlocal model of the exchange energy in ferromagnet materials. J. Integr.
Equat. Appl. 3(1), 85–127 (1991)
57. W. Rudin, Real and Complex Analysis, 3rd edn. (McGraw-Hill, New York, 1987)
58. W. Rudin, Functional Analysis, 2nd edn. (McGraw-Hill, New York, 1991)
59. H. Royden, Real Analysis, 3rd edn. (Prentice Hall India, New Delhi, 2006)
60. O. Sigmund, A 99 line topology optimization code written in Matlab. Struc. Muldisc. Optim.
21, 120–127 (2001)
61. J.J.A. Silva, A. Alvin, M. Vilhena, C.Z. Petersen, B. Bodmann, On a Closed form Solu-
tion of the Point Kinetics Equations with Reactivity Feedback of Temperature. International
Nuclear Atlantic Conference- INAC- ABEN, Belo Horizonte, MG, Brazil, October 24–28,
2011, ISBN: 978 -85-99141-04-05
62. E.M. Stein, R. Shakarchi, Real Analysis, Princeton Lectures in Analysis III, (Princeton
University Press, 2005)
63. J.C. Strikwerda, Finite Difference Schemes and Partial Differential Equations, 2d edn. SIAM
(Philadelphia, 2004)
64. D.R.S. Talbot, J.R. Willis, Bounds for the effective constitutive relation of a nonlinear
composite. Proc. R. Soc. Lond. 460, 2705–2723 (2004)
65. J.J. Telega, On the complementary energy principle in non-linear elasticity. Part I: Von Karman
plates and three dimensional solids, C.R. Acad. Sci. Paris, Serie II 308, 1193–1198; Part II:
Linear elastic solid and non-convex boundary condition. Minimax approach, C.R. Acad. Sci.
Paris, Serie II 308, 1313–1317 (1989)
66. R. Temam, Navier-Stokes Equations (AMS, Chelsea, reprint 2001)
67. J.F. Toland, A duality principle for non-convex optimisation and the calculus of variations.
Arch. Rath. Mech. Anal. 71(1), 41–61 (1979)
68. J.L. Troutman, Variational Calculus and Optimal Control, 2nd edn. (Springer, New York,
1996)
Index
F. Botelho, Functional Analysis and Applied Optimization in Banach Spaces: Applications 557
to Non-Convex Variational Models, DOI 10.1007/978-3-319-06074-3,
© Springer International Publishing Switzerland 2014
558 Index
Convex set, 7, 8, 48, 208, 251–260, 291, 302, Gram–Schmidt orthonormalization, 39–40
307, 383 Graph, 30, 88, 91, 271, 529
Countably compact set, 10, 16
H
D Hahn–Banach theorem, 41–56, 62, 185, 303
Dense set, 20, 26, 40, 52, 145 Hahn decomposition, 108–109
Distribution, 167–174, 183, 357, 456, 460, Hausdorff space, 7, 9, 48, 149–153
462, 463, 484, 486, 546, 550, 552 Helly lemma, 51
du Bois–Reymond lemma, 243–246 Hilbert space, 31–37, 39, 40, 59, 64, 66, 75,
Duality, 261–264, 280–285, 291–292, 77, 86, 88, 94, 110, 184
321–341, 343–362, 393–424, 437–446, Hölder continuous, 181
451–455, 465–491, 493–516, 537–552 Hyperplane, 44, 46, 48, 256, 287, 289
Duality principle, 321–339, 341, 343,
345–353, 361, 362, 366, 368, 394–400, I
425–464, 466–469, 478–484, 494, Inequality, 31, 32, 37, 47, 76, 105, 110, 136,
496–502, 516, 538–545, 548, 551 152, 176, 177, 179, 183, 184, 190,
Dual space, 42–43, 171, 427 192, 203–205, 207–209, 215, 219, 220,
Dual variational formulation, 343, 365–368, 299–303, 357, 358, 364, 366, 368, 381,
447, 476, 491, 493, 516, 537, 538, 552 409, 410, 482
Inner product, 31, 36, 110, 184
E Integral, 102–103, 106
Egorov theorem, 144–145 Integration, 99–128, 147–166, 199, 245, 248,
Eigenvalue, 73, 76, 465, 470, 474, 546, 549 322, 344, 375, 426, 465, 477, 493
Ekeland variational principle, 280–285 Interior, 5, 8, 12, 25, 26, 27, 28, 30, 73, 74,
Elasticity, 321–341, 343, 441, 447, 477–491 129, 252, 288, 289, 302
ε -net, 15, 17, 18, 191 Inverse mapping theorem, 30–31, 61, 293
Epigraph, 252, 253 Inverse operator, 73
Euler system, 517, 518
J
F Jordan, 109
Fatou’s lemma, 104–106
Finite-dimensional space, 4 K
Fluid mechanics, 517–535 Kakutani theorem, 49, 51
Fréchet derivative, 231, 312, 316
Fréchet differentiability, 231, 239, 292, 300, L
304, 309, 313 Lagrange multiplier, 288, 290, 292–305, 312,
Frequency, 550, 552 443, 466, 480, 482, 540, 547, 551
Fubini theorem, 122–128, 190, 193, 208 Lebesgue decomposition, 113–114
Function, 5, 21, 22, 99, 101–106, 138–146, Lebesgue dominated convergence theorem,
167, 180–182, 237, 240, 251–260, 272, 105–106, 126, 129, 178, 230
309–311, 434, 476, 486 Lebesgue integral, 103, 105, 181
Functional, 34, 41–45, 58, 62, 110, 153, 170, Lebesgue measurable function, 138–146
172, 179, 184, 185, 225, 233, 234, 239, Lebesgue measurable set, 129
251, 254, 257–258, 282, 288, 290–292, Lebesgue monotone convergence theorem,
359, 365, 370, 396, 425, 441, 449, 486, 103–106, 123, 125, 128, 129, 178
490, 494, 510, 547 Lebesgue points, 162–166
Lebesgue space, 175
G Legendre functional, 257
Gâteaux derivative, 226, 255 Legendre–Hadamard condition, 234–236
Gâteaux differentiability, 254–255, 311–317 Legendre transform, 257–259, 325, 500, 509
Gâteaux variation, 226–230, 313 Limit, 10, 12, 16, 38, 47, 64, 65, 70, 72, 86, 93,
Generalized method of lines, 307, 382–389, 163, 178, 189, 209, 226, 233, 235, 242,
392, 400, 404–411, 517–535 371, 372, 537
Ginzburg–Landau system, 374–377, 392–424 Limit point, 10, 12, 16, 93
Global existence, 374–377, 392 Linear function, 22, 23, 422
Index 559
Linear functional, 34, 41, 45, 58, 62, 110, 153, Nonnegative measurable functions, 103, 104,
170, 172, 179, 184, 185 109, 113, 125, 127, 141, 142
Linear mapping, 21–22 Norm, 11, 13, 30, 32, 43, 47, 52, 53, 56, 64,
Linear operator, 24, 25, 28, 30, 57–97, 185, 66, 69, 70, 72, 74, 80, 97, 175, 176, 180,
214, 255, 259, 263, 309, 312, 428 181, 183, 186, 187, 257, 273, 297, 307,
Locally convex space, 7, 170 308, 322, 345, 374, 384, 385, 469, 482
Local minimizer, 237, 299, 312 Normable space, 13
Local minimum, 226, 228–229, 234, 239, 247, Normed space, 11
292, 299, 300, 304–306, 339, 340 Null space, 21–22, 61, 293, 298
Lower semicontinuous, 151, 152, 163, Numerical, 321, 339–341, 353–362, 369–372,
251–260, 263, 267, 270, 280–282, 428, 382, 389–392, 400–411, 418, 419, 423,
432 424, 460–462, 466, 469–476, 484–491,
510–511, 529–530, 537, 546–552
M
O
Matrix version of generalized method of lines,
Open covering, 9
404–411
Open mapping theorem, 28–30
Maximal, 37, 42, 254
Open set, 4, 5, 9, 12, 44–46, 99, 100, 129, 133,
Maximize, 261, 352, 428, 431, 482, 546, 550,
137, 149–152, 155, 157, 159, 161, 163,
552
167, 175, 180, 195, 196, 198, 199, 205,
Maximum, 337
206, 228, 257, 271–273, 276
Measurability, 114–122
Operator, 23–24, 26, 28, 30, 57–97, 171, 180,
Measurable function, 99–101, 103–105, 109,
185, 199, 213, 218, 255, 259, 263, 283,
113, 125, 138–147, 257
307, 309, 312, 344, 365, 383, 385, 425,
Measurable set, 99, 102, 106, 107, 109,
426, 428, 429, 437
115–117, 121, 126, 129, 133–138, 141,
Operator topology, 57, 58
144, 166, 273
Optimal control, 377–381, 392, 456–459,
Measure, 99–166, 175, 242, 271, 322, 374,
537–552
477, 478, 485, 493
Optimal design, 456–459, 477–491, 537–552
Measure space, 101, 109, 122, 129 Optimality conditions, 300, 321, 476
Metric, 11–13, 53, 57, 282 Optimal shape, 486
Metric space, 11–19, 26, 52 Optimization, 261–264, 287–317, 349, 360,
Micro-magnetism, 493–516 452, 466, 477, 484, 486, 490, 491, 537,
Milman Pettis theorem, 55–56 546, 547
Minimizer, 41, 234, 263, 279, 344, 361, 364, Orthogonal complement, 32
370, 374, 435, 436, 516 Orthonormal basis, 36–40, 64
Minimum, 226, 228–229, 234, 239, 247, 267, Orthonormal set, 36–39
292, 299, 300, 304–306, 332, 339, 340, Outer measure, 114–122, 129–134
445, 537
Multi-well problem, 425–446 P
Phase transition, 425, 426, 437–446, 450–451
N Plate model, 321, 343–362, 456–459
Natural boundary conditions, 248–249 Positive functional, 153, 272, 445
Natural frequency, 550 Positive operator, 66–73, 79
Navier–Stokes system, 517, 519–530 Positive set, 106–108
Necessary conditions, 228–229, 239, 299, Projection, 49, 66, 82, 87, 94
304–306, 312, 340, 363
Neighborhood, 5–9, 22, 23, 46, 48, 51, 53–56, Q
150, 170, 252, 253, 255, 258, 263, 271, Quantum mechanics, 465–476
275, 278, 279, 309, 312, 313, 316, 334,
336, 338, 497, 504 R
Nested sequence, 13 Radon–Nikodym theorem, 109–114, 166
Newton’s method, 400, 402–405, 407, 411, Range, 21–22, 76, 96, 101, 161, 298
415–424 Real analysis, 3
560 Index
Real set, 225 309, 311, 322, 345, 365, 383, 384, 427,
Reflexive spaces, 48 428, 431, 494, 537
Relative compactness, 14, 18, 19, 191–195, Spectral theorem, 77–86, 94–97
220 Spectrum, 73–76
Resolvent, 73 Square root of a positive operator, 66–73
Riesz Strong operator topology, 57
lemma, 34, 89 Sufficient conditions, 234, 341, 343, 362, 365,
representation theorem, 110, 153–162, 394, 423
179–180 Symmetric operator, 92
S T
Scalar field, 3 Topological dual space, 42–43, 427
Schrödinger equation, 465, 466, 476 Topological space, 4, 5, 10, 99–101, 151
Self-adjoint operator, 67, 69, 71, 72, 76–86, Topological vector space, 3–40, 46, 168
91–97 Topology, 4, 6, 12, 13, 41–58, 167–169, 252,
Semi-linear case in micro-magnetism, 502–510 253, 257, 275, 282
Separable spaces, 52 Totally bounded set, 15, 17, 18, 194, 195
Sequentially compact, 15–18 Trace theorem, 213–216, 322, 344, 375, 379,
Set, 4–10, 12, 13, 15–19, 22, 25, 36–39, 447, 465, 477, 493
44–46, 48, 49, 51, 52, 64, 80, 99–100,
102, 106–109, 115–117, 120, 121, U
129, 133–138, 141, 144, 145, 149–155, Uniform convergence, 19, 180, 240
157–161, 163, 166–168, 170–172, 175, Uniformly convex space, 55–56
180, 182, 186, 187, 191–199, 205, 206, Unique, 3, 32–34, 89, 101, 109, 113, 153, 179,
208, 213, 217, 220, 225, 228, 248, 184, 257, 267, 307, 309, 310, 312, 355,
251–260, 267, 271–273, 276, 291, 302, 384, 409, 411, 493
307, 321, 322, 331, 337, 343, 345, 363, Upper bound, 37
375, 377, 383, 394, 403, 407, 426, 433, Upper semicontinuous, 79, 81, 82, 151, 152,
435, 437, 445, 447, 450, 452, 456, 465, 266, 267, 269, 431, 449, 459, 483
477, 484, 493, 517
Signed measures, 106–109 W
Simple function, 101–106, 127, 141, 142, 145, Weak convergence, 41, 381
272 Weakly closed, 48, 51, 52, 253
Sobolev imbedding theorem, 175, 186–213, Weakly compact, 51, 267–269
357, 363, 364, 375, 447, 465, 477, 493 Weak operator topology, 58
Sobolev space, 43, 175–220, 322, 494, 537 Weak-star topology, 48, 53
Space, 3–44, 46–53, 55–61, 63–64, 66, 73–75, Weak topology, 41–56, 252
77, 86, 88, 94, 99–101, 109, 110, 122, Weierstrass–Erdmann conditions, 245–247
129, 149–153, 168, 170, 171, 175–220, Weierstrass necessary condition, 239–243
225, 226, 231, 233, 234, 252, 253, 255,
257, 261, 263, 264, 266, 269, 282, 283, Z
287, 288, 292–293, 297–300, 304, 307, Zorn’s lemma, 37, 42