The Finite Element Method: Linear Static and Dynamic Finite Element Analysis
5/5
()
About this ebook
Related to The Finite Element Method
Related ebooks
Finite Element Methods Rating: 0 out of 5 stars0 ratingsNon-Linear Elastic Deformations Rating: 4 out of 5 stars4/5Finite Elements and Approximation Rating: 5 out of 5 stars5/5Theory of Elasticity and Stress Concentration Rating: 0 out of 5 stars0 ratingsIntroduction to Finite Element Analysis and Design Rating: 0 out of 5 stars0 ratingsFinite Elements of Nonlinear Continua Rating: 4 out of 5 stars4/5Theory of Thermal Stresses Rating: 5 out of 5 stars5/5Energy Principles and Variational Methods in Applied Mechanics Rating: 0 out of 5 stars0 ratingsAdvanced Mathematics for Engineers and Scientists Rating: 4 out of 5 stars4/5Introduction to the Explicit Finite Element Method for Nonlinear Transient Dynamics Rating: 0 out of 5 stars0 ratingsIntroduction to Differential Geometry for Engineers Rating: 0 out of 5 stars0 ratingsDifferential Equations for Engineers and Scientists Rating: 0 out of 5 stars0 ratingsStructural Reliability Analysis and Prediction Rating: 0 out of 5 stars0 ratingsModeling and Dimensioning of Structures: An Introduction Rating: 0 out of 5 stars0 ratingsComplex Analysis Rating: 3 out of 5 stars3/5Steel Connection Analysis Rating: 0 out of 5 stars0 ratingsStochastic Structural Dynamics: Application of Finite Element Methods Rating: 0 out of 5 stars0 ratingsFoundations of Applied Mathematics Rating: 3 out of 5 stars3/5Nonlinear Dynamics: Exploration Through Normal Forms Rating: 5 out of 5 stars5/5Complex Variable Methods in Elasticity Rating: 0 out of 5 stars0 ratingsAnalysis and Design of Machine Elements Rating: 0 out of 5 stars0 ratingsMethods of Applied Mathematics Rating: 3 out of 5 stars3/5The Finite Element Method: Fundamentals and Applications in Civil, Hydraulic, Mechanical and Aeronautical Engineering Rating: 0 out of 5 stars0 ratingsIntroduction to Numerical Analysis: Second Edition Rating: 4 out of 5 stars4/5Analytical Fracture Mechanics Rating: 0 out of 5 stars0 ratingsAdvanced Strength of Materials Rating: 3 out of 5 stars3/5Theory of Elastic Stability Rating: 5 out of 5 stars5/5Applied Partial Differential Equations Rating: 5 out of 5 stars5/5Nonlinear Finite Element Analysis of Solids and Structures Rating: 0 out of 5 stars0 ratingsModern Mathematics for the Engineer: Second Series Rating: 0 out of 5 stars0 ratings
Civil Engineering For You
Barracoon: The Story of the Last "Black Cargo" Rating: 4 out of 5 stars4/5Elon Musk: Tesla, SpaceX, and the Quest for a Fantastic Future Rating: 4 out of 5 stars4/5Small Gas Engine Repair Rating: 4 out of 5 stars4/5Life After Dead Pool: Lake Powell's Last Days and the Rebirth of the Colorado River Rating: 0 out of 5 stars0 ratingsThe End of the River Rating: 4 out of 5 stars4/5Rocks and Minerals of The World: Geology for Kids - Minerology and Sedimentology Rating: 5 out of 5 stars5/5Two-Stroke Engine Repair and Maintenance Rating: 0 out of 5 stars0 ratingsThe Off Grid Solar Power Bible For Beginners Rating: 0 out of 5 stars0 ratingsWater 4.0: The Past, Present, and Future of the World's Most Vital Resource Rating: 4 out of 5 stars4/5Troubleshooting and Repair of Diesel Engines Rating: 2 out of 5 stars2/5Killed by a Traffic Engineer: Shattering the Delusion that Science Underlies our Transportation System Rating: 4 out of 5 stars4/5Summary of Neil Postman's Amusing Ourselves to Death Rating: 4 out of 5 stars4/5The Lost Subways of North America: A Cartographic Guide to the Past, Present, and What Might Have Been Rating: 0 out of 5 stars0 ratingsHow Do Race Cars Work? Car Book for Kids | Children's Transportation Books Rating: 0 out of 5 stars0 ratingsA Trucker's Tale: Wit, Wisdom, and True Stories from 60 Years on the Road Rating: 3 out of 5 stars3/5Fracking 101 Rating: 5 out of 5 stars5/5Civil Engineering Rating: 4 out of 5 stars4/5The Things We Make: The Unknown History of Invention from Cathedrals to Soda Cans Rating: 0 out of 5 stars0 ratingsA Picture History of the Brooklyn Bridge Rating: 4 out of 5 stars4/5Structures Failures Reasons and Mitigation Rating: 0 out of 5 stars0 ratingsEngineering in Plain Sight: An Illustrated Field Guide to the Constructed Environment Rating: 0 out of 5 stars0 ratingsStructural Members and Frames Rating: 5 out of 5 stars5/5Engineering: The Way I See It Rating: 0 out of 5 stars0 ratingsSaudi America: The Truth About Fracking and How It's Changing the World Rating: 4 out of 5 stars4/5Rising Tide: The Great Mississippi Flood of 1927 and How It Changed America Rating: 4 out of 5 stars4/5Summary of Jayne Buxton's The Great Plant-Based Con Rating: 0 out of 5 stars0 ratingsCollecting and Identifying Rocks - Geology Books for Kids Age 9-12 | Children's Earth Sciences Books Rating: 0 out of 5 stars0 ratingsA History and Philosophy of Fluid Mechanics Rating: 5 out of 5 stars5/5Field Book for Quality Control in Earthwork Operations: Project Construction Management Book Rating: 0 out of 5 stars0 ratingsDesign of Buildings for Wind: A Guide for ASCE 7-10 Standard Users and Designers of Special Structures Rating: 4 out of 5 stars4/5
Reviews for The Finite Element Method
3 ratings1 review
- Rating: 5 out of 5 stars5/5exallent
Book preview
The Finite Element Method - Thomas J. R. Hughes
1
Fundamental Concepts; a Simple One-Dimensional Boundary-Value Problem
1.1 INTRODUCTORY REMARKS AND PRELIMINARIES
The main constituents of a finite element method for the solution of a boundary-value problem are
i. The variational or weak statement of the problem; and
ii. The approximate solution of the variational equations through the use of finite element functions.
To clarify concepts we shall begin with a simple example.
Suppose we want to solve the following differential equation for u:
where a comma stands for differentiation (i.e., u, xx = d²u/dxis a given smooth, scalar-valued function defined on the unit interval. We write
where [0, 1] stands for the unit interval (i.e., the set of points x such that 0 ≤ x stands for the real numbers. In words, (1.1.2) states that for a given x (xto mean in
or a member of.
Thus for each x (x.). Also, [0, 1] is said to be the domain of is its range.
, we want it to be a smooth curve without discontinuities or kinks. We do this to avoid technical difficulties. Right now we do not wish to elaborate further as this would divert us from the main theme. At some point prior to moving on to the next chapter, the reader may wish to consult Appendix 1.I, An Elementary Discussion of Continuity, Differentiability and Smoothness,
for further remarks on this important aspect of finite element work. The exercise in Sec. 1.16 already uses a little of the language described in Appendix 1.I. The terminology may be somewhat unfamiliar to engineering and physical science students, but it is now widely used in the finite element literature and therefore it is worthwhile to become accustomed to it.
Equation (1.1.1) is known to govern the transverse displacement of a string in tension and also the longitudinal displacement of an elastic rod. In these cases, physical parameters, such as the magnitude of tension in the string, or elastic modulus in the case of the rod, appear in (1.1.1). We have omitted these parameters to simplify subsequent developments.
Before going on, we introduce a few additional notations and terminologies. Let ]0, 1[ denote the unit interval without end points (i.e., the set of points x such that 0 < x < 1). ]0, 1[ and [0, 1] are referred to as open and closed unit intervals, respectively. To simplify subsequent writing and tie in with notation employed later on in multidimensional situations, we shall adopt the definitions
See Fig. 1.1.1.
Figure 1.1.1
At this point, considerations such as these may seem pedantic. Our purpose, however, is to develop a language for the precise articulation of boundary-value problems, which is necessary for good finite element work.
1.2 STRONG, OR CLASSICAL, FORM OF THE PROBLEM
A boundary-value problem for (1.1.1) involves imposing boundary conditions on the function u. There are a variety of possibilities. We shall assume u is required to satisfy
are given constants. Equations (1.2.1) and (1.2.2) require that u at x = 1 and the derivative of u at x = 0, respectively. This set of boundary conditions will later enable us to illustrate certain key features of variational formulations. For obvious reasons, boundary conditions of the type (1.2.1) and (1.2.2) lead to so-called two-point boundary-value problems.
The strong form of the boundary-value problem, (S), is stated as follows:
When we write u,xx = 0 on Ω we mean u,xx (x) = 0 for all x Ω. Of course, the exact solution of (S) is trivial to obtain, namely,
where y and z are used to denote dummy variables. However, this is not the main concern here. We are interested in developing schemes for obtaining approximate solutions to (S) that will be applicable to much more complex situations in which exact solutions are not possible.
Some methods of approximation begin directly with the strong statement of the problem. The most notable example is the finite difference method (e.g., see [1]). The finite element method requires a different formulation, which is treated in the next section.
1.3 WEAK, OR VARIATIONAL, FORM OF THE PROBLEM
To define the weak, or variational, counterpart of (S), we need to characterize two classes of functions. The first is to be composed of candidate, or trial, solutions. From the outset, we shall require these possible solutions to satisfy the boundary condition u. The other boundary condition will not be required in the definition. Furthermore, so that certain expressions to be employed make sense, we shall require that the derivatives of the trial solutions be square-integrable. That is, if u is a trial solution, then
Functions that satisfy (1.3.1) are called H¹-functions; we write u H¹. Sometimes the domain is explicitly included, i.e., u H¹([0, 1]).
Thus the collection of trial solutions, at x = 1. This is written as follows:
is a collection, or set, of objects is indicated by the curly brackets (called braces) in (1.3.2). The notation for the typical member of the set, in this case u, comes first inside the left-hand curly bracket. Following the vertical line ( | ) are the properties satisfied by members of the set.
The second collection of functions is called the weighting functions, or variations. -boundary condition. That is, we require weighting functions, w, to satisfy wand defined by
’s.)
In terms of the preceding definitions, we may now state a suitable weak form, (W), of the boundary-value problem.
Formulations of this type are often called virtual work, or virtual displacement, principles in mechanics. The w’s are the virtual displacements.
Equation (1.3.4) is called the variational equation, or (especially in mechanics) the equation of virtual work.
The solution of (W) is called the weak, or generalized, solution. The definition given of a weak formulation is not the only one possible, but it is the most natural one for the problems we wish to consider.
1.4 EQUIVALENCE OF STRONG AND WEAK FORMS; NATURAL BOUNDARY CONDITIONS
Clearly, there must be some relationship between the strong and weak versions of the problem, or else there would be no point in introducing the weak form. It turns out that the weak and strong solutions are identical. We shall establish this assuming all functions are smooth. This will allow us to proceed expeditiously without invoking technical conditions with which the reader is assumed to be unfamiliar. Proofs
of this kind are sometimes euphemistically referred to as formal proofs.
The intent is not to be completely rigorous but rather to make plausible the truth of the proposition. With this philosophy in mind, we shall prove
the following.
Proposition
a. Let u be a solution of (S). Then u is also a solution of (W).
b. Let u be a solution of (W) Then u is also a solution of (S).
Another result, which we shall not bother to verify but is in fact easily established, is that both (S) and (W) possess unique solutions. Thus, by (a) and (b), the strong and weak solutions are one and the same. Consequently, (W) is equivalent to (S).
Formal Proof
a. Since u is assumed to be a solution of (S), we may write
for any w . Integrating (1.4.1) by parts results in
Rearranging and making use of the fact that −u, x(0) = h and w(1) = 0 results in
Furthermore, since u is a solution of (S), it satisfies u. Finally, since u also satisfies (1.4.3) for all w , u satisfies the definition of a weak solution given by (W).
b. Now u is assumed to be a weak solution. Thus u ; consequently u, and
for all w . Integrating by parts and making use of the fact w(1) = 0 results in
To prove u is a solution of (S) it suffices to show that (1.4.4) implies¹
i. u,xx = 0 on Ω; and
ii. u,x= 0
First we shall prove (i). Define w in (1.4.4) by
where ϕ is smooth; ϕ(x) > 0 for all x Ω = ]0, 1[; and ϕ(0) = ϕ(1) = 0. For example, we can take ϕ(x) = x(1 – x), which satisfies all the stipulated requirements (see Figure 1.4.1). It follows that w(1) = 0 and thus w . Substituting (1.4.5) into (1.4.4) results in
Figure 1.4.1
Since ϕ > 0 on Ω, it follows from (1.4.6) that (i) must be satisfied.
Now that we have established (i), we may use it in (1.4.4) to prove (ii), namely,
That w puts no restriction whatsoever on its value at x = 0. Therefore, we may assume that the w in (1.4.7) is such that w
Remarks
1. The boundary condition –u,xis not explicitly mentioned in the statement of (W). From the preceding proof, we see that this boundary condition is, however, implied by the satisfaction of the variational equation. Boundary conditions of this type are referred to as natural boundary conditions. On the other hand, trial solutions are explicitly required to satisfy the boundary condition u. Boundary conditions of this type are called essential boundary conditions. The fact that solutions of the variational equation satisfy natural boundary conditions is extremely important in more complicated situations which we will consider later on.
2. The method used to prove part (b) of the proposition goes under the name of the fundamental lemma in the literature of the calculus of variations. In essence, it is the methodology that enables us to deduce the differential equations and boundary conditions implied by the weak formulation. To develop correct weak forms for complex, multidimensional problems, it is essential to have a thorough understanding of these procedures.
Now we see that to obtain approximate solutions to the original boundary-value problem we have alternative starting points, i.e., the strong or weak statements of the problem. Finite element methods are based upon the lattercontain infinitely many functions.) The variational equations are then solved in this finite-dimensional context. An explicit example of how to go about this is the subject of the next section. However, we first introduce some additional notations to simplify subsequent writing.
Let
In terms of (1.4.8) and (1.4.9), the variational equation takes the form
Here, a(·, ·) and (·, ·) are examples of symmetric, bilinear forms. What this means is as follows: Let c1 and c2 be constants and let u, υ, and w be functions. Then the symmetry property is
Bilinearity means linearity in each slot
; for example,
Exercise 1. Use the definitions of a(·, ·) and (·, ·) to verify the properties of symmetry and bilinearity.
The above notations are very concise; at the same time they capture essential mathematical features and thus are conducive to a mathematical understanding of variational and finite element methods. Diverse classes of physical problems can be written in essentially similar fashion to (1.4.10). Thus ideas developed and results obtained are seen at once to have very broad applicability.
1.5 GALERKIN'S APPROXIMATION METHOD
We shall now describe a method of obtaining approximate solutions to boundary-value problems based upon weak formulations. Our introduction to this subject is somewhat of an abstract treatment. However, the meaning should be significantly reinforced by the remaining sections of the chapter. It may be worthwhile for the reader to consult this section again after completing the rest of the chapter to make sure a full comprehension of the material is attained.
h hh h with a mesh, or discretization, of the domain Ω, which is parameterized by a characteristic length scale h. h h , respectively. This is written as
where the precise meaning is given in parentheses.² Consequences of (1.5.1) and (1.5.2) are (respectively) that if uh h and wh h, then
hh, are often referred to as function spaces. The terminology space in mathematics usually connotes a linear structure. This has the following meaning: If c1 and c2 are constants and v and w , then c1v + c2w h h due to the inhomogeneous boundary condition. For example, if u1 and u2 , then u1 + u, since u1(l) + uh.
(Bubnov-) Galerkin Method
h is given. Then, to each member vh h, we construct a function uh h by
h is a given function satisfying the essential boundary condition, i.e.,
Note that (1.5.5) satisfies the requisite boundary condition also:
Thus (hh hh h are composed of identical collections of functions. This property will be shown later on to have significant consequences for certain classes of problems.
We now write a variational equation, of the form of (1.4.10), in terms of wh h and uh h:
This equation is to be thought of as defining an approximate (weak) solution, uh.
Substitution of (1.5.5) into (1.5.8), and the bilinearity of a(·, ·) enables us to write
, and h). Equation (1.5.9) is to be used to define vh, the unknown part of uh.
The (Bubnov-) Galerkin form of the problem, denoted by (G), is stated as follows:
Note that (G) is just a version of (Wh.
h h have to be explicitly defined. Before doing this, it is worthwhile to mention a larger class of approximation methods, called Petrov-Galerkin methods, in which vh h. Recent attention has been paid to methods of this type, especially in the context of fluid mechanics. For the time being, we will be exclusively concerned with the Bubnov-Galerkin method. The Bubnov-Galerkin method is commonly referred to as simply the Galerkin method, terminology we shall adopt henceforth. Equation (1.5.9) is sometimes referred to as the Galerkin equation.
Approximation methods of the type considered are examples of so-called weighted residual methods. The standard reference dealing with this subject is Finlayson [3]. For a more succinct presentation containing an interesting historical account, see Finlayson and Scriven [4].
1.6 MATRIX EQUATIONS; STIFFNESS MATRIX K
hh consist of all linear combinations of given functions denoted by N, where A = 1, 2,…, n. By this we mean that if wh h , then there exist constants cA, A = 1, 2,…, n, such that
The NA’s are referred to as shape, basis, or interpolation functions. We require that each NA satisfies
from which it follows by (1.6.1) that whh is said to have dimension n, for obvious reasons.
h h. To this end, we introduce another shape function, Nn , which has the property
(Note Nn hh is given by
and thus
With these definitions, a typical uh h may be written as
where the dA’s are constants and from which it is apparent that uh.
Substitution of (1.6.1) and (1.6.6) into the Galerkin equation yields
By using the bilinearity of a(·, ·) and (·, ·), (1.6.7) becomes
where
Now the Galerkin equation is to hold for all wh h. By (1.6.1), this means for all cA’s, A = 1, 2,…, n. Since the cA’s are arbitrary in (1.6.8), it necessarily follows that each GA, A = 1, 2,…, n, must be identically zero, i.e., from (1.6.9)
Note that everything is known in (1.6.10) except the dB’s. Thus (1.6.10) constitutes a system of n equations in n unknowns. This can be written in a more concise form as follows:
Let
Then (1.6.10) becomes
Further simplicity is gained by adopting a matrix notation. Let
and
Now (1.6.13) may be written as
The following terminologies are frequently applied, especially when the problem under consideration pertains to a mechanical system:
A variety of physical interpretations are of course possible.
At this point, we may state the matrix equivalent, (M), of the Galerkin problem.
The solution of (M) is, of course, just d = K− 1F (assuming the inverse of K, K− 1 exists). Once d is known, the solution of (G) may be obtained at any point x by employing (1.6.6), viz.,
Likewise, derivatives of uh, if required, may be obtained by term-by-term differentiation. It should be emphasized, that the solution of (G) is an approximate solution of (W). Consequently, the differential equation and natural boundary condition are only approximately satisfied. The quality of the approximation depends upon the specific choice of NA’s and the number n.
Remarks
1. The matrix K is symmetric. This follows from the symmetry of a(·, ·) and use of Galerkin’s method (i.e., the same shape functions are used for the variations and trial solutions):
In matrix notation
where the superscript T denotes transpose. The symmetry of K has important computational consequences.
2. Let us schematically retrace the steps leading to the matrix problem, as they are typical of the process one must go through in developing a finite element method for any given problem:
The only apparent approximation made thus far is in approximately solving (W) via (G, and h may be approximated, as well as the domain Ω, calculation of integrals, and so on. Convergence proofs and error analyses involve consideration of each approximation.
3. It is sometimes convenient to write
where dn .
1.7 EXAMPLES: 1 AND 2 DEGREES OF FREEDOM
In this section we will carry out the detailed calculations involved in formulating and solving the Galerkin problem. The functions employed are extremely simple, thus expediting computations, but they are also primitive examples of typical finite element functions.
Example 1 (1 degree of freedom)
In this case n = 1. Thus wh = c1N1 and uh = vh h = d1NN2. The only unknown is d1. The shape functions must satisfy N1(1) = 0 and N2(1) = 1 (see (1.6.2) and (1.6.3)). Let us take N1(x) = 1 − x and N2(x) = x. These are illustrated in Fig. 1.7.1 and clearly satisfy the required conditions. Since we are dealing with only 1 degree of freedom, the matrix paraphernalia collapses as follows:
Consequently
In (1.7.7), y .
Figure 1.7.1 Functions for the 1 degree of freedom examples. (These functions are secretly the simplest finite element interpolation functions in a one-element context.)
Figure 1.7.2 The Galerkin solution for the 1 degree of freedom example.
i. = 0. Then
That is, the approximate solution is exact. In fact, it is clear by inspecting (≠ 0).
ii. (x, a constant. Then the particular solutions take the form
and
Equations (1.7.9) and (1.7.10) are compared in Fig. 1.7.3. Note that uhpart is exact at x = 0 and x = 1 and that uhpart,x is exact at x .(It should be clear that it is impossible for uhpart to be exact at all x in the present circumstances. The exact solution, (1.7.9), contains a quadratic term in x, whereas the approximate solution is restricted to linear variation in x by the definitions of N1 and N2.)
Figure 1.7.3 Comparison of exact and Galerkin particular solutions, Example 1, case (ii).
iii. (x) = qxleads to
and
which are compared in Fig. 1.7.4. Again we note that the uhpart is exact at x = 0 and x = 1. There is one point, x , at which uhpart, x is exact.
Let us summarize what we have observed in this example:
a. The homogeneous part of uh is exact in all cases.
, uh is exact at x = 0 and x = 1.
c. For each case, there is at least one point at which uh,x is exact.
Figure 1.7.4 Comparison of exact and Galerkin particular solutions, Example 1, case (iii).
Example 2 (2 degrees of freedom)
In this case n = 2. Thus wh = c1N1 + c2N2, where N1(l) = N2(l) = 0, and uh = d1N1 + d2NN3. where N3(l) = 1. Let us define the NA’s as follows
The shape functions are illustrated in Fig. 1.7.5. Typical wh h and uh h and their derivatives are shown in Fig. 1.7.6. Since n = 2, the matrix paraphernalia takes the following form:
Figure 1.7.5 Functions for the 2 degree of freedom examples. (These functions are secretly the simplest finite element functions in a two-element context.)
Figure 1.7.6 Typical weighting function and trial solution for the 2 degree of freedom example.
Note that due to the shape functions’ discontinuities in slope at x , 1] (e.g., see (1.7.12) and (1.7.15)). We need not worry about the value of the derivative of NA at x (it suffers a discontinuity there and thus is not well-defined classically) since it has no effect on the integrals in (1.7.12). This amounts to employing the notion of a generalized derivative.
We shall again analyze the three cases considered in Example 1.
= 0.
This results in
Again, the exact homogeneous solution is obtained. (The reason for this is that the exact solution is linear, and our trial solution is capable of exactly representing any linear function. Galerkin’s method will give the exact answer whenever possible—that is, whenever the collection of trial solutions contains the exact solution among its members.)
ii(x= constant.
The solution takes the form
The approximate particular solution is compared with the exact in Fig. 1.7.7, from which we see that agreement is achieved at x and 1, and derivatives coincide at x .
iii= qx, q = constant.
Figure 1.7.7 Comparison of exact and Galerkin particular solutions, Example 2, case (ii).
Again uh may be expressed in the form (1.7.25), where
A comparison is presented in Fig. 1.7.8. The Galerkin solution is seen to be exact once again at x , and 1, and the derivative is exact at two points.
Let us summarize the salient observations of Example 2:
a. The homogeneous part of uh is exact in all cases, as in Example 1. (A rationale for this is given after Equation (1.7.21).)
b. The Galerkin solution is exact at the endpoints of each subinterval for all cases.
c. In each case, there is at least one point in each subinterval at which uh,x is exact.
Figure 1.7.8 Comparison of exact and Galerkin particular solutions, Example 2, case (iii).
After generalizing to the case of n subintervals in the following section, we shall show in Sec. 1.10 that the above observations are not accidental.
Exercise 1. If the reader has not had experience with calculations of the type presented in this section, it would be worthwhile to reproduce all results, providing all omitted details.
1.8 PIECEWISE LINEAR FINITE ELEMENT SPACE
h h h is n-dimensional, we partition the domain [0, 1] into n nonoverlapping subintervals. The typical subinterval is denoted by [xA, xA + 1], where xA < xA + 1 and A = 1, 2,…, n. We also require x1 = 0 and xn + 1 = 1. The xA’s are called nodal points, or simply nodes. (The terminologies joints and knots are also used.) The subintervals are sometimes referred to as the finite element domains, or simply elements. Notice that the lengths of the elements, hA = xA + 1 − xA, are not required to be equal. The mesh parameter, h, is generally taken to be the length of the maximum subinterval (i.e., h = max hA, A= 1, 2,…, n). The smaller h, the more refined
is the partition, or mesh. If the subinterval lengths are equal, then h = l/n.
The shape functions are defined as follows: Associated to a typical internal node (i.e., 2 ≤ A ≤ n )
whereas for the boundary nodes we have
The shape functions are sketched in Fig. 1.8.1. For obvious reasons, they are referred to variously as hat,
chapeau,
and roof
functions. Note that NA(xB) = δAB, where
Figure 1.8.1 Basis functions for the piecewise linear finite element space.
δAB is the Kronecker delta (i.e., δAB = 1 if A = B, whereas δAB = 0 if A ≠ B). In words, NA takes on the value 1 at node A and is 0 at all other nodes. Furthermore, NA is nonzero only in the subintervals that contain xA.
A typical member wh h and appears as in Fig. 1.8.2. Note that wh is continuous but has discontinuous slope across each element boundary. For this reason, wh,x, the generalized derivative of wh, will be piecewise constant, experiencing discontinuities across element boundaries. (Such a function is sometimes called a generalized step function.) Restricted to each element domain, wh is a linear polynomial in x. In respect to the homogeneous essential boundary condition, wh(1) = 0. Clearly, wh is identically zero if and only if each cA = 0, A = 1,2,…, n.
Figure 1.8.2 A typical member wh h.
h h Nn h. This ensures that uh .
The piecewise linear finite element functions are the simplest and most widely used finite element functions for one-dimensional problems.
Exercise 1. Consider the weak formulation of the one-dimensional model problem:
where w and u are assumed to be smooth on element interiors (i.e., on]xA, xA + 1[, A = 1, 2,…, n), but may suffer slope discontinuities across element boundaries. (Functions of this class contain the piecewise linear finite element space described earlier.) From (1.8.4) and the assumed continuity of the functions, show that:
Arguing as in Sec. 1.4, it may be concluded that the Euler-Lagrange conditions of (1.8.5) are
i. u,xx(x(x) = 0, where x ]xA, xA + 1[ and A = 1, 2,…, n,
ii. − u,x(0+) = h; and
iii. u,x = u, x , where A = 2, 3,…, n.
Observe that (i) is the differential equation restricted to element interiors, and (iii) is a continuity condition across element boundaries. This may be contrasted with the case in which the solution is assumed smooth. In this case the continuity condition is identically satisfied and the summation of integrals over element interiors may be replaced by an integral over the entire domain (see Sec. 1.4).
In the Galerkin finite element formulation, an approximate solution of (i)–(iii) is obtained.
1.9 PROPERTIES OF K
The shape functions NA, A = 1,2,…, n + 1, are zero outside a neighborhood of node A. As a result, many of the entries of K are zero. This can be seen as follows. Let B > A + 1. Then (see Fig. 1.9.1)
The symmetry of K implies, in addition, that (1.9.1) holds for A > B + 1. One says that K is banded (i.e., its nonzero entries are located in a band about the main diagonal). Figure 1.9.2 depicts this property. Banded matrices have significant advantages in that the zero elements outside the band neither have to be stored nor operated upon in the computer. The stiffness matrix arising in finite element analysis is, in general, narrowly banded, lending itself to economical formation and solution.
Figure 1.9.1 If B > A + 1, the nonzero portions of NB and NA do not overlap.
Figure 1.9.2 Band structure of K.
Definition. An n × n matrix A is said to be positive definite if
i. cTAc ≥ 0 for all n-vectors c; and
ii. cTAc = 0 implies c = 0.
Remarks
1. A symmetric positive-definite matrix posesses a unique inverse.
2. The eigenvalues of a positive-definite matrix are real and positive.
Theorem. The n × n matrix K defined by (1.6.11) is positive definite.
Proof
i. Let cA, A = 1, 2,…, n, be the components of c (i.e., c = {cA}), an arbitrary vector. Use these cAh, wh . where the NAh. Then
ii. Assume cTKc = 0. By the proof of part (i),
and consequently wh must be constant. Since wh h, wh(1) = 0. Combining these facts, we conclude that wh(x) = 0 for all x [0, 1], which is possible only if each cA = 0, A = 1, 2,…, n. Thus c =
Note that part (ii) depended upon the definition of K and h.
Summary. K, defined by (1.6.11), is
i. Symmetric
ii. Banded
iii. Positive-definite
The practical consequence of the above properties is that a very efficient computer solution of Kd = F may be performed.
1.10 MATHEMATICAL ANALYSIS
In this section we will show that the observations made with reference to the example problems of Sec. 1.7 are, in fact, general results. To establish these facts rigorously requires only elementary mathematical techniques.
Our first objective is to establish that the Galerkin finite element solution uh is exact at the nodes. To do this we must introduce the notion of a Green’s function.
Let δy(x) = δ(x − y) denote the Dirac delta function. The Dirac function is not a function in the classical sense but rather an operator defined by its action on (continuous) functions. Let wmay be solved by w be continuous on [0, 1]; then we write
By (1.10.1), we see why attention is restricted to continuous functions—δy sifts out the value of w at y. If w were discontinuous at y, its value would be ambiguous. In mechanics, we think of δy visually as representing a concentrated force of unit amplitude located at point y.
The Green’s function problem corresponding to (S) may be stated as follows: Find a function g (i.e., the Green’s function) such that
Note that (1.10.2)–(1.10.4) are simply (Sis replaced by δy and h are taken to be zero.
This problem may be solved by way of formal calculations with distributions, or generalized functions, such as δy. (The theory of distributions is dealt with in Stakgold [5]. A good elementary account of formal calculations with distributions is presented in Popov [9]. This latter reference is recommended to readers having had no previous experience with this topic.) To this end we note that the (formal) integral of δy is the Heaviside, or unit step, function:
The integral of Hy is the Macaulay bracket:
The preceding functions are depicted in Fig. 1.10.1.
Figure 1.10.1 Elementary generalized functions (distributions).
To solve the Green’s function problem, (1.10.2) is integrated, making use of (1.10.5), to obtain:
where c1 is a constant of integration. A second integration and use of (1.10.6) yields
where c2 is another constant of integration. Evaluation of c1 and c2 is performed by requiring (1.10.7) and (1.10.8) to satisfy the boundary conditions. This results in (see Fig. 1.10.2)
Observe that g is piecewise linear. Thus if y = xA (i.e., if y is a node), g h.
Figure 1.10.2 Green’s function.
In the ensuing analysis we will need the variational equation corresponding to the Green’s function problem. This can be deduced from (W) by replacing u by gby δy, and h by 0, viz.,
Equation (1.10.10) holds for all continuous w . The square-integrability of derivatives of functions w actually implies the continuity of all w by a well-known theorem in analysis due to Sobolev. (This result is true only in one dimension. The square-integrability of second derivatives is also required to ensure the continuity of functions defined on two- and three-dimensional domains.)
Theorem. uh(xA) = u(xA), A = 1, 2,…, n + 1 (i.e., uh is exact at the nodes). To prove the theorem, we need to establish two preliminary results.
Lemma 1. a(u − uh, wh) = 0 for all wh h.
Proofh , so we may replace w by wh in the variational equation:
Equation (1.10.11) holds for all wh h. Recall that the Galerkin equation is identical to (1.10.11) except that uh appears instead of u
Lemma 2. u(y) − uh(y) = a(u − uh, g), where g is the Green’s function.
Proof
Note that line 2 is true since u − uh
Proof of Theorem. As we have remarked previously, if y = xA, a node, g h. Let us take this to be the case. Then
The theorem is valid for A = 1, 2,…, n + 1. Strang and Fix [6] attribute this argument to Douglas and Dupont. Results of this kind, embodying exceptional accuracy characteristics, are often referred to as superconvergence phenomena. However, the reader should appreciate that, in more complicated situations, we will not be able, in practice, to guarantee nodal exactness. Nevertheless, as we shall see later on, weighted residual procedures provide a framework within which optimal accuracy properties of some sort may often be guaranteed.
Accuracy of the Derivatives
In considering the convergence properties of the derivatives, certain elementary notions of numerical analysis arise. The reader should make sure that he or she has a complete understanding of these ideas as they subsequently arise in other contexts. We begin by introducing some preliminary mathematical results.
Taylor’s Formula with Remainder
Let fpossess k continuous derivatives and let y and z be two points in [0, 1]. Then there is a point c between y and z such that
The proof of this formula may be found in [7]. Equation (1.10.12) is sometimes called a finite Taylor expansion.
Mean-Value Theorem
The mean-value theorem is a special case of (1.10.12) which is valid as long as k ≥ 1 (i.e., f is continuously differentiable):
Consider a typical subinterval [xA, xA + 1]. We have already shown that uh is exact at the endpoints (see Fig. 1.10.3). The derivative of uh in ]xA, xA + 1[ is constant:
Figure 1.10.3
Theorem. Assume u is continuously differentiable. Then there exists at least one point in ]xA, xA + 1[ at which (1.10.14) is exact.
Proof. By the mean value theorem, there exists a point c ]xA, xA + l[ such that
(We have used (1.10.13) with u, xA, and xA + l, in place of f, y, and z, respectively.) Since u(xA) = uh(xA) and u(xA + 1) = uh(xA + 1), we may rewrite (1.10.15) as
Comparison of (
Remarks
1. This result means that the constant value of u,hx must coincide with u,x somewhere on ]xA, xA + 1[; see Fig. 1.10.4.
2. Without knowledge of u we have no way of determining the locations at which the derivatives are exact. The following results are more useful in that they tell us that the midpoints are, in a sense, optimally accurate, independent of u.
Figure 1.10.4
Let
the error in the derivative at α [xA, xA + 1]. To establish the superiority of the midpoints in evaluating the derivatives, we need a preliminary result.
Lemma. Assume u is three times continuously differentiable. Then
where c1 and c2 are in [xA, xA + 1].
Proof. Expand u(xA + 1) and u(xA) in finite Taylor expansions about α [xA, xA + l], viz.,
Subtracting and dividing through by hA yields
Replacing u(xA + 1) by uh(xA + 1) and u(xA) by uh(xA
Discussion
To determine what (1.10.17) tells us about the accuracy of the derivatives, we wish to think of the situation in which the mesh is being systematically refined (i.e., we let hA approach zero). In this case h²A will be much smaller than hA. Thus, for a given u, if the right-hand side of (1.10.17) is O(hA²),³ the error in the derivatives will be much smaller than if the right-hand side is only O(hA). The exponent of hA is called the order of convergence or order of accuracy. In the former case we would have second-order convergence of the derivative, whereas in the latter case we would have only first-order convergence.
As an example, assume α → xA. Then
As hA → 0, the first term dominates. (We have seen from the example calculations in Sec. 1.8 that the endpoints of the subintervals are not very accurate for the derivatives.)
Clearly any point α [xA, xA + 1] achieves first-order accuracy. We are thus naturally led to asking the question, are there any values of α at which higher-order accuracy is achieved?
Corollary. Let xA + 1/2 ≡ (xA + xA + 1)/2 (i.e., the midpoint). Then
Proof. By (1.10.17)
By the continuity of u,xxx, there is at least one point c between c1 and c2 such that
Remarks
1. From the corollary we see that the derivatives are second-order accurate at the midpoints.
2. If the exact solution is quadratic (i.e., consists of a linear combination of the monomials 1, x, x²), then u,xxx = 0 and—by ((x= constant.
3. In linear elastic rod theory, the derivatives are proportional to the stresses. The midpoints of linear elements
are sometimes called the Barlow stress points, after Barlow [8], who first noted that points of optimal accuracy existed within elements.
Exercise 1. Assume the mesh length is constant (i.e., hA = h, A = 1,2,…, n). Consider the standard finite difference stencil
for u,xx = 0 at a typical internal node, namely,
varies in piecewise linear fashion and so can be expanded as
A, set up the finite element equation associated with node A and contrast it with (?) Set up the finite element equation associated with node 1, accounting for nonzero h. Discuss this equation from the point of view of finite differences. (For further comparisons along these lines, the interested reader is urged to consult [6], Chapter 1.)
Summary. The Galerkin finite element solution uh, of the problem (S), possesses the following properties:
i. It is exact at the nodes.
ii. There exists at least one point in each element at which the derivative is exact.
iii. The derivative is second-order accurate at the midpoints of the elements.
1.11 INTERLUDE: GAUSS ELIMINATION;
HAND-CALCULATION VERSION
It is important for anyone who wishes to do finite element analysis to become familiar with the efficient and sophisticated computer schemes that arise in the finite element method. It is felt that the best way to do this is to begin with the simplest scheme, perform some hand calculations, and gradually increase the sophistication as time goes on.
To do some of the problems we will need a fairly efficient method of solving matrix equations by hand. The following scheme is applicable to systems of equations Kd = F in which no pivoting (i.e., reordering) is necessary. For example, symmetric, positive-definite coefficient matrices never require pivoting. The procedure is as follows:
Gauss Elimination
• Solve the first equation for d1 and elminate d1 from the remaining n − 1 equations.
• Solve the second equation for d2 and eliminate d2 from the remaining n − 2 equations.
• Solve the n − 1st equation for dn − 1 and eliminate dn − 1 from the nth equation.
• Solve the n–th equation for dn.
The preceding steps are called forward reduction. The original matrix is reduced to upper triangular form. For example, suppose we began with a system of four equations as follows:
The augmented matrix corresponding to this system is
After the forward reduction, the augmented matrix becomes
corresponding to the upper triangular system Ud = F′.⁴ It is a simply verified fact that if K is banded, then U will be also.
Employing the reduced augmented matrix, proceed as follows:
• Eliminate dn from equations n − 1, n − 2,…, 1.
• Eliminate dn − 1 from equations n − 2, n − 3,…,