Discover millions of ebooks, audiobooks, and so much more with a free trial

From $11.99/month after trial. Cancel anytime.

The Finite Element Method: Linear Static and Dynamic Finite Element Analysis
The Finite Element Method: Linear Static and Dynamic Finite Element Analysis
The Finite Element Method: Linear Static and Dynamic Finite Element Analysis
Ebook1,324 pages9 hours

The Finite Element Method: Linear Static and Dynamic Finite Element Analysis

Rating: 5 out of 5 stars

5/5

()

Read preview

About this ebook

This text is geared toward assisting engineering and physical science students in cultivating comprehensive skills in linear static and dynamic finite element methodology. Based on courses taught at Stanford University and the California Institute of Technology, it ranges from fundamental concepts to practical computer implementations. Additional sections touch upon the frontiers of research, making the book of potential interest to more experienced analysts and researchers working in the finite element field. In addition to its examination of numerous standard aspects of the finite element method, the volume includes many unique components, including a comprehensive presentation and analysis of algorithms of time-dependent phenomena, plus beam, plate, and shell theories derived directly from three-dimensional elasticity theory. It also contains a systematic treatment of "weak," or variational, formulations for diverse classes of initial/boundary-value problems. Directed toward students without in-depth mathematical training, the text incorporates introductory material on the mathematical theory of finite elements and many important mathematical results, making it an ideal primer for more advanced works on this subject.
LanguageEnglish
Release dateMay 23, 2012
ISBN9780486135021
The Finite Element Method: Linear Static and Dynamic Finite Element Analysis

Related to The Finite Element Method

Related ebooks

Civil Engineering For You

View More

Related articles

Reviews for The Finite Element Method

Rating: 5 out of 5 stars
5/5

3 ratings1 review

What did you think?

Tap to rate

Review must be at least 10 words

  • Rating: 5 out of 5 stars
    5/5
    exallent

Book preview

The Finite Element Method - Thomas J. R. Hughes

1

Fundamental Concepts; a Simple One-Dimensional Boundary-Value Problem

1.1  INTRODUCTORY REMARKS AND PRELIMINARIES

The main constituents of a finite element method for the solution of a boundary-value problem are

i. The variational or weak statement of the problem; and

ii. The approximate solution of the variational equations through the use of finite element functions.

To clarify concepts we shall begin with a simple example.

Suppose we want to solve the following differential equation for u:

where a comma stands for differentiation (i.e., u, xx = d²u/dxis a given smooth, scalar-valued function defined on the unit interval. We write

where [0, 1] stands for the unit interval (i.e., the set of points x such that 0 ≤ x stands for the real numbers. In words, (1.1.2) states that for a given x (xto mean in or a member of. Thus for each x (x.). Also, [0, 1] is said to be the domain of is its range.

, we want it to be a smooth curve without discontinuities or kinks. We do this to avoid technical difficulties. Right now we do not wish to elaborate further as this would divert us from the main theme. At some point prior to moving on to the next chapter, the reader may wish to consult Appendix 1.I, An Elementary Discussion of Continuity, Differentiability and Smoothness, for further remarks on this important aspect of finite element work. The exercise in Sec. 1.16 already uses a little of the language described in Appendix 1.I. The terminology may be somewhat unfamiliar to engineering and physical science students, but it is now widely used in the finite element literature and therefore it is worthwhile to become accustomed to it.

Equation (1.1.1) is known to govern the transverse displacement of a string in tension and also the longitudinal displacement of an elastic rod. In these cases, physical parameters, such as the magnitude of tension in the string, or elastic modulus in the case of the rod, appear in (1.1.1). We have omitted these parameters to simplify subsequent developments.

Before going on, we introduce a few additional notations and terminologies. Let ]0, 1[ denote the unit interval without end points (i.e., the set of points x such that 0 < x < 1). ]0, 1[ and [0, 1] are referred to as open and closed unit intervals, respectively. To simplify subsequent writing and tie in with notation employed later on in multidimensional situations, we shall adopt the definitions

See Fig. 1.1.1.

Figure 1.1.1

At this point, considerations such as these may seem pedantic. Our purpose, however, is to develop a language for the precise articulation of boundary-value problems, which is necessary for good finite element work.

1.2  STRONG, OR CLASSICAL, FORM OF THE PROBLEM

A boundary-value problem for (1.1.1) involves imposing boundary conditions on the function u. There are a variety of possibilities. We shall assume u is required to satisfy

are given constants. Equations (1.2.1) and (1.2.2) require that u at x = 1 and the derivative of u at x = 0, respectively. This set of boundary conditions will later enable us to illustrate certain key features of variational formulations. For obvious reasons, boundary conditions of the type (1.2.1) and (1.2.2) lead to so-called two-point boundary-value problems.

The strong form of the boundary-value problem, (S), is stated as follows:

When we write u,xx = 0 on Ω we mean u,xx (x) = 0 for all x Ω. Of course, the exact solution of (S) is trivial to obtain, namely,

where y and z are used to denote dummy variables. However, this is not the main concern here. We are interested in developing schemes for obtaining approximate solutions to (S) that will be applicable to much more complex situations in which exact solutions are not possible.

Some methods of approximation begin directly with the strong statement of the problem. The most notable example is the finite difference method (e.g., see [1]). The finite element method requires a different formulation, which is treated in the next section.

1.3  WEAK, OR VARIATIONAL, FORM OF THE PROBLEM

To define the weak, or variational, counterpart of (S), we need to characterize two classes of functions. The first is to be composed of candidate, or trial, solutions. From the outset, we shall require these possible solutions to satisfy the boundary condition u. The other boundary condition will not be required in the definition. Furthermore, so that certain expressions to be employed make sense, we shall require that the derivatives of the trial solutions be square-integrable. That is, if u is a trial solution, then

Functions that satisfy (1.3.1) are called H¹-functions; we write u H¹. Sometimes the domain is explicitly included, i.e., u H¹([0, 1]).

Thus the collection of trial solutions, at x = 1. This is written as follows:

is a collection, or set, of objects is indicated by the curly brackets (called braces) in (1.3.2). The notation for the typical member of the set, in this case u, comes first inside the left-hand curly bracket. Following the vertical line ( | ) are the properties satisfied by members of the set.

The second collection of functions is called the weighting functions, or variations. -boundary condition. That is, we require weighting functions, w, to satisfy wand defined by

’s.)

In terms of the preceding definitions, we may now state a suitable weak form, (W), of the boundary-value problem.

Formulations of this type are often called virtual work, or virtual displacement, principles in mechanics. The w’s are the virtual displacements.

Equation (1.3.4) is called the variational equation, or (especially in mechanics) the equation of virtual work.

The solution of (W) is called the weak, or generalized, solution. The definition given of a weak formulation is not the only one possible, but it is the most natural one for the problems we wish to consider.

1.4  EQUIVALENCE OF STRONG AND WEAK FORMS; NATURAL BOUNDARY CONDITIONS

Clearly, there must be some relationship between the strong and weak versions of the problem, or else there would be no point in introducing the weak form. It turns out that the weak and strong solutions are identical. We shall establish this assuming all functions are smooth. This will allow us to proceed expeditiously without invoking technical conditions with which the reader is assumed to be unfamiliar. Proofs of this kind are sometimes euphemistically referred to as formal proofs. The intent is not to be completely rigorous but rather to make plausible the truth of the proposition. With this philosophy in mind, we shall prove the following.

Proposition

a. Let u be a solution of (S). Then u is also a solution of (W).

b. Let u be a solution of (W) Then u is also a solution of (S).

Another result, which we shall not bother to verify but is in fact easily established, is that both (S) and (W) possess unique solutions. Thus, by (a) and (b), the strong and weak solutions are one and the same. Consequently, (W) is equivalent to (S).

Formal Proof

a. Since u is assumed to be a solution of (S), we may write

for any w . Integrating (1.4.1) by parts results in

Rearranging and making use of the fact that −u, x(0) = h and w(1) = 0 results in

Furthermore, since u is a solution of (S), it satisfies u. Finally, since u also satisfies (1.4.3) for all w , u satisfies the definition of a weak solution given by (W).

b. Now u is assumed to be a weak solution. Thus u ; consequently u, and

for all w . Integrating by parts and making use of the fact w(1) = 0 results in

To prove u is a solution of (S) it suffices to show that (1.4.4) implies¹

i. u,xx = 0 on Ω; and

ii. u,x= 0

First we shall prove (i). Define w in (1.4.4) by

where ϕ is smooth; ϕ(x) > 0 for all x Ω = ]0, 1[; and ϕ(0) = ϕ(1) = 0. For example, we can take ϕ(x) = x(1 – x), which satisfies all the stipulated requirements (see Figure 1.4.1). It follows that w(1) = 0 and thus w . Substituting (1.4.5) into (1.4.4) results in

Figure 1.4.1

Since ϕ > 0 on Ω, it follows from (1.4.6) that (i) must be satisfied.

Now that we have established (i), we may use it in (1.4.4) to prove (ii), namely,

That w puts no restriction whatsoever on its value at x = 0. Therefore, we may assume that the w in (1.4.7) is such that w

Remarks

1. The boundary condition –u,xis not explicitly mentioned in the statement of (W). From the preceding proof, we see that this boundary condition is, however, implied by the satisfaction of the variational equation. Boundary conditions of this type are referred to as natural boundary conditions. On the other hand, trial solutions are explicitly required to satisfy the boundary condition u. Boundary conditions of this type are called essential boundary conditions. The fact that solutions of the variational equation satisfy natural boundary conditions is extremely important in more complicated situations which we will consider later on.

2. The method used to prove part (b) of the proposition goes under the name of the fundamental lemma in the literature of the calculus of variations. In essence, it is the methodology that enables us to deduce the differential equations and boundary conditions implied by the weak formulation. To develop correct weak forms for complex, multidimensional problems, it is essential to have a thorough understanding of these procedures.

Now we see that to obtain approximate solutions to the original boundary-value problem we have alternative starting points, i.e., the strong or weak statements of the problem. Finite element methods are based upon the lattercontain infinitely many functions.) The variational equations are then solved in this finite-dimensional context. An explicit example of how to go about this is the subject of the next section. However, we first introduce some additional notations to simplify subsequent writing.

Let

In terms of (1.4.8) and (1.4.9), the variational equation takes the form

Here, a(·, ·) and (·, ·) are examples of symmetric, bilinear forms. What this means is as follows: Let c1 and c2 be constants and let u, υ, and w be functions. Then the symmetry property is

Bilinearity means linearity in each slot; for example,

Exercise 1.    Use the definitions of a(·, ·) and (·, ·) to verify the properties of symmetry and bilinearity.

The above notations are very concise; at the same time they capture essential mathematical features and thus are conducive to a mathematical understanding of variational and finite element methods. Diverse classes of physical problems can be written in essentially similar fashion to (1.4.10). Thus ideas developed and results obtained are seen at once to have very broad applicability.

1.5  GALERKIN'S APPROXIMATION METHOD

We shall now describe a method of obtaining approximate solutions to boundary-value problems based upon weak formulations. Our introduction to this subject is somewhat of an abstract treatment. However, the meaning should be significantly reinforced by the remaining sections of the chapter. It may be worthwhile for the reader to consult this section again after completing the rest of the chapter to make sure a full comprehension of the material is attained.

h hh h with a mesh, or discretization, of the domain Ω, which is parameterized by a characteristic length scale h. h h , respectively. This is written as

where the precise meaning is given in parentheses.² Consequences of (1.5.1) and (1.5.2) are (respectively) that if uh h and wh h, then

hh, are often referred to as function spaces. The terminology space in mathematics usually connotes a linear structure. This has the following meaning: If c1 and c2 are constants and v and w , then c1v + c2w h h due to the inhomogeneous boundary condition. For example, if u1 and u2 , then u1 + u, since u1(l) + uh.

(Bubnov-) Galerkin Method

h is given. Then, to each member vh h, we construct a function uh h by

h is a given function satisfying the essential boundary condition, i.e.,

Note that (1.5.5) satisfies the requisite boundary condition also:

Thus (hh hh h are composed of identical collections of functions. This property will be shown later on to have significant consequences for certain classes of problems.

We now write a variational equation, of the form of (1.4.10), in terms of wh h and uh h:

This equation is to be thought of as defining an approximate (weak) solution, uh.

Substitution of (1.5.5) into (1.5.8), and the bilinearity of a(·, ·) enables us to write

, and h). Equation (1.5.9) is to be used to define vh, the unknown part of uh.

The (Bubnov-) Galerkin form of the problem, denoted by (G), is stated as follows:

Note that (G) is just a version of (Wh.

h h have to be explicitly defined. Before doing this, it is worthwhile to mention a larger class of approximation methods, called Petrov-Galerkin methods, in which vh h. Recent attention has been paid to methods of this type, especially in the context of fluid mechanics. For the time being, we will be exclusively concerned with the Bubnov-Galerkin method. The Bubnov-Galerkin method is commonly referred to as simply the Galerkin method, terminology we shall adopt henceforth. Equation (1.5.9) is sometimes referred to as the Galerkin equation.

Approximation methods of the type considered are examples of so-called weighted residual methods. The standard reference dealing with this subject is Finlayson [3]. For a more succinct presentation containing an interesting historical account, see Finlayson and Scriven [4].

1.6  MATRIX EQUATIONS; STIFFNESS MATRIX K

hh consist of all linear combinations of given functions denoted by N, where A = 1, 2,…, n. By this we mean that if wh h , then there exist constants cA, A = 1, 2,…, n, such that

The NA’s are referred to as shape, basis, or interpolation functions. We require that each NA satisfies

from which it follows by (1.6.1) that whh is said to have dimension n, for obvious reasons.

h h. To this end, we introduce another shape function, Nn , which has the property

(Note Nn hh is given by

and thus

With these definitions, a typical uh h may be written as

where the dA’s are constants and from which it is apparent that uh.

Substitution of (1.6.1) and (1.6.6) into the Galerkin equation yields

By using the bilinearity of a(·, ·) and (·, ·), (1.6.7) becomes

where

Now the Galerkin equation is to hold for all wh h. By (1.6.1), this means for all cA’s, A = 1, 2,…, n. Since the cA’s are arbitrary in (1.6.8), it necessarily follows that each GA, A = 1, 2,…, n, must be identically zero, i.e., from (1.6.9)

Note that everything is known in (1.6.10) except the dB’s. Thus (1.6.10) constitutes a system of n equations in n unknowns. This can be written in a more concise form as follows:

Let

Then (1.6.10) becomes

Further simplicity is gained by adopting a matrix notation. Let

and

Now (1.6.13) may be written as

The following terminologies are frequently applied, especially when the problem under consideration pertains to a mechanical system:

A variety of physical interpretations are of course possible.

At this point, we may state the matrix equivalent, (M), of the Galerkin problem.

The solution of (M) is, of course, just d = K1F (assuming the inverse of K, K1 exists). Once d is known, the solution of (G) may be obtained at any point x by employing (1.6.6), viz.,

Likewise, derivatives of uh, if required, may be obtained by term-by-term differentiation. It should be emphasized, that the solution of (G) is an approximate solution of (W). Consequently, the differential equation and natural boundary condition are only approximately satisfied. The quality of the approximation depends upon the specific choice of NA’s and the number n.

Remarks

1. The matrix K is symmetric. This follows from the symmetry of a(·, ·) and use of Galerkin’s method (i.e., the same shape functions are used for the variations and trial solutions):

In matrix notation

where the superscript T denotes transpose. The symmetry of K has important computational consequences.

2. Let us schematically retrace the steps leading to the matrix problem, as they are typical of the process one must go through in developing a finite element method for any given problem:

The only apparent approximation made thus far is in approximately solving (W) via (G, and h may be approximated, as well as the domain Ω, calculation of integrals, and so on. Convergence proofs and error analyses involve consideration of each approximation.

3. It is sometimes convenient to write

where dn .

1.7  EXAMPLES: 1 AND 2 DEGREES OF FREEDOM

In this section we will carry out the detailed calculations involved in formulating and solving the Galerkin problem. The functions employed are extremely simple, thus expediting computations, but they are also primitive examples of typical finite element functions.

Example 1 (1 degree of freedom)

In this case n = 1. Thus wh = c1N1 and uh = vh h = d1NN2. The only unknown is d1. The shape functions must satisfy N1(1) = 0 and N2(1) = 1 (see (1.6.2) and (1.6.3)). Let us take N1(x) = 1 − x and N2(x) = x. These are illustrated in Fig. 1.7.1 and clearly satisfy the required conditions. Since we are dealing with only 1 degree of freedom, the matrix paraphernalia collapses as follows:

Consequently

In (1.7.7), y .

Figure 1.7.1    Functions for the 1 degree of freedom examples. (These functions are secretly the simplest finite element interpolation functions in a one-element context.)

Figure 1.7.2    The Galerkin solution for the 1 degree of freedom example.

i. = 0. Then

That is, the approximate solution is exact. In fact, it is clear by inspecting (≠ 0).

ii. (x, a constant. Then the particular solutions take the form

and

Equations (1.7.9) and (1.7.10) are compared in Fig. 1.7.3. Note that uhpart is exact at x = 0 and x = 1 and that uhpart,x is exact at x .(It should be clear that it is impossible for uhpart to be exact at all x in the present circumstances. The exact solution, (1.7.9), contains a quadratic term in x, whereas the approximate solution is restricted to linear variation in x by the definitions of N1 and N2.)

Figure 1.7.3    Comparison of exact and Galerkin particular solutions, Example 1, case (ii).

iii. (x) = qxleads to

and

which are compared in Fig. 1.7.4. Again we note that the uhpart is exact at x = 0 and x = 1. There is one point, x , at which uhpart, x is exact.

Let us summarize what we have observed in this example:

a. The homogeneous part of uh is exact in all cases.

, uh is exact at x = 0 and x = 1.

c. For each case, there is at least one point at which uh,x is exact.

Figure 1.7.4    Comparison of exact and Galerkin particular solutions, Example 1, case (iii).

Example 2 (2 degrees of freedom)

In this case n = 2. Thus wh = c1N1 + c2N2, where N1(l) = N2(l) = 0, and uh = d1N1 + d2NN3. where N3(l) = 1. Let us define the NA’s as follows

The shape functions are illustrated in Fig. 1.7.5. Typical wh h and uh h and their derivatives are shown in Fig. 1.7.6. Since n = 2, the matrix paraphernalia takes the following form:

Figure 1.7.5    Functions for the 2 degree of freedom examples. (These functions are secretly the simplest finite element functions in a two-element context.)

Figure 1.7.6    Typical weighting function and trial solution for the 2 degree of freedom example.

Note that due to the shape functions’ discontinuities in slope at x , 1] (e.g., see (1.7.12) and (1.7.15)). We need not worry about the value of the derivative of NA at x (it suffers a discontinuity there and thus is not well-defined classically) since it has no effect on the integrals in (1.7.12). This amounts to employing the notion of a generalized derivative.

We shall again analyze the three cases considered in Example 1.

= 0.

This results in

Again, the exact homogeneous solution is obtained. (The reason for this is that the exact solution is linear, and our trial solution is capable of exactly representing any linear function. Galerkin’s method will give the exact answer whenever possible—that is, whenever the collection of trial solutions contains the exact solution among its members.)

ii(x= constant.

The solution takes the form

The approximate particular solution is compared with the exact in Fig. 1.7.7, from which we see that agreement is achieved at x and 1, and derivatives coincide at x .

iii= qx, q = constant.

Figure 1.7.7    Comparison of exact and Galerkin particular solutions, Example 2, case (ii).

Again uh may be expressed in the form (1.7.25), where

A comparison is presented in Fig. 1.7.8. The Galerkin solution is seen to be exact once again at x , and 1, and the derivative is exact at two points.

Let us summarize the salient observations of Example 2:

a. The homogeneous part of uh is exact in all cases, as in Example 1. (A rationale for this is given after Equation (1.7.21).)

b. The Galerkin solution is exact at the endpoints of each subinterval for all cases.

c. In each case, there is at least one point in each subinterval at which uh,x is exact.

Figure 1.7.8    Comparison of exact and Galerkin particular solutions, Example 2, case (iii).

After generalizing to the case of n subintervals in the following section, we shall show in Sec. 1.10 that the above observations are not accidental.

Exercise 1.    If the reader has not had experience with calculations of the type presented in this section, it would be worthwhile to reproduce all results, providing all omitted details.

1.8  PIECEWISE LINEAR FINITE ELEMENT SPACE

h h h is n-dimensional, we partition the domain [0, 1] into n nonoverlapping subintervals. The typical subinterval is denoted by [xA, xA + 1], where xA < xA + 1 and A = 1, 2,…, n. We also require x1 = 0 and xn + 1 = 1. The xA’s are called nodal points, or simply nodes. (The terminologies joints and knots are also used.) The subintervals are sometimes referred to as the finite element domains, or simply elements. Notice that the lengths of the elements, hA = xA + 1 xA, are not required to be equal. The mesh parameter, h, is generally taken to be the length of the maximum subinterval (i.e., h = max hA, A= 1, 2,…, n). The smaller h, the more refined is the partition, or mesh. If the subinterval lengths are equal, then h = l/n.

The shape functions are defined as follows: Associated to a typical internal node (i.e., 2 ≤ A n )

whereas for the boundary nodes we have

The shape functions are sketched in Fig. 1.8.1. For obvious reasons, they are referred to variously as hat, chapeau, and roof functions. Note that NA(xB) = δAB, where

Figure 1.8.1    Basis functions for the piecewise linear finite element space.

δAB is the Kronecker delta (i.e., δAB = 1 if A = B, whereas δAB = 0 if A ≠ B). In words, NA takes on the value 1 at node A and is 0 at all other nodes. Furthermore, NA is nonzero only in the subintervals that contain xA.

A typical member wh h and appears as in Fig. 1.8.2. Note that wh is continuous but has discontinuous slope across each element boundary. For this reason, wh,x, the generalized derivative of wh, will be piecewise constant, experiencing discontinuities across element boundaries. (Such a function is sometimes called a generalized step function.) Restricted to each element domain, wh is a linear polynomial in x. In respect to the homogeneous essential boundary condition, wh(1) = 0. Clearly, wh is identically zero if and only if each cA = 0, A = 1,2,…, n.

Figure 1.8.2    A typical member wh h.

h h Nn h. This ensures that uh .

The piecewise linear finite element functions are the simplest and most widely used finite element functions for one-dimensional problems.

Exercise 1.    Consider the weak formulation of the one-dimensional model problem:

where w and u are assumed to be smooth on element interiors (i.e., on]xA, xA + 1[, A = 1, 2,…, n), but may suffer slope discontinuities across element boundaries. (Functions of this class contain the piecewise linear finite element space described earlier.) From (1.8.4) and the assumed continuity of the functions, show that:

Arguing as in Sec. 1.4, it may be concluded that the Euler-Lagrange conditions of (1.8.5) are

i. u,xx(x(x) = 0, where x ]xA, xA + 1[ and A = 1, 2,…, n,

ii. − u,x(0+) = h; and

iii. u,x = u, x , where A = 2, 3,…, n.

Observe that (i) is the differential equation restricted to element interiors, and (iii) is a continuity condition across element boundaries. This may be contrasted with the case in which the solution is assumed smooth. In this case the continuity condition is identically satisfied and the summation of integrals over element interiors may be replaced by an integral over the entire domain (see Sec. 1.4).

In the Galerkin finite element formulation, an approximate solution of (i)–(iii) is obtained.

1.9  PROPERTIES OF K

The shape functions NA, A = 1,2,…, n + 1, are zero outside a neighborhood of node A. As a result, many of the entries of K are zero. This can be seen as follows. Let B > A + 1. Then (see Fig. 1.9.1)

The symmetry of K implies, in addition, that (1.9.1) holds for A > B + 1. One says that K is banded (i.e., its nonzero entries are located in a band about the main diagonal). Figure 1.9.2 depicts this property. Banded matrices have significant advantages in that the zero elements outside the band neither have to be stored nor operated upon in the computer. The stiffness matrix arising in finite element analysis is, in general, narrowly banded, lending itself to economical formation and solution.

Figure 1.9.1    If B > A + 1, the nonzero portions of NB and NA do not overlap.

Figure 1.9.2    Band structure of K.

Definition. An n × n matrix A is said to be positive definite if

i. cTAc 0 for all n-vectors c; and

ii. cTAc = 0 implies c = 0.

Remarks

1. A symmetric positive-definite matrix posesses a unique inverse.

2. The eigenvalues of a positive-definite matrix are real and positive.

Theorem. The n × n matrix K defined by (1.6.11) is positive definite.

Proof

i. Let cA, A = 1, 2,…, n, be the components of c (i.e., c = {cA}), an arbitrary vector. Use these cAh, wh . where the NAh. Then

ii. Assume cTKc = 0. By the proof of part (i),

and consequently wh must be constant. Since wh h, wh(1) = 0. Combining these facts, we conclude that wh(x) = 0 for all x [0, 1], which is possible only if each cA = 0, A = 1, 2,…, n. Thus c =

Note that part (ii) depended upon the definition of K and h.

Summary. K, defined by (1.6.11), is

i. Symmetric

ii. Banded

iii. Positive-definite

The practical consequence of the above properties is that a very efficient computer solution of Kd = F may be performed.

1.10 MATHEMATICAL ANALYSIS

In this section we will show that the observations made with reference to the example problems of Sec. 1.7 are, in fact, general results. To establish these facts rigorously requires only elementary mathematical techniques.

Our first objective is to establish that the Galerkin finite element solution uh is exact at the nodes. To do this we must introduce the notion of a Green’s function.

Let δy(x) = δ(x y) denote the Dirac delta function. The Dirac function is not a function in the classical sense but rather an operator defined by its action on (continuous) functions. Let wmay be solved by w be continuous on [0, 1]; then we write

By (1.10.1), we see why attention is restricted to continuous functions—δy sifts out the value of w at y. If w were discontinuous at y, its value would be ambiguous. In mechanics, we think of δy visually as representing a concentrated force of unit amplitude located at point y.

The Green’s function problem corresponding to (S) may be stated as follows: Find a function g (i.e., the Green’s function) such that

Note that (1.10.2)–(1.10.4) are simply (Sis replaced by δy and h are taken to be zero.

This problem may be solved by way of formal calculations with distributions, or generalized functions, such as δy. (The theory of distributions is dealt with in Stakgold [5]. A good elementary account of formal calculations with distributions is presented in Popov [9]. This latter reference is recommended to readers having had no previous experience with this topic.) To this end we note that the (formal) integral of δy is the Heaviside, or unit step, function:

The integral of Hy is the Macaulay bracket:

The preceding functions are depicted in Fig. 1.10.1.

Figure 1.10.1    Elementary generalized functions (distributions).

To solve the Green’s function problem, (1.10.2) is integrated, making use of (1.10.5), to obtain:

where c1 is a constant of integration. A second integration and use of (1.10.6) yields

where c2 is another constant of integration. Evaluation of c1 and c2 is performed by requiring (1.10.7) and (1.10.8) to satisfy the boundary conditions. This results in (see Fig. 1.10.2)

Observe that g is piecewise linear. Thus if y = xA (i.e., if y is a node), g h.

Figure 1.10.2    Green’s function.

In the ensuing analysis we will need the variational equation corresponding to the Green’s function problem. This can be deduced from (W) by replacing u by gby δy, and h by 0, viz.,

Equation (1.10.10) holds for all continuous w . The square-integrability of derivatives of functions w actually implies the continuity of all w by a well-known theorem in analysis due to Sobolev. (This result is true only in one dimension. The square-integrability of second derivatives is also required to ensure the continuity of functions defined on two- and three-dimensional domains.)

Theorem. uh(xA) = u(xA), A = 1, 2,…, n + 1 (i.e., uh is exact at the nodes). To prove the theorem, we need to establish two preliminary results.

Lemma 1. a(u uh, wh) = 0 for all wh h.

Proofh , so we may replace w by wh in the variational equation:

Equation (1.10.11) holds for all wh h. Recall that the Galerkin equation is identical to (1.10.11) except that uh appears instead of u

Lemma 2. u(y) − uh(y) = a(u uh, g), where g is the Green’s function.

Proof

Note that line 2 is true since u uh

Proof of Theorem. As we have remarked previously, if y = xA, a node, g h. Let us take this to be the case. Then

The theorem is valid for A = 1, 2,…, n + 1. Strang and Fix [6] attribute this argument to Douglas and Dupont. Results of this kind, embodying exceptional accuracy characteristics, are often referred to as superconvergence phenomena. However, the reader should appreciate that, in more complicated situations, we will not be able, in practice, to guarantee nodal exactness. Nevertheless, as we shall see later on, weighted residual procedures provide a framework within which optimal accuracy properties of some sort may often be guaranteed.

Accuracy of the Derivatives

In considering the convergence properties of the derivatives, certain elementary notions of numerical analysis arise. The reader should make sure that he or she has a complete understanding of these ideas as they subsequently arise in other contexts. We begin by introducing some preliminary mathematical results.

Taylor’s Formula with Remainder

Let fpossess k continuous derivatives and let y and z be two points in [0, 1]. Then there is a point c between y and z such that

The proof of this formula may be found in [7]. Equation (1.10.12) is sometimes called a finite Taylor expansion.

Mean-Value Theorem

The mean-value theorem is a special case of (1.10.12) which is valid as long as k ≥ 1 (i.e., f is continuously differentiable):

Consider a typical subinterval [xA, xA + 1]. We have already shown that uh is exact at the endpoints (see Fig. 1.10.3). The derivative of uh in ]xA, xA + 1[ is constant:

Figure 1.10.3

Theorem. Assume u is continuously differentiable. Then there exists at least one point in ]xA, xA + 1[ at which (1.10.14) is exact.

Proof. By the mean value theorem, there exists a point c ]xA, xA + l[ such that

(We have used (1.10.13) with u, xA, and xA + l, in place of f, y, and z, respectively.) Since u(xA) = uh(xA) and u(xA + 1) = uh(xA + 1), we may rewrite (1.10.15) as

Comparison of (

Remarks

1. This result means that the constant value of u,hx must coincide with u,x somewhere on ]xA, xA + 1[; see Fig. 1.10.4.

2. Without knowledge of u we have no way of determining the locations at which the derivatives are exact. The following results are more useful in that they tell us that the midpoints are, in a sense, optimally accurate, independent of u.

Figure 1.10.4

Let

the error in the derivative at α [xA, xA + 1]. To establish the superiority of the midpoints in evaluating the derivatives, we need a preliminary result.

Lemma. Assume u is three times continuously differentiable. Then

where c1 and c2 are in [xA, xA + 1].

Proof. Expand u(xA + 1) and u(xA) in finite Taylor expansions about α [xA, xA + l], viz.,

Subtracting and dividing through by hA yields

Replacing u(xA + 1) by uh(xA + 1) and u(xA) by uh(xA

Discussion

To determine what (1.10.17) tells us about the accuracy of the derivatives, we wish to think of the situation in which the mesh is being systematically refined (i.e., we let hA approach zero). In this case h²A will be much smaller than hA. Thus, for a given u, if the right-hand side of (1.10.17) is O(hA²),³ the error in the derivatives will be much smaller than if the right-hand side is only O(hA). The exponent of hA is called the order of convergence or order of accuracy. In the former case we would have second-order convergence of the derivative, whereas in the latter case we would have only first-order convergence.

As an example, assume α xA. Then

As hA → 0, the first term dominates. (We have seen from the example calculations in Sec. 1.8 that the endpoints of the subintervals are not very accurate for the derivatives.)

Clearly any point α [xA, xA + 1] achieves first-order accuracy. We are thus naturally led to asking the question, are there any values of α at which higher-order accuracy is achieved?

Corollary. Let xA + 1/2 ≡ (xA + xA + 1)/2 (i.e., the midpoint). Then

Proof. By (1.10.17)

By the continuity of u,xxx, there is at least one point c between c1 and c2 such that

Remarks

1. From the corollary we see that the derivatives are second-order accurate at the midpoints.

2. If the exact solution is quadratic (i.e., consists of a linear combination of the monomials 1, x, x²), then u,xxx = 0 and—by ((x= constant.

3. In linear elastic rod theory, the derivatives are proportional to the stresses. The midpoints of linear elements are sometimes called the Barlow stress points, after Barlow [8], who first noted that points of optimal accuracy existed within elements.

Exercise 1. Assume the mesh length is constant (i.e., hA = h, A = 1,2,…, n). Consider the standard finite difference stencil for u,xx = 0 at a typical internal node, namely,

varies in piecewise linear fashion and so can be expanded as

A, set up the finite element equation associated with node A and contrast it with (?) Set up the finite element equation associated with node 1, accounting for nonzero h. Discuss this equation from the point of view of finite differences. (For further comparisons along these lines, the interested reader is urged to consult [6], Chapter 1.)

Summary. The Galerkin finite element solution uh, of the problem (S), possesses the following properties:

i. It is exact at the nodes.

ii. There exists at least one point in each element at which the derivative is exact.

iii. The derivative is second-order accurate at the midpoints of the elements.

1.11 INTERLUDE: GAUSS ELIMINATION;

HAND-CALCULATION VERSION

It is important for anyone who wishes to do finite element analysis to become familiar with the efficient and sophisticated computer schemes that arise in the finite element method. It is felt that the best way to do this is to begin with the simplest scheme, perform some hand calculations, and gradually increase the sophistication as time goes on.

To do some of the problems we will need a fairly efficient method of solving matrix equations by hand. The following scheme is applicable to systems of equations Kd = F in which no pivoting (i.e., reordering) is necessary. For example, symmetric, positive-definite coefficient matrices never require pivoting. The procedure is as follows:

Gauss Elimination

• Solve the first equation for d1 and elminate d1 from the remaining n − 1 equations.

• Solve the second equation for d2 and eliminate d2 from the remaining n − 2 equations.

• Solve the n − 1st equation for dn − 1 and eliminate dn − 1 from the nth equation.

• Solve the n–th equation for dn.

The preceding steps are called forward reduction. The original matrix is reduced to upper triangular form. For example, suppose we began with a system of four equations as follows:

The augmented matrix corresponding to this system is

After the forward reduction, the augmented matrix becomes

corresponding to the upper triangular system Ud = F′.⁴ It is a simply verified fact that if K is banded, then U will be also.

Employing the reduced augmented matrix, proceed as follows:

• Eliminate dn from equations n − 1, n − 2,…, 1.

• Eliminate dn − 1 from equations n − 2, n − 3,…,

Enjoying the preview?
Page 1 of 1