Diff Eq Notes
Diff Eq Notes
NOTES
In the first two examples above, we implicitly assume that the variables x and t vary over all possible
inputs in the domain R, while in the third example we assume (x, y)√varies over all possible inputs in the
domain R2 . For some specific examples of functions, e.g. f (x) = x − 5, we assume that the domain
includes√only real numbers x for which it makes sense to plug x into the formula, i.e., the domain of
f (x) = x − 5 is just the interval [5, ∞).
Whenever it does not lead to confusion, we will refer to functions just by their names f , y, F , etc.,
and omit reference to the input variable. This shorthand can occasionally be cause for confusion, and
so the student will be asked to parse which letters are input variables and which are functions from the
context. For example, in the differential equation
x2
y0 = y2 cos y,
the use of the single-variable derivative notation y 0 implies that y = y(x) is a function of the single input
variable x. On the other hand in the differential equation
fxx + fyy = 0,
we note the use of partial derivative notation, and assume that f = f (x, y) is a function of the two
variables x and y. (More comments on derivative notation below.)
dy d2 d4 y δf δ 2
, 2 f (x), 4 , , f (x, y), etc.
dx dx dt δx δyδx
1
2 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
The former group is often referred to as the Lagrange notation, while the latter group is the
Leibniz notation. Both are acceptable and commonly used in practice, and have distinct advantages
and disadvantages over one another. The major advantage of the Leibniz notation is that it is completely
explicit about which variable the given derivative respects. However, the Leibniz notation is notoriously
a source of great confusion for calculus students (this fact may be mostly blamed on obsolete textbook
authorship), and so I ask the student to thoroughly understand its meaning before we proceed to the
main content of the course.
dy
The major drawback of the notation dx is that it is misleading: it looks like a fraction. The derivative
dy
dx is in no way a fraction: if y = y(x) is a real-valued function of a single-variable, then the derivative
dy dy
dx = dx (x) is a real-valued function of a single variable.
Moreover, the components of the notation, dy and dx, are not rigorously defined mathematical objects
of their own. Many textbooks refer to dy, dx, and similar symbols as “infinitesimals,” “differentials”
or “elements” (Isaac Newton called them “fluxions”!). This is an outdated misuse of terminology that
dy
should be abolished from the student’s brain and vocabulary. The notation dx is to be regarded as a
single unified notation which cannot be separated into constituent parts. For instance, we may write
dy
dx = 5x, which is a meaningful mathematical statement, but we do not consider it meaningful to
“multiply on both sides by dx” to obtain the statement “dy = 5xdx.” Unfortunately many authors
choose to use expressions like dy = 5xdx on a regular basis. Whenever we see a statement of the latter
form, we will regard it mentally as just a shorthand for a statement of the former form.
Why has the “fraction notation” become so entrenched when we are not really working with fractions
at all? One popular feature of the Leibniz notation is that it synergizes well with the chain rule for
derivatives and the substitution rule for integrals, in the sense that it makes them look nice visually.
Recall that if y(x) and u(x) are real-valued functions of a single variable x, then the composition y(u(x))
is a real-valued function of x, and the chain rule says that its derivative is equal to
d
dx y(u(x)) = y 0 (u(x)) · u0 (x).
People often switch to the Leibniz notation to write the chain rule in the following condensed form:
dy dy du
= · .
dx du dx
This form is aesthetically pleasing because it looks like cancellation of fractions (again, a source of
confusion). But there is clearly no cancellation happening; the above statement is of the form “function
dy
equals function times function.” Moreover, the statement above is ambiguous because the notation du ,
where y and u are both functions, is rarely if ever defined carefully at the introductory calculus level,
and the input variables, which are of crucial importance, are omitted. The useful content of the chain
dy
rule is that the derivative of the composition y ◦ u may be found by computing dx and du
dx separately,
and then adjusting the input variables. In other words, a correct version of the chain rule, written in
Leibniz notation, should look like:
dy dy du
(x) = (u(x)) · (x),
dx dx dx
Rb R g(b)
a
y 0 (g(x)) · g 0 (x)dx = g(a)
y 0 (u)du.
Rb
(Crucially recall here that the string of symbols a dx means “take the integral of over the
interval [a, b] with respect to x”; the dx is again an inseparable part of the notation and should not be
regarded as a mathematical object all on its own.)
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 3
The substitution rule is usually applied in practice by means of the following mnemonic device: “Set
0
u = g(x). Then du
dx = g (x), and u runs over [g(a), g(b)] as x runs over [a, b], so
Rb Rb R g(b)
a
y 0 (g(x)) · g 0 (x)dx = a
y 0 (u) du
dx dx = g(a)
y 0 (u)du.”
Again we obtain the illusion of fraction cancellation, caused by the judicious omission of input vari-
ables, together with a highly suspect duplication of the symbol u as both a function and a dummy
variable of integration.
The moral of the story here is that the Leibniz notation can be useful (particularly as a mnemonic
device for certain calculus techniques) but it comes with many misleading characteristics, and so I will
ask the student to be very careful mentally when reading and working with these symbols. Especially
be aware, please, when an author is using “differentials” (dx, dy, dt etc.) as a mnemonic device or as a
non-rigorous shorthand notation.
As a last remark, the rigorousness of infinitesimals and infinitesimal notation was a major debate for
many decades and has become an interesting and important part of mathematical history. Our modern
approach to calculus avoids reference to infinitesimals by instead relying on the notion of a limit; this
approach is essentially due to Karl Weierstrass. It is actually possible to make things like differentials
and infinitesimals dy, dx, etc. rigorous notions (for example, google: “differential geometry” or “non-
standard analysis”) but these tools are well beyond the scope of our course and don’t really shed any
light on our present purposes. So we will regard dx, dy, etc. purely as notational conveniences, at their
best, and confusing encumbrances at their worst.
ODE’s PDE’s
dv
dt= −32 uxx + uyy = 0
0
y = 3x2 − 4x fx + f fyy = 0
y 00 + y = 0 δu δu
δx δy =u
d2 x
dt2+ 2 dx
+ 3x = sin t
dt Fx = 5x2 y − 3y
00 0
y + y − 2y = x3
Example 3.3. (1) Find all solutions to the differential equation y 0 = 3x2 − 4x.
d7 x
(5) dt7 + 2 dx
dt + 3x = sin t (degree 7)
Example 3.9. (1) Show that y = c1 sin t + c2 cos t satisfies the ODE y 00 + y = 0 for all real numbers
c1 , c2 .
(2) Find a solution to the IVP y 00 + y = 0, y(0) = 0, y 0 (0) = 1.
Definition 3.10. An ODE (involving the function y of the independent variable x) of order n is called
linear if it may be written in the form
an (x)y (n) + an−1 (x)y (n−1) + ... + a2 (x)y 00 + a1 (x)y 0 + a0 (x)y = f (x),
where a0 , a1 , ..., an and f are functions of x and an is not the zero function. If f is identically the 0
function, then the ODE is called homogeneous. Note that the function y(x) = 0 is a solution to any
homogeneous ODE, called the trivial solution.
Example 3.11. Determine if the following ODE’s are (a) linear or nonlinear, and (b) homogeneous or
nonhomogeneous.
dy
(1) dx + x2 y = x (linear nonhomogeneous)
dy
(2) dx = x3 (linear nonhomogeneous)
d2 x
(7) dt2 + sin x = 0 (nonlinear)
Example 3.12. Solve the following differential equations.
(1) y 0 = cos x (easy)
x
(2) y 0 = √ (substitution rule)
x2 +1
1
(3) y 0 = (trig substitution)
x2 + 16
(4) y 0 = xex (integration by parts)
1
(5) y 0 = (partial fractions)
4 − x2
(6) y 0 = y (remembering what we already know + cleverness)
Definition 3.13 (Informal). Given an ODE (involving a function y of an independent variable x), an
explicit solution is a function y = y(x) which satisfies the equation. An implicit solution is an equation
in y and x, such that if a given differentiable function satisfies the equation, then the function also
satisfies the given ODE.
Example 3.14. Show that an implicit solution to the ODE
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 5
2y − 4x − 5
y0 =
2y − 2x
is given by the equation 2x2 + y 2 − 2xy + 5x = 0. (Hint: implicit differentiation on the solution.)
Example 3.15 (Visualizing Solutions to ODEs). Sketch the slope field associated to the differential
2 2
equation y 0 = e−x . Sketch a possible solution. (For context, recall that e−x does not have a nice
explicit antiderivative.)
If f has first-order partial derivatives which are continuous on an open set containing (t0 , y0 ), then
there exists an open interval (t0 − h, t0 + h) on which there exists one and only one function which is a
solution to the IVP.
Example 4.2. Verify that the IVP y 0 = y, y(0) = 1 has a unique solution.
5 Separable Equations
Definition 5.1. A first-order ODE is called separable if it may be written in the form
g(y)y 0 = f (t),
for some functions g, f in one variable. (Note: some authors will use the shorthand g(y)dy = f (t)dt; see
Section 2 of these notes.)
Example 5.2. We wish to model the population P of a group of apes in the forest, as a function of
time t measured in years. Suppose that the birth rate of the apes is proportional to the population, so
that the birth rate may be written bP (t) for some constant b. Likewise suppose the mortality rate is
proportional to the population, say mP (t) for some constant m. Setting k = b − m, then the rate of
change of the ape population is modeled by
dP
=(Birth Rate)−(Mortality Rate)= bP − mP = kP .
dt
Assuming that the initial population of apes is P0 , find a model P (t) for the population by solving
the above first-order ODE.
Solution. First note that the equation is separable, as the equation P 0 = kP may be rewritten:
1
· P 0 = k.
P
Note that the left-hand side is a substitution rule problem (mentally set u = P (t)). Integration gives us
ln |P (t)| + C1 = kt + C2
6 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
where C1 and C2 are arbitrary constants. Before finishing the problem, let us make a simple observation
that will serve us the remainder of the course: there is no reason to write two arbitrary constants when
one will do. Set C3 = C2 − C1 (so C3 is still arbitrary) and we get
ln |P (t)| = kt + C3 .
Setting C = eC3 (so C is still an arbitrary constant) we find the implicit (general) solution |P (t)| =
Cekt . From the context we know P must be positive, so we write P (t) = Cekt . Lastly since P (0) = y0 ,
we get that C = Cek(0) = y0 , so the solution to our IVP is
P (t) = y0 ekt .
The student should verify that this is indeed a solution to the IVP.
Technique 5.3 (General Method for Solving Separable First-Order ODE’s). (1) First separate all
the y-terms onto the one side of the equation and all of the t-terms onto the other side, so the
equation is in the form
g(y)y 0 = f (t).
dy
(If you wish, you may write y 0 = dt and use the mnemonic device g(y)dy = f (t)dt.)
(2) Integrate both sides of the equation with respect to t. The left-hand side is a substitution-rule
integration, so the y 0 -term ends up disappearing.
(3) Solve the equation for y if possible to obtain a general explicit solution. (You may be stuck with
just an implicit solution.) Replace arbitrary constants as necessary to keep things simple.
et
Example 5.6. Solve the IVP y 0 = , y(0) = −2.
y+1
Example 5.7. Solve the ODE y 0 = sin(t − y) + sin(t + y).
Definition 5.8. An autonomous first-order ODE is one of the form y 0 = f (y), i.e., an an autonomous
first-order ODE is one where the independent variable t does not explicitly appear. An equilibrium
solution to an autonomous ODE is a constant solution, i.e. a function of the form y(t) = C which
satisfies the equation.
y 0 + p(t)y = q(t)
for some functions p and q. The corresponding homogeneous equation is the ODE:
y 0 + p(t)y = 0,
i.e. we obtain the corresponding homogeneous equation by just dropping the q(t) term from the original
equation.
Recall from a previous course that if F1 and F2 are two functions for which F10 = y and F20 = y, then
F2 = F1 + C for some constant C. In other words, any two antiderivatives of a given function y differ
only by a constant. By analogy, the next theorem says that any two solutions to a first-order linear
ODE differ only by a solution to the corresponding homogeneous equation.
Theorem 6.2. Let y 0 + p(t)y = q(t) be a first-order linear ODE, and suppose y1 and y2 are two partic-
ular solutions. Then
y2 = y1 + yh ,
y 0 + p(t)y = 0.
This is immediate because y1 and y2 are solutions to the original ODE; check that:
yh0 + p(t)y = (y20 − y10 ) + p(t)(y2 − y1 ) = (y20 + p(t)y2 ) − (y10 + y1 p(t)) = q(t) − q(t) = 0.
Corollary 6.3 (General Solution for First-Order Linear ODEs, Part I). Let yp be any particular solution
to the first-order linear ODE y 0 + p(t)y = q(t). Then every solution to the ODE is of the form
yp + yh ,
dy 3t t
+ 2
y= .
dt 1+t (1 + t2 )2
1
(1) Show that yp = 1+t 2 is a particular solution.
3t
y0 + y = 0.
1 + t2
8 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
1 3t
y0 = − .
y 1 + t2
Integrating both sides (substitution rule on both sides, with u = 1 + t2 on the right-hand side) yields
ln |y| = − 32 ln |1 + t2 | + C.
2
|y| = eC e−(3/2) ln |1+t | = C|1 + t2 |−3/2 .
Noting that (1 + t2 ) is always positive, and that the positivity or negativity of y may be absorbed
into the arbitrary constant C, we drop absolute value signs and arrive at the general solution
y = C(1 + t2 )−3/2 .
1
Since a particular solution to the original ODE is 1+t2 , we conclude that a general solution to the
original ODE looks like
1
y= + C(1 + t2 )−3/2 .
1 + t2
where c is an arbitrary constant in the domain of p. In other words, µ is the exponentiation of any
antiderivative of p.
Theorem 6.6. Let y 0 + p(t)y = q(t) be a first-order linear ODE, where p(t) is continuous. Then the
general solution to the corresponding homogeneous equation y 0 + p(t)y = 0 is
y = Cµ(t)−1 ,
where µ is any integrating factor and C is any constant. Using the indefinite integral notation, we may
write the general solution as
R
y = e− p(t)dt
.
1
y 0 = −p(t),
y
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 9
R
ln |y| = − p(t)dt.
Let g(t) be any antiderivative of p(t), and set µ(t) = eg(t) , so µ is an integrating factor. Since any
two antiderivatives of p(t) differ only by a constant, we may write
ln |y| = −g(t) + C.
Exponentiating both sides, and consolidating arbitrary constants, yields the desired result:
y = Ce−g(t) = Cµ(t)−1 .
Technique 6.7 (General Method for Solving First-Order Linear ODEs). (1) Given a first-order lin-
0
ear ODE ofR
the form y + p(t)y = q(t), choose a representative µ(t) from the class of integrating
factors e− p(t)dt .
(2) Multiply both sides of the differential equation by µ(t) to obtain y 0 µ(t) + µ(t)p(t)y = µ(t)q(t).
Note that since µ0 (t) = µ(t)p(t), this equation may be rewritten as
(3) Integrate both sides with respect to t. By the product rule, the left-hand side has y(t)µ(t) for
an antiderivative.
y = µ(t)−1
R
q(t)µ(t)dt.
Example 6.9. Solve the IVP y 0 + 5t4 y = t4 , y(0) = 7.
dr π
Example 6.10. Find the general solution to the ODE dt = sin t − r tan t, 0 < t < 2.
Example 6.11. Show that the IVP ty 0 − y = t2 cos t, y(0) = 0 has infinitely many solutions. Why does
this not contradict the Picard-Lindelöf Theorem 4.1?
3t
Example 6.12. Solve y 0 − 2 y = t.
(t − 4)
(To make a connection with multivariable calculus, note that an ODE is exact if and only if the vector
field (M, N ) is conservative, i.e. (M, N ) = ∇f for some potential function f .)
Theorem 7.2 (Test for Exactness). Suppose M and N have continuous partial derivatives on an open
domain D. The first-order ODE M (t, y) + N (t, y)y 0 = 0 is exact if and only if
δ δ
δy M = δt N .
(2) t2 y + 5ty 2 y 0 = 0
Theorem 7.4 (Multivariable Chain Rule). Let f (a, b) be a differentiable function of two variables and
let g(t) and h(t) be differentiable functions of one variable. Then
d
dt f (g(t), h(t)) = fa (g(t), h(t))g 0 (t) + fb (g(t), h(t))h0 (t).
Technique 7.5 (General Method for Solving First-Order Exact ODEs). (1) Given an exact equa-
tion M (t, y) + N (t, y)y 0 = 0, find a potential function f (a, b) for which M (t, y) = fa (t, y) and
N (t, y) = fb (t, y), and rewrite the equation as
fa (t, y) · 1 + fb (t, y) · y 0 = 0.
d
(2) Note, using the multivariable chain rule, that the above is of the form dt f (t, y(t)) = 0. Therefore
a general solution to the ODE is f (t, y) = C for an arbitrary constant C.
y 0 + p(t)y = q(t)y n
(2) Make the substitution w = y 1−n , so w0 = (1 − n)y −n y 0 . This transforms the equation into
w0 + (1 − n)p(t)w = (1 − n)q(t).
(3) The new equation as above is linear; so solve for w using Technique 6.7.
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 11
(4) Once you have found w, un-substitute and solve for y by taking y = wn−1 . (Warning: if n is
odd, and hence w is an even power of y, then you will probably be stuck with just an implicit
solution.)
y
Example 8.3. Solve y 0 + x = 1
xy 2 , x > 0.
Our use of the term homogeneous here bears absolutely no relation to its meaning in Definition 3.10.
With sympathy for the reader, we remark that homogeneous is just one of those words which over time
has accumulated an excess of different meanings in mathematical writing (see also for instance regular
or normal ), and this is sadly beyond the author’s power to change.
Technique 9.4 (General Method for Solving Homogeneous Equations). (1) The strategy here is to
make a clever substitution and thereby reduce to the separable case. Given an ODE of the form
where both M and N are homogeneous of degree n, choose one of the following substitutions:
y
u= t (hence y = ut)
or
t
v= y (hence t = vy).
Example 9.5. Solve t3 + y 3 − ty 2 y 0 = 0 twice, using each of the substitutions y = ut and t = vy.
Assume that t > 0 and y > 0.
(1) Separable Equations - Of the form g(y)y 0 = f (t). To solve: Separate and integrate.
(2) Linear Equations - Of the form y + p(t)y 0 = q(t). To solve: Multiply by an integrating
factor.
(3) Exact Equations - Of the form M (t, y) + N (t, y)y 0 = 0 where M , N are the partial derivatives
of some potential function. To solve: Find a potential function.
(4) Bernoulli Equations - Of the form y 0 + p(t)y = q(t)y n . To solve: Substitute and reduce
to the linear case.
(5) Homogeneous Equations - Of the form M (t, y)+N (t, y)y 0 = 0 where M , N are homogeneous.
To solve: Substitute and reduce to the separable case.
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 13
where y is population at time t, and y0 is some initial population. Recall that the solution to the above
IVP is
y = y0 ekt .
Despite the simplicity and effectiveness of the Malthus model, it has drawbacks. Primarily, the
Malthus model assumes that population growth is always directly proportional to population size, i.e.
y 0 = ky for constant k, and thus all population growth must be exponential. This is a situation that is
usually not encountered in the long term in reality, as typically the growth of a population depends on
the natural resources available in a given system. One way to try to reflect this reality is by changing
our fundamental assumption: we still assume that the growth y 0 is proportional to the total population
y, but now as the population y grows larger, the population growth will be inhibited due to scarcity of
resources. In other words, instead of assuming that y 0 is a constant k times population size at any time,
we assume that y 0 is a function (k − ay) times population size, where k and a are constant; as y increases
the proportion of y 0 to y decreases. This new assumption leads to consideration of the following famous
differential equation.
Definition 10.2. The logistic equation is the ODE
y 0 = (k − ay)y = ky − ay 2 ,
Example 11.3. A pie is removed from a 350◦ Fahrenheit oven and placed to cool in a room with
temperature 75◦ Fahrenheit. In 15 minutes, the pie has a temperature of 150◦ Fahrenheit. Determine
the time required to cool the pie to a temperature of 80◦ .
14 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
where V (t) is the volume of liquid in the tank at time t. This function V is modelled by the ODE
V 0 = R1 − R2 .
Example 12.2. Suppose S1 = 1, R1 = 4, and R2 = 3. In addition, suppose the tank initially holds 500
gallons of liquid and 250 pounds of salt.
(1) Find the amount of salt contained in the tank at any time.
(2) Determine how much time is required for 1000 pounds of salt to accumulate.
(3) If the tank holds a maximum of 800 gallons, can 1000 pounds of salt be present in the tank at
any time before the tank reaches its maximum capacity?
We visualize the complex numbers as a set by identifying each complex number a + bi with the point
(a, b) in R2 ; this gives a one-to-one correspondence between C and R2 and leads us to commonly refer
to C as the complex plane.
Utilizing this visual representation, the modulus of a complex number z = a + bi, denoted |z|, is the
Cartesian distance from (a, b) to the origin, i.e.
√
|z| = a2 + b2 .
In addition, by identifying C with R2 , we can formally talk about limits, continuity, etc. in C by
declaring that a limit exists in C if and only if the limit of the associated points exists in R2 (limits in
R2 were studied in the student’s multivariable calculus course).
Fact 13.2. Given any non-zero complex number z = a+bi, there exists a unique angle θ, with 0 ≤ θ < 2π,
called the argument of z, for which
By associating a + bi and c + di in C with the points (a, b) and (c, d) in R2 in the usual way, it
is easy to see that complex addition corresponds exactly to vector addition in R2 . However, complex
multiplication (defined below) clearly distinguishes C from R2 :
It is straightforward to verify that the usual commutative properties of addition and multiplication,
associative properties of addition and multiplication, and distributive property of multiplication are true
in C just like they are in R.
Example 13.4. Let z = 1 + i and w = 2 + 3i. Compute z + w and zw.
Definition 13.5. The complex conjugate of a complex number z = a + bi, denoted z, is the number
z = a − bi.
Example 13.6. Show that for any complex number z, we have zz = |z|2 .
Example 13.7. Show that for any complex number z = a + bi and any complex number w = c + di 6= 0,
the quotient wz exists and is a unique complex number. Therefore division makes sense in C.
Initially the operations of multiplication and division in C may seem fairly mysterious. We will
understand them better after we observe Euler’s formula below, but in order to get Euler’s formula we
need to recall the following facts, developed in a previous calculus course.
Fact 13.8 (Maclaurin Series for ex , sin x, and cos x). The functions ex , sin x, and cos x have the fol-
lowing Maclaurin series representations:
∞
x
X xk x2 x3 x4
(1) e = =1+x+ + + + ...
k! 2! 3! 4!
k=0
∞
X x2k+1 x3 x5 x7
(2) sin x = (−1)k =x− + − + ...
(2k + 1)! 3! 5! 7!
k=0
∞
X x2k x2 x4 x6
(3) cos x = (−1)k =1− + − + ...
(2k)! 2! 4! 6!
k=0
Moreover each of the series above have interval of convergence (−∞, ∞) = R. In other words, the
equalities above hold for all real numbers x.
Until now we have always regarded the functions ex , sin x, and cos x as real-valued functions of a real
variable. The following fact says that there is a meaningful way to extend the functions ex , sin x, and
cos x to be complex-valued functions of a complex variable, that is, to have domain C and codomain
C. This fact is incredibly useful and non-obvious, and the proof is not very difficult to understand,
but is a little beyond the scope of our course and so we omit it. (One typically sees this proof in an
undergraduate complex analysis course.)
Fact 13.9. Each of the Maclaurin series in Fact 13.8 above converges not just for all real numbers x in
R, but in fact for all complex numbers x in C. Of course if x is real then ex , sin x, and cos x are real,
but if x is complex non-real then ex , sin x, and cos x are complex non-real.
Moreover, the function ex , now regarded as a complex function of a complex variable, still obeys the
exponentiation rule: for any two complex numbers z and w, ez+w = ez ew .
Theorem 13.10 (Euler’s Formula). For all real numbers x,
Now it is time to harvest the corollaries. The first is a cute fact relating some of our favorite mathe-
matical constants.
Corollary 13.11. eπi = −1.
Proof. Take x = π in Euler’s formula.
The next two shed a great deal of light on the connection between the planar representation of C and
the multiplication in C.
Corollary 13.12. Suppose z is a complex number with argument θ. Then z = |z|eiθ .
Corollary 13.13. Let z be a complex number with argument θ1 , so z = |z|eiθ1 , and let w be a complex
number with argument θ2 , so w = |w|eiθ2 . Then the product zw is uniquely determined by the following
facts:
(1) zw has modulus |zw| = |z||w|; and
Recall also that a second-order linear ODE in the above form is homogeneous if f (t) = 0. A
second-order linear homogeneous equation has constant coefficients if p(t) = p and q(t) = q are just
constants.
Example 14.2. Consider the ODE y 00 − 2y 0 − 15y = 0.
(1) Show that the equation has at least two solutions of the form y = eλ1 t , y = eλ2 t , where λ1 and
λ2 are real numbers.
(2) Show that any function of the form y = C1 eλ1 t +C2 eλ2 t , where C1 and C2 are arbitrary constants,
are also solutions to the ODE.
Solution - Characteristic Polynomial with Distinct Real Roots. (a) Let us drop subscripts for a moment
and consider functions of the form y = eλt . Differentiating once and twice, we get
y 0 = λeλt ;
y 00 = λ2 eλt .
Judging from the above, y = eλt will be a solution if either λ2 − 2λ − 15 = 0 or eλt = 0. Since the
latter equality never happens (eλt is always positive), we need λ to be a root of the quadratic polynomial
λ2 − 2λ − 15 = (λ + 3)(λ − 5). So it suffices to take λ1 = −3 and λ2 = 5, i.e. y = e−3t and y = e5t are
solutions to the ODE (this is easy to verify directly).
(b) Suppose y1 = e−3t and y2 = e5t , so y1 and y2 are solutions to the ODE, and C1 and C2 are
arbitrary constants. Then taking y = C1 y1 + C2 y2 , we get
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 17
So y = C1 y1 + C2 y2 is a solution as well.
In the above example, given a second-order linear homogeneous ODE with constant coefficients, we
found infinitely many solutions of the form y = C1 eλ1 t + C2 eλ2 t . We will see in the next section that in
fact all of its solutions have this form.
The example and remarks above suggest that given a second-order linear homogeneous ODE with
constant coefficients ay 00 + by 0 + cy = 0, the task of finding solutions y = eλt may reduce in general to
the task of finding solutions to the associated polynomial equation aλ2 + bλ + c = 0.
Definition 14.3. Given a second-order linear homogeneous ODE with constant coefficients ay 00 + by 0 +
cy = 0, the associated characteristic polynomial is
aλ2 + bλ + c.
aλ2 + bλ + c = 0.
Theorem 14.4. Let ay 00 + by 0 + cy = 0 be a second-order linear homogeneous ODE with constant
coefficients. If r is a solution to the characteristic equation, then y = ert is a solution to the ODE.
In our previous example, the characteristic polynomial happened to have two distinct real roots; what
happens if the polynomial has complex roots, or a single repeated real root? The next two examples
(and their preceding observations) address these situations.
Theorem 14.5. Let ay 00 + by 0 + cy = 0 be a second-order linear homogeneous ODE with constant
coefficients. A complex-valued function y is a solution to the ODE if and only if both Re y and Im y are
solutions.
Proof. Write f = Re y and g = Im y, so y = f + gi. Plugging in and using the linearity of the derivative,
we get
The left-hand side will be 0 if and only if both the real part and the complex part of the right-hand
side are 0. So y is a solution if and only if both f and g are.
Example 14.6. Solve y 00 + 4y 0 + 20y = 0.
Solution - Characteristic Polynomial with Complex Conjugate Roots. We proceed as in the previous ex-
ample by assuming that y has solutions of the form y = eλt . Differentiating to find y 0 = λeλt and
y 00 = λ2 eλt , we can plug into the differential equation and solve to obtain (as in the previous example):
(λ2 + 4λ + 20)eλt = 0.
Since eλt above is never zero, we have λ2 + 4λ + 20 = 0 and we wish to solve for λ. The quadratic
formula gives us:
λ = −2 + 4i or λ = −2 − 4i.
18 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
We may safely conclude that the complex-valued functions of a complex variable y = e(−2+4i)t and
y = e(−2−4i)t are solutions to the given differential equation. But we are interested in finding only
real-valued functions of a real variable which satisfy the ODE. Now given a complex function y, it is
easy to check that y satisfies a given differential equation if and only if both its real part Re y and its
imaginary part Im y satisfy the same equation. Thus to find some real-valued solutions, it suffices to
take the real and complex parts of our complex solutions using Euler’s formula. Recall that
and likewise
Thus both of the functions y = e−2t cos(4t) and y = e−2t sin(4t) are solutions of the equation. (The
student should verify this manually.) And as in the previous example, it will turn out that all solutions
to the ODE are of the form
where C1 and C2 are arbitrary constants. (We will observe this rigorously in the next section.)
Theorem 14.7. Let ay 00 + by 0 + cy = 0 be a second-order linear homogeneous ODE with constant
coefficients. Suppose the characteristic polynomial has a single repeated real root, i.e. aλ2 + bλ + c) =
a(λ − r)2 for some real number r. Then y = ert is a solution, and y = tert is also a solution.
Proof. The fact that y = ert is a solution follows from what we’ve already done. The fact that y = tert
is a clever non-obvious observation. Let us see that it’s true. Compute by the product rule:
and
λ2 + 2λ + 1 = 0.
The above factors into (λ + 1)2 = 1, and thus has only one real solution λ = −1. So y = e−t and
y = te−t are solutions. We conclude that y = C1 e−t + C2 tet is a general solution to the ODE.
Technique 14.9 (General Method for Solving Second-Order Linear Homogeneous ODE’s with Constant Coefficients).
Given an ODE of the form ay 00 + by 0 + c = 0, solve the associated characteristic equation
aλ2 + bλ + c = 0.
(2) If the characteristic equation has two distinct real solutions λ1 and λ2 , then a general solution
to the ODE is given by y = C1 eλ1 t + C2 eλ2 t .
(3) If the characteristic equation has one repeated real solution λ, then a general solution to the
ODE is given by y = C1 eλt + C2 teλt .
(4) If the characteristic equation has two complex conjugate solutions z = a + bi and z = a − bi,
then a general solution to the ODE is given by y = C1 eat cos(bt) + C2 eat sin(bt).
Then there exists one and only one solution to the IVP on the interval I.
As we proceed, we will need to recall some terminology from linear algebra.
Definition 15.2. Let f1 , f2 , ..., fn be a finite collection of functions of an independent variable t. If
c1 , c2 , ..., cn are any constants, then the function
c1 f1 + c2 f2 + ... + cn fn
The functions f1 , ..., fn are called linearly dependent if there exist constants c1 , c2 , ..., cn , at least
one of which is non-zero, for which
c1 f1 + c2 f2 + ... + cn fn = 0.
If f1 , ..., fn are not linearly dependent, they are called linearly independent.
Example 15.3. Determine if the following pairs of functions are linearly dependent or linearly inde-
pendent.
(1) f1 (t) = 2t; f2 (t) = 4t
y1 (t0 ) y 0 (t0 )
= 10 .
y2 (t0 ) y2 (t0 )
y 0 (t )
Let C = yy12 (t0) 1 0
(t0 ) = y20 (t0 ) . Now consider the function y = y1 − Cy2 . By the Principle of Superposition
Theorem 15.5, y is a solution to the given ODE. In addition, notice that
y(t0 ) = y1 (t0 ) − Cy2 (t0 ) = 0 and y 0 (t0 ) = y10 (t0 ) − Cy20 (t0 ) = 0.
In other words, y = y1 −Cy2 is a solution to the IVP y 00 +p(t)y 0 +q(t)y = 0, y(t0 ) = 0, y 0 (t0 ) = 0. But
the constant zero function 0 is also a solution to this IVP! Therefore by the Existence and Uniqueness
Theorem 15.1, we must have y1 − Cy2 = 0. It follows that y1 = Cy2 , whence y1 and y2 are a linearly
dependent pair, as claimed.
Lemma 15.9. Let y1 and y2 be functions. Then d
dt W (y1 , y2 ) = y1 y200 − y100 y2 .
Proof.
d d
W (y1 , y2 ) = (y1 y20 − y10 y2 )
dt dt
= (y1 y200 + y10 y20 ) − (y10 y20 + y100 y2 )
= y1 y200 − y100 y2 .
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 21
Lemma 15.10 (Abel’s Theorem). Let y1 and y2 be two solutions of the second-order linear homogeneous
ODE y 00 + p(t)y 0 + q(t)y = 0, where p and q are continuous functions of t. Let P (t) be an antiderivative
of p(t). If y1 and y2 are linearly independent, then
Therefore,
0 = y2 · qy1 − y1 · qy2
= y2 (−y100 − py10 ) − y1 (−y200 − py20 )
= (y1 y200 − y100 y2 ) + p(y1 y20 − y10 y2 )
d
= W (y1 , y2 ) + pW (y1 , y2 ).
dt
The last line above follows from the definition of the Wronskian plus Lemma 15.9. So the above
computation yields a separable ODE in t and W (t) = W (y1 , y2 )(t):
W 0 + p(t)W = 0.
Let P (t) be some antiderivative of p(t), so P 0 = p. By Theorem 6.6, all solutions to the ODE above
have the form W (t) = Ce−P (t) , where C is a constant. Since y1 and y2 are independent, by Theorem
15.8, we must have C 6= 0.
Corollary 15.11. Let y1 and y2 be two solutions of the second-order linear homogeneous ODE y 00 +
p(t)y 0 + q(t)y = 0, where p and q are continuous functions of t. If y1 and y2 are linearly independent,
then W (y1 , y2 )(t) 6= 0 for all t, i.e. W (y1 , y2 ) is nowhere zero.
Theorem 15.12 (General Solutions to Second-Order Linear Homogeneous ODE’s). Let y1 and y2 be two
linearly independent solutions to the second-order linear homogeneous ODE y 00 + p(t)y 0 + q(t)y = 0 where
p and q are continuous functions of t. Then all solutions to the ODE are of the form y = C1 y1 + C2 y2
for some constants C1 and C2 .
Proof. Let Y be an arbitrary solution to the ODE; we will in fact compute the constants C1 and C2
for which Y = C1 y1 + C2 y2 ! Fix any real number t0 in the domain of Y . By Corollary 15.11, we have
W (y1 , y2 )(t0 ) 6= 0. Therefore it makes sense to set:
W (Y, y2 )(t0 )
C1 =
W (y1 , y2 )(t0 )
and
W (y1 , Y )(t0 )
C2 = .
W (y1 , y2 )(t0 )
22 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
Set Z = C1 y1 +C2 y2 . Note that Z is a solution to the ODE by the Principle of Superposition Theorem
15.5. If we can show Y = Z, then we will have shown that Y is a linear combination of y1 and y2 , and
hence the proof will be complete. Let us compute Z(t0 ) and Z 0 (t0 ):
Thus we have shown that Y and Z are two solutions to the ODE which agree on the initial conditions
y(t0 ) = Y (t0 ) = Z(t0 ) and y 0 (t0 ) = Y 0 (t0 ) = Z 0 (t0 ). But by the Existence and Uniqueness Theorem
15.1, there is only one such function! Therefore Y = Z, and this completes the proof.
The theorem above justifies the general solutions we gave in Technique 14.9, and motivates the
following definition.
Definition 15.13. A pair of functions {y1 , y2 } is called a fundamental set of solutions to a linear
homogeneous ODE y 00 + p(t)y 0 + q(t)y = 0 if y1 and y2 are both solutions and if they are linearly
independent. If {y1 , y2 } is a fundamental set of solutions, then all solutions to the ODE are linear
combinations of y1 and y2 .
where yh is some solution to the corresponding homogeneous equation. In other words, two solutions
differ only by a solution to the corresponding homogeneous equation.
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 23
Proof. This proof mimics exactly the proof we gave long ago in the first-order case. With y1 and y2
given, set yh = y2 − y1 . Then plug yh into the corresponding homogeneous equation and see that it is a
solution.
Example 16.3. Consider the ODE y 00 − 4y = −3 sin t.
(1) Verify that yp (t) = 53 sin t is a solution.
(2) Find all solutions.
Solution. Checking (a) is easy, so we proceed with (b). Since we know a particular solution (from part
(a)), by the previous theorem we can find all possible solutions simply by finding all solutions to the
corresponding homogeneous equation:
y 00 − 4y = 0.
For this we consider the characteristic equation λ2 − 4 = 0. This has two distinct roots 2 and −2, so
a general solution is given by
y = C1 e2t + C2 e−2t .
Therefore all possible solutions to the original ODE have the form
y= 3
5 sin t + C1 e2t + C2 e−2t .
The next technique is much more easily understood via practice than by reading the description, but
we will attempt to describe the process in words anyway. See the examples following the technique.
Technique 16.4 (Guessing a Particular Solution to a Second-Order Linear Non-Homogeneous ODE
with Constant Coefficients). Given an ODE of the form ay 00 + by 0 + cy = f (t), we look at the function
f (t) on the right-hand side. This method is only effective if f (t) is a linear combination of functions
of the forms tm , tm eαt , tm eαt cos(βt), and tm eαt sin(βt), where m is a positive integer and α, β are real
numbers.
(1) If f is indeed a linear combination of functions of this form, then we consider the following sets
of functions:
(c) F = {eαt cos(βt), teαt cos(βt), ..., tm eαt cos(βt), eαt sin(βt), teαt sin(βt), ..., tm eαt sin(βt)} if
tm eαt cos(βt) or tm eαt sin(βt) appears in f .
(2) If none of the functions listed above appear as solutions to the corresponding homogeneous equa-
tion, then we guess that a particular solution yp to the equation will be a linear combination of
the functions as above.
(3) On the other hand, if some functions in F appear as solutions to the corresponding homogeneous
equation, then we multiply F by tr to obtain a new set tr F , where r is the smallest integer for
which no function in tr F is a solution to the corresponding homogeneous equation. In this case
we guess that yp is a linear combination of the functions above, but with F replaced by tr F .
Technique 16.5 (Method of Undetermined Coefficients). Given an ODE of the form ay 00 + by 0 + cy =
f (t):
(1) Solve the corresponding homogeneous equation, to obtain a general solution yh .
24 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
(2) Guess the form of a particular solution yp to the ODE, as in the previous technique.
(3) With yp written as a general linear combination (with undetermined coefficients) of functions,
compute yp0 and yp00 and plug into the ODE. Equate the coefficients of like terms, and solve for
the coefficients. This gives a particular solution yp .
(4) A general solution to the original ODE is yp + yh (where yp is a particular function, and yh
denotes infinitely many functions).
Example 16.6. Solve y 00 + 5y 0 + 6y = 2et + 4 cos t.
Solution. First we go ahead and solve the corresponding homogeneous equation:
y 00 + 5y 0 + 6y = 0.
By the usual method of considering the characteristic equation, it is easy to see that a fundamental
set of solutions to the corresponding homogeneous equation is {e2t , e3t }. So a general solution is of the
form yh = C1 e2t + C2 e3t .
We note the 2et and the 4 cos t on the right-hand side, and we guess that a particular solution yp
will be a linear combination of the functions et , cos t, and sin t. (Because none of these functions is a
solution to the corresponding homogeneous equation, there are no wrinkles here- for remarks on what
to do if one of our guesses appears as a solution to the corresponding homogeneous equation, see the
next example). In other words, we guess that
Aet − B cos t − C sin t + 5(Aet − B sin t + C cos t) + 6(Aet + B cos t + C sin t) = 2et + 4 cos t.
Solution. First we solve the corresponding homogeneous equation y 00 − 3y 0 = 0 by considering the char-
acteristic equation λ2 − 3λ = λ(λ − 3) = 0. This has roots 0 and 3, so a fundamental set of solutions is
{e0t , e3t } = {1, e3t }, and a general solution has the form
yh = C1 + C2 e3t .
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 25
Now looking at the right-hand side of our original ODE, we are inclined to guess that a particular so-
lution yp will look like some linear combination of the functions e3t , 1, and t. However, e3t and 1 already
appear as solutions to the corresponding homogeneous equation; therefore, we multiply everything by t
(in other words by t1 , where 1 is the least power which will prevent all functions from being solutions to
the corresponding homogeneous equation). So we guess that yp is a linear combination of the functions
te3t , t, and t2 :
Regrouping, we have
It follows from the above that 3A = −1 and −6C = −2, whence A = − 31 and C = 13 . Since
2C − 3B = 0, we can also deduce that B = 29 . So a particular solution to the ODE is given by
yp = − 13 te3t + 29 t + 13 t2 .
We conclude that any solution to the ODE must have the form
y = − 13 te3t + 29 t + 13 t2 + C1 + C2 e3t .
Now it remains only to solve for the initial conditions. We first pin down C2 by considering y 0 :
8
9 = y 0 (0) = 3C2 − 1
3 + 2
9
whence C2 = 13 . Now we can use our other initial condition y(0) = 0 to find C1 :
1
0 = y(0) = C1 + C2 = C1 + 3
26 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
y(t) = − 31 te3t + 29 t + 31 t2 − 1
3 + 13 e3t .
u01 y1 + u02 y2 = 0
y 0 = u1 y10 + u2 y20 .
u1 y100 + u01 y10 + u2 y200 + u02 y20 + p(u1 y10 + u2 y20 ) + q(u1 y1 + u2 y2 ) = f .
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 27
u1 [y100 + py10 + qy1 ] + u2 [y200 + py20 + qy2 ] + u01 y10 + u02 y20 = f .
Since y1 and y2 are solutions to the corresponding homogeneous equation, the above reduces to the
following boxed equation:
With these computations completed, the question “Is the theorem true?” reduces to the question
“Do there exist differentiable functions u1 and u2 which satisfy the two boxed equations above?”, in the
sense that if the answer to the latter question is “Yes” then so is the answer to the former question.
Regard u01 and u02 as unknowns, and y1 and y2 as given- then the two boxed equations give a linear
system which is equivalent to the following matrix equation:
u01
y1 y2 0
= .
y10 y20 u02 f
(In the above we have omitted the argument t from each of the functions, but we are implicitly
considering the
equation above for all possible values of t.)
y1 y2
Let A = . Since y1 and y2 are linearly independent, the determinant det A = W (y1 , y2 )
y10 y20
is non-zero for all t, and hence the matrix A(t) is invertible for all t. Moreover, we can compute the
inverse using the familiar formula for 2 × 2-matrix inverses:
y20
−1 1 −y2
A = .
W (y1 , y2 ) −y10 y1
Multiplying on the left of our matrix equation by A−1 on both sides, we get
u01 y20
1 −y2 0 1 −y2 f
= = .
u02 W (y1 , y2 ) −y10 y1 f W (y1 , y2 ) y1 f
y2 f y1 f
In other words the equation has a solution and we get u01 = − and u02 = . The
W (y1 , y2 ) W (y1 , y2 )
functions u1 and u2 may now be obtained through integration:
−y2 f
Z Z
y1 f
u1 = dt and u2 = dt.
W (y1 , y2 ) W (y1 , y2 )
Technique 17.3 (General Method for Solving Second-Order Linear ODE’s). Given an ODE of the form
y 00 + p(t)y 0 + q(t)y = f (t):
(1) Find a fundamental set of solutions {y1 , y2 } to the corresponding homogeneous equation y 00 +
p(t)y 0 + q(t)y = 0.
(4) A particular solution to the ODE is given by yp = u1 y1 + u2 y2 , where u1 and u2 are chosen as
above.
Note that the solution set to this 6-th order linear homogeneous ODE is a 6-dimensional vector space.
This nice correspondence always happens.
Example 18.6. Solve 2y (6) − 7y (5) − 4y (4) = 0.
Theorem 18.7 (DeMoivre’s Formula). For any real numbers x and and n, we have
Note that the ωk ’s form the vertices of a regular n-sided polygon inscribed within the unit circle of the
complex plane. Also note that ωk = (ω1 )k for each k and that ωn = 1.
Proof. This is an easy consequence of DeMoivre’s formula (or Euler’s formula), if you note that e2πki = 1
for any integer value of k.
Example 18.9. Solve y (8) − y = 0.
Solution. The characteristic equation is λ8 − 1 = 0, or λ8 = 1. Thus the solutions are the 8-th roots of
unity:
√ √ √ √
2 2 2 2
λ = 1, −1, ±i, 2 ± 2 i, − 2 ± 2 i.
The above roots yield the following fundamental set of solutions to the ODE:
√ √ √ √ √ √ √ √
{et , e−t , cos t, sin t, e 2t/2
cos( 2
2 t), e
2t/2
sin( 2
2 t), e
− 2t/2
cos( 2
2 t), e
− 2t/2
sin( 2
2 t)}.
Now we will make an educated guess as to the form of a particular solution. Since we see an e−t on
the right-hand side of the ODE, and e−t is not a solution to the corresponding homogeneous equation,
we guess that some multiple of e−t will appear in our solution. Since we see a linear function 48t − 6 on
the right-hand side, we are inclined to guess that some linear combination of the functions t and 1 might
yield a solution; however, both of these are solutions to the corresponding homogeneous equation. So
30 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
we scale up by t3 (where the exponent 3 is the smallest choice that will prevent our guesses from being
solutions to the corresponding homogeneous equation) and guess a linear combination of t4 = t3 · t and
t3 = t4 · 1 instead. Thus we assume that we will be able to find a solution of the form
The above implies −5A = −10, 96B = 48, and 24C = −6, which in turn imply that A = 2, B = 21 ,
and C = − 41 . This gives us a particular solution to the ODE. And since we already know a fundamental
set of solutions to the corresponding homogeneous equation, we obtain a full general solution to the ODE:
Theorem 18.12. Let y1 , ..., yn all be solutions to the n-th order linear homogeneous ODE y (n) +
an−1 (t)y (n−1) + ... + a2 (t)y 00 + a1 (t)y 0 (t) + a0 (t)y = 0, where a0 , a1 , ..., an−1 are all continuous func-
tions of t. Then y1 , ..., yn are linearly dependent if and only if W (y1 , ..., yn ) = 0.
Theorem 18.13 (Variation of Parameters). Consider an n-th order linear ODE y (n) + an−1 (t)y (n−1) +
... + a2 (t)y 00 + a1 (t)y 0 + a0 (t)y = f (t). Let {y1 , y2 , ..., yn } be a fundamental set of solutions to the corre-
sponding homogeneous equation. Let A denote the n×n matrix defined in the definition of the Wronskian
above, i.e., let A be the matrix for which W (y1 , ..., yn ) = det A. For each integer k with 1 ≤ k ≤ n, let
Ak be the (n − 1) × (n − 1) matrix obtained by omitting the k-th column and the bottom row from A. Let
Wk (y1 , ..., yn ) = det Ak . Then there exist functions u1 (t), u2 (t), ..., un (t) for which
yp = u1 y1 + u2 y2 + ... + un yn
Wk (y1 , ..., yn ) · f
u0k = (−1)n .
W (y1 , ..., yn )
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 31
Example 18.14. Solve ty 000 +3y 00 = −t−3 , given that a fundamental set of solutions to the corresponding
homogeneous equation is {1, t, t−1 }.
(4) f (t) = t
(6) f (t) = tt
Z ∞
Lf (s) = e−st · 1dt
0
t=b
1 −st
= lim − e
b→∞ s
t=0
1 −sb 1
= lim − e +
b→∞ s s
1
= .
s
Note that the above computation is only true for s > 0; if s ≤ 0, then the integral in the definition of
the Laplace transform diverges.
R∞ R∞
(b) If f (t) = C, then Lf (s) = 0
e−st · Cdt = C 0
e−st dt = CL{1}(s) = C
s, by our computation in
part (a).
Z ∞
Lf (s) = e−st eat dt
Z0 ∞
= e−(s−a)t dt
0
t=b
1 −(s−a)t
= lim − e
b→∞ s−a t=0
1
= .
s−a
As in part (a), we note that the computation above is only valid for s > a.
Z ∞
Lf (s) = e−st tdt
0
t=b Z b
t −st 1 −st
= lim − e + e dt
b→∞ s t=0 0 s
t=b t=b
t −st 1 −st
= lim − e + − 2e
b→∞ s t=0 s t=0
1
=0+ 2
s
1
= 2.
s
(e) For f (t) = sin(kt), we need to use integration by parts twice. (Students may remember integrating
functions of the form eat sin(bt) as a challenging integration by parts exercise in Cal II.)
Z ∞
Lf (s) = e−st sin(kt)dt
0
t=a
k b −st
Z
1
= lim − e−st sin(kt) + e cos(kt)dt
b→∞ s t=0 s 0
k ∞ −st
Z
=0+ e cos(kt)dt
s 0
" t=a #
k a −st
Z
k 1 −st
= lim − e cos(kt) − e sin(kt)dt
s a→∞ s t=0 s 0
k 1 k ∞ −st
Z
= − e sin(kt)dt
s s s 0
2
k k
= 2 − 2 Lf (s).
s s
k2
k
1+ 2 Lf (s) =
s s2
and therefore
k
Lf (s) = .
s2 + k 2
k
sin(kt)
s2 + k2
k
sinh(kt) = 12 (ekt − e−kt ) ,s>k
s2 − k 2
1
t ,s>0
s2
Definition 19.7. A function f of an independent variable t has exponential order b, where b is a real
number, if there exist constants C > 0 and T > 0 for which
Theorem 19.8. If f (t) is a piecewise continuous function of exponential order b, then the Laplace
transform Lf (s) exists for s > b.
We will state the next lemma without proof, but drawing an appropriate picture lends it a lot of
credibility.
Rb Rb
Lemma 19.9. For any integrable function g(t) on any interval [a, b], | a g(t)dt| ≤ a |g(t)|dt.
Theorem 19.10. If f (t) is a piecewise continuous function of exponential order b and F = Lf , then
lim F (s) = 0.
s→∞
Proof. It suffices (by the squeeze theorem) to prove that lim |F (s)| = 0. Since f is of exponential order
s→∞
b, there are constants C and T such that |f (t)| ≤ Cebt for all t > T . Then by the previous lemma, for
all t > T we have
Z ∞
|F (s)| = e−st f (t)dt
0
Z ∞
≤ |e−st f (t)|dt
0
Z ∞
= e−st |f (t)|dt
0
Z ∞
≤ e−st Cebt dt
0
Z ∞
=C e−st ebt dt
0
= CL{ebt }(s)
1
=C· .
s−b
C
It follows then that lim |F (s)| ≤ lim = 0. This completes the proof.
s→∞ s→∞ s − b
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 35
for s > 0.
Theorem 20.3 (Shifting Property of the Laplace Transform). Let f (t) be a function and suppose
F (s) = Lf (s) exists for s > b. Then
L{eat f (t)}(s) = F (s − a)
for s > b + a.
Proof. This is just the definition of the Laplace transform again- plug it in and see.
Example 20.4. Compute Lf where
(1) f (t) = e−2t sin t.
(2) f (t) = 4te3t .
Solution. (a) By the shifting property,
Theorem 20.5 (Laplace Transform of a Derivative). Let f (t) be a differentiable function of exponential
order b and let F = Lf . Then
for s > b.
Z ∞
0
L{f (t)}(s) = e−st f 0 (t)dt
0
t=a
Z a
= lim e−st f (t) t=0 + se−st f (t)dt
a→∞ 0
Z a
−sa
= lim e f (a) − f (0) + s e−st f (t)dt
a→∞ 0
= 0 − f (0) + Lf (s)
= sF (s) − f (0).
The fact that the third line is equal to the fourth line, i.e. that e−sa f (a) approaches 0 as a → ∞, is
true when s > b because f is of exponential order b.
Solution. Note that if f (t) = sin(kt) then f 0 (t) = k cos(kt). So we apply the previous theorem (and also
use the linearity of L):
1
L{cos(kt)}(s) = L{k cos(kt)}(s)
k
1
= [sL{sin(kt)}(s) − sin(k · 0)]
k
1 k
= ·s 2
k s + k2
s
= 2 .
s + k2
Theorem 20.7 (Derivatives of the Laplace Transform). If f (t) is a piecewise continuous function of
exponential order b, and F = Lf is its Laplace transform, then for s > b,
n
−st
Proof Sketch. First note that, by an easy computation, ds d
ne = (−1)n tn e−st . Now the proof consists
of the following string of equalities which turn out to be true:
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 37
dn
F (n) (s) = L(f (t))(s)
dsn Z
n ∞
d
= n e−st f (t)dt
ds 0
Z ∞ n
d −st
= e f (t) dt
dsn
Z0 ∞
= [(−1)n tn e−st ]f (t)dt
0
Z ∞
= (−1) n
e−st tn f (t)dt
0
= (−1)n L{tn f (t)}(s).
All of the equalities between the lines above should be clear to a differential equations student, except
the equality between lines 2 and 3. The fact that the n-th derivative (with respect to s) can be pulled
inside the integral (with respect to t) without changing the value of the expression is not obvious, and
is in fact not true for all functions of two variables s and t- however, it is true for a for a special class of
functions which includes the function e−st f (t), because f (t) is of exponential order. The student may
see a proof of this fact sometime later in a good real analysis course.
(2) t2 e−3t
(3) tn
Solution. (a) Using the previous theorem,
1 d
L{t cos(2t)}(s) = (−1) L{cos(2t)}(s)
ds
d s
=−
ds s2 + 4
s + 4 − 2s2
2
=−
(s2 + 4)2
2
s −4
= 2 .
(s + 4)2
d2
L{t2 e−3t }(s) = (−1)2 L{e−3t }(s)
ds2
d2 1
=
ds2 s + 3
2
= .
(s + 3)3
38 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
dn 1 n!
(c) Note that dsn s = (−1)n sn+1 . Then:
dn
n n
L{t }(s) = (−1) L{1}(s)
dsn
n
d 1
= (−1)n
dsn s
n!
= (−1)n · (−1)n n+1
s
n!
= n+1 .
s
Example 20.9. Compute the Laplace transforms of f , f 0 , and f 00 for f (t) = (3t − 1)3 .
Fact 20.10 (More Laplace Transforms of Common Functions).
f (t) F (s) = Lf (s)
n!
tn ,s>0
sn+1
s
cos(kt)
s + k2
2
s
cosh(kt) = 21 (ekt − e−kt ) ,s>k
s2 − k 2
k
eat sin(kt)
(s − a)2 + k 2
s−a
eat cos(kt) ,s>0
(s − a)2 + k 2
k
eat sinh(kt)
(s − a)2 − k 2
s−a
eat cosh(kt) ,s>0
(s − a)2 − k 2
1
(1) F (s) = s−6
2
(2) F (s) = s2 +4
6
(3) F (s) = s4
6
(4) F (s) = (s+2)4
Theorem 21.4 (Linearity of the Inverse Laplace Transform). The inverse Laplace transform operator
L is linear. That is,
7
(2) − s2 +16
1 7
(3) s3 − s2 +16
5 2
(4) s − s−10
2s−9
Example 21.6. Find the inverse Laplace transform of F (s) = s2 +25 .
−1 −1 s −1 1
L F (t) = 2L (t) − 9L (t)
s2 + 25 s2 + 25
s 9 5
= 2L−1 (t) − L−1 (t)
s2 + 25 5 s2 + 25
9
= 2 cos(5t) − sin(5t).
5
Example 21.7. Compute L−1 { s2 +2s+5
s
}. (Hint: Rewrite the denominator in the form (s − a)2 + k 2 .)
L(f (n) )(s) = sn F (s) − sn−1 f (0) − ... − sf (n−2) (0) − f (n−1) (0).
So:
1
L{sin2 (kt)}(s) = L{k sin(2kt)}(s)
s
k
= L{sin(2kt)}(s)
s
k 2k
= · 2
s s + 4k 2
2k 2
= .
s(s2 + 4k 2 )
0 4t
Example 22.3. Solve the IVP y − 4y = e , y(0) = 0.
40 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
Solution. Let Y = Ly. Take the Laplace transform of both sides of the ODE:
By the linearity of L and by Theorem 22.1, the above translates into the following equation:
1
(sY − y(0)) − 4Y = s−4
1
(s − 4)Y = s−4
1
It follows that Ly = Y = (s−4)2 . Then
n o
y = L−1 Y = L−1 1
(s−4)2 = te4t
and the problem is finished. (Check that this answer is actually a solution to the IVP.)
(s2 − 4s)Y − 3s + 4 = 0
y = L−1 Y
−1 1 2
=L +
s s−4
−1 1 −1 1
=L + 2L
s s−4
= 1 + 2e4t .
Solution. Note that this example is interesting because it is a second-order ODE where the coefficients
are not constant. So the methods we developed earlier in the course for second-order ODE’s are not
obviously useful. But the Laplace transform turns out to be adequate to the task.
As usual, we let Y = Ly and take Laplace transforms of both sides of the ODE. The tricky thing here
is computing L{3ty 0 }. By Theorem 20.7, we have
d
L{3ty 0 } = − L{3y 0 }
ds
d
= −3 · (sY − y(0))
ds
d
= −3 · (sY )
ds
= −3(Y + sY 0 ).
Thus we have reduced the problem to a first-order linear ODE, which we can solve using an integrating
factor. Rewrite the ODE in standard form as follows:
Y 0 + (− 31 s + 3s−1 )Y = − s12 .
(−(1/3)s+3s−1 )ds 2
R
Set µ(s) = e = e−(1/6)s +3 ln s
, and multiply on both sides of the equation by µ. This
gives us
d −(1/6)s2 +3 ln s 1 2
(e Y ) = − 2 e−(1/6)s +3 ln s
ds s
1 2
= − 2 e−(1/6)s · s3
s
2
= −se−(1/6)s .
It follows that
2 2 2
e−(1/6)s +3 ln s
Y = (−se(1/6)s )ds = 3e−(1/6)s + C
R
and hence
2 2
C (1/6)s2
Y = (3e−(1/6)s + C) · e(1/6)s −3 ln s
= 3
s3 + s3 e .
42 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
Now consider the expression above; we have determined that the function Y = Ly is one of an infinite
class of functions,
which depend
on the choice of constant C. But notice that if C is not equal to zero,
3 C (1/6)s2
then lim + 3e is either ∞ or −∞ (depending on the sign of C); this is impossible if Y is
s→∞ s3 s
the Laplace transform of y, since we must have lim Y = 0. So we assume C = 0 and hence Y = s33 . It
s→∞
follows then that
This completes the problem. The reader should verify that y is actually a solution to the IVP.
(
0 if t < 0;
U(t) =
1 if t ≥ 0.
We will frequently work with horizontal shifts of the Heaviside function, i.e. functions of the form
U(t − a) where a is a real number. Note that
(
0 if t < a;
U(t − a) =
1 if t ≥ a.
We will also find it convenient to work with sums of the form U(t−a)−U(t−b), where a < b. Note that
0 if t < a;
U(t − a) − U(t − b) = 1 if a ≤ t < b;
0 if t ≥ b.
and
Z ∞
L{f (t − a)U(t − a)} = e−st f (t − a)U(t − a)dt
0
Z a Z ∞
= e−st f (t − a)U(t − a)dt + e−st f (t − a)U(t − a)dt
Z0 a Z ∞ a
−st
= e f (t − a) · 0dt + e−st f (t − a) · 1dt
0 a
Z ∞
= e−st f (t − a)dt.
a
Z ∞
L{f (t − a)U(t − a)} = e−st f (t − a)dt
Za∞
= e−s(u+a) f (u)du
0
Z ∞
= e−as e−su f (u)du
0
= e−as F (s).
The second equality in the theorem is just a re-writing of the first equality.
Example 23.4. Find
(1) L{U(t − a)}
1
Lf (s) = L{1 − U(t − π)} = L{1} − L{(U )(t − π)} = s − eπs · 1s .
s2 Y − sy(0) − y 0 (0) + 9Y = 1
s − eπs · 1s .
By applying the initial conditions y(0) = y 0 (0) = 0 and collecting like terms, the above reduces to:
or
e−πs
Y = 1
s(s2 +9) (1 − e−πs ) = 1
s(s2 +9) − s(s2 +9) .
Now to finish the problem it suffices to compute L−1 of the above, since y = L−1 Y . Let us do the
two terms on the right-hand separately. (This is ok because L−1 is linear.)
1 1 s
The method of partial fractions reveals that s(s2 +9) = 9s − 9(s2 +9) , and hence
−1 1 1 −1 1 1 −1 s
L 2
= L − L
s(s + 9) 9 s 9 s2 + 9
1 1
= − cos(3t).
9 9
e−πs
−1 −1 1
L =L (t − π) · U(t − π)
s(s2 + 9) s(s2 + 9)
1 1
= − cos(3(t − π)) U(t − π).
9 9
y = L−1 Y = 1
9 − 1
9 cos(3t) − ( 91 − 1
9 cos(3(t − π)))U(t − π).
The above solution to the IVP may be written as a piecewise function as follows:
(
1
− 91 cos(3t) if 0 < t < π;
9
y=
− 92 cos(3t) if t ≥ π.
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 45
Lemma 24.3 (Fubini’s Theorem). Let F (t, y) be an integrable function on the rectangular region R =
[a, b] × [c, d]. Then f is integrable, and
RR RdRb RbRd
R
F (t, y)d(t, y) = c a F (t, y)dtdy = a c F (t, y)dydt.
The student should recall that Fubini’s theorem is actually stronger than what is stated above, since
it is true not just for rectangular regions of integration but actually for much more general regions. We
will not benefit much at the moment from attempting to formulate a precise version of this statement,
but we will use the stronger version of Fubini’s theorem to prove the next fact.
Theorem 24.4 (Convolution Theorem). Suppose f (t) and g(t) are both piecewise continuous functions
of exponential order b. Let F = Lf and G = Lg. Then
L−1 {F (s)G(s)} = f ∗ g.
Proof. Write the product F (s)G(s) using the definition of the Laplace transform:
R∞ R∞
F (s)G(s) = 0
e−sx f (x)dx · 0
e−sy g(y)dy.
R∞R∞
F (s)G(s) = 0 0
e−s(x+y) f (x)g(y)dxdy.
R∞
Now briefly consider the inner integral 0 e−s(x+y) f (x)g(y)dx in the equation above (imagining y
held constant). Applying the substitution rule with t = x + y, we get:
R∞ R∞
0
e−s(x+y) f (x)g(y)dx = y
e−st f (t − y)g(y)dx
and therefore
R∞R∞
F (s)G(s) = 0 y
e−st f (t − y)g(y)dtdy.
Note that the region of integration in the above is exactly the set {(t, y) : 0 ≤ t ≤ y, 0 ≤ y ≤ ∞},
which corresponds to the “infinite triangle” bounded below by the line y = 0 and on the left by the line
y = t in the ty-plane. We can rewrite this region of integration as {(t, y) : 0 ≤ t ≤ ∞, 0 ≤ y ≤ t}, and
use Fubini’s theorem to exchange the order of the iterated integral:
46 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
Z ∞ Z t
F (s)G(s) = e−st f (t − y)g(y)dydt
0 0
Z ∞ Z t
−st
= e f (t − y)g(y)dy dt
Z0 ∞ 0
= e−st (f ∗ g)(t)dt
0
= L{(f ∗ g)(t)}(s).
f ∗g =g∗f
f ∗ g = L−1 (F G)
= L−1 (GF )
= g ∗ f.
The next corollary is of great theoretical significance. It gives a direct method for producing a solution
to any linear ODE with constant coefficients, provided the function f (t) on the right-hand side is Laplace
transformable. By careful analysis of characteristic polynomials, one may use Duhamel’s principle to
deduce the method of variation of parameters.
Corollary 24.7 (Duhamel’s Principle). Let a0 , a1 , ..., an be constants, and let f (t) be a function of ex-
ponential order b for some b. Consider the linear IVP:
p(s) = an sn + ... + a2 s2 + a1 s + a0 .
1
Therefore Y = p(s) · Lf , and hence by the convolution theorem,
y = L−1 Y
−1 1
=L · Lf
p(s)
1
= L−1 ∗ L(L−1 f )
p(s)
−1 1
=L ∗ f.
p(s)
Γ(x + 1) = xΓ(x).
Z ∞
Γ(x + 1) = e−u ux du
0
Z ∞
−u x ∞
= −e u u=0 + e−u (xux−1 )du
0
Z ∞
=0+x e−u ux−1 du
0
= xΓ(x).
Corollary 25.4. The function Γ(x + 1) is an extension of the factorial function to complex inputs x
with Re(x) > −1, i.e. for every nonnegative integer n,
Γ(n + 1) = n!.
Moreover, the Gamma function is continuous, in fact infinitely differentiable, in fact infinitely complex-
differentiable.
Proof. The claim in the last sentence of the corollary is a bit beyond our scope to verify at the moment.
However, we have done enough to verify that Γ extends the factorial function. First check it for n = 0
by using the definition of the Gamma function:
48 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
Z ∞
Γ(0 + 1) = e−u u0 du
Z0 ∞
= e−u du
0
= [−e−u ]∞
u=0
= 0 − (−1)
=1
= 0!
Now use the functional equation given by Theorem 25.3 to check it for the remaining integers n:
Γ(1 + 1) = 1 · Γ(1) = 1 · 1 = 1 = 1!
Γ(2 + 1) = 2 · Γ(2) = 2 · 1 = 2 = 2!
Γ(3 + 1) = 3 · Γ(3) = 3 · 2 = 6 = 3!
Γ(4 + 1) = 4 · Γ(4) = 4 · 6 = 24 = 4!
Γ(5 + 1) = 5 · Γ(5) = 5 · 24 = 120 = 5!
...
Since the Gamma function extends the factorial function, the next theorem is an extension of the rule
n!
L{tn } = n+1 which we computed earlier.
s
Theorem 25.5. Let a be a positive real number. Then
Γ(a + 1)
L{ta } = .
sa+1
Proof. The trick here is simply to look at the definition of the Laplace transform of ta , and use substi-
tution rule with u = st, which gives t = s−1 u and dt = s−1 du. We proceed:
Z ∞
L{ta }(s) = e−st ta dt
0
Z ∞
= e−u (s−1 u)a · s−1 du
0
Z ∞
= s−a · s−1 · e−u ua du
0
Γ(a + 1)
= .
sa+1
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 49
The next theorem is a very useful computational tool about the Gamma function (and is also surpris-
ing, beautiful, and strange). Its usual proof relies on a famous theorem called Weierstrass’ factorization
theorem, which is beyond our scope.
Theorem 25.6 (Euler’s Reflection Formula). For complex numbers x which are not integers,
π
Γ(1 − x)Γ(x) = .
sin(πx)
Example 25.7. Use Euler’s reflection formula to compute:
(1) Γ( 21 )
(2) Γ(− 21 )
Solution. (1) For x = 12 , we have
2
1 1 1
Γ =Γ 1− Γ
2 2 2
π
=
sin(π/2)
=π
√
and therefore since Γ( 21 ) must be positive, we have Γ( 21 ) = π.
√
π
Γ( 23 ) = 12 Γ( 12 ) = 2 .
1 3 3 3
Γ − Γ =Γ 1− Γ
2 2 2 2
π
=
sin(3π/2)
= −π
√
and therefore Γ(− 12 ) = −π/Γ( 32 ) = −π · √2
π
= −2 π.
Example 25.8. Compute:
√
(1) L{ t}
(2) L{ √1t }
Solution. Using the results of the previous example, we get:
√
Γ( 23 ) π
(1) L{t1/2 } = 3/2
= 3/2
, and
s 2s
Γ( 21 )
r
π
(2) L{t−1/2 } = 1/2 = .
s s
50 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
where each aij (1 ≤ i, j ≤ n) is a function of an independent variable t, and each variable y1 , ..., yn
represents a differentiable function of t.
Example 26.2. Rewrite the following ODE’s as systems of first-order ODE’s.
(1) y 00 − 5y 0 + 6y = 0
yn (t)
yn
derivative ~y 0 of ~y to be
y10
y20
~y 0 =
... .
yn0
~y 0 = A~y ,
where
a11 (t) a12 (t) ... a1n (t)
a21 (t) a22 (t) ... a2n (t)
A=
...
an1 (t) an2 (t) ... ann (t)
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 51
Fact 26.4. Every system of n linear homogeneous ODE’s in y1 , ..., yn (with constant coefficients)
cor-
y1
y2
responds uniquely to a linear homogeneous ODE ~y 0 = A~y of the vector-valued function ~y =
...
yn
(where A is a matrix with real entries).
y10 = −y1
y20 = 2y2
Solution. It is easy to see that the general solution to the system above is y1 = C1 e−t , y2 = C2 e2t , where
C1 , C2 are arbitrary constants.
Note that in this problem, y10 depends only on y1 and y20 depends only on y2 . This system corresponds
to the matrix equation
−1 0
~y 0 = A~y , where A =
0 2
Definition 26.6. A systen of n linear homogeneous ODE’s is called decoupled if each equation in the
system is of the form yi0 = aii (t)yi . Decoupled systems correspond to matrix equations ~y 0 = D~y with
diagonal matrices D.
Theorem 26.7. Let ~y ‘ = D~y be the matrix representation of a decoupled system of first-order lin-
ear ODE’s with constant coefficients. (In other words D is a diagonal matrix with real entries.) Let
λ1 , λ2 , ..., λn denote the entries along the diagonal of D, and assume that the entries are distinct real
numbers. For each 1 ≤ i ≤ n, let ~ei denote the i-th standard basis vector of Rn , i.e. ~ei has a 1 for the
i-th entry any all 0’s in the other entries. Then any vector-valued function of the form
Proof. For each 1 ≤ i ≤ n, the i-th equation in the system is yi0 = λi yi , which has general solution
y = C1 eλi yi . So a general solution to the vector-valued ODE is
C1 eλ1 t
C2 eλ2 t
~y = .
...
λn t
Cn e
~y 0 = A~y , ~y (t0 ) = ~y .
T (~y ) = ~y (0).
Note that if C1 and C2 are constants and ~y1 and ~y2 are in S, then
so T is a linear transformation. T is onto by the “existence” part of the Existence and Uniqueness
Theorem 27.1: if ~y0 is any vector in Rn , then there is a solution ~y in S with ~y (0) = T (~y ) = ~y0 . T is also
one-to-one by the “uniqueness” part of Theorem 27.1: if ~y0 is a vector in Rn , there is at most one ~y in
S with ~y (0) = T (~y ) = ~y0 .
We have shown T is a vector space isomorphism (in a natural way!) of the space of solutions S with
the space of initial conditions Rn . So S is n-dimensional as claimed.
Corollary 27.3. Let ~y 0 = A~y be a system of n linear first-order homogeneous ODE’s, where the entries
of A are continuous functions of an independent variable t. Let {y1 , y2 , ..., yn } be any set of linearly
independent solutions to the ODE. Then all solutions to the ODE are of the form ~y = C1 ~y1 + ... + Cn ~yn
for some constants C1 , ..., Cn .
Proof. This is because any set of n-linearly independent solutions must be a basis for the n-dimensional
vector space of all solutions.
Definition 27.4. Let ~y 0 = A~y be a system of n linear first-order homogeneous ODE’s, where the entries
of A are continuous functions of an independent variable t. A fundamental set of solutions to the
ODE is any set {y1 , ..., yn } of n linearly independent solutions.
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 53
A~v = λ~v .
Any vector ~v for which A~v = λ~v is called an eigenvector of A corresponding to the eigenvalue λ.
Fact 28.2. A number λ0 is an eigenvalue of an n × n matrix A if and only if λ0 is a root of the
characteristic polynomial det(A − λI).
Fact 28.3. A vector ~v is an eigenvector of A corresponding to the eigenvalue λ if and only if ~v is in the
null space of the matrix A − λI.
4 −6
Example 28.4. Find the eigenvalues and eigenvectors of A = .
3 −7
4−λ −6
Solution. The characteristic polynomial is det(A − λI) = det = λ2 + 3λ − 10 =
3 −7 − λ
(λ + 5)(λ − 2), so the eigenvalues of A are the roots −5 and 2.
To find an eigenvector
corresponding
to the eigenvalue −5, we are looking for a vector ~v in the null
9 −6 3 −2
space of A + 5I = . This matrix row-reduces to , which has the same null space.
3 −2 0 0
A vector ~v gets sent to 0 by the latter
matrix
ifand only if 3v1 − 2v2 = 0, if and only if v1 = 32 v2 , if and
2/3 2/3
only if ~v is in the span of . So is an eigenvector corresponding to −5 (and in fact all
1 1
other eigenvectors are a scalar multiple of it).
3
A similar computation yields for instance (or any scalar multiple of it) as an eigenvector
1
corresponding to 2.
Theorem 28.5 (Diagonalizing Matrices with Distinct Eigenvalues). Let A be an n × n matrix with
distinct real eigenvalues λ1 , ..., λn and corresponding non-zero eigenvectors ~v1 , ..., ~vn respectively. Let C
be the matrix
whose columns are the eigenvectors of A. Then C is invertible, and A = CDC −1 and D = C −1 AC,
where D is a diagonal matrix whose entries along the diagonal are λ1 , ..., λn in order.
Theorem 28.6. Let A be an n × n matrix with real entries, and suppose that A has n distinct real
eigenvalues λ1 , ..., λn . Let ~v1 , ..., ~vn be eigenvectors corresponding to λ1 , ..., λn respectively. Then a fun-
damental set of solutions to the ODE ~y 0 = A~y is
C −1 ~y 0 = DC −1 ~y .
54 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
~x 0 = D~x.
The above represents a decoupled system, and hence by Theorem 26.7, a general solution to it is
Each of the terms in the linear combination on the right-hand side above are obviously linearly inde-
pendent with respect to one another. So a fundamental set of solutions to ~x 0 = D~x is
It follows from our change of variables that for each 1 ≤ i ≤ n, the function
~y = C~x
= C e~i eλi t
= ~vi eλi t
is a set of solutions. Moreover, the functions above are linearly independent since they are just the
images of the invertible matrix C of the fundamental set of solutions to ~x 0 = D~x, already a linearly
independent set. So they comprise a fundamental set of solutions and the theorem is proved.
Example 28.7. Solve the system y10 = 4y1 − 6y2 , y20 = 3y1 − 7y2 .
0 4 −6
Solution. This corresponds to the matrix system ~y = A~y , where ~y = . In a previous exam-
3 −7
ple, we computed that the eigenvalues of A are
λ1 = −5, λ2 = 2,
2/3 3
~v1 = , ~v2 = ,
1 1
respectively. So by our previous theorem, a fundamental set of solutions to the system is exactly
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 55
(2/3)e−5t
2t
(2/3) 3 3e
e−5t , e2t = , .
1 1 e−5t e2t
Thus all possible solutions to the ODE are linear combinations of the above, i.e. a general solution is
2/3 −5t 3
~y = C1 e + C2 e2t .
1 1
Since A has real entries and ~v1 , ~v2 are real vectors, it follows that
In that case,
and
Since the ODE is linear, both the real and imaginary parts of the above have to be solutions to the
ODE as well as ~y1 . (We will not repeat the usual argument here- see Theorem 14.5 to remember how it
56 DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES
goes.) This proves the theorem. (One may check that the same pair of solutions falls out when repeating
the argument with a − bi and ~v1 − ~v2 i instead of a + bi and ~v1 + ~v2 i, as we have done.)
0 3 −2
Example 29.3. Find a general solution to ~y = A~y , where A = .
4 −1
Solution. First we compute eigenvalues and eigenvectors of A. The characteristic polynomial of A is
The quadratic formula yields λ = 1 ± 2i as roots to the above, so 1 + 2i and 1 − 2i are the complex
conjugate eigenvalues of A.
To apply our previous theorem and solve the ODE, we actually only need to find an eigenvector
corresponding to 1 + 2i. So we consider the matrix
2 − 2i −2
A − (1 + 2i)I = .
4 −2 − 2i
8 −4 + 4i 2 −1 + i
The above row-reduces to , and then to . The null-space of this matrix
4 −2 − 2i 0 0
v1
is all vectors ~v = for which 2v1 + (−1 − i)v2 = 0, i.e. for which v1 = ( 21 + 12 i)v2 . Thus an
v2
eigenvector corresponding to 1 + 2i is
1/2 + (1/2)i 1/2 1/2
= + i.
1 1 0
Lastly, we apply our previous theorem to determine that a pair of linearly independent solutions to
~y 0 = A~y is given by
t 1/2 1/2 t 1/2 1/2
e cos 2t − sin 2t , e cos 2t + sin 2t .
1 0 0 1
Since A is 2 × 2, the above is actually a complete basis for the solution space. So a general solution
has the form
t 1/2 1/2 t 1/2 1/2
~y = C1 e cos 2t − sin 2t + C2 e cos 2t + sin 2t .
1 0 0 1
1 0 0
Example 29.4. Solve ~y 0 = 0 1 −1 .
0 1 1
1 0 0
Solution. The characteristic polynomial of A = 0 1 −1 is det(A−λI) = (1−λ)[(1−λ)2 −1(−1)] =
0 1 1
−(λ3 − 3λ2 + 4λ − 2) = −(λ − 1)(λ − 2λ + 2). This polynomial has a real root λ1 = 1, and two complex
conjugate roots λ2 = 1 + i, λ3 = 1 − i. These are the eigenvalues of A.
DIFFERENTIAL EQUATIONS MATH 266 SPRING 2014 (COHEN) LECTURE NOTES 57
0 0 0
For the real eigenvalue λ1 = i, compute an eigenvector: we have A − I = 0 0 −1 , which row-
0 1 0
0 1 0 v1
reduces to 0 0 1 . This matrix will send a vector ~v = v2 to ~0 if and only if v2 = v3 = 0. So
0 0 0 v3
1 1
a good choice of non-zero eigenvector is, for instance, ~v1 = 0 . This yields a solution ~y1 = 0 et
0 0
to the ODE.
For the complex
conjugate roots 1 ± i, we compute aneigenvector corresponding to 1+ i. We have
−i 0 0 1 0 0 v1
A − (1 + i)I = 0 −i −1 , which row-reduces to 0 1 −i . For ~v = v2 to be in the
0 1 −i 0 0 0 v3
null
space
of the latter
matrix,
we need v 1 = 0 and v 2 − iv 3 = 0, i.e. v 2 = iv 3 . Choosing v3 = 1, we get
0 0 0
i = 0 + 1 i as such a vector.
1 1 0
0 0 0 0
This yields the two functions ~y2 = et 0 cos t − 1 sin t and ~y3 = et 1 cos t + 0 sin t
1 0 0 1
as solutions to the ODE. The colletion {~y1 , ~y2 , ~y3 } are linearly independent, so we have found a funda-
mental set of solutions. A general solution looks like:
1 0 0 0 0
~y = C1 0 et + C2 et 0 cos t − 1 sin t + C3 et 1 cos t + 0 sin t.
0 1 0 0 1
Technique 29.5 (General Strategy for Solving Linear Homogeneous Systems with Constant Coefficients). (1)
Given a system ~y 0 = A~y , find the eigenvalues λ1 , ..., λn of A, which are precisely the roots of
the characteristic polynomial det(A − λI). Assume all of the eigenvalues are distinct.
(2) For each real eigenvector λi , find a corresponding non-zero eigenvector ~vi . This may be done by
row reducing the matrix A−λi I and picking a vector ~vi from its null space. Then a corresponding
solution to the ODE is ~yi = ~vi eλi t .
(3) For each pair of complex conjugate eigenvectors λi = a + bi, λi+1 = a − bi, find a non-zero
eigenvector ~vi corresponding to a + bi. Again, this may be done by row reducing A − λi I and
picking a vector ~vi from its null space. This ~vi will be complex, so write ~vi = ~v1 + ~v2 i. Then two
solutions to the ODE are ~yi = eat (~v1 cos bt − ~v2 sin bt) and ~yi+1 = eat (~v2 cos bt + ~v1 sin bt).
(4) The solutions {~y1 , ..., ~yn } obtained in parts (2) and (3) above comprise a linearly independent
set, and thus are a fundamental set of solutions to the system. Therefore a general solution is
~y = C1 ~y1 + ... + Cn ~yn .
Remark 29.6. Although the technique above goes a long way, we have not completely solved the
problem of finding solutions to ODE’s of the form ~y 0 = A~y where A is a real matrix, for the following
reason: it is possible for A to have repeated eigenvalues λi , i.e. roots of the characteristic polynomial
det(A − λI) with multiplicity k > 1. When this happens, the situation is a bit complicated because it
breaks into two cases. In one case, the repeated eigenvalue λi actually has k many linearly independent
eigenvectors (matching its multiplicity of k). This is the easy case and we proceed to give solutions as
in the technique above. In the other case, however, it is possible for an eigenvalue λi with multiplicity k
to have strictly fewer than k linearly independent eigenvectors. In this case, a bit more work remains to
be done to find k many solutions to the ODE. Unfortunately we have run out of time for the semester!
Feel free to ask me about this if you are curious, and the details will probably be added to a future
incarnation of the notes.