Multivariatecalculus
Multivariatecalculus
1
The linear transformation λ is denoted Df (x) and called the derivative
(or differential or total derivative) of f at x. The matrix of Df (x) : Rn → Rm
is a m × n matrix and is called the Jacobian matrix of f at x. If f : Rn → R,
then the acobian matrix is a row vector.
2
Proposition 3 If f : Rn → Rm is differentiable at a, then Dj fi exists for
1 ≤ i ≤ m, 1 ≤ j ≤ n and Df (x) is the m × n matrix Dj fi (x).
3
Example 2 Let
xy 2
(
(x2 +y 4 )
if x 6= 0,
f (x, y) =
0 if x = 0.
∇f (x) = Df (x)> .
4
2 Taylor’s Formula
The basic objective in differentiating a function is to obtain a linear approx-
imation of the function under analysis. This allows for an easier analysis
of the function locally around the point at which it is differentiated. For
some questions, the linear approximation contains all the information that is
really needed. For example, in the study of dynamical systems (which form
an essential part of modern macroeconomics), it is often enough to linearize
the system around a steady state of the system in order to analyze the local
behavior of the system around that steady state. For other questions, the
information contained in the linear approximation is insufficient. The pri-
mary instance of this is optimization theory. In this area, the basic task is
to locate the points in the domain of a function where that function achieves
its maximal (or minimal) value. Linear approximations are used in locating
these points, but in order to tell apart the points where the maximum is
obtained from the points where the function is minimized, more information
is needed.
In this section, a function is approximated around a fixed point x by a
polynomial. This approximation will obviously be at least as accurate as the
linear approximation obtained in the theory of differentiation (since linear
functions are also polynomials). The information contained in the polynomial
approximation of a function is sufficient for the purposes of finding minima
nad maxima.
We start by recalling some facts from eelementary calculus.
5
and each continuous at the endpoints a and b. Assume also that there is no
interior point x at which both f ’(x) and g’(x) are infinite. Then for some
interior point c we have
These two results can be used to prove the famous approximation result
for continuously differentiable real valued functions on the real line.
for certain sums that arise in Taylor’s formula. These play the role of higher
order directional derivatives, and they are defined as follows. If x is a point in
Rn where all second-order partial derivatives of f exist, and if t = (t1 , ..., tn )
is an arbitrary point in Rn , we write
n X
X n
00
f (x; t) = Dij f (x) tj ti .
i=1 j=1
We also define n X
n X
n
X
000
f (x; t) = Dijk f (x) tk tj ti
i=1 j=1 k=1
6
if all third-order partial derivatives exist at x. The symbol f (m) (x; t) is
similarly defined if all m − th order partials exist.
y = f (x) ,
where the exogenous variable x appeared on the right hand side and the
endogenous variable appeared on the left hand side. When endogenous and
exogenous variables cannot be separated as in
f (x, y) = c, (4)
y = g (x) ,
for some function g. We say that g is defined implicitly by (4) . The problem
assumes a more general form when we have a system of several equations
7
involving several variables in terms of the remaining variables. An important
special case is the familiar problem in algebra of solving n linear equations
of the form n
X
aij yj = ti , (i = 1, 2, ...., n) , (5)
j=1
Using the results from linear algebra, we may try to solve explicitly for the
y 0 s in terms of the x0 s. For this purpose, write
A = (Ay , Ax )
where
a11 · · · a1n a1(n+1) · · · a1(n+k)
. ... ..
.. .. ..
Ay = . . and Ax = . .
. . .
an1 · · · ann an(n+1) · · · an(n+k)
Ay y+Ax x = 0,
y= −A−1
y Ax x.
8
Clearly, this can only be done if Ay is invertible, i.e. if det (Ay ) 6= 0.
In this section, our goal is to extend this analysis beyond linear func-
tions. Unfortunately, it is in general impossible to find explicit solutions for
n non-linear equations in n unknowns (the y above). Nevertheless, we can
sometimes achieve a more modest goal by looking at the set of equations
locally around a solution to the system. Recall that the basic method of
local analysis was to linearize the function in question around a given point
and analyze the linear approximation. The derivative of the function gives
us the desired approximation, and we can use a similar line of argument to
get local results as we did in thelinear example above.
More specifically, for a given implicit function f (y, x) we want to know
the answers to the following two questions:
9
−1
(d) Dx g (x0 ) = − (Dy f (y 0 , x0 )) Dx f (y 0 , x0 ) , where
∂f1 ∂f1 ∂f1 ∂f1
∂y1
· · · ∂yn ∂x1
··· ∂xk
. .
. . . . .. and Dx f y 0 , x0 = .. . . .
. ..
Dy f y 0 , x 0 =
. .
.
∂fn ∂fn ∂fn ∂fn
∂y1
· · · ∂yn ∂x1
··· ∂xk
Another important result from multivariate calculus that uses the local
linear approximation of the original function by its derivative is the following.
Proposition 10
Theorem 11 (Inverse Function Theorem). Assume f = (f1 , ..., fn ) ∈
C 1 on an open set S in Rn , and let T = f (S). If the Jacobian determinant
Jf (a) 6= 0 for some point a ∈ S, then there are two open sets X ⊆ S and
Y ⊆ T and a uniquely determined function g such that:
(a) a ∈ X, f (y) ∈ Y,
(b) Y = f (X) ,
(c) f is one-to-one on X,
(d) g is defined on Y , g (Y ) = X, and g [f (x)] = x, for every x∈X,
(e) g ∈ C 1 on Y.
10
x∗ satisfying f (x∗ ) = c. Implicit Function Theorem shows that whenever
∂f (x∗ )
∂x1
6= 0, there is a function g : Rn−1 → R and an ε > 0 such that
i) x∗1 = g (x∗2 , ..., x∗n ) ,
ii) For all (x2 , ..., xn ) ∈ Bε ((x∗2 , ..., x∗n )) , f (g (x2 , ..., xn ) , (x2 , ..., xn )) = c,
∂f (x∗ )
∂g ((x∗2 ,...,x∗n ))
iii) ∂xi
= − ∂f∂x(xi∗ ) .
∂x1
Observe that the last conclusion follows from ii) by the chain rule. The
interpretation of the result is the following. Suppose that one of the variables,
xi is increased by a small amount dxi . The change in x1 needed in order to
keep the value of the function f constant at c is then given by
∂f (x∗ )
− ∂f∂x i
(x∗ )
.
∂x1
i and j.
11
the consumers’ optimal oil purchasing decisions, we can derive the demand
function for oil:
y1 = M (y3 , x1 ) .
From the producers’ optimal production decisions, we can derive the supply
function for oil:
y2 = S (y3 , x2 ) .
y1 = y2 .
Using the notation from the implicit function theorem, we can write
f1 = M (y3 , x1 ) − y1 ,
f2 = S (y3 , x2 ) − y2 ,
f3 = y1 − y2 .
f y 0 , x0 = f1 y 0 , x0 , f2 y 0 , x0 , f3 y 0 , x0 = 0
12
thus nonzero. As a result, the implicit function theorem is applicable. Hence
we may write around the initial point (y 0 , x0 ) y = g (x0 ) , where g : R2 → R3 .
Since
∂M (y0 ,x0 )
∂x1
0
∂S (y0 ,x0 )
Dx f y 0 , x 0 = 0
,
∂x2
0 0
1
Dx g x0 = − ∂M (y0 ,x0 ) ∂S(y0 ,x0 ) ×
∂y3
− ∂y3
( ) ∂M (y0 ,x0 ) ∂M (y0 ,x0 ) ∂M (y0 ,x0 )
∂S y0 ,x0
− 0
∂S ∂y3 ∂y3 ∂y3 ∂x1
(y0 ,x0 ) ∂M (y0 ,x0 ) ∂S (y0 ,x0 ) ∂S (y0 ,x0 )
− ∂y3 0
∂y3 ∂y3 ∂x2
1 −1 1 0 0
4 Exercises
1. A function f : R2 → R2 is given. Verify that the inverse function
theorem is applicable and find the inverse function f −1 .
(i) y1 = x1 , y2 = x21 + x2 ,
(ii) y1 = 2x1 − 3x2 , y2 = x1 + 2x2 .
2. In each of the next two problems a function F and a point P are given.
Verify that the implicit function theorem is applicable. Denoting the
implicitly defined function by g, find the values of all the first partial
derivatives of g at P .
13
(i) F = (F 1 , F 2 ) , P = (0, 0, 0, 0) , where F 1 = 2x1 −3x2 +y1 −y2 , F 2 =
x1 + 2x2 + y1 − 2y2 ,
(ii) F = (F 1 , F 2 ) , P = (0, 0, 0, 0) , where F 1 = 2x1 −x2 +2y1 −y2 , F 2 =
3x1 + 2x2 + y1 + 2y2 .
f1 (y1 , y2 ; α, β, q) = y1 + αy22 − q = 0
f2 (y1 , y2 ; α, β, q) = βy12 + y2 − q = 0
14
(i.e. the limit on the right hand side must be well defined). Obviously
all polynomials are analytic functions.
Define the exponential function as follows:
∞
X xn
ex = .
n=0
n!
Show that the series on the right converges for all x and hence the
exponential function is also analytic.
It is a well known theorem for analytic functions that the derivatives
of f can be obtained by differentiating the series of functions term by
term, i.e.
∞
X
0
f (x) = nan xn−1 . (7)
n=0
Use this fact to evaluate the derivative of the exponential function and
show that the exponential function has a well defined inverse called
the logarithmic function. Evaluate the derivative of the logarithmic
function.
Deduce from 6 and 7 Taylor’s theorem for the case where f is analytic:
∞
X f (n) (a)
f (x) = (x − a)n . (8)
n=0
n!
Notice that all analytic functions are in class C ∞ since their derivatives
exist for all orders. Finally, show that there are C ∞ functions that are
not analytic. To achieve this, evaluate the derivatives of the following
function ( 1
e− x2 if x 6= 0,
f (x, y) =
0 if x = 0.
Conclude that 8 fails for a = 0.
15
6. Let
xy 2
(
(x2 +y 4 )
if x 6= 0,
f (x, y) =
0 if x = 0.
Prove that f has well defined finite directional derivatives in all direc-
tions in R2 and prove also that f is not continuous at 0.
16