Curves in R: Graphs Vs Level Sets
Curves in R: Graphs Vs Level Sets
{(x, y) ∈ R2 | y = f (x)}.
Example: When we say “the curve y = x2 ,” we really mean: “The graph of the function
f (x) = x2 .” That is, we mean the set {(x, y) ∈ R2 | y = x2 }.
Note: Every graph is a level set (why?). But not every level set is a graph.
Graphs must pass the vertical line test. (Level sets may or may not.)
Again: Every graph is a level set (why?). But not every level set is a graph.
Graphs must pass the vertical line test. (Level sets may or may not.)
f (x, y) = x2 + y 2 .
The level sets of f are curves in R2 . Level sets are {(x, y) ∈ R2 x2 + y 2 = c}.
The graph of f is a surface in R3 . Graph is {(x, y, z) ∈ R3 z = x2 + y 2 }.
f (x, y) = −x2 − y 2 .
The level sets of f are curves in R . Level sets are {(x, y) ∈ R −x2 −y 2 = c}.
2 2
f (x, y) = x2 − y 2 .
The level sets of f are curves in R2 . Level sets are {(x, y) ∈ R2 x2 − y 2 = c}.
The graph of f is a surface in R3 . Graph is {(x, y, z) ∈ R3 z = x2 − y 2 }.
Note: The columns are the partial derivatives with respect to x1 , then to
x2 , etc. The rows are the gradients of the component functions F 1 , F 2 , etc.
∇F1
∂F
DF (x1 , . . . , xn ) = ∂x ∂F
· · · ∂x = .. .
1 n
.
∇Fm
F (x, y) = (x + 2y, sin(x), ey ) = (F1 (x, y), F2 (x, y), F3 (x, y)).
Goals: We will:
◦ Interpret the derivative of F as the “best linear approximation to F .”
◦ State a Chain Rule for multivariable derivatives.
Introduction: Gradient of Functions Rn → R
Def: Let f : Rn → R be a function.
Recall: The derivative of f : Rn → R at the point x = (x1 , . . . , xn ) is the
1 × n matrix h i
∂f ∂f
Df (x) = ∂x1 · · · ∂xn .
The gradient of f : Rn → R at the point x = (x1 , . . . , xn ) is the vector
h i
∂f ∂f
∇f (x) = ∂x1 , . . . , ∂xn .
Dv f (x) = ∇f (x) · v.
vT Hf (x) v.
∆y ≈ dy.
Therefore:
f (x) ≈ f (a) + f 0 (a)(x − a).
Observations:
(1) For this to make sense, we must have: co-domain(f ) = domain(g).
(2) Composition is not generally commutative: that is, f ◦ g and g ◦ f are
usually different.
(3) Composition is always associative: (h ◦ g) ◦ f = h ◦ (g ◦ f ).
(m × n)(n × k) → (m × k).
Notice that n appears twice here to “cancel out.” That is, we need the number
of rows of A to equal the number of columns of B – otherwise, the product
AB makes no sense.
Many texts describe the chain rule in the following more classical form.
While there is a “classical” form in the general case of functions g : Rm → Rp ,
we will keep things simple and only state the case of functions g : Rm → R
with codomain R.
Example 1: Let z = g(u, v), where u = h(x, y) and v = k(x, y). Then the
chain rule reads:
∂z ∂z ∂u ∂z ∂v
= +
∂x ∂u ∂x ∂v ∂x
and
∂z ∂z ∂u ∂z ∂v
= + .
∂y ∂u ∂y ∂v ∂y
(1) Uniqueness: Are inverses unique? That is, must a function f have at
most one inverse f −1 , or is it possible for f to have several different inverses?
Answer: Yes.
(2) Existence: Do inverses always exist? That is, does every function f
have an inverse function f −1 ?
Answer: No. Some functions have inverses, but others don’t.
Question: If inverse functions “undo” our original functions, can they help
us solve equations? Yes! That’s the entire point:
In other words, non-square matrices are never invertible. But square ma-
trices may or may not be invertible. Which ones are invertible? Well:
Normal vectors help us find tangent planes to level sets (see the handout
“Tangent Lines/Planes...”) But there’s another reason we like normal vectors.
Namely, ∂f
∂x is the directional derivative of f in the e1 -direction, while
∂f
∂y
is the directional derivative in the e2 -direction.
Determinants: Algebra
Prop 17.3: An n × n matrix A is invertible ⇐⇒ det(A) 6= 0.
Moreover: if A is invertible, then
1
det(A−1 ) = .
det(A)
Properties of Determinants (17.2, 17.4):
(1) (Multiplicativity) det(AB) = det(A) det(B).
(2) (Alternation) Exchanging two rows of a matrix reverses the sign of the
determinant.
(3) (Multilinearity): First:
a1 a2 · · · an b1 b2 · · · bn a1 + b1 a2 + b2 · · · an + bn
c21 c22 · · · c2n c21 c22 · · · c2n c21 c22 ··· c2n
det . + det = det
.. .. .. .. .. . . .. .. .. .. ..
..
. . . . . . . . . . .
cn1 cn2 · · · cnn cn1 cn2 · · · cnn cn1 cn2 ··· cnn
and similarly for the other rows; Second:
ka11 ka12 · · · ka1n a11 a12 · · · a1n
a21 a22 · · · a2n a21 a22 · · · a2n
det . = k det
. .
. . . .
. .. .. . . ..
. . . . . . . .
an1 an2 · · · ann an1 an2 · · · ann
and similarly for the other rows. Here, k ∈ R is any scalar.
Determinants: Geometry
Prop 17.5: Let A be any 2 × 2 matrix. Then the area of the parallelogram
generated by the columns of A is |det(A)|.
v = c1 v1 + · · · + ck vk .
Note: Levandosky (L21: p 145-149) explains all this very clearly, in much
more depth than this review sheet provides. The examples are also quite
good: make sure you understand all of them.
C = v1 v2 · · · vk .
v = c1 v1 + · · · + ck vk .
In other words:
v = C[v]B .
This formula tells us how to go between the standard coordinates for v and
the B-coordinates of v.