Professional Documents
Culture Documents
An angle is a figure formed by two rays (half-lines) with common vertex (end-point).
Two lines are perpendicular if they intersect and form right angles. A line L and a plane
M are perpendicular if they intersect at a single point P and L is perpendicular to every
line lying in M and going through P.
Projections
In E
2
, for a given line L and any point P, there exists a unique line L through P
perpendicular to L. The intersection point of L and L is called the projection of point P
on L. For any set S of points and any line L, the set of all projections on L of points in S
is called the projection of S on L.
In E
3
, for a given plane M and any point P, there exists a unique line going through P and
perpendicular to M. The intersection point of this line and M is called the projection of the
point P on the plane M. For any set S of points, the set of all projections on M of points in
S is called the projection of S on M.
For a given line L and any point P, there is a unique plane going through P and
perpendicular to L. The intersection point of L with this plane is called the projection of
the point P on the line L. For any set S of points, the set of all projections on L of points
in S is called the projection of S on L.
For more definitions and elementary facts in E
2
and E
3
, read the second part of Chapter 1
of the textbook.
|
|
|
Lecture II
Vectors and Vector Algebra
A set S of rays is called a direction if it satises the following laws:
(1) Any two rays in S have the same direction.
(2) Every ray that has the same direction as some member of S is in S.
A vector A consists of a non-negative real number, called the magnitude of
the
| A by A .
A vector
c,weobtainavectordenotedbycA. Themagnitudeoftheresultisgiven
A
A if by cA =|c||
B suchthatitsstartpointisthe
end-pointof
A+
A. Then
Bwillbethevectorhavingthesamestart-point
as A and the same end-point as
B.
3 Scalar product (Dot product)
The scalar product of two vectors A and
B is a scalar quantity denoted
B. Its value is
by
B=
|
B if we translate
and
A
.
A
.
B
A
A+
B
Figure 1: Vector addition
The magnitude of A is equal to the square root of the dot product of A
with itself:
|A|=
A A
Hence
|A|
4 Vector product (Cross product)
The cross product of two vectors
B is a vector denoted by
A and
AB.
The magnitude of the cross product is given by:
B = A B sin. |A
| |
|
Let A and B have the same start-point P and end-points Q
1
and Q
2
,
respectively. Let M be the plane of
B. The direction of
A and
AB is
normal to M in the manner established by the right-hand rule: if a right
handisplacedatP andthengersarecurlingfromPQ
1
toTQ
2
through
AB. theanglesmallerthan,thenthethumbindicatesthedirectionof
The following are some properties of these basic vector operations.
C =aC+b
1. (a+b)
C
C =a(b
2. (ab)
C)
B+
3. a(
C)=aB+aC
2
B
| | | | | | | |
|
A
.
A
.
B =B+
4. A+
A
B+
5. (
C)= B+ C
B
6. AB =
A
3
A
.
A
.
A
.
Lecture III
Vector Algebra in Cartesian Coordinates
Let us construct a Cartesian coordinates system in E
3
. First we choose a
point O, called the origin. Then we chose three mutually perpendicular rays
starting from O. These rays are called the positive x axis, positive y axis, and
positive z axis. Consider the lines containing these rays. For any of these lines,
everypointonitcanbeidentiedwitharealnumber: ifthepointisontheray,
the real number is the distance to O, if its not on the ray, the number is the
distance to O times 1. Let us denote these three lines by X,Y, and Z. Let P
beapointinspace. Considertheprojection-pointsofP onX,Y,andZ. These
pointsgivetheCartesiancoordinatesofP,denotedx
P
,y
P
,andz
P
. Anytriplet
ofrealnumbersformsthecoordinatesforsome pointP. Dierentpointshave
dierent coordinates.
The three unit vectors in the directions of the positive x,y, and z axes are
customarilydenotedby
j,and
AbeavectorinE
3
andletP bethepoint i,
k. Let
suchthatOP =A. Let(a
1
,a
2
,a
3
)bethecoordinatesofP. Considerthevectors
A
1
=a
1
i, A
2
=a
2
j, and A
3
=a
3
k. by vector addition and multiplication with
scalars, one obtains the following expression:
A
2
+
A=A
1
+
A
3
=a
1
i+a
2
j+a
3
k
Then a
1
,a
2
, and a
3
are called the scalar components of A, and a
1
i,a
2
j, and
a
3
k are called the vector components of
i,
k unit vectors, we
A. By using the
j,
i+(ca
2
)
k,
j+(ca
3
)
i+b
2
j+b
3
k, then A=a
1
i+a
2
j+a
3
k and
A+
j+(a
3
+b
3
)
B =(a
1
+b
1
)
i+(a
2
+b
2
)
k.
3. Dot product
B = b
1
i+b
2
j+b
3
k, considering that
i = If A = a
1
i+a
2
j+a
3
k and
j
i=
i k=
i=
k j =0,wegetthat
k k=1and
i j =
k
jj =
jk=
B =a
1
b
1
+a
2
b
2
+a
3
b
3
.
4. Cross product
Let
B =b
1
i+b
2
j+b
3
k. To compute the cross A=a
1
i+a
2
j+a
3
k and
product, we use the following equalities:
i=
k
i jj =
k=0, and
ij =
i,
i=
ik=
k,
k=kj =
j. j
i=
j
We get the following formula:
AB =(a
2
b
3
a
3
b
2
)
i+(a
3
b
1
a
1
b
3
)
k.
j+(a
1
b
2
a
2
b
1
)
i j k
AB = a
1
a
2
a
3
b
1
b
2
b
3
i=
k k=1, we get that
jj =
A=( i)
j+( k)
i+( j)
k,
which is know as the frame identity.
Triple products
1. Scalar triple product
2
A
.
A
.
A
.
A
.
n
C = c
1
i+c
2
j+c
3
k. Let A = a
1
i+a
2
j+a
3
k, B = b
1
i+b
2
j+b
3
k,and
The triple product (
A B) is called a scalar triple product, since it is
a scalar quantity. Its value is given by:
a
1
a
2
a
3
B)
.
C = (
C)= b
1
b
2
b
3
.
c
1
c
2
c
3
Hencewecansimplywrite(
C]withoutspecifyingthe A B) as[A, ,
positions for the cross and dot signs. We will use the following equalities
in computing quadruple products:
B,
C,
A,
C,
B,
A,
[A,
C] = [B,
A] = [C,
B] =[A,
B] =[C,
A] =[B,
C].
2. Vector triple product
The cross product is not associative so we will give two formulas:
AB)C =(
A)
C
.
A, and (
C
.
B(
B)
A(
B( B)
BC)=( C)
C
Quadruple products
1. Scalar quadruple product
Theexpression(
B)
.
(
D)iscalledaquadruplescalarproduct,and A
C
by applying the formulas for triple products, we get the value:
(
B)
.
(
C
.
B
.
C
.
A
C D)=(
A)(
D)(
B)( D)
2. Vector quadruple product
The expression (
B)(
D,
D,
(
C D)= [C,
A]B[C,
B]A
3
A
.
C
.
C
.
A
.
A
.
A
.
B
Lecture I
Analytic Geometry in E
2
and E
3
First we review some basic facts of analytic geometry in E
2
. Let us consider a
Cartesian coordinate system in E
2
. We denote by F[x, y] an algebraic formula
in the variables x and y. Any equation of the form F
1
[x, y] = F
2
[x, y] can
be reduced to an equation of the form F[x, y] = 0. We say that an equation
F[x, y] = 0 determines a set S if S is the set of all points in E
2
whose coordinates
satisfy the equation.
In E
2
, an equation is called linear if has the form Ax +By +C = 0. where
A, B, and C are real coecients, with A
2
+ B
2
> 0. Also, the following facts
are true:
(a) Every linear equation determines a unique straight line.
(b) Every straight line is determined by some linear equation.
Let
N = A
i +B
j. Then
N is normal to the line L determined by Ax+By +
C = 0. The vector
N =
N
|
N|
is a unit vector normal to the same line. Let O
be the origin of the Cartesian system, let P be a point position vector
R
P
.
Denote by d the distance from O to the line L. The following lemma holds:
Lemma 1 If P is on the line L then |
R
P
N| = d.
This result follows quickly from the equality
R
P
N = |
R
P
| cos , where is the
angle between
R
P
and
N.
We can use lemma 1 to nd the distance from O to the line L. Let R = (x, y)
be a point on L and let
R = x
i +y
R
N| = |
N|d =
A
2
+B
2
d. Hence we get the following result:
Theorem 1 The distance d between the origin O and the line determined by
Ax +By +C = 0 is given by d =
|C|
A
2
+B
2
.
1
V
with
In E
2
, an equation F[x, y] = is called a second-degree equation if F[x, y] =
Ax
2
+ By
2
+ Cxy + Dx + Ey + F, where A, B, C, D, E, and F are real and
A
2
+B
2
+C
2
> 0.
As we go from E
2
to E
3
, here is how the basic facts change:
An equation is called linear if it has the form Ax+By +Cz +D = 0, where
A, B, C, and D are real coecients, with A
2
+ B
2
+ C
2
> 0. In E
3
, a linear
equation determine a unique plane, not a line.
The vectors
N = A
i + B
j + C
k and
N =
N
|
N|
are normal to the plane M
determined by Ax + y+Cz +D = 0. Let d be the distance between the origin
O and the plane M. Consider a point P with position vector
R
P
. The following
properties hold:
Lemma 2 If P is in the plane M then |
R
P
N| = d.
Theorem 2 The distance d between the origin O and the plane M is given by:
d =
|D|
A
2
+B
2
+C
2
.
Let M and M
x +B
y +C
z +D
= 0. Then:
(a) If M and M
,
B = cB
, and C = cC
.
(b) If M and M
x +B
y +C
z +D
= 0.
In E
3
, a system of two linear equations for two non-parallel planes determines
a unique line. But more frequently parametric equations are used to determine
lines in E
3
.
The system of equations
x = a
1
+tb
1
, y = a
2
+tb
2
, z = a
3
+tb
3
with t in (, +) is called a system of linear scalar parametric equations
on (, +) in E
3
.
2
s
B
with
We say that a gure S is determined by a system of parametric equations
on (, +) if S is the set of all points in E
3
whose coordinates satisfy the
equations as the parameter t takes all real values. Then system of linear
scalar parametric equations in E
3
determines a unique line.
The vector interperpretation of this system of equations gives a clearer rea-
soning the property above. Let
R
0
be the position vector of the point P
0
coordinates (a
1
, a
2
, a
3
), and let
A = b
1
i +b
2
j +b
3
i + y
j + z
k such that
R =
R
0
+ t
A real t is a
line in E
3
.
3
each
for with
with for
Lecture V
Calculus of One-Variable Functions
Let us rst review some denitions in calculus on real numbers.
In order to dene limit on the real numbers we will use the concept of funnel
functions.
Denition 1 A function (t) on [0, d] is called a funnel function if it has the
following properties:
1. It is strictly increasing.
2. For every a > 0, there exists t, 0 < t d, such that 0 < (t) < a.
Denition 2 Let f(x) be a scalar function and let l be a real number. We say
that the limit of f at c is l and we denote this by
lim f(x) = l
x c
if for some d > 0 there exists a funnel function (t) on [0, d] such that for every
x with 0 < x c d it follows that f(x) l ( x c ). | | | | | | |
f (c +h)f (c )
For a real function f and c, h real numbers, the fraction
h
is called
dierence quotient and is denoted by
f
.
x
f (c +h)f (c )
Denition 3 For a real function f, if lim
h 0
h
exists and has a real
df
value, we call this value the derivative of f at c and denote it by f
(c) or
dt
|
c
.
We say that f is dierentiable at c.
Denition 4 A real function f is continuous at c if lim
x c
f(x) = f(c).
1. lim
x c
(f + g)(x) = l
1
+ l
2
.
2. lim
x c
(fg)(x) = l
1
l
2
.
(c) = f
(c) + g
(c).
2. (fg)
(c) = f(c)g
(c) + g(c)f
(c).
Denition 5 Let f(t) and F(t) be two real function. The function F(t) is
called the indenite integral of f(t) if F(t) is dierentiable and F
b
called the denite integral of f on [a, b] and it is denoted by
a
f(t)dt.
We will now review how the denitions above extend to one-variable vector
functions.
Let
1
+ L
2
. 1. lim
t c
(
A(t)) =
lL
1
.
2
3. lim
t c
(
A(t) B(t)) = L
1
L
2
.
4. lim
t c
(
A(t) B(t)) = L
1
L
2
.
A vector function
A(t) is a continuous
vector function.
Denition 8 Let
A
A(t). If
t
at c. This is denoted by
A
A(t) is dierentiable at c. D =
d
dt
|
c
. We say that
A function
i + a
2
(t)
+ a
3
(t)
A(t) is dierentiable
on [a, b]. In this case,
d
A da
1
t
da
2
t
da
3
t
= i + j + k.
dt dt dt dt
Indenite and denite integrals for one-variable vector functions are dened
in the exact same manner as for scalar functions.
3
Lecture VI
Calculus of Vector Functions
R d
2
R
Recall that
d
R(t) = a
1
(t)
i + a
2
(t)
+ a
3
(t)
functions:
d
= R(t) and
d
2
= R(t). Let
j k.
dt dt
2
Then the following dierentiation rules stand:
j + a
3
(t)
1. R(t) = a
1
(t)
i + a
2
(t)
k.
a
1
(t)
i + j a
3
(t)
2. R(t) = a
2
(t)
+ k.
A(t) and
Let
3.
d
A(t) + B(t)) = A + B.
4.
d
A(t)) = a (t)
A(t) + a(t)A(t).
dt
(a(t)
dt
(
A B. 5.
d
A(t) B(t)) = A B +
6.
d
A(t)
B +
B(t)) = A
A B.
dt
(
R(t) =
u(t) for all t. Then:
u is perpendicular to that of 1. The direction of
=
dt
=
d
3
a(P ),d(P )]
Denition 3 The torsion is a function dened by (P) =
[
| a(P )|
2
.
v(P )
The curve C is uniquely determined by the functions k and .
2
v
a
Lecture VII
Paths and Curves
First we go through several basic notions about paths. Let
R(t) on [a, b] be a
given path.
Denition 1 R(t) is called elementary if for every pair (t
1
, t
2
), with t
1
and t
2
distinct in [a, b], R(t
1
) = R(t
2
).
Denition 2 R(t) is called simple if for every pair (t
1
, t
2
), with t
1
and t
2
distinct in [a, b], except possibly for the pair (a, b), R(t
1
) = R(t
2
).
i+Rsin t
j.
Let C be the curve generated by a point P xed on a circle of radius a rolling
without slipping along the x axis. Suppose the curve begins with P at O and
ends when P again touches the x axis. The circle makes a full rotation, so it
would be advisable to take , the angle through which the circle rotates, as
the variable for a path for C. The coordinates of the center of the circle are
(a, a), so the coordinates of P are (a( sin ), a(1 cos )). Hence the path
R() = a( sin )
i + a(1 cos )
is
j, with from 0 to 2.
Let
R be a path for the curve C and let P be the point with position vector
R(t0)
=
R d
R(t
0
). Recall that lim
t 0
R(t0+t)
dt
| = v(P) is a tangent vector
t
t0
to C at P.
1
R
R(t
0
). If
d
dt
| exists and
t0
is not equal to 0, then we denote by T
P
the unit tangent vector at P, the vector
t0
d
R d
R
T
P
=
dt dt t0
t0 d
R
Let s be the arc length of C, s(t
0
) =
a
|
dt
|dt. Then
d
R R ds R
v = = and T
=
d
, so
dt ds
dt ds
v = T
ds
= T
s
dt
dT
dT
. at P, N
P
=
ds
|
ds
|
N =
dT
Clearly, k
ds
and N
P
is orthogonal to T
P
. Since v = sT
, we get that
a =
sT)
N
d(
= sT
+ sT. Since T
=
dT
=
dT
ds
= k
s, we have that
dt dt ds dt
a = sT
+ sT
= sT
+ ks
2
N
Denition 8 Let P be a point where T
P
and N
P
both exist. The vector
B
P
=
P
N
P
is called the unit binormal vector at P.
B
P
form a frame at P, that is called the moving frame The vectors T
P
, N
P
, and
at P.
2
| | | |
|
1
Lecture VIII
Scalar Fields
Cylindrical Coordinates
Scalar Fields
Denition 1 Let D be a subset of E
3
. A function f that associates each point
P in D to a real number f(P) is called a scalar eld. D is called the domain
of f.
Denition 2 Let f be a scalar eld on a domain D. Let u be a xed unit vector
and let P be a xed point. For any point Q such that PQ is parallel to
u, let
f(P)F(Q)
s(P, Q) = PQ u. Consider the limit lim
s 0
s(P,Q)
. If this limit exists,
then we denote it
df f
ds
|
u,P
= lim
s 0 s
and call it the directional derivative of f in the direction u at P.
Denition 3 Let f be a scalar eld on D, and let P(a, b) be a point in D.
The partial derivative of f with respect to x at P is the derivative at a of the
function f(x, b). It is denoted by
f
P
.
x
|
The partial derivative with respect to x equals the directional derivative in the
i :
f df
direction
x
|
P
=
ds
|
i,P
. Also, the partial derivative with respect to y equals
df
j :
f
the directional derivative in the direction
y
|
P
=
ds
|
j,P
. Since they are
dened on subsets of D, the partial derivatives can be viewed as scalar elds.
f
Let f(x, y) = x
2
+ cos xy. Then
x
= 2x y sin xy is dened on the same
domain as f.
Denition 4 Let f be a scalar eld on D and let P be a point in D. We say
the limit of f at P is l and denote this by lim
QP
f(Q) = l, if there exists a one-
variable function on a domain [0, d] such that (t) decreases to 0 as t decreases
to 0, and for all points Q = P such that |QP d, then f(Q) l ( QP ).
1
2 Cylindrical Coordinates
Let us dene a non-Cartesian coordinates system that will prove usefull in ap-
plying calculus techniques to certain functions: the cylindrical coordinates sys-
tem.The coordinates of a point P in this system are denoted r, and z. let us
rst take a Cartesian system in E
3
. To nd the cylindrical coordinates of a
point P, we rst let P
|
and is the angle between Ox and OP
R, we write
j + zk R = x
i + yj + zk = (r cos )
i + (r sin )
2
Lecture IX
Linear Approximation
1 One-variable Functions
Let f be a one-variable function on a domain D. If f is dierentiable at x = c,
we say f has a linear approximation, which we dene in the following way.
Denition 1 If f is a function dierentiable at c, for any x dene f =
f(x) f(c) and x = x c. There exist a scalar A
c
and a function
c
(x) such
that
f = A
c
x +
c
(x)x and lim
c
(x) = 0.
x c
Then
f
A
c
= f
(c) and
c
(x) =
x
f
(c).
We say that f has a linear approximation at c and that the expression f
app
=
f
1
(Q) = 0 and lim
QP
2
(Q) = 0.
Theorem 1 If f has a linear approximation at P, then f has partial derivatives
at P and
f
y
|
x
|
P
= A,
f
P
= B.
1
Theorem 2 If f has continuous partial derivatives on all the domain, then f
has a linear approximation.
For a multivariable function f that has a linear approximation at P, we
f
dene f
app
=
f
P
x +
y
|
P
y, the linear approximation formula for f at
x
|
P.
f f
For a three-variable function f, we say that f has a linear approximation
at P if there exist A, B, C scalars and
1
,
2
,
3
functions such that f =
Ax + By + Cz +
1
(Q)x +
2
(Q)y +
3
(Q)z and lim
QP
1
(Q) = 0,
lim
QP
2
(Q) = 0, lim
QP
3
(Q) = 0. If f has a linear approximation at P,
then f has partial derivatives at P, and A =
x
|
P
, B =
f
P
, C =
z
|
P
. The
y
|
f
P
x+
f
P
y+
f
linear approximation formula for f at P is f
app
=
x
|
y
|
z
|
P
z.
2
1
Lecture X
Linear Approximation
Chain Rule
Linear Approximation; Gradient
We say that a function has a linear approximation on a domain D if it has a
linear approximation at any point P D.
Theorem 1 If f has a linear approximation on a domain D, then f is contin-
uous on D.
Theorem 2 If f has continuous partial derivatives on D, then f has linear
approximation on D.
Denition 1 Let f be a function with partial derivatives at P. The vector
G
P
=
i +
f
j + k
x y z
is called the gradient at P and is usually denoted by
P
. f|
Theorem 3 (Gradient Theorem) Let f have domain D in E
3
and let P be
a point in D. If u is a xed unit vector, then
df
=
f
i +
f f
j +
k =
u f
P
|
ds u,P x P y P z P
Chain Rule
Physical variables, quantities that can be measured in a given physical system,
are related in various ways. For example, pressure p is a function of volume V
and temperature T, p = h(V, T). Also, T depends on V and internal energy U,
T = t(V, U). From these to equalities, we get that p is a function of volume V
1
2
and internal energy U, p = h(V, T) = h(V, t(V, U)) = g(V, U). So to nd the
partial derivative of p with respect to V , we have to speciy the function we
=
h
are referring to: h or g. We write (
p
=
g
and (
p
. Below we
V
)
U
V V
)
T
V
will see how such partial derivatives are related. By (
f
x
)
y
we mean the partial
derivative of f with respect to x when f is seen as a function of variables x and
y.
Theorem 4 (Chain Rule) Let w, u, v, x, y be physical variables such that there
exist dierentiable functions f, g, h such that w = f(u, v), u = g(x, y) and
v = h(x, y). Then
w
=
f g
+
f h
x
y
u x v x
w
=
f g
+
f h
y
x
u y v y
If we take w = f(u, v), u = g(x, y), and v = h(x), then applying the Chain Rule
we get
w f g
= ,
y
x
u y
h
since = 0.
y
2
Lecture XI
Chain Rule: Elimination Method
Let w = f(x, y) be a dierentiable function of x and y. The linear approximation
of f is given by
f
app
= f
x
(x, y)x + f
y
(x, y)y.
We introduce a new notation, the dierential notation for the increments f,
x, y, namely we write df, dx, dy instead: df = f
x
dx + f
y
dy. This expression
is called the dierential of f. For example, the dierential of
2
w = x
2
+ y 1 is dw = 2xdx + 2ydy.
For any function w = f(x, y), the equality
w w
dw = dx + dy
x
y
y
x
holds. By the elimination method, we can nd (
w
x
)
y
and (
w
if w is not given
y
)
x
directly as a function of x and y but can be reduced to such a function. We
ilustrate this in the following example. Consider the these two equalities:
e
xy
w = f(x, y, z) = xyz, z = g(x, y) = .
Then the dierentials of w and z are
dw = yzdx + xzdy + xydz and dz = ye
xy
dx + xe
xy
dy.
Substituting dz in the rst equality, we get
2 2
dw = (yz + xy e
xy
)dx + (xz + yx e
xy
)dy.
Then the derivative of w with respect to x when w is seen as a function of x
and y is precisely th term of dx in the equality above:
w
= yz + xy
2
e
xy
and
w
= xz + yx
2
e
xy
.
x
y
y
x
1
Lecture XII
Terminology for Point-Sets in Euclidean Spaces and
Minimum-Maximum Theorems
First let us take a short look at a problem that was on the exam. We are
given a level curve (in E
2
) or a surface(in E
3
) and a point P on that curve or
surface. How to nd a vector normal to that curve or surface at that point P?
Let us consider the case of E
3
. The answer lies in the gradient of the function
dening the surface. The graph of the surface is given by z = f(x, y), hence
z f(x, y) = g(x, y, z) = 0. The gradient of g at P is a vector normal to the
surface at P:
f f
P
=
P
j +
k. g|
x
|
P
i
y
|
Now let us take a look at some basic notions of point-set topology of Eu-
clidean spaces.
Denition 1 Given a point P, we dene a neighborhood of P in the following
manner:
(i) For the 1-dimensional space, a neighborhood is an interval of the form
[c r, c + r] for some r > 0, where c is the point P.
(ii) For the 2-dimensional space, a neighorhood is a disc of points with center
at P and radius r.
(iii) For the 3-dimensional space, a neighborhood is a solid ball of radius r and
center P.
In all three cases, a neighborhood is a set U for which there exists r R such
that U is the set of all points Q with PQ r.
Let D be a given set of points.
Denition 2 A point P is an interior point of D if there is some neighborhood
of P which is entirely contained in D.
1
Denition 3 A point P is a boundary point of D if every neighborhood of P
contains at least one point in D and at least one point not in D.
Denition 4 D is called open if every point in D is an interior point of D.
D is called closed if it contains all its boundary points.
Note: D is open if and only if D contains none of its boundary points.
Denition 5 D is bounded if there is some point P and some neighborhood of
P such that the neighborhood contains the set D.
Denition 6 D is compact if D is closed and bounded.
If f is a path for D, how do we nd points P in D where f has a maximum value
or a minimum value? We investigate this question, providing some methods of
solving it. In doing so, we will consider the input(i.e. the points we are working
with) to be of the form x, contained in the domain of f, and not of the form P,
contained in D.
Denition 7 Let f be dened on [(a, b)] and let x be a point in its denition
domain.
(i) x is a global maximum point if f(x) f(y) for all y in the denition
domain.
(ii) x is a global minimum point if f(x) f(y) for all y in the denition
domain.
(iii) x is a local maximum point if there exists a neighobrhood U of x such
that f(x) f(y) for all y in intersection of U and the denition domain.
(iv) x is a local minimum point if there exists a neighobrhood U of x such that
f(x) f(y) for all y in intersection of U and the denition domain.
All maximum and minimum points, global and local, are called extreme points
of f. The two following theorems are useful in nding extreme points.
Theorem 1 (Min-Max Existence Theorem) Let f be a continuous func-
tion, having D as its domain. Assume D is compact.Then f has at least one
global maximum point and at least one global minimum point in D.
2
Denition 8 Let f be a dierentiable function. If
P
= 0, then P is called f|
a critical point.
Theorem 2 (Critical Point Theorem) Let f be a dierentiable function with
a domain D and let P be an interior point of D. If P is an extreme point, then
P is a critical point.
3
Lecture XIII
Two-Variable Test
Constrained Maximum-Minimum Problems
1 The two-variable test
Recall that by the Critical Point Theorem, onlyif the gradient of a function f
at P is 0 (i.e. P is a critical point for f), can P be an extreme point of f.
Let f(x, y) = x
3
+ xy + y
3
. Then
f = (3x
2
+ y)
i + (x + 3y
2
)
j,
so if P = (x, y) is a critical point, then
2 2
x = 3y , y = 3x ,
1 1
hence x + 27x
4
= 0, so the only critical points are P
1
(0, 0) and P
2
(
3
,
3
).
Now we have to nd out if these points are extreme points. For this we can use
the following test.
Theorem 1 (The two-variable test) Let f be a two-variable C
2
function,
and let P be an interior critical point for f. We dene the functions H
1
and
H
2
, called Hessian functions in the following manner:
H
1
(P) = f
xx
(P),
H
2
(P) =
f
xx
(P) f
xy
(P)
f
yx
(P) f
yy
(P)
= f
xx
(P)f
yy
(P) f
2
xy
(P)
Then:
(a) If H
1
(P) > 0 and H
2
(P) > 0 then P is a local minimum point for f.
(b) If H
1
(P) < 0 and H
2
(P) > 0 then P is a local maximum point for f.
1
(c) If H
2
(P) < 0 then P is a saddle point for f.
Going back to the function f(x, y) = x
3
+ xy + y
3
, we nd that H
1
(x, y) = 6x
and H
2
(x, y) = 36xy 1. Hence H
1
(P
1
) = 0, H
2
(P
1
) = 1, so P
1
(0, 0) is a
1 1
saddle point. Also, H
1
(P
2
) = 2 and H
2
(P
2
) = 3, so P
2
(
3
,
3
) is a local
maximum point.
2 Constrained problems
Suppose f is a function dened on a region D in E
3
and C is a surface or a
curve in D. We can think of C as a new domain for f and try to nd extreme
point for f in C. Such a problem is a called a constrained problem.
Let f and g be C
1
and let D
be the set of all points P such that g(P) = c.
Assume
P
= 0 at all points P in D
. Let D
be the new restricted domain f|
of f. Such a case is called a one constraint case, since the restricted domain is
given by one function, g.
Denition 1 Let P be a point in D
P
. for f on D
if there is a scalar
P
such that
g|
f|
P
= 0, then
|
P
is normal
to D
at P.
Theorem 2 If P is a extreme point for f on D
.
The one-constraint Lagrange method of nding extreme points on D
uses the
theorem above in the following manner. First we dene a function h on D
,
h(x, y, z, ) = f(x, y, z) g(x, y, z). Then we form the system of equations
h
x
(x, y, z, ) = 0, h
y
(x, y, z, ) = 0, h
z
(x, y, z, ) = 0, g(x, y, z) = c
and solve it. Its solutions are the constrained critical points of f. Finally, among
these points we identify the extreme points for f on D
.
The case of two constraints can be described as follows. Let f, g
1
and g
2
be C
1
functions. Let D
be the set of all points P such that g
1
(P) = c
1
and
g
2
(P) = c
2
. Assume
g
2 P
= 0 for all points P in D
. g
1
|
P
|
2
Denition 2 Let P be a point in D
1
g
1 P
+
P
P
.
1 2
|
P
=
P
|
2
g
2
|
Theorem 3 If P is a extreme point for f on D
.
The Lagrange method for two constraints is similar to the method for one con-
straint. We dene h(x, y, z,
1
,
2
) = f(x, y, z)
1
g
1
(x, y, z)
2
g
2
(x, y, z).
We form the system of equations
h
x
(x, y, z,
1
,
2
) = 0, h
y
(x, y, z,
1
,
2
) = 0, h
z
(x, y, z,
1
,
2
) = 0,
g
1
(x, y, z) = c
1
g
2
(x, y, z) = c
2
and solve it. The solutions will be the constrained critical points. From among
these points, we identify the extreme points of f on D
.
3
Lecture XIV
Multiple Integrals
1 Integrals of one-variable functions
For a real-valued function f dened on an interval [a, b], the integral of f over
b
[a, b], denoted by
a
fdx, is dened as follows:
n
n,max xi0
a
i=1
where a = x
1
< . . . < x
i
< x
i+1
< x
n
= b, x
i
= x
i
x
i1
and <
x
i
[x
i1
, x
i
] for all i 2, i n. We will now give a few properties and
theorems about integrals in one-variable calculus.
Theorem 1 (First Existence Theorem) Let f be a one-variable function.
b
If f is continuous on [a, b], then
a
fdx exists.
b
Let f and g be two functions dened on [a, b] such that
a
fdx and gdx exist.
a
Then the following properties hold:
1. Let c be such that a < c < b. Then the following integrals exist and the
equality holds:
b
fdx = fdx + fdx.
a a c
2. The integral in the left side of the equality exists and the equality holds:
b
(f + g)dx = fdx + gdx.
a a a
3. If c is a constant, the integral in the left side of the equality exists and the
equality holds:
b
(cf)dx = c fdx.
a a
1
2
4. If m
1
and m
2
are two constants such that m
1
f(x) m
2
for all x [a, b],
then
b
m
1
(b a) fdx m
2
(b a).
a
Scalar-valued integrals in E
2
and E
3
If f is a scalar-valued function dened on E
2
or E
3
, how do we dene the
integrals of f on a region R? We can only do this on a special kind of region,
called regular, which we will dene in section 5.
Denition 1 Let R be a regular region in E
2
and f a scalar eld on R. The
integral of f on R, denoted by
R
fdA, is dened as follows:
n
fdA = lim f(P
i
)A
i
,
R
nmax di0
i=1
where A
1
, . . . , A
n
form a subdivision of R into elementary regions, A
i
is the
area of A
i
, P
i
is a point in A
i
and d
i
is the diameter of A
i
, i.e. the longest
distance between two points in A
i
, for all i 1, i n.
Theorem 2 (Second Existence Theorem) If R is a regular region in E
2
and f is continuous on R, then
R
fdA exists.
For E
3
, the denition of the integral is very similar.
Denition 2 Let R be a regular region in E
3
and f a scalar eld on R. The
integral of f on R, denoted by
R
fdV , is dened as follows:
n
fdV = lim f(P
i
)V
i
,
R
nmax di0
i=1
where V
1
, . . . , V
n
form a subdivision of R into elementary regions, V
i
is the
volume of V
i
, P
i
is a point in V
i
and d
i
is the diameter of V
i
, for all i 1, i n.
Theorem 3 (Third Existence Theorem) If R is a regular region in E
3
and
f is continuous on R, then
R
fdV exists.
Remark: If we consider a constant function f(x, y) = 1 over a region R, then
fdA = dA is equal to the area of R.
R R
2
Sometimes it is useful not to use the Cartesian coordinate system. For
example, let R be the disc bounded by the curve x
2
+ y
2
= 4. Let f(x, y) =
2
x
2
+ y . We want to nd the integral
R
f(x, y)dA. We divide this disc into
n concentric rings, and further divide these rings by radial segments. For each
ring j , we choose a r
j
r
j
, where r
j
j
is the dierence between the radii of rings j and j 1. Then we can choose a
point P
) = (r
j
)
2
. Hence the
m
Reimann sum for this choice of subdivision is 2(r
j
)
3
r
j
. Now we can
j=1
view this sum as the Reimann sum of a function of r. Hence
2
x
2
+ y
2
dA = 2r
3
dr = 8.
R 0
3 Properties of multiple integrals
1. Let R be a regular region in E
3
and let R
1
and R
2
be two regular disjoint
regions such that R = R
1
R
2
. If fdV exists, then the other
R
integrals in the following equality exists, and the equality holds:
fdV = fdV + fdV.
R R1 R2
The equivalent for E
2
holds as well.
Here is an application of this property. Let f(x, y) = x for all x, y, and let
R be the disc bounded by x
2
+y
2
= 1. Let R
1
be the part of R that is left
of Oy, and let R
2
be the part of R that is right of Oy. Then
R
fDA =
R1
fdA + fdA. But by symmetry,
R1
xdA =
R2
xdA, so
R2
fDA = 0.
R
2. Let R be a regular region in E
3
and f a function on R such that
R
fdV
exists. If m
1
and m
2
are two constants such that m
1
f(P) m
2
for all
points P in R, then the following inequalities hold:
m
1
V fdV =m
2
V,
R
where V is the volume of R. The equivalent for E
2
also holds.
As an application of this property, consider the region R to be the square
having (0, 0) and (2, 2) opposite vertices. Let f(x, y) = 2 +
xy
for all
100
3
i+
M(x, y, z)
k. The integral of
j + N(x, y, z)
i + MdV j + NdV k.
4
5
As an application, consider an object that ocupies a region R in space and has
density (x, y, z) at any point P(x, y, z). The center of mass of this object s
given by
1
C
M
= (x, y, z)OP(x, y, z)dV,
M
R
where M is the total mass of the object, i. e. M =
R
(x, y, z)dV . Hence
C
M
=
1
(x, y, z)(x
i + yj + zk)dV =
M
R
1
= x(x, y, z)dV
i+
M
R
1
j + k. + y(x, y, z)dV
1
z(x, y, z)dV
M
R
M
R
Further remarks
In the existence theorems, the condition that f be continuous can be replaced
with piecewise continuous.
Denition 3 A function f is said to be piecewise continuous on a regular region
R if it is dened on R and it is possible to divide R in a nite number of regular
subregions R
1
, . . . , R
n
such that for each subregion R
i
, the restiction of f to the
interior of R
i
can be extended to a function f
i
dened on the boundary of R
i
as
well, such that f
i
is continuous. It is not necessary that f
i
coincide with f on
the boundary of R
i
.
Now let us dene the concept of regular region. For this we will need to dene
rst what an elementary region is.
Denition 4 A region R is called elementary if there is a piecewise smooth,
simple, closed curve such that R is the union of the curve and its interior.
Denition 5 A region R is called regular if it can be divided into a nite num-
ber of elementary regions in the following fashion: the regions form a connected
whole and neighboring regions touch along a unique segment of common bound-
ary curve, that has positive length.
5
2
Let R be the region in E
2
dened by all points (x, y) such that x
2
+ y 1
and both x and y are irrational numbers. Then R is an example of a region
that is not regular.
6
Lecture XV
Iterated Integrals
In this lecture we look at methods to compute multiple integrals, introducing
iterated integrals. First, let us dene the type of regions for which it can be used.
These regions are called simple regions.
Denition 1 A region R in E
2
is called simple if it has one of the following
properties:
(i) There exists an interval [a, b] and there exist continuous functions g
1
(x)
and g
2
(x) dened on [a, b] such that g
1
(x) g
2
(x) for all x [a, b] and
the region R is the set of points (x, y) such that a x b and g
1
(x)
y g
2
(y). A region with this property is called y-simple.
(ii) There exists an interval [c, d] and there exist continuous functions h
1
(y)
and h
2
(y) dened on [c, d] such that h
1
(y) h
2
(y) for all x [c, d] and
the region R is the set of points (x, y) such that c y d and h
1
(y)
x h
2
(x). A region with this property is called x-simple.
Denition 2 Let R be an y-simple region in E
2
, with g
1
(x) and g
2
(x) its lower
and upper boundary functions, as in Denition 1(i). Let f be a continuous
function on R. For each x [a, b] we dene h(x) by
g2(x)
h(x) = f(x, y)dy
g1(x)
b
Then we compute the integral of h on [a, b],
a
h(x)dx. This two-step operation
is called a double iterated integral. The order of integration is rst y, then x.
Alternatively, we can denote the iterated integral by
g2(x)
f(x, y)dydx.
a g1(x)
In the same manner we can dene iterated integrals for x-simple regions. We
say then that the order of integration is rst x then y.
1
Let the region R be the set of all points (x, y) with 0 x 1, x
2
y x. We
dene f as the function on R such that f(x, y) = xy for all (x, y) R. Let us
compute the iterated integral of f:
1
1
x
x
3
x
5
1 1 1
xydydx = dx =
2
2 8
12
=
24
.
We can see that R is also x-simple, with 0 y 1 and h
1
(y) = y x
y =
h
2
(y). Hence we can compute the iterated integral of f in the order x, then y.
1
0 x
2
0
y
2
y
3
1 1 1
xydxdy = dy =
2
2 6
8
=
24
.
0 y 0
It is not by chance that the two integrals have the same value. In fact this can
be proved for any continuous function f on a x, y-simple region R.
Theorem 1 (Fubinis Theorem) Let R be a x, y-simple region and let f be
a continuous function on R. Then
b
g2(x)
h2(y)
fdA = f(x, y)dydx = f(x, y)dxdy,
R a g1(x) c h1(y)
where a, b, c, d and the functions g
1
, g
2
, h
1
, h
2
are as in Denition 1.
The notion of iterated integrals can be extended to E
3
. Consider a region R
in E
3
such that the projection R
g2(x)
h2(x,y)
f(x, y, z)dz dy dx.
a g1(x) h1(x,y)
In E
3
, we can use two methods to compute triple integrals.
(i) Method of slices
Let f be a function on a region R. We dene R
z
to be the intersection
of R with the plane parallel to the xy plane, that intersects the z axis at
z. If the double integral
Rz
f(x, y, z)dA is a continuous function of f,
then
c2
fdV = f(x, y, z)dA dz,
R c1 Rz
where c
1
and c
2
are the minimum and maximum values for z in R.
2
be the projection of R in
the xy plane. For each (x, y) R
we dene h
!
(x, y) and h
2
(x, y) to be
the minimal and maximal value that z takes when (x, y, z) R. If R
coincides wtih the set of all points (x, y, z) with (x, y) R
and h
1
(x, y)
z h
@
(x, y), then
h2(x,y)
fdV = f(x, y, z)dz dA.
R R
h1(x,y)
3
Lecture XVI
Integrals in Polar, Cylindrical, or Spherical Coordinates
Usually, we write functions in the Cartesian coordinate system. Hence we
write and compute multiple integrals in Cartesian coordinates. But there are
other coordinate systems that can help us compute iterated integrals faster. We
analyze bellow three such coordinate systems.
1 Polar coordinates
In E
2
, the polar coordinates system is often used along with a Cartesian system.
The polar coordinates of a point P(x, y) are r and where r is the distance from
the origin O and is the angle done by OP and Ox measured counter-clockwise
from Ox. The equations we use to switch from one system to the other are:
x = r cos , y = r sin , r = x
2
+ y
2
.
To transform an integral
R
f(x, y)dA from Cartesian coordinates to polar
coordinates, we take a look at Reimann sums. By consider a particular type of
subregions, we get that dA = rdrd, so we can write:
f(x, y)dA = f(r cos , r sin )rdrd,
R R
where
R is
r-simple, we can compute the iterated integral in polar coordinates.
For example, let us compute the area of the region R enclosed by the curve
r = 1 + cos , where goes from 0 to 2. By looking at the graph of the curve,
we can see that the area is equal to twice the area enclosed when goes from
0 to . In polar coordinates, we can see that
1+cos
dA = 2 rdrd
R 0 0
1
2 1+cos
= (r | )d
0
0
= 1 + 2 cos + cos
2
d
0
= + cos
2
d
0
Knowing that for any n Z, the following equality holds
n
cos
2
d = sin
2
d,
0 0
and that cos
2
+ sin
2
= 1 for all , we get that
cos
2
d =
2
, so
0
3
dA = + cos
2
d =
2
.
R 0
2 Cylindrical coordinates
In E
3
, it is easier sometimes to use the cylindrical coordinates system. We can
view this system as the extension of the polar coordinates system in E
3
. Let
P(x, y, z) be a point in E
3
and let P
R R
x
2
+ y
2
+ z
2
, = arccos
z
.
x
2
+ y
2
+ z
2
To move integrals from the Cartesian system to the spherical system, we use
the following equality:
dV =
2
sin ddd.
Hence in terms of integrals, the following equality holds
f(x, y, z)dV =
R
= f( sin cos , sin sin , cos )
2
sin ddd,
Lecture XVII
Curvilinear Coordinates; Change of Variables
As we saw in lecture 16, in E
2
we can use the polar coordinates system.
In this system, we have a xed point O and a xed ray Ox. The coordinates
of a point P are given by r, the distance from P to O, and the angle made
by
Ox to OP. We can OP and Ox, as measured going counterclockwise from
change the system of coordiantes from polar to Cartesian through a system of
equations:
x = r cos , y = r sin .
This is called the dening system. To go from the Cartesian system to the polar
one, we use the inverse dening system:
r = x
2
+ y
2
, = arctan
y
,
x
for (x, y) in the rst quadrant. Generally, we can introduce new non-Cartesian
coordinates u, v by writing x, y as functions of these new coordiantes:
x = g(u, v), y = h(u, v).
These equations form the dening system for the new coordiantes. They can be
summarized by writting the position vector
j. R(u, v) = x
i + yj = g(u, v)
i + h(u, v)
R g h
P
=
P
j,
u
|
u
|
P
i +
u
|
and we form the unit vector
R
v
P
= u
P
=
R
P
/|
P
. Similarly, we dene
u
|
u
| |
R
R
v
|
P
/|
v
|
P
. For example, recall that for polar coordinates, the dening system |
is x = r cos , y = r sin . The position vector is:
R = r cos
i + r cos
j,
1
R
j = cos
i + sin
= r,
r
R
j
= sin
i + cos
= .
In the uv plane, consider the rectangle determined by the vectors vv uu and
with point P as their tail, where u and v are two small positive quantities.
Now consider the point P
R R
u
P
|
P
v
| uv
R R
P
u
u
| |
v
This vector product can be easily computed, since
R g
R
=
g
i +
h
j and
= i + j.
u u u v v v
Hence
R R
u
v
= =
(x, y)
(u, v)
.
x x
u v
y y
u v
This expression is called the Jacobian of x and y with respect to u and v. For
polar coordinates, the Jacobian is
(x, y)
(u, v)
=
cos r sin
sin r cos
= r cos
2
+ r sin
2
= r.
When switching from the Cartesian system to a curvilinear one, u, v might be
given as functions of x, y, so the following equality is useful:
(u, v) (x, y)
= 1 .
(x, y) (u, v)
The notion of Jacobian can be extended to E
3
in the following manner. If x, y, z
are functions of u, v, w, the Jacobian is
(x, y, z)
(u, v, w)
=
x x x
u v w
y y y
u v w
z z z
u v w
2
.
R R
(x, y)
(u, v)
dudv.
3
1
Lecture XVIII
Change of Variables; Vector Fields
Change of Variables
Recall from lecture 17 that we change variables in integrals by the following
formula:
f(x, y)dxdy = f(x(u, v), y(u, v))
R R
x x
u v
y y
dudv,
u v
where R and
R corresponding to R in
the uv plane is the rectangle of vertices (1, 1), (3, 1), (3, 2), and (1, 2). Changing
variables to u, v, the area of R is given by
(x, y)
(u, v)
dudv =
2
3
dxdy =
2
3
(u, v)
(x, y)
1 dudv =
R 1 1 1 1
2
3
2
3
1 y x
= 1 dudv = dudv =
xy
2
y
2
2xy 1 1 1 1
=
2
3
1
dudv =
2
2
dv = 2 ln 2.
1 1
v
1
v
2 Vector Fields
Denition 1 A vector eld in E
2
is a function dened on a region R in E
2
that gives as outputs vectors in E
2
.
1
Clearly all vector operations apply to vector elds as well. We can write a vector
eld as following:
F(x, y) = f(x, y)
i + g(x, y)
j.
An important category of vector elds that we will be working with is the
category of gradient elds. We will see that not all vector elds are gradient
elds.
Denition 2 Let
F(x, y) = f(x, y)
i + g(x, y)
F be a vector eld,
j. F can be a
gradient eld only if f
y
= g
x
.
This test doesnt always tell us if F is a gradient led or not, it only states
that if the condition
f
=
g
isnt satised, then we can be sure that F isnt a
y x
gradient eld.
Proof of Lemma 1:
F
Let
= f
h
x
i + h
y
j.
Hence f = h
x
and g = h
y
, and since (h
x
)
y
= (h
y
)
x
, we get that f
y
= g
x
.
2
1
Lecture XIX
Visualizing Vector Fields; Line Integrals
Visualizing Vector Fields
Recall that a vector eld in E
2
is a function of the form
F(x, y) = f
1
(x, y)
i + f
2
(x, y)
j.
We dene two concepts that help us visualize vector elds.
Denition 1 Let C be a directed smooth curve in E
2
, and let
F be a vector
eld in E
2
. Then C is called an integral curve of F if, at any point P on C,
F(P) = 0 and
as T
P
, the unit tangent vector to C at P .
Recall that C is the class of all continuous scalar functions and that C
1
is the
class of all dierentiable functions with continuous partial derivatives. We say
that the vector eld
F as dened above is in C
1
if f
1
, f
2
are in C
1
.
Lemma 1 For any point P such that F(P) = 0, there exists an integral curve
for F through P.
F
Let us take
= x
i + yj = r
Denition 2 Let
2 Line Integrals
Denition 3 Let C be a nite curve in E
2
and let f be a scalar eld dened
on C. We divide C into n pieces of arc-length s
1
, s
2
, . . . , s
n
and from each
n
piece we choose a point P
i
)s
i
. If the
i=1
limit
n
lim f(P
i
)s
i
n
max si0 i=1
exists, then it is called the scalar line integral of f on C. It is denoted by
C
fds.
ds R
Remember that s is the arc length of C and that
dt
= |
d
. We can evaluate
dt
|
line integrals by two methods.
1. Evaluation by denition
In this method, we simply use the denition of the integral. For example,
let C be the circle of radius 2 with center at the origin, and let f(x, y) =
x
2
+ y
2
. Then
fds = (x
2
+ y
2
)ds = 4ds = 4 4 = 16,
C C C
since the arc length of C is 4.
Let us extend the notion of line integral to vector elds. If
F(x, y) =
f
1
(x, y)
i + f
2
(x, y)
Fds = f
1
ds
i + f
2
ds
j.
C C C
F
Let us take
= x
i +y(t)
j, with t going
from a to b. Then
=
d
R
dt
=
d
R dx
ds
j
i +
dy
i + j, x x
2
+ y
2
. = = y so
dt dt dt dt
2
R
dt
dt =
b
fds = f
C a
f(x(t), y(t)) x
2
(t) + y
2
(t)dt.
a
3
1
Lecture XX
Vector Line Integrals; Conservative Fields
Vector line integrals
Let
F be a vector eld of domain D. Let D be connected, i.e. for any two points
P, P
in D there is a curve C contained in D that goes from P to P
. Let C be
a nite directed curve contained in D that is rectiable, i.e. it has a length. We
d
F R = lim F(P
i
T(P
i
)s
i
,
C
n
max si0 i=1
where T
, s
i
are dened in the
same way as for scalar line integrals. Let us give two basic laws for vector line
integrals:
1. Let
d
d
d
F R = F R + F R.
C C1 C2
F and
2. Let
d
G d
R. aF + b
R = a F R + b G d
C C C
There are two ways of evaluating vector line integrals. First, we can evaluate
using the denition of the integral as the limit of a Riemann sum. We can
also evaluate by parameter. Let R(t) be a path for C, with t from a to b,
d
R
j + z(t)
x(t)
i + y(t)
k. Since T
=
d
R
R(t) and
ds
dt
dR
, we get that = =
dt dt dt
b
d
d
F R = F T
ds = F dt
dt
C C a
1
2 Conservative Fields
Denition 1 Let D be a domain for a vector eld F. If for any points A, B D
curve C D, F R = 0, then
1
C
2
Lecture XXI
Line Integrals; Conservative Fields
Line integrals
Let us recapitulate the basic notions reering to line integrals. For both scalar
and vector elds, we can dene line integrals. For a scalar eld f on a curve
fds.
C
For a vector eld
R
dt
dt,
b
b
d
d
F R = F fds = f dt.
dt
C C a a
For vector elds we can also dene other line integrals as well. If
F(x, y) =
R
Fds
M(x, y)
i + N(x, y)
j, then
i +
j. Also, is
C
R
Fds = Mds Nds
C C C ds
C
the average value of
R F d
=
C
R
dt.
C
F
dt
Conservative elds
F. We say that
F has independence of
F R = 0. If
d
then
= F.
Theorem 1 (Conservative-eld Theorem) Let
F be a continuous vector
eld on a domain D. Then
F is a gradient
eld.
Proof:
1
Since
d
R = R(t))dt = f(
R(a)). F R = f d
d
f(
R(b)) f(
dt
C C a
Let Q be a xed point in D. For any point P D, we dene f(P) =
F is conservative,
C
F is a F, so
gradient eld.
If we know that
F.
1. The vector line integral method uses the conservative eld theorem. More
F.
2. The indenite integral method is often simpler than the line integral
method. Let
j F(x, y) = M(x, y)
i + N(x, y)
= (2x3y4)
i+(4y3x+2)
Then f = x
2
3xy 4x + C(y), so
f
= 3x + C
(y) = 4y 3x + 2.
y
Hence C
(y) = 4y + 2, so C(y) = 2y
2
+ 2y + c. It follows that f(x, y) =
x
2
3xy 4x + 2y
2
+ 2x + c is a scalar potential for
F.
2
Lecture XXII
Surfaces
Recall that an elementary region
D contained in E
2
is called convex if for
D, the line conecting them is contained in
any points in
R. map
D give dierent values of
D
with values in E
3
. The surface S is the set of all points that have
R(u, v) as
position vector for some (u, v)
D.
For example, in the uv plane we consider the rectangle 0 u , 0 v 2,
and the function
at O. Also, if we take
R(u, v) =
k dened on D, then the surface given by
ui +vj +
1 u
2
v
2
R is the sphere
of radius 1 centered at O.
Denition 2 Suppose
be an
u
|
P
v
|
P
vector given by P
. Taking P = R(P
), we can denote P) as w
P
and call it
w(
in
S. More precisely, if
D, and U is the
corresponding region around P, then surface area (U) w
p
area (
| | U).
|
Lecture XXIII
Surface Integrals
Remember that the parametric expression for a curve is given by a vector func-
j + z(t)
i + y(t)
i + y(u, v)
j +
z(u, v)
U around P
to the surface
area U around P.
We can use the Jacobian to nd the paramteric nromal vector:
d
R d
R (y, z)
(x, z)
(x, y)
i + j +
k. w(u, v) = =
P
(u, v) (u, v) (u, v) du dv P
Consider the semicylinder S of radius a, with 0 z b and y 0. In cylindrical
coordinates, it is described by 0 , 0 z b, r = a. Clearly, since w
P
is normal to S,
w
P
= r. Since z and in the z plane correspond to z
|w
P
|
and a on the semicylinder, |w
P
= a. Hence w
P
= ar. We now give several
denitions concerning types of surfaces.
Denition 1 A surface is said to be elementary if its representation
R is a
one-to-one mapping.
Denition 2 A surface is nite if it can be divided into a nite number of
elementary surfaces, where any two of these surfaces intersect at most along
their common boundary, and no segment of boundary is contained in more than
two of these surfaces.
Denition 3 A nite surface S is one-sided if there exists a continuous loop
on S such that if we start from a point P and consider the parametric normal
1
vector on the loop, when arriving at P again the vector has direction opposite
of the one from the start. If such a loop does not exist, then we say the surface
is two-sided.
Denition 4 A nite surface that has no boundary is called closed.
Denition 5 Let D be a connected region in E
3
. D is simply-connected if for
every loop C contained in D, there exists a two-sided surface S such that C is
the boundary of S and S is contained in D.
Denition 6 Let D be a connected region in E
3
. D is two-connected if for
every closed surface S in D, the interior of S is also contained in D.
We will now dene surface integrals. There are scalar and vector surface inte-
grals.
Denition 7 Let f be a scalar eld dened on a nite surface S. The scalar
surface integral of f on S is denoted by
S
fd and is the limit of a Riemann
sum:
n
fd = lim f(P
i
)
i
,
n
S
max di0 i=1
where S
1
, . . . , S
n
form a subdivision of S intro elementary surfaces, P
is a
i
point in S
i
,
i
is the surface area of S
i
, and d
i
is the diameter of S
i
for
all i n.
Denition 8 A directed surface S is a two-sided surface with all its nonzero
normal vectors pointing away from the same side. The normal vectors are said
to give a direction to S.
Denition 9 Let
) n(P
i
)]
i
,
S
n
max di0 i=1
where S
1
, . . . , S
n
form a subdivision of S intro elementary surfaces, P
is
a point in S
i
,
i
is the surface area of S
i
, d
i
is the diameter of S
i
, and
n(P
i
. To evaluate surface
integrals, the following formula are useful:
fd = f(x(u, v), y(u, v), z(u, v))|w(u, v) dudv
S D
F d
= F w(u, v)dudv
S D
3
Lecture XXIV
Measures
Remember that an elementary region R in E
2
is a region that has a simple,
closed, piecewise smooth curve C as its boundary. A regular region R is a region
that is either regular or can be divided into nitely many regular regions. In the
second case, the boundary of R consists of two or more simple, closed, piecewise
smooth curves, of which one is the outer boundary curve, denoted by C, and the
others are interior boundary curves C
1
, . . . , C
n
. Then R consists of the points
in the interior of C, excluding those in the interior of C
1
, . . . , C
n
.
Similarly, in E
3
a region is regular if its boundary is a simple, closed, piece-
wise smooth surface S. An elementary region in E
3
is either regular or can
be divided into nitely many regular regions. The boundary of an elementary
region that is not regular is formed by one or more closed, but not simple piece-
wise smooth surfaces. In the case that the boundary is formed by two or more
surfaces, one of them is the outer boundary surface S, and the other are interior
boundary surfaces S
1
, . . . , S
n
. Then R consists of the points in the interior of
S, excluding those in the interior of S
1
, . . . , S
n
.
Denition 1 Let D be a region in E
2
or E
3
. Let be a real-valued function
that has as inputs the regular subregions of D. The function is called a nite
measure on D if for every regular subregion R and for every division of R into
regular subregions R
1
and R
2
we have
(R
1
) + (R
2
) = (R).
In E
2
, an example of a nite measure is the function that assigns to every
regular region its area. Similarly, in E
3
, the function that assignes to every
regular region its volume is a nite measure. More generally, in E
2
, if f is a
continuous scalar eld on a region D, for every regular subregion R of D we
dene
f
(R) = fdA.
R
1
Then
f
is a measure on D. Similarly, in E
3
,
f
(R) =
R
fdV is a measure
for every continuous scalar eld f.
Denition 2 Let
F
in the following manner:
1. if R is elementary,
=
C
F d
R, where C is the counterclockwise
F
directed boundary of R.
2. if R is not elementary,
R+
C1
F d
+
Cn
F d
R, where
F C
F d
R+
C is the counterclockwise directed outer boundary of R and C
1
, . . . , C
n
are
the clockwise directed inner boundaries of R.
It is easily veriable that
F.
Denition 3 Let
S
F d
, where S is
F
S
F d +
F
F d + +
Sn
F d, where S is directed outward, and S
1
, . . . , S
n
S1
are directed inward.
It is easy to prove that
f
is a nite measure. It is called the ux measure on
F
D given by
F.
Denition 4 Let D be a region in E
3
. Let
s
be a real-valued function that
has as inputs the nite, piecewise smooth, directed surfaces in D. The function
s
is called a nite surface measure on D if for every nite, piecewise smooth,
directed surface S and for every division of S into nite, piecewise smooth,
directed surfaces S
1
and S
2
we have
s
(S
1
) +
s
(S
2
) =
s
(S).
Denition 5 Let
F
= 0.
2. if S is not closed, let C
1
, . . . , C
n
be the piecewise smooth, simple, closed
boundary curves of S, with directions coherent with the direction of S.
F R +
F R. Then
c
F
=
C1
+
Cn
F
lation measure on D given by
F.
3
Lecture XXV
Greens Theorem
Let us dene a new type of derivative, called rotational derivative, applicable to
vector elds in E
2
. For such a vector eld
F on a domain D in E
2
, let us dene
the rotational derivative at interior points of D. Here, a point P in D is called
an interior point if there exists a circle of center P and radius a > 0 that has
its interior contained in D.
Denition 1 Let
F be a C
1
vector eld on D in E
2
. Let P be an interior point
of D and let C(P, a) be the circle of center P and radius a for all a > 0. The
rotational derivative, denoted by rot
P
, is given by the following formula: F|
rot
F
1
F R. = lim
P
a 0 a
2
C(P,a)
F| For boundary points Q of D, if lim
PQ,P
rot
P
= q exists, we say that rot
Q
F|
F is C
1
, then rot
|
P
exists at every point P on the exists and its value is q. If
F
domain D of
G) = a(rot
G). rot(aF + b
F) + b(rot
F be a vector eld on
D in E
2
for which there exists a scalar eld f and a unit vector w such that
F = f u be the unit vector
F
df
=
P
ds v,P
F j be a vector eld in E
2
with Cartesian coordinates. Using the Let
= L
i +M
i and M
j, we get the
formula for rot
F
M
P
= M
x
(P) L
y
(P) =
P
.
x y P
1
F For example, if F = yi+xj, then rot
= M
x
L
y
= 1(1) = 2. The follow-
ing theorem links integrals on regions in E
2
with integrals on their boundaries,
with the help of rotational derivatives.
Theorem 2 (Greens theorem) Let
F be a vector eld on D in E
2
. Let R
be a regular region in D, with a counterclockwise directed outer boundary C and
clockwise directed interior boundaries C
1
, . . . , C
n
. The following equality holds:
d
d
d
rot
FdA = F R + F R + + F R.
R C C1 Cn
By Greens theorem, if R is a regular region in D, with C its counterclockwise
directed boundary, then
rot
FdA = F R.
R
F
Using this equality, if we let
C R R
2
Lecture XXVI
The Divergence Theorem
In this lecture, we will dene a new type of derivative for vector elds on E
3
,
called divergence. Let
F through S(P, a) is
3
F d. Consider the limit
S(P,a)
lim F d.
a 0 4a
3
S(P,a)
If this limit exists, then it is called the divergence of
F at P, and it is denoted
by div
|
P
. F
Below are two important properties of divergence.
F is C
1
on D, then the divergence of
1. Existence: If
F exists at every
interior point of D.
2. Linearity: If
G) = a(div
G). div(aF + b
F) + b(div
F be a vector eld on
D in E
3
such that there exists a scalar eld on D and a constant unit vector w
F w on D. Such a vector eld
for which
F .
df
=
P
ds w,P
F j +N
Let
= L
i +M
k be a C
1
vector eld on a domain D in E
3
. By using the
linearity of divergence and applying the parallel ow theorem to L
i, M
j, and
N
F = L
x
(P) + M
y
(P) + N
z
(P) =
L
+
M
+
N
x P y P z P
P
Through the following theorem, we can use divergence to compute surface inte-
grals more easily.
Theorem 2 (The divergence theorem) Let
F be a C
1
vector eld on D in
E
3
. Let R be a regular region in D, with outward directed outer boundary surface
S and inward directed inner boundary surfaces S
1
, . . . , S
n
. Then
FdV = F d + F d div
+ + F d.
R
S S1 Sn
Let R be a regular region in D, and let S be its boundary. Then, by the
divergence theorem, we have that
div
FdV = F d.
R
S
2
Lecture XXVII
Stokess Theorem
In the previous lecture, we saw how Greens theorem deals with integrals on
elementary regions in E
2
and their boundaries. We also saw how the divergence
thoerem deals with elementary regions in E
3
and their boundaries. We will now
look at Stokess theorem, which applies to integrals on elementary surfaces in
E
3
and their boundaries.
First let us extend the concept of rotational derivatives to E
3
.
Denition 1 Let
lim F R
a 0 a
2
Ca
If the limit exists, then it is called the rotational derivative of F in the direction
F
|
u,P
.
F and F
If
|
u,P
can
i +M(x, y, z)
k and let
N(x, y, z)
u = u
1
i + u
2
j + u
3
k. Then
F|
rot
u,P
= (N
j
M
z
)u
1
+ (L
z
N
x
)u
2
+ (M
x
L
y
)u
3
.
Denition 2 Let
j k and let P be a
F. The vector
F at P and denoted
F
F = L(x, y, z)
i + M(x, y, z)
+ N(x, y, z)
C
P
, called the curl of
by curl
|
P
, is dened by the formula
F| j + (M
x
L
y
)
curl
P
= (N
y
M
z
)
i + (L
z
N
x
)
k
1
The curl of
F curl
=
i j k
x y z
L M N
The following equality results immediately from the formulas given above.
Theorem 1 (Curl theorem)
F
F| rot
|
u,P
= curl
P
u
From the equality, since u is a unit vector, it follows that the magnitude of
curl
|
P
is the maximum value of rot
u,P
as
F F|
|
P
.
Theorem 2 (Stokess theorem) Let S be a two-sided directed surface whose
boundary consists of the simple, closed curves C
1
, , C
k
. Let
F be a vector
eld whose domain contains S. If we take the direction of each curve C
i
to be
coherent with the direction of S, then
d
d
F curl
d = F R + + F R
S C1 C
k
2
1
Lecture XXVIII
Measures; Irrotational elds
Circulation and ux measures
Let us rst see what the theorems from the past lectures say about the circula-
tion and ux measures. From Greens theorem, we get that, if
F is a C
1
vector
eld on D in E
2
, then
R
F
(R) = rot
FdA
for every regular region R of D, where
F
is the circulation measure given by
F. In E
3
, from the divergence theorem we have that
f
F
(R) = div
FdV,
R
f
F
is the ux measure for every regular region R of D, where D is in E
3
and
given by
F , (S) = curl
d
c
F
S
is the circulation measure given by
F. Remember that in E
2
, if where
c
F
F = L
i + M
j, then
L
rot
=
M
F
x
y
F j + N
In E
3
, if
= L
i + M
k, then
M N
div
=
L
+ + F
x y z
F curl
=
i j k
z
y
x
L M N
1
2 Irrotational elds
Denition 1 A vector eld F on D in E
2
such that rot
F = 0 everywhere is
called an irrotational eld on D.
If
F is an irrotational eld in E
2
and C
1
and C
2
are two simple, closed, piecewise
smooth curves having the same direction, such that C
2
is contained in C
1
, then
by Greens theorem we have
d
d
F R = F R.
C1 C2
Denition 2 A vector eld F on D in E
3
such that div
F = 0 everywhere is
called an divergenceless eld on D.
Let F be a divergenceless eld in E
3
. Let S
1
and S
2
be two directed nite
surfaces with identical boundaries, such that their directions are coherent with
the directions of their boundaries. Then by the divergence theorem,
F d = F d.
S1 S2
Denition 3 A vector eld
F F on D in E
3
such that curl
= 0 everywhere is
called an irrotational eld on D.
Let
F be an irrotational eld in E
3
. Let S be a piecewise C
2
smooth, two-sided
surface S with C and C
d
d
F R = F R.
C C
In E
3
, let us introduce the dierential operator del, denoted by
and dened
by
i +
d
j + k.
dx dy dz
F j + N
If
= L
i + M
k, then
dL dM dN
F = + + = div
dx dy dz
By Stokess theorem, if
F is a C
2
vector eld in E
3
, then
F curl
= 0.
2
Lecture XXIX
Mathematical Applications
1 Leibnitzs Rule
Leibnitzs Rule : Let f(x, t) be a C
1
function dened for a x b. Then
b
d f
f(x, t)dx = f
t
(x, t)dx, where f
t
(x, t) = .
dt t
a a
b
In other words, if we dene g(t) =
A
f(x, t)dx, then
dg
=
b
f
t
(x, t)dx.
dt a
Leibnitzs rule also works for functions that have more parameters. For
example, we have
b
d
f(x, y, z, t)dx = f
t
(x, y, z, t)dx.
dt
a a
The rule can be generalized and applied to gradients. Indeed, let f(x, y, z, t) be
a scalar eld dened for all a t b and (x, y, z) D. Then if P is a point in
D, we have that
P
fdt =
P
fdt.
a A
Let f(x, y, z, t) be a time-dependent mass-density function on a region D. Then,
by Leibnitzs rule, we have
d f
fdV = dV.
dt
D
t
D
This means that the rate of change of the mass( the left-hand side of the equality)
equals the integral of the rate of change of density( the right-hand side of the
equality).
2 Formulas in Spherical Coordinates
The following are the formulas for gradient, divergence, and curl in spherical
coordinates.
1
3
Let g(, , ) be a scalar eld and let G(, , ) = g
1
(, , ) +
+g
2
(, , )
g
3
(, , )
sin
1 (
2
g
1
) 1 (g
2
sin )
G =
G = + +
1 g
3
div
2
sin sin
=
1
2
sin sin
g
1
g
2
g
3
sin
curl
G =
G =
1 (g
3
sin ) 1 g
2
+ =
sin sin
1 g
1
1 (g
3
) 1 (g
2
) 1 g
1
+ +
sin
The Laplacian
The Laplacian is a dierential operator on scalar and vector elds. It is denoted
by
2
. For a scalar eld f, the Laplacian is dened in the following manner:
2
f
2
f
2
f
2
f =
f = +
y
2
+ .
x
2
z
2
F j + N
= L
i + M
2
L)
i + (
2
M)
k.
j + (
2
N)
2
f = 0, called Laplaces equation.
While for scalar elds we have
2
f =
F)
2
F).
Theorem 1 Let
main D. Then
F =
G and
F =
G +
H =
F; H) =
(iii)
G +
G =
F and H) =
G +
(iv) F (
1
Lecture XXX
n-Vectors and Matrices
n-Vectors
We dene E
n
to be the n-dimensional Euclidean space, and R
n
to be the set
of points in E
n
. Hence R
n
is the set of all ordered n-tuples of real numbers
(x
1
, . . . , x
n
). Ordered n-tuples allow repetitions, i.e. x
i
= x
j
is possible for i =
j, but if x
i
= x
j
, then (x
1
, . . . , x
i
, . . . , x
j
, . . . , x
n
) = (x
1
, . . . , x
j
, . . . , x
i
, . . . , x
n
).
These n-tuples can be viewed as vectors in the n-dimensional Euclidean space.
Hence we will call them n-vectors.
Denition 1 Let P = (p
1
, p
2
, . . . , p
n
) and Q(q
1
, q
2
, . . . , q
n
). We say that the
distance between P and Q is d(P, Q) = (p
1
q
1
)
2
+ (p
2
q
2
)
2
+ + (p
n
q
n
)
2
.
We dene
Let
A = (ka
1
, ka
2
, . . . , ka
n
).
2. Vector addition
A +
Let A = (a
1
, a
2
, . . . , a
n
) and B = (b
1
, b
2
, . . . , b
n
). Then
B = (a
1
+
b
1
, a
2
+ b
2
, . . . , a
n
+ b
n
).
3. Dot product
A
Let A = (a
1
, a
2
, . . . , a
n
) and B = (b
1
, b
2
, . . . , b
n
). Then
B = a
1
b
1
+
a
2
b
2
+ + a
n
b
n
.
4. Magnitude of a vector
Let A = (a
1
, a
2
, . . . , a
n
). We say the magnitude of
|
A is |A = A = A
2 2
a + a
2
+ + a
2
.
1 n
1
| |
|
| |
|
5. Unit vectors
A
1
Let
A =
0. Then
A, B are orthogonal if
B = 0.
For example, in E
4
, e
1
= (1, 0, 0, 0), e
2
= (0, 1, 0, 0), e
3
= (0, 0, 1, 0) and e
4
=
(0, 0, 0, 1) play the role of
j, and
i,
k from E
3
. They are mutually orthogonal
and any 4-vector as a linear combination of these vectors.
Denition 3 Let
A
1
, A
2
, . . . , A
n
be mutually orthogonal n-vectors. We say that
they form a frame.
Theorem 1 (Cauchy-Schwartz inequality) For any given n-vectors A and
B, |A B 1. |
Theorem 2 (Triangle inequality) For any given n-vectors A and B, |A +
B| |A + B .
Proof of Theorem 1:
A A2
B B = 22
Since (
B) (
B) 0, we get that
A B+
AB 0,
hence
B) (
B) 0, we get that A
A +
A B 1. Hence A B 1. |
|
Proof of Theorem 2:
From Theorem 1, we have that
A B
A B 1, hence
|A| |B|
1, so A A B|.
B |
Hence
A B +
|
2
= ( A + B
B|
2
= A A + 2
B B |A
2
+ 2 B + |B |A +
| |A||
| |
| |
|)
2
So |A +
B| |A + B .
| A and
Since |A B 1 and
A
B to be cos
1
(
B).
2
2 Matrices
Denition 4 An mn matrix is a rectangular array of m rows and n columns
of numbers.
One-column matrices are called column-vectors and one-row matrices are called
row-vectors. Generally, we write an m n matrix A in the following way:
A =
a
11
a
12
a
1n
a
21
a
22
a
2n
. . .
. . .
. . .
a
m1
a
m2
a
mn
, then A
T
=
1 3 4
3 1 0
3 1 .
4 0
The m n matrix all whose entries are 0 is denoted by 0
m,n
. An n n matrix
is called a square matrix. An n n matrix that has all elements on the main
diagonal (from upper left to lower right) equal to 1 and all other elements equal
to 0 is called an identity matrix, and is denoted by I
n
. Let us dene a new
operation on matrices, multiplication of a matrix with a column-vector :
3
Let A = (a
i,j
) be an m n matrix, and let B = (b
i
) be an n 1 matrix.
Then we dene the m 1 matrix AB as follows:
a
11
b
1
+ a
12
b
2
+ + a
1n
b
n
a
21
b
1
+ a
22
b
2
+ + a
2n
b
n
AB =
a
m1
b
1
+ a
m2
b
2
+ + a
mn
b
n
4
Lecture XXXI
Linear Equation Systems
As we saw in the previous lecture, we can multiply mn matrices by column
n-vectors. Consider the rows of an m n matrix A to be n-vectors:
r
1
c
1
r
1
C
r
2
.
.
.
, C =
c
2
.
.
.
then AC =
r
2
C
.
.
A =
.
C r
m
c
n
r
m
For example, if
7
1 2 1
27
22
, C = , then AC =
9 A =
4 0 3
2
We can use multiplication by a column vector to solve equation systems. An
m n equation system has the form
a
11
x
1
+ a
12
x
2
+ + a
1n
x
n
= d
1
a
21
x
1
+ a
22
x
2
+ + a
2n
x
n
= d
2
a
m1
x
1
+ a
m2
x
2
+ + a
mn
x
n
= d
m
This system has m equations and n unknowns: x
1
, x
2
, , x
n
. An example of a
2 3 system is
4x
1
+ 3x
2
x
3
= 1
x
1
x
2
+ x
3
= 6
This system has innitely many solutions.
An example of a 3 2 system is
x
1
+ x
2
= 4
1
X
2x
1
x
2
= 6
x
1
x
2
= 1
This system has no solutions.
An example of a 2 2 system is
2x
1
x
2
= 4
x
1
+ x
2
= 0
4 4
This system has an unique solution, namely x
1
=
3
, x
2
= .
3
A system that has all the coecients on the right-hand side equal to zero, i.e.
d
1
= d
2
= = d
m
= 0, is called a homogeneous system.
An m n system with the general form given above can by written as a matrix
equality. Let D be an m 1 column-vector whose entries are d
1
, d
2
, , d
m
from top to bottom. Let A = (a
ij
) be the m n matrix those entries are the
coecients on the left-hand side of the system. Let X be an n1 column-vector
whose entries are x
1
, x
2
, , x
n
from top to bottom. Then the left-hand side
of the system is given by the elements of AX, so the system can be written as
AX = D.
x
1
d
1
a
11
a
12
a
1n
=
x
2
.
.
.
, D =
d
2
.
.
.
, A =
a
21
a
22
a
2n
. . .
. . .
. . .
hence ,
x
n
d
m
a
m1
a
m2
a
mn
a
11
x
1
+ a
12
x
2
+ + a
1n
x
n
a
21
x
2
+ a
22
x
2
+ + a
2n
x
n
d
1
AX =
d
2
.
.
.
= D
a
m1
x
1
+ a
m2
x
2
+ + a
mn
x
n d
m
We dene three elementary operations on the equations of a system that do not
modify its solution-set:
() multiplying an equation by a non-zero scalar;
() adding to an equation some multiple of a dierent equation;
2
() interchanging two equations.
By using these operations through the method of row-reduction, we can simplify
the equation system without altering its solution-set. In the next lecture will
dene row-reduction, which is also known as Gauss-Jordan reduction.
3
1
Lecture XXXII
Row Reduction; Determinants
Row Reduction
Recall the 3 elementary operations we will use to solve the system of equations
AX = D:
() multiplying an equation by a non-zero scalar;
() adding to an equation some multiple of a dierent equation;
() interchanging two equations.
We call the matrix [A : D] the augmented matrix of the system AX = D. We
will use , , and to bring B = [A : D], or, in the case of homogeneous
systems, B = A to a row-reduced form.
Denition 1 The row-reduced form of a matrix B is a matrix obtained from
B by applying the elementary operations that has the following properties:
1. Every non-zero row has 1 as its rst non-zero element from left to right.
This element is called the pivot of the row.
2. Each pivot occurs in a column to the right of all pivots from upper rows.
3. All zero rows are at the bottom of the matrix.
4. For any pivot, all other elements in its column are equal to zero.
We obtain this form through the method of row-reduction, which is described
below:
1. Choose the rst non-zero column from left to right, then choose a row
having a non-zero element on this column.
1
1 0 7 0
0 1 3 2
0 0 0 0
Then we can write the equations of the system:
x
1
7x
3
= 0, hence x
1
= 7x
3
x
2
+ 3x
3
= 2, 2 3x
3
hence x
2
=
x
1
0 7
+ t
x
2
x
3
0 1
B, where Hence X = A + t
0 7
A =
and B =
0 1
In this case the solution set forms a line in 3-dimensional space. Generally, for
systems with 3 unknowns, the solution set is given by X = C + t
1
B
1
+ t
2
B
2
,
where C,
B
1
, B
2
are 3-vectors, possibly equal to
0.
2
2 Determinants
Consider a square n n matrix A. The determinant of the matrix is a scalar
depending on A, denoted by A . For a 1 1 matrix, the determinant of the | |
matrix is simply the value of its element. Using the notion of minors, we can
dene a reccurent formula for determinants.
Denition 2 The minor of an element a
ij
of a square matrix is the determi-
nant of the matrix formed from the initial matrix excluding the row i and the
column j containing this element.
We denote by m
ij
the minor of a
ij
. Consider the nn matrix having as elements
the signs + and , in a checkerboard pattern:
+ +
+
+ +
We call the cofactor of an element a
ij
, the minor of a
ij
multiplied by +1 or 1
, depending on the position of the element in the above matrix. The cofactor
of a
ij
has the value (1)
i+j
m
ij
. For any column or row of A, the determinant
of A is given by the formula
A| = (1)
i+j
a
ij
m
ij
, |
where the sum is on the column or row, i.e. j is constant and i goes from 1 to
m, or i is constant and j goes from 1 to n.
3
1
Lecture XXXIII
Determinants; Matrix Algebra
Determinants
For a square matrix A, the determinant of A has the following properties:
1. Interchanging two rows of the matrix multiplies the value of the determi-
nant by -1.
2. If there exists two identical rows, then the value of the determinant is 0.
3. If we multiply a row by a scalar c, then the value of the determinant is
also multiplied by c.
4. Applying the operation to the matrix leaves the value of the determinant
unchanged.
5. If the determinant is 0, then there exists a row that can be written as a
linear combination of the other rows.
6. Properties (1) through (5) also hold for columns instead of rows.
In addition to the Laplace expansion method of nding determinants, there is
a second method that is faster in most cases. This is a simplied row-reduction
that brings the matrix in to a form called row-echelon matrix. In this method
we use operations and . The row-echelon matrix is dierent from the row-
reduced matrix in that:
1. pivots can have other values than 1, since we do not use operations;
2. in the column of a pivot, we only need to get 0s below the pivot.
1
|
2
If each column has a pivot, then the determinant of the inital matrix A is equal
to the product of the pivots times (1)
k
, where k is the number of operations
we applied to obtain the row-echelon form. If a column has no pivots, then there
will be a row containing only 0s, so the determinant will be 0. Let us dene
the rank of A as the number of pivots in the row-echelon form. The following
sentences hold:
1. If |A = 0, then rank(A) < n. In this case we say that A is a singular |
matrix.
2. If |A = 0, then rank(A) = n. In this case we say that A is a non-singular
matrix.
Matrix Algebra
We saw in the previous lecture that we can multiply matrices with column
vectors. Let us now dene in a larger sense matrix multiplication. In order to
dene the product AB of matrices A and B, A must have dimensions m n
and B must have dimensions n p. In other words, he number of columns
in A must equal the number of rows in B. Then multiplying A with the i-th
column of B, we obtain a column m-vector. This column-vector will be the i-th
column of AB. Hence AB is an m p matrix. It can be easily observed that
BA exists as well only if m = p. In this case AB is an m m matrix and BA
is an n n matrix. So even if AB and BA both exist, they may not have the
same size. Furthermore, even if AB and BA have the same size, it is possible
that AB = BA. Hence the multiplication of matrices is non-commutative. Let
A be an m n matrix, let B be an n p matrix, and let C be a p l matrix.
Then (AB)C and A(BC) both exist and furthermore
(AB)C = A(BC)
Hence, the multiplication of matrices is associative. The following properties
also hold:
1. AI
n
= A and I
n
A = A, if the products exists, where I
n
is the identity
n n matrix we dened in the previous lecture.
2
)
t
, and
1 1
multiply it by
detA
to obtain the inverse of A:
det A
(A
)
t
. The folowing theorem
is an important determinant property.
Theorem 1 Let A and B be two square matrices of the same size. Then
AB|= . | |A||B|
Denition 1 A set V is a vector subspace of R
n
if it satises the following
properties:
(i) 0 is in V .
x, x + (ii) For any y in V , y is in V .
x in V and any scalar c in R, c (iii) For any x is in V .
Clearly, the set containing only the origin is a vector subspace. In R
3
, a vector
subspace must be either the set previously mentioned, a line through the origin,
a plane through the origin, or all of R
3
.
Denition 2 Let X be a vector in R
n
and let V be a vector subspace of R
n
.
Let us denote by A = X +V the set of all the vectors X +Y , with Y in V . We
say that A is an ane subspace of R
n
.
Let A be a given ane subspace that is not formed by a single point only. Let
X be a vector and let V be a vector space such that A = X + V . Then the
choice of V is unique, where X can be any vector in A.
1
Denition 3 Let c
1
, , c
n
be scalars and let Y
1
, , Y
n
be contained in a vector
subspace V . The vector c
1
Y
1
+ c
2
Y
2
+ +c
n
Y
n
is called a linear combination
of vectors Y
1
, , Y
n
.
Theorem 2 Any linear combination of vectors from a vector subspace V is
contained in V .
Denition 4 Let
n
be contained in a vector subspace V . We call the Y
1
, , Y
span of Y
1
, , Y
n
and denote by span(Y
1
, , Y
n
) the set of all linear combina-
tions of Y
1
, , Y
n
.
For any vectors Y
1
, , Y
n
in V , span(Y
1
, , Y
n
) is a vector subspace contained
in V .
Denition 5 Vectors Y
1
, , Y
k
are linearly independent if a linear combina-
tion c
1
Y
1
+ + c
k
Y
k
is equal to 0 if and only if c
1
= = c
k
= 0.
Denition 6 If Y
1
, , Y
k
V are linearly independent and the span of Y
1
, , Y
k
is V , then the set Y
1
, , Y
k
is called a basis for V . Any point in V can be ex-
pressed as a linear combination of Y
1
, , Y
k
in a unique way.
2