You are on page 1of 11

Topics from Tensoral Calculus

Jay R. Walton
Fall 2006
1 Preliminaries
These notes are concerned with topics from tensoral calculus, i.e. generalizations of calculus
to functions dened between two tensor spaces. To make the discussion more concrete, the
tensor spaces are dened over ordinary Euclidean space, R
N
, with its usual inner product
structure. Thus, the tensor spaces inherit a natural inner product, the tensor dot-product,
from the underlying vector space, R
N
.
2 The Derivative of Tensor Valued Functions
Let
F : T T
r
T
s
(1)
be a function with domain T a subset of T
r
taking values in the tensor space T
s
, where T
is assumed to be an open subset of T
r
.
1
Continuity. The function F is said to be Continuous at A T provided for every > 0
there exists a > 0 such F(B) B(F(A), ) whenever B B(A, ), i.e. F maps the ball
of radius centered at A into the ball of radius centered at F(A). The function F is said
to be continuous on all of T provided it is continuous at each A T. There are two useful
alternative characterizations of continuity. The rst is that F is continuous on T provided
it maps convergent sequences onto convergent sequences. That is, if A
n
T is a sequence
converging to A T (lim
n
A
n
= A), then lim
n
F(A
n
) = F(A). The second alternative
characterization is that the inverse function F
1
maps open subsets of T
s
to open subsets
of T.
Derivative. The function F is said to be Dierentiable at A T provided there exists a
tensor L T
r+s
such that
F(A + H) = F(A) + L[H] + o(H) as [H[ 0 (2)

Copyright c _ 2006 by J. R. Walton. All rights reserved.


1
A subset T T
r
is said to be open provided for every element A T there exists an open ball centered
at A that is wholly contained in T.
1
where [H[ denotes the norm of the tensor H T
r
. If such a tensor L exists satisfying (2),
it is called the Derivative of F at A and denoted DF(A). Thus, (2) can be rewritten
F(A + H) = F(A) + DF(A)[H] + o(H). (3)
Recall that o([H[) is the Landau little oh symbol which is used to denote a function
depending upon H that tends to zero faster that [H[, i.e.
lim
|H|0
o(H)
[H[
= 0.
If the derivative DF(A) exists at each point in T, then it denes a function
DF() : T T
r
T
r+s
.
Moreover, if the function DF() is dierentiable at A T
r
, then its derivative is a tensor in
T
2r+s
, denoted by D
2
F(A), called the second derivative of F at A T. Continuing in this
manner, derivatives of F of all orders can be dened.
Example. Let
() : T
1

= R
N
T
0

= R. (4)
Thus, () is a real-valued function of N-real variables and its graph is a surface in R
N+1
.
In the denition of the derivative (3), it is more customary in this context to let H = hu
where u is a unit vector in R
N
. Dening equation (3) then becomes
(a + hu) = (a) + D(a)[hu] + o(hu). (5)
From the linearity of the tensor D(a)[hu], one concludes that
D(a)[u] = lim
h0
(a + hu) (a)
h
, (6)
which is the familiar directional derivative of () at the point a in the direction u. Thus,
being dierentiable at a point implies the existence of directional derivatives, and hence
partial derivatives, in all directions. However, the converse is not true. That is, there exist
functions with directional derivatives existing at a point in all possible directions but which
are not dierentiable at the point. For such an example, consider the function () : R
2

R given by
(x, y) =

x
3
y
3
x
2
+y
2
when (x, y) ,= (0, 0),
0 when (x, y) = (0, 0).
(7)
One then shows easily that if u = (cos(), sin())
T
, the directional derivative of () at the
origin (0,0) in the direction u equals cos()
3
sin()
3
. However, () is not dierentiable at
the origin in the sense of (5). (Why?)
Consider further the function () in (4). If () is dierentiable in the sense of (5),
then its derivative D(a)[] T
1
is a linear transformation from R
N
to R, and as such is
representable by dot-product with a vector in R
N
. Specically, there exists a unique vector,
denoted by (a) R
N
such that
D(a)[u] = (a) u for all u R
N
. (8)
2
The vector (a) is called the Gradient of () at a.
The component forms for the derivative D and the gradient are easily constructed.
In particular, let B = e
1
, . . . , e
N
be the natural orthonormal basis for R
N
. Then the 1N
matrix representation for D and the N-tuple vector representation for the gradient are
given by
[D(a)]
B
= [
x
1
(a), . . . ,
x
N
(a)] and [(a)]
B
=

x
1
(a)
.
.
.

x
N
(a)

(9)
where
x
i
(a) denotes the partial derivative of with respect to x
i
at the point a

x
i
(a) = lim
h0
(a + he
i
) (a)
h
.
Example. Another important example is provided by functions F() : T
0
T
s
, i.e.
s-ordered tensor valued functions of a single scalar variable. Since in continuum mechanics
the scalar independent variable is usually time, that variable is given the special symbol t
and the derivative of such functions is represented by
DF(t)[] =

F(t) T
s
(10)
where

F(t) = lim
h0
F(t + h) F(t)
h
.
In component form, if the tensor valued function F() has the component representation
[F
i
1
,...,is
] with respect to the natural basis for T
s
, then the component representation for the
tensor

F(t) is
[

F(t)] = [

F
i
1
,...,is
]. (11)
Example. A Vector Field is a function a() : T T
1

= R
N
T
1
. Its derivative denes
a second order tensor Da(x)[] T
2
. Its component form, with respect to the natural basis
on R
N
for example, is
[Da(x)] = [
x
j
a
i
(x)] i, j = 1, . . . , N (12)
where [a
i
(x)], i = 1, . . . , N gives the component representation of a(x). The right hand side
of (??) is the familiar Jacobian matrix.
Product Rule. Various types of products of tensor functions occur naturally in tensor
calculus. Rather than proving a separate product rule formula for every product that arises,
it is much more expedient and much cleaner to prove one product rule formula for a general,
abstract notion of product. To that end, the appropriate general notion of product is provided
by a general bi-linear form. More specically, suppose that F() : T T
r
T
p
and
G() : T T
r
T
q
are two dierentiable functions with the same domain set T in T
r
but dierent range spaces. Let (, ) : T
p
T
q
T
s
denote a bi-linear function (i.e.
(, ) is linear in each of its variables separately) with values in T
s
. One then denes the
product function E() : T T
r
T
s
by
E(A) := (F(A), G(A)) for A T.
3
Since F and G are assumed to be dierentiable at A T, it is not dicult to show that E
is also dierentiable at A with
DE(A)[H] = (DF(A)[H], G(A)) + (F(A), DG(A)[H]) for all H T
r
. (13)
Notice that (13) has the familiar form (fg)

= f

g + fg

from single variable calculus.


Example. Let A() : T
0

= R T
r
and B() : T
0
T
s
be dierentiable tensor valued
functions of the single scalar variable t. Then their tensor product E(t) := A(t) B(t) is
dierentiable with

E(t) =

A(t) B(t) + A(t)

B(t).
Chain Rule The familiar chain rule from single variable calculus has a straight forward
generalization to the tensor setting. Specically, suppose F() : T T
r
T
q
is dier-
entiable at A T and G() : ( T
q
T
s
(with ( being an open set on which G() is
dened) is dierentiable at F(A) ( F(T), then the composite function E() := G F()
is also dierentiable at A T with
DE(A)[H] = DG(F(A))[DF(A)[H]] for all H T
r
. (14)
The right hand side of (14) is the composition of the tensor DG(F(A))[] T
q+s
with
the tensor DF(A)[] T
r+q
producing a the tensor DG(F(A)) DF(A)[] T
r+s
. This
generalizes the familiar chain rule formula g(f(x))

= g

(f(x))f

(x) from single variable


calculus.
Example. An important application of the chain rule is to composite functions of the form
E(t) = G F(t) = G(F(t)), i.e. functions for which the inner function is a function of the
single scalar variable t. The chain rule then yields the result

E(t) = DG(F(t))[

F(t)].
For example, let A(t) be a dierentiable function taking values in T
2
, i.e. A() : R
T
2
. Then the composite real valued function, (t) = det(A(t)) is dierentiable provided
det(A(t)) ,= 0 with

(t) = det(A(t))tr(

A(t)A(t)
1
).
3 Div, Grad, Curl
Classical vector calculus concerns special cases of (1) with r = 0, 1 and s = 0, 1, 2. It is in that
setting that the operators gradient, divergence and curl are usually dened. In elementary
calculus, these operators are most often dened through component representations with
respect to the natural orthonormal basis for R
N
. Here they are given intrinsic denitions
irrespective of any chosen basis for R
N
.
Gradient. The denition of the gradient of a scalar valued function dened on T
1

= R
N
has been given previously and wont be repeated here. It is also customary to dene the
gradient of a vector eld a() : T
1
T
1
as
Grad(a(x)) = a(x) := Da(x).
4
Thus, for vector elds on R
N
, the gradient is just another name for the previously dened
derivative.
Divergence. For a vector eld, a(x), dened on T R
N
, the Divergence is dened by
Div(a(x)) := tr(a(x)). (15)
Thus, the divergence of a vector eld is a scalar valued function. The reader should verify
that the component form of Div(a(x)) with respect to the natural basis on R
N
is
Div(a(x)) =
x
1
a
1
(x) + . . . +
x
N
a
N
= a
i,i
(x).
It is also useful to dene the divergence for second order tensor valued functions dened
on T R
N
. To that end, if A() : T R
N
T
2
, then its divergence, denotes Div(A(x)),
is that unique vector eld satisfying
Div(A(x)) a = Div(A(x)
T
a) (16)
for all constant vectors a R
N
. In component form, if [A(x)] = [a
ij
(x)], then
[Div(A(x))] = [a
ij,j
(x)]
T
.
Curl. The operator Curl is dened for vector elds on R
3
. To give it an implicit, component
independent denition, one proceeds as follows. For a vector eld v() : T R
3
R
3
,
Curl(v(x)) is dened to be the unique vector eld satisfying
(v(x) (v(x))
T
)a = Curl(v(x)) a (17)
for all constant vectors a R
3
. The right hand side of (17) is the vector cross product of the
two vectors Curl(v(x)) and a. An important observation from (17) is that Curl(v(x)) = 0
if and only if v(x) is a symmetric second order tensor.
Useful Formulas. In the following formulas, (x) denotes a scalar valued eld, u(x) and
v(x) denote vector elds and A(x) denotes a second order tensor valued eld. One can then
readily show from the general product rule (13) that
(v) = v +v
Div(v) = Div(v) +v
(u v) = (v)
T
u + (u)
T
v
Div(u v) = uDiv(v) + (u)v (18)
Div(A
T
v) = A v +v Div(A)
Div(A) = Div(A) + A.
Another useful formula shows that the operators Div and do not commute. Specically,
the reader should verify that
(Div(v)) = Div((v)
T
). (19)
5
4 Integral Theorems
In this section are cataloged a selection of results concerning integrals of various dierential
operators acting on tensor elds. They all may be thought of as multidimensional general-
izations of the Fundamental Theorem of Single Variable Calculus

b
a
f(x) dx = F(b) F(a)
where F(x) is any anti-derivative of f(x), i.e. F

(x) = f(x). All of the theorems presented


have versions valid in R
N
. However, attention is restricted here to R
3
to keep the mathemat-
ical complications to a minimum. The reader is assumed to be familiar with the denition
of line, surface and volume integral in R
3
as well as with techniques for actually computing
them. In what follows, 1 R
3
is assumed to be a bounded open set with piecewise smooth
boundary 1, and n denotes the piecewise smooth eld on 1 of outward pointing unit
normal vectors. Moreover, the natural basis for R
3
is denoted by e
1
, e
2
, e
3
. Also, volume
integrals over a region 1 R
3
are denoted

R
f dV
with dV denoting volume measure, and surface integrals over a surface o in R
3
are denoted

S
f dA
with dA denoting surface area measure.
4.1 Divergence Theorem
The Divergence Theorem has four fundamental versions in R
3
given in
Divergence Theorem. Let () : 1 R, v() : 1 T
1

= R
3
and S() : 1 T
2
denote smooth scalar, vector and 2
nd
-order tensor elds on 1. Then

R
dV =

R
ndA (20)

R
Div(v) dV =

R
v ndA (21)

R
v dV =

R
v ndA (22)

R
Div(S) dV =

R
SndA. (23)
Proof: The reader has undoubtedly seen (20) and (21) proved in a vector calculus text.
(22) and (23) are simple consequences of (20) and (21). In particular, for (23), let a be a
6
constant vector in R
3
. Then,
a

R
SndA =

R
a (Sn) dA
=

R
(S
T
a) ndA
=

R
Div(S
T
a) dV
=

R
(Div(S)) a dV = a

R
Div(S) dV.
Since the vector a is arbitrary, (23) now follows easily. (22) can be proved in similar fashion.
The following important application of the Divergence Theorem provides a useful inter-
pretation of the divergence operator. Specically, one can easily show
Theorem. Let B[a, r] 1 denote the ball centered at a 1 of radius r > 0. Then
Div(v(a)) = lim
r0+
1
vol(B[a, r])

B[a,r]
v ndA (24)
Div(S(a)) = lim
r0+
1
vol(B[a, r])

B[a,r]
SndA. (25)
Proof: Supplied by the reader.
Thus, one sees from (25) that the Div(v(a)) equals the (outward) ux per unit volume of
v through an innitesimally small sphere centered at a. Whereas, if the second order tensor
S is thought of as a stress tensor, for example, one sees from (25) that Div(S(a)) gives the
total contact force per unit volume acting on the boundary of an innitesimally small ball
centered at a.
4.2 Green Identities
The Green Identities are important applications of the Divergence Theorem. They are proved
from the following multidimensional generalization of the single variable calculus integration
by parts formula

b
a
f(x)g

(x)dx = f(b)g(b) f(a)g(a)

b
a
f

(x)g(x)dx.
If is a smooth scalar eld and v is a smooth vector eld dened on 1 R
3
, then it follows
from the Divergence Theorem and the product rule formula (18b) that

R
(Div(v) +v ) dV =

R
Div(v) dV =

R
v ndA.
Letting v = , where is another smooth scalar eld dened on 1, one obtains the
First Green Identity. Let and denote smooth scalar elds dened on the region
1 R
3
. Then

R
dV =

d
dn
dA

R
dV (26)
7
where
d
dn
:= n
denotes the (outward) normal derivative of on the boundary 1 of 1.
From the First Green Identity one easily derives the Second Green Identity. Again
let and denote smooth scalar elds dened on 1 R
3
. Then

R
( ) dV =

R
(
d
dn

d
dn
) dA. (27)
Application to Poisson Equation. The Green Identities are work-horse tools in applied
mathematics. As an illustration, consider the classical mixed boundary value problem for
the Poisson equation. Specically, let 1 R
3
be a bounded open region with piecewise
smooth boundary 1. Suppose further that the boundary is the union of two surfaces with
non-empty (2-dimensional) interiors and with a piecewise smooth common boundary curve
. More specically, 1 = o
1
o
2
with

o
1

o
2
= , o
1
o
2
= and

o
1
,

o
2
,= where ,
the common boundary to the surfaces o
1
and o
2
is a piecewise smooth closed curve.
Mixed Boundary Value Problem. The classical mixed boundary value problem for the
Poisson equation requires nding a smooth function u(x) satisfying the Poisson equation
u = f, for x 1 (28)
subject to the mixed boundary conditions
u = g for x o
1
(29)
du
dn
= h for x o
2
. (30)
Boundary condition (29) is called a Dirichlet Boundary Condition while (30) is called a
Neumann Boundary Condition.
Data Compatibility. Letting (x) 1 and (x) = u(x) in (26) yields

R
f dV =

R
u dV =

R
du
dn
. (31)
In particular, for the pure Neumann problem for which o
1
= , one has the equilibrium
requirement on the data h and f

R
hdA +

R
f dV = 0.
Uniqueness of Smooth Solutions. Suppose there are two smooth solutions, u
1
(x) and
u
2
(x), of (28,29,30). Dene u(x) := u
1
(x) u
2
(x). Then u(x) satises
u = 0, for x 1 (32)
subject to the mixed boundary conditions
u = 0 for x o
1
and
du
dn
= 0 for x o
2
. (33)
8
Multiplying (32) by u(x), integrating the resulting equation over 1 and then integrating by
parts (i.e. using the First Green Identity (26)) making use of (33), one obtains the identity

R
[u[
2
dV = 0. (34)
It follows from (34) that [u(x)[ 0 on 1, and hence that u(x) is a constant function. Since
the boundary conditions require u(x) 0 on o
1
, u(x) 0 on all of 1. Thus, u
1
(x) = u
2
(x)
on 1 as was required to be shown.
4.3 Potential Theory
This section presents a few results from classical potential theory. Specically, the issue
studied is conservative vector elds on R
N
.
Conservative Field. A vector eld v(x) dened on a region 1 R
N
is called Conservative
provided that work is path independent. That is, given any two piecewise smooth paths

1
[a, b],
2
(a, b] 1 connecting the two points a, b 1, one has

1
[a,b]
v dx =

2
[a,b]
v dx (35)
where

[a,b]
v x denotes the line integral of v(x) along the path [a, b]. Alternatively,
the eld v(x) is conservative provided its line integral around any piecewise smooth, closed
curve in 1 is zero.
The main result of the subject asserts that a vector eld v(x) is conservative if and only
if it is the gradient of some scalar eld (x). If such a scalar eld exists, it (or possibly
its negation in some applications) is called a Potential for the eld v(x). If the region 1 is
connected, i.e. if every two points a, b 1 can be connected by a path [a, b] 1, then
any two potential functions dier by at most an additive constant. More formally, this result
is called the
Potential Theorem. A vector eld v(x) dened on a connected region 1 R
N
is conser-
vative if and only if there exists a scalar eld (x) such that v(x) = (x). Any two such
scalar elds dier by a constant.
Proof: Suppose that v(x) = (x) and let [a, b] be a path in 1 joining the two points a
and b. Suppose c(r), 0 r 1 is a parametric representation of [a, b] with c(0) = a and
c(1) = b. Then

[a,b]
v(x) dx =

1
0
v(c(r)) c

(r) dr
=

1
0
(c(r)) c

(r) dr
=

1
0
d
dr
(c(r)) dr
= (c(1)) (c(0)) = (b) (a).
9
Thus, the line integral depends only upon a and b, not the particular path joining them. It
follows that v(x) is a conservative vector eld on 1. Conversely, assume v(x) is conservative.
Let a 1 be xed. Then for any x 1, dene (x; a) by
(x; a) :=

[a,x]
v(z) dz (36)
where [a, x] is any path in 1joining a to x. By path independence, (x; a) is unambiguously
dened. To show that v(x) = (x; a), it suces to demonstrate that
d
de
(x; a) =
d
ds
(x + s e; a)[
s=0
= v(x) e (37)
where e is any unit vector. Since the left most term in (37) is the directional derivative of
(x; a) at x in the direction e, which must also be given by
d
de
(x; a) = (x; a) e,
one concludes that (x, a) = v(x). To verify (37), let > 0 be chosen small enough so
that x + s e 1 for all [s[ < . Let [a, x e] denote any path in 1 joining a and x e
and let [x e, x + s e] denote the straight line path connecting x e to x + s e. Then
(x + s e; a) =

[a,xe]
v(z) dz +

[xe,x+s e]
v(z) dz. (38)
The rst integral on the right-hand-side of (38) is constant with respect to s whereas for the
second,

[xe,x+s e]
v(z) dz =

v(x + r e) e dr.
It now follows that
d
ds
(x + s e; a)[
s=0
= v(x + s e) e[
s=0
= v(x) e
as required. Finally, let
1
(x) and
2
(x) be two scalar elds whose gradients equal v(x).
Dene (x) :=
1
(x)
2
(x). Then (x) 0 on 1. Since 1 is assumed to be connected,
a standard theorem in calculus shows that (x) is identically constant on 1.
This result is somewhat unsatisfying as a test for whether or not a given vector eld v(x) is
conservative since it requires either showing that a suitable scalar potential exists or showing
that there is a closed curve on which the line integral of v(x) does not vanish. However, for
simply connected regions 1 R
3
, there is a convenient test for conservative elds involving
the Curl operator. More specically, a region 1 R
3
is called Simply Connected provided
every closed path is Homotopic to a Point. Roughly speaking, homotopic to a point means
that a closed path can be continuously deformed to a point without leaving the region 1.
To be more precise, a closed curve c(s), 0 s 1 is said to be homotopic to a point if
there exists a continuous function h(s, r) dened for 0 s, r 1 sasfying h(s, 0) = c(s)
for 0 s 1, h(0, r) = h(1, r) for all 0 r 1, h(s
1
, 1) = h(s
2
, 1) for all 0 s
1
, s
2
1.
10
The function h(, ), which is called a homotopy, is a one parameter family of closed paths,
h(, r) 0 r 1, that continuously deforms the given closed curve, c(s), to the one point
(constant) curve h(s, 1). The reader should show that the region between two co-axial
cylinders is not simply connected in R
3
whereas the region between to concentric spheres is.
A simple test for conservative vector elds on simply connected regions in R
3
is given by
the following
Theorem. Let 1 R
3
be simply connected. Then a smooth vector eld v(x) on 1 is
conservative if and only if Curl(v(x)) 0 on 1.
Proof: The theorem can be proved with the aid of the classical
Stokes Theorem. Let v(x) be a smooth vector eld on a region 1 R
3
and let o 1 be
a smooth, orientable surface whose boundary o = is a a piecewise smooth, simple closed
curve (no self-intersections). Then

R
Curl(v) ndA =

v(x) dx. (39)


where n(x) is a continuous normal vector eld on o and the direction for the line integral
on the right-hand-side of (39) is chosen by a parametric representation c(s), 0 s 1 for
satisfying ( c(0) c(s)) n(c(0)) > 0, 0 < s < for some 0 < . (Proof: Omitted since
the reader is assumed to have seen the proof in a previous calculus course.)
The main theorem is proved by showing that the line integral of v(x) along any simple closed
curve in 1 vanishes. To that end, let c(s), 0 s 1 be a parametric representation for a
smooth, simple closed curve in 1. Since 1 is assumed to be simply connected, there exists
a homotopy h(s, r), 0 s, r 1 smoothly deforming to a point in 1. The homotopy
h(s, r), 0 s, r 1 may be thought of as giving a parametric representation for a smooth
orientable surface o 1 having as boundary and with a smooth normal vector eld
dened by n(x) :=
m(x)
|m(x)|
where
m(x) :=
s
h(s, r)
r
h(s, r).
Applying Stokes Theorem one concludes that

v(x) dx =

S
Curl(v(x)) n(x) dA = 0.
The integral over the surface o vanishes because Curl(v(x)) is assumed to be identically zero
in 1.
Conversely, if v(x) is conservative, then v(x) = (x) for some scalar eld (x).
It follows that v(x) = (x) is a symmetric second order tensor, and hence that
Curl(v(x)) 0 on 1. This completes the proof of the theorem.
11

You might also like