Discover millions of ebooks, audiobooks, and so much more with a free trial

Only $11.99/month after trial. Cancel anytime.

Introduction to Matrices and Linear Transformations: Third Edition
Introduction to Matrices and Linear Transformations: Third Edition
Introduction to Matrices and Linear Transformations: Third Edition
Ebook734 pages10 hours

Introduction to Matrices and Linear Transformations: Third Edition

Rating: 3 out of 5 stars

3/5

()

Read preview

About this ebook


This versatile undergraduate text can be used in a variety of courses in linear algebra. It contains enough material for a two-semester course, and it also serves as a support text and reference. Chapter Ten, on linear programming, will be of special interest to students of business and economics. A balanced combination of formal theory and related computational techniques, this treatment begins with the familiar problem of solving a system of linear equations. Subsequent chapters explore linear spaces and mappings, matrices, determinants, inner product spaces, scalar-valued functions, and linear differential equations. The author introduces metric notions of Euclidean space at an early stage and employs the computational technique of Gaussian elimination throughout the book. Solutions to selected exercises appear at the end.
LanguageEnglish
Release dateOct 21, 2013
ISBN9780486279664
Introduction to Matrices and Linear Transformations: Third Edition

Related to Introduction to Matrices and Linear Transformations

Titles in the series (100)

View More

Related ebooks

Mathematics For You

View More

Related articles

Reviews for Introduction to Matrices and Linear Transformations

Rating: 3 out of 5 stars
3/5

1 rating0 reviews

What did you think?

Tap to rate

Review must be at least 10 words

    Book preview

    Introduction to Matrices and Linear Transformations - Daniel T. Finkbeiner

    TRANSFORMATIONS

    CHAPTER 1

    LINEAR

    EQUATIONS

    Linear algebra is concerned primarily with mathematical systems of a particular type (called vector spaces), functions of a particular type (called linear mappings, so at least superficially the study of linear algebra appears to be a natural extension and generalization of your previous studies. But you should be forewarned that the degree of generalization is substantial and the methods of linear algebra are significantly different from those of calculus.

    A glance at the Table of Contents will reveal many terms and topics that might be unfamiliar to you at this stage in your mathematical development. Therefore, as you study this material you will need to pay close attention to the definitions and theorems, assimilating each idea as it arises, gradually building your mathematical vocabulary and your ability to utilize new concepts and techniques. You are urged to make a practice of reading all the exercises and noting the results they contain, whether or not you solve them in detail.

    The contents of this book are a blend of formal theory and computational techniques related to that theory. We begin with the problem, familiar from secondary school algebra, of solving a system of linear equations, thereby introducing the idea of a vector space informally. Vector spaces are not defined formally until Section 3 of Chapter 2. At that point, and from time to time thereafter, you are urged to study Appendix A.1, where algebraic systems are explained briefly but generally. You might not need that much generality to understand the concept of a vector space, but firm familiarity with the notion of an algebraic system will greatly accelerate your ability to feel comfortable with the ideas of linear algebra.

    Individuals acquire mathematical sophistication and maturity at different rates, and you should not expect to achieve instant success in assimilating some of the more subtle concepts of this course. With patience, persistence, and plenty of practice with specific examples and exercises, you can anticipate steady progress in developing your capacity for abstract thought and careful reasoning. Moreover, you will greatly enhance your insight into the nature of mathematics and your appreciation of its power and beauty.

    1.1 SYSTEMS OF LINEAR EQUATIONS

    The central focus of this book is the concept of linearity. Persons who have studied mathematics through a first course in calculus already are familiar with examples of linearity in elementary algebra, coordinate geometry, and calculus, but they probably are not yet aware of the extent to which linear methods pervade mathematical theory and application. Such awareness will develop gradually throughout this book as we explore the properties and significance of linearity in various mathematical settings.

    We begin with the familiar example of a line L in the real coordinate plane, which can be described algebraically by a linear equation in two variables:

    A point (x0, y0) of the plane lies on the line L if and only if the real number ax0 + by0 has the value d. The formal expression

    is called a linear combination of x and y.

    By analogy a linear combination of three variables has the form

    where a, b, and c are constants. Any equation of the form

    is called a linear equation in three variables. If you have studied the geometry of three-dimensional space, you will recall that the graph of a linear equation in three variables is a plane, rather than a line. This is a significant observation: the word linear refers to the algebraic form of an equation rather than to the geometric object that is its graph. The two meanings coincide only for the case of two variables—that is, for the coordinate plane. In general, a linear equation in n variables has the form

    where at least one ci ≠ 0. For n > 3 the graph of this equation in n-dimensional space is called a hyperplane.

    Applications of mathematics to science and social science frequently lead to the need to solve a system of several linear equations in several variables, the coefficients being real numbers:

    The number m of equations might be less than, equal to, or greater than the number n of variables. A solution of the System 1.1 is an ordered n-tuple (c1,..., cn) of real numbers having the property that the substitution

    simultaneously satisfies each of the m equations of the system. The solution of (1.1) is the set of all solutions, and to solve the system means to describe the set of all solutions. As we shall see, this set can be finite or infinite.

    This problem is considered in algebra courses in secondary school for the case m = 2 = n, and sometimes for other small values of m and n. But a large scale linear model in contemporary economics might require the solution of a system of perhaps 83 equations in 115 unknowns. Hence we need to find very efficient procedures for solving (1.1), regardless of the values of m and n, in a finite number of computational steps. Any fixed set of instructions that is guaranteed to solve a particular type of problem in a finite number of steps is called an algorithm. Many algorithms exist for solving systems of linear equations, but one of the oldest methods, introduced by Gauss, is also one of the most efficient. Gaussian elimination, and various algorithms related to it, operate on the principle of exchanging the given system (1.1) for another system (1.1 A) that has precisely the same set of solutions but one that is easier to solve. Then (1.1 A) is exchanged for still another system (1.1B) that has the same solutions as (1.1) but is even easier than (1.1A) to solve. By increasing the ease of solution at each step, after m or fewer exchanges we obtain a system with the same solutions as (1.1) and in an algebraic form that easily produces the solution. For convenience, we say that two systems of linear equations are equivalent if and only if each solution of each system is also a solution of the other.

    We first illustrate this idea with a specific example. Soon we shall be able to verify that the following two systems are equivalent, and for the moment we shall assume that they are.

    Obviously, we would prefer to solve the second system. To do so we let x4 be any number, say c. Then

    and we conclude that for any number c the ordered quadruple

    is a solution of the second system and hence of the first. Furthermore, it is easy to see that any solution of the second system must be of that form, and therefore we have produced the complete solution of the first system. There are infinitely many solutions because each value of c produces a different solution. When a system has infinitely many solutions, a complete description of all solutions involves one, two, or more arbitrary constants.

    The second system is easy to solve because of its special algebraic form: one of the variables (x1) appears with nonzero coefficient in the first equation but in no subsequent equation, another variable (x2) appears with nonzero coefficient in the second equation but in no subsequent equation, and so on. A system of this nature is said to be in echelon form. To solve a system that already is in echelon form we first consider the last equation; we solve for the first variable of that equation in terms of the constant term and the subsequent variables. Each subsequent variable may be assigned an arbitrary value. In this case

    Then we consider the next to last equation; we solve for the first variable of that equation, assigning an arbitrary value to any subsequent variable whose value is not already assigned. For this example,

    Continuing in the same way with each preceding equation, we eventually obtain the complete solution of the system.

    What we need, therefore, is a process that leads from a given system of linear equations to an equivalent system that is in echelon form. And that is precisely the process that Gaussian elimination provides, as we now shall see. Beginning with a system in the form (1.1), we can assume that x1 has a nonzero coefficient in at least one of the m equations. Furthermore, because the solution of a system does not depend on the order in which the equations are written, we can assume further that a11 ≠ 0. Thus we can solve the first equation for x1 in terms of the other variables:

    We then replace x1 by this expression in each of the other equations. The resulting equations then contain variables x2 through xm, and after collecting the coefficients of each of these variables we obtain the equivalent system

    At this stage we need not be concerned with explicit formulas for the new coefficient bij and the new constants ei, where i ≥ 2 and j ≥ 2. Such formulas result immediately from a bit of routine algebra, and we record the results here for future reference.

    The system (1.1A) is said to be obtained from (1.1) by means of a pivot operation on the nonzero entry a11.

    The second stage of Gaussian elimination leaves the first equation of (1.1A) untouched but repeats the pivot process on the reduced system of m − 1 equations in n − 1 variables:

    Conceivably each coefficient bi2 is zero; if so, we look at the coefficients bi3, in order, and continue in this way until we find the first nonzero coefficient, say brs. Again because we can write these equations in any order without changing the solutions, we can assume that r = 2. Then we pivot on b2s; that is, we solve for xs as

    and substitute this expression for xs into each of the last m − 2 equations.

    Together with the original first equation the new system, equivalent to (1.1) and to (1.1A), is of this form:

    Then the pivot process is repeated again on the last m− 2 equations of (1.1B), leaving the first two equations untouched. Continuing in this manner, we eventually obtain a system that is equivalent to (1.1) and is in echelon form.

    To illustrate the method of Gaussian elimination we return to our previous example of three equations in four unknowns. The first equation is

    We pivot on the coefficient 6 by solving for x1,

    substituting this expression in the last two equations, and collecting like terms. The result, which you should verify on scratch paper, is the equivalent system,

    Now we pivot on the coefficient 1 by solving the second equation for x2,

    substituting this expression for x2 in the third equation, and collecting like terms. Again you should verify that the result is

    Although this new system is in echelon form, we can improve its appearance by multiplying each side of the second equation by 2 and each side of the third equation by 6, obtaining an equivalent system in echelon form:

    The last equation contains two variables. We assign arbitrary values to all but one, say x4 = c. Then x3 − 2 + c. Using these values for x3 and x4 in the second equation, we have x2 = −2c, and then from the first equation we obtain x1 = −1, which agrees with our previous solution.

    Suppose we now replace the second equation of this system with a new equation, obtained by adding the two left-hand members and the two right-hand members of the second and third equations,

    or equivalently

    The resulting system is then

    and it is equivalent to the preceding system. Now we replace the first equation by the equation obtained by subtracting the third equation from the first equation,

    and then immediately replace that equation by the equation obtained by twice subtracting the second equation from it,

    or in simpler form

    Then the new system, also in echelon form, is

    Note that this is precisely the system that we solved when this example was originally introduced.

    Let us summarize what we have observed:

    (1) A system of m linear equations in n variables is easily solved if that system is in echelon form.

    (2) Gaussian elimination is a systematic procedure for replacing a given system of linear equations by an equivalent system that is in echelon form.

    (3) Two equivalent systems of linear equations can both be in echelon form and still not be identical; that is, different methods of reducing a system of linear equations to echelon form can produce different (but equivalent) systems of equations in echelon form.

    In the next section we shall use these observations to simplify and to formalize Gaussian elimination as a practical computational method for solving systems of linear equations. In Section 1.3 we shall analyze the various types of solutions that can occur; these types are illustrated in the following exercises.

    EXERCISES 1.1

    1. Use the method of Gaussian elimination to solve each of the following systems of linear equations.

    2. In the following system of linear equations the symbol b represents a number whose value is unspecified.

      (i) Use Gaussian elimination to find an equivalent system that is in echelon form.

     (ii) What value must b have in order that the system have a solution?

    (iii) If b is assigned the value determined in (ii), does the system have more than one solution? Write the complete solution.

    3. Consider the system (1.1) of m linear equations in n variables.

    (i) Let (1.1C) denote the system obtained by replacing the first equation of (1.1) by

    where k is any nonzero constant. Explain why (1.1) and (1.1C) are equivalent. Also explain why (1.1) and (1.1C) are not necessarily equivalent if k = 0.

    (ii) Let (1.1D) denote the system obtained by replacing the first equation of (1.1) by

    Explain why (1.1) and (1.1D) are equivalent.

    (iii) Let (1.1E) denote the system obtained by interchanging the positions of the first two equations of (1.1). Explain why (1.1) and (1.1E) are equivalent.

    4. A system of two linear equations in two unknowns,

    can be interpreted geometrically as two lines in the real coordinate plane. The solution of the system consists of all points that lie simultaneously on both lines. By considering the possible points of intersection of two lines, show that this linear system can have no solutions, exactly one solution, or infinitely many solutions. Are these the only possibilities?

    5. As a special case of the system (1.1), suppose that d1 = d= dm = 0; let the ordered n-tuples U = (u1,..., un) and V = (v1,..., vn) denote two solutions.

      (i) Show that (u1 + v1 ,..., un + vn) is a solution.

     (ii) Show that (bu1 ,..., bun) is a solution for any constant b.

    (iii) Deduce that for any constants b and c,

    is a solution. (This last n-tuple can also be denoted by bU + cV, and it is therefore referred to as a linear combination of the solutions U and V.)

    1.2 MATRIX REPRESENTATION OF A LINEAR SYSTEM

    After solving a few systems of linear equations by hand, we recognize that a lot of unnecessary writing is involved, even for small values of m and n. However, if we agree to arrange the work so that the symbols xj for the n variables always appear in the natural order, we can dispense with writing the symbols for those variables because the required computations involve only the constants in each equation. Thus we can represent the essential information of the linear system (1.1) in skeleton form by the array of numbers

    This rectangular array has m horizontal rows and n + 1 vertical columns. The first row displays the coefficients of the variables and the constant term of the first equation; the second row displays the corresponding information of the second equation, and so on. The first column displays all of the coefficients of x1 in the order in which the equations of the system are written; the second column displays the coefficients of x2, and so on. The last column displays the constant terms on the right-hand sides of the equations, and the vertical line replaces the equality signs. Such an array is called a matrix.

    Definition 1.1  A rectangular array A of real numbers arranged in r rows and s columns is called a real r-by-s matrix. The number in row i and column j is denoted aij; the first subscript is the row index and the second subscript is the column index. A one-by-n matrix is called a row vector, and an m-by-one matrix is called a column vector.

    From the preceding section we recall that our strategy for solving a linear system was to replace the original system by an equivalent system in echelon form. This was accomplished by a succession of m or fewer exchanges, each exchange producing a new system equivalent to the original system but closer to echelon form than the previous system. In this section we shall develop the matrix analogue of that strategy.

    We have previously observed (Exercise 1.1-3) that each of the following operations on a linear system produces a new system that is equivalent to the one on which the operation was performed:

    (M) Multiply each side of any equation of the system by the same nonzero constant.

    (R) Replace any equation of the system by the equation obtained by adding to each side of the given equation a multiple c of the corresponding side of some other equation of the system.

    (P) Permute the equations of the system; that is, rewrite the same equations in different order.

    Because each row of the matrix (1.2) represents the corresponding equation in the system (1.1), these operations on the equations of a linear system give rise to three types of elementary row operations on a matrix:

    Mi(c): Multiply each entry in row i by the nonzero constant c,

    Ri, i+cj: Replace row i by the sum of row i and c times row j, j i,

    Pij: Permute (interchange) row i and row j, j i.

    A matrix that results from applying any one of these operations to the rows of a given matrix will represent a new linear system that is equivalent to the system represented by the given matrix. Hence we are led to adopt the following terminology.

    Definition 1.2  Let A and B denote two r-by-s matrices. B is said to be row equivalent to A if and only if B can be derived from A by applying successively a finite number of elementary row operations.

    By the same analogy we arrive at a definition of row echelon form for matrices.

    Definition 1.3  An r-by-s matrix is said to be in row echelon form if and only if

    (a) for some k r, each of the first k rows contains a nonzero entry, and each entry of the last r k rows is zero, and

    (b) the first nonzero entry in each nonzero row is 1, and it occurs in a column to the right of the first nonzero entry in any preceding row.

    The matrix E, shown below, is in row echelon form with r = 4, s = 5, and k = 3:

    Perhaps you noticed that our definition of row echelon form for matrices is not quite an exact analogue of our description of a linear system in echelon form, because previously we did not require the first nonzero coefficient in each equation to be 1. But by dividing each side of each equation by the first nonzero coefficient of that equation, we obtain an equivalent system in echelon form and having 1 as the first nonzero coefficient of each equation. The reason for standardizing echelon forms in this way will be apparent later in our study.

    Next we demonstrate the use of matrices and elementary row operations in solving a system of linear equations, returning to the example introduced in Section 1.1. The matrix that represents the original system is

    Using Ri, to denote row i of the current matrix and an arrow to denote is replaced by, we perform the following elementary row operations in succession to produce zero in the first column of the second and third rows.

    Be sure that you understand the notation. Each row operation produces a new current matrix, and the rows of that new matrix are used in the next operation. Thus the second operation, R2 → R2 − 3R1, instructs you to replace the second row of the current matrix

    by the sum of the second row and −3 times the first row of that matrix. If you perform these calculations on scratch paper you will verify that the resulting matrix is

    Next we apply row operations on A1 to produce 0 in the second column of the third row, which can be accomplished as follows:

    : replace the third row by the sum of R.

    Again you should verify that the result is

    Finally multiply R3 of A2 by 6 to produce a matrix that is row equivalent to A and is in row echelon form.

    The solution of the linear system represented by A3 is easily obtained from A3. The last row represents the equation

    Let x4 = c, an arbitrary number. Then x3 = 2 + c. The second row represents the equation

    so

    Similarly,

    Hence the solution can be written as the column vector

    If you are not familiar with the method of adding numerical vectors (component-by-component) and the method of multiplying a vector by a number (component-by-component), you may either accept these calculations tentatively, pending our study of vector algebra in the next chapter, or read Section 2.1 at this time.

    It is important to notice that the matrix A1 represents the system obtained in the previous section by a pivot operation on the coefficient 6 in the first equation. And A2 represents the system obtained previously by a pivot on the coefficient 1 in the second equation of the system represented by A1. Hence we see that the process of Gaussian elimination was carried out in this example by a systematic use of elementary row operations, rather than by referring to the formulas for pivot operations, cited in Section 1.1.

    For emphasis we now describe the Gaussian procedure for reducing a matrix to row echelon form by means of elementary row operations. Given an r-by-s matrix

    we examine in order the entries of the first column to locate the first nonzero entry in the first column. If all entries in the first column are 0 we move to the next column and continue the search. If each entry of the matrix is 0, the matrix is already in row echelon form. Otherwise, we thus locate the first nonzero entry, say apq. If p ≠ 1 we then interchange R1 and Rp. Then we have a matrix of the form

    where b1q = apq ≠ 0. The matrix B is simply the matrix A with R1 and Rp interchanged. We now divide R1 by b1q, and for each i > 1 we add to Ri, the required multiple of R1 to produce 0 in row i and column q. The resulting matrix is

    where the exact values for the entries marked with an asterisk are not now of concern. We now repeat this entire procedure for the matrix composed of the last r − 1 rows of C. After r or fewer such repetitions, we obtain a matrix that is row equivalent to A and is in row echelon form.

    It is easy to see that the required multiple of R1 to be added to Ri (after dividing R1 by b1q) is −biq. Hence in C the entry in row i > 1 and column j > q is

    which for q = 1 confirms the formula stated in Section 1.1 for the process of Gaussian elimination.

    To recapitulate, in this section we have seen how the work of solving a system of linear equations can be lessened somewhat by representing the system in matrix form. Furthermore, we have formally expressed the process of Gaussian elimination in terms of elementary row operations on matrices. Stated concisely, a linear system can be solved by first reducing the corresponding matrix to row echelon form and then solving the linear system represented by that matrix.

    Actually, additional row operations can be applied to a matrix in row echelon form to simplify further the procedure for solving a linear system. We illustrate with the matrix E that appears immediately following Definition 1.3. The idea is to use the leading 1 in each nonzero row to produce 0 in that column in each preceding row. Starting with the leading 1 in R3 of E we perform the operations

    The result is

    Now moving to the leading 1 in R2 of E1 we let

    to obtain

    The matrix E2 is in row echelon form and it satisfies the additional property

    (c) the first nonzero entry in each nonzero row is the only nonzero entry in that column.

    A matrix that satisfies (c) in addition to (a) and (b) of Definition 1.3 is said to be in reduced echelon form. It can be proved that for each matrix A there is one and only one matrix E in reduced echelon form that is row equivalent to E. We shall return to this idea in Sections 4.3 and 6.1.

    If we regard E as representing a system of four linear equations in four unknowns, the solution is immediately obvious from E2. The third row of E2 tells us that x4 = 1, the second row requires that x3 = −1, and the first row yields x2 = c (arbitrary) and x1 = −1 − 2c. Hence the solution, written in column vector notation, is

    The reduced echelon form of a matrix results from a modified form of the Gaussian method of solving linear systems; the modified method is called Gauss-Jordan elimination.

    EXERCISES 1.2

    1. For each of the following matrices use elementary row operations to find a matrix in row echelon form that is row equivalent to the given matrix.

    2. Determine the reduced echelon forms of the matrices in Exercise 1.

    3. Solve each of the following systems of linear equations by first reducing a corresponding matrix to echelon form or to reduced echelon form.

    4. Show that each of the three elementary row operations can be undone by an elementary row operation. (For example, if we apply Pik to B to obtain B1, and then apply Pik to B1; the result is B. The corresponding question for Mi(c) also is answered readily, and the question for Ri, i + cj is only slightly more difficult.)

    1.3 SOLUTIONS OF A LINEAR SYSTEM

    Up to this point our attention has been focused mainly on the mechanics of solving a system of m linear equations in n variables. Now we want to describe the solutions of such a system. Any one solution is an ordered n-tuple of numbers, which we can write as a column vector

    having the property that the substitution xi = ci, i = 1,..., n, satisfies each of the m equations of the system. The complete solution S of the system is the set of all individual solutions, and we seek to describe the set S. Can S be the void set; that is, can a linear system fail to have any solution? If the system has one solution, can it have more than one? How many more? And so on.

    We begin by distinguishing two types of linear systems. The linear system (1.1) is said to be homogeneous if and only if d1 = d= dm = 0. If any di is nonzero, the system is nonhomogeneous. We first consider the homogeneous case:

    There exists at least one solution; namely, when each xi = 0. Suppose that there are two solutions,

    Then for each i = 1, ..., m,

    Adding these equations yields

    Since column vectors of the same length can be added, component by component, we deduce the following significant fact.

    (A) If each of two column vectors U and V is a solution of (1.3), then so is their sum, U + V.

    Furthermore for any fixed number k, and for each i = 1, ..., m,

    Thus

    (B) If a column vector U is a solution of (1.3) and if k is any number, then kU is also a solution.

    By combining Statements A and B we can conclude:

    (C) If U and V are solutions of (1.3) and if b and c are any numbers, then bU + cV is also a solution.

    Statement C can be rephrased as follows: any linear combination of solutions of a homogeneous system is also a solution of the system. So now we have a concise description of the set S of all solutions of a homogeneous system: the zero n-tuple is a solution, and any linear combination of solutions is a solution. Conceivably the zero solution is the only solution. But suppose there exists a solution B in which at least one component is nonzero. Then each real number k produces a solution kB. We conclude that every homogeneous system has either exactly one solution or infinitely many solutions.

    For the case n = 3 let us look at the system geometrically. The graph in three-dimensional space of a homogeneous linear equation in three variables,

    is a plane through the origin. A system of m such equations defines m planes, each passing through the origin. Those planes have at least the origin O as a point of intersection. If there is another point P common to all of those planes, then each point on the line through O and P lies on each of those planes. If a point Q not on the line through O and P lies on each plane, then every point on the plane determined by the three points O, P, and Q also lies on each of the m planes of the system. Hence the solution set S of a homogeneous system of m equations in three variables is either a single point (the origin), or an entire line through the origin, or an entire plane through the origin.

    Now let us turn to the general, nonhomogeneous system. For m = n = 2 this problem was described geometrically in Exercise 1.1-4. You should review that exercise at this time before reading on.

    One of the possibilities in the nonhomogeneous case is that no solution exists. This circumstance will be revealed when the matrix of the system is brought into echelon form; at least one of the rows will be of the form

    This corresponds to the equation

    which is a contradiction for all choices of x1,..., xn.

    Now assume that at least one solution W of the nonhomogeneous system exists:

    And let U denote any solution of the associated homogeneous system obtained by replacing each di in the nonhomogeneous system by 0. Then for each i we have

    It follows that

    That is, W + U is a solution of the nonhomogeneous system. Conversely let Y be any solution of the nonhomogeneous system, and let W be a known solution, as before. Then for each i,

    Hence Y W is a solution of the associated homogeneous system. That is, Y = W + U for some solution U of the homogeneous system. Hence if the nonhomogeneous system has any solution W, the complete solution can be obtained by adding W to each vector in the set of all solutions of the associated homogeneous equation.

    Geometrically, for n = 3 a single linear equation,

    represents a plane in three-dimensional space. That plane passes through the origin if and only if d = 0. Several such equations represent several planes, which might have no points in common. The equations of the associated homogeneous system represent planes that pass through the origin, and each is parallel to the corresponding plane of the nonhomogeneous system. The set of all solutions of the homogeneous system is a point, or a line, or a plane, each containing the origin. If the planes of the nonhomogeneous system do intersect at a point W, then the complete set of solutions of the nonhomogeneous system is the point W, or a line through W, or a plane through W, according to the nature of the complete set of solutions of the homogeneous system.

    These observations demonstrate that vectors and matrices provide a natural and convenient language in which to study systems of linear equations. In the next three chapters we shall develop the necessary algebraic and geometric groundwork for further study of systems of linear equations and many other aspects of mathematics in which linearity plays a crucial role. Frequently in that development the answers to quite different questions will depend critically on properties of the solution set of a system of linear equations. After we have developed the basic tools of linear mathematics we shall reexamine some of the ideas of this chapter.

    EXERCISES 1.3

    1. Describe geometrically the set of all solutions to each of the following systems of linear equations.

      (i) The system in Exercise 1.2-3(i).

     (ii) The system in Exercise 1.2-3(ii).

    (iii) The system in Exercise 1.2-3(iii).

    (iv)

    2. Let A denote the matrix that corresponds to a homogeneous system of three equations in three variables. Suppose that A is row equivalent to a matrix in the form

    Describe geometrically the nature of the set of all solutions of that system.

    3. Consider the general system of two linear equations in two variables,

    Assume that a is nonzero and at least one of b and d is nonzero, but otherwise make no other assumptions about the constants.

      (i) Under what conditions on the constants will the system have no solution (that is, the solution set is void)?

     (ii) Under what conditions on the constants will the system have exactly one solution (that is, the solution set contains exactly one element)?

    (iii) Under what conditions on the constants will the solution set be infinite?

    4. Determine a relationship among the constants a, b, and c which is a necessary and sufficient condition that the following system has a solution:

    If that condition is satisfied, will there necessarily be more than one solution? Explain.

    CHAPTER 2

    LINEAR SPACES

    n

    The purpose of this chapter is to develop a general algebraic structure for the study of linearity and to interpret that structure in geometric terms that are familiar from our past experience with the real coordinate plane and three-dimensional space. That basic algebraic structure is called a linear space, or more commonly a vector space.

    In the preceding chapter we used the word vector to refer to an ordered n-tuple of real numbers, written either as a row or a column. Later we shall see that there are many other mathematical entities that justifiably can be called vectors, so it is important to realize that, although any real n-tuple can be regarded as a vector, not every object that we will call a vector can be represented as an ordered n-tuple.

    ², and the set of all ordered nn(a real number) will be denoted by a lowercase letter, such as a, b, x, yn (a real n-tuple) will be denoted by a capital letter, such as A, B, X, Y⁴ might be denoted by A, and if we need to exhibit the numerical components of A, we will write either

    Proper use of notation is important in any mathematical study, particularly in subjects like linear algebra that involve several types of mathematical objects. You are strongly

    Enjoying the preview?
    Page 1 of 1