Explore 1.5M+ audiobooks & ebooks free for days

From $11.99/month after trial. Cancel anytime.

Finite-Dimensional Vector Spaces: Second Edition
Finite-Dimensional Vector Spaces: Second Edition
Finite-Dimensional Vector Spaces: Second Edition
Ebook415 pages4 hoursDover Books on Mathematics

Finite-Dimensional Vector Spaces: Second Edition

Rating: 0 out of 5 stars

()

Read preview

About this ebook

A fine example of a great mathematician's intellect and mathematical style, this classic on linear algebra is widely cited in the literature. The treatment is an ideal supplement to many traditional linear algebra texts and is accessible to undergraduates with some background in algebra.
"This is a classic but still useful introduction to modern linear algebra. It is primarily about linear transformations … It's also extremely well-written and logical, with short and elegant proofs. … The exercises are very good, and are a mixture of proof questions and concrete examples. The book ends with a few applications to analysis … and a brief summary of what is needed to extend this theory to Hilbert spaces." — Allen Stenger, MAA Reviews, maa.org, May, 2016.
"The theory is systematically developed by the axiomatic method that has, since von Neumann, dominated the general approach to linear functional analysis and that achieves here a high degree of lucidity and clarity. The presentation is never awkward or dry, as it sometimes is in other 'modern' textbooks; it is as unconventional as one has come to expect from the author. The book contains about 350 well-placed and instructive problems, which cover a considerable part of the subject. All in all this is an excellent work, of equally high value for both student and teacher." — Zentralblatt für Mathematik.
LanguageEnglish
PublisherDover Publications
Release dateMay 24, 2017
ISBN9780486822266
Finite-Dimensional Vector Spaces: Second Edition

Related to Finite-Dimensional Vector Spaces

Titles in the series (100)

View More

Related ebooks

Mathematics For You

View More

Reviews for Finite-Dimensional Vector Spaces

Rating: 0 out of 5 stars
0 ratings

0 ratings0 reviews

What did you think?

Tap to rate

Review must be at least 10 words

    Book preview

    Finite-Dimensional Vector Spaces - Paul R. Halmos

    CHAPTER I

    SPACES

    §

    1. Fields

    In what follows we shall have occasion to use various classes of numbers (such as the class of all real numbers or the class of all complex numbers). Because we should not, at this early stage, commit ourselves to any specific class, we shall adopt the dodge of referring to numbers as scalars. The reader will not lose anything essential if he consistently interprets scalars as real numbers or as complex numbers; in the examples that we shall study both classes will occur. To be specific (and also in order to operate at the proper level of generality) we proceed to list all the general facts about scalars that we shall need to assume.

    (A) To every pair, α and β, of scalars there corresponds a scalar α + β, called the sum of a and β, in such a way that

    (1) addition is commutative, α + β = β + α,

    (2) addition is associative, α + (β + γ) = (α + β) + γ,

    (3) there exists a unique scalar 0 (called zero) such that α + 0 = α for every scalar a, and

    (4) to every scalar α there corresponds a unique scalar — α such that α+ (–α) = 0.

    (B) To every pair, α and β, of scalars there corresponds a scalar αβ, called the product of α and β, in such a way that

    (1) multiplication is commutative, αβ = βα,

    (2) multiplication is associative, α(βγ) = (αβ)γ,

    (3) there exists a unique non-zero scalar 1 (called one) such that α1 = α for every scalar α, and

    (4) to every non-zero scalar a there corresponds a unique scalar α–1 such that αα–¹ = 1.

    (C) Multiplication is distributive with respect to addition, α(β +γ) = αβ + αγ.

    If addition and multiplication are defined within some set of objects (scalars) so that the conditions (A), (B), and (C) are satisfied, then that set (together with the given operations) is called a field. Thus, for example, the set of all rational numbers (with the ordinary definitions of sum and product) is a field, and the same is true of the set of all real numbers and the set of all complex numbers.

    EXERCISES

    1. Almost all the laws of elementary arithmetic are consequences of the axioms defining a field. Prove, in particular, that if is a field, and if α, β, and γ belong to , then the following relations hold.

    (a) 0 + α = α.

    (b) If α + β = α + γ, then β = γ.

    (c) α + (β α) = β. (Here β α = β + (–α).)

    (d) α·0 = 0·α = 0. (For clarity or emphasis we sometimes use the dot to indicate multiplication.)

    (e) (–1)α = –α.

    (f) (–α)(–β) = αβ.

    (g) If αβ = 0, then either α = 0 or β = 0 (or both).

    2. (a) Is the set of all positive integers a field? (In familiar systems, such as the integers, we shall almost always use the ordinary operations of addition and multiplication. On the rare occasions when we depart from this convention, we shall give ample warning. As for positive, by that word we mean, here and elsewhere in this book, greater than or equal to zero. If 0 is to be excluded, we shall say strictly positive.)

    (b) What about the set of all integers?

    (c) Can the answers to these questions be changed by re-defining addition or multiplication (or both)?

    3. Let m be an integer, m 2, and let m be the set of all positive integers less than m, m = {0, 1, ···, m – 1}. If α and β are in m, let α + β be the least positive remainder obtained by dividing the (ordinary) sum of α and β by m, and, similarly, let αβ be the least positive remainder obtained by dividing the (ordinary) product of α and β by m. (Example: if m = 12, then 3 + 11 = 2 and 3 11 = 9.)

    (a) Prove that m is a field if and only if m is a prime.

    (b) What is –1 in 5?

    (c) What is in 7?

    4. The example of p (where p is a prime) shows that not quite all the laws of elementary arithmetic hold in fields; in 2, for instance, 1 + 1 = 0. Prove that if is a field, then either the result of repeatedly adding 1 to itself is always different from 0, or else the first time that it is equal to 0 occurs when the number of summands is a prime. (The characteristic of the field is defined to be 0 in the first case and the crucial prime in the second.)

    5. Let ( ) be the set of all real numbers of the form α + β , where α and β are rational.

    (a) Is ( ) a field?

    (b) What if α and β are required to be integers?

    6. (a) Does the set of all polynomials with integer coefficients form a field?

    (b) What if the coefficients are allowed to be real numbers?

    7. Let be the set of all (ordered) pairs (α, β) of real numbers.

    (a) If addition and multiplication are defined by

    and

    does become a field?

    (b) If addition and multiplication are defined by

    and

    is a field then?

    (c) What happens (in both the preceding cases) if we consider ordered pairs of complex numbers instead?

    §

    2. Vector spaces

    We come now to the basic concept of this book. For the definition that follows we assume that we are given a particular field ; the scalars to be used are to be elements of .

    DEFINITION. A vector space is a set of elements called vectors satisfying the following axioms.

    (A) To every pair, x and y, of vectors in there corresponds a vector x + y, called the sum of x and y, in such a way that

    (1) addition is commutative, x + y = y + x,

    (2) addition is associative, x + (y + z) = (x + y) + z,

    (3) there exists in a unique vector 0 (called the origin) such that x + 0 = x for every vector x, and

    (4) to every vector x in there corresponds a unique vector – x such that x + (–x) = 0.

    (B) To every pair, α and x, where α is a scalar and x is a vector in , there corresponds a vector αx in , called the product of α and x, in such a way that

    (1) multiplication by scalars is associative, α(βx) = (αβ)x, and

    (2) 1x = x for every vector x.

    (C) (1) Multiplication by scalars is distributive with respect to vector addition, α(x + y) = αx + αy, and

    (2) multiplication by vectors is distributive with respect to scalar addition, (α + β)x = αx + βx.

    These axioms are not claimed to be logically independent; they are merely a convenient characterization of the objects we wish to study. The relation between a vector space and the underlying field is usually described by saying that is a vector space over . If is the field of real numbers, is called a real vector space; similarly if is or if is , we speak of rational vector spaces or complex vector spaces.

    §

    3. Examples

    Before discussing the implications of the axioms, we give some examples. We shall refer to these examples over and over again, and we shall use the notation established here throughout the rest of our work.

    (1) Let ¹(= ) be the set of all complex numbers; if we interpret x + y and αx as ordinary complex numerical addition and multiplication, ¹ becomes a complex vector space.

    (2) Let be the set of all polynomials, with complex coefficients, in a variable t. To make into a complex vector space, we interpret vector addition and scalar multiplication as the ordinary addition of two poly nomials and the multiplication of a polynomial by a complex number; the origin in is the polynomial identically zero.

    Example (1) is too simple and example (2) is too complicated to be typical of the main contents of this book. We give now another example of complex vector spaces which (as we shall see later) is general enough for all our purposes.

    (3) Let n, n = 1, 2, · · ·, be the set of all n-tuples of complex numbers. If x = (ξ1, · · ·, ξn) and y = (η1, · · ·, ηn) are elements of n, we write, by definition,

    It is easy to verify that all parts of our axioms (A), (B), and (C), § 2, are satisfied, so that n is a complex vector space; it will be called n-dimensional complex coordinate space.

    (4) For each positive integer n, let n be the set of all polynomials (with complex coefficients, as in example (2)) of degree n – 1, together with the polynomial identically zero. (In the usual discussion of degree, the degree of this polynomial is not defined, so that we cannot say that it has degree n – 1.) With the same interpretation of the linear operations (addition and scalar multiplication) as in (2), n is a complex vector space.

    (5) A close relative of n is the set n of all n-tuples of real numbers. With the same formal definitions of addition and scalar multiplication as for n, except that now we consider only real scalars α, the space n is a real vector space; it will be called n-dimensional real coordinate space.

    (6) All the preceding examples can be generalized. Thus, for instance, an obvious generalization of (1) can be described by saying that every field may be regarded as a vector space over itself. A common generaliza tion of (3) and (5) starts with an arbitrary field and forms the set n of n-tuples of elements of ; the formal definitions of the linear operations are the same as for the case = .

    (7) A field, by definition, has at least two elements; a vector space, however, may have only one. Since every vector space contains an origin, there is essentially (i.e., except for notation) only one vector space having only one vector. This most trivial vector space will be denoted by .

    (8) If, in the set of all real numbers, addition is defined as usual and multiplication of a real number by a rational number is defined as usual, then becomes a rational vector space.

    (9) If, in the set of all complex numbers, addition is defined as usual and multiplication of a complex number by a real number is defined as usual, then becomes a real vector space. (Compare this example with (1); they are quite different.)

    §

    4. Comments

    A few comments are in order on our axioms and notation. There are striking similarities (and equally striking differences) between the axioms for a field and the axioms for a vector space over a field. In both cases, the axioms (A) describe the additive structure of the system, the axioms (B) describe its multiplicative structure, and the axioms (C) describe the connection between the two structures. Those familiar with algebraic terminology will have recognized the axioms (A) (in both § 1 and § 2) as the defining conditions of an abelian (commutative) group; the axioms (B) and (C) (in § 2) express the fact that the group admits scalars as operators. We mention in passing that if the scalars are elements of a ring (instead of a field), the generalized concept corresponding to a vector space is called a module.

    Special real vector spaces (such as ² and ³) are familiar in geometry. There seems at this stage to be no excuse for our apparently uninteresting insistence on fields other than , and, in particular, on the field of complex numbers. We hope that the reader is willing to take it on faith that we shall have to make use of deep properties of complex numbers later (conjugation, algebraic closure), and that in both the applications of vector spaces to modern (quantum mechanical) physics and the mathematical generalization of our results to Hilbert space, complex numbers play an important role. Their one great disadvantage is the difficulty of drawing pictures; the ordinary picture (Argand diagram) of ¹ is indistinguishable from that of ², and a graphic representation of ² seems to be out of human reach. On the occasions when we have to use pictorial language we shall therefore use the terminology of n in n, and speak of ², for example, as a plane.

    Finally we comment on notation. We observe that the symbol 0 has been used in two meanings: once as a scalar and once as a vector. To make the situation worse, we shall later, when we introduce linear functional and linear transformations, give it still other meanings. Fortunately the relations among the various interpretations of 0 are such that, after this word of warning, no confusion should arise from this practice.

    EXERCISES

    1. Prove that if x and y are vectors and if α is a scalar, then the following relations hold.

    (a) 0 + x = x.

    (b) –0 = 0.

    (c) α·0 = 0.

    (d)0·x = 0. (Observe that the same symbol is used on both sides of this equation; on the left it denotes a scalar, on the right it denotes a vector.)

    (e) If αx = 0, then either α = 0 or x = 0 (or both).

    (f) –x = (1)x.

    (g) y + (x–y) = x. (Here x – y = x + (–y).)

    2. If p is a prime, then is a vector space over p (cf. § 1, Ex. 3); how many vectors are there in this vector space?

    3. Let be the set of all (ordered) pairs of real numbers. If x = (ξ1, ξ2) and y = (η1, η2) are elements of , write

    Is a vector space with respect to these definitions of the linear operations? Why?

    4. Sometimes a subset of a vector space is itself a vector space (with respect to the linear operations already given). Consider, for example, the vector space ³ and the subsets of ³ consisting of those vectors (ξ1, ξ2, ξ3) for which

    (a) ξ1 is real,

    (b) ξ1 = 0,

    (c) either ξ1 = 0 or ξ2 = 0,

    (d) ξ1 + ξ2 = 0,

    (e) ξ1 + ξ2 = 1.

    In which of these cases is a vector space?

    5. Consider the vector space and the subsets of consisting of those vectors (polynomials) x for which

    (a) x has degree 3,

    (b) 2x(0) = x(1),

    (c) x(t) 0 whenever 0 t 1,

    (c) x(t) = x(1 – t) for all t.

    In which of these cases is a vector space?

    §

    5. Linear dependence

    Now that we have described the spaces we shall work with, we must specify the relations among the elements of those spaces that will be of interest to us.

    We begin with a few words about the summation notation. If corresponding to each of a set of indices i there is given a vector xi, and if it is not necessary or not convenient to specify the set of indices exactly, we shall simply speak of a set {xi} of vectors. (We admit the possibility that the same vector corresponds to two distinct indices. In all honesty, therefore, it should be stated that what is important is not which vectors appear in {xi}, but how they appear.) If the index-set under consideration is finite, we shall denote the sum of the corresponding vectors by ∑i xi (or, when desirable, by a more explicit symbol such as . In order to avoid frequent and fussy case distinctions, it is a good idea to admit into the general theory sums such as ∑i xi even when there are no indices i to be summed over, or, more precisely, even when the index-set under consideration is empty. (In that case, of course, there are no vectors to sum, or, more precisely, the set {xi} is also empty.) The value of such an empty sum is defined, naturally enough, to be the vector 0.

    DEFINITION. A finite set {xi} of vectors is linearly dependent if there exists a corresponding set {αi} of scalars, not all zero, such that

    If, on the other hand, ∑i αi xi = 0 implies that αi = 0 for each i, the set {xi} is linearly independent.

    The wording of this definition is intended to cover the case of the empty set; the result in that case, though possibly paradoxical, dovetails very satisfactorily with the rest of the theory. The result is that the empty set of vectors is linearly independent. Indeed, if there are no indices i, then it is not possible to pick out some of them and to assign to the selected ones a non-zero scalar so as to make a certain sum vanish. The trouble is not in avoiding the assignment of zero; it is in finding an index to which something can be assigned. Note that this argument shows that the empty set is not linearly dependent; for the reader not acquainted with arguing by vacuous implication, the equivalence of the definition of linear independence with the straightforward negation of the definition of linear dependence needs a little additional intuitive justification. The easiest way to feel comfortable about the assertion "∑i αixi = 0 implies that αi = 0 for each i," in case there are no indices i, is to rephrase it this way: "if ∑i αixi = 0, then there is no index i for which αi ≠ 0." This version is obviously true if there is no index i at all.

    Linear dependence and independence are properties of sets of vectors; it is customary, however, to apply the adjectives to vectors themselves, and thus we shall sometimes say a set of linearly independent vectors instead of a linearly independent set of vectors. It will be convenient also to speak of the linear dependence and independence of a not necessarily finite set, , of vectors. We shall say that is linearly independent if every finite subset of is such; otherwise is linearly dependent.

    To gain insight into the meaning of linear dependence, let us study the examples of vector spaces that we already have.

    (1) If x and y are any two vectors in ¹, then x and y form a linearly dependent set. If x = y = 0, this is trivial; if not, then we have, for example, the relation yx + (–x)y = 0. Since it is clear that every set containing a linearly dependent subset is itself linearly dependent, this shows that in ¹ every set containing more than one element is a linearly dependent set.

    (2) More interesting is the situation in the space . The vectors x, y, and z, defined by

    are, for example, linearly dependent, since x + y – z = 0. However, the infinite set of vectors x0, x1, x2, · · ·, defined by

    is a linearly independent set, for if we had any relation of the form

    then we should have a polynomial identity

    whence

    (3) As we mentioned before, the spaces n are the prototype of what we want to study; let us examine, for example, the case n = 3. To those familiar with higher-dimensional geometry, the notion of linear dependence in this space (or, more properly speaking, in its real analogue ³) has a concrete geometric meaning, which we shall only mention. In geometrical language, two vectors are linearly dependent if and only if they are collinear with the origin, and three vectors are linearly dependent if and only if they are coplanar with the origin. (If one thinks of a vector not as a point in a space but as an arrow pointing from the origin to some given point, the preceding sentence should be modified by crossing out the phrase with the origin both times that it occurs.) We shall presently introduce the notion of linear manifolds (or vector subspaces) in a vector space, and, in that connection, we shall occasionally use the language suggested by such geometrical considerations.

    §

    6. Linear combinations

    We shall say, whenever x = ∑i αixi, that x is a linear combination of {xi}; we shall use without any further explanation all the simple grammatical implications of this terminology. Thus we shall say, in case x is a linear combination of {xi}, that x is linearly dependent on {xi}; we shall leave to the reader the proof that if {xi} is linearly independent, then a necessary and sufficient condition that x be a linear combination of {xi} is that the enlarged set, obtained by adjoining x to {xi}, be linearly dependent. Note that, in accordance with the definition of an empty sum, the origin is a linear combination of the empty set of vectors; it is, moreover, the only vector with this property.

    The following theorem is the fundamental result concerning linear dependence.

    THEOREM. The set of non-zero vectors x1, · · ·, xn is linearly dependent if and only if some xk, 2 k n, is a linear combination of the preceding ones,

    PROOF. Let us suppose that the vectors x1, · · ·, xn are linearly dependent, and let k be the first integer between 2 and n for which x1, · · ·, xk are linearly dependent. (If worse comes to worst, our assumption assures us that k = n will do.) Then

    for a suitable set of α’s (not all zero); moreover, whatever the α’s, we cannot have αk = 0, for then we should have a linear dependence relation among x1, · · ·, xk–1, contrary to the definition of k. Hence

    as was to be proved. This proves the necessity of our condition; sufficiency is clear since, as we remarked before, every set containing a linearly dependent set is itself such.

    §

    7. Bases

    DEFINITION. A (linear) basis (or a coordinate system) in a vector space is a set of linearly independent vectors such that every vector in is a linear combination of elements of . A vector space is finite-dimensional if it has a finite basis.

    Except for the occasional consideration of examples we shall restrict our attention, throughout this book, to finite-dimensional vector spaces.

    For examples of bases we turn again to the spaces and n. In , the set {xn}, where xn(t) = tn, n = 0, 1, 2, · · ·, is a basis; every polynomial is, by definition, a linear combination of a finite number of xn. Moreover has no finite basis, for, given any finite set of polynomials, we can find a polynomial of higher degree than any of them; this latter polynomial is obviously not a linear combination of the former ones.

    An example of a basis in n is the set of vectors xi, i = 1, · · ·, n, defined by the condition that the j-th coordinate of xi is δij. (Here we use for the first time the popular Kronecker δ; it is defined by δij = 1 if i = j and δij = 0 if i j.) Thus we assert that in ³ the vectors x1 = (1, 0, 0), x2 = (0, 1, 0), and x3 = (0, 0, 1) form a basis. It is easy to see that they are linearly independent; the formula

    proves that every x in

    Enjoying the preview?
    Page 1 of 1