Finite-Dimensional Vector Spaces: Second Edition
()
About this ebook
"This is a classic but still useful introduction to modern linear algebra. It is primarily about linear transformations … It's also extremely well-written and logical, with short and elegant proofs. … The exercises are very good, and are a mixture of proof questions and concrete examples. The book ends with a few applications to analysis … and a brief summary of what is needed to extend this theory to Hilbert spaces." — Allen Stenger, MAA Reviews, maa.org, May, 2016.
"The theory is systematically developed by the axiomatic method that has, since von Neumann, dominated the general approach to linear functional analysis and that achieves here a high degree of lucidity and clarity. The presentation is never awkward or dry, as it sometimes is in other 'modern' textbooks; it is as unconventional as one has come to expect from the author. The book contains about 350 well-placed and instructive problems, which cover a considerable part of the subject. All in all this is an excellent work, of equally high value for both student and teacher." — Zentralblatt für Mathematik.
Related to Finite-Dimensional Vector Spaces
Titles in the series (100)
First-Order Partial Differential Equations, Vol. 1 Rating: 5 out of 5 stars5/5Fourier Series and Orthogonal Polynomials Rating: 0 out of 5 stars0 ratingsTopology for Analysis Rating: 4 out of 5 stars4/5Laplace Transforms and Their Applications to Differential Equations Rating: 5 out of 5 stars5/5Counterexamples in Topology Rating: 4 out of 5 stars4/5The History of the Calculus and Its Conceptual Development Rating: 4 out of 5 stars4/5Infinite Series Rating: 4 out of 5 stars4/5Analytic Inequalities Rating: 5 out of 5 stars5/5Mathematics for the Nonmathematician Rating: 4 out of 5 stars4/5Methods of Applied Mathematics Rating: 3 out of 5 stars3/5An Adventurer's Guide to Number Theory Rating: 3 out of 5 stars3/5How to Gamble If You Must: Inequalities for Stochastic Processes Rating: 0 out of 5 stars0 ratingsCalculus Refresher Rating: 3 out of 5 stars3/5Optimization Theory for Large Systems Rating: 5 out of 5 stars5/5A First Course in Partial Differential Equations: with Complex Variables and Transform Methods Rating: 5 out of 5 stars5/5Statistical Inference Rating: 4 out of 5 stars4/5Calculus: An Intuitive and Physical Approach (Second Edition) Rating: 4 out of 5 stars4/5A Catalog of Special Plane Curves Rating: 2 out of 5 stars2/5History of the Theory of Numbers, Volume II: Diophantine Analysis Rating: 0 out of 5 stars0 ratingsDifferential Forms Rating: 5 out of 5 stars5/5Theory of Approximation Rating: 0 out of 5 stars0 ratingsChebyshev and Fourier Spectral Methods: Second Revised Edition Rating: 4 out of 5 stars4/5First-Order Partial Differential Equations, Vol. 2 Rating: 0 out of 5 stars0 ratingsApplied Multivariate Analysis: Using Bayesian and Frequentist Methods of Inference, Second Edition Rating: 0 out of 5 stars0 ratingsNumerical Methods Rating: 5 out of 5 stars5/5The Foundations of Statistics Rating: 0 out of 5 stars0 ratingsDynamic Probabilistic Systems, Volume II: Semi-Markov and Decision Processes Rating: 0 out of 5 stars0 ratingsAn Introduction to Lebesgue Integration and Fourier Series Rating: 0 out of 5 stars0 ratingsAdvanced Calculus: Second Edition Rating: 5 out of 5 stars5/5Differential Forms with Applications to the Physical Sciences Rating: 5 out of 5 stars5/5
Related ebooks
Differential Topology: An Introduction Rating: 0 out of 5 stars0 ratingsPrinciples of Algebraic Geometry Rating: 5 out of 5 stars5/5Rotations, Quaternions, and Double Groups Rating: 3 out of 5 stars3/5Introduction to Topological Groups Rating: 0 out of 5 stars0 ratingsElements of Abstract Algebra Rating: 4 out of 5 stars4/5The Theory of Spinors Rating: 0 out of 5 stars0 ratingsTheory of Linear Operators in Hilbert Space Rating: 0 out of 5 stars0 ratingsLectures on Ergodic Theory Rating: 0 out of 5 stars0 ratingsElementary Number Theory: An Algebraic Approach Rating: 0 out of 5 stars0 ratingsMatrix Representations of Groups Rating: 0 out of 5 stars0 ratingsElementary Functional Analysis Rating: 4 out of 5 stars4/5Introduction to Hilbert Space and the Theory of Spectral Multiplicity: Second Edition Rating: 0 out of 5 stars0 ratingsGeometric Algebra Rating: 5 out of 5 stars5/5Algebraic Geometry Rating: 0 out of 5 stars0 ratingsIntroduction to Topology: Second Edition Rating: 4 out of 5 stars4/5The Red Book of Mathematical Problems Rating: 0 out of 5 stars0 ratingsTheory of Lie Groups Rating: 0 out of 5 stars0 ratingsFundamental Concepts of Abstract Algebra Rating: 5 out of 5 stars5/5Abstract Analytic Number Theory Rating: 0 out of 5 stars0 ratingsAbelian Varieties Rating: 0 out of 5 stars0 ratingsIntroduction to Algebraic Geometry Rating: 4 out of 5 stars4/5Group Theory I Essentials Rating: 0 out of 5 stars0 ratingsTopological Transformation Groups Rating: 3 out of 5 stars3/5Analytic Inequalities Rating: 5 out of 5 stars5/5Algebra: Polynomials, Galois Theory and Applications Rating: 0 out of 5 stars0 ratingsFinite Field Fun: A lightweight introduction to finite fields and their applications for engineers, computer scientists, and others Rating: 0 out of 5 stars0 ratingsLinear Algebra Rating: 3 out of 5 stars3/5The Theory of Algebraic Numbers Rating: 4 out of 5 stars4/5Elementary Algebraic Geometry: Second Edition Rating: 0 out of 5 stars0 ratings
Mathematics For You
Fluent in 3 Months: How Anyone at Any Age Can Learn to Speak Any Language from Anywhere in the World Rating: 3 out of 5 stars3/5Quantum Physics for Beginners Rating: 4 out of 5 stars4/5The Little Book of Mathematical Principles, Theories & Things Rating: 3 out of 5 stars3/5What If? 10th Anniversary Edition: Serious Scientific Answers to Absurd Hypothetical Questions Rating: 4 out of 5 stars4/5Calculus Made Easy Rating: 4 out of 5 stars4/5Seeing Further: The Story of Science and the Royal Society Rating: 4 out of 5 stars4/5Mental Math Secrets - How To Be a Human Calculator Rating: 5 out of 5 stars5/5Basic Math & Pre-Algebra Workbook For Dummies with Online Practice Rating: 3 out of 5 stars3/5A Mind For Numbers: How to Excel at Math and Science (Even If You Flunked Algebra) Rating: 4 out of 5 stars4/5Basic Math & Pre-Algebra For Dummies Rating: 4 out of 5 stars4/5Pre-Calculus For Dummies Rating: 5 out of 5 stars5/5My Best Mathematical and Logic Puzzles Rating: 4 out of 5 stars4/5The Everything Guide to Algebra: A Step-by-Step Guide to the Basics of Algebra - in Plain English! Rating: 4 out of 5 stars4/5Algebra I For Dummies Rating: 4 out of 5 stars4/5Algebra I Workbook For Dummies Rating: 3 out of 5 stars3/5Algebra - The Very Basics Rating: 5 out of 5 stars5/5Math Magic: How To Master Everyday Math Problems Rating: 3 out of 5 stars3/5Introducing Game Theory: A Graphic Guide Rating: 4 out of 5 stars4/5The Math of Life and Death: 7 Mathematical Principles That Shape Our Lives Rating: 4 out of 5 stars4/5Standard Deviations: Flawed Assumptions, Tortured Data, and Other Ways to Lie with Statistics Rating: 4 out of 5 stars4/5Geometry For Dummies Rating: 4 out of 5 stars4/5Alan Turing: The Enigma: The Book That Inspired the Film The Imitation Game - Updated Edition Rating: 4 out of 5 stars4/5Calculus Essentials For Dummies Rating: 5 out of 5 stars5/5Why Machines Learn: The Elegant Math Behind Modern AI Rating: 3 out of 5 stars3/5Calculus For Dummies Rating: 4 out of 5 stars4/5Limitless Mind: Learn, Lead, and Live Without Barriers Rating: 4 out of 5 stars4/5Mental Math: Tricks To Become A Human Calculator Rating: 2 out of 5 stars2/5Flatland Rating: 4 out of 5 stars4/5
Reviews for Finite-Dimensional Vector Spaces
0 ratings0 reviews
Book preview
Finite-Dimensional Vector Spaces - Paul R. Halmos
CHAPTER I
SPACES
§
1. Fields
In what follows we shall have occasion to use various classes of numbers (such as the class of all real numbers or the class of all complex numbers). Because we should not, at this early stage, commit ourselves to any specific class, we shall adopt the dodge of referring to numbers as scalars. The reader will not lose anything essential if he consistently interprets scalars as real numbers or as complex numbers; in the examples that we shall study both classes will occur. To be specific (and also in order to operate at the proper level of generality) we proceed to list all the general facts about scalars that we shall need to assume.
(A) To every pair, α and β, of scalars there corresponds a scalar α + β, called the sum of a and β, in such a way that
(1) addition is commutative, α + β = β + α,
(2) addition is associative, α + (β + γ) = (α + β) + γ,
(3) there exists a unique scalar 0 (called zero) such that α + 0 = α for every scalar a, and
(4) to every scalar α there corresponds a unique scalar — α such that α+ (–α) = 0.
(B) To every pair, α and β, of scalars there corresponds a scalar αβ, called the product of α and β, in such a way that
(1) multiplication is commutative, αβ = βα,
(2) multiplication is associative, α(βγ) = (αβ)γ,
(3) there exists a unique non-zero scalar 1 (called one) such that α1 = α for every scalar α, and
(4) to every non-zero scalar a there corresponds a unique scalar α–1 such that αα–¹ = 1.
(C) Multiplication is distributive with respect to addition, α(β +γ) = αβ + αγ.
If addition and multiplication are defined within some set of objects (scalars) so that the conditions (A), (B), and (C) are satisfied, then that set (together with the given operations) is called a field. Thus, for example, the set of all rational numbers (with the ordinary definitions of sum and product) is a field, and the same is true of the set of all real numbers and the set of all complex numbers.
EXERCISES
1. Almost all the laws of elementary arithmetic are consequences of the axioms defining a field. Prove, in particular, that if is a field, and if α, β, and γ belong to , then the following relations hold.
(a) 0 + α = α.
(b) If α + β = α + γ, then β = γ.
(c) α + (β – α) = β. (Here β – α = β + (–α).)
(d) α·0 = 0·α = 0. (For clarity or emphasis we sometimes use the dot to indicate multiplication.)
(e) (–1)α = –α.
(f) (–α)(–β) = αβ.
(g) If αβ = 0, then either α = 0 or β = 0 (or both).
2. (a) Is the set of all positive integers a field? (In familiar systems, such as the integers, we shall almost always use the ordinary operations of addition and multiplication. On the rare occasions when we depart from this convention, we shall give ample warning. As for positive,
by that word we mean, here and elsewhere in this book, greater than or equal to zero.
If 0 is to be excluded, we shall say strictly positive.
)
(b) What about the set of all integers?
(c) Can the answers to these questions be changed by re-defining addition or multiplication (or both)?
3. Let m be an integer, m 2, and let m be the set of all positive integers less than m, m = {0, 1, ···, m – 1}. If α and β are in m, let α + β be the least positive remainder obtained by dividing the (ordinary) sum of α and β by m, and, similarly, let αβ be the least positive remainder obtained by dividing the (ordinary) product of α and β by m. (Example: if m = 12, then 3 + 11 = 2 and 3 11 = 9.)
(a) Prove that m is a field if and only if m is a prime.
(b) What is –1 in 5?
(c) What is in 7?
4. The example of p (where p is a prime) shows that not quite all the laws of elementary arithmetic hold in fields; in 2, for instance, 1 + 1 = 0. Prove that if is a field, then either the result of repeatedly adding 1 to itself is always different from 0, or else the first time that it is equal to 0 occurs when the number of summands is a prime. (The characteristic of the field is defined to be 0 in the first case and the crucial prime in the second.)
5. Let ( ) be the set of all real numbers of the form α + β , where α and β are rational.
(a) Is ( ) a field?
(b) What if α and β are required to be integers?
6. (a) Does the set of all polynomials with integer coefficients form a field?
(b) What if the coefficients are allowed to be real numbers?
7. Let be the set of all (ordered) pairs (α, β) of real numbers.
(a) If addition and multiplication are defined by
and
does become a field?
(b) If addition and multiplication are defined by
and
is a field then?
(c) What happens (in both the preceding cases) if we consider ordered pairs of complex numbers instead?
§
2. Vector spaces
We come now to the basic concept of this book. For the definition that follows we assume that we are given a particular field ; the scalars to be used are to be elements of .
DEFINITION. A vector space is a set of elements called vectors satisfying the following axioms.
(A) To every pair, x and y, of vectors in there corresponds a vector x + y, called the sum of x and y, in such a way that
(1) addition is commutative, x + y = y + x,
(2) addition is associative, x + (y + z) = (x + y) + z,
(3) there exists in a unique vector 0 (called the origin) such that x + 0 = x for every vector x, and
(4) to every vector x in there corresponds a unique vector – x such that x + (–x) = 0.
(B) To every pair, α and x, where α is a scalar and x is a vector in , there corresponds a vector αx in , called the product of α and x, in such a way that
(1) multiplication by scalars is associative, α(βx) = (αβ)x, and
(2) 1x = x for every vector x.
(C) (1) Multiplication by scalars is distributive with respect to vector addition, α(x + y) = αx + αy, and
(2) multiplication by vectors is distributive with respect to scalar addition, (α + β)x = αx + βx.
These axioms are not claimed to be logically independent; they are merely a convenient characterization of the objects we wish to study. The relation between a vector space and the underlying field is usually described by saying that is a vector space over . If is the field of real numbers, is called a real vector space; similarly if is or if is , we speak of rational vector spaces or complex vector spaces.
§
3. Examples
Before discussing the implications of the axioms, we give some examples. We shall refer to these examples over and over again, and we shall use the notation established here throughout the rest of our work.
(1) Let ¹(= ) be the set of all complex numbers; if we interpret x + y and αx as ordinary complex numerical addition and multiplication, ¹ becomes a complex vector space.
(2) Let be the set of all polynomials, with complex coefficients, in a variable t. To make into a complex vector space, we interpret vector addition and scalar multiplication as the ordinary addition of two poly nomials and the multiplication of a polynomial by a complex number; the origin in is the polynomial identically zero.
Example (1) is too simple and example (2) is too complicated to be typical of the main contents of this book. We give now another example of complex vector spaces which (as we shall see later) is general enough for all our purposes.
(3) Let n, n = 1, 2, · · ·, be the set of all n-tuples of complex numbers. If x = (ξ1, · · ·, ξn) and y = (η1, · · ·, ηn) are elements of n, we write, by definition,
It is easy to verify that all parts of our axioms (A), (B), and (C), § 2, are satisfied, so that n is a complex vector space; it will be called n-dimensional complex coordinate space.
(4) For each positive integer n, let n be the set of all polynomials (with complex coefficients, as in example (2)) of degree n – 1, together with the polynomial identically zero. (In the usual discussion of degree, the degree of this polynomial is not defined, so that we cannot say that it has degree n – 1.) With the same interpretation of the linear operations (addition and scalar multiplication) as in (2), n is a complex vector space.
(5) A close relative of n is the set n of all n-tuples of real numbers. With the same formal definitions of addition and scalar multiplication as for n, except that now we consider only real scalars α, the space n is a real vector space; it will be called n-dimensional real coordinate space.
(6) All the preceding examples can be generalized. Thus, for instance, an obvious generalization of (1) can be described by saying that every field may be regarded as a vector space over itself. A common generaliza tion of (3) and (5) starts with an arbitrary field and forms the set n of n-tuples of elements of ; the formal definitions of the linear operations are the same as for the case = .
(7) A field, by definition, has at least two elements; a vector space, however, may have only one. Since every vector space contains an origin, there is essentially (i.e., except for notation) only one vector space having only one vector. This most trivial vector space will be denoted by .
(8) If, in the set of all real numbers, addition is defined as usual and multiplication of a real number by a rational number is defined as usual, then becomes a rational vector space.
(9) If, in the set of all complex numbers, addition is defined as usual and multiplication of a complex number by a real number is defined as usual, then becomes a real vector space. (Compare this example with (1); they are quite different.)
§
4. Comments
A few comments are in order on our axioms and notation. There are striking similarities (and equally striking differences) between the axioms for a field and the axioms for a vector space over a field. In both cases, the axioms (A) describe the additive structure of the system, the axioms (B) describe its multiplicative structure, and the axioms (C) describe the connection between the two structures. Those familiar with algebraic terminology will have recognized the axioms (A) (in both § 1 and § 2) as the defining conditions of an abelian (commutative) group; the axioms (B) and (C) (in § 2) express the fact that the group admits scalars as operators. We mention in passing that if the scalars are elements of a ring (instead of a field), the generalized concept corresponding to a vector space is called a module.
Special real vector spaces (such as ² and ³) are familiar in geometry. There seems at this stage to be no excuse for our apparently uninteresting insistence on fields other than , and, in particular, on the field of complex numbers. We hope that the reader is willing to take it on faith that we shall have to make use of deep properties of complex numbers later (conjugation, algebraic closure), and that in both the applications of vector spaces to modern (quantum mechanical) physics and the mathematical generalization of our results to Hilbert space, complex numbers play an important role. Their one great disadvantage is the difficulty of drawing pictures; the ordinary picture (Argand diagram) of ¹ is indistinguishable from that of ², and a graphic representation of ² seems to be out of human reach. On the occasions when we have to use pictorial language we shall therefore use the terminology of n in n, and speak of ², for example, as a plane.
Finally we comment on notation. We observe that the symbol 0 has been used in two meanings: once as a scalar and once as a vector. To make the situation worse, we shall later, when we introduce linear functional and linear transformations, give it still other meanings. Fortunately the relations among the various interpretations of 0 are such that, after this word of warning, no confusion should arise from this practice.
EXERCISES
1. Prove that if x and y are vectors and if α is a scalar, then the following relations hold.
(a) 0 + x = x.
(b) –0 = 0.
(c) α·0 = 0.
(d)0·x = 0. (Observe that the same symbol is used on both sides of this equation; on the left it denotes a scalar, on the right it denotes a vector.)
(e) If αx = 0, then either α = 0 or x = 0 (or both).
(f) –x = (–1)x.
(g) y + (x–y) = x. (Here x – y = x + (–y).)
2. If p is a prime, then is a vector space over p (cf. § 1, Ex. 3); how many vectors are there in this vector space?
3. Let be the set of all (ordered) pairs of real numbers. If x = (ξ1, ξ2) and y = (η1, η2) are elements of , write
Is a vector space with respect to these definitions of the linear operations? Why?
4. Sometimes a subset of a vector space is itself a vector space (with respect to the linear operations already given). Consider, for example, the vector space ³ and the subsets of ³ consisting of those vectors (ξ1, ξ2, ξ3) for which
(a) ξ1 is real,
(b) ξ1 = 0,
(c) either ξ1 = 0 or ξ2 = 0,
(d) ξ1 + ξ2 = 0,
(e) ξ1 + ξ2 = 1.
In which of these cases is a vector space?
5. Consider the vector space and the subsets of consisting of those vectors (polynomials) x for which
(a) x has degree 3,
(b) 2x(0) = x(1),
(c) x(t) 0 whenever 0 t 1,
(c) x(t) = x(1 – t) for all t.
In which of these cases is a vector space?
§
5. Linear dependence
Now that we have described the spaces we shall work with, we must specify the relations among the elements of those spaces that will be of interest to us.
We begin with a few words about the summation notation. If corresponding to each of a set of indices i there is given a vector xi, and if it is not necessary or not convenient to specify the set of indices exactly, we shall simply speak of a set {xi} of vectors. (We admit the possibility that the same vector corresponds to two distinct indices. In all honesty, therefore, it should be stated that what is important is not which vectors appear in {xi}, but how they appear.) If the index-set under consideration is finite, we shall denote the sum of the corresponding vectors by ∑i xi (or, when desirable, by a more explicit symbol such as . In order to avoid frequent and fussy case distinctions, it is a good idea to admit into the general theory sums such as ∑i xi even when there are no indices i to be summed over, or, more precisely, even when the index-set under consideration is empty. (In that case, of course, there are no vectors to sum, or, more precisely, the set {xi} is also empty.) The value of such an empty sum
is defined, naturally enough, to be the vector 0.
DEFINITION. A finite set {xi} of vectors is linearly dependent if there exists a corresponding set {αi} of scalars, not all zero, such that
If, on the other hand, ∑i αi xi = 0 implies that αi = 0 for each i, the set {xi} is linearly independent.
The wording of this definition is intended to cover the case of the empty set; the result in that case, though possibly paradoxical, dovetails very satisfactorily with the rest of the theory. The result is that the empty set of vectors is linearly independent. Indeed, if there are no indices i, then it is not possible to pick out some of them and to assign to the selected ones a non-zero scalar so as to make a certain sum vanish. The trouble is not in avoiding the assignment of zero; it is in finding an index to which something can be assigned. Note that this argument shows that the empty set is not linearly dependent; for the reader not acquainted with arguing by vacuous implication,
the equivalence of the definition of linear independence with the straightforward negation of the definition of linear dependence needs a little additional intuitive justification. The easiest way to feel comfortable about the assertion "∑i αixi = 0 implies that αi = 0 for each i," in case there are no indices i, is to rephrase it this way: "if ∑i αixi = 0, then there is no index i for which αi ≠ 0." This version is obviously true if there is no index i at all.
Linear dependence and independence are properties of sets of vectors; it is customary, however, to apply the adjectives to vectors themselves, and thus we shall sometimes say a set of linearly independent vectors
instead of a linearly independent set of vectors.
It will be convenient also to speak of the linear dependence and independence of a not necessarily finite set, , of vectors. We shall say that is linearly independent if every finite subset of is such; otherwise is linearly dependent.
To gain insight into the meaning of linear dependence, let us study the examples of vector spaces that we already have.
(1) If x and y are any two vectors in ¹, then x and y form a linearly dependent set. If x = y = 0, this is trivial; if not, then we have, for example, the relation yx + (–x)y = 0. Since it is clear that every set containing a linearly dependent subset is itself linearly dependent, this shows that in ¹ every set containing more than one element is a linearly dependent set.
(2) More interesting is the situation in the space . The vectors x, y, and z, defined by
are, for example, linearly dependent, since x + y – z = 0. However, the infinite set of vectors x0, x1, x2, · · ·, defined by
is a linearly independent set, for if we had any relation of the form
then we should have a polynomial identity
whence
(3) As we mentioned before, the spaces n are the prototype of what we want to study; let us examine, for example, the case n = 3. To those familiar with higher-dimensional geometry, the notion of linear dependence in this space (or, more properly speaking, in its real analogue ³) has a concrete geometric meaning, which we shall only mention. In geometrical language, two vectors are linearly dependent if and only if they are collinear with the origin, and three vectors are linearly dependent if and only if they are coplanar with the origin. (If one thinks of a vector not as a point in a space but as an arrow pointing from the origin to some given point, the preceding sentence should be modified by crossing out the phrase with the origin
both times that it occurs.) We shall presently introduce the notion of linear manifolds (or vector subspaces) in a vector space, and, in that connection, we shall occasionally use the language suggested by such geometrical considerations.
§
6. Linear combinations
We shall say, whenever x = ∑i αixi, that x is a linear combination of {xi}; we shall use without any further explanation all the simple grammatical implications of this terminology. Thus we shall say, in case x is a linear combination of {xi}, that x is linearly dependent on {xi}; we shall leave to the reader the proof that if {xi} is linearly independent, then a necessary and sufficient condition that x be a linear combination of {xi} is that the enlarged set, obtained by adjoining x to {xi}, be linearly dependent. Note that, in accordance with the definition of an empty sum, the origin is a linear combination of the empty set of vectors; it is, moreover, the only vector with this property.
The following theorem is the fundamental result concerning linear dependence.
THEOREM. The set of non-zero vectors x1, · · ·, xn is linearly dependent if and only if some xk, 2 k n, is a linear combination of the preceding ones,
PROOF. Let us suppose that the vectors x1, · · ·, xn are linearly dependent, and let k be the first integer between 2 and n for which x1, · · ·, xk are linearly dependent. (If worse comes to worst, our assumption assures us that k = n will do.) Then
for a suitable set of α’s (not all zero); moreover, whatever the α’s, we cannot have αk = 0, for then we should have a linear dependence relation among x1, · · ·, xk–1, contrary to the definition of k. Hence
as was to be proved. This proves the necessity of our condition; sufficiency is clear since, as we remarked before, every set containing a linearly dependent set is itself such.
§
7. Bases
DEFINITION. A (linear) basis (or a coordinate system) in a vector space is a set of linearly independent vectors such that every vector in is a linear combination of elements of . A vector space is finite-dimensional if it has a finite basis.
Except for the occasional consideration of examples we shall restrict our attention, throughout this book, to finite-dimensional vector spaces.
For examples of bases we turn again to the spaces and n. In , the set {xn}, where xn(t) = tn, n = 0, 1, 2, · · ·, is a basis; every polynomial is, by definition, a linear combination of a finite number of xn. Moreover has no finite basis, for, given any finite set of polynomials, we can find a polynomial of higher degree than any of them; this latter polynomial is obviously not a linear combination of the former ones.
An example of a basis in n is the set of vectors xi, i = 1, · · ·, n, defined by the condition that the j-th coordinate of xi is δij. (Here we use for the first time the popular Kronecker δ; it is defined by δij = 1 if i = j and δij = 0 if i ≠ j.) Thus we assert that in ³ the vectors x1 = (1, 0, 0), x2 = (0, 1, 0), and x3 = (0, 0, 1) form a basis. It is easy to see that they are linearly independent; the formula
proves that every x in