Baixe o app para aproveitar ainda mais
Prévia do material em texto
A Concise Text on Advanced Linear Algebra This engaging textbook for advanced undergraduate students and beginning graduates covers the core subjects in linear algebra. The author motivates the concepts by drawing clear links to applications and other important areas. The book places particular emphasis on integrating ideas from analysis wherever appropriate and features many novelties in its presentation. For example, the notion of determinant is shown to appear from calculating the index of a vector field which leads to a self-contained proof of the Fundamental Theorem of Algebra; the Cayley–Hamilton theorem is established by recognizing the fact that the set of complex matrices of distinct eigenvalues is dense; the existence of a real eigenvalue of a self-adjoint map is deduced by the method of calculus; the construction of the Jordan decomposition is seen to boil down to understanding nilpotent maps of degree two; and a lucid and elementary introduction to quantum mechanics based on linear algebra is given. The material is supplemented by a rich collection of over 350 mostly proof-oriented exercises, suitable for readers from a wide variety of backgrounds. Selected solutions are provided at the back of the book, making it ideal for self-study as well as for use as a course text. www.pdfgrip.com www.pdfgrip.com A Concise Text on Advanced Linear Algebra YISONG YANG Polytechnic School of Engineering, New York University www.pdfgrip.com University Printing House, Cambridge CB2 8BS, United Kingdom Cambridge University Press is part of the University of Cambridge. It furthers the University’s mission by disseminating knowledge in the pursuit of education, learning and research at the highest international levels of excellence. www.cambridge.org Information on this title: www.cambridge.org/9781107087514 c© Yisong Yang 2015 This publication is in copyright. Subject to statutory exception and to the provisions of relevant collective licensing agreements, no reproduction of any part may take place without the written permission of Cambridge University Press. First published 2015 Printed in the United Kingdom by Clays, St Ives plc A catalogue record for this publication is available from the British Library Library of Congress Cataloguing in Publication data Yang, Yisong. A concise text on advanced linear algebra / Yisong Yang, Polytechnic School of Engineering, New York University. pages cm Includes bibliographical references and index. ISBN 978-1-107-08751-4 (Hardback) – ISBN 978-1-107-45681-5 (Paperback) 1. Algebras, Linear–Textbooks. 2. Algebras, Linear–Study and teaching (Higher) . 3. Algebras, Linear–Study and teaching (Graduate). I. Title. II. Title: Advanced linear algebra. QA184.2.Y36 2015 512′.5–dc23 2014028951 ISBN 978-1-107-08751-4 Hardback ISBN 978-1-107-45681-5 Paperback Cambridge University Press has no responsibility for the persistence or accuracy of URLs for external or third-party internet websites referred to in this publication, and does not guarantee that any content on such websites is, or will remain, accurate or appropriate. www.pdfgrip.com For Sheng, Peter, Anna, and Julia www.pdfgrip.com www.pdfgrip.com Contents Preface page ix Notation and convention xiii 1 Vector spaces 1 1.1 Vector spaces 1 1.2 Subspaces, span, and linear dependence 8 1.3 Bases, dimensionality, and coordinates 13 1.4 Dual spaces 16 1.5 Constructions of vector spaces 20 1.6 Quotient spaces 25 1.7 Normed spaces 28 2 Linear mappings 34 2.1 Linear mappings 34 2.2 Change of basis 45 2.3 Adjoint mappings 50 2.4 Quotient mappings 53 2.5 Linear mappings from a vector space into itself 55 2.6 Norms of linear mappings 70 3 Determinants 78 3.1 Motivational examples 78 3.2 Definition and properties of determinants 88 3.3 Adjugate matrices and Cramer’s rule 102 3.4 Characteristic polynomials and Cayley–Hamilton theorem 107 4 Scalar products 115 4.1 Scalar products and basic properties 115 vii www.pdfgrip.com viii Contents 4.2 Non-degenerate scalar products 120 4.3 Positive definite scalar products 127 4.4 Orthogonal resolutions of vectors 137 4.5 Orthogonal and unitary versus isometric mappings 142 5 Real quadratic forms and self-adjoint mappings 147 5.1 Bilinear and quadratic forms 147 5.2 Self-adjoint mappings 151 5.3 Positive definite quadratic forms, mappings, and matrices 157 5.4 Alternative characterizations of positive definite matrices 164 5.5 Commutativity of self-adjoint mappings 170 5.6 Mappings between two spaces 172 6 Complex quadratic forms and self-adjoint mappings 180 6.1 Complex sesquilinear and associated quadratic forms 180 6.2 Complex self-adjoint mappings 184 6.3 Positive definiteness 188 6.4 Commutative self-adjoint mappings and consequences 194 6.5 Mappings between two spaces via self-adjoint mappings 199 7 Jordan decomposition 205 7.1 Some useful facts about polynomials 205 7.2 Invariant subspaces of linear mappings 208 7.3 Generalized eigenspaces as invariant subspaces 211 7.4 Jordan decomposition theorem 218 8 Selected topics 226 8.1 Schur decomposition 226 8.2 Classification of skewsymmetric bilinear forms 230 8.3 Perron–Frobenius theorem for positive matrices 237 8.4 Markov matrices 242 9 Excursion: Quantum mechanics in a nutshell 248 9.1 Vectors in Cn and Dirac bracket 248 9.2 Quantum mechanical postulates 252 9.3 Non-commutativity and uncertainty principle 257 9.4 Heisenberg picture for quantum mechanics 262 Solutions to selected exercises 267 Bibliographic notes 311 References 313 Index 315 www.pdfgrip.com Preface This book is concisely written to provide comprehensive core materials for a year-long course in Linear Algebra for senior undergraduate and beginning graduate students in mathematics, science, and engineering. Students who gain profound understanding and grasp of the concepts and methods of this course will acquire an essential knowledge foundation to excel in their future aca- demic endeavors. Throughout the book, methods and ideas of analysis are greatly emphasized and used, along with those of algebra, wherever appropriate, and a delicate balance is cast between abstract formulation and practical origins of various subject matters. The book is divided into nine chapters. The first seven chapters embody a traditional course curriculum. An outline of the contents of these chapters is sketched as follows. In Chapter 1 we cover basic facts and properties of vector spaces. These include definitions of vector spaces and subspaces, concepts of linear dep- endence, bases, coordinates, dimensionality, dual spaces and dual bases, quotient spaces, normed spaces, and the equivalence of the norms of a finite- dimensional normed space. In Chapter 2 we cover linear mappings between vector spaces. We start from the definition of linear mappings and discuss how linear mappings may be con- cretely represented by matrices with respect to given bases. We then introduce the notion of adjoint mappings and quotient mappings. Linear mappings from a vector space into itself comprise a special but important family of mappings and are given a separate treatment later in this chapter. Topics studied there include invariance and reducibility, eigenvalues and eigenvectors, projections, nilpotent mappings, and polynomials of linear mappings. We end the chapter with a discussion of the concept of the norms of linear mappings and use it to show that being invertible is a generic property of a linear mapping and ix www.pdfgrip.com x Preface then to show how the exponential of a linear mapping may be constructed and understood. In Chapter 3 we cover determinants. As a non-traditional but highly motivating example, we show that the calculation of the topological degree of a differentiable map from a closed curve into the unit circle in R2 involves computing a two-by-two determinant, and the knowledge gained allows us to prove the Fundamental Theorem of Algebra. We then formulate the definition of a general determinant inductively, withoutresorting to the notion of permu- tations, and establish all its properties. We end the chapter by establishing the Cayley–Hamilton theorem. Two independent proofs of this important theorem are given. The first proof is analytic and consists of two steps. In the first step, we show that the theorem is valid for a matrix of distinct eigenvalues. In the second step, we show that any matrix may be regarded as a limiting point of a sequence of matrices of distinct eigenvalues. Hence the theorem follows again by taking the limit. The second proof, on the other hand, is purely algebraic. In Chapter 4 we discuss vector spaces with scalar products. We start from the most general notion of scalar products without requiring either non-degeneracy or positive definiteness. We then carry out detailed studies on non-degenerate and positive definite scalar products, respectively, and elaborate on adjoint mappings in terms of scalar products. We end the chapter with a discussion of isometric mappings in both real and complex space settings and noting their subtle differences. In Chapter 5 we focus on real vector spaces with positive definite scalar products and quadratic forms. We first establish the main spectral theorem for self-adjoint mappings. We will not take the traditional path of first using the Fundamental Theorem of Algebra to assert that there is an eigenvalue and then applying the self-adjointness to show that the eigenvalue must be real. Instead we shall formulate an optimization problem and use calculus to prove directly that a self-adjoint mapping must have a real eigenvalue. We then present a series of characteristic conditions for a symmetric bilinear form, a symmetric matrix, or a self-adjoint mapping, to be positive definite. We end the chapter by a discussion of the commutativity of self-adjoint mappings and the useful- ness of self-adjoint mappings for the investigation of linear mappings between different spaces. In Chapter 6 we study complex vector spaces with Hermitian scalar products and related notions. Much of the theory here is parallel to that of the real space situation with the exception that normal mappings can only be fully understood and appreciated within a complex space formalism. In Chapter 7 we establish the Jordan decomposition theorem. We start with a discussion of some basic facts regarding polynomials. We next show how www.pdfgrip.com Preface xi to reduce a linear mapping over its generalized eigenspaces via the Cayley– Hamilton theorem and the prime factorization of the characteristic polynomial of the mapping. We then prove the Jordan decomposition theorem. The key and often the most difficult step in this construction is a full understanding of how a nilpotent mapping is reduced canonically. We approach this problem inductively with the degree of a nilpotent mapping and show that it is crucial to tackle a mapping of degree 2. Such a treatment eases the subtlety of the subject considerably. In Chapter 8 we present four selected topics that may be used as materi- als for some optional extra-curricular study when time and interest permit. In the first section we present the Schur decomposition theorem, which may be viewed as a complement to the Jordan decomposition theorem. In the second section we give a classification of skewsymmetric bilinear forms. In the third section we state and prove the Perron–Frobenius theorem regarding the prin- cipal eigenvalues of positive matrices. In the fourth section we establish some basic properties of the Markov matrices. In Chapter 9 we present yet another selected topic for the purpose of optional extra-curricular study: a short excursion into quantum mechanics using gadgets purely from linear algebra. Specifically we will use Cn as the state space and Hermitian matrices as quantum mechanical observables to for- mulate the over-simplified quantum mechanical postulates including Bohr’s statistical interpretation of quantum mechanics and the Schrödinger equation governing the time evolution of a state. We next establish Heisenberg’s uncer- tainty principle. Then we prove the equivalence of the Schrödinger description via the Schrödinger equation and the Heisenberg description via the Heisen- berg equation of quantum mechanics. Also provided in the book is a rich collection of mostly proof-oriented exercises to supplement and consolidate the main course materials. The diversity and elasticity of these exercises aim to satisfy the needs and inter- ests of students from a wide variety of backgrounds. At the end of the book, solutions to some selected exercises are presented. These exercises and solutions provide additional illustrative examples, extend main course materials, and render convenience for the reader to master the subjects and methods covered in a broader range. Finally some bibliographic notes conclude the book. This text may be curtailed to meet the time constraint of a semester-long course. Here is a suggested list of selected sections for such a plan: Sec- tions 1.1–1.5, 2.1–2.3, 2.5, 3.1.2, 3.2, and 3.3 (present the concept of adjugate matrices only), Section 3.4 (give the second proof of the Cayley–Hamilton the- orem only, based on an adjugate matrix expansion), Sections 4.3, 4.4, 5.1, 5.2 www.pdfgrip.com xii Preface (omit the analytic proof that a self-adjoint mapping must have an eigenvalue but resort to Exercise 5.2.1 instead), Sections 5.3, 6.1, 6.2, 6.3.1, and 7.1–7.4. Depending on the pace of lectures and time available, the instructor may decide in the later stage of the course to what extent the topics in Sections 7.1–7.4 (the Jordan decomposition) can be presented productively. The author would like to take this opportunity to thank Patrick Lin, Thomas Otway, and Robert Sibner for constructive comments and suggestions, and Roger Astley of Cambridge University Press for valuable editorial advice, which helped improve the presentation of the book. West Windsor, New Jersey Yisong Yang www.pdfgrip.com Notation and convention We use N to denote the set of all natural numbers, N = {0, 1, 2, . . . }, and Z the set of all integers, Z = {. . . ,−2,−1, 0, 1, 2, . . . }. We use i to denote the imaginary unit √−1. For a complex number c = a + ib where a, b are real numbers we use c = a − ib to denote the complex conjugate of c. We use �{c} and �{c} to denote the real and imaginary parts of the complex number c = a + ib. That is, �{c} = a, �{c} = b. We use i, j, k, l,m, n to denote integer-valued indices or space dimen- sion numbers, a, b, c scalars, u, v,w, x, y, z vectors, A,B,C,D matrices, P,R, S, T mappings, and U,V,W,X, Y,Z vector spaces, unless otherwise stated. We use t to denote the variable in a polynomial or a function or the transpose operation on a vector or a matrix. When X or Y is given, we use X ≡ Y to denote that Y , or X, is defined to be X, or Y , respectively. Occasionally, we use the symbol ∀ to express ‘for all’. Let X be a set and Y,Z subsets of X. We use Y \ Z to denote the subset of elements in Y which are not in Z. xiii www.pdfgrip.com www.pdfgrip.com 1 Vector spaces In this chapter we study vector spaces and their basic properties and structures. We start by stating the definition and a discussion of the examples of vector spaces. We next introduce the notions of subspaces, linear dependence, bases, coordinates, and dimensionality. We then consider dual spaces, direct sums, and quotient spaces. Finally we cover normed vector spaces. 1.1 Vector spaces A vector space is a non-empty set consisting of elements called vectors which can be added and multiplied by some quantities called scalars. In this section, we start with a study of vector spaces. 1.1.1 Fields The scalars to operate on vectors in a vector space are required to form a field, which may be denoted by F, where two operations, usually called addition, denoted by ‘+’, and multiplication, denoted by ‘·’ or omitted, over F are per- formed between scalars, such that the following axioms aresatisfied. (1) (Closure) If a, b ∈ F, then a + b ∈ F and ab ∈ F. (2) (Commutativity) For a, b ∈ F, there hold a + b = b + a and ab = ba. (3) (Associativity) For a, b, c ∈ F, there hold (a + b)+ c = a + (b + c) and a(bc) = (ab)c. (4) (Distributivity) For a, b, c ∈ F, there hold a(b + c) = ab + ac. (5) (Existence of zero) There is a scalar, called zero, denoted by 0, such that a + 0 = a for any a ∈ F. (6) (Existence of unity) There is a scalar different from zero, called one, denoted by 1, such that 1a = a for any a ∈ F. 1 www.pdfgrip.com 2 Vector spaces (7) (Existence of additive inverse) For any a ∈ F, there is a scalar, denoted by −a or (−a), such that a + (−a) = 0. (8) (Existence of multiplicative inverse) For any a ∈ F \ {0}, there is a scalar, denoted by a−1, such that aa−1 = 1. It is easily seen that zero, unity, additive and multiplicative inverses are all unique. Besides, a field consists of at least two elements. With the usual addition and multiplication, the sets of rational numbers, real numbers, and complex numbers, denoted by Q, R, and C, respectively, are all fields. These fields are infinite fields. However, the set of integers, Z, is not a field because there is a lack of multiplicative inverses for its non-unit elements. Let p be a prime (p = 2, 3, 5, . . . ) and set pZ = {n ∈ Z | n = kp, k ∈ Z}. Classify Z into the so-called cosets modulo pZ, that is, some non-overlapping subsets of Z represented as [i] (i ∈ Z) such that [i] = {j ∈ Z | i − j ∈ pZ}. (1.1.1) It is clear that Z is divided into exactly p cosets, [0], [1], . . . , [p − 1]. Use Zp to denote the set of these cosets and pass the additive and multiplicative operations in Z over naturally to the elements in Zp so that [i] + [j ] = [i + j ], [i][j ] = [ij ]. (1.1.2) It can be verified that, with these operations, Zp becomes a field with its obvi- ous zero and unit elements, [0] and [1]. Of course, p[1] = [1] + · · · + [1] (p terms)= [p] = [0]. In fact, p is the smallest positive integer whose multipli- cation with unit element results in zero element. A number of such a property is called the characteristic of the field. Thus, Zp is a field of characteristic p. For Q,R, and C, since no such integer exists, we say that these fields are of characteristic 0. 1.1.2 Vector spaces Let F be a field. Consider the set of n-tuples, denoted by Fn, with elements called vectors arranged in row or column forms such as⎛⎜⎜⎝ a1 ... an ⎞⎟⎟⎠ or (a1, . . . , an) where a1, . . . , an ∈ F. (1.1.3) Furthermore, we can define the addition of two vectors and the scalar multipli- cation of a vector by a scalar following the rules such as www.pdfgrip.com 1.1 Vector spaces 3 ⎛⎜⎜⎝ a1 ... an ⎞⎟⎟⎠+ ⎛⎜⎜⎝ b1 ... bn ⎞⎟⎟⎠ = ⎛⎜⎜⎝ a1 + b1 ... an + bn ⎞⎟⎟⎠ , (1.1.4) α ⎛⎜⎜⎝ a1 ... an ⎞⎟⎟⎠ = ⎛⎜⎜⎝ αa1 ... αan ⎞⎟⎟⎠ where α ∈ F. (1.1.5) The set Fn, modeled over the field F and equipped with the above operations, is a prototype example of a vector space. More generally, we say that a set U is a vector space over a field F if U is non-empty and there is an operation called addition, denoted by ‘+’, between the elements of U , called vectors, and another operation called scalar mul- tiplication between elements in F, called scalars, and vectors, such that the following axioms hold. (1) (Closure) For u, v ∈ U , we have u + v ∈ U . For u ∈ U and a ∈ F, we have au ∈ U . (2) (Commutativity) For u, v ∈ U , we have u+ v = v + u. (3) (Associativity of addition) For u, v,w ∈ U , we have u + (v + w) = (u+ v)+ w. (4) (Existence of zero vector) There is a vector, called zero and denoted by 0, such that u+ 0 = u for any u ∈ U . (5) (Existence of additive inverse) For any u ∈ U , there is a vector, denoted as (−u), such that u+ (−u) = 0. (6) (Associativity of scalar multiplication) For any a, b ∈ F and u ∈ U , we have a(bu) = (ab)u. (7) (Property of unit scalar) For any u ∈ U , we have 1u = u. (8) (Distributivity) For any a, b ∈ F and u, v ∈ U , we have (a+b)u = au+bu and a(u+ v) = au+ av. As in the case of the definition of a field, we see that it readily follows from the definition that zero vector and additive inverse vectors are all unique in a vector space. Besides, any vector multiplied by zero scalar results in zero vector. That is, 0u = 0 for any u ∈ U . Other examples of vector spaces (with obviously defined vector addition and scalar multiplication) include the following. (1) The set of all polynomials with coefficients in F defined by P = {a0 + a1t + · · · + antn | a0, a1, . . . , an ∈ F, n ∈ N}, (1.1.6) www.pdfgrip.com 4 Vector spaces where t is a variable parameter. (2) The set of all real-valued continuous functions over the interval [a, b] for a, b ∈ R and a < b usually denoted by C[a, b]. (3) The set of real-valued solutions to the differential equation an dnx dtn + · · · + a1 dx dt + a0x = 0, a0, a1, . . . , an ∈ R. (1.1.7) (4) In addition, we can also consider the set of arrays of scalars in F consisting of m rows of vectors in Fn or n columns of vectors in Fm of the form (aij ) = ⎛⎜⎜⎜⎜⎝ a11 a12 . . . a1n a21 a22 . . . a2n . . . . . . . . . . . . am1 am2 . . . amn ⎞⎟⎟⎟⎟⎠ , (1.1.8) where aij ∈ F, i = 1, . . . , m, j = 1, . . . , n, called an m by n or m × n matrix and each aij is called an entry or component of the matrix. The set of all m × n matrices with entries in F may be denoted by F(m, n). In particular, F(m, 1) or F(1, n) is simply Fm or Fn. Elements in F(n, n) are also called square matrices. 1.1.3 Matrices Here we consider some of the simplest manipulations on, and properties of, matrices. Let A be the matrix given in (1.1.8). Then At , called the transpose of A, is defined to be At = ⎛⎜⎜⎜⎜⎝ a11 a21 . . . am1 a12 a22 . . . am2 . . . . . . . . . . . . a1n a2n . . . amn ⎞⎟⎟⎟⎟⎠ . (1.1.9) Of course, At ∈ F(n,m). Simply put, At is a matrix obtained from taking the row (column) vectors of A to be its corresponding column (row) vectors. For A ∈ F(n, n), we say that A is symmetric if A = At , or skew-symmetric or anti-symmetric if At = −A. The sets of symmetric and anti-symmetric matrices are denoted by FS(n, n) and FA(n, n), respectively. It is clear that (At )t = A. www.pdfgrip.com 1.1 Vector spaces 5 It will now be useful to introduce the notion of dot product. For any two vectors u = (a1, . . . , an) and v = (b1, . . . , bn) in Fn, their dot product u·v ∈ F is defined to be u · v = a1b1 + · · · + anbn. (1.1.10) The following properties of dot product can be directly examined. (1) (Commutativity) u · v = v · u for any u, v ∈ Fn. (2) (Associativity and homogeneity) u · (av + bw) = a(u · v)+ b(u · w) for any u, v,w ∈ Fn and a, b ∈ F. With the notion of dot product, we can define the product of two matrices A ∈ F(m, k) and B ∈ F(k, n) by C = (cij ) = AB, i = 1, . . . , m, j = 1, . . . , n, (1.1.11) where cij is the dot product of the ith row of A and the j th column of B. Thus AB ∈ F(m, n). Alternatively, if we use u, v to denote column vectors in Fn, then u · v = utv. (1.1.12) That is, the dot product of u and v may be viewed as a matrix product of the 1× n matrix ut and n× 1 matrix v as well. Matrix product (or matrix multiplication) enjoys the following properties. (1) (Associativity of scalar multiplication) a(AB) = (aA)B = A(aB) for any a ∈ F and any A ∈ F(m, k), B ∈ F(k, n). (2) (Distributivity) A(B + C) = AB + AC for any A ∈ F(m, k) and B,C ∈ F(k, n); (A+B)C = AC+BC for any A,B ∈ F(m, k) and C ∈ F(k, n). (3) (Associativity) A(BC) = (AB)C for any A ∈ F(m, k), B ∈ F(k, l), C ∈ F(l, n). Alternatively, if we express A ∈ F(m, k) and B ∈ F(k, n) as made of m row vectors and n column vectors, respectively, rewritten as A = ⎛⎜⎜⎝ A1 ... Am ⎞⎟⎟⎠ , B = (B1, . . . , Bn), (1.1.13) www.pdfgrip.com 6 Vector spaces then, formally, we have AB = ⎛⎜⎜⎜⎜⎝ A1 · B1 A1 · B2 · · · A1 · Bn A2 · B1 A2 · B2 · · · A2 · Bn · · · · · · · · · · · · Am · B1 Am · B2 · · · Am · Bn ⎞⎟⎟⎟⎟⎠ = ⎛⎜⎜⎝ A1 ... Am ⎞⎟⎟⎠ (B1, . . . , Bn) = ⎛⎜⎜⎜⎜⎝ A1B1 A1B2 · · · A1Bn A2B1A2B2 · · · A2Bn · · · · · · · · · · · · AmB1 AmB2 · · · AmBn ⎞⎟⎟⎟⎟⎠ , (1.1.14) which suggests that matrix multiplication may be carried out with legitimate multiplications executed over appropriate matrix blocks. If A ∈ F(m, k) and B ∈ F(k, n), then At ∈ F(k,m) and Bt ∈ F(n, k) so that BtAt ∈ F(n,m). Regarding how AB and BtAt are related, here is the conclusion. Theorem 1.1 For A ∈ F(m, k) and B ∈ F(k, n), there holds (AB)t = BtAt . (1.1.15) The proof of this basic fact is assigned as an exercise. Other matrices in F(n, n) having interesting properties include the following. (1) Diagonal matrices are of the form A = (aij ) with aij = 0 whenever i = j . The set of diagonal matrices is denoted as FD(n, n). (2) Lower triangular matrices are of the form A = (aij ) with aij = 0 when- ever j > i. The set of lower triangular matrices is denoted as FL(n, n). (3) Upper triangular matrices are of the form A = (aij ) with aij = 0 when- ever i > j . The set of upper triangular matrices is denoted as FU(n, n). There is a special element in F(n, n), called the identity matrix, or unit matrix, and denoted by In, or simply I , which is a diagonal matrix whose diagonal entries are all 1 (unit scalar) and off-diagonal entries are all 0. For any A ∈ F(n, n), we have AI = IA = A. www.pdfgrip.com 1.1 Vector spaces 7 Definition 1.2 A matrix A ∈ F(n, n) is called invertible or nonsingular if there is some B ∈ F(n, n) such that AB = BA = I. (1.1.16) In this situation, B is unique (cf. Exercise 1.1.7) and called the inverse of A and denoted by A−1. If A,B ∈ F(n, n) are such that AB = I , then we say that A is a left inverse of B and B a right inverse of A. It can be shown that a left or right inverse is simply the inverse. In other words, if A is a left inverse of B, then both A and B are invertible and the inverses of each other. If A ∈ R(n, n) enjoys the property AAt = AtA = I , then A is called an orthogonal matrix. For A = (aij ) ∈ C(m, n), we adopt the notation A = (aij ) for taking the complex conjugate of A and use A† to denote taking the complex conjugate of the transpose of A, A† = At , which is also commonly referred to as taking the Hermitian conjugate of A. If A ∈ C(n, n), we say that A is Hermitian symmetric, or simply Hermitian, if A† = A, and skew-Hermitian or anti-Hermitian, if A† = −A. If A ∈ C(n, n) enjoys the property AA† = A†A = I , then A is called a unitary matrix. We will see the importance of these notions later. Exercises 1.1.1 Show that it follows from the definition of a field that zero, unit, additive, and multiplicative inverse scalars are all unique. 1.1.2 Let p ∈ N be a prime and [n] ∈ Zp. Find −[n] and prove the existence of [n]−1 when [n] = [0]. In Z5, find −[4] and [4]−1. 1.1.3 Show that it follows from the definition of a vector space that both zero and additive inverse vectors are unique. 1.1.4 Prove the associativity of matrix multiplication by showing that A(BC) = (AB)C for any A ∈ F(m, k), B ∈ F(k, l), C ∈ F(l, n). 1.1.5 Prove Theorem 1.1. 1.1.6 Let A ∈ F(n, n) (n ≥ 2) and rewrite A as A = ( A1 A2 A3 A4 ) , (1.1.17) where A1 ∈ F(k, k), A2 ∈ F(k, l), A3 ∈ F(l, k), A4 ∈ F(l, l), k, l ≥ 1, k + l = n. Show that At = ( At1 A t 3 At2 A t 4 ) . (1.1.18) www.pdfgrip.com 8 Vector spaces 1.1.7 Prove that the inverse of an invertible matrix is unique by showing the fact that if A,B,C ∈ F(n, n) satisfy AB = I and CA = I then B = C. 1.1.8 Let A ∈ C(n, n). Show that A is Hermitian if and only if iA is anti- Hermitian. 1.2 Subspaces, span, and linear dependence Let U be a vector space over a field F and V ⊂ U a non-empty subset of U . We say that V is a subspace of U if V is a vector space over F with the inherited addition and scalar multiplication from U . It is worth noting that, when checking whether a subset V of a vector space U becomes a sub- space, one only needs to verify the closure axiom (1) in the definition of a vector space since the rest of the axioms follow automatically as a conse- quence of (1). The two trivial subspaces of U are those consisting only of zero vector, {0}, and U itself. A nontrivial subspace is also called a proper subspace. Consider the subset Pn (n ∈ N) of P defined by Pn = {a0 + a1t + · · · + antn | a0, a1, . . . , an ∈ F}. (1.2.1) It is clear that Pn is a subspace of P and Pm is subspace of Pn when m ≤ n. Consider the set Sa of all vectors (x1, . . . , xn) in F n satisfying the equation x1 + · · · + xn = a, (1.2.2) where a ∈ F. Then Sa is a subspace of Fn if and only if a = 0. Let u1, . . . , uk be vectors in U . The linear span of {u1, . . . , uk}, denoted by Span{u1, . . . , uk}, is the subspace of U defined by Span{u1, . . . , uk} = {u ∈ U | u = a1u1 + · · · + akuk, a1, . . . , ak ∈ F}. (1.2.3) Thus, if u ∈ Span{u1, . . . , uk}, then there are a1, . . . , ak ∈ F such that u = a1u1 + · · · + akuk. (1.2.4) We also say that u is linearly spanned by u1, . . . , uk or linearly dependent on u1, . . . , uk . Therefore, zero vector 0 is linearly dependent on any finite set of vectors. If U = Span{u1, . . . , uk}, we also say that U is generated by the vectors u1, . . . , uk . For Pn defined in (1.2.1), we have Pn = Span{1, t, . . . , tn}. Thus Pn is generated by 1, t, . . . , tn. Naturally, for two elements p, q in www.pdfgrip.com 1.2 Subspaces, span, and linear dependence 9 Pn, say p(t) = a0 + a1t + · · · + antn, q(t) = b0 + b1t + · · · + bntn, we identify p and q if and only if all the coefficients of p and q of like powers of t coincide in F, or, ai = bi for all i = 0, 1, . . . , n. In Fn, define e1 = (1, 0, 0, . . . , 0), e2 = (0, 1, 0, . . . , 0), en = (0, 0, . . . , 0, 1). (1.2.5) Then Fn = Span{e1, e2, . . . , en} and Fn is generated by e1, e2, . . . , en. Thus, for S0 defined in (1.2.2), we have (x1, x2, . . . , xn) = −(x2 + · · · + xn)e1 + x2e2 + · · · + xnen = x2(e2 − e1)+ · · · + xn(en − e1), (1.2.6) where x2, . . . , xn are arbitrarily taken from F. Therefore S0 = Span{e2 − e1, . . . , en − e1}. (1.2.7) For F(m, n), we define Mij ∈ F(m, n) to be the vector such that all its entries vanish except that its entry at the position (i, j) (at the ith row and j th column) is 1, i = 1, . . . , m, j = 1, . . . , n. We have F(m, n) = Span{Mij | i = 1, . . . , m, j = 1, . . . , n}. (1.2.8) The notion of spans can be extended to cover some useful situations. Let U be a vector space and S be a (finite or infinite) subset of U . Define Span(S) = the set of linear combinations of all possible finite subsets of S. (1.2.9) It is obvious that Span(S) is a subspace of U . If U = Span(S), we say that U is spanned or generated by the set of vectors S. As an example, we have P = Span{1, t, . . . , tn, . . . }. (1.2.10) Alternatively, we can also express P as P = ∪∞n=0Pn. (1.2.11) The above discussion motivates the following formal definition. Definition 1.3 Let u1, . . . , um be m vectors in the vector space U over a field F. We say that these vectors are linearly dependent if one of them may be written as a linear span of the rest of them or linearly dependent on the rest www.pdfgrip.com 10 Vector spaces of them. Or equivalently, u1, . . . , um are linearly dependent if there are scalars a1, . . . , am ∈ F where (a1, . . . , am) = (0, . . . , 0) such that a1u1 + · · · + amum = 0. (1.2.12) Otherwise u1, . . . , um are called linearly independent. In this latter case, the only possible vector (a1, . . . , am) ∈ Fn to make (1.2.12) fulfilled is the zero vector, (0, . . . , 0). To proceed further, we need to consider the following system of linear equations ⎧⎪⎨⎪⎩ a11x1 + · · · + a1nxn = 0, . . . . . . . . . . . . . . . . . . . . . . . . am1x1 + · · · + amnxn = 0, (1.2.13) over F with unknowns x1, . . . , xn. Theorem 1.4 In the system (1.2.13), if m < n, then the system has a nontrivial solution (x1, . . . , xn) = (0, . . . , 0). Proof We prove the theorem by using induction on m+ n. The beginning situation is m+n = 3 when m = 1 and n = 2. It is clear that we always have a nontrivial solution.Assume that the statement of the theorem is true when m + n ≤ k where k ≥ 3. Let m+n = k+1. If k = 3, the condition m < n implies m = 1, n = 3 and the existence of a nontrivial solution is obvious. Assume then k ≥ 4. If all the coefficients of the variable x1 in (1.2.13) are zero, i.e. a11 = · · · = am1 = 0, then x1 = 1, x2 = · · · = xn = 0 is a nontrivial solution. So we may assume one of the coefficients of x1 is nonzero. Without loss of generality, we assume a11 = 0. If m = 1, there is again nothing to show. Assume m ≥ 2. Dividing the first equation in (1.2.13) by a11 if necessary, we can further assume a11 = 1. Then, adding the (−ai1)-multiple of the first equation into the ith equation, in (1.2.13), for i = 2, . . . , m, we arrive at⎧⎪⎪⎪⎪⎨⎪⎪⎪⎪⎩ x1+ a12x2 + · · · + a1nxn = 0, b22x2 + · · · + b2nxn = 0, . . . . . . . . . . . . . . . . . . . . . . . . bm2x2 + · · · + bmnxn = 0. (1.2.14) The system below the first equation in (1.2.14) contains m − 1 equations and n − 1 unknowns x2, . . . , xn. Of course, m − 1 < n − 1. So, in view of the www.pdfgrip.com 1.2 Subspaces, span, and linear dependence 11 inductive assumption, it has a nontrivial solution. Substituting this nontrivial solution into the first equation in (1.2.14) to determine the remaining unknown x1, we see that the existence of a nontrivial solution to the original system (1.2.13) follows. The importance of Theorem 1.4 is seen in the following. Theorem 1.5 Any set of more than m vectors in Span{u1, . . . , um} must be linearly dependent. Proof Let v1, . . . , vn ∈ Span{u1, . . . , um} be n vectors where n > m. Consider the possible linear dependence relation x1v1 + · · · + xnvn = 0, (1.2.15) for some x1, . . . , xn ∈ F. Since each vj ∈ Span{u1, . . . , um}, j = 1, . . . , n, there are scalars aij ∈ F (i = 1, . . . , m, j = 1, . . . , n) such that vj = m∑ i=1 aijui, j = 1, . . . , n. (1.2.16) Substituting (1.2.16) into (1.2.15), we have n∑ j=1 xj ( m∑ i=1 aijui ) = 0. (1.2.17) Regrouping the terms in the above equation, we arrive at m∑ i=1 ⎛⎝ n∑ j=1 aij xj ⎞⎠ ui = 0, (1.2.18) which may be fulfilled by setting n∑ j=1 aij xj = 0, i = 1, . . . , m. (1.2.19) This system of equations is exactly the system (1.2.13) which allows a non- trivial solution in view of Theorem 1.4. Hence the proof follows. We are now prepared to study in the next section several fundamental prop- erties of vector spaces. www.pdfgrip.com 12 Vector spaces Exercises 1.2.1 Let U1 and U2 be subspaces of a vector space U . Show that U1 ∪U2 is a subspace of U if and only if U1 ⊂ U2 or U2 ⊂ U1. 1.2.2 Let Pn denote the vector space of the polynomials of degrees up to n over a field F expressed in terms of a variable t . Show that the vectors 1, t, . . . , tn in Pn are linearly independent. 1.2.3 Show that the vectors in Fn defined in (1.2.5) are linearly independent. 1.2.4 Show that S0 defined in (1.2.2) may also be expressed as S0 = Span{e1 − en, . . . , en−1 − en}, (1.2.20) and deduce that, in Rn, the vectors( 1, 1 2 , . . . , 1 2n−2 , 2 [ 1 2n−1 −1 ]) , e1−en, . . . , en−1 − en, (1.2.21) are linearly dependent (n ≥ 4). 1.2.5 Show that FS(n, n), FA(n, n), FD(n, n), FL(n, n), and FU(n, n) are all subspaces of F(n, n). 1.2.6 Let u1, . . . , un (n ≥ 2) be linearly independent vectors in a vector space U and set vi−1 = ui−1 + ui, i = 2, . . . , n; vn = un + u1. (1.2.22) Investigate whether v1, . . . , vn are linearly independent as well. 1.2.7 Let F be a field. For any two vectors u = (a1, . . . , an) and v = (b1, . . . , bn) in F n (n ≥ 2), viewed as matrices, we see that the ma- trix product A = utv lies in F(n, n). Prove that any two row vectors of A are linearly dependent. What happens to the column vectors of A? 1.2.8 Consider a slightly strengthened version of the second part of Exercise 1.2.1 above: Let U1, U2 be subspaces of a vector space U , U1 = U , U2 = U . Show without using Exercise 1.2.1 that there exists a vec- tor in U which lies outside U1 ∪ U2. Explain how you may apply the conclusion of this exercise to prove that of Exercise 1.2.1. 1.2.9 (A challenging extension of the previous exercise) Let U1, . . . , Uk be k subspaces of a vector space U over a field of characteristic 0. If Ui = U for i = 1, . . . , k, show that there is a vector in U which lies outside ∪ki=1Ui . 1.2.10 For A ∈ F(m, n) and B ∈ F(n,m) with m > n show that AB as an element in F(m,m) can never be invertible. www.pdfgrip.com 1.3 Bases, dimensionality, and coordinates 13 1.3 Bases, dimensionality, and coordinates Let U be a vector space over a field F, take u1, . . . , un ∈ U , and set V = Span{u1, . . . , un}. Eliminating linearly dependent vectors from the set {u1, . . . , un} if necessary, we can certainly assume that the vectors u1, . . . , un are already made linearly independent. Thus, any vector u ∈ V may take the form u = a1u1 + · · · + anun, a1, . . . , an ∈ F. (1.3.1) It is not hard to see that the coefficients a1, . . . , an in the above representation must be unique. In fact, if we also have u = b1u1 + · · · + bnun, b1, . . . , bn ∈ F, (1.3.2) then, combining the above two relations, we have (a1 − b1)u1 + · · · + (an − bn)un = 0. Since u1, . . . , un are linearly independent, we obtain a1 = b1, . . . , an = bn and the uniqueness follows. Furthermore, if there is another set of vectors v1, . . . , vm in U such that Span{v1, . . . , vm} = Span{u1, . . . , un}, (1.3.3) then m ≥ n in view of Theorem 1.5. As a consequence, if v1, . . . , vm are also linearly independent, then m = n. This observation leads to the following. Definition 1.6 If there are linearly independent vectors u1, . . . , un ∈ U such that U = Span{u1, . . . , un}, then U is said to be finitely generated and the set of vectors {u1, . . . , un} is called a basis of U . The number of vectors in any basis of a finitely generated vector space, n, is independent of the choice of the basis and is referred to as the dimension of the finitely generated vector space, written as dim(U) = n. A finitely generated vector space is also said to be of finite dimensionality or finite dimensional. If a vector space U is not finite dimensional, it is said to be infinite dimensional, also written as dim(U) = ∞. As an example of an infinite-dimensional vector space, we show that when R is regarded as a vector space over Q, then dim(R) = ∞. In fact, recall that a real number is called an algebraic number if it is the zero of a polyno- mial with coefficients in Q. We also know that there are many non-algebraic numbers in R, called transcendental numbers. Let τ be such a transcenden- tal number. Then for any n = 1, 2, . . . the numbers 1, τ, τ 2, . . . , τ n are lin- early independent in the vector space R over the field Q. Indeed, if there are r0, r1, r2, . . . , rn ∈ Q so that r0 + r1τ + r2τ 2 + · · · + rnτn = 0, (1.3.4) www.pdfgrip.com 14 Vector spaces and at least one number among r0, r1, r2, . . . , rn is nonzero, then τ is the zero of the nontrivial polynomial p(t) = r0 + r1t + r2t2 + · · · + rntn, (1.3.5) which violates the assumption that τ is transcendental. Thus R is infinite dimensional over Q. The following theorem indicates that it is fairly easy to construct a basis for a finite-dimensional vector space. Theorem 1.7 Let U be an n-dimensional vector space over a field F. Any n linearly independent vectors in U form a basis of U . Proof Let u1, . . . , un ∈ U be linearly independent vectors. We only need to show that they span U . In fact, take any u ∈ U . We know that u1, . . . , un, u are linearly dependent. So there is a nonzero vector (a1, . . . , an, a) ∈ Fn+1 such that a1u1 + · · · + anun + au = 0. (1.3.6) Of course, a = 0, otherwise it contradicts the assumption that u1, . . . , un are linearly independent. So u = (−a−1)(a1u1 + · · · + anun). Thus u ∈ Span{u1, . . . , un}. Definition 1.8 Let {u1, . . . , un} be a basis of the vector space U . Given u ∈ U there are unique scalars a1, . . . , an ∈ F such that u = a1u1 + · · · + anun.(1.3.7) These scalars, a1, . . . , an are called the coordinates, and (a1, . . . , an) ∈ Fn the coordinate vector, of the vector u with respect to the basis {u1, . . . , un}. It will be interesting to investigate the relation between the coordinate vec- tors of a vector under different bases. Let U = {u1, . . . , un} and V = {v1, . . . , vn} be two bases of the vector space U . For u ∈ U , let (a1, . . . , an) ∈ Fn and (b1, . . . , bn) ∈ Fn be the coordinate vectors of u with respect to U and V , respectively. Thus u = a1u1 + · · · + anun = b1v1 + · · · + bnvn. (1.3.8) On the other hand, we have vj = n∑ i=1 aijui, j = 1, . . . , n. (1.3.9) www.pdfgrip.com 1.3 Bases, dimensionality, and coordinates 15 The n× n matrix A = (aij ) is called a basis transition matrix or basis change matrix. Inserting (1.3.9) into (1.3.8), we have n∑ i=1 aiui = n∑ i=1 ⎛⎝ n∑ j=1 aij bj ⎞⎠ ui. (1.3.10) Hence, by the linear independence of the basis vectors, we have ai = n∑ j=1 aij bj , i = 1, . . . , n. (1.3.11) Note that the relation (1.3.9) between bases may be formally and conve- niently expressed in a ‘matrix form’ as (v1, . . . , vn) = (u1, . . . , un)A, (1.3.12) or concisely V = UA, or ⎛⎜⎜⎝ v1 ... vn ⎞⎟⎟⎠ = At ⎛⎜⎜⎝ u1 ... un ⎞⎟⎟⎠ , (1.3.13) where multiplications between scalars and vectors are made in a well defined manner. On the other hand, the relation (1.3.11) between coordinate vectors may be rewritten as ⎛⎜⎜⎝ a1 ... an ⎞⎟⎟⎠ = A ⎛⎜⎜⎝ b1 ... bn ⎞⎟⎟⎠ , (1.3.14) or (a1, . . . , an) = (b1, . . . , bn)At . (1.3.15) Exercises 1.3.1 Let U be a vector space with dim(U) = n ≥ 2 and V a subspace of U with a basis {v1, . . . , vn−1}. Prove that for any u ∈ U \ V the vectors u, v1, . . . , vn−1 form a basis for U . 1.3.2 Show that dim(F(m, n)) = mn. 1.3.3 Determine dim(FS(n, n)), dim(FA(n, n)), and dim(FD(n, n)). 1.3.4 Let P be the vector space of all polynomials with coefficients in a field F. Show that dim(P) = ∞. www.pdfgrip.com 16 Vector spaces 1.3.5 Consider the vector space R3 and the bases U = {e1, e2, e3} and V = {e1, e1 + e2, e1 + e2 + e3}. Find the basis transition matrix A from U into V satisfying V = UA. Find the coordinate vectors of the given vector (1, 2, 3) ∈ R3 with respect to the bases U and V , respectively, and relate these vectors with the matrix A. 1.3.6 Prove that a basis transition matrix must be invertible. 1.3.7 Let U be an n-dimensional vector space over a field F where n ≥ 2 (say). Consider the following construction. (i) Take u1 ∈ U \ {0}. (ii) Take u2 ∈ U \ Span{u1}. (iii) Take (if any) u3 ∈ U \ Span{u1, u2}. (iv) In general, take ui ∈ U \ Span{u1, . . . , ui−1} (i ≥ 2). Show that this construction will terminate itself in exactly n steps, that is, it will not be possible anymore to get un+1, and that the vectors u1, u2, . . . , un so obtained form a basis of U . 1.4 Dual spaces Let U be an n-dimensional vector space over a field F. A functional (also called a form or a 1-form) f over U is a linear function f : U → F satisfying f (u+ v) = f (u)+ f (v), u, v ∈ U ; f (au) = af (u), a ∈ F, u ∈ U. (1.4.1) Let f, g be two functionals. Then we can define another functional called the sum of f and g, denoted by f + g, by (f + g)(u) = f (u)+ g(u), u ∈ U. (1.4.2) Similarly, let f be a functional and a ∈ F. We can define another functional called the scalar multiple of a with f , denoted by af , by (af )(u) = af (u), u ∈ U. (1.4.3) It is a simple exercise to check that these two operations make the set of all functionals over U a vector space over F. This vector space is called the dual space of U , denoted by U ′. Let {u1, . . . , un} be a basis of U . For any f ∈ U ′ and any u = a1u1+ · · · + anun ∈ U , we have f (u) = f ( n∑ i=1 aiui ) = n∑ i=1 aif (ui). (1.4.4) www.pdfgrip.com 1.4 Dual spaces 17 Hence, f is uniquely determined by its values on the basis vectors, f1 = f (u1), . . . , fn = f (un). (1.4.5) Conversely, for arbitrarily assigned values f1, . . . , fn in (1.4.5), we define f (u) = n∑ i=1 aifi for any u = n∑ i=1 aiui ∈ U. (1.4.6) It is clear that f is a functional. That is, f ∈ U ′. Of course, such an f satisfies (1.4.5). Thus, we have seen a well-defined 1-1 correspondence U ′ ↔ Fn, f ↔ (f1, . . . , fn). (1.4.7) Especially we may use u′1, . . . , u′n to denote the elements in U ′ correspond- ing to the vectors e1, . . . , en in F n given by (1.2.5). Then we have u′i (uj ) = δij = { 0, i = j, 1, i = j, i, j = 1, . . . , n. (1.4.8) It is clear that u′1, . . . , u′n are linearly independent and span U ′ because an element f of U ′ satisfying (1.4.5) is simply given by f = f1u′1 + · · · + fnu′n. (1.4.9) In other words, {u′1, . . . , u′n} is a basis of U ′, commonly called the dual basis of U ′ with respect to the basis {u1, . . . , un} of U . In particular, we have seen that U and U ′ are of the same dimensionality. Let U = {u1, . . . , un} and V = {v1, . . . , vn} be two bases of the vec- tor space U . Let their dual bases be denoted by U ′ = {u′1, . . . , u′n} and V ′ = {v′1, . . . , v′n}, respectively. Suppose that the bases U ′ and V ′ are related through u′j = n∑ i=1 a′ij v′i , j = 1, . . . , n. (1.4.10) Using (1.3.9) and (1.4.10) to evaluate u′i (vj ), we obtain u′i (vj ) = u′i ( n∑ k=1 akjuk ) = n∑ k=1 akju ′ i (uk) = n∑ k=1 akj δik = aij , (1.4.11) u′i (vj ) = n∑ k=1 a′kiv′k(vj ) = n∑ k=1 a′kiδkj = a′j i , (1.4.12) www.pdfgrip.com 18 Vector spaces which leads to a′ij = aji (i, j = 1, . . . , n). In other words, we have arrived at the correspondence relation u′j = n∑ i=1 ajiv ′ i , j = 1, . . . , n. (1.4.13) With matrix notation as before, we have (u′1, . . . , u′n) = (v′1, . . . , v′n)At , (1.4.14) or ⎛⎜⎜⎝ u′1 ... u′n ⎞⎟⎟⎠ = A ⎛⎜⎜⎝ v′1 ... v′n ⎞⎟⎟⎠ . (1.4.15) Besides, for any u′ ∈ U ′ written as u′ = a′1u′1 + · · · + a′nu′n = b′1v′1 + · · · + b′nv′n, (1.4.16) the discussion in the previous section and the above immediately allow us to get b′i = n∑ j=1 ajia ′ j , i = 1, . . . , n. (1.4.17) Thus, in matrix form, we obtain the relation (b′1 . . . , b′n) = (a′1 . . . , a′n)A, (1.4.18) or ⎛⎜⎜⎝ b′1 ... b′n ⎞⎟⎟⎠ = At ⎛⎜⎜⎝ a′1 ... a′n ⎞⎟⎟⎠ . (1.4.19) Comparing the above results with those established in the previous section, we see that, with respect to bases and dual bases, the coordinates vectors in U and U ′ follow ‘opposite’ rules of correspondence. For this reason, coordinate vectors in U are often called covariant vectors, and those in U ′ contravariant vectors. Using the relation stated in (1.4.8), we see that we may naturally view u1, . . . , un as elements in (U ′)′ = U ′′ so that they form a basis of U ′′ dual to {u′1, . . . , u′n} since www.pdfgrip.com 1.4 Dual spaces 19 ui(u ′ j ) ≡ u′j (ui) = δij = { 0, i = j, 1, i = j, i, j = 1, . . . , n. (1.4.20) Thus, for any u ∈ U ′′ satisfying u(u′i ) = ai (i = 1, . . . , n), we have u = a1u1 + · · · + anun. (1.4.21) In this way, we see that U ′′ may be identified with U . In other words, we have seen that the identification just made spells out the relationship U ′′ = U, (1.4.22) which is also referred to as reflectivity of U or U is said to be reflective. Notationally, for u ∈ U and u′ ∈ U ′, it is often convenient to rewrite u′(u), which is linear in both the u and u′ arguments, as u′(u) = 〈u′, u〉. (1.4.23) Then our identification (1.4.22), made through setting u′(u) = u(u′), u ∈ U, u′ ∈ U ′, (1.4.24) simply says that the ‘pairing’ 〈·, ·〉 as given in (1.4.23) is symmetric: 〈u′, u〉 = 〈u, u′〉, u ∈ U, u′ ∈ U ′. (1.4.25) For any non-empty subset S ⊂ U , the annihilator of S, denoted by S0, is the subset of U ′ given by S0 = {u′ ∈ U ′ | 〈u′, u〉 = 0,∀u ∈ S}. (1.4.26) It is clear that S0 is always a subspace of U ′ regardless of whether S is a subspace of U . Likewise, for any nonempty subset S′ ⊂ U ′, we can define the annihilator of S′, S′0, as the subset S′0 = {u ∈ U | 〈u′, u〉 = 0,∀u′ ∈ S′} (1.4.27) of U . Of course, S′0 is always a subspace of U . Exercises 1.4.1 Let F be a field. Describe thedual spaces F′ and (F2)′. 1.4.2 Let U be a finite-dimensional vector space. Prove that for any vectors u, v ∈U (u = v) there exists an element f ∈U ′ such that f (u) = f (v). 1.4.3 Let U be a finite-dimensional vector space and f, g ∈ U ′. For any v ∈ U , f (v) = 0 if and only if g(v) = 0. Show that f and g are linearly dependent. www.pdfgrip.com 20 Vector spaces 1.4.4 Let F be a field and V = {(x1, . . . , xn) ∈ Fn | x1 + · · · + xn = 0}. (1.4.28) Show that any f ∈ V 0 may be expressed as f (x1, . . . , xn) = c n∑ i=1 xi, (x1, . . . , xn) ∈ Fn, (1.4.29) for some c ∈ F. 1.4.5 Let U = P2 and f, g, h ∈ U ′ be defined by f (p) = p(−1), g(p) = p(0), h(p) = p(1), p(t) ∈ P2. (1.4.30) (i) Show that B′ = {f, g, h} is a basis for U ′. (ii) Find a basis B of U which is dual to B′. 1.4.6 Let U be an n-dimensional vector space and V be an m-dimensional subspace of U . Show that the annihilator V 0 is an (n−m)-dimensional subspace of U ′. In other words, there holds the dimensionality equation dim(V )+ dim(V 0) = dim(U). (1.4.31) 1.4.7 Let U be an n-dimensional vector space and V be an m-dimensional subspace of U . Show that V 00= (V 0)0=V . 1.5 Constructions of vector spaces Let U be a vector space and V,W its subspaces. It is clear that V ∩W is also a subspace of U but V ∪ W in general may fail to be a subspace of U . The smallest subspace of U that contains V ∪W should contain all vectors in U of the form v + w where v ∈ V and w ∈ W . Such an observation motivates the following definition. Definition 1.9 If U is a vector space and V,W its subspaces, the sum of V and W , denoted by V +W , is the subspace of U given by V +W ≡ {u ∈ U | u = v + w, v ∈ V,w ∈ W }. (1.5.1) Checking that V +W is a subspace of U that is also the smallest subspace of U containing V ∪W will be assigned as an exercise. www.pdfgrip.com 1.5 Constructions of vector spaces 21 Now let B0 = {u1, . . . , uk} be a basis of V ∩W . Expand it to obtain bases for V and W , respectively, of the forms BV ={u1, . . . , uk, v1, . . . , vl}, BW = {u1, . . . , uk, w1, . . . , wm}. (1.5.2) From the definition of V +W , we get V +W = Span{u1, . . . , uk, v1, . . . , vl, w1, . . . , wm}. (1.5.3) We can see that {u1, . . . , uk, v1, . . . , vl, w1, . . . , wm} is a basis of V +W . In fact, we only need to show that the vectors u1, . . . , uk, v1, . . . , vl, w1, . . . , wm are linearly independent. For this purpose, consider the linear relation a1u1 + · · · + akuk + b1v1 + · · · + blvl + c1w1 + · · · + cmwm=0, (1.5.4) where a1, . . . , ak, b1, . . . , bl, c1, . . . , cm are scalars. We claim that w = c1w1 + · · · + cmwm = 0. (1.5.5) Otherwise, using (1.5.4) and (1.5.5), we see that w ∈ V . However, we already have w ∈ W . So w ∈ V ∩W , which is false since u1, . . . , uk, w1, . . . , wm are linearly independent. Thus (1.5.5) follows and c1 = · · · = cm = 0. Applying (1.5.5) in (1.5.4), we immediately have a1 = · · · = ak = b1 = · · · = bl = 0. Therefore we can summarize the above discussion by concluding with the following theorem. Theorem 1.10 The following general dimensionality formula dim(V +W) = dim(V )+ dim(W)− dim(V ∩W) (1.5.6) is valid for the sum of any two subspaces V and W of finite dimensions in a vector space. Of great importance is the situation when dim(V ∩W) = 0 or V ∩W = {0}. In this situation, the sum is called direct sum, and rewritten as V ⊕W . Thus, we have dim(V ⊕W) = dim(V )+ dim(W). (1.5.7) Direct sum has the following characteristic. Theorem 1.11 The sum of two subspaces V and W of U is a direct sum if and only if each vector u in V +W may be expressed as the sum of a unique vector v ∈ V and a unique vector w ∈ W . www.pdfgrip.com 22 Vector spaces Proof Suppose first V ∩W = {0}. For any u ∈ V +W , assume that it may be expressed as u = v1 + w1 = v2 + w2, v1, v2 ∈ V, w1, w2 ∈ W. (1.5.8) From (1.5.8), we have v1 − v2 = w2 −w1 which lies in V ∩W . So v1 − v2 = w2 − w1 = 0 and the stated uniqueness follows. Suppose that any u ∈ V + W can be expressed as u = v + w for some unique v ∈ V and w ∈ W . If V ∩W = {0}, take x ∈ V ∩W with x = 0. Then zero vector 0 may be expressed as 0 = x + (−x) with x ∈ V and (−x) ∈ W , which violates the stated uniqueness since 0 = 0 + 0 with 0 ∈ V and 0 ∈ W , as well. Let V be a subspace of an n-dimensional vector space U and BV = {v1, . . . , vk} be a basis of V . Extend BV to get a basis of U , say {v1, . . . , vk, w1, . . . , wl}, where k + l = n. Define W = Span{w1, . . . , wl}. (1.5.9) Then we obtain U = V ⊕W . The subspace W is called a linear complement, or simply complement, of V in U . Besides, the subspaces V and W are said to be mutually complementary in U . We may also build up a vector space from any two vector spaces, say V and W , over the same field F, as a direct sum of V and W . To see this, we construct vectors of the form u = (v,w), v ∈ V, w ∈ W, (1.5.10) and define vector addition and scalar multiplication component-wise by u1 + u2 = (v1, w1)+ (v2, w2) = (v1 + v2, w2 + w2), v1, v2 ∈ V, w1, w2 ∈ W, (1.5.11) au = a(v,w) = (aw, av), v ∈ V, w ∈ W, a ∈ F. (1.5.12) It is clear that the set U of all vectors of the form (1.5.10) equipped with the vector addition (1.5.11) and scalar multiplication (1.5.12) is a vector space over F. Naturally we may identify V and W with the subspaces of U given by Ṽ = {(v, 0) | v ∈ V }, W̃ = {(0, w) |w ∈ W }. (1.5.13) Of course, U = Ṽ ⊕W̃ . Thus, in a well-understood sense, we may also rewrite this relation as U = V ⊕W as anticipated. Sometimes the vector space U so constructed is also referred to as the direct product of V and W and rewritten as U = V ×W . In this way, R2 may naturally be viewed as R×R, for example. www.pdfgrip.com 1.5 Constructions of vector spaces 23 More generally, let V1, . . . , Vk be any k subspaces of a vector space U . In a similar manner we can define the sum V = V1 + · · · + Vk, (1.5.14) which is of course a subspace of U . Suggested by the above discussion, if each v ∈ V may be written as v = v1 + · · · + vk for uniquely determined v1 ∈ V1, . . . , vk ∈ Vk , then we say that V is the direct sum of V1, . . . , Vk and rewrite such a relation as V = V1 ⊕ · · · ⊕ Vk. (1.5.15) It should be noted that, when k ≥ 3, extra caution has to be exerted when checking whether the sum (1.5.14) is a direct sum. For example, the naive condition Vi ∩ Vj = {0}, i = j, i, j = 1, . . . , k, (1.5.16) among V1, . . . , Vk , is not sufficient anymore to ensure (1.5.15). To illustrate this subtlety, let us consider V = F2 and take V1 = Span {( 1 0 )} , V2 = Span {( 1 1 )} , V3 = Span {( 1 −1 )} . (1.5.17) It is clear that V1, V2, V3 satisfy (1.5.16) and V = V1 + V2 + V3 but V cannot be a direct sum of V1, V2, V3. In fact, in such a general situation, a correct condition which replaces (1.5.16) is Vi ∩ ⎛⎝ ∑ 1≤j≤k,j =i Vj ⎞⎠ = {0}, i = 1, . . . , k. (1.5.18) In other words, when the condition (1.5.18) is fulfilled, then (1.5.15) is valid. The proof of this fact is left as an exercise. Exercises 1.5.1 For U = {(x1, . . . , xn) ∈ Fn| x1 + · · · + xn = 0}, V = {(x1, . . . , xn) ∈ Fn | x1 = · · · = xn}, (1.5.19) prove that Fn = U ⊕ V . www.pdfgrip.com 24 Vector spaces 1.5.2 Consider the vector space of all n×n matrices over a field F, denoted by F(n, n). As before, use FS(n, n) and FA(n, n) to denote the subspaces of symmetric and anti-symmetric matrices. Assume that the characteristic of F is not equal to 2. For any M ∈ F(n, n), rewrite M as M = 1 2 (M +Mt)+ 1 2 (M −Mt). (1.5.20) Check that 1 2 (M +Mt) ∈ FS(n, n) and 1 2 (M −Mt) ∈ FA(n, n). Use this fact to prove the decomposition F(n, n) = FS(n, n)⊕ FA(n, n). (1.5.21) What happens when the characteristic of F is 2 such as when F = Z2? 1.5.3 Show that FL(n, n) ∩ FU(n, n) = FD(n, n). 1.5.4 Use FL(n, n) and FU(n, n) in F(n, n) to give an example for the dimen- sionality relation (1.5.6). 1.5.5 Let X = C[a, b] (a, b ∈ R and a < b) be the vector space of all real- valued continuous functions over the interval [a,b] and Y = { f ∈ X ∣∣∣∣ ∫ b a f (t) dt = 0 } . (1.5.22) (i) Prove that if R is identified with the set of all constant functions over [a, b] then X = R⊕ Y . (ii) For a = 0, b = 1, and f (t) = t2 + t − 1, find the unique c ∈ R and g ∈ Y such that f (t) = c + g(t) for all t ∈ [0, 1]. 1.5.6 Let U be a vector space and V,W its subspaces such that U = V +W . If X is subspace of U , is it true that X = (X ∩ V )+ (X ∩W)? 1.5.7 Let U be a vector space and V,W,X some subspaces of U such that U = V ⊕W, U = V ⊕X. (1.5.23) Can one infer W = X from the condition (1.5.23)? Explain why or why not. 1.5.8 Let V1, . . . , Vk be some subspaces of U and set V = V1 + · · · + Vk . Show that this sum is a direct sum if and only if one of the following statements is true. (i) V1, . . . , Vk satisfy (1.5.18). (ii) If non-overlapping sets of vectors {v11, . . . , v1l1}, . . . , {vk1, . . . , vklk } (1.5.24) www.pdfgrip.com 1.6 Quotient spaces 25 are bases of V1, . . . , Vk , respectively, then the union {v11, . . . , v1l1} ∪ · · · ∪ {vk1, . . . , vklk } (1.5.25) is a basis of V . (iii) There holds the dimensionality relation dim(V ) = dim(V1)+ · · · + dim(Vk). (1.5.26) 1.6 Quotient spaces In order to motivate the introduction of the concept of quotient spaces, we first consider a concrete example in R2. Let v ∈ R2 be any nonzero vector. Then V = Span{v} (1.6.1) represents the line passing through the origin and along (or opposite to) the direction of v. More generally, for any u ∈ R2, the coset [u] = {u+ w |w ∈ V } = {x ∈ R2 | x − u ∈ V } (1.6.2) represents the line passing through the vector u and parallel to the vector v. Naturally, we define [u1] + [u2] = {x + y | x ∈ [u1], y ∈ [u2]} and claim [u1] + [u2] = [u1 + u2]. (1.6.3) In fact, let z ∈ [u1] + [u2]. Then there exist x ∈ [u1] and y ∈ [u2] such that z = x + y. Rewrite x, y as x = u1 + w1, y = u2 + w2 for some w1, w2 ∈ V . Hence z = (u1 + u2)+ (w1 + w2), which implies z ∈ [u1 + u2]. Conversely, if z ∈ [u1 + u2], then there is some w ∈ V such that z = (u1 + u2) + w = (u1+w)+u2. Since u1+w ∈ [u1] and u2 ∈ [u2], we see that z ∈ [u1]+ [u2]. Hence the claim follows. From the property (1.6.3), we see clearly that the coset [0] = V serves as an additive zero element among the set of all cosets. Similarly, we may also naturally define a[u] = {ax | x ∈ [u]} for a ∈ R where a = 0. Note that this last restriction is necessary because otherwise 0[u] would be a single-point set consisting of zero vector only. We claim a[u] = [au], a ∈ R \ {0}. (1.6.4) In fact, if z ∈ a[u], there is some x ∈ [u] such that z = ax. Since x ∈ [u], there is some w ∈ V such that x = u+w. So z = au+ aw which implies z ∈ [au]. Conversely, if z ∈ [au], then there is some w ∈ V such that z = au+w. Since z = a(u+ a−1w), we get z ∈ a[u]. So (1.6.4) is established. www.pdfgrip.com 26 Vector spaces Since the coset [0] is already seen to be the additive zero when adding cosets, we are prompted to define 0[u] = [0]. (1.6.5) Note that (1.6.4) and (1.6.5) may be collectively rewritten as a[u] = [au], a ∈ R, u ∈ R2. (1.6.6) We may examine how the above introduced addition between cosets and scalar multiplication with cosets make the set of all cosets into a vector space over R, denoted by R2/V, and called the quotient space of R2 modulo V . As investigated, the geometric meaning of R2/V is that it is the set of all the lines in R2 parallel to the vector v and that these lines can be added and multiplied by real scalars so that the set of lines enjoys the structure of a real vector space. There is no difficulty extending the discussion to the case of R3 with V a line or plane through the origin. More generally, the above quotient-space construction may be formulated as follows. Definition 1.12 Let U be a vector space over a field F and V a subspace of U . The set of cosets represented by u ∈ U given by [u] = {u+ w |w ∈ V } = {u} + V ≡ u+ V, (1.6.7) equipped with addition and scalar multiplication defined by [u] + [v] = [u+ v], u, v ∈ U, a[u] = [au], a ∈ F, u ∈ U, (1.6.8) forms a vector space over F, called the quotient space of U modulo V , and is denoted by U/V . Let BV = {v1, . . . , vk} be a basis of V . Extend BV to get a basis of U , say BU = {v1, . . . , vk, u1, . . . , ul}. (1.6.9) We claim that {[u1], . . . , [ul]} forms a basis for U/V . In fact, it is evident that these vectors span U/V . So we only need to show their linear independence. Consider the relation a1[u1] + · · · + al[ul] = 0, a1, . . . , al ∈ F. (1.6.10) Note that 0 = [0] in U/V . So a1u1 + · · · + alul ∈ V . Thus there are scalars b1, . . . , bk ∈ F such that a1u1 + · · · + alul = b1v1 + · · · + bkvk. (1.6.11) www.pdfgrip.com 1.6 Quotient spaces 27 Hence a1 = · · · = al = b1 = · · · = bk = 0 and the claimed linear indepen- dence follows. As a consequence of the afore-going discussion, we arrive at the following basic dimensionality relation dim(U/V )+ dim(V ) = dim(U). (1.6.12) Note that the construction made to arrive at (1.6.12) demonstrates a practical way to find a basis for U/V . The quantity dim(U/V ) = dim(U)− dim(V ) is sometimes called the codimension of the subspace V in U . The quotient space U/V may be viewed as constructed from the space U after ‘collapsing’ or ‘suppressing’ its subspace V . Exercises 1.6.1 Let V be the subspace of R2 given by V = Span{(1,−1)}. (i) Draw V in R2. (ii) Draw the cosets S1 = (1, 1)+ V, S2 = (2, 1)+ V. (1.6.13) (iii) Describe the quotient space R2/V . (iv) Draw the coset S3 = (−2)S1 + S2. (v) Determine whether S3 is equal to (−1, 0)+ V (explain why). 1.6.2 Let V be the subspace of P2 (set of polynomials of degrees up to 2 and with coefficients in R) satisfying∫ 1 −1 p(t) dt = 0, p(t) ∈ P2. (1.6.14) (i) Find a basis to describe V . (ii) Find a basis for the quotient space P2/V and verify (1.6.12). 1.6.3 Describe the plane given in terms of the coordinates (x, y, z) ∈ R3 by the equation ax + by + cz = d, (a, b, c) = (0, 0, 0), (1.6.15) where a, b, c, d ∈ R are constants, as a coset in R3 and as a point in a quotient space. 1.6.4 Let U be a vector space and V and W some subspaces of U . For u ∈ U , use [u]V and [u]W to denote the cosets u+ V and u+W , respectively. Show that, if V ⊂ W , and u1, . . . , uk ∈ U , then that [u1]V , . . . , [uk]V are linearly dependent implies that [u1]W, . . . , [uk]W are linearly dependent. In particular, if U is finite dimensional, then dim(U/V ) ≥ dim(U/W), if V ⊂ W , as may also be seen from using (1.6.12). www.pdfgrip.com 28 Vector spaces 1.7 Normed spaces It will be desirable to be able to evaluate the ‘length’ or ‘magnitude’ or ‘amplitude’ of any vector in a vector space. In other words, it will be use- ful to associate to each vector a quantity that resembles the notion of length of a vector in (say) R3. Such a quantity is generically called norm. In this section, we take the field F to be either R or C. Definition 1.13 Let U be a vector space over the field F. A norm over U is a correspondence ‖ · ‖ : U → R such that we have the following. (1) (Positivity) ‖u‖ ≥ 0 for u ∈ U and ‖u‖ = 0 only for u = 0. (2) (Homogeneity) ‖au‖ = |a|‖u‖ for a ∈ F and u ∈ U . (3) (Triangle inequality) ‖u+ v‖ ≤ ‖u‖ + ‖v‖ for u, v ∈ U . A vector space equipped with a norm is called a normed space. If ‖ · ‖ is the specific norm of the normed space U , we sometimes spell this fact out by stating ‘normed space (U, ‖ · ‖)’. Definition 1.14 Let (U, ‖ · ‖) be a normed space and {uk} a sequence in U . For some u0 ∈ U , we say that uk → u0 or {uk} converges to u0 as k →∞ if lim k→∞‖u0 − uk‖ = 0, (1.7.1) and the sequence {uk} is said to be a convergent sequence. The vector u0 is also said to be the limit of the sequence {uk}. The notions of convergence and limit are essential for carrying out calculus in a normed space. Let U be a finite-dimensional space. We can easily equip U with a norm. For example, assume that B = {u1, . . . , un} is a basis of U . For any u ∈ U , define ‖u‖1 = n∑ i=1 |ai |, where u = n∑i=1 aiui . (1.7.2) It is direct to verify that ‖ · ‖1 indeed defines a norm. More generally, for p ≥ 1, we may set ‖u‖p = ( n∑ i=1 |ai |p ) 1 p , where u = n∑ i=1 aiui . (1.7.3) It can be shown that ‖ · ‖p also defines a norm over U . We will not check this fact. What interests us here, however, is the limit www.pdfgrip.com 1.7 Normed spaces 29 lim p→∞‖u‖p = max{|ai | | i = 1, . . . , n}. (1.7.4) To prove (1.7.4), we note that the right-hand side of (1.7.4) is simply |ai0 | for some i0 ∈ {1, . . . , n}. Thus, in view of (1.7.3), we have ‖u‖p ≤ ( n∑ i=1 |ai0 |p ) 1 p = |ai0 |n 1 p . (1.7.5) From (1.7.5), we obtain lim sup p→∞ ‖u‖p ≤ |ai0 |. (1.7.6) On the other hand, using (1.7.3) again, we have ‖u‖p ≥ (|ai0 |p) 1p = |ai0 |. (1.7.7) Thus, lim inf p→∞ ‖u‖p ≥ |ai0 |. (1.7.8) Therefore (1.7.4) is established. As a consequence, we are motivated to adopt the notation ‖u‖∞ = max{|ai | | i = 1, . . . , n}, where u = n∑ i=1 aiui, (1.7.9) and restate our result (1.7.4) more elegantly as lim p→∞‖u‖p = ‖u‖∞, u ∈ U. (1.7.10) It is evident that ‖ · ‖∞ does define a norm over U . Thus we have seen that there are many ways to introduce a norm over a vector space. So it will be important to be able to compare norms. For calculus, it is obvious that the most important thing with regard to a norm is the notion of convergence. In particular, assume there are two norms, say ‖ · ‖ and ‖ · ‖′, equipped over the vector space U . We hope to know whether convergence with respect to norm ‖·‖ implies convergence with respect to norm ‖·‖′. This desire motivates the introduction of the following concept. Definition 1.15 Let U be a vector space and ‖ · ‖ and ‖ · ‖′ two norms over U . We say that ‖ · ‖ is stronger than ‖ · ‖′ if convergence with respect to norm ‖ · ‖ implies convergence with respect to norm ‖·‖′. More precisely, any convergent sequence {uk} with limit u0 in (U, ‖ · ‖) is also a convergent sequence with the same limit in (U, ‖ · ‖′). Regarding the above definition, we have the following. www.pdfgrip.com 30 Vector spaces Theorem 1.16 Let U be a vector space and ‖ · ‖ and ‖ · ‖′ two norms over U . Then norm ‖ · ‖ is stronger than norm ‖ · ‖′ if and only if there is a constant C > 0 such that ‖u‖′ ≤ C‖u‖, ∀u ∈ U. (1.7.11) Proof If (1.7.11) holds, then it is clear that ‖ · ‖ is stronger than ‖ · ‖′. Suppose that ‖·‖ is stronger than ‖·‖′ but (1.7.11) does not hold. Thus there is a sequence {uk} in U such that uk = 0 (k = 1, 2, . . . ) and ‖uk‖′ ‖uk‖ ≥ k, k = 1, 2, . . . . (1.7.12) Define vk = 1‖uk‖′ uk, k = 1, 2, . . . . (1.7.13) Then (1.7.12) yields the bounds ‖vk‖ ≤ 1 k , k = 1, 2, . . . . (1.7.14) Consequently, vk → 0 as k →∞ with respect to norm ‖ · ‖. However, accord- ing to (1.7.13), we have ‖vk‖′ = 1, k = 1, 2, . . . , and vk → 0 as k →∞ with respect to norm ‖ · ‖′. This reaches a contradiction. Definition 1.17 Let U be a vector space and ‖ · ‖ and ‖ · ‖′ two norms over U . We say that norms ‖ · ‖ and ‖ · ‖′ are equivalent if convergence in norm ‖ · ‖ is equivalent to convergence in norm ‖ · ‖′. In view of Theorem 1.16, we see that norms ‖ · ‖ and ‖ · ‖′ are equivalent if and only if the inequality C1‖u‖ ≤ ‖u‖′ ≤ C2‖u‖, ∀u ∈ U, (1.7.15) holds true for some suitable constants C1, C2 > 0. The following theorem regarding norms over finite-dimensional spaces is of fundamental importance. Theorem 1.18 Any two norms over a finite-dimensional space are equivalent. Proof Let U be an n-dimensional vector space and B = {u1, . . . , un} a basis for U . Define the norm ‖ · ‖1 by (1.7.2). Let ‖ · ‖ be any given norm over U . Then by the properties of norm we have www.pdfgrip.com 1.7 Normed spaces 31 ‖u‖ ≤ n∑ i=1 |ai |‖ui‖ ≤ α2 n∑ i=1 |ai | = α2‖u‖1, (1.7.16) where we have set α2 = max{‖ui‖ | i = 1, . . . , n}. (1.7.17) In other words, we have shown that ‖ · ‖1 is stronger than ‖ · ‖. In order to show that ‖ · ‖ is also stronger than ‖ · ‖1, we need to prove the existence of a constant α1 > 0 such that α1‖u‖1 ≤ ‖u‖, ∀u ∈ U. (1.7.18) Suppose otherwise that (1.7.18) fails to be valid. In other words, the set of ratios { ‖u‖ ‖u‖1 ∣∣∣∣ u ∈ U, u = 0} (1.7.19) does not have a positive infimum. Then there is a sequence {vk} in U such that 1 k ‖vk‖1 ≥ ‖vk‖, vk = 0, k = 1, 2, . . . . (1.7.20) Now set wk = 1‖vk‖1 vk, k = 1, 2, . . . . (1.7.21) Then (1.7.20) implies that wk → 0 as k →∞ with respect to ‖ · ‖. On the other hand, we have ‖wk‖1 = 1 (k = 1, 2, . . . ). Express wk with respect to basis B as wk = a1,ku1 + · · · + an,kun, a1,k, . . . , an,k ∈ F, k = 1, 2, . . . . (1.7.22) The definition of norm ‖ · ‖1 then implies |ai,k| ≤ 1 (i = 1, . . . , n, k = 1, 2, . . . ). Hence, by the Bolzano–Weierstrass theorem, there is a subsequence of {k}, denoted by {ks}, such that ks →∞ as s →∞ and ai,ks → some ai,0 ∈ F as s →∞, i = 1, . . . , n. (1.7.23) Now set w0 = a1,0u1 + · · · + an,0un. Then ‖w0 − wks‖1 → 0 as s →∞. Moreover, the triangle inequality gives us ‖w0‖1 ≥ ‖wks‖1 − ‖wks − w0‖1 = 1− ‖wks − w0‖1, s = 1, 2, . . . . (1.7.24) Thus, letting s →∞ in (1.7.24), we obtain ‖w0‖1 ≥ 1. However, substituting u = w0 −wks in (1.7.16) and letting s →∞, we see that wks → w0 as s → ∞ with respect to norm ‖ · ‖ as well which is false because we already know that wk → 0 as k →∞ with respect to norm ‖ · ‖. www.pdfgrip.com 32 Vector spaces Summarizing the above study, we see that there are some constants α1, α2 > 0 such that α1‖u‖1 ≤ ‖u‖ ≤ α2‖u‖1, ∀u ∈ U. (1.7.25) Finally, let ‖ · ‖′ be another norm over U . Then we have some constants β1, β2 > 0 such that β1‖u‖1 ≤ ‖u‖′ ≤ β2‖u‖1, ∀u ∈ U. (1.7.26) Combining (1.7.25) and (1.7.26), we arrive at the desired conclusion β1 α2 ‖u‖ ≤ ‖u‖′ ≤ β2 α1 ‖u‖, ∀u ∈ U, (1.7.27) which establishes the equivalence of norms ‖ · ‖ and ‖ · ‖′ as stated. As an immediate application of Theorem 1.18, we have the following. Theorem 1.19 A finite-dimensional normed space (U, ‖·‖) is locally compact. That is, any bounded sequence in U contains a convergent subsequence. Proof Let {u1, . . . , un} be any basis of U and introduce norm ‖ · ‖1 by the expression (1.7.2). Let {vk} be a bounded sequence in (U, ‖ · ‖). Then Theo- rem 1.18 says that {vk} is also bounded in (U, ‖ · ‖1). If we rewrite vk as vk = a1,ku1 + · · · + an,kun, a1,k, . . . , an,k ∈ F, k = 1, 2, . . . , (1.7.28) the definition of ‖ · ‖1 then implies that the sequences {ai,k} (i = 1, 2, . . . ) are all bounded in F. Thus the Bolzano–Weierstrass theorem indicates that there are subsequences {ai,ks } (i = 1, . . . , n) which converge to some ai,0 ∈ F (i = 1, . . . , n) as s →∞. Consequently, setting v0 = a1,0u1 + · · · + an,0un, we have ‖v0 − vks‖1 → 0 as s →∞. Thus ‖v0 − vks‖ ≤ C‖v0 − vks‖1 → 0 as s →∞ as well. Let (U, ‖ · ‖) be a finite-dimensional normed vector space and V a subspace of U . Consider the quotient space U/V . As an exercise, it may be shown that ‖[u]‖ = inf{‖v‖ | v ∈ [u]}, [u] ∈ U/V, u ∈ U, (1.7.29) defines a norm over U/V . Exercises 1.7.1 Consider the vector space C[a, b] of the set of all real-valued continuous functions over the interval [a, b] where a, b ∈ R and a < b. Define two norms ‖ · ‖1 and ‖ · ‖∞ by setting www.pdfgrip.com 1.7 Normed spaces 33 ‖u‖1 = ∫ b a |u(t)| dt, ‖u‖∞ = max t∈[a,b] |u(t)|, ∀u ∈ C[a, b]. (1.7.30) Show that ‖ · ‖∞ is stronger than ‖ · ‖1 but not vice versa. 1.7.2 Over the vector space C[a, b] again, define norm ‖ · ‖p(p ≥ 1) by ‖u‖p = (∫ b a |u(t)|p dt ) 1 p , ∀u ∈ C[a, b]. (1.7.31) Show that ‖ · ‖∞ is stronger than ‖ · ‖p for any p ≥ 1 and prove that lim p→∞‖u‖p = ‖u‖∞, ∀u ∈ C[a, b]. (1.7.32) 1.7.3 Let (U, ‖ · ‖) be a normed space and use U ′ to denote the dual space of U . For each u′ ∈ U ′, define ‖u′‖′ = sup{|u′(u)| | u ∈ U, ‖u‖ = 1}. (1.7.33) Show that ‖ · ‖′ defines a norm over U ′. 1.7.4 Let U be a vector space and U ′ its dual space which is equipped with a norm, say ‖ · ‖′. Define ‖u‖ = sup{|u′(u)| | u′ ∈ U ′, ‖u′‖′ = 1}. (1.7.34) Show that ‖ · ‖ defines a norm over U as well. 1.7.5 Prove thatfor any finite-dimensional normed space (U, ‖·‖) with a given subspace V we may indeed use (1.7.29) to define a norm for the quotient space U/V . 1.7.6 Let ‖ · ‖ denote the Euclidean norm of R2 which is given by ‖u‖ = √ a21 + a22, u = (a1, a2) ∈ R2. (1.7.35) Consider the subspace V = {(x, y) ∈ R2 | 2x − y = 0} and the coset [(−1, 1)] in R2 modulo V . (i) Find the unique vector v ∈ R2 such that ‖v‖ = ‖[(−1, 1)]‖. (ii) Draw the coset [(−1, 1)] in R2 and the vector v found in part (i) and explain the geometric content of the results. www.pdfgrip.com 2 Linear mappings In this chapter we consider linear mappings over vector spaces. We begin by stating the definition and a discussion of the structural properties of linear map- pings. We then introduce the notion of adjoint mappings and illustrate some of their applications. We next focus on linear mappings from a vector space into itself and study a series of important concepts such as invariance and reducibility, eigenvalues and eigenvectors, projections, nilpotent mappings, and polynomials of linear mappings. Finally we discuss the use of norms of linear mappings and present a few analytic applications. 2.1 Linear mappings A linear mapping may be regarded as the simplest correspondence between two vector spaces. In this section we start our study with the def- inition of linear mappings. We then discuss the matrix representation of a linear mapping, composition of linear mappings, and the rank and nullity of a linear mapping. 2.1.1 Definition, examples, and notion of associated matrices Let U and V be two vector spaces over the same field F. A linear mapping or linear map or linear operator is a correspondence T from U into V , written as T : U → V , satisfying the following. (1) (Additivity) T (u1 + u2) = T (u1)+ T (u2), u1, u2 ∈ U . (2) (Homogeneity) T (au) = aT (u), a ∈ F, u ∈ U . 34 www.pdfgrip.com 2.1 Linear mappings 35 A special implication of the homogeneity condition is that T (0) = 0. One may also say that a linear mapping ‘respects’ or preserves vector addition and scalar multiplication. The set of all linear mappings from U into V will be denoted by L(U, V ). For S, T ∈ L(U, V ), we define S + T to be a mapping from U into V satisfying (S + T )(u) = S(u)+ T (u), ∀u ∈ U. (2.1.1) For any a ∈ F and T ∈ L(U, V ), we define aT to be the mapping from U into V satisfying (aT )(u) = aT (u), ∀u ∈ U. (2.1.2) We can directly check that the mapping addition (2.1.1) and scalar-mapping multiplication (2.1.2) make L(U, V ) a vector space over F. We adopt the notation L(U) = L(U,U). As an example, consider the space of matrices, F(m, n). For A = (aij ) ∈ F(m, n), define TA(x) = Ax = ⎛⎜⎝ a11 · · · a1n· · · · · · · · · am1 · · · amn ⎞⎟⎠ ⎛⎜⎜⎝ x1 ... xn ⎞⎟⎟⎠ , x = ⎛⎜⎜⎝ x1 ... xn ⎞⎟⎟⎠ ∈ Fn. (2.1.3) Then TA ∈ L(Fn,Fm). Besides, for the standard basis e1, . . . , en of Fn, we have TA(e1) = ⎛⎜⎜⎝ a11 ... am1 ⎞⎟⎟⎠ , . . . , TA(en) = ⎛⎜⎜⎝ a1n ... amn ⎞⎟⎟⎠ . (2.1.4) In other words, the images of e1, . . . , en under the linear mapping TA are sim- ply the column vectors of the matrix A, respectively. Conversely, take any element T ∈ L(Fn,Fm). Let v1, . . . , vn ∈ Fm be images of e1, . . . , en under T such that T (e1) = v1 = ⎛⎜⎜⎝ a11 ... am1 ⎞⎟⎟⎠ = m∑ i=1 ai1ei, . . . , www.pdfgrip.com 36 Linear mappings T (en) = vn = ⎛⎜⎜⎝ a1n ... amn ⎞⎟⎟⎠ = m∑ i=1 ainei, (2.1.5) where {e1, . . . , em} is the standard basis of Fm. Then any x = x1e1+· · ·+xnen has the image T (x) = T ⎛⎝ n∑ j=1 xj en ⎞⎠ = n∑ j=1 xjvj = (v1, . . . , vn) ⎛⎜⎜⎝ x1 ... xn ⎞⎟⎟⎠ = ⎛⎜⎝ a11 · · · a1n· · · · · · · · · am1 · · · amn ⎞⎟⎠ ⎛⎜⎜⎝ x1 ... xn ⎞⎟⎟⎠ . (2.1.6) In other words, T can be identified with TA through the matrix A consisting of column vectors as images of e1, . . . , en under T given in (2.1.5). It may also be examined that mapping addition and scalar-mapping multi- plication correspond to matrix addition and scalar-matrix multiplication. In this way, as vector spaces, F(m, n) and L(Fn,Fm) may be identified with each other. In general, let BU = {u1, . . . , un} and BV = {v1, . . . , vm} be bases of the finite-dimensional vector spaces U and V over the field F, respectively. For any T ∈ L(U, V ), we can write T (uj ) = m∑ i=1 aij vi, aij ∈ F, i = 1, . . . , m, j = 1, . . . , n. (2.1.7) Since the images T (u1), . . . , T (un) completely determine the mapping T , we see that the matrix A = (aij ) completely determines the mapping T . In other words, to each mapping in L(U, V ), there corresponds a unique matrix in F(m, n). Conversely, for any A = (aij ) ∈ F(m, n), we define T (u1), . . . , T (un) by (2.1.7). Moreover, for any x ∈ U given by x = n∑ j=1 xjuj for x1, . . . , xn ∈ F, we set T (x) = T ⎛⎝ n∑ j=1 xjuj ⎞⎠ = n∑ j=1 xjT (uj ). (2.1.8) It is easily checked that this makes T a well-defined element in L(U, V ). www.pdfgrip.com 2.1 Linear mappings 37 Thus we again see, after specifying bases for U and V , that we may identify L(U, V ) with F(m, n) in a natural way. After the above general description of linear mappings, especially their iden- tification with matrices, we turn our attention to some basic properties of linear mappings. 2.1.2 Composition of linear mappings Let U,V,W be vector spaces over a field F of respective dimensions n, l,m. For T ∈ L(U, V ) and S ∈ L(V,W), we can define the composition of T and S with the understanding (S ◦ T )(x) = S(T (x)), x ∈ U. (2.1.9) It is obvious that S ◦ T ∈ L(U,W). We now investigate the matrix of S ◦ T in terms of the matrices of S and T . To this end, let BU ={u1, . . . , un},BV ={v1, . . . , vl},BW ={w1, . . . , wm} be the bases of U,V,W , respectively, and A = (aij ) ∈ F(l, n) and B = (bij ) ∈ F(m, l) be the correspondingly associated matrices of T and S, respec- tively. Then we have (S ◦ T )(uj ) = S(T (uj )) = S ( l∑ i=1 aij vi ) = l∑ i=1 aij S(vi) = l∑ i=1 m∑ k=1 aij bkiwk = m∑ i=1 l∑ k=1 bikakjwi. (2.1.10) In other words, we see that if we take C = (cij ) ∈ F(m, n) to be the matrix associated to the linear mapping S ◦ T with respect to the bases BU and BW , then C = BA. Hence the composition of linear mappings corresponds to the multiplication of their associated matrices, in the same order. For this reason, it is also customary to use ST to denote S ◦T , when there is no risk of confusion. The composition of linear mappings obviously enjoys the associativity property R ◦ (S ◦ T ) = (R ◦ S) ◦ T , R ∈ L(W,X), S ∈ L(V,W), T ∈ L(U, V ), (2.1.11) as can be seen from (R ◦ (S ◦ T ))(u) = R((S ◦ T )(u)) = R(S(T (u))) (2.1.12) www.pdfgrip.com 38 Linear mappings and ((R ◦ S) ◦ T )(u) = (R ◦ S)(T (u)) = R(S(T (u))) (2.1.13) for any u ∈ U . Thus, when applying (2.1.11) to the situation of linear mappings between finite-dimensional vector spaces and using their associated matrix representa- tions, we obtain another proof of the associativity of matrix multiplication. 2.1.3 Null-space, range, nullity, and rank For T ∈ L(U, V ), the subset N(T ) = {x ∈ U | T (x) = 0} (2.1.14) in U is a subspace of U called the null-space of T , which is sometimes called the kernel of T , and denoted as Kernel(T ) or T −1(0). Here and in the sequel, note that, in general, T −1(v) denotes the set of preimages of v ∈ V under T . That is, T −1(v) = {x ∈ U | T (x) = v}. (2.1.15) Besides, the subset R(T ) = {v ∈ V | v = T (x) for some x ∈ U} (2.1.16) is a subspace of V called the range of T , which is sometimes called the image of U under T , and denoted as Image(T ) or T (U). For T ∈ L(U, V ), we say that T is one-to-one, or 1-1, or injective, if T (x) = T (y) whenever x, y ∈ U with x = y. It is not hard to show that T is 1-1 if and only if N(T ) = {0}. We say that T is onto or surjective if R(T ) = V . A basic problem in linear algebra is to investigate whether the equation T (x) = v (2.1.17) has a solution x in U for a given vector v ∈ V . From the meaning of R(T ), we see that (2.1.17) has a solution if and only if v ∈ R(T ), and the solution
Compartilhar