Birkhäuser Advanced Texts Basler Lehrbücher
Andreas Rosén
Geometric Multivector Analysis From Grassmann to Dirac
Bir...

Author:
Andreas Rosen

This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!

Birkhäuser Advanced Texts Basler Lehrbücher

Andreas Rosén

Geometric Multivector Analysis From Grassmann to Dirac

Birkhäuser Advanced Texts Basler Lehrbücher

Series editors Steven G. Krantz, Washington University, St. Louis, USA Shrawan Kumar, University of North Carolina at Chapel Hill, Chapel Hill, USA Jan Nekováˇr, Sorbonne Université, Paris, France

More information about this series at http://www.springer.com/series/4842

Andreas Rosén

Geometric Multivector Analysis From Grassmann to Dirac

Andreas Rosén Department of Mathematical Sciences Chalmers University of Technology and the University of Gothenburg Gothenburg, Sweden

ISSN 1019-6242 ISSN 2296-4894 (electronic) Birkhäuser Advanced Texts Basler Lehrbücher ISBN 978-3-030-31411-8 (eBook) ISBN 978-3-030-31410-1 https://doi.org/10.1007/978-3-030-31411-8 Mathematics Subject Classification (2010): 15-01, 15A72, 15A66, 35-01, 35F45, 45E05, 53-01, 58A10, 58A12, 58J20 © Springer Nature Switzerland AG 2019 This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmission or information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed. The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant protective laws and regulations and therefore free for general use. The publisher, the authors, and the editors are safe to assume that the advice and information in this book are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or the editors give a warranty, express or implied, with respect to the material contained herein or for any errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. This book is published under the imprint Birkhäuser, www.birkhauser-science.com by the registered company Springer Nature Switzerland AG The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland

Contents Preface 1

2

3

4

ix

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

1 2 5 9 13 17 21

Exterior Algebra 2.1 Multivectors . . . . . . . . . . . . . 2.2 The Grassmann Cone . . . . . . . 2.3 Mapping Multivectors . . . . . . . 2.4 Oriented Measure . . . . . . . . . . 2.5 Multicovectors . . . . . . . . . . . 2.6 Interior Products and Hodge Stars 2.7 Mappings of Interior Products . . . 2.8 Anticommutation Relations . . . . 2.9 The Pl¨ ucker Relations . . . . . . . 2.10 Comments and References . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

23 24 34 40 44 47 51 61 63 67 69

Clifford Algebra 3.1 The Clifford Product . . . . . . . . . 3.2 Complex Numbers and Quaternions 3.3 Abstract Clifford Algebras . . . . . . 3.4 Matrix Representations . . . . . . . 3.5 Comments and References . . . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

73 . 74 . 82 . 89 . 93 . 102

Prelude: Linear Algebra 1.1 Vector Spaces . . . . . . . . . . 1.2 Duality . . . . . . . . . . . . . 1.3 Inner Products and Spacetime . 1.4 Linear Maps and Tensors . . . 1.5 Complex Linear Spaces . . . . 1.6 Comments and References . . .

. . . . . .

Rotations and M¨obius Maps 105 4.1 Isometries and the Clifford Cone . . . . . . . . . . . . . . . . . . . 106 4.2 Infinitesimal Rotations and Bivectors . . . . . . . . . . . . . . . . . 113 v

vi

Contents 4.3 4.4 4.5 4.6 4.7

5

6

7

8

9

Euclidean Rotations . . . . . . . Spacetime Rotations . . . . . . . Fractional Linear Maps . . . . . Mappings of the Celestial Sphere Comments and References . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

117 125 134 142 150

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

153 154 161 167 172 183

Interlude: Analysis 6.1 Domains and Manifolds . . . 6.2 Fourier Transforms . . . . . . 6.3 Partial Differential Equations 6.4 Operator Theory . . . . . . . 6.5 Comments and References . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

185 186 191 197 200 206

Multivector Calculus 7.1 Exterior and Interior Derivatives . . 7.2 Pullbacks and Pushforwards . . . . . 7.3 Integration of Forms . . . . . . . . . 7.4 Vector Fields and Cartan’s Formula 7.5 Poincar´e’s Theorem . . . . . . . . . 7.6 Hodge Decompositions . . . . . . . . 7.7 Comments and References . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

209 211 216 224 235 239 242 253

Hypercomplex Analysis 8.1 Monogenic Multivector Fields 8.2 Spherical monogenics . . . . . 8.3 Hardy Space Splittings . . . . 8.4 Comments and References . .

Spinors in Inner Product Spaces 5.1 Complex Representations . 5.2 The Complex Spinor Space 5.3 Mapping Spinors . . . . . . 5.4 Abstract Spinor Spaces . . 5.5 Comments and References .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

255 257 265 277 283

Dirac Wave Equations 9.1 Wave and Spin Equations . . . . 9.2 Dirac Equations in Physics . . . 9.3 Time-Harmonic Waves . . . . . . 9.4 Boundary Value Problems . . . . 9.5 Integral Equations . . . . . . . . 9.6 Boundary Hodge Decompositions 9.7 Maxwell Scattering . . . . . . . . 9.8 Comments and References . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

285 287 291 303 309 319 327 332 339

. . . .

vii

Contents 10 Hodge Decompositions 10.1 Nilpotent operators . . . . . . . . . 10.2 Half-Elliptic Boundary Conditions 10.3 Hodge Potentials . . . . . . . . . . 10.4 Bogovski˘ı and Poincar´e Potentials ˇ 10.5 Cech Cohomology . . . . . . . . . 10.6 De Rham Cohomology . . . . . . . 10.7 Comments and References . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

343 344 350 354 362 367 372 381

11 Multivector and Spinor Bundles 11.1 Tangent Vectors and Derivatives . 11.2 Multivector Calculus on Manifolds 11.3 Curvature and Bivectors . . . . . . 11.4 Conformal Maps and ON-Frames . 11.5 Weitzenb¨ock Identities . . . . . . . 11.6 Spinor Bundles . . . . . . . . . . . 11.7 Comments and References . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

383 385 390 398 405 408 413 421

12 Local Index Theorems 12.1 Fredholm Dirac Operators . . . . . . 12.2 Normal Coordinates . . . . . . . . . 12.3 The Chern–Gauss–Bonnet Theorem 12.4 The Atiyah–Singer Index Theorem . 12.5 Comments and References . . . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

423 425 431 434 441 448

Bibliography

451

Index

459

Preface I guess all mathematicians have had their defining moments, some events that led them to devote much of their lives and energy to mathematics. Myself, I vividly recall the spring and summer of 1997, spending my days reading about Clifford algebras in David Hestenes’s inspirational books and listening to the Beatles. Don’t misunderstand me. To a Swede, there is nothing that beats ABBA, but that summer it happened that the Clifford algebras were enjoyed in this particular way. I was a fourth-year undergraduate student at Link¨oping University studying the civil engineering program of applied physics and electrical engineering, and the very last course I took there came to change my life in a way that no one could have anticipated. The course was on “applied mathematics”, and we were supposed to pursue a math project of our choice, typically to solve some differential equation. One odd topic proposed was learning Clifford algebras, and it appealed to me. I fell deeply in love with the beauty of it all, and I read and I read. I found the biographies [34, 37] about Hermann Grassmann, and I learned what an unfortunate turn mathematics had taken since the 1800s. During my university studies I had had a sense of something missing in the vector calculus that we were taught. I remember students asking me in the linear algebra sessions that I taught how the vector product could have area as dimension while at the same time being a vector. I discovered that Grassmann had figured it all out more than 150 years ago, and now it was all strangely hidden to us students of mathematics, all but the one-dimensional vectors. No one had told me anything about vector products in dimensions other than three, or about determinants of rectangular matrices. My personal relations with the vector product had in fact begun some five years earlier, when I borrowed a telescope from my high school for a science project on satellites. Using Kepler’s laws, I calculated a formula for the altitude of a satellite’s orbit, using as input two observations of the satellite’s position and the time elapsed between the two observations. Of course you don’t need a telescope for this, it’s just to look for a slow falling star, but I did other things as well. As you may guess, I stumbled upon a curious expression involving three mixed products, for the plane of rotation of the satellite. It was only the following year, when I had started my university studies, that I learned in the linear algebra lectures that this intriguing formula was called a vector product. ix

x

Preface

A second defining moment occurred two years later, around May 1999. I was spending a Saturday or Sunday in the library at the mathematics department in Lund, and stumbled upon a friend. We started a discussion that led to a search on this rather new thing called the internet, where I found the perfect PhD supervisor, Alan McIntosh, from Australia, one of the giants in harmonic analysis and operator theory. It was a perfect match, since he was doing real analysis, singular integrals and operator theory, as well as mixing in the algebras of Clifford and Grassmann when needed. And so I ended up down under in Canberra, and spent three years applying singular integrals and Clifford algebra to solve Maxwell boundary value problems on Lipschitz domains with Alan McIntosh. The publications [11, 8, 9, 7, 14, 10] related to my thesis work are perhaps the real starting point for this book. To shed light on the confusion: Axelsson = Ros´en before 2011. The reason for telling this story is not that I think the reader is more interested in my personal story than in the subject of the book. I certainly hope not. But nothing is without context, and it may help to know the background to understand this book. The basic algebra is not new; it goes back to the pioneering works of Hermann Grassmann, first published in 1843, whose exterior algebra of multivectors is the topic of Chapter 2, and of William Kingdon Clifford from 1878, whose geometric algebra is the topic of Chapter 3. Although these algebras are geometric and useful enough that one would expect them to fit into the mainstream mathematics curriculum at a not too advanced level, this has not really happened. But over the last century, they have been rediscovered over and over ´ Cartan developed his calculus of again. Inspired by the Grassmann algebra, Elie differential forms in the early 1900s. He was also the first to discover spinors in general in 1913, which is the topic of Chapter 5. In 1928, Paul Dirac formulated his famous equation that describes massive spin 1/2 particles in relativistic quantum mechanics, which we discuss in Section 9.2, and which makes use of spacetime spinors and matrix representations of Clifford’s algebra. In 1963, Michael Atiyah and Isadore Singer rediscovered and generalized the Dirac operator to Riemannian manifolds in connection with their celebrated index theorem, which is the topic of Chapter 12. There are also works by Marcel Riesz from 1958 on spacetime isometries and by Lars Ahlfors from 1985 on M¨obius maps, using Clifford algebra, which is the topic of Chapter 4. Mentioned above, David Hestenes has been advocating the use of Clifford algebra, in particular in mathematical physics, since the 1960s. There is also a research field of Clifford analysis, where a higher-dimensional complex analysis using Clifford algebras has been developed, starting from around 1980 and which is the topic of Chapter 8. Included in this book are also some more recent results related to my own research. The material in Sections 9.3 to 10.4 on Dirac integral equations and Hodge decompositions originates with my early thesis work with Alan McIntosh in 2000– 2002, and most of the key ideas there are an inheritance from him. Since then, the material covered in this book has been a continued source of inspiration for my research. The following publications of mine in particular make use, explicitly

Preface

xi

or implicitly, of the algebras of Grassmann and Clifford in real analysis: Axelsson, Keith, and McIntosh [12]; Auscher, Axelsson, and Hofmann [4]; Auscher, Axelsson, and McIntosh [5]; Axelsson, Kou, and Qian [13]; Ros´en [82, 83]; Bandara, McIntosh, and Ros´en [17]; Bandara and Ros´en [18]; and Ros´en [84, 80]. This book was written in four stages. The first part, on the algebras of Grassmann and Clifford, was written around 2008 at Stockholm University and was used as material for a graduate course given there. In the second stage I wrote basically Chapters 7, 8, and 10 for a graduate course given in Link¨oping in 2010. In the third stage I wrote Chapters 11 and 12 for a graduate course in Gothenburg 2014. In between and after these writing periods, the manuscript was collecting dust until I decided, upon returning to mathematics after an extended period of parental leave in 2018, to prepare this final version for publication. Having been away from math for a while gave me new perspectives on things, and this final preparation turned into a major rewriting of the whole book, which I hope will benefit the reader. A number of mathematicians and friends deserve a sincere thanks for being helpful, directly or indirectly, in the creation of this book. Those who untimely have passed away by now, Peetre, McIntosh, and Passare, will always be remembered fondly by me. In mainly chronological order the following people come to mind. Hans Lundmark, who was my mentor for that very first Clifford algebra project in Link¨ oping. I wonder whether and where I would have discovered this mathematics had he not proposed this project to me. Mats Aigner in Link¨oping, whom I first met in Lund and with whom I have had uncountably many interesting discussions about the algebras of Clifford and Grassmann. Jaak Peetre, who encouraged me and provided interesting discussions on the subject. Wulf Staubach at Uppsala University, that friend from the library who changed my life by being well read and knowing about Alan McIntosh. Alan Mcintosh at the Australian National University, my mathematical father from whom I have learned so much. I doubt very much that I will ever again meet someone with as deep an understanding of life and mathematics as he possessed. Mikael Passare at Stockholm University, who supported me at a critical stage. Erik Duse, who was a student attending that first course that I gave in Stockholm, and who more recently himself gave a course based on the third version of this book in Helsinki, and who has given me valuable feedback, including some exercises contained in this book. The book is organized so that the reader finds in the introduction to each chapter a description of and a road map to the material in that chapter. Comments and references are collected in the final section of each chapter. There are two parts of the book. In the first part, the affine multivector and spinor algebra and geometry are explained. A key idea here is the principle of abstract algebra, as explained in the introduction to Chapter 1. In the second part, we use multivectors and spinors in analysis, first in affine space and later on manifolds. A key idea here is that of splittings of function spaces, as explained in the introduction to Chapter 6. My intention is that the material covered should be accessible to basically anyone with mathematical maturity corresponding to that of an advanced undergraduate

Preface

xii

student, with a solid understanding of standard linear algebra, multi-variable and vector calculus, and complex analysis. My hope is that you will find this beautiful mathematics as useful and inspiring as I have.

Andreas Ros´en G¨ oteborg, August 2019

...the horrible “Vector analysis”, which we now see as a complete perversion of Grassmann’s best ideas. (It is limited to 3 dimensions, replaces bivectors by the awful “vector product” and trivectors by the no less awful “mixed product”, notions linked to the euclidean structure and which have no decent algebraic properties!) / J. Dieudonn´e

Chapter 1

Prelude: Linear Algebra Road map: This chapter is not where to start reading this book, which rather is Chapter 2. The material in the present chapter is meant to be used as a reference for some background material and ideas from linear algebra, which are essential to this book, in particular to the first part of it on algebra and geometry consisting of Chapters 2 through 5. The main idea in this part of the book is what may be called the principle of abstract algebra: It is not important what you calculate with, it is only important how you calculate. Let us explain by examples. Consider √ for example the complex numbers x + iy, where you of course ask what is i = −1 when you first encounter this mathematical construction. But that uncomfortable feeling of what this strange imaginary unit really is fades away as you get more experienced and learn that C is a field of numbers that is extremely useful, to say the least. You no longer care what kind of object i is but are satisfied only to know that i2 = −1, which is how you calculate with i. It is this principle of abstract algebra that one needs to bear in mind for all our algebraic constructions in this book: the exterior algebra of multivectors in Chapter 2, Clifford algebras in Chapter 3, and spinors in Chapter 5. In all cases the construction starts by specifying how we want to calculate. Then we prove that there exist objects that obey these rules of calculation, and that any two constructions are isomorphic. Whenever we know the existence and uniqueness up to isomorphism of the objects, we can regard them as geometric objects with an invariant meaning. Which concrete representation of the objects we have becomes irrelevant. In this chapter, Sections 1.1, 1.2, and 1.4 contain background material for Chapter 2, whereas Sections 1.3 and 1.5 are mainly relevant for Chapters 4 and 5 respectively. © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_1

1

Chapter 1. Prelude: Linear Algebra

2

1.1

Vector Spaces

Two general notations which we use throughout this book are the following. By X := Y or Y =: X we mean that X is defined to be / is assigned the value of Y . By A ↔ B we denote a one-to-one correspondence, or an isomorphism between A and B, depending on context. We shall distinguish the concept of a vector space from the more general concept of a linear space. Except for function spaces, which we use later in part two of the book, we shall assume that our linear spaces are finite-dimensional. The difference between linear spaces and vector spaces is only a conceptual one, though. Indeed, any linear space V is naturally an affine space (V, V ), where V acts on itself through the addition in V ; see below. Thus, strictly mathematically speaking, a linear space is the same thing as a vector space. The difference between linear and vector spaces lies in the geometric interpretation of their objects, and we want to make this distinction clear to start with, since we are going to work with linear spaces whose objects are not to be interpreted as geometric vectors. Definition 1.1.1 (Linear space). A real linear space (L, +, ·) is an abelian group (L, +) together with a scalar multiplication R × L → L that is bilinear with respect to addition and a group action of the multiplicative group R∗ = R \ {0} on L. We recall that a group is a set equipped with a binary associative multiplication, containing an identity element and an inverse to each element. For an abelian group, we assume commutativity and write the binary operation as addition. In a linear space, we sometimes write a product xv of x ∈ R and v ∈ L as vx. Since the product of real numbers is commutative, this presents no problem. On the other hand, by a vector space V we mean a linear space consisting of geometric vectors, that is, “one-dimensional directed objects”, which we refer to as vectors. More precisely, this means that V is the space of translations in some affine space X as follows. Definition 1.1.2 (Vector space). An affine space (X, V ) is a set X on which a real linear space V , the space of translations/vectors in X, acts freely and transitively by addition, that is, there exists an addition by vectors map X × V → X that is a (left or right) action of (V, +) on X such that for all x, y ∈ X there exists a unique v ∈ V , the vector denoted by y − x, for which x + v = y. If x, y ∈ X, then the vector v = y − x has the interpretation of a onedimensional arrow starting at x and ending at y. Starting at a different point x0 ∈ X, the same vector v also appears as the arrow from x0 to x0 + v. Thus a vector v is characterized by its orientation and length, but not its position in X. In general affine spaces, the notion of lengths, and more generally k-volumes, have only a relative meaning when we do not have access to an inner product on the space to measure angles and absolute lengths. Thus in general affine spaces, only

1.1. Vector Spaces

3

the relative lengths of two parallel vectors v1 and v2 can be compared: if v1 = λv2 , then v1 is λ times longer than v2 . In practice, one often identifies the affine space X and its vector space V . The difference is the origin 0: X is V , but where we have “forgotten” the origin. Given an origin point x0 ∈ X, we can identify the vector v ∈ V with the point x0 + v ∈ X. In particular, x0 ∈ X is identified with 0 ∈ V . The reader will notice that in Chapters 2 and 7 we carefully distinguish between X and its vector space V , but that in the later chapters, we become more pragmatic and often identify X =V. Definition 1.1.3 (Rn ). The vector space Rn is the set of n-tuples Rn := {(x1 , . . . , xn ) ; xi ∈ R}, with the usual addition and multiplication by scalars. This linear space has a distinguished basis, the standard basis {ei }, where ei := (0, . . . , 0, 1, 0, . . . , 0), with coordinate 1 at the ith position. We adopt the practical convention that we identify row vectors with column vectors, as is often done in doing analysis in Rn . More precisely, Rn should be the space of column vectors, since matrix multiplication is adapted to this convention. However, whenever no matrix multiplication is involved, it is more convenient x1 t to write x1 . . . xn than ... = x1 . . . xn , where ·t denotes matrix xn transpose. We will not distinguish between parentheses (·) and brackets · . Note the decreasing generality of the notions: an affine space is homogeneous and isotropic, that is, without any distinguished points or directions. A linear space is isotropic, but has a distinguished point: the origin 0. The linear space Rn is neither homogeneous or isotropic: it has an origin and a distinguished basis, the standard basis. Whenever we have fixed a basis {ei } in a vectorP space V , there is a xi ei corresponds natural identification between V and Rn , where a vector v = to the coordinate tuple x = (x1 , . . . , xn ) ∈ Rn . Recall the notion of direct sums of linear spaces. Define the sum of subspaces V1 + V2 := {v1 + v2 ; v1 ∈ V1 , v2 ∈ V2 } when V1 and V2 are two subspaces of a linear space V . When V1 ∩ V2 = {0}, we write V1 ⊕ V2 and call the sum a direct sum. This is an intrinsic direct sum. In contrast, suppose that we are given two linear spaces V1 and V2 , without any common embedding space V . In this case we define the (extrinsic) direct sum of these spaces as V1 ⊕ V2 := {(v1 , v2 ) ∈ V1 × V2 ; v1 ∈ V1 , v2 ∈ V2 }.

4

Chapter 1. Prelude: Linear Algebra

In a natural way, V1 ⊕ V2 is a linear space that contains both spaces V1 , V2 , under suitable identifications. As an example, Rn is the exterior direct sum of n copies of the one-dimensional linear space R. Recall the notions of linear independence of a set S ⊂ V and its linear span span(S) ⊂ V . For concrete calculations in a given linear space V , it is often needed to fix a basis {e1 , . . . , en } ⊂ V, with n = dim V being the dimension of V . It is conceptually important to understand that a basis in general is an unordered set. But often bases for vector spaces are linearly ordered e1 , e2 , e3 , . . . by the positive integers and considered as ordered sets. In particular, this is needed in order to represent v ∈ V , x1 n X . xi ei = e1 . . . en .. , v = x1 e1 + · · · + xn en = i=1 xn by its coordinates (x1 , . . . , xn ) ∈ Rn , and in order to represent a linear map T : V1 → V2 between linear spaces V1 , V2 , a1,1 · · · a1,n x1 m X n X 0 . .. . 0 . 0 . . . ei ai,j xj = e1 · · · em . T (x1 e1 +· · ·+xn en ) = . . . , i=1 j=1 xn am,1 · · · am,n by its matrix A = (ai,j ) relative to the bases {ej } for V1 and {e0i } for V2 . However, many fundamental types of bases used in mathematics do not come with any natural linear order. Indeed, this will be the usual situation in this book, where the basic linear spaces of multivectors, tensors, and spinors have standard bases that are not linearly ordered but rather have some sort of lattice ordering, meaning that the basis elements naturally are indexed by subsets of integers or tuples of integers. Another central theme in this book is that many basic linear spaces that appear are not only linear spaces, but associative algebras in the sense that they come equipped with an associative, but in general noncommutative, product. Definition 1.1.4 (Associative algebra). A real associative algebra (A, +, ∗, 1), with identity, is a linear space over R equipped with a bilinear and associative product ∗, with identity element 1. Scalars λ ∈ R are identified with multiples λ1 ∈ A of the identity, and it is assumed that (λ1) ∗ v = λv = v ∗ (λ1) for all v ∈ A. Let (A1 , +1 , ∗1 , 11 ) and (A2 , +2 , ∗2 , 12 ) be two algebras. Then a map T : A1 → A2 is said to be an algebra homomorphism if it is linear and satisfies T (v1 ∗1 v2 ) = T (v1 ) ∗2 T (v2 ) for all v1 , v2 ∈ A1 and if T (11 ) = 12 . An invertible homomorphism is called an algebra isomorphism.

1.2. Duality

5

Exercise 1.1.5. Let A be an associative algebra. Define the exponential function exp(x) :=

∞ X 1 k x , k!

x ∈ A.

k=0

Show that exp(x + y) = exp(x) exp(y), provided that x and y commute, that is, if xy = yx. If φ ∈ R, show that if j 2 = −1, cos φ + j sin φ, exp(φj) = cosh φ + j sinh φ, if j 2 = 1, 1 + φj, if j 2 = 0.

1.2

Duality

There are several reasons for us to consider inner products and dualities more general than Euclidean ones. A first reason is that we want to study the geometry of multivectors in Minkowski spacetimes, the closest relative to Euclidean spaces among inner product spaces, which are modeled by an indefinite inner product as in Section 1.3. A second reason is that we want to study real Clifford algebras where the fundamental representation Theorem 3.4.2 involves inner product spaces of signature zero. A third reason is that we want to study spinor spaces, where more general nonsymmetric dualities may appear. Definition 1.2.1 (Duality and inner product). A duality of two linear spaces V1 and V2 is a bilinear map V1 × V2 → R : (v1 , v2 ) 7→ hv1 , v2 i that is non-degenerate in the sense that hv1 , v2 i = 0 for all v1 ∈ V1 only if v2 = 0, and hv1 , v2 i = 0 for all v2 ∈ V2 only if v1 = 0. In the case V1 = V2 = V , we speak of a duality on V . If a duality on V is symmetric in the sense that hv1 , v2 i = hv2 , v1 i for all v1 , v2 ∈ V , then we call the duality an inner product and V an inner product space. We use the notation hvi2 := hv, vi ∈ R. A vector v such that hvi2 = 0 is called singular. If an inner product has the additional property that hvi2 > 0 for all 0 6= v ∈ V , then we call it a Euclidean inner product, and V is called a Euclidean space. In this case, we define the norm p |v| := hvi2 ≥ 0, so that hvi2 = |v|2 . If a duality on V is skew-symmetric in the sense that hv1 , v2 i = −hv2 , v1 i for all v1 , v2 ∈ V , then we call the duality a symplectic form and V a symplectic space. Note carefully that in general, hvi2 may be negative, as compared to the square of a real number. We do not define any quantity hvi, and the square in the notation hvi2 is only formal.

Chapter 1. Prelude: Linear Algebra

6

Exercise 1.2.2. Show that an inner product is Euclidean if hvi2 ≥ 0 for all v ∈ V . Let V be a linear space. There is a canonical linear space V ∗ and duality hV , V i, namely the dual space of V defined as ∗

V ∗ := {linear functionals θ : V → R}. Given such a scalar-valued linear function θ ∈ V ∗ , its value θ(v) ∈ R at v ∈ V will be denoted by hθ, vi := θ(v) ∈ R. Note that this is indeed a duality: if θ(v) = 0 for all v ∈ V , then θ = 0 by definition. On the other hand, if θ(v) = 0 for all θ, then it follows that v = 0, since otherwise, we can take a complementary subspace V 0 ⊂ V so that V = span{v} ⊕ V 0 and define the linear functional θ(αv + v 0 ) := α, α ∈ R, v 0 ∈ V 0 for which θ(v) 6= 0. If V is a vector space with a geometric interpretation of v ∈ V as in Section 1.1, then θ ∈ V ∗ , which we refer to as a covector, is best described in V by its level sets {v ∈ V ; hθ, vi = C}, for different fixed values of C ∈ R. Since θ is linear, these level sets are parallel hyperplanes. The following observation is fundamental in understanding dualities. Proposition 1.2.3 (Representation of dual space). Fix a linear space V . Then there is a one-to-one correspondence between dualities hV 0 , V i and invertible linear maps g : V 0 → V ∗ : v 7→ θ, given by hg(v 0 ), vi := hv 0 , vi,

v ∈ V.

Here the pairing on the left is the functional value g(v 0 )v, whereas the pairing on the right is as in Definition 1.2.1. If V 0 = V , then V is an inner product/symplectic space if and only if g : V → V ∗ is a symmetric/antisymmetric linear map. With Proposition 1.2.3 in mind, we write a duality between two linear spaces as hV ∗ , V i, where V ∗ not necessarily is the dual space of V , but rather a linear space dual to V in the sense of Definition 1.2.1. By Proposition 1.2.3 this abuse of notation presents no problem. In particular, when we have a duality or inner product on V , we shall write θ=v to mean θ = g(v).

7

1.2. Duality

Definition 1.2.4 (Orthogonal complement). Consider a linear space V and a duality hV ∗ , V i. If hv 0 , vi = 0, then we say that v 0 ∈ V ∗ and v ∈ V are orthogonal. The orthogonal complement of a set S 0 ⊂ V ∗ is the subspace (S 0 )⊥ := {v ∈ V ; hv 0 , vi = 0, for all v 0 ∈ S 0 } ⊂ V. For S ⊂ V we similarly define the orthogonal complement S ⊥ := {v 0 ∈ V ∗ ; hv 0 , vi = 0, for all v ∈ S} ⊂ V ∗ . Definition 1.2.5 (Dual basis).P Let {e1 , . . . , en } be a basis for V . Then each v ∈ V can be uniquely written v = j xj ej , and we define covectors e∗j by he∗j , vi := xj = the jth coordinate of v. We call {e∗1 , . . . , en∗ } ⊂ V ∗ the dual basis of {e1 , . . . , en } ⊂ V . Note that the dual basis {e∗1 , . . . , e∗n } is indeed a basis for V ∗ whenever {e1 , . . . , en } is a basis for V , and is characterized by the property ( 1, i = j, he∗i , ej i = 0, i 6= j. When we have a duality on V , then the dual basis is another basis for V . Exercise 1.2.6. Consider V = R2 , the Euclidean plane with its standard inner product. Find the dual basis to {(3/2, 0), (1/4, 1/2)} and draw the two bases. Example 1.2.7 (Crystal lattices). Let {e1 , e2 , e3 } be the standard basis for R3 . In solid-state physics one studies crystal structures. These have the atoms arranged/packed in a regular pattern that repeats itself, a lattice, which may be different for different crystals. Mathematically a crystal lattice is described by a basis {v1 , v2 , v3 }, which is such that the atoms in the crystal are located at the lattice points {n1 v1 + n2 v2 + n3 v3 ; n1 , n2 , n3 ∈ Z}. Two commonly occurring crystal structures are the body-centered cubic lattice, which has basis { 21 (−e1 + e2 + e3 ), 12 (e1 − e2 + e3 ), 21 (e1 + e2 − e3 )}, and the face-centered cubic lattice, which has basis { 12 (e2 + e3 ), 12 (e1 + e3 ), 21 (e1 + e2 )}. Except for a factor 2, these two bases are seen to be dual bases: one speaks of reciprocal lattices for crystal lattices. The names of these lattices are clear if one draws the basis vectors in relation to the unit cube {0 ≤ x1 , x2 , x3 ≤ 1} and its integer translates.

Chapter 1. Prelude: Linear Algebra

8

Example 1.2.8 (Basis FEM functions). When solving partial differential equations numerically using the finite element method (FEM), the following problem appears. For a three-dimensional computation we consider simplices D, the closed convex hull of four points. Using one corner as the origin 0, and vectors {v1 , v2 , v3 } along the edges to the other three corners, we wish to construct linear functions fk : D → R such that fk (vk ) = 1 and fk = 0 on the opposite face of D, for k = 1, 2, 3. Using the dual basis {v1∗ , v2∗ , v3∗ }, we immediately obtain fk (x) = hvk∗ , xi. For practical calculations in an inner product space, we prefer to use the simplest bases: the ON-bases. Definition 1.2.9 (ON-bases). Let h·, ·i be a duality on V . Then {ei } is called an ON-basis if hei , ej i = 0 when i 6= j and if hei i2 = ±1 for all i. In terms of dual bases, a basis {ei } is an ON-basis if and only if e∗i = ±ei ,

i = 1, . . . , n.

In particular, for a Euclidean space, a basis is an ON-basis if and only if it coincides with its dual basis. Proposition 1.2.10 (Existence of ON-bases). Consider a linear space V with a duality hV, V i. Then V is an inner product space if and only if there exists an ON-basis for V . Proof. Clearly V is an inner product space if an ON-basis exists. Conversely, fix any basis {vi } for V , and define the matrix A = (ai,j ) of hV, V i in this basis by ai,j := hvi , vj i. If V is an inner product space, then A is a symmetric matrix. Using the spectral theorem, we can write D = M ∗ AM , for some invertible matrix diagonal matrix D with ±1 as diagonal elements. The basis {ei } M = (mi,j ) and P defined by ei := j vj mj,i is seen to be an ON-basis. For symplectic spaces, the following is the analogue of ON-bases. Let h·, ·i be k k a duality on V , with dim V = 2k. Then {ei }i=1 is called a Darboux basis ∪ {e0i }i=1 if 0 0 1 ≤ i, j ≤ k, hei , ej i = 0 = hei , ej i, 0 0 i 6= j, 1 ≤ i, j ≤ k, hei , ej i = 0 = hei , ej i, 0 hei , ei i = 1 = −hei , e0i i, 1 ≤ i ≤ k. In terms of dual bases, a basis is clearly a Darboux basis if and only if ei∗ = e0i ,

(ei0 )∗ = −ei ,

for each i = 1, . . . , n.

Exercise 1.2.11 (Existence of Darboux bases). Consider a linear space V with a duality hV, V i. Adapt the proof of Proposition 1.2.10 and prove that V is a symplectic space if and only if there exists a Darboux basis for V . Hint: The spectral theorem for normal complex linear operators applies.

1.3. Inner Products and Spacetime

9

1.3 Inner Products and Spacetime In this section we consider non-Euclidean inner product spaces, and in particular Minkowski spacetimes, the mathematical model for special relativity theory. Definition 1.3.1. Let V be an inner product space. Let n+ be the maximal dimension of a subspace V+ ⊂ V such that hvi2 > 0 for all v ∈ V+ \ {0}, and let n− be the maximal dimension of a subspace V− ⊂ V such that hvi2 < 0 for all v ∈ V− \ {0}. The signature of V is the integer n+ − n− . We say that a subspace V1 ⊂ V is degenerate if there exists 0 6= v1 ∈ V1 such that hv1 , vi = 0 for all v ∈ V1 . Otherwise, V1 is called nondegenerate. If hu, vi = 0 for all u, v ∈ V1 , then V1 is called totally degenerate. Note that a subspace of an inner product space is itself an inner product space if and only if the subspace is nondegenerate. Also, a subspace of an inner product space is totally degenerate if and only if all its vectors are singular, as is seen through polarization, that is, the identity hu + vi2 − hu − vi2 = 4hu, vi. A nonzero singular vector spans a one-dimensional totally degenerate subspace. Proposition 1.3.2 (Sylvester’s law of inertia). Let h·, ·i be an inner product on an n-dimensional vector space V , and let n+ and n− be as in Definition 1.3.1. For every ON basis {ei } for V , the number of basis vectors with hei i2 = 1 equals n+ , and the number of basis vectors with hei i2 = −1 equals n− . If n0 denotes the maximal dimension of a totally degenerate subspace V0 ⊂ V , then n+ + n− = n,

min(n+ , n− ) = n0 .

Proof. Let V+ , V− , and V0 be any Euclidean, anti-Euclidean and totally degenerate subspaces, respectively. Then clearly V+ ∩ V− = V+ ∩ V0 = V− ∩ V0 = {0}, and it follows that n+ + n− ≤ n, n+ + n0 ≤ n, and n− + n0 ≤ n. Fix an ON-basis {ei } for V and choose V± := span{ei ; hei i2 = ±1}. Then dim V+ +dim V− = n and dim V± ≤ n± . It follows that n± = dim V± and n+ +n− = n. From n+ + n− = n, it follows that n0 ≤ min(n − n+ , n − n− ) = min(n− , n+ ) =: m. To see that equality is attained, let V0 := {ei1 − ej1 , . . . , eim − ejm }, where heik i2 = 1 and hejk i2 = −1. Then V0 is seen to be totally degenerate. Exercise 1.3.3. Generalize Proposition 1.3.2 to degenerate bilinear and symmetric forms B(·, ·). Let Rad(V ) := {v ∈ V ; B(v, v 0 ) = 0 for all v 0 ∈ V } be the radical of V , and let n00 := dim Rad(V ). Show that n+ + n− + n00 = n and n0 = n00 + min(n+ , n− ).

10

Chapter 1. Prelude: Linear Algebra

Geometrically, the most important difference between a general inner product space and Euclidean spaces concerns orthogonal complements. For any subspace V1 of a Euclidean space V , we always have a direct sum decomposition V = V1 ⊕ V1⊥ , since V1 ∩ V1⊥ = {0}, because there are no singular vectors. This is not always true in general inner product spaces, but we have the following general result. Proposition 1.3.4 (Orthogonal sums). Let V1 be a k-dimensional subspace in an n-dimensional inner product space V . Then dim V1⊥ = n − k and (V1⊥ )⊥ = V1 , and V1 is a nondegenerate subspace if and only if V1 ∩ V1⊥ = {0}, or equivalently, V = V1 ⊕ V1⊥ . In particular, if V1 is one-dimensional and is spanned by a vector v, then V = span{v} ⊕ span{v}⊥ if and only if v is a nonsingular vector. For the remainder of this section, we study the following non-Euclidean inner product spaces. Definition 1.3.5 (Spacetime). An inner product space (W, h·, ·i) is said to be a Minkowski spacetime, or spacetime for short, with n space dimensions if dim W = 1 + n and the signature is n − 1. We always index spacetime ON-bases as {e0 , e1 , . . . , en }, where he0 i2 = −1. Note that in spacetime coordinates, hx0 e0 + x1 e1 + · · · + xn en i2 = −x20 + x21 + · · · + x2n . To describe the geometry given by such an inner product, we use the following terminology. See Figure 1.1. • The double cone Wl := {v ∈ W ; hvi2 = 0} consisting of all singular vectors v is referred to as the light cone in spacetime. Vectors v ∈ Wl are called light-like. We make a choice and declare one of these two cones to be the future light cone Wl+ , and the other cone Wl− is the past light cone. Thus Wl = Wl+ ∪ Wl− and Wl+ ∩ Wl− = {0}. • We denote the interior of the light cone by Wt := {v ∈ W ; hvi2 < 0}, and it contains the time-like vectors. Since Wt is disconnected, we write it as the disjoint union of the future time-like vectors Wt+ , which is the interior of the future light cone, and the past time-like vectors Wt− , which is the interior of the past light cone. We always assume that e0 ∈ Wt+ , that is, that e0 is a future-pointing time-like vector. • We denote the exterior of the light cone by Ws := {v ∈ W ; hvi2 > 0}, and it contains the space-like vectors. Except when the space dimension is n = 1,

1.3. Inner Products and Spacetime

11

Ws is connected. The whole spacetime thus can be written as the disjoint union W = Wt+ ∪ Wt− ∪ Ws ∪ Wl+ ∪ Wl− , except for the origin. • The analogue of the Euclidean unit sphere is the spacetime unit hyperboloid H(W ) := {v ∈ W ; hvi2 = ±1}. Except for space dimension n = 1, this hyperboloid has three connected components: the future time-like part H(Wt+ ) := H(W ) ∩ Wt+ , the past time-like part H(Wt− ) := H(W ) ∩ Wt− , and the space-like part H(Ws ) := H(W ) ∩ Ws = {v ∈ W ; hvi2 = +1}.

Figure 1.1: The lightcone partition of spacetime, and the straight line representing an inertial observer. Exercise 1.3.6. Let {e0 , e1 , e2 } be an ON-basis for a Minkowski spacetime W . Calculate the dual basis {v1 , v2 , v3 } ⊂ W to {e0 +e1 , e2 , e0 −e1 }. If instead {e0 , e1 , e2 } were an ON-basis for a Euclidean space V , what would this dual basis be?

12

Chapter 1. Prelude: Linear Algebra

A main reason for considering Minkowski spacetime is that it is the mathematical model for Einstein’s special relativity theory, when n = 3. Fix an ON-basis {e0 , e1 , e2 , e3 } with he0 i2 = −1. Once an origin is fixed, points in W are identified with vectors x0 e 0 + x1 e 1 + x2 e 2 + x3 e 3 . The coordinates xi are lengths, and we shall use the meter [m] as the unit of length. We shall write the time coordinate x0 as x0 = ct, where t is time measured in seconds [s] and c = 299792458 [m/s] is the exact speed of light. In relativity theory, the points in spacetime are referred to as events, at time t and position x. The entire life of an observer forms a curve γ(s) ∈ W , s ∈ R, containing all the events that he is present at, at least if he has lived and will live forever. For each s ∈ R, the tangent vector γ 0 (s) ∈ Wt+ will be future-pointing and time-like, since the observer always moves at a speed less than that of light. An observer moving without acceleration is called an inertial observer, and is described p by a straight line in spacetime W spanned by a time-like vector. The quantity −hvi2 /c for a time-like vector v has the meaning of time elapsed as measured by an inertial observer present at two events separated by v in spacetime. We refer to the physics literature for further details on relativity theory. See Section 1.6. In the literature, one often models spacetime as an inner product space with signature 1 − 3, as opposed to the signature convention 3 − 1 used here. An advantage is that the important time-like vectors then have hvi2 > 0. A disadvantage is that in this case, spacetimes are close relatives to the anti-Euclidean space, rather than the Euclidean spaces. Of course, these differences are minor technical ones rather than real geometrical or physical ones. A geometric result about spacetime subspaces that we need is the following. Proposition 1.3.7. Let W be a spacetime and let V ⊂ W be a subspace. Then V is of exactly one of the following types. (i) A space-like subspace. In this case V is nondegenerate and is a Euclidean space, whereas V ⊥ is a spacetime. (ii) A time-like subspace. In this case V is nondegenerate and is a spacetime, whereas V ⊥ is a Euclidean space. (iii) A light-like subspace. In this case V is a degenerate subspace and contains a unique one-dimensional subspace V0 spanned by a light-like vector. The hyperplane V0⊥ in W is the tangent space to the light cone Wl along the line V0 and V0 ⊂ V ⊂ V0⊥ . If V 0 is a complement of V0 in V , so that V = V0 ⊕V 0 , then V 0 is space-like.

1.4. Linear Maps and Tensors

13

0 Proof. Consider first the case that V is nondegenerate, and let n± be the signature indices for V as in Proposition 1.3.2. If n+ = n and n− = 1 are the indices for W , then clearly n0− ≤ n− = 1 and n0+ ≤ n+ . Thus two cases are possible. Either 0 n0− = 0, in which case V is a Euclidean space, or n− = 1, in which case V is a 00 00 = n− , which spacetime. Furthermore, if n± are the indices for V ⊥ , then n0− + n− proves the statement about V ⊥ . On the other hand, if V is a degenerate subspace, write n000 and n00 for the dimensions of the radical and a maximal totally degenerate subspace in V as in Exercise 1.3.3. Then

1 ≤ n000 ≤ n00 ≤ n0 = min(n− , n+ ) = 1. 0 0 0 ) = n00 − n00 = 1 − 1 = 0, and also n0− ≤ n− = 1. We claim Therefore min(n+ , n− 0 0 = 0, that n− = 0. To prove this, assume on the contrary that n0− = 1. Then n+ 0 0 0 so that dim V = n00 + n+ + n− = 1 + 0 + 1 = 2. Let v− ∈ V be a time-like vector, and consider the splitting W = span{v− } ⊕ span{v− }⊥ . If v0 ∈ Rad(V ) \ {0}, then v0 = αv− + v+ , which shows that V contains a space-like vector v+ = v0 − αv− by (ii). This contradicts n0+ = 0. We have proved that 0 0 = dim V − 1. = n00 = 1, n+ n0− = 0, n00

Write V0 := Rad(V ). Then V0 ⊂ V ⊂ V0⊥ . Let t 7→ v(t) ∈ Wl be a curve on the light cone such that v(0) ∈ V0 \ {0}. Then 0 = ∂t hv(t), v(t)i|t=0 = 2hv 0 (0), v(0)i. This shows that the hyperplane V0⊥ must contain the tangent space to Wl along V0 . Since the dimensions are equal, this proves the proposition.

1.4

Linear Maps and Tensors

We denote the set of linear operators between two given linear spaces V1 and V2 by L(V1 ; V2 ) := {T : V1 → V2 ; T is linear}, which itself forms a linear space of dimension dim V1 × dim V2 . For V1 = V2 = V , we write L(V ). The null space of a linear map T is denoted by N(T ), and its range is denoted by R(T ) = T V1 . In this section we discuss a less well known generalization that is essential to this book: the tensor product of linear spaces. Just as a linear operator can be represented by its matrix, a two-dimensional rectangular scheme of numbers, general tensor products can be represented by k-dimensional schemes of numbers. However, we shall restrict ourselves to k = 2 and the relation between operators and tensors. The construction of tensors uses the following maps. Definition 1.4.1 (Multilinearity). A map M : V1 × · · · × Vk → V , where V1 , . . . , Vk and V are linear spaces, is called multilinear, or more precisely k-linear, if for each 1 ≤ j ≤ k, the restricted map Vj 3 vj 7→ M (v1 , . . . , vj , . . . , vk ) ∈ V

Chapter 1. Prelude: Linear Algebra

14

is linear for every fixed vi ∈ Vi , i 6= j. When k = 2, we use the name bilinear. The construction of tensors is very similar to that of multivectors in Section 2.1, but is less geometrically transparent. Following the principle of abstract algebra, we proceed as follows to construct the tensor product V ⊗ V 0 of two given linear spaces V and V 0 . • We first note that there exist a linear space VM and a bilinear map M : V × V 0 → VM such that for two given bases {ei }1≤i≤n and {e0j }1≤j≤n0 for V and V 0 respectively, the set {M (ei , e0j )}1≤i≤n,1≤j≤n0 forms a basis for VM . To see this, just let VM be any linear space of dimension nn0 and define M (ei , ej ) to be some basis for VM . Then extend M to a bilinear map. • We next note that if {M (ei , ej0 )}ij is a basis, then {M (fi , fj0 )}ij is also a basis for VM , for any other choice of bases {fi }i and {fj0 }j for V and V 0 respectively. Indeed, using the bilinearity one checks that {M (fi , fj0 )}ij is a linearly independent set in VM . • If M : V × V 0 → VM maps bases onto bases as above, we note the following. If N : V × V 0 → VN is any other bilinear map, then since {M (ei , e0j )}ij is a basis, setting T (M (ei , ej0 )) := N (ei , ej0 ),

1 ≤ i ≤ n, 1 ≤ j ≤ n0 ,

we have the existence of a unique linear map T : VM → VN such that N = T ◦ M . If M has the property that every other bilinear map factors through it in this way, we say that M has the universal property (U). We shall encounter universal properties for other constructions, so more precisely, this is the universal property for tensor products. Conversely, if a given bilinear map M satisfies (U), then it must map bases onto bases as above. Indeed, take any bilinear map N : V × V 0 → VN such that {N (ei , ej0 )}ij is a basis. We now have a unique linear map T : VM → VN mapping {M (ei , ej0 )}ij onto a basis. This is possible only if {M (ei , e0j )}ij is a basis. Definition 1.4.2 (Tensor product). Let V and V 0 be linear spaces. Fix any bilinear map M : V × V 0 → VM satisfying (U). The tensor product of V and V 0 is the linear space V ⊗ V 0 := VM . We call elements in V ⊗ V 0 tensors and we write u ⊗ v := M (u, v).

1.4. Linear Maps and Tensors

15

Note that if some other bilinear map N : V × V 0 → VN satisfies (U), then the linear map T : VM → VN given by the universal property for M has inverse T −1 : VN → VM given by the universal property for N . Therefore, T provides a unique identification of VM and VN . By the principle of abstract algebra, our definition of V ⊗ V 0 makes sense. If {ei } and {e0j } are bases for V and V 0 , then a general tensor in V ⊗ V 0 is of the form XX αij ei ⊗ e0j , j

i

for some αij ∈ R. Proposition 1.4.3 (Operator=tensor). Let V1 and V2 be linear spaces and consider a duality hV1∗ , V1 i. Then there is a unique invertible linear map V2 ⊗ V1∗ → L(V1 ; V2 ) such that v ⊗ θ 7→ T , where T x := hθ, xiv, x ∈ V1 . Proof. Consider the bilinear map V2 × V1∗ → L(V1 ; V2 ) : (v, θ) 7→ T, where T (x) := hθ, xiv for all x ∈ V1 . According to the universal property for V2 ⊗ V1∗ , there exist a unique linear map V2 ⊗ V1∗ → L(V1 ; V2 ) such that v ⊗ θ 7→ T . Let {ei0 } be a basis for V2 , and let {ej } be a basis for V1 with dual basis {e∗i } for V1∗ . Then we see that the tensor X αij e0i ⊗ ej∗ ij

maps onto the linear operator with matrix {αij }ij . This proves the invertibility. The following shows how this translation between tensors and linear operators works. • If T = v ⊗ θ : V1 → V2 and T 0 = v 0 ⊗ θ0 : V2 → V3 , then the composed operator T 0 ◦ T : V1 → V3 corresponds to the tensor (v 0 ⊗ θ0 )(v ⊗ θ) = hθ0 , viv 0 ⊗ θ. This yields a multiplication of tensors, which is referred to as a contraction. • Let T : V → V be a linear operator on a linear space V . Applying the universal property to the pairing V × V ∗ → R : (v, θ) 7→ hθ, vi,

Chapter 1. Prelude: Linear Algebra

16 we get a canonical linear map

Tr : L(V ) = V ⊗ V ∗ → R. R is called the trace of the operator T . If {ei } The obtained number Tr(T ) ∈ P is a basis for V , then Tr(T ) = i αii if {αij } is the matrix for T . • If V1 and V2 are two linear spaces, then there is a natural swapping map S : V2 ⊗ V1∗ → V1∗ ⊗ V2 : v ⊗ θ 7→ θ ⊗ v, defined using the universal property. Identifying V2 ⊗ V1∗ = L(V1 ; V2 ) and V1∗ ⊗ V2 = L(V2∗ ; V1∗ ), this map S of tensors corresponds to the operation of taking adjoints of linear operators. Recall that the adjoint, or dual, of a linear operator T ∈ L(V1 ; V2 ) is T ∗ ∈ L(V2∗ ; V1∗ ) given by hT ∗ θ, vi = hθ, T vi,

θ ∈ V2∗ , v ∈ V1 .

• Let V be a Euclidean space, and let T = T ∗ be a symmetric operator. By the spectral theorem, there exists an ON-basis {ei } for V in which T has a diagonal matrix. Translated to tensors, this result means that if a tensor w ∈ V ⊗ V is fixed by the above swapping map S, then there is an ON-basis in which X αi ei ⊗ ei , w= i

where we as usual identify V and V ∗ through the inner product. • Let V and V 0 be two Euclidean spaces, and w ∈ V ⊗ V 0 . Then there exist ON-bases {ej } for V and {ej0 } for V 0 , and µj ∈ R such that w=

n X

µj ej

0 ⊗ ej .

j=1

This follows, by translation to tensors, from the spectral theorem and Proposition 1.4.4 for operators, where µj are the singular values of the corresponding operator. Proposition 1.4.4 (Polar decomposition). Let V1 , V2 be Euclidean spaces, and consider an invertible linear map T ∈ L(V1 , V2 ). Then there exists a unique symmetric map S ∈ L(V1 ) such that hSu, ui > 0 for all u ∈ V1 \ {0}, and a unique isometric map U ∈ L(V1 , V2 ) such that T = U S. Similarly, there exists a unique factorization T = S 0 U 0 of T , where S 0 is positive symmetric on V2 and U 0 : V1 → V2 is isometric. We have U 0 = U and S 0 = U SU ∗ .

1.5. Complex Linear Spaces

17

Proof. For such S, U we have T ∗ T = S(U ∗ U )S = S 2 . Thus S = (T ∗ T )1/2 , so S and U are uniquely determined by T . To show existence, define S := (T ∗ T )1/2 and U := T (T ∗ T )−1/2 . Then S is positive, T = U S, and hU x, U yi = hT S −1 x, T S −1 yi = hS 2 S −1 x, S −1 yi = hx, yi. Similarly, U 0 = (T T ∗ )−1/2 T = T (T ∗ T )−1/2 = = (V AV −1 )−1/2 for every positive A and invertible V .

1.5

U , since V A−1/2 V −1

Complex Linear Spaces

The fundamental constructions of exterior algebras and Clifford algebras in this book can be made for linear spaces over more general fields than the real numbers R. We will consider only the field of complex numbers C besides R, which is particularly useful in analysis. We write the complex conjugate of z ∈ C as z c . Given a complex matrix A = (aij )ij , its conjugate transpose is A∗ := (acji )ij , as compared to its transpose At := (aji )ij . Definition 1.5.1. A complex linear space (V, +, ·) is an abelian group (V, +) together with a scalar multiplication C × V → V that is bilinear with respect to the addition operations and a group action of the multiplicative group C∗ = C \ {0} on V. By a complex vector space we shall mean simply a complex linear space, without any interpretation like that in Definition 1.1.2, since this concerns the additive structure of the vector space. Before proceeding with the algebra, an example is in order, to show why complex linear spaces are natural and very useful in analysis. Example 1.5.2 (Time-harmonic oscillations). Consider a quantity f (t, x) that depends on time t ∈ R and position x in some space X. We assume that f takes values in some real linear space. Fixing a basis there, we can assume that f (t, x) ∈ RN . One example is the electromagnetic field in which case N = 6, since it consists of a three-dimensional electric field and a three-dimensional magnetic field. The most convenient way to represent f oscillating at a fixed frequency ω ∈ R is to write f (t, x) = Re(F (x)e−iωt ), for a function F : X → CN , where the real part is taken componentwise. In this way, each component fk (t, x), k = 1, . . . , N , at each point x will oscillate at frequency ω. The complex-valued function F has a very concrete meaning: the absolute value |Fk (x)| is the amplitude of the oscillation of component k at the point x, and the argument arg Fk (x) is the phase of this oscillation. Note that we

Chapter 1. Prelude: Linear Algebra

18

do not assume that the oscillations at different points have the same phase; this happens only for standing waves. Since the complex field has two automorphisms, the identity and complex conjugation, there are two types of dualities that are natural to consider. These correspond to linear and antilinear identification of V 0 and the dual space V ∗ = {θ : V → C ; θ is complex linear} of V. • A complex bilinear duality of two complex linear spaces V 0 and V is a complex bilinear map V 0 × V → C : (v 0 , v) 7→ hv 0 , vi that is nondegenerate. When V 0 = V, we refer to a bilinear duality as a complex bilinear inner product if it is symmetric, that is, if hx, yi = hy, xi. A main difference is that notions like signature are not present in the complex bilinear case since we can normalize −hx, xi = hix, ixi. • A complex sesquilinear duality of V 0 and V, is a nondegenerate pairing (·, ·i such that (v 0 , ·i is complex linear for each v 0 ∈ V 0 and (·, vi is complex antilinear for each v ∈ V . Note the difference in left and right parantheses, which we use to indicate the sesquilinearity. When V 0 = V, we refer to a sesquilinear duality as a complex inner c product if it is symmetric, that is, if (x, yi = (y, xi. A complex inner product is called Hermitian if it is positive definite, that is (u, ui > 0 for all p u ∈ V \{0}. The norm associated with a Hermitian inner product is |u| := (u, ui. The existence of the following types of canonical bases can be derived from the spectral theorem for normal complex linear operators. Proposition 1.5.3 (Complex ON-bases). Let V be a complex linear space. (i) A sesquilinear duality (·, ·i is symmetric if and only if there exists a basis {ei } that is ON in the sense that (ei , ej i = 0 when i 6= j and (ei , ei i = ±1. (ii) A bilinear duality h·, ·i is symmetric in the sense that hv1 , v2 i = hv2 , v1 i if and only if there exists a basis {ei } that is ON in the sense that (ei , ej i = 0 when i 6= j and (ei , ei i = 1. Exercise 1.5.4. (i) Prove that a sesquilinear duality (x, yi is skew-symmetric, that c is, (x, yi = −(y, xi, if and only if i(x, yi is an inner product. (ii) Prove that a bilinear duality h·, ·i is skew-symmetric in the sense that hv1 , v2 i = −hv2 , v1 i if and only if dim V = 2k and there exists a Darboux basis, that is, a basis {ei }ki=1 ∪{e0i }ki=1 in which the only nonzero pairings are he0i , ei i = 1, hei , e0i i = −1, i = 1, . . . , k.

1.5. Complex Linear Spaces

19

We next consider the relation between real and complex linear spaces. We first consider how any complex linear space can be turned into a real linear space, and how to reverse this process. • Let V be a complex linear space. Simply forgetting about the possibility of scalar multiplication by nonreal numbers, V becomes a real linear space, which we denote by V = V. Note that dimC V = 2 dimR V . Besides this real linear structure, V also is equipped with the real linear operator J : V → V : v 7→ iv, which has the property that J 2 = −I. A complex linear map T : V1 → V2 is the same as a real linear map T : V1 → V2 between these spaces regarded as real linear spaces, for which T J1 = J2 T . Given a complex functional θ ∈ V ∗ , the real linear functional V 3 v 7→ Re θ(v) ∈ R belongs to V ∗ . This gives a real linear one-to-one correspondence between V ∗ and V ∗ . In particular, if (·, ·i is a complex inner product on V, taking the real part of the antilinear identification V → V ∗ , we obtain a real inner product hv 0 , viR := Re(v 0 , vi on V , and h·, ·iR is a Euclidean inner product if and only if (·, ·i is a Hermitian inner product. It is possible but less useful to start with a complex bilinear inner product, since this always leads to a real inner product with signature zero. • We can reverse the above argument. Let V be a real linear space equipped with a complex structure, that is, a real linear operator J : V → V such that J 2 = −I. Then (α + βi)v := αv + βJ(v),

v ∈ V, α, β ∈ R,

defines a complex scalar multiplication, which turns V into a complex linear space V. If dim V is odd, then no such J exists, since we would then have (det J)2 = det(−I) = (−1)n = −1, which is unsolvable over R. If dim V is even, there are infinitely many complex structures among which to choose. Indeed, if {e1 , . . . , e2k } is any basis, then J

k k X X (α2j−1 e2j−1 + α2j e2j ) = (−α2j e2j−1 + α2j−1 e2j ) j=1

is one such complex structure.

j=1

Chapter 1. Prelude: Linear Algebra

20

If furthermore the complex structure J on V is an isometry J ∗ J = I, or equivalently skew-adjoint, then polarizing hv 0 , viR = Re(v 0 , vi recovers the sesquilinear duality (v 0 , vi = hv 0 , viR − ihv 0 , JviR . We next consider how any real linear space can be embedded in a complex linear space, and how to reverse this process. • Let V be a real linear space. Define the real linear space V ⊕ V , and consider V as a subspace of V ⊕ V by identifying v ∈ V and (v, 0) ∈ V ⊕ V . Define the standard complex structure J(v1 , v2 ) := (−v2 , v1 ),

(v1 , v2 ) ∈ V ⊕ V.

Then the complex linear space Vc := (V ⊕ V, J) is called the complexification of V . The complex vector (v1 , v2 ) is usually written as the formal sum v1 +iv2 , so that complex scalar multiplication becomes (α + βi)(v1 + iv2 ) = (αv1 − βv2 ) + i(αv2 + βv1 ). The complexification Vc of a real linear space V is a complex linear space V, with dimC Vc = dimR V , which comes with two canonical real linear c subspaces. Defining a complex conjugation operator (x + iy) := x−iy, this is a complex antilinear operation that fixes V ⊂ Vc and squares to the identity. A real linear map T : V → V 0 extends to a complex linear map Tc : Vc → Vc0 by complexification: Tc (v1 + iv2 ) := T v1 + iT v2 . The complexification (V ∗ )c of the real dual can in a natural way be identified with the complex dual (Vc )∗ of the complexification, through the complex linear invertible map given by hθ1 +iθ2 , v1 +iv2 i := hθ1 , v1 i−hθ2 , v2 i+ i(hθ1 , v2 i + hθ2 , v1 i). In particular, if h·, ·i is a duality on V , by complexifying the linear identification V → V ∗ , we obtain a complex bilinear inner product h·, ·iC on Vc , described by Vc 7→ (V ∗ )c = (Vc )∗ . Concretely, hu0 + iv 0 , u + iviC := hu0 , ui − hv 0 , vi + i(hv 0 , ui + hu0 , vi). Alternatively, we may equip Vc with the complex (sesquilinear) inner product (u0 + iv 0 , u + iviC := hu0 , ui + hv 0 , vi + i(−hv 0 , ui + hu0 , vi), which is Hermitian if h·, ·i is Euclidean. We can also complexify a real associative algebra (A, +, ∗, 1), by complexifying the linear space A as well as the bilinear product ∗, to obtain an associative algebra Ac over the complex field.

1.6. Comments and References

21

• We can reverse the above argument. Let V be any complex linear space equipped with a real structure, that is, a complex antilinear operator V → V : z 7→ z c c

such that (z c ) = z. Then V is isomorphic to the complexification Vc of the real subspace V := {z ∈ V ; z c = z} through 1 V 3 z = x + iy ←→ (x, y) = 21 (z + z c ), 2i (z − z c ) ∈ Vc . Clearly, on any complex linear space there are infinitely many real structures. A important advantage over the real theory is that every complex linear operator has an eigenvector, by the fundamental theorem of algebra. For a normal operator, that is, if T ∗ T = T T ∗ on a Hermitian space, we can iterate this result on the orthogonal complement, yielding an ON-basis of eigenvectors. If we apply these results to the complexification of a real linear operator, we obtain the following real result. • Every real linear map T : V → V has either an eigenvector or an invariant two-dimensional subspace. More precisely, in the latter case there exist α, β ∈ R, with β 6= 0, and linearly independent vectors v1 , v2 ∈ V such that T (v1 ) = αv1 − βv2 ,

T (v2 ) = βv1 + αv2 .

• Let T : V → V be a real linear normal operator, that is, T ∗ T = T T ∗ , on a Euclidean space. Then, there exists an ON-basis in which the matrix for T is block diagonal, with 2 × 2 and 1 × 1 blocks along the diagonal. Examples include isometries and skew-symmetric maps.

1.6

Comments and References

1.1 A reference for basic algebraic structures such as groups, rings, fields, vector spaces, and algebras is Nicholson [73]. 1.2 I thank Mats Aigner, Link¨oping University, for suggesting the notation for dualities used in this book, which incorporates the dual space of linear functionals as a special case. 1.3 Spacetime in the sense of Definition 1.3.5 was first constructed by Hermann Minkowski (1864–1909), for Maxwell’s equations. He had Albert Einstein as a student and realized later when Einstein created his special theory of relativity that this could be modeled mathematically by a four-dimensional spacetime. A reference for the theory of relativity is Rindler [79]. The most common sign convention for spacetime in the literature is + − −−, that is, opposite to the sign − + ++ used in this book.

22

Chapter 1. Prelude: Linear Algebra

1.4 Tensors and tensor products appear in the work of J.W. Gibbs (1839–1903), although some specific examples of tensors such as the Cauchy stress tensor and the Riemann curvature tensor had been found earlier. A reference for our construction of tensor products, using the universal property, is Greub [46]. 1.5 We use of the word Hermitian as the complex analogue of Euclidean, with a meaning of positivity. However, in many contexts in the literature, Hermitian refers to the conjugate-symmetry, without any implied positivity. The proof of Proposition 1.5.3(ii) uses a variant of the spectral theorem known as the Autonne–Takagi factorization. An equivalent way to define the complexification Vc of a real linear space V , which is standard but not used in this book, is as the tensor product Vc := V ⊗ C of real linear spaces.

Chapter 2

Exterior Algebra Prerequisites: This chapter is where this book starts, and everything else in the book depends on it, except for Section 2.9, which is not needed elsewhere. Chapter 1 is meant to be used as a reference while reading this and later chapters. Otherwise, a solid background in linear algebra should suffice. Section 2.4 requires a small amount of analysis. Road map: We all know the algebra of vectors, the one-dimensional oriented/directed arrows. Here we construct and develop the algebra for bivectors, the two-dimensional oriented objects, and 3-vectors, the three-dimensional oriented objects, and so on, which live in n-dimensional affine space. In total we obtain a linear space of dimension 2n containing all the multivectors in the space, referred to as the exterior algebra. Algebraically, multivectors are in some sense nothing but rectangular determinants, but it is important to understand the geometry to be able to use the theory. Sections 2.2 and 2.4 aim to convey the geometric meaning of multivectors to the reader. Most applications use Euclidean space, but for a number of practical reasons, including applications to Minkowski spacetime, we allow for more general inner product spaces and dualities. The exterior product u ∧ v can be seen as a higherdimensional generalization of the vector product, but in a more fundamental way, so that it corresponds to the direct sum [u] ⊕ [v] of subspaces [u] and [v]. Since ∧ is noncommutative, two different but closely related dual products come into play, the right and left interior products v x u and u y v, which geometrically correspond to the orthogonal complement [u]⊥ ∩ [v] of subspace [u] in a larger subspace [v]. When the larger space is the whole space, we have the Hodge star map, which corresponds to taking orthogonal complements of subspaces. © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_2

23

Chapter 2. Exterior Algebra

24

Developing the algebra of these products of multivectors, we obtain a geometric birds-eye view on various algebraic results in linear algebra such as identities for the vector product, Cramer’s rule, and the cofactor formula for inverses of linear maps, and expansion rules for determinants. Highlights: • Simple k-vectors ↔ k-dimensional subspaces: 2.2.3 • Factorization algorithm for k-vectors: 2.2.8 • Geometry of Cramer’s rule: 2.3.6 • Algebra for interior product: 2.6.3 • Geometry of cofactor formula: 2.7.1 • Anticommutation relation between exterior and interior products: 2.8.1

2.1

Multivectors

Let us fix an affine space (X, V ) of dimension 1 ≤ n < ∞. The letter n will be the standard notation for the dimension of the vector space V . We set out to construct, for any 0 ≤ k ≤ n, a linear space ∧k V of k-vectors in X. A k-vector w ∈ ∧k V is to be interpreted as an affine k-dimensional object in X determined by its orientation and k-volume. When k = 1, then ∧1 V := V and 1-vectors are simply vectors in X, or oriented 1-volumes. We build k-vectors from vectors using certain multilinear maps. See Definition 1.4.1. Lemma 2.1.1. For a multilinear map M : V × · · · × V → L, the following are equivalent: (i) M (v1 , . . . , vk ) = 0 whenever {v1 , . . . , vk } are linearly dependent. (ii) M (v1 , . . . , vk ) = 0 whenever vi = vj for some i 6= j. (iii) M is alternating, that is, for all 1 ≤ i < j ≤ k and vectors {vm }, we have M (v1 , . . . , vi , . . . , vj , . . . , vk ) = −M (v1 , . . . , vj , . . . , vi , . . . , vk ). Proof. That (i) implies (ii) is clear, as is (iii) implies (ii). P For (ii) implies (i), recall that if {v1 , . . . , vk } are linearly dependent, then vj = i6=j xi vi for some j. Doing this substitution and expanding with multilinearity shows that all terms have two identical factors. This proves (i), using (ii). Finally, to prove (ii) implies (iii), note that 0 = M (v1 , . . . , vi + vj , . . . , vi + vj , . . . , vk ) = M (v1 , . . . , vi , . . . , vj , . . . , vk ) + M (v1 , . . . , vj , . . . , vi , . . . , vk ), from which (iii) follows.

2.1. Multivectors

25

The theory of k-vectors can be thought of as a theory of rectangular determinants. Let us start with a definition of the usual concept of a (quadratic) determinant from linear algebra. Proposition 2.1.2 (Determinant). There exists a unique multilinear map det : Rn × · · · × Rn → R, where the number of copies of Rn is n, with the following properties. (A) If the vectors {v1 , . . . , vn } are linearly dependent, then det(v1 , . . . , vn ) = 0. (B) If {ei } is the standard basis, then det(e1 , . . . , en ) = 1. Let us sketch the proof of this well-known fact. P If det exists, then (A), (B), and multilinearity show that for any vectors vj = i αi,j ei , we must have det(v1 , . . . , vn ) =

n X s1 =1

···

n X

αs1 ,1 · · · αsn ,n (s1 , . . . , sn ),

(2.1)

sn =1

where (s1 , . . . , sn ) is zero if an index is repeated and otherwise denote the sign of the permutation (s1 , . . . , sn ) 7→ (1, . . . , n). Hence uniqueness is clear. Note now that if such det exists, then necessarily it must satisfy (2.1). Thus all that remains is to take (2.1) as the definition and verify properties (A) and (B). Note carefully this frequently useful technique to prove existence, using inspiration from a uniqueness proof. P If vj = i αi,j ei and A = (αi,j ), then we use the standard notation α1,1 .. det(v1 , . . . , vn ) = det(A) = . αn,1

··· .. . ···

α1,n .. . . αn,n

We now generalize this construction to fewer than n vectors, replacing the range R by a more general linear space L. Proposition 2.1.3. Let 2 ≤ k ≤ n and let {e1 , . . . , en } be a basis for V . Then there exist a linear space L and a multilinear map ∧k : V × · · · × V → L, where the number of copies of V is k, that satisfy the following properties. (A) If the {v1 , . . . , vk } are linearly dependent, then ∧k (v1 , . . . , vk ) = 0. (B) The set {∧k (es1 , . . . , esk )}s1 0, [b] = W1 , and T v = exp(b/2)v exp(−b/2) for all v ∈ W . (iii) A parabolic rotation: T has one eigenvector along the light cone. In this case W0 ⊂ W0⊥ ⊂ W , where W0 is a line spanned by a singular vector fixed by T , and W0⊥ is the plane tangent to the light cone along the b 2 W such line W0 , and W0⊥ is invariant under T . In this case there is b ∈ 4 2 ⊥ that b = 0, [b] = W0 , and T v = exp(b/2)v exp(−b/2) for all v ∈ W .

130

Chapter 4. Rotations and M¨obius Maps

Proof. Let T ∈ SO+ (W ) \ {I}. By the Cartan–Dieudonn´e theorem (Theorem 4.1.3), and since T 6= I is a rotation, there are linearly independent vectors v1 , v2 such that T x = v1 v2 x(v1 v2 )−1 . Note that since v1 v2 is orthochronous, we have hv1 , v2 i2 − (v1 ∧ v2 )2 = hv1 , v2 i2 + hv1 ∧ v2 i2 > 0. Three cases are possible: (i) The space [v1 ∧ v2 ] is a space-like plane. In this case, let {e0 , e1 , e2 } be an ON-basis such that {e1 , e2 } is an ON-basis for [v1 ∧ v2 ]. Then there exist α, φ ∈ R such that exp(φe12 ) = cos φ + e12 sin φ = αv1 v2 . This gives conclusion (i) with b := φe12 , W1 := [e0 ], and W2 := [e12 ]. (ii) The space [v1 ∧ v2 ] is a time-like plane. In this case, let {e0 , e1 , e2 } be an ON-basis such that {e0 , e1 } is an ON-basis for [v1 ∧ v2 ]. Then there exist α, φ ∈ R such that exp(φe01 ) = cosh φ + e01 sinh φ = αv1 v2 . This gives conclusion (ii) with b := φe01 , W1 := [e01 ], and W2 := [e2 ]. (iii) The space [v1 ∧ v2 ] is a light-like plane. In this case, let {e0 , e1 , e2 } be an ON-basis such that {e0 + e1 , e2 } is a basis for [v1 ∧ v2 ]. Then there exist α, φ ∈ R such that exp(φ(e0 + e1 )e2 ) = 1 + φ(e0 + e1 )e2 = αv1 v2 . This gives conclusion (iii) with b := φ(e0 + e1 )e2 , W0 := [e0 + e1 ], and W0 = [(e0 + e1 ) ∧ e2 ]. Our next objective is to show that every orthochronous rotation decomposes into two-dimensional elliptic and hyperbolic rotations and three-dimensional parabolic rotations. Proposition 4.4.7 (Spacetime rotations). Let T ∈ SO+ (W ) be an orthochronous rotation in a spacetime W . Then there exist a time-like subspace W−1 of dimension two or three, two-dimensional space-like subspaces W1 , . . . , Wk , and a space-like subspace W0 such that W splits as a direct sum of orthogonal subspaces W = W−1 ⊕ W1 ⊕ · · · ⊕ Wk ⊕ W0 , where all the subspaces are invariant under T , T |W−1 is a Lorentz boost if dim W−1 = 2 or a parabolic rotation if dim W−1 = 3, T |Wj is a Euclidean rotation for j ≥ 1, and T |W0 = I. ˜ := N(T − I). If W ˜ is space- or Proof. We first split off the subspace W0 . Define W ˜ ˜ of ˜ time-like, let W0 := W . If W is light-like, take any space-like subspace W0 ⊂ W ˜ one dimension less than W , which is possible by Proposition 1.3.7. Looking into W0⊥ , replacing W by W0⊥ , we assume from now on that N(T − I) is either {0} or is a line along the light cone. Complexification as in Section 1.5 shows that there exists a one- or twodimensional subspace W 0 that is invariant, and we have the following possible cases:

4.4. Spacetime Rotations

131

1. A space-like plane W 0 , where det T |W 0 = 1. 2. A space-like line W 0 , where T = −I due to the assumption on N(T − I). 3. A time-like plane W 0 , where det T |W 0 = 1. A time-like line is not possible because T |W 0 6= I, by the assumption on N(T − I), and T 6= −I, since T is orthochronous. 4. A light-like plane or line. However, in a plane, any vector along the intersection with the light cone must be an eigenvector, since T is an isometry. Assume, therefore, that T v1 = λv1 , where v1 ∈ Wl . Since T is orthochronous, λ > 0, and we may assume that λ = 1. If this is not the case, since 0 = det(T − λ) = det(T ∗ − λ) = det(T −1 − λ); we see that also λ−1 must be an eigenvalue to T , giving another eigenvector not parallel to v1 . But λ−1 6= 1 can be an eigenvalue of an isometry only if the eigenvector belongs to Wl , which would give us two eigenvectors on Wl , and hence an invariant time-like plane, and we are in the previous case. To summarize, in this last case, we may assume that there exists a line along Wl spanned by v1 such that T v1 = v1 . In all but the last case, the invariant subspace is nondegenerate, and we can split W orthogonally into a direct sum W = W 0 ⊕ (W 0 )⊥ , as in Proposition 1.3.4, with both subspaces invariant under T . In the last case, we need to find a larger invariant and nondegenerate subspace W 0 containing v1 . Let E1 := span{v1 } = N(T − I), and note as in Proposition 1.3.7 that E1⊥ is the tangent hyperplane to the light cone Wl along E1 . We have R(T − I) = N(T ∗ − I)⊥ = N(T −1 − I)⊥ = E1⊥ 3 v1 . Therefore we have v2 such that (T − I)v2 = v1 . Let E2 := span{v1 , v2 }, and note that v2 is not parallel to v1 , since T v1 = v1 . We note that 0 = hv1 i2 = hT v2 − v2 i2 = 2hv2 i2 − 2hT v2 , v2 i = −2hv1 , v2 i, so v2 ∈ E1⊥ . By Proposition 1.3.7, E2 is a degenerate subspace, and we proceed by taking v3 such that (T −I)v3 = v2 , which is possible, since v2 ∈ E1⊥ = R(T −I). Let E3 := span{v1 , v2 , v3 }. We note that T −I : E2 → E1 , so that T ∗ −I = (I −T )T −1 : / E1⊥ . Since / E2⊥ because hv2 i2 6= 0, it follows that T v3 ∈ E1⊥ → E2⊥ . Since v2 ∈ ⊥ ⊥ ⊥ T : E1 → E1 , we have v3 ∈ / E1 . Thus we have constructed an invariant and nondegenerate subspace W 0 = E3 by Proposition 1.3.7. We proceed recursively and split the invariant complements (W 0 )⊥ . Collecting all these orthogonal subspaces produced, we note that the only subspace for which det TW 0 = −1 is that of the space-like lines from 2. Since det T = 1, the number of such lines must be even, and we can evenly collect them in invariant

Chapter 4. Rotations and M¨obius Maps

132

2-planes. Note that these planes are nondegenerate and cannot be time-like, since T is orthochronous. Hence they are space-like, and T acts in them as Euclidean rotation through an angle π. This proves the orthogonal splitting of W . As for Euclidean rotations, this orthogonal decomposition implies the following representation theorem for rotations of spacetime. Theorem 4.4.8 (SO+ surjectivity). Let W be a spacetime. If T ∈ SO+ (W ), then there exist A ∈ SO(W ) and b ∈ Spin(W ) such that T v = exp(A)v = exp(b/2)v exp(−b/2),

v ∈ W.

Thus each T ∈ SO+ (W ) belongs to a one-parameter group of rotations T (φ)v := exp(φA)v = exp(φb/2)v exp(−φb/2), and the exponential map exp : SO(W ) → SO+ (W ) is surjective. Proof. Let T ∈ SO+ (W ), and split spacetime W = W−1 ⊕W1 ⊕· · ·⊕Wk ⊕W0 as in Proposition 4.4.7. By Proposition 4.4.6 there are bj ∈ Spin(Wj ) such that T |Wj v = exp(bj /2)v exp(−bj /2), v ∈ Wj , j = −1, 1, 2, . . . , k. The bivectors commute, so letting b := b−1 + b1 + · · · + bk , we have T v = exp(b−1 /2) exp(b1 /2) · · · exp(bk /2)v exp(−bk /2) · · · exp(−b1 /2) exp(−b−1 /2) = exp(b/2)v exp(−b/2). Also, if Av := b x v, then exp(A) = T .

We end this section with a number of exercises, which completes the extension of the Euclidean results to spacetime. Exercise 4.4.9. Prove the following by modifying the proofs of Proposition 4.4.7, letting N(A) play the role of N(T − I). Let A ∈ SO(W ) be a skew-symmetric map in a spacetime W . Then there exist a time-like subspace W−1 of dimension two or three, two-dimensional spacelike subspaces W1 , . . . , Wk , and a space-like subspace W0 such that W splits as a direct sum of orthogonal subspaces W = W−1 ⊕ W1 ⊕ · · · ⊕ Wk ⊕ W0 , where all the subspaces are invariant under T . If dim W−1 = 2, then the matrix 0 φ of A in an ON-basis {e0 , e1 } is , for some φ ∈ R, that is, Av = φe0,1 x v. If φ 0 dim W−1 = 3, then there is an ON-basis {e0 , e1 , e2 } in which the matrix of A is 0 0 φ 0 0 φ φ −φ 0 for some φ ∈ R, that is, Av = φ(e0 + e1 )e2 x v. The restrictions T |Wj , j ≥ 1, are all Euclidean skew maps, and T |W0 = I.

4.4. Spacetime Rotations

133

Exercise 4.4.10. Deduce the following from the above results for spacetime rotations and skew maps, similar to what was done in the Euclidean case in Proposition 4.3.9. Let W be a spacetime. (i) If b ∈ 42 W is a bivector, then there exists an ON-basis {ei } in which b has the form b = φ0 e0 e1 + φ1 e2 e3 + φ2 e4 e5 + · · · + φk e2k e2k+1 or b = φ0 (e0 + e1 )e2 + φ1 e3 e4 + φ2 e5 e6 + · · · φk e2k+1 e2k+2 . (ii) If q ∈ Spin+ (W ) is an orthochronous rotor, then there exist an ON-basis {ei }, a sign = ±1, and angles 0 < φj ≤ π such that q has the form q = cosh(φ0 /2) + e0,1 sinh(φ0 /2) cos(φ1 /2) + e2,3 sin(φ1 /2) · · · cos(φk /2) + e2k,2k+1 sin(φk /2) , or q = 1 + φ20 (e0 + e1 )e2 cos(φ1 /2) + e3,4 sin(φ1 /2) · · · cos(φk /2) + e2k+1,2k+2 sin(φk /2) . In contrast to Theorem 4.4.8, the exponential map exp : Spin(W ) → Spin+ (W ) is surjective in spacetime only when dim W ≥ 5. In dimensions two and three, exp is far from being surjective, whereas in dimension 4, it is only half of the orthochronous rotors representing parabolic rotations that cannot be represented by a bivector, as the following clarifies. Exercise 4.4.11. Let W be a two-dimensional spacetime, and let j ∈ 42 W be such that j 2 = 1. Show that n o p Spin+ (W ) = α + βj ; α = ± 1 + β 2 n o p % exp(Spin(W )) = α + βj ; α = 1 + β 2 . Exercise 4.4.12. Let W be a three-dimensional spacetime. Then Spin+ (W ) = {α + β1 j1 + β2 j2 + β3 j3 ; α2 + β32 = β12 + β22 + 1} is a connected Lie group by Proposition 4.4.4, where {e0 , e1 , e2 } is an ON-basis with e20 = −1 and j1 := e01 , j2 := e02 , and j3 := e12 . Show that exp(Spin(W )) contains all orthochronous rotors except those α + β1 j1 + β2 j2 + β3 j3 ∈ Spin+ (W ) for which β32 ≤ β12 + β22 and α ≤ −1.

Chapter 4. Rotations and M¨obius Maps

134

Exercise 4.4.13. Let W be a four-dimensional spacetime as in relativity. Show that all q ∈ Spin+ (W ) can be written q = exp(b/2) for some b ∈ Spin(W ), except those q that are of the form q = −(1 + b), where b ∈ 42 W satisfies b2 = 0. Theorem 4.4.14 (Spin+ surjectivity). Let W be a spacetime with dim W ≥ 5. If q ∈ Spin+ (W ), then there exists b ∈ Spin(W ) such that q = exp(b/2). Thus each q ∈ Spin+ (W ) belongs to a one-parameter group of rotors q(φ) := exp(φb/2), and the exponential map exp : Spin(W ) → Spin+ (W ) is surjective. Proof. If q ∈ Spin+ (W ), consider the corresponding rotation T v := qvq −1 . Theorem 4.4.8 shows that T v = exp(b/2)v exp(−b/2), where b = b−1 + b1 + · · · + bk for some commuting simple bivectors bj . Thus q = ± exp(b/2). In case of a minus psign, we can eliminate this as follows. If k ≥ 1, then b2k < 0 and we let φ := 2π/ −b2k . This gives exp((b − φbk )/2) = (−q)(−1) = q, since b and bk commute. If b = b−1 , then since dim W ≥ 5, there exists b1 ∈ 42 W that commutes with b and satisfies b21 = −π 2 . In this case exp((b − b1 )/2) = (−q)(−1) = q also.

4.5

Fractional Linear Maps

In this section we study a fundamental class of maps of a Euclidean space (X, V ) that are closely related to isometries. Definition 4.5.1 (Conformal map). Let (X, V ) be a Euclidean space and let D ⊂ X be an open set. A differentiable map f : D → X is called conformal in D if at each point x ∈ D, the derivative f x : V → V is a nonzero multiple of an isometry, that is, if there exists λ : D → R \ {0} such that λ(x)f x is an isometry for all x ∈ D. Example 4.5.2. We have the following four classes of basic conformal maps in a Euclidean space X. (i) Fix an origin in X, and identify X and V . Then the isometry x 7→ qxq −1 , b . x ∈ X, is conformal for every q ∈ 4V (ii) For every v ∈ V , translation x 7→ x + v, x ∈ X, by the vector v is a conformal map. (iii) Fix an origin in X, and identify X and V . For every c ∈ R \ {0}, dilation x 7→ cx, x ∈ X, with scale factor c is a conformal map. (iv) Fix an origin in X, and identify X and V . Then inversion in the unit sphere x 7→ is a conformal map.

1 x = , x |x|2

x ∈ X \ {0},

4.5. Fractional Linear Maps

135

Exercise 4.5.3. Let f (x) = x−1 . Show that |x|2 f x is reflection in the hyperplane orthogonal to x, and in particular that f is conformal with derivative f x (h) = −x−1 hx−1 ,

h ∈ V, x 6= 0.

Clearly compositions of conformal maps are conformal, on appropriate domains of definition. Thus we have the following group consisting of conformal maps. Definition 4.5.4 (Fractional linear map). Let (X, V ) be a Euclidean space. A map f : X → X is said to be a fractional linear map if it is a finite composition of isometries, translations, dilations, and inversions. In studying fractional linear maps it is convenient to extend the Euclidean space V by adding the point at infinity ∞. More precisely, we embed V in a Euclidean space V∞ that is one dimension larger. A vector e∞ ∈ V∞ with |e∞ | = 1 is fixed and V is identified with the hyperplane orthogonal to e∞ . The extended Euclidean space V is the unit sphere in V∞ , where the north pole e∞ is identified with the point at infinity ∞. The remaining points V \ {e∞ } are identified with V through stereographic projection. Definition 4.5.5 (Stereographic projection). Stereographic projection is the map V \ {e∞ } 3 x = x0 + x∞ e∞ 7→ x ∈ V, for x0 ∈ V and x∞ ∈ [−1, 1), where the projection x is defined as the intersection of V and the straight line in V∞ that intersects V at e∞ and x. See Figure 4.4. We may also refer to the inverse map x 7→ x as stereographic projection. Exercise 4.5.6. Prove the following explicit formulas for the stereographic projection and its inverse. x=

x0 , 1 − x∞

x=

2 |x|2 − 1 x+ 2 e∞ . +1 |x| + 1

|x|2

In particular, the south pole −e∞ projects onto 0 ∈ V and the equator x∞ = 0 projects onto the unit sphere |x| = 1. After adding the point at infinity ∞ to V , the inversion map x 7→ x−1 becomes a smooth map V → V. Indeed, it is seen to correspond to the equatorial reflection x = x0 + x∞ e∞ 7→ x0 − x∞ e∞ = eb∞ xe∞ , where 0 ↔ −e∞ is mapped onto ∞ ↔ e∞ and vice versa. We also extend isometries, translations, and dilation to continuous maps of the extended space V, identified by V through stereographic projection, that fixes the point at infinity e∞ . Thus all fractional linear maps are homeomorphisms as maps V → V.

136

Chapter 4. Rotations and M¨obius Maps

Figure 4.4: Stereographic projection. Proposition 4.5.7 (Mapping of spheres). Let f : V ∪{∞} → V ∪{∞} be a fractional linear map, and let S ⊂ V ∪ {∞} be either a hypersphere, that is, of the form S = {x ∈ V ; |x − a|2 = r2 }, or a hyperplane, that is, of the form S = {x ∈ V ; hx − a, bi = 0} ∪ {∞}. Then the image f (S) is a hypersphere if f −1 (∞) ∈ / S, and f (S) is a hyperplane if f −1 (∞) ∈ S. Proof. It suffices to show that a hypersphere or hyperplane is mapped onto either a hypersphere or a hyperplane, since hyperspheres cannot be unbounded and hyperplanes cannot be bounded. It also suffices to prove this for isometries, translations, dilations, and inversions, and only the last case needs proof. Consider x 7→ x−1 = y and a hypersphere |x − a|2 = r2 . This is mapped onto the points satisfying |1 − ay|2 = r2 |y|2 , where we have used Lagrange’s identity for the Clifford product. This yields 1 + |a|2 |y|2 − 2ha, yi = r2 |y|2 . If |a| = r, that is, ∞−1 = 0 ∈ S, then this is a hyperplane ha, yi = 1/2. If |a| 6= r, then this is the hypersphere |y − a/(|a|2 − r2 )|2 = r2 /(|a|2 − r2 )2 . On the other hand, consider a hyperplane ha, xi = b. This is mapped onto the points satisfying ha, yi = b|y|2 . If b = 0, this is a hyperplane, and if b 6= 0, this is the hypersphere |y − a/(2b)|2 = |a|2 /(4b2 ). This proves the proposition. This result can also be used in the larger space V∞ , where it turns out that the stereographic projection is a restriction of a fractional linear map. Proposition 4.5.8. The map V∞ \ {e∞ } → V∞ \ {e∞ } : y 7→ (e∞ y + 1)(y − e∞ )−1

(4.4)

is a self-inverse fractional linear map. Its restriction to V coincides with the stereographic projection V → V , and its restriction to V coincides with the inverse V → V of the stereographic projection.

4.5. Fractional Linear Maps

137

Under the identification V ∪ {∞} ↔ V, hyperspheres in V correspond to hyperspheres in V not passing through e∞ , and hyperplanes in V correspond to hyperspheres in V passing through e∞ . By a hypersphere in V we mean a nontangential intersection of V and a hyperplane in V∞ . Proof. The map (4.4) can be written y 7→ (e∞ (y − e∞ ) + 2)(y − e∞ )−1 = e∞ + 2(y − e∞ )−1 , from which it is seen that it is a self-inverse fractional linear map. If y = x ∈ V \ {e∞ }, then |x − e∞ |2 = 2 − 2hx, e∞ i = 2(1 − x∞ ), and thus (e∞ x + 1)(x − e∞ ) e∞ − e∞ xe∞ + x − e∞ = 2(1 − x∞ ) 2(1 − x∞ ) 0 x = . 1 − x∞

(e∞ x + 1)(x − e∞ )−1 =

On the other hand, if y = x ∈ V , then |x − e∞ |2 = |x|2 + 1, and thus (e∞ x + 1)(x − e∞ ) |x|2 + 1 2 e∞ |x| + x + x − e∞ = |x|2 + 1 2 |x|2 − 1 = x + e∞ . |x|2 + 1 |x|2 + 1

(e∞ x + 1)(x − e∞ )−1 =

To prove the mapping properties of the stereographic projection, we use Proposition 4.5.7 on the map (4.4). Given a hypersphere S : |x − a|2 = r2 in V , we view this as the intersection of V and the hypersphere S : |y − (a + te∞ )|2 = r2 + t2 passing through e∞ , where t := 12 (|a|2 + 1 − r2 ). Then (4.4) maps S onto a hyperplane not passing through e∞ . Similarly, given a hyperplane S : ha, xi = b in V , we view this as the intersection of V and the hyperplane S : ha + be∞ , yi = b passing through e∞ . Then (4.4) maps S onto a hyperplane passing through e∞ . This proves the proposition. Exercise 4.5.9. Find a fractional linear map of three-dimensional space that maps the unit sphere onto itself and maps (0, 0, 1/2) to the origin. Definition 4.5.10 (M¨obius map). A diffeomorphism f : V → V is said to be a M¨ obius map if it maps hyperspheres in V onto hyperspheres in V. A map f : V → V is called a M¨ obius map if it extends to a M¨obius map of V. We denote the group of M¨ obius maps by M¨ob(V), or equivalently M¨ob(V ). Definition 4.5.11 (Global conformal map). A differentiable map f : V → V, or the corresponding map f : V ∪{∞} → V ∪{∞}, is said to be a global conformal map if f is conformal at each p ∈ V. That f : V ∪ {∞} → V ∪ {∞} is globally conformal

138

Chapter 4. Rotations and M¨obius Maps

means that each point in V has a neighborhood, in which either f (x) or 1/f (x) is a well-defined conformal map, and that 0 ∈ V has a neighborhood where either f (1/x) or 1/f (1/x) is a well-defined conformal map. We here let 1/0 = ∞ and 1/∞ = 0, and well defined at x means that f (x) 6= ∞. Theorem 4.5.12. Assume dim V ≥ 2. For a diffeomorphism f : V → V, the following are equivalent: (i) f is a fractional linear map. (ii) f is a M¨ obius map. (iii) f is a global conformal map. Note that this is a somewhat remarkable result, in that (i) is an algebraic statement, (ii) is a geometric statement, and (iii) is an analytic statement. Proof. We have seen that fractional linear maps are M¨obius maps as well as global conformal maps. To show that all M¨obius maps are conformal, assume that f : V → V maps hyperspheres to hyperspheres. Let a ∈ V. Composing f by a suitable fractional linear map, we may assume that a, f (a) 6= ∞. We have f (a + h) − f (a) → f a (h),

→ 0,

uniformly for |h| = 1. Therefore the linear image f a (S) of the unit sphere S ⊂ V is the uniform limit of spheres, and is there itself a sphere. This is possible only if f a is a multiple of an isometry, proving that f is conformal. To show that all global conformal maps are fractional linear, assume that f : V → V is conformal. For dim V ≥ 3 it follows from Liouville’s theorem (Theorem 11.4.2) on conformal maps that f is a fractional linear map. If dim V = 2, by composing f with a suitable fractional linear map, we may assume that f (∞) = ∞ and that f is orientation preserving. Identifying V and C, we have in this case a bijective entire analytic function f : C → C, and 1/f (1/z) is analytic around z = 0. It is well known from complex analysis that this implies that f (z) = az + b, z ∈ C, for some a, b ∈ C, a 6= 0. Our next objective is to develop the algebra of fractional linear maps. Exercise 4.5.13. Generalizing the algebra of fractional linear maps in complex analysis, we represent a map 4V 3 w 7→ (aw + b)(cw + d)−1 ∈ 4V, a where a, b, c, d ∈ 4V are constants, by the matrix c composition of matrices corresponds to composition of Be careful: the algebra is not commutative!

w ∈ V, b ∈ 4V (2). Show that d the corresponding maps.

4.5. Fractional Linear Maps

139

The following definition gives conditions on a, b, c, d such that (aw + b)(cw + d)−1 preserves the vectors V ⊂ 4V , analogous to the Clifford cone for isometries. In Section 4.6 we will show that this rather ad hoc looking definition is very natural indeed and closely related to the Clifford group for spacetime. Definition 4.5.14 (Vahlen matrices). Let V be a Euclidean space. A matrix M = a b ∈ 4V (2) is called a Vahlen matrix if c d b ∪ {0}, (i) a, b, c, d ∈ 4V (ii) ab, cd ∈ V = 41 V , (iii) ∆(M ) := ad − bc ∈ R \ {0} = 40 V \ {0}. b (2) to be the set of Vahlen matrices in 4V (2). Define the Vahlen cone 4V b (2) form a multiplicative group in 4V (2), Lemma 4.5.15. The Vahlen matrices 4V b and the determinant satisfies ∆(M1 M2 ) = ∆(M1 )∆(M 2 ) for M1 , M2 ∈ 4V (2). If d −b b (2) is a Vahlen matrix, then M −1 = 1 M ∈ 4V is a Vahlen matrix ∆(M ) −c a and ab = ba ∈ V, ab = ba ∈ V,

cd = dc ∈ V, cd = dc ∈ V,

ac = ca ∈ V, ac = ca ∈ V,

bd = db ∈ V, bd = db ∈ V,

ad − bc = da − cb = da − bc = ad − cb ∈ R \ {0}. Thus either a, d ∈ 4ev V and b, c ∈ 4od V , or a, d ∈ 4od V and b, c ∈ 4ev V . Proof. (i) For the inverse of a Vahlen matrix M , the stated matrix is seen to be a right inverse, hence a left inverse also, which shows that ∆(M ) 0 d −b a b da − bc db − bd = = . 0 ∆(M ) c d −c a −ca + ac −cb + ad The diagonal entries show that ∆(M ) = da − bc. Applying the reversion proves the remaining two determinant formulas. Clearly ba = ab ∈ V . To prove ab ∈ V , we may assume that a 6= 0. In this b and a−1 = a/|a|2 = a/(aa). case V 3 a−1 (ba)a = a−1 b(aa) = ab, since a ∈ 4V Repeating this argument, it suffices to prove that ac ∈ V and bd ∈ V . For the first we may assume that a, c 6= 0. Consider the equation ad − bc = λ ∈ R \ {0}. Multiplying by a−1 from the left and c−1 from the right, we get c−1 d − a−1 b = λ(ca)−1 , from which ca ∈ V follows and thus ac ∈ V . The proof of bd ∈ V is similar. b (2) is closed under multiplication. Consider a (ii) Next we show that 4V product a1 b1 a2 b2 a1 a2 + b1 c2 a1 b2 + b1 d2 M1 M2 = = . c1 d1 c2 d2 c1 a2 + d1 c2 c1 b2 + d1 d2

Chapter 4. Rotations and M¨ obius Maps

140

b ∪ {0}, we may assume that a1 , c2 6= 0. In this case To show that a1 a2 + b1 c2 ∈ 4V −1 a1 a2 + b1 c2 = a1 (a2 c−1 2 + a1 b1 )c2 ,

b ∪ {0}. The proofs for the other three entries are where all factors belong to 4V similar. To show that (b2 a1 +d2 b1 )(a1 a2 +b1 c2 ) ∈ V , it suffices to show that b2 a1 b1 c2 + d2 b1 a1 a2 ∈ V , where we may assume that a2 6= 0, for otherwise c2 is parallel to b2 , by the determinant condition on M2 , and the result follows. We need to show that a2 (b2 a1 b1 c2 + d2 b1 a1 a2 )a2 = (a2 b2 )(a1 b1 )(c2 a2 ) + (a2 d2 )(b1 a1 )(a2 a2 ) = (a2 b2 )(2ha1 b1 , c2 a2 i − (c2 a2 )(a1 b1 )) + (a2 d2 )(a2 a2 )(b1 a1 ) = 2ha1 b1 , c2 a2 i(a2 b2 ) + a2 (−b2 c2 + d2 a2 )a2 (a1 b1 ) is a vector, which is clear. A similar calculation shows that (c1 b2 + d1 d2 )(c1 a2 + d1 c2 ) ∈ V . Finally, we calculate the determinant ∆(M1 M2 ) = (a1 a2 + b1 c2 )(b2 c1 + d2 d1 ) − (a1 b2 + b1 d2 )(a2 c1 + c2 d1 ) = a1 (a2 b2 − b2 a2 )c1 + b1 (c2 d2 − d2 c2 )d1 + a1 (a2 d2 − b2 c2 )d1 + b1 (c2 b2 − d2 a2 )c1 = a1 ∆(M2 )d1 − b1 ∆(M2 )c1 = ∆(M1 )∆(M2 ). Theorem 4.5.16 (Representation by Vahlen matrices). Let V be an n-dimensional a b Euclidean space. If M = is a Vahlen matrix, then c d T : V → V : x 7→ (ax + b)(cx + d)−1 is a well-defined fractional linear map. The map M 7→ T is a surjective homob (2) to the group of fractional linear maps. Its morphism from the Vahlen cone 4V kernel consists of the Vahlen matrices 1 0 (−1)n+1 en 0 , α0 , α1 0 en 0 1 where 0 6= α0 ∈ R, 0 6= α1 ∈ R. Proof. To show that T is a fractional linear map, assume first that c 6= 0. Then (ax + b)(cx + d)−1 = ac−1 (cx + d) + (b − ac−1 d) (cx + d)−1 = ac−1 + (b − acd|c|−2 )(x + c−1 d)−1 c−1 = ac−1 + (bc − ad)c|c|−2 (x + c−1 d)−1 c−1 = ac−1 − (∆(M )/|c|2 ) c(x + c−1 d)−1 c−1 ,

4.5. Fractional Linear Maps

141

using that cd = cd = dc, since cd ∈ V . Thus T is a composition of a translation, an inversion, an isometry, a dilation, and finally a translation. If c = 0, then d−1 = |d|−2 d = |d|−2 ∆(M )a−1 , and (ax + b)(cx + d)−1 = (∆(M )/|d|2 )axa−1 + bd−1 , which is an isometry, a dilation, and a translation. In each case we have a fractional linear map. It follows from Exercise 4.5.13 that M 7→ T is a homomorphism. Surjectivity is clear, since the Vahlen matrices qˆ 0 1 v α 0 0 1 , , , , 0 q 0 1 0 1 1 0 represent isometries, translations, dilations, and inversions respectively. To find the kernel, assume that T x = x for all x ∈ V , that is, (ax+b) = x(cx+ d). Letting x = 0 shows that b = 0, and x = ∞ shows that c = 0. The equation ax = xd can hold for all vectors only if a = α0 + α1 en and d = α0 + (−1)n−1 α1 en , as is seen by expressing a and d in an induced ON-basis and using the fact that if s 6= ∅ and s 6= n, then there exist i ∈ s and j ∈ / s such that es has different commutation relations with ei and ej . Checking the conditions ad ∈ R \ {0} and b in the four possible different dimensions modulo 4 shows that either a, d ∈ 4V α0 = 0 or α1 = 0. This proves the theorem. Example 4.5.17. To find a fractional linear map that has certain desired mapping properties in an n-dimensional Euclidean space, one proceeds as in complex analysis, the only difference being that circles and lines are replaced by hyperspheres and hyperplanes. For example, to find the map (e∞ y + 1)(y − e∞ )−1 = e∞ + 2(y − e∞ )−1 from the properties that it should map the sphere V to the hyperplane V in V∞ , one composes the following maps. First the translation x 7→ x − e∞ , which maps e∞ to 0, followed by inversion x 7→ 1/x, which further maps this point to ∞, and also the translated sphere to the hyperplane x∞ = −1/2. Finally, one dilates and translates by x 7→ 2x + e∞ to obtain the image V of the sphere. In total, this gives 2(y − e∞ )−1 + e∞ = 2 + e∞ (y − e∞ ) (y − e∞ )−1 = (e∞ y + 1)(y − e∞ )−1 . Exercise 4.5.18. Extend Exercise 4.5.3 to cover general fractional linear maps. Use the factorization T in the proof of Theorem 4.5.16 to show that every fractional linear map T is conformal, with derivative T x (h) =

∆(M ) (cx + d)h(cx + d)−1 . |cx + d|2

142

Chapter 4. Rotations and M¨obius Maps

The general theory for fractional linear maps of Euclidean spaces can be expressed with complex numbers in the plane and with quaternions in threedimensional space using the standard geometric representations of C and H from Section 3.2. Example 4.5.19 (2D M¨obius maps). Let dim V = 2 and fix an ON-basis {e1 , e2 }. This gives an identification V ∈ x = e1 z ←→ e1 x = z ∈ C = 4ev V between vectors x and complex numbers z. On the one hand, if (ax + b)(cx + d)−1 is a fractional linear map such that a, d ∈ C and b, c ∈ V , then in the complex representation of vectors this corresponds to z 7→ e1 (ae1 z + b)(ce1 z + d)−1 = (w1 z + w2 )(w3 z + w4 )−1 , where w1 := e1 ae1 = a, w2 := e1 b, w3 := ce1 , and w4 := d. The conditions on a, b, c, d translate to wi ∈ C and da − cb = w4 w1 − w3 e1 w2 e1 = w1 w4 − w2 w3 ∈ R \ {0}. On the other hand, if a, d ∈ V and b, c ∈ C, then multiplying the two parentheses by e1 from the right shows that the map is z 7→ e1 (a(e1 ze1 ) + be1 )(c(e1 ze1 ) + de1 )−1 = (w1 z + w2 )(w3 z + w4 )−1 , where w1 := e1 a, w2 := b, w3 := c, and w4 := de1 , and it follows that wi ∈ C and w1 w4 − w2 w3 ∈ R \ {0}. These two cases correspond to orientation-preserving and orientation-reversing M¨obius maps respectively. Example 4.5.20 (3D M¨obius maps). Let dim V = 3 and fix a volume element J ∈ 43 V such that J 2 = −1. This gives an identification V 3 x = −Jz = −zJ ←→ z = Jx = xJ ∈ H ∩ 42 V between vectors and pure quaternions, with the Hodge star map. Let (ax + b)(cx + d)−1 be a fractional linear map. In this case we may assume that a, d ∈ 4ev V and b, c ∈ 4od ∈ V , after possibly multiplying all coefficients by J. In quaternion representation, the fractional linear map is z 7→ J(−aJz + b)(−cJz + d)−1 = (q1 z + q2 )(q3 z + q4 )−1 , where q1 := a, q2 := Jb, q3 := −cJ, and q4 := d. The conditions on a, b, c, d translate to qi ∈ H and q1 q 4 + q2 q 3 ∈ R \ {0}. Orientation-preserving maps arise when q1 q 4 + q2 q 3 > 0, and orientation-reversing maps when q1 q 4 + q2 q 3 < 0.

4.6

Mappings of the Celestial Sphere

In this section we prove that there is a 2−1 homomorphism taking spacetime isometries to Euclidean M¨obius maps. To see the connection with Lorentz isometries,

4.6. Mappings of the Celestial Sphere

143

we identify the higher-dimensional Riemann sphere V with the celestial sphere in a spacetime W . Definition 4.6.1 (Celestial sphere). Given a vector space W , the projective space P (W ) is P (W ) := {[x] ; 0 = 6 x ∈ W }, so that an object in P (W ) is a one-dimensional line through the origin in W . Given a subspace V∞ ⊂ W of codimension one and a vector e0 ∈ W \ V∞ , we identify V∞ and the subset P (V∞ − e0 ) ⊂ P (W ) with the injective map V∞ → P (V∞ − e0 ) : v 7→ [v − e0 ]. The subset P (V∞ ) := {[v] ; 0 6= v ∈ V∞ } is the complement of V∞ in P (W ), and is referred to as the hyperplane at infinity. If W is a spacetime, we assume that e0 ∈ Wt+ and that V∞ = [e0 ]⊥ is a spacelike hyperplane. The celestial sphere is the image P (Wl ) of the light cone Wl , which coincides with the unit sphere V in V∞ under the identification P (V∞ −e0 ) ↔ V∞ . See Figure 4.5. The relativity theory interpretation of this is as follows. Consider an inertial observer O with world line spanned by the future-pointing vector e0 . At the event in spacetime represented by the origin, what O can observe is all light sent out by past events, traveling exactly at the speed of light and reaching the eyes of O at the origin. For this to happen, the past events must lie on the past light cone Wl− . If we write the past event as v − te0 ∈ Wl− , v ∈ V∞ , then v represents the direction from which the light ray reaches O, and t = |v| means that the time it took the light to reach O is proportional to the space distance to the past event. Thus the point [v −te0 ] ∈ P (Wl ) represents all superimposed past events, the light from which O sees in direction v. In this way, the celestial sphere P (Wl ) ⊂ V∞ represents what O observes of the universe W at the origin event. Definition 4.6.2 (Induced map of the sphere). Let W be a spacetime. Fix a futurepointing vector e0 ∈ Wt+ and the Euclidean subspace V∞ := [e0 ]⊥ , and identify the celestial sphere P (Wl ) and the unit sphere V in V∞ . Let T ∈ O(W ) be an isometry. Then T maps the light cone Wl onto itself, and thus induces a diffeomorphism TV : V → V. We define this induced map of the celestial sphere by [TV (v) − e0 ] = [T (v − e0 )],

v ∈ V.

As in Section 4.5, we fix a zenith direction e∞ ∈ V, write V := V∞ ∩ [e∞ ]⊥ , and identify V and V ∪ {∞} under stereographic projection. Write TV for the map in V corresponding to TV . Proposition 4.6.3. If T : W → W is an isometry, then the induced map TV : V → obius map, or equivalently, TV V of the celestial sphere is well defined and is a M¨ is a fractional linear map.

144

Chapter 4. Rotations and M¨obius Maps

Figure 4.5: The celestial sphere and the past spacetime light cone.

Proof. Take any hyperplane P ⊂ V∞ , and let PW be the unique hyperplane in W passing through the origin in W and intersecting V∞ − e0 along P . Then the intersection of PW and Wl corresponds to the hypersphere in V, the intersection of P and V. Since T is linear, it maps PW onto another hyperplane T (PW ) passing through the origin in W and intersecting V∞ − e0 along a hyperplane P 0 . Then by the definition of TV , the intersection P 0 ∩ V is the image TV (P ∩ V). This proves that TV maps hyperspheres to hyperspheres. We wish to obtain an algebraic expression for TV , given the spacetime rotor representing T . Proposition 4.6.4 (M¨obius reflection). Fix an ON-basis {e0 , e1 , . . . , en , e∞ } for spacetime W , and consider the reflection T v = −(a + a∞ e∞ + a0 e0 )v(a + a∞ e∞ + a0 e0 )−1 ,

v∈W

in the hyperplane orthogonal to a + a∞ e∞ + a0 e0 ∈ W , where a ∈ V , V = [e∞ ∧

4.6. Mappings of the Celestial Sphere

145

e0 ]⊥ ⊂ V∞ , and a20 6= |a|2 + a2∞ . Then TV (x) = (−ax + a− )(a+ x + a)−1 ,

x ∈ V,

where a+ := a∞ + a0 and a− := a∞ − a0 . Proof. Consider first the induced map on TV and write a0 := a + a∞ e∞ . Let x ∈ V and consider the point −e0 + x ∈ Wl− . This is mapped by T to − (a0 e0 + a0 )(−e0 + x)(a0 e0 + a0 )−1 = λ(a0 + a0 e0 x − a0 e0 + a0 x)(a0 e0 + a0 ) = λ(a20 e0 + a20 x + a0 a0 + a0 a0 xe0 + a0 a0 + a0 e0 xa0 + |a0 |2 e0 + a0 xa0 ) = λ((a20 + 2a0 ha0 , xi + |a0 |2 )e0 + (a20 x + 2a0 a0 + a0 xa0 )) ∈ Wl , where λ ∈ R \ {0}. By normalizing the e0 coordinate to −1, this means that TV (x) = −

a20 x + 2a0 a0 + a0 xa0 (a0 x + a0 )(a0 x + a0 ) =− 2 0 0 2 a0 + 2a0 ha , xi + |a | |a0 x + a0 |2

= −(a0 x + a0 )(a0 x + a0 )−1 . Using matrix representation as in Exercise 4.5.13, through stereographic projection, in V this corresponds to the map 0 e∞ 1 a a0 e∞ 1 1 −e∞ −a0 −a0 1 −e∞ 0 0 e ae −a e∞ a0 + a0 e∞ − 2a0 −2a 2(a∞ − a0 ) = 0 ∞ ∞ 0 = . a e∞ + e∞ a + 2a0 a0 − e∞ a0 e∞ 2(a∞ + a0 ) 2a This proves that the map in V is x 7→ (−ax+a− )(a+ x+a)−1 , which is a fractional linear map by Theorem 4.5.16. We next consider some applications to special relativity theory. Let {e0 , e1 , . . . , en , e∞ } be an ON-basis for an inertial observer O. This means that the e0 coordinate is the time of an event that O measures and that the e∞ coordinate measures how far in the zenith direction in space the event lies as measured by O, and similarly for the other space coordinates. Assume that O0 is another inertial observer passing by O at the origin event at relativistic speed, having an ON-basis {e00 , e01 , . . . , e0n , e0∞ } relative to which he measures time and space. Denote by T the isometry that maps T (ei ) = e0i , andPlet A = (αi,j ) denote the matrix of T relative to the basis {ei }, that is, T (ei ) = j ej αj,i . Consider an event with coordinates X = {xi } in the basis {ei } as observed by O. The same event observed by O0 has coordinates Y = {yi } in the basis {e0i }, where Y = A−1 X. If O and O0 compare their observations by identifying ei = e0i , then passing from O’s observation to O0 ’s observation defines the experienced map X 7→ A−1 X,

146

Chapter 4. Rotations and M¨obius Maps

in the basis {ei }, that is, T −1 . In particular, the map taking O’s observation of the celestial sphere to O0 ’s observation of the celestial sphere is TV−1 . Compare this result to how one experiences a usual rotation T in threedimensional Euclidean space: if one does not realize that one has rotated by T , it looks as if space has been rotated by T −1 . Sometimes these two points of view are referred to as active and passive transformations. The above result is the analogue for the hyperbolic rotations of spacetime. Example 4.6.5 (Watching a Lorentz boost). Consider four dimensional spacetime in relativity, and an inertial observer O with ON-basis {e0 , e1 , e2 , e∞ }. Consider another observer O’, viewed by O as travelling with speed tanh(φ) > 0 toward the north pole e∞ . As in Example 4.4.1, the Lorentz boost taking O:s ON-basis to O’:s ON-basis is T v = exp(φe0∞ /2)v exp(−φe0∞ /2) −1 = (cosh(φ/2)e∞ + sinh(φ/2)e0 )e∞ ve−1 . ∞ (cosh(φ/2)e∞ + sinh(φ/2)e0 )

Computing the induced fractional linear map on V = span{e1 , e2 } by two applications of Proposition 4.6.4, we get x 7→ x−1 7→ (cosh(φ/2) − sinh(φ/2))/(cosh(φ/2) + sinh(φ/2))x = exp(−φ)x. The experienced M¨obius map of the celestial sphere V, going from O’s to O0 ’s observation, is the inverse of this map corresponding to dilation x 7→ exp(φ)x on V . This means that the faster an observer travels in the direction e∞ , the more he will see the stars move in this direction! See Figure 4.6. Note that the two fixed points of the celestial sphere, ±e∞ , correspond to the two eigenvectors of the Lorentz boost on the light cone. Example 4.6.6 (Relativistic sphere paradox). Consider a Lorentz boost from O to O0 as in Example 4.6.5, and consider a spherical object at rest relative to O0 described by the equation (x0∞ )2 + (x01 − a1 )2 + (x02 − a2 )2 = 1. As the observer O0 passes O, O will see the object passing by at the same speed v as O0 in the direction e∞ . However, according to O’s measurements, it will not be a spherical object but rather an ellipsoid. Indeed, with the Lorentz boost in Example 4.4.1 in the {e0 , e∞ } plane, O will at time t describe the object by the equation p ((x∞ − vt)/ 1 − v 2 )2 + (x1 − a1 )2 + (x2 + a2 )2 = 1, which is an ellipsoid that is shorter in the e∞ direction. However, an amazing phenomenon occurs due to the finite speed of light. Even though O measures the object to be an ellipsoid after taking into account the finite propagation speed of light, the image he sees of the object is a circular shape, just like O0 , although any pattern on the surface of the sphere would be distorted. This is clear from Proposition 4.6.3, since the circular shape O0 sees of the object is mapped by

4.6. Mappings of the Celestial Sphere

147

Figure 4.6: A selection of stars and constellations on the northern celestial sphere, with the Polar Star at (0, 0) and scale 1 corresponding to angle π/4. (a) Our view from Earth showing Cygnus, Cepheus, Cassiopeia, Perseus, Ursa Minor, Auriga, and Ursa Major. (b) The same view for an inertial observer passing Earth at 2/3 the speed of light towards the Polar Star, showing also Aquila, Pegasus, Andromeda, Pisces, Taurus, Orion, Gemini, Leo, Virgo, Bo¨otes, and Hercules. Note that since the constellations have changed by a conformal map, there is little distortion between the views. Note also that at speeds close enough to that of light, even the Southern Cross will move up from behind and be visible next to the Polar Star. a M¨ obius map to another circular shape of the object on O’s celestial sphere! This phenomenon occurs only for spherical objects. Differently shaped objects will become distorted by the M¨obius map. Example 4.6.7 (Watching a parabolic rotation). Consider four-dimensional spacetime in relativity with ON-basis {e0 , e1 , e2 , e∞ } and the parabolic rotation T v = exp(φ(e0 − e∞ )a/2)v exp(−φ(e0 − e∞ )a/2), where a = a1 e1 + a2 e2 is a unit vector in V . The rotor here is q = exp(φ(e0 − e∞ )a/2) = (a+ φ2 (e0 −e∞ ))a. Proposition 4.6.4 applied twice shows that T induces a map of the celestial sphere that corresponds to the fractional linear map x 7→ −axa−1 7→ (−a(−axa−1 ) − φ)a−1 = x − φa of V . Thus a parabolic spacetime rotation translates the celestial sphere, fixing only the north pole e∞ , which corresponds to the eigenvector of the rotation on the light cone. We next turn to a more detailed study of the fractional linear maps appearing in Proposition 4.6.4. Collecting the four coefficients in a matrix, we have a map of

148

Chapter 4. Rotations and M¨obius Maps

spacetime vectors a: C : a + a∞ e∞ + a0 e0 7→

−a a∞ + a0

a∞ − a0 . a

(4.5)

In fact, we already used such a map in the proof of Lemma 3.4.14, so the following should not come as a surprise. Proposition 4.6.8 (Vahlen spacetime algebra). Let W be a spacetime, and fix an orthogonal splitting W = V ⊕ [e∞ ] ⊕ [e0 ] as above. Let 4V (2) denote the algebra of 2 × 2 matrices with coefficients in 4V . Then (4V (2), C), where C is defined in (4.5), is a Clifford algebra for W . Thus there is a unique algebra isomorphism 4W → 4V (2) that identifies a + a∞ e∞ + a0 e0 ∈ W ⊂ 4W and C(a + a∞ e∞ + a0 e0 ) ∈ 4V (2). Proof. The basis vectors map to 0 1 0 −1 C(e∞ ) = , C(e0 ) = , 1 0 1 0

−ej C(ej ) = 0

0 , ej

j = 1, . . . , n.

The Clifford condition (C) is straightforward to verify. To verify (U), which is needed only when n ≡ 3 (mod 4), it suffices by Proposition 3.3.3 to show that the volume element is not scalar. We have (−1)n+1 en 0 C(e0 )C(e∞ )C(e1 ) · · · C(en ) = , 0 en which is not scalar.

To understand the connection between spacetime isometries and fractional linear operators, we need to identify the Clifford cone in 4V (2). We use the following. a b Lemma 4.6.9. Let M = ∈ 4V (2). Then the involution and reversion in c d the Clifford algebra 4V (2) = 4W are # \ " "b b# a b b a −bb a b = , = d b . c d c d bc b −b c db a Proof. Recall that the involution and reversion act by (−1)k and (−1)k(k−1)/2 on 4k W . The corresponding subspace in 4V (2) is spanned by eb 0 0 et 0 −et0 −b et00 0 es = s , e∞ et = , e0 et0 = , e0 e∞ et00 = , 0 es ebt 0 ebt0 0 0 et00 where |s| = k, |t| = |t0 | = k − 1, and |t00 | = k − 2, and the result follows by inspection.

4.6. Mappings of the Celestial Sphere

149

b (2) is isomorProposition 4.6.10 (Vahlen = Clifford cone). The Vahlen cone 4V b phic to the Clifford cone 4W under the isomorphism of Clifford algebras 4V (2) = b (2) can be writ4W determined by (4.5). In particular, every Vahlen matrix in 4V ten as a product of at most dim V + 2 matrices of the form (4.5), with a ∈ V , a0 , a∞ ∈ R, and a02 6= a2∞ + |a|2 . b b (2), note that spacetime vectors belong to 4V b (2) ⊂ 4V Proof. To show that 4W b (2) is closed under multiplication by Lemma 4.5.15. Next consider and that 4V a b b b the converse inclusion 4V (2) ⊂ 4W . Let M = be a Vahlen matrix, and c d thus invertible. By Proposition 4.1.5 it suffices to prove, for all v ∈ V , α, β ∈ R, that # " b a −bb −v α d −b β v −c a c db −b is a spacetime vector, that is, the off-diagonal entries are scalars and the diagonals are ± a vector. By linearity, it suffices to check the cases in which onlyone of v, α, −b ac b aa and β is nonzero. If v = β = 0, α = 1, then the product is , which is a b cc −b ca spacetime vector since b ac = −b ac ∈ V . The case v = α = 0, β = 1 abc = −b ca, since b is similar. For α = β = 0, we get " # −b avd + bbvc b avb − bbva . b b b cvd − dvc −b cvb + dva

To show that the lower left entry is a scalar, we may assume that d 6= 0. In this case, the question is whether b db = (db b − (dd)v(c b b = (dd)((db b b = (dd)((db b d(b cvd − dvc) c)v(dd) d) c)v − v(cd)) c)v + v(db c)) b c, since is scalar. This is clear, since db c ∈ V and ddb ∈ R. Note that cdb = −cdb = −db cdb ∈ V . To show that the upper left entry is a vector, we may assume that b 6= 0, for otherwise, a and d are parallel, and the result follows. We need to show that b(−b avd + bbvc)bb = −(bb a)v(dbb) + (bbb)v(cbb) = (v(bb a) − 2hv, bb ai)(dbb) + v(bbb)cbb = v(bb a)(dbb) + v(bbb)cbb − 2hv, bb ai(dbb) = −vbb(ad − bc)bb − 2hv, bb ai(dbb) is a vector, which is clear. The right entries are shown to be scalars and vectors similarly. Finally, d c b −b avd + bbvc = b avd − bbvc = b cvb − dva. b (2) = 4W b , from which the factorization result for This completes the proof of 4V Vahlen matrices follows by Proposition 4.1.5.

150

Chapter 4. Rotations and M¨obius Maps

We end by summarizing the relation between spacetime isometries and M¨obius maps. Let W be an n-dimensional spacetime, and let V ↔ V ⊂ V∞ ⊂ W represent the celestial sphere as above. Then we have group homomorphisms b q ∈ 4W ρ0 y

pW

−−−−→ O(W ) 3 T ρ1 y

b (2) −−p−V−→ M¨ob(V ) 3 f M ∈ 4V such that ρ1 ◦ pW = pV ◦ ρ0 . • The map ρ0 from Proposition 4.6.10 is an isomorphism. • The map pW as in Proposition 4.1.9 is surjective and has kernel 40 W \ {0}. • The map pV from Theorem 4.5.16 is surjective and has kernel ρ0 (40 W \ {0}) ∪ ρ0 (4n W \ {0}). • The map ρ1 from Proposition 4.6.3 is surjective and has kernel ±1. That ρ1 ◦pW = pV ◦ρ0 is straightforward to verify. We have seen the mapping properties for ρ0 , pW , and pV , and those for ρ1 follow from this. Note that if we b b (2) correspondingly, then pW normalize 4W to Pin(W ), and also normalize 4V and ρ1 are 2 − 1 maps, pV is a 4 − 1 map, and ρ0 is a 1 − 1 map.

4.7

Comments and References

4.1 Our proof of the Cartan–Dieudonn´e theorem (Theorem 4.1.3) is from Grove [49], with a minor simplification using a continuity argument. b is usually referred to as the Clifford What we here call the Clifford cone 4V group in the literature. Our terminology has been chosen to go together with b V . The orthogonal and special orthogonal the notion of the Grassmann cone ∧ groups and their abbreviations are standard, and the name of the spin group is due to its connection to physics. J.-P. Serre introduced the name Pin by removing the S in Spin, in analogy with the orthogonal group. The terminology rotor for an object in the spin group is not used in the literature. 4.2 Example 4.2.6, showing that the exponential map is not surjective for an ultrahyperbolic inner product space, is taken from [78]. 4.4 The source of inspiration for the treatment of rotations in space and spacetime with Clifford algebra is M. Riesz [78], where most of the results are found. 4.5 Our treatment of M¨obius maps with Clifford algebra follows L. V. Ahlfors [1]. Theorem 4.5.12 is from Hertrich–Jeromin [54].

4.7. Comments and References

151

4.6 I want to thank Malcolm Ludvigsen, whom I had as a teacher in a course in cosmology at Link¨oping University as an undergraduate student and who shared with us students many fascinating insights into relativity theory, including the sphere paradox described in Example 4.6.6.

Chapter 5

Spinors in Inner Product Spaces Prerequisites: This chapter builds on Chapters 3 and 4, and uses the material in Sections 1.4 and 1.5. Any knowledge of representation theory is helpful, but the presentation is self-contained and should be accessible to anyone with a solid background in linear algebra. Road map: In a certain sense, one can form a square root p 4V / = 4V of the Clifford algebra of a given inner product space V . Indeed, we have seen in Section 3.4 that 4V is isomorphic to a matrix algebra. For example, the Clifford algebra of spacetime W with three space dimensions is isomorphic to L(R4 ) = R4 ⊗ R4 , √ so in this sense, R4 = 4W . Such spaces are referred to as spinor spaces, and have they deep applications in both physics and mathematics. Two problems with this construction need to be addressed, though. The first is that depending on the dimension and signature of the inner product space, by Theorem 3.4.13 the coefficients in the matrices may belong to R, R2 , C, H, or H2 . However, the standard construction of spinor spaces is over the complex field, and indeed, complex Clifford algebras are always isomorphic to matrix algebras over C or C2 depending on the parity of the dimension, which simplifies matters. The second and more fundamental problem is that isomorphisms 4Vc ↔ L(S) are a priori not unique. To show that a spinor space S has an invariant geometric meaning, we need to show that different choices of matrices used in setting up the isomorphism amount to only a renaming of the elements in S. This © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_5

153

Chapter 5. Spinors in Inner Product Spaces

154

is the purpose of Section 5.2, which, through the principle of abstract algebra, completes the construction of the complex spinor space S = 4V / of an inner product space V . In Section 5.3 we show how to map spinors between different spaces, something that becomes important, for example, in considering spinors on manifolds. / is a fundamental construction for the celebrated Atiyah– The spinor space 4V Singer index theorem for Dirac operators, which we look at in Chapter 12. In physics, spinors are famous for describing a certain intrinsic state for some elementary particles such as electrons in quantum mechanics, which we look at briefly in Section 9.2. Spinor spaces generalize to higher dimensions, a topological feature of the complex square root √ z 7→ z. As we know, the √ complex square root has two possible values differing by sign. If we √ start at 1 = 1 √ and move continuously around the unit circle, then we have e2πi = −1 and e4πi = 1. This means that it takes two full rotations for the square root to return to its original value. In higher dimensions, this is the characteristic behavior of spinor spaces: two full rotations of the physical, or vector, space are needed to return the spinors to their original positions. Beyond the standard linear representation 4V / , there are countably infinitely many nonisomorphic spinor spaces possible for a given inner product space. This is the topic of Section 5.4. Highlights: • The standard representation of spinors: 5.1.5 • The main invariance theorem for spinors: 5.2.3 • Induced spinor maps: 5.3.5 • Finding all abstract spinor spaces in three and four dimensions: 5.4.2, 5.4.8

5.1

Complex Representations

We start with the following general idea of a spinor space. Definition 5.1.1 (Group representation). Let V be a real inner product space, and consider the spin group Spin(V ) ⊂ 4ev V . A complex Spin(V ) representation is a smooth group homomorphism ρ : Spin(V ) → L(S) into the space of linear operators on a complex linear space S. Consider −1 ∈ Spin(V ). If ρ(−1) = −I, then S is called an abstract spinor space for V . If ρ(−1) = I, then S is called an abstract tensor space for V .

5.1. Complex Representations

155

As spaces, abstract tensor and spinor spaces are nothing more than linear spaces in general. The point is that there is a coupled action of rotations in V and its various abstract tensor and spinor spaces. To explain the topological idea, let T (t) ∈ SO(V ), t ∈ [0, 1], be a loop of rotations of V . Lift T (t) to a path q(t) in Spin(V ) such that p(q(t)) = T (t) with the covering map p from Proposition 4.1.9. Assuming that T (0) = T (1) = I and that q(0) = 1, it may happen that q(1) = 1 or that q(1) = −1. See Theorem 4.3.4. Assume that T (t) is such that q(1) = −1, that is, that T (t) is a loop of rotations that is not homotopic to the constant loop. If S is an abstract tensor space for V , then ρ(−q) = ρ(q). This means that ρ is just a representation of SO(V ), and as we rotate V by T (t), the associated linear operators ρ(q(t)) rotate the abstract tensor space S a full loop. If S is instead an abstract spinor space for V , then ρ(−q) = −ρ(q). This means that as we rotate V by T (t), the associated linear operators ρ(q(t)) rotate the abstract tensor space S. But when we have completed one full turn of V at T (1) = I, the space S is not in its initial position, since ρ(−1) = −I. See Figure 5.1. On performing the same rotation of V by T (t) one more time, however, the associated linear operators ρ(q(t)) will return the abstract spinor space S to its original position.

Figure 5.1: The coupled rotation of vector space V (left) and a spinor space S (right), where one full rotation of V corresponds to half a rotation of S. In this concrete example V = R3 and S = 4R / 3 ↔ C2 , as in Definition 5.2.4, where 2 2 only the real subspace R ⊂ C is shown to the right, and the spinor rotation is furnished by the two real Pauli matrices as in Examples 5.1.5 and 5.1.7.

156

Chapter 5. Spinors in Inner Product Spaces

Example 5.1.2. We list some abstract tensor and spinor spaces for a given Euclidean space V that we encounter elsewhere in this book. (i) The exterior powers ∧k V of V are (real) abstract tensor spaces for V , with the action of rotations T being the induced rotations ρ(T )w = T (w) as in Definition 2.3.1. In terms of the rotor q representing T , we have ρ(q)w = qwq −1 as in Proposition 4.1.10. (ii) Similarly, the tensor product V ⊗ V from Definition 1.4.2 is a (real) abstract tensor space for V , with the action of a rotation T being the linear map induced by the bilinear map (v1 , v2 ) 7→ T (v1 ) ⊗ T (vk ) with the universal property of tensor algebras. (iii) The Clifford algebra 4V , as well as the even subalgebra 4ev V , is an abstract spinor space for V , with the action of a rotor q being ρ(q)w = qw. (iv) The spaces Pksh of scalar-valued k-homogeneous harmonic polynomials from Definition 8.2.1 are (real) abstract tensor spaces for V with the action of a rotation T being P (x) 7→ P (T −1 (x)). Note that the rotation invariance of the Laplace equation shows that the rotated polynomial P ◦ T −1 will be harmonic. (v) The spaces Pkem of k-homogeneous monogenic polynomials with values in the even subalgebra from Definition 8.2.1 are (real) abstract spinor spaces for V with the action of a rotor q being P (x) 7→ qP (q −1 xq). Note that Proposition 8.1.14 shows that the so obtained polynomial will be monogenic, and that −1 acts by −I. There are infinitely many nonisomorphic abstract tensor and spinor spaces for a given inner product space V , as we demonstrate in Section 5.4 in low dimension. We will mainly be concerned with one particular example of a spinor space, which builds on the following related notion. Definition 5.1.3 (Vector representation). Let V be a real inner product space. A complex V representation is a real linear map ρ : V → L(S) into the space of complex linear operators on a complex linear space S such that ρ(v)2 = hvi2 I,

v ∈ V.

We note that such V representations are nothing but algebra homomorphisms.

5.1. Complex Representations

157

Lemma 5.1.4 (Vector=algebra representation). Let V be a real inner product space, and consider its real Clifford algebra 4V with complexification 4Vc . Then every complex V representation ρ : V → L(S) extends in a unique way to a homomorphism 4Vc → L(S) of complex algebras. Conversely, every such homomorphism restricts to a complex V representation. Proof. Clearly, for any algebra homomorphism we have ρ(v)2 = ρ(v 2 ) = ρ(hvi2 1) = hvi2 ρ(1) = hvi2 I,

v ∈ V.

Given a complex V representation, it follows from the universal property of Clifford algebras as in Section 3.3 that it extends uniquely to a real algebra homomorphism 4V → L(S). As in Section 1.5, this complexifies in a unique way to a complex algebra homomorphism. A complex V representation is a special case of a complex Spin(V ) representation. Indeed, if we restrict an algebra homomorphism 4Vc → L(S) to the embedded Lie group Spin(V ) ⊂ 4ev V ⊂ 4Vc , we obtain a smooth group homomorphism. Moreover, since ρ(−1) = −ρ(1) = −I for every algebra homomorphism, S will be an abstract spinor space for V . We also note that in general, a complex Spin(V ) representation does not arise in this way from a complex V representation, not even for abstract spinor spaces. Indeed, in general, a complex Spin(V ) representation is not the restriction to Spin(V ) of a linear map on 4ev V . In Example 5.1.2(v) above, this happens only for k = 0. The spinor spaces that we mainly will use are the following. Example 5.1.5 (The standard representation). We set out to construct a complex V representation, for a given Euclidean space V , which will be the most important example for us of a spinor space. The basic idea of this construction is Theorem 3.4.2. (i) Consider first the case that dim V = n = 2m is even. Fix a complex structure J on V , as in Section 1.5, which is isometric. This turns V into a complex linear space V = (V, J). In fact, this is a Hermitian inner product space with complex inner product (·, ·i such that Re(·, ·i is the original Euclidean inner product. Define the complex exterior algebra S := ∧V, which has complex dimension 2m . Generalizing the real theory from Chapter 2 to complex linear spaces, we obtain a complex bilinear exterior product w1 ∧ w2

158

Chapter 5. Spinors in Inner Product Spaces

and a dual complex sesquilinear product w1 y w2 on ∧V. Define a real linear map ρ : V → L(S) by ρ(v)w := v y w + v ∧ w, w ∈ S. Since v y (v y w + v ∧ w) + v ∧ (v y w + v ∧ w) = v y (v ∧ w) + v ∧ (v y w) = (v, viw, by the complex analogue of Theorem 2.8.1, where (v, vi = hv, vi = hvi2 , this ρ is a complex V representation. To make the construction above more concrete, choose a complex ON-basis m for V. This means that this together with the vectors e−k := Jek , k = {ek }k=1 1, . . . , m, forms a real ON-basis for V . Then ρ(ek )ψ = ek y ψ + ek ∧ ψ, ρ(e−k )ψ = i(−ek y ψ + ek ∧ ψ),

ψ ∈ S.

(ii) Consider now the case that dim V = n = 2m + 1 is odd. In this case, we fix a unit vector e0 and consider V 0 := [e0 ]⊥ . Proceeding as in (i) with the evendimensional space V 0 , fixing an isometric complex structure J on V 0 , we obtain a complex V 0 representation on S = ∧V 0 . We extend this to a real linear map of V by defining b ψ ∈ S, ρ(e0 )ψ := ψ, where ψb denotes the complex analogue of the involution from Definition 2.1.18. Note that we define this as a complex linear map. This yields a complex V representation. Indeed, ρ(e0 ) anticommutes with all ρ(v), v ∈ V 0 , since these latter operators swap ∧ev V 0 and ∧od V 0 . Proposition 5.1.6 (Minimal vector representations). Let V be a real inner product space of dimension n = 2m or n = 2m + 1. Then dimC S ≥ 2m for every complex V representation ρ : V → L(S). There exists such a complex V representation with dimC S = 2m . If n is even, such a minimal representation is an isomorphism of complex algebras 4Vc → L(S). If n is odd, then 4ev Vc → L(S) is an algebra isomorphism. Proof. Consider a complex V representation ρ, and let {ek } be an ON-basis for V . As in the proof of Proposition 3.3.3, it follows that {ρ(es )}s⊂n are linearly independent operators when n is even, and that {ρ(es )}|s| is even are linearly independent operators when n is odd. This shows that dimC L(S) ≥ 22m and therefore that dimC S ≥ 2m . It also proves the statements about algebra isomorphisms. The existence of minimal V representations follows from Example 5.1.5 when V is a Euclidean space. For a non-Euclidean inner product space (V, h·, ·i) we write V = V+ ⊕ V− , where V+ and V− are orthogonal subspaces on which the inner

5.1. Complex Representations

159

product is positive and negative definite respectively. Write Ve for V made into a Euclidean space by changing the sign on the inner product on V− , and let ρ be a standard Ve representation as in Example 5.1.5. Then it is straightforward to verify that V = V+ ⊕ V− → L(S) : v+ + v− 7→ ρ(v+ ) + iρ(v− ) is a complex V representation.

We end this section by showing what these matrix representations look like in low-dimensional spaces. Example 5.1.7 (2D and 3D matrices). (i) Let V be a two-dimensional Euclidean space. Fix an ON-basis {e−1 , e1 } and a complex structure J such that Je1 = e−1 . We take e1 as an ON-basis for the one-dimensional Hermitian space V, and for S = ∧V we fix the ON-basis {1, e1 }. In this basis, the basis multivectors act as follows: 0 1 0 −i , ρ(e1 ) = , ρ(e−1 ) = 1 0 i 0 and thus ρ(1) =

1 0 , 0 1

i 0 . 0 −i

ρ(e1 e−1 ) =

(ii) Next add a basis vector e0 and consider three-dimensional Euclidean space with ON-basis {e−1 , e0 , e1 }. With the same spinor space and basis as in two dimensions, we have the action 1 0 . ρ(e0 ) = 0 −1 Note that {ρ(e1 ), ρ(e −1 ),ρ(e0 )} are the Pauli matrices from Example 3.4.19, and i 0 . that ρ(e1 e−1 e0 ) = 0 i (iii) Let W be a spacetime with one space dimension. Fix an ON-basis {e0 , e1 } and write {ie0 , e1 } for the associated Euclidean ON-basis inside Wc . Identifying {ie0 , e1 } and {e−1 , e1 }, we obtain from (i) the representation 0 −1 0 1 , , ρ(e1 ) = ρ(e0 ) = 1 0 1 0 of W . This coincides with the representation we obtained in Example 3.4.18. (iv) Consider next a spacetime with two space dimensions. Adding a Euclidean ON-basis vector e2 , with 1 0 ρ(e2 ) = , 0 −1 we obtain a representation of this spacetime.

160

Chapter 5. Spinors in Inner Product Spaces

Example 5.1.8 (4D and 5D matrices). (i) Let V be a four-dimensional Euclidean space. Fix an ON-basis {e−2 , e−1 , e1 , e2 } and a complex structure J such that Je1 = e−1 and Je2 = e−2 . We take {e1 , e2 } as ON-basis for the two-dimensional Hermitian space V, and for S = ∧V we fix the ON-basis {1, e12 , e1 , e2 }. In this basis, the basis vectors {e1 , e2 , e−1 , e−2 } act by matrices 0 0 0 1 0 0 0 −i 0 0 1 0 0 0 −i 0 0 i , 0 0 −i 0 0 0 0 1, 0 0 −1 0, 0 0 1 0 0 0 0 −1 0 0 i 0 0 i 0 0 0 0 0 1 0 0 0 −i 0 0 0 1 0 0 0 i 0 0 respectively. (ii) Next add a basis vector e0 and consider five-dimensional Euclidean space with ON-basis {e−2 , e−1 , e0 , e1 , e2 }. With the basis for S from (i), we have the action 0 1 0 0 0 1 0 0 ρ(e0 ) = 0 0 −1 0 . 0 0 0 −1 Example 5.1.9 (Dirac’s γ-matrices). Let W be a spacetime with three space dimensions. Fix an ON-basis {e0 , e1 , e2 , e3 } and write {ie0 , e1 , e2 , e3 } for the associated Euclidean ON-basis inside Wc . Identifying {ie0 , e1 , e2 , e3 } and {e1 , e−2 , −e2 , e−1 }, we obtain from Example 5.1.8(i) the representation 0 −i 0 0 0 0 0 −i 0 0 −i , ρ(e1 ) = 0 0 −i 0 , 0 ρ(e0 ) = 0 i 0 0 0 0 −i 0 0 −i 0 0 i 0 0 0 0 0 −i 0 0 0 0 −1 0 0 0 0 1 0 0 i ρ(e2 ) = 0 1 0 0 , ρ(e3 ) = i 0 0 0 0 −i 0 0 −1 0 0 0 for W . The relation to standard representations used in quantum mechanics is the f as an inner product space of signature following. Consider instead spacetime W 1 − 3. Here the Dirac gamma matrices 0 σ3 I 0 σ1 σ2 0 0 3 0 1 2 , , γ = , γ = γ = , γ = 0 −I −σ2 0 −σ3 0 −σ1 0 f , where {σ1 , σ2 , σ3 } are the Pauli matrices from Exrepresent an ON-basis for W ample 3.4.19. Another important matrix is the so-called fifth gamma matrix 0 I 5 0 1 2 3 . γ := iγ γ γ γ = I 0

5.2. The Complex Spinor Space

161

f. In quantum mechanics this represents chirality, since γ 5 ∈ 44 W However, matrix representations of the ON-basis vectors of W are quite arf, bitrary. Another representation is the Weyl representation of an ON-basis for W namely {γ 5 , γ 1 , γ 2 , γ 3 }. f representation, since (γ 5 )2 = I and γ 5 anticomThis gives another complex W 2 3 1 mutes with γ , γ , γ . Comparing to our basis {e0 , e1 , e2 , e3 } for spacetime with signature 3 − 1, we see that e0 = −iγ 5 ,

e1 = −iγ 1 ,

e2 = −iγ 2 ,

e3 = −iγ 3 .

5.2 The Complex Spinor Space To formulate the main theorem for spinor spaces, we need the following terminology. Definition 5.2.1 (Main reflectors). Let V be a real inner product space of dimension n, and consider its complexified Clifford algebra 4Vc . The main reflectors in 4Vc are the two n-vectors wn ∈ 4n Vc satisfying wn2 = 1. For a given choice of main reflector wn , define wn+ := 21 (1 + wn ) and wn− := 12 (1 − wn ). Note that if we fix 0 6= w0 ∈ 4n V , then any w ∈ 4n Vc can be written w = λw0 for some λ 6= 0. The equation 1 = λ2 w02 has exactly two solutions over the complex field, which yield the two main reflectors. Lemma 5.2.2. Fix a main reflector wn ∈ 4n Vc in a real inner product space V , and let ρ : V → L(S) be a complex V representation. Then ρ(wn+ ) and ρ(wn− ) are complementary projections in the sense that ρ(wn± )2 = ρ(wn± ) and ρ(wn+ )+ρ(wn− ) = I. If dim V is even, then the dimensions of the ranges of the two projections ρ(wn± ) are equal. If dim V = 2m + 1 is odd and if dimC S = 2m , then either ρ(wn+ ) = 0 or ρ(wn− ) = 0. Equivalently, ρ(wn ) = I or ρ(wn ) = −I. Proof. That ρ(wn± ) are complementary projections is a consequence of ρ being an algebra homomorphism. In even dimension, we fix a unit vector v ∈ V and note that ρ(v)2 = I and therefore ρ(v) is self-inverse. Since vw = −wv, we deduce that ρ(wn− ) = ρ(v)ρ(wn+ )ρ(v)−1 , from which it follows in particular that ρ(wn± ) projects onto subspaces of S of equal dimension. In odd dimension, it follows as in the proof of Proposition 3.3.3 for a minimal complex V representation that ρ(wn ) and ρ(1) = I must be linearly dependent. Since ρ(wn )2 = I, we have in fact ρ(wn ) = ±I.

162

Chapter 5. Spinors in Inner Product Spaces

The main result needed for the geometric construction of spinors is the following uniqueness result. This is similar to the universal property for the exterior algebra from Proposition 2.1.4, and the universal property for Clifford algebras from Definition 3.3.1, in that it provides a useful way to construct mappings of spinors. See Section 5.3. Theorem 5.2.3 (Uniqueness of minimal representations). Let V be a real inner product space of dimension n = 2m or n = 2m + 1. Assume that ρj : V → L(Sj ) are two complex V representations, both with dimension dimC Sj = 2m , j = 1, 2. (i) If n = 2m, then there exists an invertible linear map T : S1 → S2 such that w ∈ 4Vc .

ρ2 (w)T = T ρ1 (w),

If Te is a second such map, then Te = λT for some λ ∈ C \ {0}. (ii) If n = 2m + 1, then there exists an invertible linear map T : S1 → S2 such that either ρ2 (w)T = T ρ1 (w), ρ2 (w)T = T ρ1 (w), b

w ∈ 4Vc , w ∈ 4Vc .

or

The first case occurs when ρ1 (wn ) = ρ2 (wn ) for the main reflectors wn , and the second case occurs when ρ1 (wn ) = −ρ2 (wn ). In particular, ρ2 (w)T = T ρ1 (w) for all w ∈ 4ev Vc . If T ∈ L(S1 ; S2 ) is a second map with this property, then Te = λT for some λ ∈ C \ {0}. This section is devoted to the proof of Theorem 5.2.3, but before embarking on this, we make use of this result to give a proper invariant geometric definition of spinors and the complex spinor space of V . To this end, recall the principle of abstract algebra, as discussed in the introduction of Chapter 1. At this stage the reader hopefully is so comfortable using multivectors that he or she has forgotten that we never actually defined a k-vector to be a specific object, but rather as an element in the range space of some arbitrarily chosen multilinear map with properties (A) and (U) in Definition 2.1.6. Proposition 5.1.6 and Theorem 5.2.3 now allow us to define spinors in a similar spirit, as elements in some arbitrarily chosen representation space S of minimal dimension 2m . Indeed, Theorem 5.2.3 shows that any other choice of ρ and S amounts only to a renaming of the objects in the space S, and does not affect how 4Vc acts on S. Definition 5.2.4 (The complex spinor space). Let (X, V ) be a real inner product space of dimension n = 2m or n = 2m + 1. Fix one complex V representation / , and refer to (4V, / ρ) ρ : V → L(S) with dimC S = 2m . We denote this S by 4V as the complex spinor space for V . We also use the shorthand notation w.ψ := ρ(w)ψ

163

5.2. The Complex Spinor Space

for the action of w ∈ 4Vc on ψ ∈ 4V / . + Fix a main reflector wn ∈ 4n Vc . When dim V is even, we denote by 4 / V − and 4 / V the ranges of ρ(wn± ), so that +

−

/ =4 / V ⊕4 / V. 4V When dim V is odd, we write ρ− (w)ψ := w.ψ ˆ for the second nonisomorphic action / . of 4Vc on 4V

Figure 5.2: Multivector action on spinors in a four-dimensional inner product space. We have constructed the complex spinor space 4V / for V in a way very similar in spirit to the construction of the exterior algebra ∧V for V . There is one big difference between the constructions of multivectors and spinors, though: with the map ∧k , a set of k vectors single out a certain k-vector, but vectors are not factors in spinors in this way. Rather vectors, and more generally multivectors, act as linear operators on spinors. +

−

Exercise 5.2.5. Assume that dim V is even. Show that ρ(v) maps 4 / V →4 / V − + and 4 / V , for every vector v ∈ 41 V = V . Using the standard represen/ V →4 + − tation from Example 5.1.5(i), show that 4 / V and 4 / V coincide with ∧ev V and od ∧ V.

164

Chapter 5. Spinors in Inner Product Spaces

We now turn to the proof of Theorem 5.2.3. The idea is roughly that two minimal complex V representations yield an isomorphism L(S1 ) → L(S2 ) of complex algebras. Given an invertible map T ∈ L(S1 ; S2 ), L(S1 ) → L(S2 ) : X → T XT −1 is such an algebra isomorphism. We prove below that every isomorphism L(S1 ) → L(S2 ) arises in this way, and we deduce from this fact Theorem 5.2.3. For this proof, it does not matter whether we work over the real or complex field. Definition 5.2.6 (Ideals). Let A be a complex associative algebra, as in Definition 1.1.4 but replacing R by C. A linear subspace I ⊂ A is called a left ideal if xy ∈ I whenever x ∈ A and y ∈ I. If yx ∈ I whenever x ∈ A and y ∈ I, then I is called a right ideal. A linear subspace I that is both a left and right ideal is called a two-sided ideal. The notion of ideal is important in identifying the T representing a given algebra isomorphism as above. We also recall from Proposition 1.4.3 that there is a natural isomorphism S ⊗ S ∗ ↔ L(S), which identifies the simple tensor v ⊗ θ ∈ S ⊗ S ∗ and the rank-one linear operator x 7→ hθ, xiv. We use both these views on linear operators below. Proposition 5.2.7 (Matrix ideals). Consider the algebra L(S) of all linear operators on a linear space S. There is a one-to-one correspondence between linear subspaces of S ∗ and left ideals in L(S) that identifies a subspace U ⊂ S ∗ and the left ideal o nX vk ⊗ θk ; vk ∈ S, θk ∈ U . IUl := {T ∈ L(S) ; R(T ∗ ) ⊂ U } = k

Similarly there is a one-to-one correspondence between linear subspaces of S and right ideals in L(S) that identifies a subspace U ⊂ S and the right ideal nX o IUr := {T ∈ L(S) ; R(T ) ⊂ U } = vk ⊗ θk ; vk ∈ U, θk ∈ S ∗ . k

The only two-sided ideals in L(S) are the two trivial ones, namely {0} and L(S) itself. Note that in terms of matrices, a left ideal consists of all matrices with all row vectors in a given subspace, whereas a right ideal consists of all matrices with all column vectors in a given subspace.

5.2. The Complex Spinor Space

165

Proof. Clearly IUl is a left ideal for any subspace U ⊂ S ∗ . Note that R(T ∗ ) ⊂ U is equivalent to U ⊥ ⊂ N(T ). To see that all left ideals are of the form IUl , consider any left ideal I ⊂ L(S). Let U ⊂ S ∗ be such that \ N(T ). U ⊥ := T ∈I

Then I ⊂ IUl . For the converse we claim that there exists T0 ∈ I such that N(T0 ) = U ⊥ . From this it follows that IUl ⊂ I. Indeed, if N(T0 ) ⊂ N(T ), then T = AT0 ∈ I for some operator A. One way to prove the claim is to observe that N(T1 ) ∩ N(T2 ) = N(T1∗ T1 + T2∗ T2 ), and that T1∗ T1 + T2∗ T2 ∈ I whenever T1 , T2 ∈ I. Here the adjoints are with respect to any auxiliary Euclidean inner product. Using this observation a finite number of times, it follows that there exists T0 ∈ I with minimal null space. The proof for right ideals is similar. Finally, if I is a two-sided ideal, then I = IUl = IVr for some subspaces U ⊂ S ∗ and V ⊂ S. This can happen only if U = {0} = V or if U = S ∗ and V = S. Thus I = {0} or I = L(S). The following is the key result in the proof of uniqueness of minimal spinor representations. Proposition 5.2.8 (Completeness of matrix algebras). The algebra L(S) of all linear operators on a linear space S is complete in the sense that every algebra automorphism φ : L(S) → L(S) is inner, that is, there exists a linear invertible map T ∈ L(S) such that φ(X) = T XT −1 for all X ∈ L(S). Such T are unique up to scalar multiples. Proof. By Proposition 5.2.7, every minimal left ideal, that is, a left ideal I such that no left ideal J such that {0} $ J $ I exists, must be of the form Lθ := {v ⊗ θ ; v ∈ S},

for some θ ∈ S ∗ \ {0}.

Similarly, all minimal right ideals are of the form Rv := {v ⊗ θ ; θ ∈ S ∗ },

for some v ∈ S \ {0}.

Let φ : L(S) → L(S) be an algebra automorphism, and fix θ0 ∈ S ∗ and v 0 ∈ S such that hθ0 , v 0 i = 1. Since φ only relabels the objects in L(S) without changing the algebraic structure, it is clear that φ(Lθ0 ) = Lθ00 for some θ00 ∈ S ∗ \ {0}. It follows that there is an invertible linear operator T1 ∈ L(S) such that φ(v ⊗ θ0 ) = T1 (v) ⊗ θ00 ,

for all v ∈ S.

(5.1)

Similarly, by considering the mapping of minimal right ideals, there exist an invertible linear operator T2 ∈ L(S ∗ ) and v 00 ∈ S such that φ(v 0 ⊗ θ) = v 00 ⊗ T2 (θ),

for all θ ∈ S ∗ .

166

Chapter 5. Spinors in Inner Product Spaces

Since φ is an automorphism, it follows that φ(v ⊗ θ) = φ((v ⊗ θ0 )(v 0 ⊗ θ)) = (T1 (v) ⊗ θ00 )(v 00 ⊗ T2 (θ)) = hθ00 , v 00 iT1 (v) ⊗ T2 (θ), for all v ∈ S, θ ∈ S ∗ . For example, by inversely rescaling T1 and θ00 , we may assume that hθ00 , v 00 i = 1. Furthermore, for all v1 , v2 ∈ S and θ1 , θ2 ∈ S ∗ we have hθ1 , v2 iφ(v1 ⊗ θ2 ) = φ((v1 ⊗ θ1 )(v2 ⊗ θ2 )) = (T1 (v1 ) ⊗ T2 (θ1 ))(T1 (v2 ) ⊗ T2 (θ2 )) = hT2 (θ1 ), T1 (v2 )iT1 (v1 ) ⊗ T2 (θ2 ), so (T2 )−1 = (T1 )∗ . Thus, with T = T1 , we get that φ(v ⊗ θ) = T (v) ⊗ (T ∗ )−1 (θ), and therefore φ(X) = T XT −1 by linearity. The uniqueness result is a consequence of the fact that Z(L(S)) = span{I}. Indeed, if TeX Te−1 = T XT −1 for all X, then T −1 Te ∈ Z(L(S)). Note that formula (5.1) can be used to calculate T for a given automorphism ∗ φ. Fixing a basis for S and dual00 basis for S , we work with matrices and may 0 assume θ = 1 0 . . . 0 and θ = a1 a2 . . . ak . Then we have for vectors v ∈ S the matrix identity φ v 0 . . . 0 = a1 T (v) a2 T (v) . . . ak T (v) . Since at least one aj is nonzero, we find that T (v) can be defined as a nonzero column of the matrix φ(v ⊗ θ0 ). Proof of Theorem 5.2.3. (i) Assume dim V = 2m. Consider two representations ρ1 and ρ2 with dim S1 = 2m = dim S2 , which by Proposition 5.1.6 are algebra isomorphisms 4Vc → L(Si ). Then ρ2 ρ−1 1 : L(S1 ) → L(S2 ) is an algebra isomorphism. Take any linear invertible map T0 : S1 → S2 and consider the induced −1 algebra isomorphism ρ0 : L(S1 ) → L(S2 ) : X 7→ T0 XT0−1 . Then ρ−1 is an 0 ρ2 ρ 1 automorphism of L(S1 ), and Proposition 5.2.8 shows the existence of T1 ∈ L(S1 ), unique up to scalar multiples, such that −1 −1 ρ−1 0 ρ2 ρ1 (X) = T1 XT1 ,

for all X ∈ L(S1 ).

Letting T := T0 T1 : S1 → S2 , this means that ρ2 (w) = T ρ1 (w)T −1 ,

for all w ∈ 4Vc .

(ii) Assume dim V = 2m + 1. Consider two representations ρ1 and ρ2 with dim S1 = 2m = dim S2 . Then ρi : 4ev V → L(S) are both isomorphisms, so as in (i), we get from the algebra isomorphism (ρ2 |4ev V )(ρ1 |4ev V )−1 : L(S1 ) → L(S2 ) the existence of T ∈ L(S1 ; S2 ) such that ρ2 (w) = T ρ1 (w)T −1 ,

for all w ∈ 4ev Vc ,

5.3. Mapping Spinors

167

unique up to multiples. Consider next an arbitrary multivector w ∈ 4Vc . This can be uniquely written as w = w1 + wn w2 , with w1 , w2 ∈ 4ev Vc , if we fix a main reflection wn ∈ 4n Vc . If ρ1 (wn ) = ρ2 (wn ), then ρ2 (w)T = (ρ2 (w1 ) + ρ2 (wn )ρ2 (w2 ))T = T (ρ1 (w1 ) + ρ1 (wn )ρ1 (w2 )) = T ρ1 (w), since w1 , w2 ∈ 4ev Vc and ρi (wn ) = ±I. If ρ1 (wn ) = −ρ2 (wn ), then ρ2 (w)T = (ρ2 (w1 ) + ρ2 (wn )ρ2 (w2 ))T = T (ρ1 (w1 ) − ρ1 (wn )ρ1 (w2 )) = T ρ1 (w). b

5.3

Mapping Spinors

Consider two vector spaces V1 and V2 and a linear map T : V1 → V2 . In Section 2.3, we saw that this induces a unique linear map T = T∧ : ∧V1 → ∧V2 of multivectors, which is in fact a homomorphism with respect to the exterior product. When V1 and V2 are inner product spaces, we saw in Proposition 4.1.10 that T∧ : 4V1 → 4V2 will be a homomorphism with respect to Clifford products if and only if T is an isometry. In this section, we study in what sense a linear map T : V1 → V2 induces a linear map of spinors / 2. T4 / 1 → 4V / : 4V To avoid extra technicalities, we consider only Euclidean spinors in this section. Consider first an invertible isometry T between Euclidean spaces. We have fixed complex Vi representations ρi : Vi → L(4V / i ), i = 1, 2. This means that on V = V1 , we have the two complex V representations ρ1 and ρ2 T . It follows from Theo/ 2 ) such that rem 5.2.3 that there exists an invertible map T4 / 1 ; 4V / ∈ L(4V T4 / (w.ψ) = (T w).(T4 / ψ),

/ 1. w ∈ 4V1 , ψ ∈ 4V

In odd dimension, some care about how the main reflectors map is needed. See Proposition 5.3.5. This construction of induced maps T4 / of spinors leads to the following two questions, which we address in this section. • How unique can we make the spinor map T4 / induced by the vector map T ? • Is there a natural way to define a spinor map T4 / for more general invertible vector maps T that are not isometries? / 2 ) is Concerning the first question, the problem is that T4 / 1 ; 4V / ∈ L(4V unique only as a projective map T4 / 1 /C → 4V / 2 /C, / : 4V

Chapter 5. Spinors in Inner Product Spaces

168

that is, T4 / , λ ∈ C \ {0}. When, for / is unique only up to complex multiples λT4 example, we are constructing and working with spinors over manifolds as we do in Chapter 12, this presents problems. To this end, we next define two additional natural structures, an inner product and a conjugation, on spinor spaces that allow us to obtain induced maps of spinors that are unique only up to sign ±T4 / . Such sign ambiguity will always be present, but this discrete nonuniqueness will not cause any problems. Proposition 5.3.1 (Spinor inner product). Let V be a Euclidean space, with complex spinor space 4V / . Then there exists a Hermitian complex inner product (·, ·i on / such that 4V (ψ1 , v.ψ2 i = (v.ψ1 , ψ2 i, ψ1 , ψ2 ∈ 4V, (5.2) / for all vectors v ∈ V . If (·, ·i0 is another Hermitian inner product for which (5.2) holds, then there is a constant λ > 0 such that (ψ1 , ψ2 i0 = λ(ψ1 , ψ2 i for all ψ1 , ψ2 ∈ 4V / . Proof. Fix a basis for 4V / , view ψi as column vectors and ρ(v) as matrices. Then a sesquilinear duality is uniquely represented by an invertible matrix M such that (ψ1 , ψ2 i = ψ1∗ M ψ2 . Condition (5.2) translated to M is that M ρ(v) = ρ(v)∗ M for all vectors v ∈ V . We note that v 7→ ρ(v)∗ is a second complex V representation, which extends to the algebra homomorphism w 7→ ρ(wc )∗ . Therefore the existence of M , unique up to complex nonzero multiples, follows from Theorem 5.2.3. Note that in odd dimensions, these two representations coincide on 4n Vc . It remains to see that M can be chosen as a self-adjoint positive definite matrix. We note that when we are using the standard representation from Example 5.1.5, all matrices ρ(v) are self-adjoint. Hence M = I can be used in this basis. Proposition 5.3.2 (Spinor conjugation). Let V be a Euclidean space of dimension n, with spinor space 4V / . Then there exists an antilinear map 4V / → 4V / : ψ 7→ ψ † such that ( † (v.ψ) = v.ψ † , n 6≡ 3 mod 4, (5.3) † (v.ψ) = −v.ψ † , n ≡ 3 mod 4, for all vectors v ∈ V and spinors ψ ∈ 4V / , and satisfying ( † (ψ † ) = ψ, n ≡ 0, 1, 2, 7 mod 8, †

(ψ † ) = −ψ, 0

n ≡ 3, 4, 5, 6

mod 8.

If ψ 7→ ψ † is another such map for which this holds, then there is λ ∈ C, |λ| = 1, 0 such that ψ † = λψ † for all ψ ∈ 4V / .

5.3. Mapping Spinors

169

Note that in even dimensions n ≡ 0, 2 mod 8, this spinor conjugation provides a real structure on the complex spinor space, as in Section 1.5, in accordance with Theorem 3.4.13, which shows that in these dimensions the real-Euclidean Clifford algebras are isomorphic to real matrix algebras. The technicalities about signs and dimension in the statement of Proposition 5.3.2 are best understood from its proof. Proof. Fix a basis for 4V / , view ψ as column vectors and ρ(w) as matrices. Write N c for the componentwise complex conjugation of a matrix N . Every antilinear / : ψ 7→ ψ † can be written / → 4V map 4V c

ψ † = (N ψ) for some matrix N . The condition †

(v.ψ) = v.(ψ † )

(5.4) c

c

is equivalent to N ρ(v) = (ρ(v)) N , for all vectors v. We note that that v 7→ (ρ(v)) is a second complex V representation, which extends to the algebra homomorphism c w 7→ (ρ(wc )) , where wc denotes the real structure on 4Vc . Existence of antilinear maps satisfying (5.4), unique up to λ ∈ C \ {0}, follows from Theorem 5.2.3, provided n is even, or if c

c

ρ(e1 ) · · · ρ(en ) = ρ(e1 ) · · · ρ(en ) when n is odd. Using a standard complex representation from Example 5.1.5, we see that this holds unless n ≡ 3 mod 4, since all but m of the matrices representing the basis vectors are real. When n ≡ 3 mod 4, by Theorem 5.2.3 the correct c b = (ρ(wc )) N . relation is N ρ(w) To complete the proof, we claim that using a standard representation, we may choose ρ(e−1 · · · e−m ), n = 2m ≡ 0 mod 4, n = 2m + 1 ≡ 1 mod 4, ρ(e−1 · · · e−m ), N := n = 2m ≡ 2 mod 4, ρ(e1 · · · em ), n = 2m + 1 ≡ 3 mod 4. ρ(e1 · · · em e0 ), c

c

Indeed, we note that ρ(ek ) = ρ(ek ), k ≥ 0, and ρ(ek ) = −ρ(ek ), k < 0, and we verify that N c N = I when n ≡ 0, 1, 2, 7 mod 8 and N c N = −I when n ≡ 3, 4, 5, 6 † † mod 8. This completes the proof, since λ(λψ † ) = |λ|2 (ψ † ) . Definition 5.3.3 (Normed spinor space). Let V be a Euclidean space, with spinor space 4V / . Fix a spinor inner product, by which we mean a Hermitian inner product on 4V / such that all vectors act as self-adjoint maps as in Proposition 5.3.1. This amounts to a choice of the parameter λ > 0.

170

Chapter 5. Spinors in Inner Product Spaces

Fix also a spinor conjugation, by which we mean an antilinear map on 4V / with properties as in Proposition 5.3.2. This amounts to a choice of the parameter |λ| = 1. We refer to the triple (4V, / (·, ·i, ·† ) as a normed spinor space. Lemma 5.3.4 (Compatibility). A spinor inner product and a spinor conjugation are compatible in the sense that c

(ψ1 , ψ2 i = (ψ1 † , ψ2 † i,

ψ1 , ψ2 ∈ 4V. /

Proof. Note that (ψ1 , ψ2 i0 := (ψ1 † , ψ2 † i

c

defines a second Hermitian spinor inner product. By uniqueness in Proposition c 5.3.1, we have (ψ1 , ψ2 i = λ(ψ1 † , ψ2 † i for some λ > 0. In particular, †

†

(ψ, ψi = λ(ψ † , ψ † i = λ2 ((ψ † ) , (ψ † ) i = λ2 (ψ, ψi, so λ = 1.

We can now answer the first question posed above, concerning uniqueness of induced spinor maps. For simplicity we write the inner product and conjugation in both spaces below with the same symbols. Proposition 5.3.5 (Uniqueness of spinor maps). Let V1 , V2 be Euclidean spaces, with normed spinor spaces 4V / 1 and 4V / 2 respectively. Assume that T : V1 → V2 is an invertible isometry. When dim V1 is odd, we assume that the main reflectors wn and T wn in V1 and V2 respectively both act as +I or as −I. Then there exists † † an isometric complex linear map T4 / 1 → 4V / 2 such that (T4 / : 4V / ψ) = T4 / (ψ ) and T4 ψ ∈ 4V, / v ∈ V. (5.5) / (v.ψ) = (T v).(T4 / ψ), 0 0 If T4 /. / is another such map, then T4 / = ±T4

Proof. We saw at the beginning of this section how Theorem 5.2.3 implies the existence of T4 / satisfying (5.5), and every other such map is of the form λT4 / for some λ ∈ C \ {0}. To see that T4 / can be chosen to be isometric, consider the Hermitian inner product (ψ1 , ψ2 i0 := (T4 / 1 . We calculate / ψ1 , T4 / ψ2 i on 4V 0 (ψ1 , v.ψ2 i0 = (T4 / ψ1 , T v.T4 / ψ2 i = (T v.T4 / ψ1 , T4 / ψ2 i = (v.ψ1 , ψ2 i ,

for v ∈ V . Proposition 5.3.1 shows that (T4 / ψ1 , T4 / ψ2 i = µ(ψ1 , ψ2 i for some µ > 0, −1/2 so λT4 will be isometric if |λ| = µ . / To see that T4 / can be chosen to be compatible with spinor conjugation, † −1 consider the antilinear map ψ 7→ T4 / 1 . We calculate / ψ) on 4V / (T4 †

†

†

†

−1 −1 −1 −1 T4 / (v.ψ)) = T4 / ψ)) = T4 / ψ) ) = v.T4 / ψ) , / (T4 / ((T v).(T4 / ((T v).(T4 / (T4

5.3. Mapping Spinors

171

for v ∈ V , where = ±1 depending on the dimension as in Proposition 5.3.2. Since also the square of this antilinear map coincides with the square of spinor † −1 † conjugation on 4V / 1 , we conclude from Proposition 5.3.2 that T4 / ψ) = σψ , / (T4 iα for some σ ∈ C, |σ| = 1. Therefore e T4 / will be compatible with spinor conjugation if e−2iα = σ. These two equations for the modulus and argument of λ have exactly two solutions differing by sign, which completes the proof. We next consider the second question posed above, concerning how to define a map of spinors T4 / , when the map of vectors T : V1 → V2 is not an isometry. Recall that when T is an isometry, writing ρk : Vk → L(4V / k ), k = 1, 2, for the complex Vk representations defining the spinor spaces, we used that ρ2 T was a second complex V1 representation. Comparing this to ρ1 , the existence of T4 / followed from Theorem 5.2.3. When T is not an isometry, this argument breaks down, since ρ2 T is not a complex V1 representation. Indeed, ρ2 (T v)2 = hT vi2 I 6= hvi2 I. What we do in this case is to produce an isometry U : V1 → V2 from T by polar factorization. Proposition 1.4.4 shows that there is a unique isometry U : V1 → V2 such that T = U S1 = S2 U, for some positive symmetric maps S1 ∈ L(V1 ) and S2 ∈ L(V2 ). The formula for this U , which we refer to as the polar isometric factor of T , is U := T (T ∗ T )−1/2 = (T T ∗ )−1/2 T. Definition 5.3.6 (Induced spinor map). Let V1 , V2 be Euclidean spaces, with normed spinor spaces 4V / 1 and 4V / 2 respectively. Assume that T : V1 → V2 is an invertible linear map. Denote by U : V1 → V2 the polar isometric factor of T . If dim V1 is odd, we assume that the main reflectors wn and U wn in V1 and V2 respectively both act as +I or as −I. Then we refer to the two maps / 2 / 1 → 4V T4 / := U4 / : 4V constructed from U as in Proposition 5.3.5 as the spinor maps induced by T . Exercise 5.3.7. Consider the Euclidean plane V , with ON-basis {e1 , e−1 }, and con/ from Examsider the standard representation of the complex spinor space 4V ple 5.1.5(i), equipped with the spinor duality and conjugation from the proofs of Propositions 5.3.1 and 5.3.2. Calculate the two spinor maps T4 / / → 4V / : 4V induced by the linear map T : V → V with matrix 1 0 T = 3 2 in the basis {e1 , e−1 }.

Chapter 5. Spinors in Inner Product Spaces

172

Recall from Section 2.3 that if vector maps T induce multivector maps T∧ , then (T −1 )∧ = (T∧ )−1

and

(T2 ◦ T1 )∧ = (T2 )∧ ◦ (T1 )∧ ,

but (λT )∧ 6= λT∧ and (T1 + T2 )∧ 6= (T1 )∧ + (T2 )∧ . The corresponding result for induced spinor maps holds for isometries, but in general the composition rule fails for non-isometries. Exercise 5.3.8 (Failure of transitivity). (i) Let T : V1 → V2 be an invertible linear map with polar isometric factor U : V1 → V2 . Show that T −1 has polar isometric factor U −1 . Conclude that −1 (T4 /) are the spinor maps induced by T −1 . (ii) Let T1 = T be as in (i), and let T2 : V2 → V3 be an invertible linear map with polar isometric factor U2 : V2 → V3 . Construct T1 and T2 such that U2 U1 is not the polar isometric factor of T2 T1 . Conclude that (T2 )4 / ◦ (T1 )4 / are not in general the spinor maps induced by T2 ◦ T1 . Show, however, that this is the case when at least one of the maps T1 and T2 is an isometry. For the action of general multivectors on spinors, we note the following somewhat surprising result. Proposition 5.3.9 (Polar factorization of induced maps). Let T : V1 → V2 be an invertible linear map between Euclidean spaces, and define its polar isometric factor U : V1 → V2 as above. Let T∧ : ∧V1 → ∧V2 be the ∧ homomorphism induced by T from Proposition 2.3.2, and let U∧ = U4 : 4V1 → 4V2 be the induced 4 (as well as ∧) homomorphism induced by U . Then the polar isometric factor of T∧ equals U∧ . Proof. We have T∧ ((T∧ )∗ T∧ )−1/2 = T∧ ((T ∗ )∧ T∧ )−1/2 = T∧ ((T ∗ T )∧ )−1/2 = T∧ ((T ∗ T )−1/2 )∧ = (T (T ∗ T )−1/2 )∧ = U∧ . For the first equality, see Section 2.7. The third equality uses (A2 )∧ = ((A−1 )∧ )−2 for A = (T ∗ T )1/2 , which is true.

5.4

Abstract Spinor Spaces

In Section 3.3 we introduced the notion of abstract Clifford algebras, among which we treat 4V = (∧V, 4) as the standard Clifford algebra. Similarly, we introduced the concept of abstract spinor spaces in Definition 5.1.1, for which it takes two full

5.4. Abstract Spinor Spaces

173

V -rotations two complete one full rotation of the spinor space. Among these we / from Section 5.2 as the standard spinor space. However, treat the spinor space 4V there is one important difference: all abstract Clifford algebras are isomorphic to 4V , but there are infinitely many nonisomorphic abstract spinor spaces, as we shall see. The goal of the present section is to identify all possible abstract complex spinor and tensor spaces of three and four-dimensional Euclidean space, up to isomorphism. We start by collecting the basic tools from the theory of representations of compact Lie groups, which we need. The only groups we use are G = SO(V ) and G = Spin(V ), where V is a given Euclidean space. • Let G denote a compact Lie group. A complex representation (S, ρ) of G is a complex linear space S together with a smooth homomorphism ρ : G → L(S). If (S, ρ) is a representation of G, and if S 0 ⊂ S is a subspace such that ρ(g)w ∈ S 0 ,

for all w ∈ S 0 , g ∈ G,

then (S 0 , ρ0 ) is said to be a subrepresentation of (S, ρ), where ρ0 (g) denotes the restriction of ρ(g) to the invariant subspace S 0 . If (S, ρ) has no nontrivial subrepresentations, that is none besides S 0 = {0} and S 0 = S, then we say that (S, ρ) is an irreducible representation of G. • Let (S1 , ρ1 ) and (S2 , ρ2 ) be two representations of G. We write LG (S1 , S2 ) := {T ∈ L(S1 , S2 ) ; ρ2 (g)T = T ρ1 (g), g ∈ G} and call T ∈ LG (S1 , S2 ) a G-intertwining map. The representations (S1 , ρ1 ) and (S2 , ρ2 ) are isomorphic if there exists a bijective map T ∈ LG (S1 , S2 ). b be the set of equivalence classes of mutually isomorphic irreducible Let G representations of G. Schur’s lemma shows that if (S1 , ρ1 ), (S2 , ρ2 ) are irreducible representations of G, then (i) LG (S1 , S2 ) = {0} if they are nonisomorphic, and (ii) LG (S1 , S2 ) = {λT ; λ ∈ C \ {0}} if T ∈ LG (S1 , S2 ) is an isomorphism. The proof follows from the observation that if T ∈ LG (S1 , S2 ), then N(T ) is a subrepresentation of S1 and R(T ) is a subrepresentation of S2 . Irreducibility allows us to conclude. • Given a representation (S, ρ) of G, we can write S = S1 ⊕ S2 ⊕ · · · ⊕ Sk ,

(5.6)

where each Sj is an invariant subspace of ρ(g) for all g ∈ G, and each Sj is an irreducible representation of G. For the proof, we construct an auxiliary Hermitian inner product (·, ·i such that G acts isometrically on S, that is, (ρ(g)wi2 = (wi2 ,

for all w ∈ S, g ∈ G.

174

Chapter 5. Spinors in Inner Product Spaces Such invariant inner products are not unique. The existence follows from the well-known result in measure theory that there exists a Haar measure on G, that is, a Borel measure dµ with total measure µ(G) = 1 that is left and right invariant in the sense that Z Z Z f (gx)dµ(x) = f (x)dµ(x) = f (xg)dµ(x), G

G

G

for all Borel measurable functions f : G → C and g ∈ G. Starting from any Hermitian inner product (·, ·i0 on S, we see that Z (w1 , w2 i := (ρ(g)w1 , ρ(g)w2 i0 dµ(g) G

defines a Hermitian inner product that is invariant under G. To obtain a decomposition of the form (5.6), we simply note that if S1 is a subrepresentation, then so is S1⊥ , using a complex analogue of Lemma 4.1.2. We continue to split S1 and S1⊥ further until irreducible subrepresentations have been obtained. • Given a representation (S, ρ) of G, the number of irreducible subrepresentab that it contains is well defined. To tions from each equivalence class α ∈ G see this, assume that S = S1 ⊕ S2 ⊕ · · · ⊕ Sk = S10 ⊕ S20 ⊕ · · · ⊕ Sl0 b and are two decompositions of S into irreducible representations. Fix α ∈ G consider any Si ∈ α appearing in the first decomposition, and p0j : Si → Sj0 : w 7→ p0j (w), where p0j denotes orthogonal projection onto Sj0 . Schur’s lemma implies that Si ⊂ ⊕Sj0 ∈α Sj0 . With this and the reverse result obtained by swapping the roles of Si and Sj0 , we get k = l and ⊕Si ∈α Si = ⊕Sj0 ∈α Sj0 ,

b α ∈ G.

b are present in • A tool for identifying which irreducible representations α ∈ G a given representation (S, ρ), and their multiplicities, is the character χS of (S, ρ). This is the function χS : G → C : g 7→ χS (g) := Tr(ρS (g)), where Tr denotes the trace functional as in Section 1.4. Since Tr(T ρ(g)T −1 ) = Tr(ρ(g)), isomorphic representations have the same character. The Peter– b Weyl theorem in representation theory shows that the characters χα , α ∈ G, of the irreducible representations form an ON-basis for −1 Lcl ) = f (x), for all x, g ∈ G}. 2 (G) := {f ∈ L2 (G) ; f (gxg

5.4. Abstract Spinor Spaces

175

Such functions f are referred to as class functions on G. We shall not use the general fact that the class functions span L2cl (G). To see the orthogonality of χS1 and χS2 for two nonisomorphic irreducible representations, we consider the auxiliary representation of G on S := L(S1 ; S2 ) given by ρ(g)T := ρS2 (g) ◦ T ◦ ρS1 (g −1 ), g ∈ G. c

This is seenR to have character χ(g) = χS1 (g) χS2 (g). Define the linear operator P := G ρ(g)dµ(g) ∈ L(S), using componentwise integration, and note that Z Z Z Z 2 ρ(gh)dµ(g)dµ(h) ρ(h)dµ(h) = P = ρ(g)dµ(g) G G G G Z Z = ρ(g)dµ(g) dµ(h) = P. G

G

Thus P is a projection, and we check that its range is LG (S1 ; S2 ), which gives Z c χ1 (g) χS2 (g)dµ(g) = TrP = dim LG (S1 ; S2 ). G

The orthonormality of characters therefore follows from Schur’s lemma. • As a corollary of the Peter–Weyl theorem, it follows that there are at most countably many nonisomorphic irreducible representations of G. Moreover, the number of R irreducible subrepresentations that a given representation contains equals G |χS |2 . We now apply these tools from Lie group representation theory, starting with the three-dimensional spin group. Fixing an ON-basis for V , we write Spin(n) := Spin(Rn ),

and SO(n) := SO(Rn ).

Recall from Theorem 3.4.13 that 4R3 is isomorphic to C(2). By restricting such an isomorphism to Spin(3) ⊂ 4R3 , this Lie group is seen to be isomorphic to the complex isometries on C2 . Definition 5.4.1. Let the special unitary group SU(2) in two complex dimensions be n a o b 2 2 SU(2) := c c ; a, b ∈ C, |a| + |b| = 1 . −b a By the standard isomorphism SU(2) ↔ Spin(3), we mean a b ↔ a1 − a2 j1 − b1 j2 − b2 j3 ∈ Spin(3) = S 3 ⊂ H, SU(2) 3 −bc ac where j1 = e23 , j2 = e31 , j3 = e12 , and a = a1 + ia2 and b = b1 + ib2 . Here i ∈ C is the algebraic imaginary unit, not related to H.

Chapter 5. Spinors in Inner Product Spaces

176

We set out to find the characters of all irreducible representations of SU(2) = Spin(3) = S 3 . Note that if f is a class function on this group, then it is uniquely determined by its values at points exp(−it) 0 ↔ cos t + j1 sin t = exp(tj1 ) ∈ Spin(3), SU(2) 3 exp(it) 0 for t ∈ [0, π]. Indeed, as noted at the end of Section 3.2, all rotations of pure quaternions ∧2 V can be represented x 7→ qxq −1 , q ∈ Spin(3). In particular, we can rotate any pure quaternion to the line [j1 ] in this way. From hqxi2 = hxi2 = hxqi2 , for all q ∈ S 3 , x ∈ H, it is clear that Lebesgue surface measure on Spin(V ) is invariant under R Spin(3), and therefore equals the Haar measure, modulo the normalizing factor S 3 |dˆ x| = 2π 2 . Consider representations (Vk , ρk ) of SU(2), where Vk := {polynomials P : C2 → C ; P (λz, λw) = λk P (z, w), λ ∈ R, z, w ∈ C}, and ρk (T )P := P ◦ T −1 . Concretely, Vk is spanned by {z k−j wj ; 0 ≤ j ≤ k} and a b c c . (ρk (T )P )(z, w) = P (a z − bw, b z + aw) if T = −bc ac Proposition 5.4.2 (Finding all 3D spinor spaces). The irreducible representations d of the Lie group SU(2) = Spin(3) are indexed by the natural numbers Spin(3) = d {0, 1, 2, 3, . . .}. The representations in the equivalence class k ∈ Spin(3) are (k+1)dimensional and are uniquely determined by the character values k/2 X 1 + 2 cos(2mt), k even, m=1 χk (exp(tj1 )) = (k−1)/2 2 X cos((2m + 1)t), k odd. m=0

With the terminology of Definition 5.1.1, these irreducible representations are abstract spinor spaces if k is odd and abstract tensor spaces if k is even. Proof. Note that z k−m wm is an eigenvector to ρk (exp(tj1 )) with eigenvalue eit(k−2m) . Summing these gives the stated characters. We calculate Z π 1 |χk (exp(tj1 ))|2 4π sin2 tdt |χk (g)| dµ(g) = 2 2π 3 S 0 2 Z π X Z k 2 2 π 1 it(k−2m) it −it = e (e − e ) dt = | sin((k + 1)t)|2 dt = 1, π 0 2i m=0 π 0

Z

2

5.4. Abstract Spinor Spaces

177

so these representations are irreducible. Moreover, it is known from Fourier theory ∞ is dense L2 , so these are all the irreducible representations. For that {cos(jx)}j=0 the last statement, we note that −1 ∈ Spin(V ) acts by multiplication with (−1)k on Vk . Example 5.4.3 (4R3 ). The standard representation of Spin(V ) on the three-dimensional vector space Vc itself is given by ρ = p from Proposition 4.1.9. The rotor q = exp(tj1 ) acts by rotation the angle 2t and with plane of rotation [j1 ]. Summing the diagonal elements of the matrix for this rotation in the standard basis yields d ) the character χ(exp(tj1 )) = 1 + 2 cos(2t). Thus the tensor space 2 ∈ Spin(V is the standard representation Vc = ∧1 Vc . Through the Hodge star map and Proposition 4.1.10, ∧2 Vc is an isomorphic representation. Even more trivial: the d ), where representations on ∧0 Vc and ∧3 Vc are both representatives of 0 ∈ Spin(V all rotors act as the identity on a one-dimensional space. / for three-dimensional Eu/ 3 ). Consider the spinor space 4V Example 5.4.4 (4R clidean space V from Definition 5.2.4, and restrict this to a complex Spin(V ) representation. Using the matrices from Example 5.1.7(ii), we have 0 1 0 −i i 0 , ρ(j1 ) = ρ(e1 e−1 ) = = 1 0 i 0 0 −i it e 0 tj1 = and therefore the character for this representation is χ(e ) = Tr 0 e−it d ) is the 2 cos t. We conclude that the smallest abstract V -spinor space 1 ∈ Spin(V standard spinor space 4V / . The following two examples build on Section 8.2, but are not used elsewhere and may be omitted and returned to after Section 8.2 has been read. Example 5.4.5 (Harmonic polynomials). Consider the abstract V -tensor space of scalar (complexified) k-homogeneous harmonic polynomials Pksh from Example 5.1.2(iii). To avoid dealing with bases for these spaces, we recall Proposition 8.2.3, which amounts to a statement about decomposition of the representation on all k-homogeneous polynomials into subrepresentations. We have the representation ρ(q)P (x) = P (q −1 xq) on s Pks = Pksh ⊕ |x|2 Pk−2 , s s where both the terms are subrepresentations, and |x|2 ∈ LSpin(3) (Pk−2 , |x|2 Pk−2 ) sh is an isomorphism. It follows that the character for the representation Pk is the s difference between the characters of the two representations Pks and Pk−2 . To avoid unnecessarily technical trace computations, we choose the convenient basis β γ {x1α z β z γ }α+β+γ=k = {xα 1 (x2 + ix3 ) (x2 − ix3 ) }α+β+γ=k

for Pks . Here {x1 , x2 , x3 } are the coordinates in the ON-basis {e1 , e2 , e3 } for V , that is, the dual basis for V ∗ = P1s . From ρ(etj1 )P (x) = P (e−tj1 xetj1 ) we see that

Chapter 5. Spinors in Inner Product Spaces

178

{x1α z β z γ } is an eigenbasis of the operator ρ(etj1 ) with eigenvalues e2it(β−γ) . This gives the character X X χ(etj1 ) = e2it(β−γ) = e2it(β−γ) α+β+γ=k

β+γ≤k

for the representation Pks , and therefore the character χ(etj1 ) =

X

e2it(β−γ) = 1 +

k X

cos(2mt)

m=0

β+γ=k−1,k

for the representation Pksh . We conclude that Pksh is the irreducible V -tensor space d ). 2k ∈ Spin(V Example 5.4.6 (Monogenic polynomials). Consider the abstract V -spinor space of k-homogeneous monogenic polynomials Pkm from Example 5.1.2(iv), for a threedimensional Euclidean space V . On this real linear space of dimension 8(k + 1), there is a rather natural complex structure, namely multiplication by the orientation e123 ∈ 43 V , which belongs to the center of the algebra. In this way, we consider Pkm as a 4(k + 1)-dimensional complex spinor space with rotors acting as ρ(q)P (x) = qP (q −1 xq). Similar to Example 5.4.5, having Proposition 8.2.3 in mind, we first compute the character of the spinor space Pk (also considered as a complex linear space) with same same action. To handle the multivector-valued polynomials, we note that we have an isomorphism of representations Pk ↔ H ⊗ Pks , where q ∈ Spin(V ) acts on H as x 7→ qx, and on Pks as P (x) 7→ P (q −1 xq). It follows that the character is X χPk (etj1 ) = χH (etj1 )χPks (etj1 ) = 4 cos t e2it(β−γ) , β+γ≤k

from which we get χPkm (etj1 ) = χPk (etj1 ) − χPk−1 (etj1 ) = 4 cos t

k X j=0

e2it(k−2j) = 4

k X

cos((2j + 1)t).

j=0

Proposition 5.4.2 now shows that the V -spinor space Pkm is not irreducible but d contains two copies of the irreducible spinor space 2k + 1 ∈ Spin(3). We can find such a subrepresentation as a minimal left ideal S ⊂ 4V . Then the subspace of monogenic polynomials Pkm with values in S will be such an irreducible spinor subrepresentation. Note that such subrepresentations are not unique. And indeed, there are infinitely many left ideals S that can be used.

5.4. Abstract Spinor Spaces

179

We next study representations of the group Spin(4) of rotors in four-dimensional Euclidean space V . It is a fortunate fact that we have an isomorphism Spin(4) ↔ Spin(3) × Spin(3), as we shall see. This should not come as a surprise, since 4ev V , by Proposition 3.3.5, is isomorphic to the Clifford algebra for R3 with negative definite inner product, which in turn is isomorphic to H ⊕ H according to Theorem 3.4.13. To make these isomorphisms explicit, fix an ON-basis {e1 , e2 , e3 , e4 } and an orientation e1234 for V , and define p± := 21 (1 ± e1234 ). We have p2± = p± ,

p+ + p− = 1,

ev

and p± ∈ Z(4 V ). Therefore p± split the even subalgebra ev 4ev V = 4ev + V ⊕ 4− V, ev where 4ev ± V := p± 4 V are the two two-sided ideals in the even subalgebra. The subspaces 4ev V are real algebras in themselves, each being isomorphic to H, but ± should not be considered subalgebras of 4ev V , since p± , and not 1, is the identity ev element in 4ev ± V . To make the isomorphism 4 V = H ⊕ H explicit, we identify basis elements as 1 4ev + V 3 p+ = p+ e1234 = 2 (1 + e1234 ) ↔ e ∈ H, 1 4ev + V 3 p+ e41 = p+ e23 = 2 (e41 + e23 ) ↔ j1 = −i ∈ H, 1 4ev + V 3 p+ e42 = −p+ e13 = 2 (e42 − e13 ) ↔ j2 = −j ∈ H, 1 4ev + V 3 p+ e43 = p+ e12 = 2 (e43 + e12 ) ↔ j3 = −k ∈ H,

with notation as in Section 3.2, but writing e ∈ H for the identity in H to avoid confusion with 1 ∈ 4ev V . We then use the self-inverse automorphism 4ev V → 4ev V : w 7→ e4 we4 ev 0 0 to identify 4ev − V = 4+ V , and write e := e4 ee4 = p− and jk := e4 jk e4 , k = 1, 2, 3.

Proposition 5.4.7 (Spin(4) = Spin(3) × Spin(3)). Let V be a four-dimensional Euclidean space. We have an algebra isomorphism 4ev V → H ⊕ H : w 7→ (p+ w, e4 p− we4 ), where we identify H = 4ev , + V as above. Letting h·, ·iH = 2h·, ·i4ev + V this isomorphism is an isometry. This algebra isomorphism restricts to a group isomorphism Spin(V ) → S 3 ×S 3 , where S 3 = Spin(3) is the unit quaternion 3-sphere, as in the following diagram: / H⊕H 4evO V (5.7) O ? Spin(V ) p

SO(V )

? / S3 × S3 p˜

/ SO(H)

180

Chapter 5. Spinors in Inner Product Spaces

The standard covering map p : Spin(V ) → SO(V ) : q 7→ q(·)q −1 corresponds to the covering map p˜ : S 3 × S 3 → SO(H) : (q1 , q2 ) 7→ q1 (·)q2−1 if we use the linear identification V → H : v 7→ p+ ve4 . Proof. From the above discussion it is clear that 4ev V → H ⊕ H : w 7→ (p+ w, e4 p− we4 ) is an algebra isomorphism. To check the stated mapping of the spin group, write w ∈ 4ev V as w = (w0 1 + w1234 e1234 ) + (w12 e12 + w34 e34 ) + (w13 e13 + w24 e24 ) + (w14 e14 + w23 e23 ) = (w0 + w1234 )e + (w0 − w1234 )e0 + (−w14 + w23 )j1 + (w14 + w23 )j10 + (−w13 − w24 )j2 + (−w13 + w24 )j20 + (w12 − w34 )j3 + (w12 + w34 )j30 = (x0 e + x1 j1 + x2 j2 + x3 j3 ) + (y0 e0 + y1 j10 + y2 j20 + y3 j30 ). It follows that x20 + x21 + x22 + x32 = 1 = y02 + y12 + y22 + y32 if and only if |w| = 1 and w0 w1234 − w14 w23 + w13 w24 − w12 w34 . By Example 4.1.8 this is equivalent to w ∈ Spin(V ). Next consider the action of Spin(V ) on V with p. We have qvq −1 = qv q = (p+ q + p− q)v(p+ q + p− q) = q1 ve4 q 2 e4 + e4 q2 e4 vq 1 , where q1 := p+ q, q2 := e4 p− qe4 ∈ H = 4ev + V . Note that two terms vanish, since vp+ = p− v and p+ p− = 0. To write this action entirely in terms of quaternions, we need to identify V = H. Multiplying v ∈ V by e4 , we have ve4 ∈ 4ev V . Projecting onto the subspace 4ev + V , we verify that p+ ve4 = v1 i + v2 j + v3 k + v4 e ∈ H = 4ev +V if v = v1 e1 + v2 e2 + v3 e3 + v4 e4 ∈ V . We obtain p+ (qvq −1 )e4 = q1 (p+ ve4 )q 2 + (p+ e4 q2 e4 )vq 1 e4 = q1 (p+ ve4 )q2−1 , since p+ e4 q2 = 0, which completes the proof.

We now construct all irreducible representations of Spin(V ) for a four-dimensional Euclidean space V . Fix an ON-basis {e1 , e2 , e3 , e4 } and consider the subgroup T := {eθ1 e12 +θ2 e34 ; θ1 , θ2 ∈ R} ⊂ Spin(V ). In representation theory, such a subgroup is referred to as a maximal torus. This subgroup plays the same role for Spin(4) as did the circle, the one-dimensional

5.4. Abstract Spinor Spaces

181

torus, {etj1 ; t ∈ R} for Spin(3). Using the isomorphism from Proposition 5.4.7, we have e12 ↔ (j3 , j3 ) and e34 ↔ (−j3 , j3 ), and thus eθ1 e12 +θ2 e34 ↔ (et1 j3 , et2 j3 ),

where t1 = θ1 − θ2 , t2 = θ1 + θ2 .

The importance of the maximal torus T ⊂ Spin(V ) is that a class function f : Spin(V ) → C is uniquely determined by its values on T , in the sense that for each element q ∈ Spin(V ) there is q1 ∈ Spin(V ) such that q1 qq1−1 ∈ T . In fact, only the values on part of T are needed. Using the S 3 × S 3 characterization, this result carries over from three dimensions, but we can also prove it directly for Spin(V ). According to Proposition 4.3.9(i), which we note is a result on maximal tori in the Lie groups Spin(V ), each q ∈ Spin(V ) belongs to one such maximal torus (with e10 e20 , e03 e40 instead of e12 , e34 ). We can then find q1 ∈ Spin(V ) such that the induced rotation w 7→ q1 wq1−1 of 4V maps e10 e02 , e03 e04 to e12 , e34 , possibly after first having adjusted the angles so that the two bases have the same orientation. Proposition 5.4.8 (Finding all 4D spinor spaces). For the Lie group Spin(4) = Spin(3)×Spin(3), the irreducible representations are indexed by the pairs of natural d numbers Spin(4) = N2 = {(k, l) ; k, l = 0, 1, 2, 3, . . .}. The representations in the d are (k + 1)(l + 1)-dimensional and are uniquely equivalence class (k, l) ∈ Spin(4) determined by the character values χ(k,l) (eθ1 e12 +θ2 e34 ) = χk (e(θ1 −θ2 )j3 )χl (e(θ1 +θ2 )j3 ), d of where χk denotes the character for the irreducible representation k ∈ Spin(3) Spin(3) from Proposition 5.4.2. The irreducible abstract V -tensor spaces correspond to those pairs for which k + l is even, and the irreducible abstract V -spinor spaces correspond Rto those for which k + l is odd. The integral Spin(4) f (x)dµ(x) of a class function f on Spin(4) with respect to Haar measure equals Z πZ π 4 f (exp( 12 (t1 + t2 )e12 + 12 (−t1 + t2 )e34 )) sin2 t1 sin2 t2 dt1 dt2 . π2 0 0 Proof. It is clear that the Haar measure on Spin(3) × Spin(3) is the product measure dµ(x1 )dµ(x2 ). We see that a class function is determined by its values on the quarter {(et1 j3 , et2 j3 ) ; 0 < t1 , t2 < π} of the maximal torus, and the stated integral formula follows by translating back to Spin(V ). Now let (Vk , ρk ) be the irreducible representations of SU(2) = Spin(3) used in the proof of Proposition 5.4.2. For (k, l) ∈ N2 , define the representation ρ(k,l) of Spin(3) × Spin(3) on the tensor product space Vk ⊗ Vl by applying the universal property to the bilinear map (P1 , P2 ) 7→ (ρk (q1 )P1 ) ⊗ (ρl (q2 )P2 ) to obtain a linear map ρ(k,l) (q1 , q2 ) ∈ L(Vk ⊗ Vl ) for each (q1 , q2 ) ∈ Spin(3) × Spin(3). The character of this representation is seen to be χ(k,l) (q1 , q2 ) =

182

Chapter 5. Spinors in Inner Product Spaces

χk (q1 )χl (q2 ), and Fubini’s theorem shows that

R Spin(4)

|χ(k,l) (q1 , q2 )|2 = 1, so this is

an irreducible representation of Spin(3)× Spin(3). In particular, for q1 = e(θ1 −θ2 )j3 and q2 = e(θ1 +θ2 )j3 we obtain the stated character values. 2 2 Since {χk (etj3 )}∞ k=0 is an ON-basis for L2 ((0, π); π sin tdt), the functions t1 j3 t2 j3 ∞ {χk (e )χl (e )}k,l=0 form an ON-basis for L2 ((0, π) × (0, π); π42 sin2 t1 sin2 t2 dt1 dt2 ). Hence the representations ρ(k,l) constitute all the possible irreducible representations of Spin(3) × Spin(3) = Spin(4). Example 5.4.9 (4R4 ). The standard representation of Spin(V ) on the four-dimensional space Vc itself is given by ρ = p from Proposition 4.1.9. The rotor q = eθ1 e12 +θ2 e34 acts by rotation in the plane [e12 ] through an angle 2θ1 and rotation in the plane [e34 ] through an angle 2θ2 . This gives the character χ(eθ1 e12 +θ2 e34 ) = 2 cos(2θ1 ) + 2 cos(2θ2 ) = (2 cos(θ1 − θ2 ))(2 cos(θ1 + θ2 )), d so (1, 1) ∈ Spin(4) is the standard V -tensor space. Turning to the induced representation of Spin(V ) on ∧2 V , this is not irreducible for a four-dimensional space V . Indeed, it splits into two subrepresentations 2 ev on ∧2± V := 4ev ± V ∩ ∧ V , where 4± V are the two two-sided ideals of the even subalgebra as above. Using the basis {e41 + e23 , e42 − e13 , e43 + e12 } for ∧2+ V , we obtain the character χ∧2+ V (eθ1 e12 +θ2 e34 ) = 1 + 2 cos(2(θ1 − θ2 )), d so this is the V -tensor space (2, 0) ∈ Spin(4). On the other hand, a trace calculation with the basis {−e41 +e23 , e42 +e13 , −e43 +e12 } for ∧2− V gives χ∧2− V (eθ1 e12 +θ2 e34 ) = 1 + 2 cos(2(θ1 + θ2 )), and we conclude that ∧2− V is the V -tensor space (0, 2) ∈ d Spin(4). Note the somewhat surprising result that the standard four-dimensional V -tensor space (1, 1) is not the smallest: there are two nonisomorphic threedimensional V -tensor spaces (2, 0) and (0, 2)! Example 5.4.10 (4R / 4 ). Consider the spinor space 4V / from Definition 5.2.4 for a four-dimensional space V and the restriction of ρ to Spin(V ) ⊂ 4ev V . We pick bases {1, e12 , e1 , e2 } for 4V / and {e1 , e−1 , e2 , e−2 } for V . Using the matrices from Example 5.1.8(i), we obtain cos(θ1 − θ2 ) − sin(θ1 − θ2 ) 0 0 sin(θ1 − θ2 ) cos(θ1 − θ2 ) 0 0 . ρ(eθ1 e12 +θ2 e34 ) = 0 0 cos(θ1 + θ2 ) sin(θ1 + θ2 ) 0 0 − sin(θ1 + θ2 ) cos(θ1 + θ2 )

5.5. Comments and References

183 +

The block structure of this matrix is due to the two subrepresentations 4 / V − and 4 / V . For these we obtain the characters χS + = 2 cos(θ1 − θ2 ) and χS − = 2 cos(θ1 + θ2 ) respectively, so these two representations are the two irreducible d respectively. V -spinor spaces (1, 0) and (0, 1) ∈ Spin(4) Exercise 5.4.11. Extend Examples 5.4.5 and 5.4.6 to four-dimensional Euclidean space, and find the irreducible subrepresentations contained in (Pksh )c and (Pkem )c . Note that in this case, (Pkem )c 6= Pkm .

5.5

Comments and References

´ Cartan. Here is a quotation from 5.1 Spinors in general were discovered by Elie his book [26]: “Spinors were first used under that name, by physicists, in the field of Quantum Mechanics. In their most general form, spinors were discovered in 1913 by the author of this work, in his investigations on the linear representations of simple groups.” There exist many variations of the construction of spinors. The complex spinors that we construct here are usually referred to as Dirac spinors in physics. The method of imaginary rotations used in the proof of Proposition 5.1.6 to reduce to the case of Euclidean space is, in the case of spacetime, referred to as Wick rotation. An inspirational book on Clifford algebras and spinors is Lounesto [64]. 5.2 In Chapter 3, we recycled Grassmann’s exterior algebra ∧V , and identified the Clifford algebra 4V and the exterior algebra ∧V . This space of multivectors is the same as a linear space, but we have two different associative products 4 and ∧ on it, leading to two different algebras. In view of Ex/ as some ample 5.1.5, it is tempting to try to realize the spinor space 4V suitable subspace of a Clifford algebra. One obvious way is to view 4V / as a minimal left ideal in 4V , as first proposed by M. Riesz. However, there is no canonical choice of minimal left ideal, and there is no geometric reason for such an identification. The ideal point of view is also problematic in considering spinor bundles as in Section 11.6. The approach in this book is that the spinor space 4V / is a new independent universe, without any relation to other spaces other than that multivectors act as linear operators on spinors, and that it is unnatural to try to set up any identification between the spinor space and any other space. As before we follow the principle of abstract algebra, as explained in the introduction of Chapter 1, to construct / . However, to calculate with spinors one may choose any favorite repre4V sentation of them, which may consist in using some ad hoc identification. In this book, we choose the identification with a certain exterior algebra in Example 5.1.5.

184

Chapter 5. Spinors in Inner Product Spaces / was introduced by Feynman in physics to denote The slash notation D the representation of a vector by gamma matrices. In this book we use the notation differently. We have chosen to use the slash as a symbol for spinor objects, to distiguish them from the related Clifford algebra objects. For / for spinor space as compared to 4V for Clifford algebra, and example 4V / will denote Dirac operators acting on spinor fields as compared to D for D Dirac operators acting on multivector fields. We mainly denote objects in spinor spaces by ψ, following the tradition from quantum mechanics, where they represent the wave functions of particles.

5.3 The notion of spinor conjugation is related to charge conjugation in physics, which is a transformation that switches particles and antiparticles. See Hladik [59]. A reference for spinor inner products is Harvey [50]. Spinor inner products and conjugation are well-defined natural structures on spinor spaces also for general inner product spaces. In Section 9.2 we study the case of spacetime with three space dimensions, and explain how spinors are used in the Dirac theory in quantum mechanics. The construction of induced spinor maps using polar decomposition of non-isometries is due to Bourgignon [22]. A motivating application is to construct a map of spinor fields between nonisometric manifolds, and this was / operators in Bandara, McIntosh, used in studying perturbations of 4-Dirac and Ros´en [17]. 5.4 The representation theory for compact Lie groups used in this section can be found in many textbooks on the subject, including results for higherdimensional spin groups beyond the three- and four-dimensional examples that we limit ourselves to in this book. Our discussion of small spin groups gathers inspiration from Br¨ocker and tom Dieck [24]. See also Fulton and Harris [41] and Gilbert and Murray [42]. The spin representation from Example 5.4.6 was introduced by Sommen [86, 87, 88].

Chapter 6

Interlude: Analysis Road map: This chapter is not where to start reading the second part of this book on multivector analysis, which rather is Chapter 7. The material in the present chapter is meant to be used as a reference for some background material from analysis, which we use in the remaining chapters. A main idea in this second part is that of splittings of function spaces: H = H1 ⊕ H 2 .

(6.1)

When H1 is a k-dimensional subspace of a linear space H of finite dimension n, then every subspace H2 of dimension n − k that intersects H1 only at 0 is a subspace complementary to H1 . When H is an infinite-dimensional linear space, which is typically the case for the linear spaces of functions that we use in analysis, it is seldom of any use to have only an algebraic splitting (6.1), meaning that every x ∈ H can be written x = x1 + x2 , for unique x1 ∈ H1 and x2 ∈ H2 . Using the axiom of choice, one can show that every subspace H1 has an algebraic complement H2 . Instead, we restrict attention to topological splittings of Hilbert and Banach spaces H, meaning that we assume that H1 and H2 are closed subspaces and that we have an estimate kx1 k + kx2 k ≤ Ckxk for some C < ∞. See Definition 6.4.5. The Hilbert space interpretation of this latter reverse of the triangle inequality is that the angle between H1 and H2 is positive. We consider splittings of Banach spaces also, and it should be noted that closed subspaces H1 of a Banach space H in general do not have any topological complement H2 . Such a complementary subspace exists precisely when there exists a bounded projection onto H1 . A well-known example is the subspace c0 of the sequence space `∞ , which does not have any topological complement at all. In © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_6

185

Chapter 6. Interlude: Analysis

186

⊥ Hilbert space, though, we can always use the orthogonal √ complement H2 = H1 . In this case we have kx1 k2 + kx2 k2 = kxk2 , and C = 2 suffices. When doing analysis, not only with scalar valued functions but with more general multivector fields, we can replace the Laplace operator ∆ by more fundamental first-order partial differential operators d, δ, and D. When working with these first-order operators on domains D, splittings of function spaces occur frequently. In Section 7.6, Chapter 10, and Section 9.6 we encounter the Hodge splittings associated with the exterior and interior derivatives d and δ. These are orthogonal splittings of L2 (D), involving also a finite-dimensional third subspace. In Section 8.3 we encounter the Hardy splittings associated to the Dirac operator D. These are in general nonorthogonal, but topological, splittings of L2 (∂D). In Section 9.3 we consider generalizations of the Hardy splittings for time-harmonic waves, and in Sections 9.4 and 9.5 we show that the fundamental structure behind elliptic boundary value problems consists not only of one splitting, but two independent splittings: one related to the differential equation and one related to the boundary conditions. For Dirac’s original equation for the time evolution of spin-1/2 particles in relativistic quantum mechanics, we also see in Section 9.2 how splittings appear, for example, in the description of antiparticles and chirality. Splittings also appear in index theory for Dirac operators in Chapter 12, where the Dirac operators are considered in the splittings L2 (M ; 4M ) = + − L2 (M ; 4ev M )⊕L2 (M ; 4od M ) and L2 (M ; 4M / ) = L2 (M ; 4 / M )⊕L2 (M ; 4 / M ), respectively, on our manifold M . In the present chapter, the material in Section 6.4 should not be needed before Section 9.4, with the exception of splittings of function spaces.

6.1

Domains and Manifolds

An extremely useful notation for estimates in analysis is the following, which we use in the remainder of this book. Definition 6.1.1 (The analyst’s (in)equality). By the notation X . Y , we mean that there exists C < ∞ such that X ≤ CY for all relevant values of the quantities X and Y . By X & Y we mean Y . X, and X ≈ Y means X . Y and X & Y . We use the following standard terminology and notation concerning the regularity of functions. Consider a function f : D → L, with D an open subset of some affine space X, and L a linear space, or possibly affine. Denote by B(x, r) := {y ∈ X ; |y − x| < r} the ball with center x and radius r > 0 when X is Euclidean space. • The function f is C k -regular in D, k = 0, 1, 2, . . ., if all directional/partial derivatives ∂v1 · · · ∂vm f (x) of order m ≤ k exist as continuous functions of x ∈ D, for all directions vi ∈ V . Here ∂v f (x) := lim (f (x + hv) − f (x))/h. h→0

6.1. Domains and Manifolds

187

Given a basis {ei }, we write ∂i := ∂ei . We say that f is C ∞ -regular if it is C k -regular for all k < ∞. older regular of order 0 < α < 1 in D if • The function f is H¨ |f (x) − f (y)| . |x − y|α ,

for all x, y ∈ D,

and we write f ∈ C α (D; L). For α = 0, f ∈ C 0 (D; L) = C(D; L) means that f is continuous on D. When α = 1, we say that f is Lipschitz regular and write f ∈ C 0,1 (D; L). Note that the precise value of the implicit constant C as in Definition 6.1.1, but not the H¨older or Lipschitz property of f , depends on the choice of Euclidean norm | · | on X. • A bijective function f : D → D0 , with an open set D0 ⊂ L, is a homeomorphism if f ∈ C 0 (D; D0 ) and f −1 ∈ C 0 (D0 ; D). Lipschitz diffeomorphisms and C k -diffeomorphisms are defined similarly. A diffeomorphism refers to a C ∞ -diffeomorphism. • The support of a function defined in X is the closed set supp f := {x ∈ X ; f (x) 6= 0}. If f ∈ C ∞ (D), D an open set, then we write f ∈ C0∞ (D) if supp f is a compact subset of D. • Write C k (D) := {F |D ; F ∈ C k (X)}, and similarly for C α and C ∞ . When the range L of the function is clear from the context, we suppress L in the notation and abbreviate C k (D; L) to C k (D). Definition 6.1.2 (Total derivative). Let D ⊂ X be an open set in an affine space X with vectors V , and let (X 0 , V 0 ) be a second affine space. If ρ : D → X 0 is differentiable at x ∈ D, then we define its total derivative at x to be the unique linear map ρx : V → V 0 such that ρ(x + v) − ρ(x) = ρx (v) + o(v), where o(v) denotes a function λ(v) such that λ(v)/|v| → 0 when v → 0. With respect to bases {ei } and {e0i }, ρx has matrix ∂1 ρ1 (x) · · · ∂k ρ1 (x) .. .. .. , . . . ∂1 ρn (x) · · ·

∂k ρn (x)

P 0 where ρ = ρi ei and partial derivatives ∂i = ∂ei are with respect to ei . Equivalently, the total derivative is ρx (v) = ∂v ρ(x). Note that when ρ maps between affine spaces, then the total derivative ρx maps between the vector spaces since differences of points are vectors. To simplify notation, we shall often drop subscripts x and write ρ.

Chapter 6. Interlude: Analysis

188

The total derivative of a differential map between affine spaces extends from a map of vectors to a map of multivectors as in Section 2.3. With our notation, for example, the chain rule takes the form ρ2 ◦ ρ1 x (w) = ρ2 ρ

1 (x)

ρ1 x (w),

w ∈ ∧V1 ,

for the composition of maps ρ1 : X1 → X2 and ρ2 : X2 → X3 . Definition 6.1.3 (Jacobian). Let ρ : D → X 0 be as in Definition 6.1.2, with total derivative ρx : V → V 0 . Denote by ρx : ∧V → ∧V 0 the induced linear map. Assume that X, X 0 are oriented n-dimensional affine spaces with orientations en and e0n respectively. Then its Jacobian Jρ (x) is the scalar function representing ρx |∧n V , that is, the determinant Jρ (x) := he0∗ n , ρx (en )i of ρx . The main use of Jacobians is in the change of variables formula Z Z f (y)dy = f (ρ(x))Jρ (x)dx ρ(D)

(6.2)

D

for integrals. For Lipschitz change of variables ρ, this continues to hold. Note that in this case Jρ is well defined almost everywhere, since Lipschitz maps ρ are differentiable almost everywhere by Rademacher’s theorem. We use the following standard terminology for domains D ⊂ X. Definition 6.1.4 (Domains). Let D be a domain, that is, an open subset, in an ndimensional affine space (X, V ). We say that D is a C k -domain, k = 1, 2, . . ., if its boundary D is C k -smooth in the following sense. At each p ∈ ∂D, we assume that there exists a C k diffeomorphism ρ : Ωp → Dp between a neighborhood Ωp ⊂ Rn of 0 and a neighborhood Dp ⊂ X such that ρ({x ∈ Ωp ; xn > 0}) = Dp ∩ D, ρ({x ∈ Ωp ; xn = 0}) = Dp ∩ ∂D, and ρ({x ∈ Ωp ; xn < 0}) = Dp \ D. Lipschitz domains are defined similarly, by requiring that the local parametrizations ρ be C 0,1 diffeomorpisms. In a Euclidean space X, we denote by ν the outward-pointing unit normal vector field on ∂D. For a C k -domain, ν is a C k−1 -regular vector field defined on

6.1. Domains and Manifolds

189

all ∂D. For a Lipschitz domain, by Rademacher’s theorem, ν is well defined at almost every point p ∈ ∂D. In many cases it is important to consider domains beyond C 1 , such as Lipschitz domains. For example, the intersection and union of two C 1 domains is much more likely to be Lipschitz than C 1 . However, as the following example indicates, Lipschitz domains constitute a far wider class than domains with a finite number of corners, edges, etc. Example 6.1.5 (Lipschitz scale invariance). We consider how a function φ : R → R scales. Assume that φ(0) = 0 and let φn (x) := nφ(x/n). Thus the graph of φn represents what φ looks like around 0 through a magnifying glass that magnifies n times. If φ is C 1 regular, then |φn (x) − φ0 (0)x| ≤ n |x|, |x| < 1, where n → 0 when n → ∞. This means that φ “looks flat” on small enough scales, since it is well approximated by the straight line y = φ0 (0)x. On the other hand, if φ is a Lipschitz function, then φn is another Lipschitz function with the same Lipschitz constant C. In contrast to the C 1 case, φn will not converge to a linear function, as is seen, for example, from φ(x) = |x|, for which φn (x) = |x| for all n. However, this example is very atypical for Lipschitz functions. In general, each φn will give an entirely new function. This means that a Lipschitz function is nontrivial, that is, nonflat, on each scale, but still nondegenerate, that is, still a Lipschitz function. By the implicit function theorem, the boundary of a C k domain, k = 1, 2, . . ., is locally the graph of a C k function, in the sense that the local parametrization ρ can be written ρ(x0 , xn ) = (x0 , xn + φ(x0 )),

x0 ∈ Rn−1 ,

(6.3)

in a suitable basis for X = V , where φ : Rn−1 → R is a C k -regular function. In stark contrast, this is not true for Lipschitz domains. Example 6.1.6 (Bricks and spirals). (i) In R3 , let D1 := {(x, y, z) ; −1 < x < 0, −1 < y < 1, −1 < z < 0} and D2 := {(x, y, z) ; −1 < x < 1, −1 < y < 0, 0 < z < 1}. Placing the “brick” D2 on top of D1 , consider the two-brick domain D with D = D1 ∪ D2 . Then D is a Lipschitz domain, but at the origin ∂D is not the graph of a Lipschitz function. (ii) In polar coordinates (r, θ) in R2 , consider the logarithmic spiral D := {(r cos θ, r sin θ) ; e−(θ+a) < r < e−(θ+b) , θ > 0}, where b < a < b + 2π are two constants. Then D is a Lipschitz domain, but at the origin ∂D is not the graph of a Lipschitz function. If D is a Lipschitz domain in which all local parametrizations ρ of ∂D are of the form (6.3) with C 0,1 functions φ, then we say that D is a strongly Lipschitz domain.

190

Chapter 6. Interlude: Analysis

Exercise 6.1.7 (Star-shaped domains). We say that a domain D is star-shaped with respect to some point p ∈ D if for each x ∈ D, the line {p + t(x − p) ; t ∈ [0, 1]} in contained in D. Show that every bounded domain in a Euclidean space that is starshaped with respect to each point in some ball B(p; ) ⊂ X, > 0, is a strongly Lipschitz domain. Conversely, show that every bounded strongly Lipschitz domain is a finite union of such domains that are star-shaped with respect to some balls. Exercise 6.1.8 (Rellich fields). Let D be a bounded strongly Lipschitz domain in a Euclidean space (X, V ). Show that there exists a vector field θ ∈ C0∞ (X; V ) such that inf hν(x), θ(x)i > 0. x∈∂D

A partition of unity, see below, may be useful. Besides open subsets, that is, domains in affine space, we also make use of lower-dimensional curved surfaces. More generally, we require the notion of a manifold from differential geometry, which we now fix notation. We consider only compact manifolds, but both with and without boundary, and in many cases embedded in an affine space. For simplicity, we consider only regularity k ≥ 1. n Our notation is the following. Let H+ := {(x0 , xn ) ; x0 ∈ Rn−1 , xn ≥ 0} and n−1 0 0 n R+ := {(x , xn ) ; x ∈ R , xn > 0} denote the closed and open upper halfspaces, and identify Rn−1 and Rn−1 ×{0}. In general, let M be a compact (second countable Hausdorff) topological space, for example a compact subset of an affine space X. n , in the sense that we are • We assume that M is locally homeomorphic to H+ given a collection of charts, that is, homeomorphisms {µα : Dα → Mα }α∈I , n the atlas S for M , between open sets Dα ⊂ H+ and Mα ⊂ M such that M = α∈I Mα . By compactness, we may assume that the index set I is finite.

• Define open sets Dβα := µ−1 α (Mβ ) ⊂ Dα , and transition maps µβα : Dβα → Dαβ : x 7→ µβα (x) := µ−1 β (µα (x)) for α, β ∈ I. We say that M is a (compact) C k -manifold if µβα ∈ C k (Dβα ) for all α, β ∈ I. In this case, these transition maps are C k diffeomorphisms, ∞ since µ−1 βα = µαβ . A manifold refers to a C -manifold. If all these transition maps are orientation-preserving, then we say that M is oriented. When it is possible to find another atlas with all transition maps between its charts orientation-preserving, then we say that M is orientable. More generally, a chart for M refers to any homeomorphism µ0 : D0 → n M between open sets D0 ⊂ H+ and M 0 ⊂ M such that µ0−1 ◦ µα ∈ 0

0 C k (µ−1 α (M )) for all α ∈ I.

6.2. Fourier Transforms

191

n • If Dα ⊂ R+ for all α ∈ I, then we say that M is a closed manifold. This means that M is a compact manifold without boundary. If Dα ∩ Rn−1 6= ∅ for some α ∈ I, then we say that M is a manifold with boundary. In this case, the boundary of M , denoted by ∂M , is the closed manifold defined as follows. Let Dα0 := Aα ∩ Rn−1 , 0 := µα |Rn−1 , µα

∅, and we may and Mα0 := µ0α (Dα0 ). It suffices to consider α such that Dα0 6= S 0 assume that Dα0 ⊂ Rn−1 + . Then ∂M is the closed manifold α∈I Mα with 0 0 0 atlas {µα : Dα → Mα }α∈I . • When M is a compact n-dimensional C k -manifold that is also a subset of an affine space X, with the topology inherited from X, then we say that M is an n-surface in X if the derivative µα of µα : Dα → Mα ⊂ X is injective for all x ∈ Dα and all α ∈ I. If µα ∈ C k (Dα ; X), then we say that M is a C k -regular n-surface in X. By the inverse function theorem, an n-surface is locally the graph of a C k -regular function in n variables, in a suitably rotated coordinate system for X. As above, n-surfaces may be closed or may have a boundary. If D ⊂ X is a bounded C k -domain in an affine space X as in Definition 6.1.4, then we see that M = D is a compact C k -regular n-surface with boundary. More generally but similarly, we can consider n-surfaces M embedded in some, in general higher-dimensional, manifold N . • For a function f : M → L on a C k manifold M , with values in a linear space L, we define f ∈ C j (M ; L) to mean that f ◦ µα ∈ C j (Dα ; L) for all α ∈ I, when j ≤ k. k A partition S of unity for a C -manifold M , subordinate to a finite covering M = α∈I Mα by open sets MP α ⊂ M , is a collection {µα }α∈I of functions such that supp ηα ⊂ Mα and α∈I ηα (x) = 1 for all x ∈ M . There exists such a partition of unity with ηα ∈ C k (M ; [0, 1]) on every C k -manifold M.

The standard use of a partition of unity is to localize problems: Given a function f on M , we write X f= ηα f. α

Here supp ηα f ⊂ Mα , and by working locally in this chart, we can obtain results for ηα f , which then we can sum to a global result for f .

6.2

Fourier Transforms

This section collects computations of certain Fourier transforms that are fundamental to the theory of partial differential equations. Fix a point of origin in an

192

Chapter 6. Interlude: Analysis

oriented affine space X and identify it with its vector space V . In particular, V is an abelian group under addition, and as such it comes with a Fourier transform. This is the linear operator Z ˆ f (x)e−ihξ,xi dx, ξ ∈ V ∗ . F(f )(ξ) = f (ξ) := V

This Fourier transform maps a complex-valued function f on V to another complex-valued function on V ∗ . If instead f takes values in some complex linear space L, we let F act componentwise on f . Assuming that V is a Euclidean space and V ∗ = V , the fundamental theorem of Fourier analysis is Plancherel’s theorem, which states that F defines, modulo a constant, an L2 isometry: Z Z 1 |f (x)|2 dx = |fˆ(ξ)|2 dξ. n (2π) V V We recall that the inverse Fourier transform is given by Z F −1 (fˆ)(x) = f (x) = fˆ(ξ)eihξ,xi dξ,

x ∈ V,

V∗

and basic formulas F(∂k f (x)) = iξk fˆ(ξ), F(f (x) ∗ g(x)) = fˆ(ξ) · gˆ(ξ), where the convolution of f (x) and g(x) is the function Z (f ∗ g)(x) := f (x − y)g(y)dy. V

The most fundamental of Fourier transforms is 2

F{e−|x| 2

that is, the Gauss function e−|x|

/2

/2

} = (2π)n/2 e−|ξ|

2

/2

,

is an eigenfunction to F.

Proposition 6.2.1 (Gaussians and homogeneous functions). Let f (x) be a homogeneous polynomial of degree j that is harmonic on an n-dimensional Euclidean space V . Then for every constant s > 0, we have Fourier transforms 2

2

F{f (x)e−s|x| } = 2−j cs−(n/2+j) f (ξ)e−|ξ|

/(4s)

,

where c = π n/2 (−i)j . For every constant 0 < α < n, we have Fourier transforms F{f (x)/|x|n−α+j } = 2α c where Γ(z) :=

R∞ 0

Γ((α + j)/2) f (ξ)/|ξ|α+j , Γ((n − α + j)/2)

e−t tz−1 dt is the gamma function, with Γ(k) = (k − 1)!.

193

6.2. Fourier Transforms Proof. (i) Calculating the Fourier integral, we have Z Z 2 −s|x|2 −ihx,ξi −|ξ|2 /(4s) e f (x)e−shx+iξ/(2s)i dx f (x)e dx = e V V Z −|ξ|2 /(4s) −s|x|2 f (x − iξ/(2s))e =e dx V ! Z Z 2

= e−|ξ|

∞

2

e−sr rn−1

/(4s)

f (rω − iξ/(2s))dω dr, |ω|=1

0

where we have extended f to a polynomial of n complex variables. According to the mean value theorem for harmonic functions, Z f (rω + y)dω = σn−1 f (y), |ω|=1

for every y ∈ V , where σn−1 is the area of the unit sphere in V . By analytic continuation, this formula remains valid for all complex y ∈ Vc . Since Z ∞ Z ∞ 1 1 −sr 2 n−1 e dr = n/2 r e−u un/2−1 = n/2 Γ(n/2) 2s 2s 0 0 and σn−1 = 2π n/2 /Γ(n/2), the stated identity follows. (ii) To establish the second Fourier transform identity, we use the identity Z ∞ Z ∞ 2 1 Γ((n − α + j)/2) x(n−α+j)/2−1 e−x dx = . s(n−α+j)/2−1 e−sr ds = n−α+j r rn−α+j 0 0 2

Writing r−(n−α+j) as a continuous linear combination of functions e−sr in this way, we deduce that Z ∞ 2 1 n−α+j F{f (x)/|x| }= s(n−α+j)/2−1 F{f (x)e−s|x| }ds Γ((n − α + j)/2) 0 Z ∞ 2 1 = s(n−α+j)/2−1 2−j cs−(n/2+j) f (ξ)e−|ξ| /(4s) ds Γ((n − α + j)/2) 0 Z ∞ 2 2−j cf (ξ) s−(α+j)/2−1 e−(1/s)(|ξ|/2) ds = Γ((n − α + j)/2) 0 Γ((α + j)/2) f (ξ) = 2α c . Γ((n − α + j)/2) |ξ|α+j The following functions, or more precisely distributions in dimension ≥ 3, appear in solving the wave equation. Proposition 6.2.2 (Riemann functions). Let Rt , for t > 0, be the Fourier multiplier F(Rt f )(ξ) =

sin(t|ξ|) F(f )(ξ). |ξ|

194

Chapter 6. Interlude: Analysis

In low dimensions, the Riemann function Rt has the following expression for t > 0: Z 1 Rt f (x) = f (x − y)dy, dim V = 1, 2 |y| 0, x2 > 0}, then ρ : D1 → D2 is a diffeomorphism. Let F be the constant vector field F (y) = e1 parallel to the y1 -axis. To push forward and pull back F to the x1 x2 -plane, we calculate the derivative a cos y2 −ay1 sin y2 ρy = . b sin y2 by1 cos y2 This gives the pushed forward vector field 1 a cos y2 −ay1 sin y2 1 a cos y2 x1 ρ∗ F = = =p . 2 2 b sin y2 by1 cos y2 0 b sin y2 (x1 /a) + (x2 /b) x2 On the other hand, pulling back F by ρ−1 gives −1 −1 a cos y2 −a−1 y1−1 sin y2 1 a cos y2 (ρ−1 )∗ F = −1 = 0 b−1 sin y2 b sin y2 b−1 y1−1 cos y2 −2 1 a x1 =p . −2 (x1 /a)2 + (x2 /b)2 b x2

220

Chapter 7. Multivector Calculus

Note that ρ∗ F is tangent to the radial lines ρ({y2 = constant}), and that (ρ−1 )∗ F is normal to the ellipses ρ({y1 = constant}), in accordance with the discussion above. See Figure 7.1(b)–(c).

Figure 7.1: (a) Change of variables F ◦ ρ−1 . (b) Pushforward ρ∗ F . (c) Inverse pullback (ρ−1 )∗ F . (d) Normalized pushforward ρ˜∗ F . The field has been scaled by a factor 0.3, that is, the plots are for F = 0.3e1 . Since F is constant, it is of course divergence- and curl-free: ∇yF = 0 = ∇∧F . By direct calculation, we find that (ρ−1 )∗ F is curl-free. For the pushforward, we note that p div(ρ∗ F ) = 1/ (x1 /a)2 + (x2 /b)2 6= 0. However, the normalized pushforward

7.2. Pullbacks and Pushforwards

221

1 x1 ρ˜∗ F = ab((x1 /a)2 + (x2 /b)2 ) x2 is seen to be divergence-free. See Figure 7.1(d). This is in accordance with Theorem 7.2.9 below. We now show that in general, pullbacks commute with the exterior derivative, and dually that normalized pushforwards commute with the interior derivative. At first it seems that taking the exterior derivative of a pulled back multicovector field would give two terms, a first-order term when the derivatives hit Θ(ρ(x)) according to the chain rule and a zero-order term when the derivatives hit ρx according to the product rule. However, it turns out that the zero-order term vanishes miraculously due to the alternating property of the exterior product and the equality of mixed derivatives. Theorem 7.2.9 (The commutation theorem). Let ρ : D1 → D2 be a C 2 map between open sets D1 ⊂ X1 and D2 ⊂ X2 . (i) If Θ : D2 → ∧V2∗ is a C 1 multicovector field in D2 , then the pullback ρ∗ Θ : D1 → ∧V1∗ is C 1 and ∇ ∧ (ρ∗ Θ)(y) = ρ∗ (∇ ∧ Θ)(y) for y ∈ D1 , that is, d(ρ∗ Θ) = ρ∗ (dΘ). (ii) Further assume that ρ is a C 2 diffeomorphism. If F : D1 → ∧V1 is a C 1 multivector field in D1 , then the normalized pushforward ρ˜∗ F : D2 → ∧V2 is C 1 and ∇ y (˜ ρ∗ F )(x) = ρ˜∗ (∇ y F )(x) for x ∈ D2 , that is, δ(˜ ρ∗ F ) = ρ˜∗ (δF ). The proof uses the following lemma, the proof of which we leave as an exercise. Lemma 7.2.10. Let {ei } and {e0i } be bases for V1 and V2 , with dual {e∗i } and P bases 0∗ 0∗ {ei } respectively. Then the pullback of a covector field θ(x) = i θi (x)ei is X ρ∗ θ(y) = θi (x) ∂j ρi (y)e∗j , x = ρ(y) ∈ D2 , y ∈ D1 , i,j

P and the pushforward of a vector field v(y) = i vi (y)ei is X ρ∗ v(x) = vi (y) ∂i ρj (y)e0j , x = ρ(y) ∈ D2 , y ∈ D1 . i,j

Proof of Theorem 7.2.9. Since both y 7→ ρ∗y and y 7→ Θ(ρ(y)) are C 1 , so is ρ∗ Θ. (i) When Θ = f is a scalar field, then the formula is the chain rule. Indeed, changing variables x = ρ(y) in the scalar function f (x), for ρ∗ f = f ◦ ρ we have X ∇y (f (ρ(y))) = e∗i (∂i ρk (y))(∂xk f )(x) = ρ∗y (∇f )(x), i,k

222

Chapter 7. Multivector Calculus

using Lemma 7.2.10. P (ii) Next consider a vector field Θ = θ = i θi ei : D2 → ∧V2∗ . Fix bases {ei } and {e0i } for V1 and V2 respectively and write {e∗i } and {e0∗ i } for the dual bases and ∂i and ∂i0 for the partial derivatives. From Lemma 7.2.10 we have X ∇ ∧ (ρ∗ θ) = ∇y ∧ θi (ρ(y)) ∂j ρi (y) e∗j i,j

=

X

(∂k θi ∂j ρi + θi ∂k ∂j ρi )e∗k ∧ e∗j =

i,j,k

X

∂k θi ∂j ρi e∗k ∧ e∗j ,

i,j,k

since ∂k ∂j = ∂j ∂k and ek ∧ el = −el ∧ ek . This is the key point of the proof. On the other hand, we have X X X 0∗ 0∗ ρ∗ (∇∧θ) = ρ∗ ∂j0 θi e0∗ = ∂j0 θi ρ∗ (e0∗ ∂j0 θi ∂k ρj ∂l ρi e∗k ∧e∗l . j ∧ ei j ∧ ei ) = i,j

i,j

i,j,k,l

Note that j ∂j0 θi ∂k ρj = ∂k θi by the chain rule. Thus changing the dummy index l to j proves the formula for covector fields. (iii) Next consider a general multicovector field Θ. By linearity, we may assume that Θ(x) = θ1 (x) ∧ · · · ∧ θk (x) for C 1 covector fields θj . We need to prove P

X

e∗i ∧ ρ∗ θ1 ∧ · · · ∧ ∂i (ρ∗ θj ) ∧ · · · ∧ ρ∗ θk =

i,j

X

∗ 1 ∗ 0 j ∗ k ρ∗ e0∗ i ∧ ρ θ ∧ · · · ∧ ρ (∂i θ ) ∧ · · · ∧ ρ θ .

i,j

For this, it suffices to show that X X ∗ 0 e∗i ∧ ∂i (ρ∗ θ) = ρ∗ e0∗ j ∧ ρ (∂j θ) i

j

for all C 1 covector fields θ in D2 . But this follows from step (ii) of the proof. (iv) From the hypothesis it follows that x 7→ ρρ−1 (x) , x 7→ |Jρ (ρ−1 (x))|, and x 7→ F (ρ−1 (x)) are C 1 . Therefore the product rule shows that ρ˜∗ F is C 1 . Let Θ : D2 → ∧V2∗ be any compactly supported smooth multicovector field. Then Propositions 7.1.7(ii) and 7.2.5(ii) and step (iii) above show that Z Z Z hΘ, ∇ y ρ˜∗ F idx = − h∇ ∧ Θ, ρ˜∗ F idx = − hρ∗ (∇ ∧ Θ), F idy D2 D2 D1 Z Z Z =− h∇ ∧ ρ∗ Θ, F idy = hρ∗ Θ, ∇ y F idy = hΘ, ρ˜∗ (∇ y F )idx. D1

D1

D2

Since Θ is arbitrary, we must have ∇ y (˜ ρ∗ F ) = ρ˜∗ (∇ y F ).

Example 7.2.11 (Orthogonal curvilinear coordinates). Let ρ : R3 → X be curvilinear coordinates in three-dimensional Euclidean space X. Important examples

7.2. Pullbacks and Pushforwards

223

treated in the standard vector calculus curriculum are spherical and cylindrical coordinates. The pushforward of the standard basis vector fields are e˜i := ρ∗ ei = ∂yi ρ(y),

i = 1, 2, 3,

ei } is in general not where {ei } denotes the standard basis in R3 . The frame {˜ an ON-frame in X, but in important examples such as the two mentioned above, these frame vector fields are orthogonal at each point. Assuming that we have such orthogonal curvilinear coordinates, we define hi (y) := |ρ∗ ei | and ei := e˜i /hi , for y ∈ R3 . This gives us an ON-frame {ei (y)} in X. We now show how the well-known formulas for gradient, divergence, and curl follow from Theorem 7.2.9. Note that 0 h1 0 ρ = ρ∗ = 0 h2 0 0 0 h3 with respect to the ON-bases {ei } and {ei }. For the gradient, we have X X ∇u = (ρ∗ )−1 grad(ρ∗ u) = (ρ∗ )−1 (∂i u)ei = h−1 i (∂i u)ei . i

i

∗

Note that ρ acts on scalar functions just by changing variables, whereas ρ∗ acts on vectors by the above matrix. P For the curl of a vector field F = i Fi ei in X, we similarly obtain X ∇ ∧ F = (ρ∗ )−1 curl(ρ∗ F ) = (ρ∗ )−1 ∇ ∧ hi Fi ei i ∗ −1

= (ρ )

XX j

∂j (hi Fi )ej

∧ ei

=

i

h1 e1 1 h2 e2 = h1 h2 h3 h3 e3

XX (hi hj )−1 ∂j (hi Fi )ej j

∂1 ∂2 ∂3

∧

ei

i

h1 F1 h2 F2 . h3 F3

Note that ρ∗ acts on ∧2 V by two-by-two subdeterminants of the above matrix as in Example 2.3.4. P For the divergence of a vector field F = i Fi ei in X, we use instead the normalized pushforward to obtain X X ∇ y F = ρ˜∗ div(˜ ρ∗ )−1 F = ρ˜∗ ∇ y h1 h2 h3 h−1 F e = ρ˜∗ ∂i (h1 h2 h3 h−1 i i i i Fi ) i

i

= (h1 h2 h3 )−1 ∂1 (h2 h3 F1 ) + ∂2 (h1 h3 F2 ) + ∂3 (h1 h2 F3 ) . Note that ρ˜∗ = ρ∗ /(h1 h2 h3 ) and that ρ∗ acts on vectors by the above matrix and simply by change of variables on scalars.

Chapter 7. Multivector Calculus

224

Example 7.2.12 (Pullback of Laplace equation). To see how the Laplace operator ∆ transforms under a change of variables, let ρ : D1 → D2 be a C 2 -diffeomorphism of Euclidean domains and let u : D2 → R be harmonic, that is, ∆u = 0. Changing variables, u corresponds to a function v(y) = u(ρ(y)) = ρ∗ u(y) in D1 . According to the commutation theorem (Theorem 7.2.9), we have −1 ρ∗ )−1 (ρ∗ )−1 (∇v) v) = ∇ y (ρ∗ )−1 (∇v) = ρ˜∗ ∇ y (˜ 0 = ∇ y ∇ ∧ (ρ∗ = ρ˜∗ ∇ y (ρ∗ ρ˜∗ )−1 (∇v) = ρ˜∗ div(A−1 grad v). Since ρ˜∗ is invertible, we see that the Laplace equation transforms into the divergence-form equation div(A−1 grad v) = 0. Here the linear map A(y), for fixed y ∈ D1 , in an ON-basis {ei } has matrix elements Ai,j (y) = hei , ρ∗ ρ˜∗ ej i = |Jρ (y)|−1 hρy (ei ), ρy (ej )i = g(y)−1/2 gi,j (y), where gi,j is the metric in D1 representing the Euclidean metric in D2 and g(x) := 2 det(gi,j (x)) = det(ρ∗ ρP ∗ ) = |Jρ (x)| is the determinant of the metric matrix. Thus 2 the Laplace equation i ∂i u = 0 transforms to the divergence-form equation X √ ∂i ( g g i,j ∂j v) = 0, i,j

where g i,j denotes the inverse of the metric matrix. Example 7.2.12 is a special case of the following pullback formulas for exterior and interior derivatives. Proposition 7.2.13 (Pullback of interior derivatives). Let X1 , X2 be oriented Euclidean spaces, let ρ : D1 → D2 be a C 2 -diffeomorphism, and denote by G(y) : ∧V1 → ∧V1 , y ∈ D1 , the metric of D2 pulled back to D1 , that is, hGw1 , w2 i = hρ∗ w1 , ρ∗ w2 i for multivector fields w1 , w2 in D1 . Write g(y) = det G|∧1 V1 (y) = |Jρ (y)|2 . Then we have pullback formulas ρ∗ (dF ) = d(ρ∗ F ), ρ∗ (δF ) = (g −1/2 G)δ(g 1/2 G−1 )(ρ∗ F ).

7.3

Integration of Forms

In this section, we develop integration theory for forms over k-surfaces. To avoid technicalities concerning bundles, at this stage we limit ourselves to k-surfaces in affine spaces, but the integration theory we develop generalizes with minor changes to general manifolds.

7.3. Integration of Forms

225

Definition 7.3.1 (k-Form). A k-form defined on a subset D of an affine space (X, V ) is a function b k V → L, Θ:D×∧ with range in a finite-dimensional linear space L, such that Θ(x, λw) = λΘ(x, w),

k

b V, λ > 0. x ∈ M, w ∈ ∧

We say that Θ is a linear k-form if for each x ∈ D, w 7→ Θ(x, w) extends to a linear function of w ∈ ∧k V . The idea with k-forms is that in integrating at a point x ∈ D, the integrand also depends on the orientation at x ∈ M of the k-surface M that we integrate over. Definition 7.3.2 (Integral of form). Let M be a compact oriented C 1 -regular kb k V → L be a continuous surface in an affine space (X, V ), and let Θ : M × ∧ k-form. We define the integral of Θ over M to be Z Θ(x, dx) := M

XZ α∈I

Dα

ηα (µα (y))Θ(µα (y), µα y (e1 ∧ · · · ∧ ek ))dy,

where {ei } is the standard basis in Rk and dy is standard Lebesgue measure in Rk ⊃ Dα . Here {µα }α∈I is the atlas of M , and {ηα }α∈I denotes a partition of unity for M . Note as in Section 2.4 how the induced action of the derivative µα y : ∧k Rk → ∧k V maps the oriented volume element e1 ∧ · · · ∧ ek dy in Rk to the oriented volume element dx = µα y (e1 ∧ · · · ∧ ek dy) = µα y (e1 ∧ · · · ∧ ek )dy on M . Note that this infinitesimal k-vector is simple, and hence Θ in general, needs to be defined only on the Grassmann cone. However, it is only to linear forms that Stokes’s theorem applies, as we shall see. The following proposition shows that such integrals do not depend on the precise choice of atlas and partition of unity for M , but only on the orientation of M. Proposition 7.3.3 (Independence of atlas). Consider a compact oriented C 1 -regular k-surface M , with atlas {µα : Dα → Mα }α∈I , in an affine space (X, V ). Let {µβ : Dβ → Mβ }β∈I 0 , I 0 ∩ I = ∅, be a second atlas for M such that all transition maps between Dα and Dβ , α ∈ I, β ∈ I 0 , are C 1 -regular and orientation-preserving. Further assume that {ηα }α∈I is a partition of unity for {µα }α∈I and that {ηβ }β∈I 0

Chapter 7. Multivector Calculus

226

is a partition of unity for {µβ }β∈I 0 . Then for every continuous k-form Θ, we have XZ ηα (µα (y))Θ(µα (y), µα y (e1 ∧ · · · ∧ ek ))dy α∈I

=

Dα

XZ β∈I 0

ηβ (µβ (z))Θ(µβ (z), µβ (e1 ∧ · · · ∧ ek ))dz. z

Dβ

P Proof. Inserting 1 = β∈I 0 ηβ in the integral on the left-hand side and 1 = P α∈I ηα in the integral on the right-hand side, it suffices to show that Z Z Θαβ (µα (y), µα y (e1 ∧ · · · ∧ ek ))dy = Θαβ (µβ (z), µβ (e1 ∧ · · · ∧ ek ))dz, Dβα

Dαβ

z

where Θαβ (x, w) := ηα (x)ηβ (x)Θ(x, w), since supp ηα ηβ ⊂ Mα ∩ Mβ . Changing variables z = µβα (y) in the integral on the right-hand side, we get Z Θαβ (µβ (µβα (y)), µβ (e1 ∧ · · · ∧ ek ))Jµβα (y)dy. µβα (y)

Dβα

Since µβ ◦ µβα = µα , and therefore µβ (e1 ∧ · · · ∧ ek )Jµβα = µα (e1 ∧ · · · ∧ ek ), the stated formula follows from the homogeneity of w 7→ Θαβ (x, w). Example 7.3.4 (Oriented and scalar measure). The simplest linear k-form in an affine space is Θ(x, w) = w ∈ L := ∧k V. R R In this case, M Θ(x, dx) = M dx = ∧k (M ) is the oriented measure of M discussed in Section 2.4. In a Euclidean space, given a continuous function f : M → L, the integral of the k-form Θ(x, w) := f (x)|w| R is seen to be the standard surface integral M f (x)dx, where dx = |dx|. Note that these are R not linear k-forms. In particular, f = 1 yields the usual (scalar) measure |M | := M dx of M . Using that dx = |dx| and the usual triangle inequality for integrals, we obtain from the definitions that oriented and scalar measure satisfy the triangle inequality Z Z dx ≤ dx. M

M

We continue with Example 2.4.4, where we calculated the oriented area element dx = (e12 + 2y2 e13 − 2y1 e23 )dy1 dy2 . Hence dx = |dx| =

q 1 + 4y12 + 4y22 dy1 dy2 ,

7.3. Integration of Forms

227

giving an area of the paraboloid equal to Z Z q 1 + 4y12 + 4y22 dy1 dy2 = 2π |y| 0. In the limit of small velocities v, the Navier– Stokes equations reduce to the linear Stokes equations ( ∇p − µ∆v = f, div v = 0. We are given the external forces f , and we want to compute the velcity vector field v and the scalar pressure p. The equations express conservation of momentum and the incompressibility of the fluid respectively. Assuming that the flow takes place inside a domain D, it is natural to demand in particular the boundary condition that the velocity vector field v is tangential at ∂D. We claim that the tangential Hodge decomposition Theorem 7.6.6 contains the solvability of this boundary value problem. To see this, we recall that with Clifford algebra we have a2 v = a y (a ∧ v) + aha, vi for vectors a and v. Replacing a by ∇, we obtain by incompressibility that ∆v = ∇ y (∇ ∧ v). Given a force vector field f ∈ L2 (D; ∧1 ), we apply the tangential Hodge decomposition (7.9) and conclude when f is orthogonal to Ck (D) that there exist a unique scalar function p ∈ R(δ; ∧0 ) and a tangential and exact bivector field ω ∈ R(d; ∧2 ) such that ∇p − µ∇ y ω = f. Again as in Corollary 7.6.8, there further exists a unique tangential vector field v ∈ R(δ; ∧1 ) such that ω = ∇ ∧ v. Note in particular that v is divergence-free and tangential at ∂D, and that it is uniquely determined if we demand that it be orthogonal to Ck (D; ∧1 ). The pressure p is unique modulo Ck (D; ∧0 ). The curl ω = ∇ ∧ v is the vorticity of the flow, which in three-dimensional space can be represented by the Hodge dual vector field.

7.7. Comments and References

7.7

253

Comments and References

7.1 The standard terminology for k-covector fields is differential forms. The anal´ Cartan, and they ysis of such differential forms was first developed by Elie have become a standard tool in modern differential geometry. The notation d and δ for the exterior and interior derivatives, when considered as partial differential operators, is standard in the literature. Beware, though, that it is standard to include an implicit minus sign in the notation δ, so that δF = −∇ y F and δ = d∗ . In the literature it is also common to write df also for the total derivative f = ∇ ⊗ f . We reserve the notation df for the exterior derivative. The nabla notations ∇ · F and ∇ × F are common in vector calculus. The generalizations ∇ ∧ F , ∇ y F , ∇ ⊗ F of this notation, used in this book, are inspired by Hestenes and Sobczyk [57] and Hestenes [56]. Nilpotence of an object a usually means that ak = 0 for some positive integer k. In this book, we consider only nilpotence of order 2. 7.2 The normalized pushforward of multivector fields is rarely found in the literature. In the special case of vector fields, this operation goes under the name Piola transformation in continuum mechanics. Pullbacks and pushforwards by smooth maps act in a natural way on multivector-valued distributions. For maps that are not diffeomorphisms, the pullback acts on test functions and the normalized pushforward act on distributions. Using k-covector fields as test functions, following George de Rham one defines a k-current as a k-vector-valued distribution. A classical reference for the applications of currents to geometric measure theory is Federer [38]. 7.3 The classical Stokes theorem from Example 7.3.10 dates back to Lord Kelvin and George Stokes in 1850. The general higher-dimensional result (7.5) was ´ Cartan in 1945. Standard notation with differential forms formulated by Elie ω is Z Z dω = ω, M

∂M

where the oriented measure dx is implicit in the notation ω. The trivial extension to forms with values in a finite-dimensional linear space, as presented in Theorem 7.3.9, is rarely found in the literature. The numerous applications in this book, though, show the usefulness of having such a generalized Stokes formula ready to use. 7.4–7.5 The identity L∗v Θ = ∇ ∧ (v y Θ) + v y (∇ ∧ Θ) is often referred to as Cartan’s magic formula due to its usefulness. A reference for the material in these two sections is Taylor [91], where results on ordinary differential equations, flows of vector fields, and the proof of Poincar´e’s theorem presented here can be found.

254

Chapter 7. Multivector Calculus

7.6 The theory of Hodge decompositions was developed by William Vallance Douglas Hodge in the 1930s, as a method for studying the cohomology of smooth manifold using PDEs. For further references for the presentation given here, which builds on the survey paper by Axelsson and McIntosh [14], we refer to Section 10.7. We find it convenient to work in the full exterior algebra ∧V . However, since the exterior and interior products preserve homogeneity of multivector fields, we may rather state the results at each level ∧k V to obtain Hodge splittings of k-vector fields, which is standard in the literature. A reference for solving boundary value problems with Hodge decompositions, also on more general manifolds with boundary, but under the assumption that the boundary is smooth, can be found in Schwarz [85]. Two harmonic analysis works using multivector calculus, motivated by Example 7.6.14 and nonsmooth boundary value problems for the Stokes equations, are McIntosh and Monniaux [68] and Ros´en [83]. Techniques from multivector calculus have also been used successfully in numerical analysis. A seminal paper on finite element exterior calculus is Arnold, Falk, and Winther [2].

Chapter 8

Hypercomplex Analysis Prerequisites: A solid background in analysis of one complex variable is required for this chapter, but no knowledge of analysis in several complex variables is needed. We make use of real-variable calculus, and build some on Chapter 7. Road map: We saw in Chapter 3 that even though it is natural to view Clifford algebras as a kind of hypercomplex numbers, the analogy fails in some important aspects, and it may be more appropriate to view Clifford algebras as matrix algebras, but from a geometric point of view. Nevertheless, a great deal of one-dimensional complex analysis does generalize to a noncommutative hypercomplex analysis in n-dimensional Euclidean space, replacing complex-valued functions by multivector fields. This yields a generalization of one-variable complex analysis that is fundamentally different from the commutative theory of several complex variables. Recall from one-dimensional complex analysis the following equivalent characterizations of analytic/holomorphic functions f : C → C defined in a domain D ⊂ C, where we assume that the total derivative f z is injective. The analysis definition: f is analytic if the limit f 0 (z) = limw→0 (f (z + w) − f (z))/w exists at each z ∈ D. The partial differential equation definition: f is analytic if it satisfies the Cauchy–Riemann system of partial differential equations ∂1 f1 (z) − ∂2 f2 (z) = 0,

∂2 f1 (z) + ∂1 f2 (z) = 0,

in D, where f1 , f2 are the real component functions of f = f1 + if2 . The algebra definition: f isP analytic if around each point z ∈ D it is locally the sum of a power ∞ series f (w) = k=0 ak (w − z)k , convergent in {w ∈ D ; |w − z| < r(z)} for some r(z) > 0. The geometry definition: f is analytic if it is an (orientation-preserving) © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_8

255

Chapter 8. Hypercomplex Analysis

256

conformal map, that is if at each z ∈ D the derivative f z is of the form a −b fz = , b a where a = ∂1 f1 and b = ∂1 f2 . This means that f z is a nonzero multiple of a rotation matrix and can be expressed as complex multiplication by f 0 (z). In generalizing to a hypercomplex analysis in higher-dimensional Euclidean spaces, the partial differential equation definition turns out to be most successful, where the Cauchy–Riemann equations are replaced by a Dirac equation ∇ 4 F (x) = 0, using the nabla operator induced by the Clifford product. As in Example 7.3.13, we may express this Dirac equation in terms of an integral difference quotient. Behind the Dirac equation, a fundamental type of splitting of function spaces is lurking: splittings into Hardy subspaces. With a solid understanding of Clifford algebra, these are straightforward generalizations of the classical such splittings in the complex plane. Recall that in the complex plane, any function f : ∂D → C on the boundary of a bounded domain D is in a unique way the sum f = f + + f −, where f + is the restriction to ∂D of an analytic function in D, and f − is the restriction to ∂D of an analytic function in C \ D that vanishes at ∞. The two subspaces consisting of traces of analytic functions from the interior or the exterior domain are the Hardy subspaces, and the Cauchy integral formula Z 1 f (w) dw 2πi ∂D w − z provides the projection operators onto these subspaces. There is one important difference with the Hodge splitting from Section 7.6: the two Hardy spaces are in general not orthogonal subspaces of L2 (∂D), but the angle between them depends on the geometry of ∂D. We show in Section 8.2 that the algebraic definition can be generalized to give power series expansions in higher dimensions. This is closely related to the classical theory of spherical harmonics. Later in Section 11.4, we shall see that the geometry definition does not generalize well. The higher dimensional conformal maps are very scarce indeed: the only ones are the M¨obius maps! Highlights: • The higher dimensional Cauchy integral formula: 8.1.8 • M¨ obius pullbacks of monogenic fields: 8.1.14 • Splitting into spherical monogenics: 8.2.6 • Spherical Dirac operator: 8.2.15 • Splittings into Hardy subspaces: 8.3.6

8.1. Monogenic Multivector Fields

257

8.1 Monogenic Multivector Fields In this chapter we work in Euclidean space (X, V ), and we study the following generalization of the Cauchy–Riemann equations. Definition 8.1.1 (Monogenic fields). Let D be an open set in a Euclidean space X. If F : D → 4V is differentiable at x ∈ D, we define the Clifford derivative ∇ 4 F (x) =

n X

ei∗ 4 ∂i F (x)

i=1

{e∗i }

is the basis dual to {ei }, and ∂i is the partial as in Definition 7.1.2, where derivative with respect to the corresponding coordinate xi . The Dirac operator D : F 7→ ∇ 4 F is the nabla operator induced by Clifford multiplication. If F is a C 1 multivector field for which ∇ 4 F (x) = 0 in all of D, then F is said to be a monogenic field in D. Let {es } be an induced ON-basis for 4V and write X F (x) = Fs (x)es . s

If F is a monogenic field, then each scalar component function Fs is a harmonic function. To see this, we note that X XX XX ei ej ∂i ∂j F (x) = ∂i2 F (x) = 0 = D2 F (x) = ∂i2 Fs (x) es i

j

i

s

i

X = (∆Fs (x))es . s

This is a consequence of the defining property v 2 = |v|2 for the Clifford product, and it means that D is a first-order differential operator that is a square root of the componentwise Laplace operator. Similar to the situation for analytic functions, a monogenic multivector field consists of 2n scalar harmonic functions, which are coupled in a certain sense described by the Dirac equation DF = 0. In particular, monogenic fields are smooth, even real analytic. The Dirac derivative further combines the exterior and interior derivative. Indeed, since e∗i 4 w = ei∗ y w + e∗i ∧ w, it is clear that DF (x) = ∇ 4 F (x) = ∇ y F (x) + ∇ ∧ F (x) = δF (x) + dF (x). This means that D2 = (d + δ)2 = d2 + δ 2 + dδ + δd = dδ + δd, by nilpotence. Another way to see this is to put v = θ = ∇ in the anticommutation relation from Theorem 2.8.1.

Chapter 8. Hypercomplex Analysis

258

As in Chapter 3, in using the Dirac operator, it is in general necessary to work within the full Clifford algebra, since typically DF will not be a homogeneous multivector field even if F is. However, in some applications the fields have values in the even subalgebra, or even are homogeneous k-vector fields. Example 8.1.2 (Analytic functions). Let X be a two-dimensional Euclidean plane, and let C = 40 V ⊕ 42 V be the standard geometric representation of the complex plane as in Section 3.2. Consider the Dirac equation ∇ 4 f = 0 for a complex valued function f = u + vj : C → C, where we have fixed an origin and ON-basis {e1 , e2 } in X = V , giving the identification V ↔ C, e1 ↔ 1, e2 ↔ j = e12 . Writing out the equation, we have ∇ 4 f = (e1 ∂1 + e2 ∂2 ) 4 (u + e12 v) = (∂1 u − ∂2 v)e1 + (∂1 v + ∂2 u)e2 . Thus ∇ 4 f = 0 coincides with the Cauchy–Riemann equations, and f is monogenic if and only if it is analytic. Note also that the only functions f : C → C that satisfy ∇ y f = 0 = ∇ ∧ f are the locally constant functions, since ∇ ∧ f = grad u and ∇ y f = −j grad v. On the other hand, the complex function f corresponds to the vector field F (x) = e1 f (x) under the identification V ↔ C. Reversing this relation gives F (x) = F (x) = f (x)e1 . Since the Clifford product is associative, it follows that F (x) is a plane divergence and curl-free vector field if and only if 0 = ∇ 4 F (x) = ∇ 4 (f (x) 4 e1 ) = (∇ 4 f (x)) 4 e1 , that is, if f is antianalytic. Example 8.1.3 (3D monogenic fields). Let F, G be vector fields and let f, g be scalar functions defined in an open set D in three-dimensional oriented Euclidean space. Then the multivector field f (x) + F (x) + ∗G(x) + ∗g(x) is monogenic if and only if div F (x) = 0, ∇f (x) − ∇ × G(x) = 0, ∇ × F (x) + ∇g(x) = 0, div G(x) = 0. We note that there is no restriction to assume that a monogenic field F takes values in the even subalgebra 4ev V . Indeed, if F : D → 4V is monogenic, we write F = F ev + F od where F ev : D → 4ev V and F od : D → 4od V . Then 0 = ∇ 4 F (x) = ∇ 4 F ev + ∇ 4 F od , where ∇ 4 F ev : D → 4od V and ∇ 4 F od : D → 4ev V , so we conclude that F ev and F od each are monogenic.

8.1. Monogenic Multivector Fields

259

Example 8.1.4 (Stein–Weiss vector fields). If F : D → V = 41 V is a vector field in a general Euclidean space, then F is monogenic if and only if F is divergenceand curl-free. Thus, for vector fields the equation DF = 0 is equivalent to the first-order system ( div F (x) = 0, curl F (x) = 0. This is a consequence of the fact that ∇ y F : D → 40 V and ∇ ∧ F : D → 42 V , where 40 V ∩ 42 V = {0}. We note that Example 8.1.4 generalizes as follows. When F : D → 4k V is a homogeneous multivector field, then ∇ 4 F = 0 if and only if ( ∇ y F (x) = 0, ∇ ∧ F (x) = 0. The following proposition shows when all the homogeneous parts of a monogenic field are themselves monogenic. Proposition 8.1.5 (Two-sided monogenicity). Let F : D → 4V be a C 1 multivector field, and write F = F0 + · · · + Fn , where Fj : D → 4j V . Then the following are equivalent. (i) All the homogeneous parts Fj are monogenic fields. (ii) The field F satisfies both ∇ ∧ F = 0 and ∇ y F = 0. (iii) The field F is two-sided monogenic, that is, ∇4F = 0 and F 4∇ = ei∗ = 0.

P

∂i F (x)4

Proof. (i) implies (iii): If Fj is monogenic, then ∇ ∧ Fj = 0 = ∇ y Fj as above, and therefore ∇ 4 Fj as well as Fj 4 ∇ = (−1)j (∇ ∧ Fj − ∇ y Fj ) is zero. Adding up all Fj proves (iii). (iii) implies (ii): this is a consequence of the Riesz formulas, which show that and ∇ y F = 12 (∇ 4 F + F\ 4 ∇) 4 ∇). ∇ ∧ F = 21 (∇ 4 F − F\ (ii) implies (i): If ∇ ∧ F = 0, then 0 = (∇ ∧ F )j+1 = ∇ ∧ Fj for all j, since d maps j-vector fields to (j + 1)-vector fields. Similarly ∇ y Fj = 0 for all j. Thus ∇ 4 Fj = ∇ y Fj + ∇ ∧ Fj = 0. We next consider the fundamental solution for the Dirac operator. In order to apply the Fourier transform componentwise as in Section 6.2, we complexify the Clifford algebra 4V ⊂ 4Vc . We note that the exterior, interior, and Clifford derivatives are the Fourier multipliers \ dF (x) = iξ ∧ Fˆ (ξ), \ δF (x) = iξ y Fˆ (ξ), \ DF (x) = iξ 4 Fˆ (ξ),

and ξ ∈ V.

Chapter 8. Hypercomplex Analysis

260

From this it follows that unlike d and δ, the Dirac operator D is elliptic and has a fundamental solution Ψ(x) with Fourier transform ξ ˆ = (iξ)−1 = −i 2 . Ψ(ξ) |ξ| Using the formula for the fundamental solution Φ to the Laplace operator ∆ ˆ from Example 6.3.1, where Φ(ξ) = −1/|ξ|2 , we obtain the following formula for Ψ(x) = ∇Φ. Note that unlike the situation for Φ, the two-dimensional case does not use any logarithm. Definition 8.1.6 (Fundamental solution). The fundamental solution to the Dirac operator D in an n-dimensional Euclidean space with origin fixed, n ≥ 1, is the vector field 1 x Ψ(x) := , σn−1 |x|n R where σn−1 := |x|=1 dx = 2π n/2 /Γ(n/2) is the measure of the unit sphere in V , R ∞ −t z−1 and Γ(z) := 0 e t dt is the gamma function, with Γ(k) = (k − 1)!. Exercise 8.1.7. Show by direct calculation that ∇ ∧ Ψ(x) = 0 and ∇ y Ψ(x) = δ0 (x) in the distributional sense in V , where δ0 (x) is the Dirac delta distribution. The following application of the general Stokes theorem is central to hypercomplex analysis. Theorem 8.1.8 (Cauchy–Pompeiu formula for D). Let D be a bounded C 1 -domain in Euclidean space X. If F ∈ C 1 (D; 4V ), then Z Z F (x) + Ψ(y − x)(DF )(y)dy = Ψ(y − x)ν(y)F (y) dy, D

∂D

for all x ∈ D, where ν(y) denotes the outward-pointing unit normal vector field on ∂D. In particular, for monogenic multivector fields F , we have the Cauchy reproducing formula Z F (x) = Ψ(y − x)ν(y)F (y) dy, x ∈ D. (8.1) ∂D

Proof. For fixed x ∈ D, consider the linear 1-form θ(y, v) := Ψ(y − x) 4 v 4 F (y). For y ∈ D \ {x}, its nabla derivative is θ(y, ˙ ∇) =

n X

∂yi Ψ(y − x) 4 ei 4 F (y)

i=1

= (Ψ(y˙ − x) 4 ∇) 4 F (y) + Ψ(y − x) 4 (∇ 4 F (y)) ˙ = Ψ(y − x) 4 (DF )(y),

8.1. Monogenic Multivector Fields

261

by associativity of the Clifford product and since Ψ 4 ∇ = ∇ y Ψ − ∇ ∧ Ψ = 0 by Exercise 8.1.7. To avoid using distribution theory, we consider the domain D := D \ B(x, ), obtained by removing a small ball around x. On ∂B(x, ) the outward-pointing unit normal relative D is (x − y)/|x − y|. The Stokes formula (7.4) gives Z Ψ(y − x)(DF )(y)dy D Z Z x−y = Ψ(y − x) Ψ(y − x)ν(y)F (y) dy F (y) dy + |x − y| ∂B(x,) ∂D Z Z 1 y−x 1 =− F (y) dy + ν(y)F (y) dy. n−1 σ σ |y − x|n n−1 n−1 ∂B(x,) ∂D Upon taking limits → 0, the first term on the right-hand side will converge to −F (x), and the Cauchy–Pompeiu formula follows. Exercise 8.1.9 (Cauchy integral theorem). Apply Stokes’s theorem and prove the general Cauchy theorem Z G(y) 4 ν(y) 4 F (y) dy = 0 ∂D

for a left monogenic field F and a right monogenic G, that is,R ∇ 4 F = 0 = G(x) ˙ 4∇ = 0, in D. Deduce from this the classical Cauchy theorem ∂D f (w)dw = 0 for an analytic function f from complex analysis. See Example 7.3.12. Example 8.1.10 (Cauchy formula in C). The Cauchy formula for analytic functions in the complex plane is a special case of Theorem 8.1.8. To see this, consider an analytic function f (z) in a plane domain D. As in Example 8.1.2, we identify the vector x ∈ V and the complex number z = e1 x ∈ C = 4ev V , y ∈ V with w = e1 y ∈ C, and the normal vector ν with the complex number n = e1 ν. If f (z) : D → C = 4ev V is analytic, thus monogenic, and if x ∈ D, then Theorem 8.1.8 shows that Z 1 e1 (w − z) f (z) = (e1 n(w))f (w)|dw| σ1 ∂D |w − z|2 Z Z e1 (w − z)e1 1 f (w)dw 1 = = f (w)(jn(w)|dw|) . 2 2πj ∂D |w − z| 2πj ∂D w − z Here we have used that e1 (w − z)e1 = w − z, that complex numbers commute, and that jn is tangent to a positively oriented curve. We have written |dw| for the scalar length measure on ∂D. Note that unlike the situation for analytic functions, in higher dimensions the normal vector must be placed in the middle, between the fundamental solution and the monogenic field. This is because the Clifford product is noncommutative. For

Chapter 8. Hypercomplex Analysis

262

analytic functions, the normal infinitesimal element ν(y)dy corresponds to dw/j, which can be placed, for example, at the end of the expression, since complex numbers commute. As in the complex plane, the Cauchy formula for monogenic fields has a number of important corollaries, of which we next consider a few. Corollary 8.1.11 (Smoothness). Let F : D → 4V be a monogenic field in a domain D. Then F is real analytic, and in particular a C ∞ -regular field. Proof. Fix a ball B(x0 , ) such that B(x0 , 2) ⊂ D. Then Z 1 y−x F (x) = ν(y)F (y) dy, for all x ∈ B(x0 , ). σn−1 |y−x0 |= |y − x|n The stated regularity now follows from that of the fundamental solution x 7→ y−x |y−x|n . We also obtain a Liouville theorem for entire monogenic fields. Corollary 8.1.12 (Liouville). Let F : X → 4V be an entire monogenic field, that is monogenic on the whole Euclidean space X. If F is bounded, then F is a constant field. Proof. Let x0 ∈ X. For all R > 0 and 1 ≤ k ≤ n we have Z 1 y − x ∂xk ν(y)F (y) dy. ∂k F (0) = σn−1 |y−x0 |=R |y − x|n x=0 If F is bounded, the triangle inequality for integrals shows that Z |∂k F (0)| . R−n dy . 1/R. |y|=R

Taking limits as R → 0 shows that ∂k F (x0 ) = 0 for all k. Since x0 was arbitrary, F must be constant. Next we consider what further properties monogenic fields do and do not share with analytic functions. In contrast to analytic functions, monogenic fields do not form a multiplication algebra, that is, F (x) and G(x) being monogenic in D does not imply that x 7→ F (x) 4 G(x) is monogenic. The obstacle here is the noncommutativity of the Clifford product, which causes D(F G) = (DF )G + DF G˙ 6= (DF )G + F (DG). Although monogenic fields in general cannot be multiplied to form another monogenic field, we can do somewhat better than the real linear structure of monogenic fields. Recall that analytic functions form a complex linear space. This generalizes to monogenic fields as follows.

8.1. Monogenic Multivector Fields

263

Proposition 8.1.13 (Right Clifford module). Let D be an open set in a Euclidean space X. Then the monogenic fields in D form a right Clifford module, that is, if F (x) is monogenic, then so is x 7→ F (x) 4 w for every constant w ∈ 4V . Proof. This is a consequence of the associativity of the Clifford product, since ∇ 4 (F (x) 4 w) = (∇ 4 F (x)) 4 w = 0 4 w = 0.

In contrast to analytic functions, monogenic functions do not form a group under composition, that is, F (y) and G(x) being monogenic in appropriate domains does not imply that x 7→ F (G(x)) is monogenic. Indeed, in general the composition is not even well defined, since the range space 4V is not contained in V . Although it does not make sense to compose monogenic fields, the situation is not that different in higher dimensions. Recall that in the complex plane, analytic functions are the same as conformal maps, at least for functions with invertible derivative. In higher dimensions one should generalize so that the inner function G is conformal and the outer function F is monogenic. In this way, we can do the following type of conformal change of variables that preserves monogenicity. Sections 4.5 and 11.4 are relevant here. Proposition 8.1.14 (Conformal Kelvin transform). Let T x = (ax + b)(cx + d)−1 be a fractional linear map of a Euclidean space X = V , and let D ⊂ X be an open set such that ∞ ∈ / T (D). For a field F : T (D) → 4V , define a pulled back field KTm F : D → 4V : x 7→ Then D(KTm F )(x) =

cx + d |cx + d|n

4

F (T (x)).

det4 (T ) m K (DF )(x), |cx + d|2 T

x ∈ D,

where det4 (T ) := ad − bc. In particular, if F is monogenic, then so is KTm F . Proof. Applying the product rule as in Example 7.1.9 shows that ! n X x + c−1 d c cx + d ∇ 4 (KT F )(x) = ∇ + ei F (T (x)) ∂i (F (T (x))). |x + c−1 d|n |c|n |cx + d|n i=1 The first term is zero, since the fundamental solution is monogenic outside the origin. For the second term, we note that since T is conformal, the derivative T x will map the ON-basis {ei } onto a basis {e0i = T x ei } of orthogonal vectors of equal length. By Exercise 4.5.18, we have e0i = (det4 (T )/cx + d)ei (cx + d)−1 . The dual basis is seen to be e0∗ i = ((cx + d)/det4 (T ))ei cx + d,

Chapter 8. Hypercomplex Analysis

264 so that

ei cx + d = (det4 (T )/(cx + d))e0∗ i . According to the chain rule, the directional derivatives are ∂ei (F (T (x))) = (∂e0i F )(T (x)), so ∇ 4 (KTm F )(x) =

n det4 (T ) cx + d X 0∗ det4 (T ) m e (∂e0i F )(T (x)) = K (DF )(x). |cx + d|2 |cx + d|n i=1 i |cx + d|2 T

Specializing to the inversion change of variables T x = 1/x, we make the following definition. Definition 8.1.15 (Kelvin transform). The monogenic Kelvin transform of a field F : D → 4V is the field K m F (x) :=

x |x|n

4

F (1/x).

Similarly, using the fundamental solution for the Laplace operator, we define the harmonic Kelvin transform of a function u : D → R to be K h u(x) := |x|2−n u(1/x). For the monogenic Kelvin transform, we have shown that DK m = −|x|−2 K m D. We now use this to obtain a similar result for the harmonic Kelvin transform. Proposition 8.1.16. The harmonic Kelvin transform satisfies the commutation relation ∆(K h u)(x) = |x|−4 K h (∆u)(x). In particular, the Kelvin transform of an harmonic function is harmonic. Proof. We note that ∆ = D2 and K h u = xK m u. Thus ∆K h u = DD(xK m u) = D(nK m u + (2∂x − xD)K m u) = nDK m u + 2(DK m u + ∂x DK m u) − nDK m u − (2∂x − xD)DK m u = 2DK m u + xD2 K m u = −2|x|−2 K m Du − xD|x|−2 K m Du = −x−1 DK m Du = x−1 |x|−2 K m D2 u = |x|−4 K h ∆u. Pn Here ∂x f = j=1 xj ∂j f = hx, ∇if denotes the radial directional derivative, and Pn we have used that D∂x f = j=1 (∇xj )∂j f + ∂x Df = (1 + ∂x )Df by the product rule.

8.2. Spherical monogenics

265

Exercise 8.1.17. Consider the special case of Proposition 8.1.14 in which ρ(x) = T x = qbxq −1 is an isometry, where q ∈ Pin(V ). Investigate how the conformal Kelvin transform KTm F , the pullback ρ∗ F , the pushforward ρ∗−1 F and the normalized pushforward ρ˜∗−1 F are related. Show that all these four fields are monogenic whenever F is monogenic, and relate this result to Proposition 8.1.13.

8.2

Spherical monogenics

In our n-dimensional Euclidean space X = V with a fixed origin, we denote by S := {x ∈ X ; |x| = 1} the unit sphere. We generalize the well-known theory of Taylor series expansions of analytic functions in the plane. When n = 2, we know that a function analytic at 0 can be written as a convergent power series f (x + iy) =

∞ X

Pk (x, y),

x2 + y 2 < 2 ,

k=0

for some > 0, where Pk ∈ Pkm := {ak (x + iy)k ; ak ∈ C}. A harmonic function can be written in the same way if we allow terms Pk ∈ Pkh := {ak (x + iy)k + bk (x − iy)k ; ak , bk ∈ C}. Note that P0h and all Pkm are one-dimensional complex linear spaces, whereas Pkh are two-dimensional when k ≥ 1. The spaces Pkm and Pkh are subspaces of the space Pk of all polynomials of order k, which has dimension k + 1. A polynomial P ∈ Pk is in particular homogeneous of degree k in the sense that P (rx) = rk P (x),

for all r > 0, x ∈ R2 .

This shows that P is uniquely determined by its restriction to the unit circle |x| = 1 if the degree of homogeneity is known. In the power series for f , the term Pk describes the kth-order approximation of f around the origin. Next consider an n-dimensional space X, and the following generalization of the spaces above. Definition 8.2.1 (Spherical harmonics and monogenics). Let X = V be a Euclidean space, and let k ∈ N and s ∈ R. Define function spaces Pk := {P : X → 4V ; all component functions are homogeneous polynomials of degree k}, m Ps := {F : X \ {0} → 4V ; DF = 0, F (rx) = rs F (x), x 6= 0, r > 0}, Psh := {F : X \ {0} → 4V ; ∆F = 0, F (rx) = rs F (x), x 6= 0, r > 0}.

Chapter 8. Hypercomplex Analysis

266

Let Pks ⊂ Pk and Pssh ⊂ Psh be the subspaces of scalar functions F : X \ {0} → 40 V = R, and let Psem ⊂ Psm be the subspace of functions F : X \ {0} → 4ev V that take values in the even subalgebra. Denote by Psh (S) the space of restrictions of functions P ∈ Psh to the unit sphere S. Denote by Psm (S) the space of restrictions of functions P ∈ Psm to the unit sphere S. We refer to these functions as (multivector-valued) spherical harmonics and spherical monogenics respectively. Note that the spaces Pk and Psh essentially are spaces of scalar functions: Each function in these spaces has component functions that belong to the same space, since the conditions on the function do not involve any coupling between the component functions. Even if the definitions of Psm and Psh are quite liberal, these are essentially spaces of polynomials, as the following shows. Proposition 8.2.2. Let n := dim X. The monogenic space Psm contains nonzero functions only if s ∈ {. . . , −(n + 1), −n, −(n − 1), 0, 1, 2, . . .}. The harmonic space Psh contains nonzero functions only if s ∈ {. . . , −(n + 1), −n, −(n − 1), −(n − 2), 0, 1, 2, . . .}. If k ∈ N, then Pkm ⊂ Pkh ⊂ Pk . The Kelvin transforms give self-inverse one-to-one correspondences m , K m : Psm → P−(s+n−1)

h K h : Psh → P−(s+n−2) .

Proof. (i) First consider the monogenic spaces Psm . Apply the Cauchy formula (8.1) to P ∈ Psm in the domain D := B(0; 1) \ B(0; ) for fixed 0 < < 1. For x ∈ D, we have Z Z Ψ(y − x)yP (y)dy + Ψ(y − x)ν(y)P (y)dy. P (x) = S

|y|=

For fixed x 6= 0, the second integral is dominated by n−1 sup|y|= |P |. Letting → 0, this tends to zero if s > −(n − 1), and it follows that 0 is a removable singularity of P (x). If −(n − 1) < s < 0, Liouville’s Theorem 8.1.12 shows that P = 0. Furthermore, generalizing the proof of Liouville’s theorem by applying higher-order derivatives shows that if s ≥ 0, then P (x) must be a polynomial. Thus Psm 6= {0} m is bijective and self-inverse is straightonly if s ∈ N. That K m : Psm → P−(s+n−1) forward to verify. m (ii) Next consider the harmonic spaces Psh . If P ∈ Psh , then DP ∈ Ps−1 . If m s∈ / Z or −(n − 2) < s < 0, then (i) shows that DP = 0, so that P ∈ Ps . Again by (i), we conclude that P = 0. If s ∈ N, then the same argument shows that DP is a polynomial. Here we may assume that P is scalar-valued, so that DP = ∇P . h Integrating we find that P is a polynomial as well. That K h : Psh → P−(s+n−2) is bijective and self-inverse is straightforward to verify.

267

8.2. Spherical monogenics

We next examine the finite-dimensional linear spaces Pkm and Pkh for k ∈ N. m and As we have seen, this also gives information about the spaces P−(k+n−1) h P−(k+n−2) via the Kelvin transforms. Note that unlike the situation in the plane, there is a gap −(n − 1) < s < 0 and −(n − 2) < s < 0 respectively between the nonzero spaces, and that this gap grows with dimension. A polynomial P (x) ∈ Pk , can be written X X P (x) = Pαs xα es . s⊂n α∈Nk αk 1 Here we use multi-index notation xα = x(α1 ,...,αk ) := xα 1 · · · xk , and we shall write δi := (0, . . . , 0, 1, 0, . . . , 0), where 1 is the ith coordinate. We introduce an auxiliary inner product XX hP, Qip := α!Pαs Qαs , s

α

where α! = (α1 , . . . , αk )! := α1 ! · · · αk !. Proposition 8.2.3. With respect to the inner product h·, ·ip on Pk , we have orthogonal splittings Pk = Pkm ⊕ xPk−1 , Pk = Pkh ⊕ x2 Pk−2 , where xPk−1 := {x 4 P (x) ; P ∈ Pk−1 }, as well as m Pkh = Pkm ⊕ xPk−1 ,

k ≥ 1,

P0h = P0m .

Proof. (i) The key observation is that Pk → Pk−1 : P (x) 7→ ∇ 4 P (x)

and Pk−1 → Pk : P (x) 7→ x 4 P (x)

are adjoint maps with respect to h·, ·ip .P In fact, the inner product P is designed for this purpose. To see this, write P (x) = s,α Pα,s xα es and Q(x) = t,β Qβ,t xβ et . Pn P Then ∇ 4 P (x) = i=1 s,α Pα,s αi xα−δi (i, s)ei4s , so that X h∇ 4 P, Qip = Pα,s αi (i, s)Qβ,t hxα−δi ei4s , xβ et i i,s,α,t,β

=

X

Pα,s αi (i, s)Qα−δi ,i4s (α − δi )!

i,s,α

Pn P On the other hand, x 4 Q(x) = i=1 t,β Qβ,t xβ+δi (i, t)ei4t , so that X hP, x 4 Qip = Pα,s Qβ,t (i, t)hxα es , xβ+δi ei4t i i,s,α,t,β

=

X i,s,α

Pα,s Qα−δi ,i4s (i, i 4 s)α!.

Chapter 8. Hypercomplex Analysis

268

Since αi (α − δi )! = α! and (i, s) = (i, i 4 s), the duality follows. (ii) We note that Pkm = N(∇) and that xPk−1 = R(x). Since the maps are adjoint, these subspaces are orthogonal complements in Pk . Similarly, Pk = m Pkh ⊕x2 Pk−2 , since (∇2 )∗ = x2 . Finally, we consider the map Pkh → Pk−1 : P (x) 7→ ∇ 4 P (x). This is well defined, since ∇ 4 P is monogenic if P is harmonic. The m adjoint operator will be Pk−1 → Pkh : Q(x) 7→ x 4 Q(x), provided x 4 Q is harmonic whenever Q is monogenic. To verify that this is indeed the case, we calculate as in the proof of Proposition 8.1.16 that D2 (xQ) = D(nQ + (2∂x − xD)Q) = nDQ + 2(D + ∂x D)Q − (nDQ + (2∂x − D)DQ) = (2 + D)DQ. m This proves that Pkm = N(∇) is the orthogonal complement to xPk−1 = R(x) in h Pk .

Corollary 8.2.4 (Dimensions). Let X be an n-dimensional Euclidean space. Then k+n−1 , dim Pk = 2n dim Pks , dim Pks = n−1 s dim Pkem = 2n−1 (dim Pks − dim Pk−1 ), s , dim Pksh = dim Pks − dim Pk−2

dim Pkm = 2 dim Pkem ,

dim Pkh = 2n dim Pksh .

Proof. To find dim Pks , note that this is the number of monomials of degree k in n variables. The standard combinatorial argument is as follows. Choose n − 1 of the numbers 1, 2, 3, . . . , k + n − 1, ways. say 1 ≤ m1 < m2 < · · · < mn−1 ≤ k + n − 1. This can be done in k+n−1 n−1 Such choices {mi } are in one-to-one correspondence with monomials 3 −m2 −1 · · · xnk+n−1−mm−1 . x1m1 −1 x2m2 −m1 −1 xm 3

From Proposition 8.2.3 the remaining formulas follow.

Exercise 8.2.5 (Two and three dimensions). Let V be a two-dimensional Euclidean space. In this case dim Pksh = 2 = dim Pkem . Show that dim Pkem is a one-dimensional complex linear space with the geometric complex structure j = e12 ∈ 42 V . Find bases for these spaces using the complex powers z k = (x + jy)k . Identifying vectors and complex numbers as in Section 3.2, write the splitting m Pkh = Pkm ⊕ xPk−1 in complex notation. Let V be a three-dimensional Euclidean space. In this case, dim Pksh = 2k + 1 and dim Pkem = 4(k + 1). Find bases for the spherical harmonics Pksh and for the spherical monogenics Pkem . Note that Pkem is a right vector space over H of dimension k + 1.

8.2. Spherical monogenics

269

Recall from Fourier analysis that the trigonometric functions {eikθ }k∈Z , suitably normalized, form an ON-basis for L2 (S) on the unit circle S ⊂ C = R2 in the complex plane. Thus every f ∈ L2 (S) can be uniquely written ∞ X

f (eiθ ) =

ak eikθ .

k=−∞

For k ≥ 0, the function eikθ extends to the analytic function z k on the disk |z| < 1. For k < 0, the function eikθ extends to the analytic function z k on |z| > 1, which vanishes at ∞, or alternatively to the antianalytic and harmonic function z −k on |z| < 1. In higher dimensions, we have the following analogue. Theorem 8.2.6. Let S be the unit sphere in an n-dimensional Euclidean space. The subspaces Pkh (S), k = 0, 1, 2, . . ., of spherical harmonics are pairwise orthogonal with respect to the L2 (S) inner product Z hF, Gi := hF (x), G(x)idx. S m Moreover, within each Pkh (S), the two subspaces Pkm (S) and xPk−1 (S) are orm m thogonal, and xPk−1 (S) = P2−n−k (S). The Hilbert space L2 (S) splits into finitedimensional subspaces as

L2 (S) =

∞ M

Pkh (S) =

k=0

∞ M

−(n−1)

Pkm (S) ⊕

k=0

M

Pkm (S).

k=−∞

Proof. Let P ∈ Pkh (S) and Q ∈ Plh (S) with k 6= l. Green’s second theorem, as in Example 7.3.11, shows that Z Z h∂x P (x), Q(x)i − hP (x), ∂x Q(x)i dx = h∆P, Qi − hP, ∆Qi dx = 0. |x| 0. For the two spherical operators we have the following. Proposition 8.2.15. Let V be an n-dimensional Euclidean space, and consider the Hilbert space L2 (S) on the unit sphere S. Then DS defines a self-adjoint operator in L2 (S) with spectrum σ(DS ) = Z \ {−(n − 2), . . . , −1}. The spherical Laplace operator equals ∆S = DS (2 − n − DS ). h In the splitting into spherical harmonics, L2 (S) = ⊕∞ k=0 Pk (S), the spherical Laplace operator acts according to

∆S

∞ X k=0

∞ X fk = k(2 − n − k)fk , k=0

8.2. Spherical monogenics

275

whereas in the splitting into spherical monogenics, L2 (S) =

∞ M

−(n−1)

Pkm (S)

k=0

M

⊕

Pkm (S),

k=−∞

the spherical Dirac operator acts according to DS

∞ X k=0

−(n−1)

fk +

X

−(n−1) ∞ X X fk = kfk . kfk +

k=−∞

k=0

k=−∞

Proof. It remains to prove that ∆S = DS (2 − n − DS ). Using polar coordinates x = ry, y ∈ S, we note that D = r−1 yxD = r−1 y(∂x − DS ) = y∂r − r−1 yDS . Squaring this Euclidean Dirac operator, we get ∆ = D2 = (y∂r − r−1 yDS )2 = y∂r y∂r − y∂r r−1 yDS − r−1 yDS y∂r + r−1 yDS r−1 yDS = ∂r2 − ∂r r−1 DS − r−1 yDS y∂r + r−2 yDS yDS . Writing [A, B] = AB − BA for the commutator of operators, we have used that [∂r , y] = 0 and [DS , r] = 0. To simplify further, we compute that [∂r , r] = 1 and [∂x , DS ] = [∂x , ∂x − xD] = −[∂x , xD] = 0. Thus ∂r r−1 DS = −r−2 DS + r−1 DS ∂r , so that ∆ = ∂r2 − r−1 (DS + yDS y)∂r + r−2 (DS + yDS yDS ). Comparing this equation and (8.2), we see that n − 1 = −(DS + yDS y) and ∆S = DS + yDS yDS = DS + (1 − n − DS )DS = DS (2 − n − DS ), as claimed.

In three dimensions, it is standard to introduce spherical coordinates (r, θ, φ), and DS and ∆S can be expressed in terms of ∂θ and ∂φ . The classical expression for the spherical harmonics, obtained by separation of variables, is rk Pkm (cos θ)eimφ , where the Pkm (t) denote the associated Legendre polynomials, m = −k, . . . , −1, 0, 1, . . . , k. The optimal parametrization of the sphere S, though, uses stereographic projection, which is conformal and has only one singular point for the coordinate system.

Chapter 8. Hypercomplex Analysis

276

Proposition 8.2.16 (Stereographic projection of DS ). Fix an (n − 1)-dimensional subspace VS ⊂ V and a point p ∈ S orthogonal to VS , and consider the stereographic parametrization T : VS → S : y 7→ T (y) = (py + 1)(y − p)−1 , as in (4.4). The monogenic Kelvin transform associated to the stereographic projection T defines an isometry of Hilbert spaces 2(n−1)/2 KTm : L2 (S) → L2 (VS ), and the spherical Dirac operator corresponds to (KTm DS (KTm )−1 )G(y) = − 21 p (|y|2 + 1)Dy + (y − p) G(y),

y ∈ VS ,

where Dy denotes the Dirac operator in the Euclidean space VS . Proof. According to Proposition 8.1.14, the Kelvin transform y−p KTm F (y) = F ((py + 1)(y − p)−1 ) |y − p|n satisfies D(KTm F )(y) = −2|y − p|−2 KTm (DF )(y). From the definition of DS we get KTm (DS F ) = KTm (∂x F ) − KTm (xDF ), where KTm (xDF )(y) =

y−p (py + 1)(y − p)−1 (DF )(T (y)) |y − p|n

= (y − p)−1 (py + 1)KTm (DF )(y) = − 12 (y − p)(py + 1)D(KTm F )(y) = 12 (1 + |y|2 )pD(KTm F )(y). To rewrite KTm (∂x F ), we observe that the vertical derivative of the stereographic parametrization at y ∈ VS is T y (p) =

−2 2 (y − p)p(y − p)−1 = x 2 1 + |y| 1 + |y|2

(8.3)

according to Exercise 4.5.18. Thus the chain and product rules give y−p y − p 1 + |y|2 = (∂ F )(T (y)) ∂yn (F (T (y)) x |y − p|n |y − p|n 2 1 + |y|2 p m = ∂yn KT F (y) − F (T (y)) 2 |y − p|n 1 + |y|2 p = ∂yn KTm F (y) − (y − p)KTm F (y). 2 2 Here ∂yn is the partial derivative in the direction p. Since pD = ∂yn + pDy , we obtain the stated formula. To show that the stated map is a Hilbert space isometry, note that by (8.3) the Jacobian is JT (y) = (2/(1 + |y|2 ))n−1 , since T is conformal. Thus Z Z Z 2n−1 dy 2 2 n−1 |F (x)| dx = |F (T (y))| =2 |KTm F (y)|2 dy. (1 + |y|2 )n−1 S VS VS KTm (∂x F )(y) =

8.3. Hardy Space Splittings

277

8.3 Hardy Space Splittings Let D = D+ be a bounded Lipschitz domain in Euclidean space X, with boundary ∂D separating it from the exterior unbounded domain D− = X \ D. Let ν denote the unit normal vector field on ∂D pointing into D− . The main operator in this section is the principal value Cauchy integral Z Z Eh(x) := 2p.v. Ψ(y − x)ν(y)h(y) dy = 2 lim Ψ(y − x)ν(y)h(y) dy, →0

∂D

∂D\B(x;)

x ∈ ∂D, which appears when we let x ∈ ∂D, rather than x ∈ D, in the Cauchy reproducing formula from Theorem 8.1.8. Here we assume only suitable bounds on h, and in particular we do not assume that h is a restriction of a monogenic field. The factor 2 is a technicality that will ensure that E 2 = I. The singularity at y = x in the integral is of order |y − x|1−n on the (n − 1)-dimensional surface ∂D, which makes the definition and boundedness of E a nontrivial matter, and cancellations need to be taken into account. Due to the strong singularity at y = x, we also refer to E as the Cauchy singular integral. Ignoring these analytic problems for the moment, we first investigate by formal calculations how E is related to the two limits Z + E h(x) := lim Ψ(y − z)ν(y)h(y) dy, x ∈ ∂D, z∈D + ,z→x

and

∂D

Z

E − h(x) :=

lim −

z∈D ,z→x

Ψ(y − z)(−ν(y))h(y) dy,

x ∈ ∂D,

∂D

in the Cauchy reproducing formula (8.1) for D+ and D− respectively. Placing z = x infinitesimally close, but interior, to ∂D, we have for E + ! Z Z + E h(x) = lim Ψ(y − x)ν(y)h(x)dy + Ψ(y − x)ν(y)h(y)dy →0

=

1 2 h(x)

Σ0x

+

Σ1x

1 2 Eh(x).

where Σ0x := {y ∈ D− ; |y − x| = } and Σ1x := {y ∈ ∂D ; |y − x| > }. We have here approximated h(y) ≈ h(x), changed the integration surface from ∂D \ Σ1x to Σ0x using Stokes’s theorem, and used that Ψ(y − x)ν(y) = 1/(σn−1 n−1 ) on Σ0x in the first integral. Thus the first term h/2 appears when we integrate around the singularity y = x on an infinitesimal half-sphere. Since −ν is outward pointing from D− , a similar formal calculation indicates that E − h(x) = 12 h(x) − 12 Eh(x), and we deduce operator relations 1 2 (I 1 2 (I +

+ E) = E + , − E) = E − ,

E + E − = I, E + − E − = E.

278

Chapter 8. Hypercomplex Analysis Moreover, from Theorem 8.1.8 we conclude that E+E+ = E+, E−E− = E−,

since E ± h by definition is the restriction of a monogenic field to ∂D, no matter what h is. This shows that E + and E − are complementary projection operators. For a suitable space of multivector fields H on ∂D, these projections define a splitting H = E + H ⊕ E − H. This means that any given field h on ∂D can be uniquely written as a sum h = h+ +h− , where h+ is the restriction to ∂D of a monogenic field in D+ and h− is the restriction to ∂D of a monogenic field in D− that decays at ∞. We refer to E ± H as Hardy subspaces, and to E ± as Hardy projections. Note also the structure of the Cauchy singular integral operator E = E + − E − : it reflects the exterior Hardy subspace E − H across the interior Hardy subspace E + H. In particular, E 2 = I, as claimed.

Figure 8.1: (a) The piecewise constant vector field h : ∂D → 41 R2 which equals e1 in the second quadrant and vanishes on the rest of the curve ∂D. (b) The Hardy splitting of h as the sum of two traces of divergence- and curl-free vector fields.

Example 8.3.1 (Constant-curvature boundaries). The most natural space for the singular integral operator E is H = L2 (∂D). In the simplest case, in which D is the upper complex half-plane, with ∂D the real axis, then E is a convolution singular integral, which under the Fourier transform corresponds to multiplication by ( 1, ξ > 0, sgn(ξ) = −1, ξ < 0,

8.3. Hardy Space Splittings

279

at least if h takes values in the even subalgebra and we use the geometric imaginary unit j as in Example 8.1.10. The second simplest example is that in which L∞ D is the unit ball |x| < 1 as in Theorem 8.2.6. In this case, E + projects onto k=0 Pkm (S), whereas E − projects L−(n−1) onto k=−∞ Pkm (S). In these examples the Hardy subspaces are orthogonal and kE ± k = 1. However, unlike Hodge splittings, the splitting into Hardy subspaces is not orthogonal for more general domains D. When ∂D has some smoothness beyond Lipschitz, Fourier methods apply to prove that E is a bounded operator on L2 (∂D), which geometrically means that the angle between the Hardy subspaces, although not straight, is always positive. A breakthrough in modern harmonic analysis was the discovery that this continues to hold for general Lipschitz domains. Theorem 8.3.2 (Coifman–McIntosh–Meyer). Let D be a bounded strongly Lipschitz domain. Then the principal value Cauchy integral Eh(x) of any h ∈ L2 (∂D) is well defined for almost every x ∈ ∂D, and we have bounds Z Z |h(x)|2 dx. |Eh(x)|2 dx . ∂D

∂D

This is a deep result that is beyond the scope of this book. There exist many different proofs. A singular integral proof is to estimate the matrix for E in a wavelet basis for L2 (∂D) adapted to ∂D. A spectral proof is to identify E ± as spectral projections of a Dirac-type operator on ∂D, generalizing the spherical Dirac operator DS from Definition 8.2.13. The problem is that for general domains this operator is no longer self-adjoint, but rather has spectrum in a double sector around R, and it becomes a nontrivial matter involving Carleson measures to estimate the spectral projections corresponding to the two sectors. See Section 8.4 for references and further comments. We remark only that from Theorem 8.3.2 one can prove that for h ∈ L2 (∂D), the Cauchy extensions Z + (8.4) Ψ(y − x)ν(y)h(y) dy, x ∈ D+ , F (x) := ∂D

and −

Z Ψ(y − x)(−ν(y))h(y) dy,

F (x) :=

x ∈ D− ,

(8.5)

∂D

have limits as x → ∂D both in an L2 (∂D) sense and pointwise almost everywhere, provided that we approach ∂D in a nontangential way. In the remainder of this section, we perform a rigorous analysis of the splitting of the space of H¨older continuous multivector fields C α (∂D) = C α (∂D; 4V ),

0 < α < 1,

from Example 6.4.1, into Hardy subspaces on a bounded C 1 surface ∂D. This setup is a good starting point for studying Hardy splittings that only requires

Chapter 8. Hypercomplex Analysis

280

straightforward estimates. We exclude the endpoint cases α = 0, continuous functions, and α = 1, Lipschitz continuous functions, for the reason that typically singular integral operators like E are not bounded on these spaces. It is also not bounded on L∞ (∂D), something that could be seen in Figure 8.1 if we zoom in at the discontinuities of h. Proposition 8.3.3 (Hardy projection bounds). Let D be a bounded C 1 domain and 0 < α < 1, and assume that h ∈ C α (∂D). Define the Cauchy extensions F ± in D± as in (8.4) and (8.5). Then F + is a monogenic field in D+ , and F − is a monogenic field in D− with decay F − = O(|x|−(n−1) ) at ∞. At the boundary ∂D, the traces f + (y) :=

lim

x∈D + ,x→y

F + (x),

f − (y) :=

lim

x∈D − ,x→y

F − (x),

y ∈ ∂D,

exist, with estimates kf + kα . khkα and kf − kα . khkα . In terms of operators, this means that the Hardy projections E ± : h 7→ f ± are bounded on C α (∂D). Proof. (i) That F + and F − are monogenic is a consequence of the associativity of the Clifford product. Indeed, applying the partial derivatives under the integral sign shows that Z + ∇x 4 Ψ(y − x) 4 ν(y) 4 h(y) dy ∇ 4 F (x) = Z∂D ∇x 4 (Ψ(y − x) 4 ν(y) 4 h(y)dy = 0, = ∂D

/ ∂D. The decay at infinity follows from the fact that ∂D and h are when x ∈ bounded and the decay of the fundamental solution Ψ. (ii) We next consider the boundary trace of F + . A similar argument applies to the trace of F − . Note that in order to estimate kf + kα , it suffices to estimate |f + (x) − f + (y)| . |x − y|α khkα for |x − y| ≤ δ, provided that |f + (x)| . khkα for all x ∈ ∂D, since ∂D is bounded. Thus we may localize to a neighborhood of a point p ∈ ∂D, in which we can assume that ∂D coincides with the graph of a C 1 -function φ. We choose a coordinate system {xi } so that p is the origin and ∂D is given by xn = φ(x0 ), where x0 = (x1 , . . . , xn−1 ), in the cylinder |x0 | < r, |xn | < s. Let δ < min(r, s) and consider a point x = (x0 , xn ) ∈ D ∩ B(0, δ). We claim that (8.6) |∂j F + (x)| . khkα (xn − φ(x0 ))α−1 , j = 1, . . . , n. To show this, considerR the vertical projection z = (x0 , φ(x0 )) of R x onto ∂D, and note that F + (x) − h(z) = ∂D Ψ(y − x)ν(y)(h(y) − h(z))dy, since Ψ(y − x)ν(y)dy = 1, according to the Cauchy formula. Thus differentiation with respect to x, with z fixed, gives Z |∂j Ψ(y − x)| |y − z|α dy = khkα (I + II). |∂j F + (x)| . khkα ∂D

8.3. Hardy Space Splittings

281

Here I denotes the part of the integral inside the cylinder, and II is the part outside. Since |∂j Ψ(y − x)| . 1/|y − x|n , the term II is bounded. For the integral I, we change variable from y = (y 0 , φ(y 0 )) =: ρ(y 0 ) ∈ ∂D to y 0 ∈ Rn−1 . To find the change of (n − 1)-volume, we calculate ρy0 (e1 ∧ · · · ∧ en−1 ) = (e1 + (∂1 φ)en ) ∧ · · · ∧ (en−1 + (∂n−1 φ)en ) = e1···(n−1) + (∂1 φ)en2···(n−1) + (∂2 φ)e1n3···(n−1) + · · · + (∂n−1 φ)e123···(n−2) , p the norm of which is 1 + |∇φ|2 . Since the function φ is C 1 , we conclude that |∂j Ψ(y − x)| ≈ 1/(|y 0 − x0 | + t)n , |y − z|α ≈ |y 0 − x0 |α , and dy ≈ dy 0 , where t = xn − φ(x0 ). Therefore Z Z ∞ 0 0 −n 0 0 α 0 I. (|y − x | + t) |y − x | dy . (r + t)−n rα rn−2 dr . tα−1 . |y 0 | φ(x0 ) and consider first the vertical limit f + (y). Since Z r Z r−φ(y0 ) + 0 + 0 0 + 0 |F (y , r) − F (y , φ(y ) + t)| ≤ |∂n F (y , s)|ds . khkα sα−1 ds, φ(y 0 )+t

0

it is clear that this limit exists, since the integral is convergent. Moreover, we get the estimate |f + (y)| . khkα , since |F + (y 0 , r)| is bounded by khkα . Next we aim to show that {F + (x)} converges when x → y from D+ in general, and not only along the vertical direction. Let x1 = (x01 , t1 ), x2 = (x02 , t2 ) ∈ D ∩ B(y; ), and define t := max(t1 , t2 ) + 2(1 + k∇φk∞ ). Then Z F + (x2 ) − F + (x1 ) = hdx, ∇iF + (x), γ

where γ is the piecewise straight line from x1 to x2 viaR(x01 , t) and (x2 , t). The first and last vertical line integrals are dominated by khkα 0 tα−1 dt as above, whereas in the middle horizontal line integral, the integrand is dominated by khkα α−1 . In total we obtain the estimate |F + (x2 ) − F + (x1 )| . khkα α , when x1 , x2 ∈ D ∩ B(y, ). This shows the existence of the limit f + (y) as x → y from D+ . By taking x1 , x2 ∈ ∂D, it also shows that kf + kα . khkα , which completes the proof. Proposition 8.3.4 (Sokhotski–Plemelj jumps). Let D be a bounded C 1 domain and 0 < α < 1. Then the Cauchy principal value integral E : C α (∂D) → C α (∂D) is a well-defined and bounded linear operator. The Hardy projections E ± equal Z ± 1 E h(x) = 2 h(x) ± p.v. Ψ(y − x)ν(y)h(y) dy, x ∈ ∂D. ∂D

In terms of operators, this means that E ± = 12 (I ± E).

Chapter 8. Hypercomplex Analysis

282

Proof. We start by verifying the identity E + h(x) = 21 h(x) + 12 Eh(x) for x ∈ ∂D. As in the proof of Proposition 8.3.3, write x = (x0 , φ(x0 )) in a coordinate system in a cylinder around x. If h ∈ C α (∂D), the integrand of Z Ψ(y − (x + ten ))ν(y) h(y) − h(x) dy ∂D

is seen to be bounded by |y − x|α−(n−1) , uniformly for 0 < t ≤ t0 . Here we view h(x) as a constant function. Letting t → 0+ and applying the Lebesgue dominated convergence theorem, it follows that Z E + h(x) − h(x) = E + (h − h(x))(x) = Ψ(y − x)ν(y) h(y) − h(x) dy ∂D ! Z Z Ψ(y − x)ν(y)h(y)dy −

= lim

→0

∂D\B(x;)

Ψ(y − x)ν(y)dy h(x).

lim

→0

∂D\B(x;)

The first equality follows from the fact that the Cauchy integral of the constant field h(x) is the constant field h(x) in D+ . It suffices to show that Z lim Ψ(y − x)ν(y)dy = 12 . →0

∂D\B(x;)

formula for the domain D+ \ B(x; ) shows that it suffices Applying the Cauchy R to prove lim→0 ∂B(x;)∩D+ Ψ(y − x)ν(y)dy = 12 . But Z Ψ(y − x)

lim

→0

∂B(x;)∩D +

y−x |∂B(x; ) ∩ D+ | , dy = lim →0 |y − x| |∂B(x; )|

(8.7)

and on approximating ∂D by its tangent hyperplane at x, this limit is seen to be 1/2, since ∂D is assumed to be C 1 regular at x. To summarize, we have shown that E + = 12 (1 + E), and Proposition 8.3.3 shows that E = 2E + − I is a bounded and well-defined operator. Letting t → 0− instead, we get −E − h(x) − 0 = Eh(x) − 12 h(x), which shows that E − = 12 (I − E). Exercise 8.3.5. Generalize Proposition 8.3.3 to bounded Lipschitz domains. Show that Proposition 8.3.4 fails for bounded Lipschitz domains. Summarizing the H¨older estimates in this section, we have the following main result. Theorem 8.3.6 (Hardy subspace splitting). Let D be a bounded C 1 domain and let 0 < α < 1. Then we have a splitting of the H¨ older space C α (∂D) into Hardy subspaces C α (∂D) = E + C α ⊕ E − C α .

8.4. Comments and References

283

The Hardy subspaces are the ranges of the Hardy projections E ± : C α (∂D) → C α (∂D), which are the spectral projections E ± = 12 (I ± E) of the Cauchy singular integral operator E. α The interior Hardy subspace E + C+ consists of all traces F + |∂D of monogenic + + fields F in D that are H¨ older continuous up to ∂D. The exterior Hardy subspace E − C α consists of all traces F − |∂D of monogenic fields F − in D− that are H¨ older continuous up to ∂D and have limit limx→∞ F − (x) = 0. In fact, all such F − have decay O(1/|x|n−1 ) as x → ∞. Proof. Proposition 8.3.3 shows that E ± : C α (∂D) → C α (∂D) are bounded projection operators. Proposition 8.3.4 shows in particular that they are complementary: E + + E − = I. This shows that C α (∂D) splits into the two Hardy subspaces. It is clear from the definition and Proposition 8.3.3 that the Hardy subspaces consist of traces of H¨older continuous monogenic fields F ± in D± respectively. The decay of F − at ∞ follows from that of Ψ. Conversely, the fact that the trace of every H¨ older continuous monogenic field F + in D+ belongs to E + C α follows from Theorem 8.1.8. For the corresponding result for D− , we apply the Cauchy − reproducing formula to the bounded domain DR := D− ∩ B(0; R) for large R. We have Z Z − F − (x) = − Ψ(y − x)ν(y)F − (y)dy + Ψ(y − x)ν(y)F − (y)dy, x ∈ DR . ∂D

|y|=R

Since |∂B(0; R)| grows like Rn−1 and Ψ(x − y) decays like 1/Rn−1 as R → ∞, the last integral will vanish if limx→∞ F − (x) = 0, showing that F − is the Cauchy integral of F − |∂D , so that F − |∂D ∈ E − C α .

8.4

Comments and References

8.1 The higher-dimensional complex analysis obtained from the Dirac equation and Clifford algebra has been developed since the 1980s. This research field is referred to as Clifford analysis. The pioneering work is Brackx, Delanghe, and Sommen [23]. Further references include Gilbert and Murray [42] and Delanghe, Sommen, and Soucek [33]. Div/curl systems like those in Example 8.1.4 have been used to define higher-dimensional harmonic conjugate functions in harmonic analysis. The seminal work is Stein and Weiss [89] 8.2 This material builds on the treatment by Axler, Bourdon, and Ramey [16] of spherical harmonics. We have generalized mutatis mutandis the theory for spherical harmonics to spherical monogenics. 8.3 The classical Lp -based Hardy spaces, named after G.H. Hardy, on the real axis or the unit circle in the complex plane where introduced by F. Riesz in

284

Chapter 8. Hypercomplex Analysis 1923. The function space topologies for p ≤ 1 that they provide are fundamental in modern harmonic analysis. Theorem 8.3.2 was proved by R. Coifman, A. McIntosh, and Y. Meyer in [28] for general Lipschitz graphs in the complex plane. Earlier, A. Calder´on had obtained a proof in the case of small Lipschitz constants. The higherdimensional result in Theorem 8.3.2 is equivalent to the L2 boundedness of the Riesz transforms on Lipschitz surfaces, and this was known already in [28] to follow from the one-dimensional case by a technique called Calder´on’s method of rotations. A direct proof using Clifford algebra is in [66]. From Calder´on–Zygmund theory, also Lp boundedness for 1 < p < ∞ follows. A reference for wavelet theory, which is intimitely related to Theorem 8.3.2, is Meyer [69]. It is interesting to note that just like induced bases {es } for multivectors, wavelet bases for function spaces also do not come with a linear order of the basis functions, but these are rather ordered as a tree. For Clifford algebras and wavelets, see Mitrea [71]. Unpublished lecture notes by the author containing the wavelet proof of Theorem 8.3.2 are [81]. The basic idea behind estimating singular integrals like the Cauchy integral using wavelets is simple: the matrices of such operators in a wavelet basis are almost diagonal in a certain sense. However, the nonlinear ordering of the basis elements and the details of the estimates make the proof rather technical. There is also a much deeper extension to higher dimensions of the result in [28] that was known as the Kato square root problem. It was finally solved affirmatively by Auscher, Hofmann, Lacey, McIntosh, and Tchamitchian [6] 40 years after it was formulated by Kato, and 20 years after the one-dimensional case [28] was solved. As McIntosh used to tell the story, the works on linear operators by T. Kato and J.-L. Lions closed that field of research in the 1960s; only one problem remained open, and that was the Kato square root problem. A reference for a spectral/functional calculus approach to Theorem 8.3.2 is Axelsson, Keith, and McIntosh [12]. See in particular [12, Consequence 3.6] for a proof of Theorem 8.3.2, and [12, Consequence 3.7] for a proof of the Kato square root problem. This paper illustrates well how Dirac operators and Hodge- and Hardy-type splittings can be used in modern research in harmonic analysis.

Chapter 9

Dirac Wave Equations Prerequisites: Some familiarity with electromagnetism and quantum mechanics is useful for Section 9.2. A background in partial differential equations, see Section 6.3, and boundary value problems is useful but not necessary for the later sections. For the operator theory that we use, the reader is referred to Section 6.4. Ideally, we would have liked to place Section 9.6 after Chapter 10. But since it belongs to the present chapter, we ask the reader to consult Chapter 10 for more on Hodge decompositions when needed. Road map: Acting with the nabla symbol through the Clifford product ∇4F (x) on multivector fields, or through a representation ∇.ψ(x) on spinor fields, in Euclidean space we obtain first-order partial differential operators which are square roots of the Laplace operator ∆. However, Paul Dirac first discovered his original equation in 1928 for spin-1/2 massive particles, in the spacetime setting as a square root of the Klein–Gordon equation, that is, the wave equation with a zero-order term ∂x2 ψ + ∂y2 ψ + ∂z2 ψ − c−2 ∂t2 ψ =

m2 c2 ψ. ~2

The resulting Dirac wave equation ~∇.ψ = mcψ describing the free evolution of the wave function for the particle, a spinor field ψ : W → 4W / in physical spacetime, has been described as one of the most successful and beautiful equations ever. For example, it predicted the existence of antiparticles some years before these were experimentally found in 1932. In Section 9.2 we survey Dirac’s equation, as well as Maxwell’s equations from the early 1860s, which describes the evolution of the electrical and magnetic fields. We show how, in a very geometric way, the electromagnetic field is a multivector field and that the Maxwell equations, when written in terms of Clifford algebra, form a Dirac wave equation. The four classical © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_9

285

286

Chapter 9. Dirac Wave Equations

equations correspond to the four spaces ∧j V of homogeneous multivectors in threedimensional Euclidean space V . Motivated by applications to Maxwell’s equations, Sections 9.3 to 9.7 develop a theory for boundary value problems (BVPs) for Dirac equations, and they show how it applies to electromagnetic scattering. We consider only time-harmonic waves at a fixed frequency. Our abstract setup for a BVP is to consider two splittings of a space H of functions on the boundary ∂D of the domain D. The first splitting, H = A+ H ⊕ A− H, encodes the differential equation and generalizes the splitting into Hardy subspaces A+ H = E + H and A− H = E − H from Theorem 8.3.6. The second splitting, H = B + H ⊕ B − H, encodes the boundary conditions. Typically the projections B ± are pointwise and determined by the normal vector ν. Relating them to the classical boundary value problems for the Laplace operator, in that case B + would encode Dirichlet boundary conditions and B − would encode Neumann boundary conditions. From this point of view of functional analysis, studying BVPs amounts to studying the geometry between these two different splittings. Well-posedness of BVPs will mean that the subspaces A± H do not intersect the subspaces B ± H, and in the optimal case, the two reflection operators A and B, where A generalizes the Cauchy principal value integral, anticommute. In Section 9.5, we formulate integral equations for solving scattering problems for Dirac equations. The aim is to find singular but not hypersingular integral operators that are both bounded and invertible also on Lipschitz boundaries, whenever the scattering problem considered is well posed. A problem that we need to overcome to find an integral equation which is numerically useful is that we cannot easily discretize spaces like the Hardy spaces E ± H, which are defined by a nonlocal integral constraint. We obtain good integral equations on good function spaces for solving BVPs for the Dirac equation. To apply these to scattering problems for electromagnetic waves, we show in Sections 9.6 and 9.7 how we require a third splitting of the boundary function space H: a boundary Hodge decomposition H = R(Γk ) ⊕ R(Γ∗k ), where the Maxwell fields live in the Hodge component R(Γk ). Embedding Maxwell’s equations into the Dirac equation, and solving the BVP with a Dirac integral equation, we give examples in Section 9.7 of how this algorithm performs numerically. Highlights: • Boosting E and B using Clifford algebra: 9.2.4

9.1. Wave and Spin Equations

287

• Discovery of antiparticles: 9.2.8 • Stratton–Chu as a Clifford–Cauchy integral: 9.3.8 • Well-posedness via operator Clifford algebra 4R2 : 9.4.5 • Rellich spectral sector vs. Lipschitz geometry of boundary: 9.5.1 • Spin integral equation: 9.5.5 • Maxwell fields and boundary Hodge decompositions: 9.7.1

9.1

Wave and Spin Equations

The Dirac operator on a Euclidean space from Definition 8.1.1 generalizes in the obvious way to an inner product space of arbitrary signature. Definition 9.1.1 (4-Dirac operator). Let (X, V ) be an inner product space. The 4-Dirac operator D = DV acting on multivector fields F : D → 4V defined on some domain D ⊂ V is the nabla operator (DF )(x) := ∇ 4 F (x) =

n X

e∗j 4 ∂j F (x)

j=1

induced by the Clifford product V × 4V → 4V : (v, w) → 7 v 4 w as in Definition 7.1.2. Here ∂j are partial derivatives with respect to coordinates in a basis {ej } for V , with dual basis {e∗j }. Since 4-Dirac operators are our main type of Dirac operator, we sometimes omit 4 in the notation. When V is a Euclidean space, we speak of a harmonic Dirac operator, while if V = W is a spacetime, we speak of a wave Dirac operator. For a Euclidean space, we know from Chapter 8 that D2 = ∆. We have seen that this means that multivector fields F solving DF = 0 in particular have scalar component functions Fs that are harmonic. Turning to the wave Dirac operator, we now have D2 = , where is the d’Alembertian from Section 6.3. Indeed, in an ON-basis {ei } we have D2 F = (−e0 ∂0 + e1 ∂1 + · · · + en ∂n )2 F = (−∂02 + ∂12 + · · · + ∂n2 )F. Similar to the Euclidean case, it follows that multivector fields solving the wave Dirac equation DF = 0 have scalar component functions Fs that solve the wave equation Fs = 0.

Chapter 9. Dirac Wave Equations

288

The wave Dirac operator describes a wave propagation of multivector fields. For the harmonic Dirac operator we saw in Chapter 8 how the fundamental solution Φ to the Laplace operator yielded a fundamental solution Ψ = ∇Φ to D. Similarly, the fundamental solution to the wave equation encoded by the Riemann operators Rt from Proposition 6.2.2 and Example 6.3.2 now yield solution formulas for the wave Dirac operator. Proposition 9.1.2 (Propagation of Dirac waves). Fix a time-like unit vector e0 in a spacetime W and let V = [e0 ]⊥ be the space-like complement. Consider the initial value problem for the wave Dirac equation DW F = G for given initial data F |V = f and source G. We assume that f and Gx0 (·) = G(x0 , ·), for each fixed time x0 , belong to L2 (V ; 4W ). Then the solution Fx0 (x) = F (x0 , x) is given by Z x0 F (x0 , x) = Mx0 F0 (x) + Mx0 −s (e0 Gs )(x)ds, x0 > 0, 0

where the Fourier multiplier Mx0 on L2 (V ; 4W ) is Mx0 g := (∂0 − e0 D)Rx0 g. Proof. We apply the partial Fourier transform to DW F = G in the x-variables, ˆ x (ξ), ξ ∈ V , for each fixed x0 . We obtain the ODE −e0 ∂0 Fˆx0 (ξ) + iξ Fˆx0 (ξ) = G 0 with solution Z x0 ˆ s (ξ)ds. exp(−ie0 ξ(x0 − s))e0 G Fˆx (ξ) = exp(−ie0 ξx0 )fˆ(ξ) + 0

0

We have exp(−ie0 ξx0 ) = cos(|ξ|x0 )−ie0 ξ sin(|ξ|x0 )/|ξ| according to Exercise 1.1.5, which is seen to be the symbol of Mx0 . The inverse Fourier transformation yields the stated formula for F . It follows that the time evolution of the wave Dirac equation is quite similar to that for the scalar second-order wave equation: with our scaling, the propagation speed is 1, and in odd dimensions there is a Huygens principle. However, although evolution backward in time is well posed, the wave Dirac equation is not symmetric in the time variable, unlike the scalar wave equation. Another difference is that the only inital datum that we need is F (0, ·), and no normal derivative. The second type of Dirac operator that we consider are the following spinDirac operators. / Dirac operator). Let (X, V ) be an inner product space, with Definition 9.1.3 (4/ V acting on spinor fields / =D / . The 4-Dirac operator D / complex spinor space 4V Ψ : X → 4V / is the nabla operator / := ∇.Ψ(x) = (DΨ)(x)

n X j=1

ej∗ .∂j Ψ(x),

9.1. Wave and Spin Equations

289

which is induced by the bilinear map V × 4V → 4V / / : (θ, ψ) 7→ θ.ψ as in Definition 7.1.2. Here ∂j are partial derivatives with respect to coordinates in a basis {ej } for V , with dual basis {ej∗ }. / operator, When V is a Euclidean space, we speak of a harmonic 4-Dirac while if V = W is a spacetime, we speak of a wave 4-Dirac operator. The 4/ / Dirac operators are best known for their representations as matrix first-order partial differential operators. Such expressions are straightforward to derive using representations of Clifford algebras. See Section 5.1. 2 / = ∆, Analogous to the 4-Dirac operator, for a Euclidean space, we have D 2 / = 0 have harmonic / = , and spinor fields solving DΨ while in spacetime D functions and solutions to the wave equation as component functions, respectively. Exercise 9.1.4 (Hypercomplex spin analysis). Show how the theory from Chapter 8 for the Cauchy integral, the monogenic Kelvin transform, and spherical monogen/ operator. ics generalizes in a natural way for solutions to the harmonic 4-Dirac Explain why such solutions do not form a right Clifford module as in Proposition 8.1.13 and why the notion of two-sided monogenicity from Proposition 8.1.5 does not generalize. Show how Proposition 9.1.2 generalize to describe the free wave evolution for equations. wave 4-Dirac / We next consider how the wave Dirac equations are related to the harmonic Dirac equations. Proposition 9.1.5 (4V representation of DW ). Let W be a spacetime and fix a time-like unit vector e0 and its Euclidean orthogonal complement V = [e0 ]⊥ . Identify the even part 4ev W of the spacetime Clifford algebra, and the Euclidean Clifford algebra 4V via the isomorphism in Proposition 3.3.5. We write a general spacetime multivector w ∈ 4W as w = w+ + e0 w− , where w± ∈ 4ev W ↔ 4V . Identifying a multivector field F in W in this way with a pair (F + , F − ) of multivector fields in V , and similarly for G, the wave Dirac equation DW F = G corresponds to ( (∂0 + DV )F+ = −G− , (∂0 − DV )F− = G+ . Proof. Since DW swaps 4ev W and 4od W fields, we obtain in a spacetime ONbasis {ei } that DW F = G is equivalent to ( (−e0 ∂0 + D0 )F+ = e0 G− , (−e0 ∂0 + D0 )(e0 F− ) = G+ , Pn where D0 := j=1 ej ∂j . Multiplying the first equation by e0 and commuting e0 to the left in the second equation establishes the claim.

Chapter 9. Dirac Wave Equations

290

Changing notation, an argument as in the proof of Proposition 9.1.5 in the /WF = G / equation D case that dim W is even also shows that the wave 4-Dirac corresponds to ( / V )F+ = −G− , (∂0 + D / V )F− = G+ . (∂0 − D +

+

/ W : v 7→ Here we use that 4V / = 4 / W via the representation ρ : V → 4 e0 v.(·), and write the spacetime spinor field as F = F+ + e0 F− in the splitting − + + − 4 / W . Note that e0 : 4 / W is invertible. / W ⊕4 / W →4 We end this section by considering how these Dirac operators are related to exterior and interior derivative operators d and δ. For any inner product space, it is clear from the definitions that the 4-Dirac operators are D = d + δ.

(9.1)

This holds true also for the wave 4-Dirac operator. We note the following refinement of Proposition 9.1.5. Proposition 9.1.6 (∧V representation of dW , δW ). Let W be a spacetime and use notation as for D in Proposition 9.1.5. Then the differential equations dW F = G and δW F = G correspond to ( ( (T − ∂0 + δV )F+ = −G− , (T + ∂0 + dV )F+ = −G− , − (T ∂0 − dV )F− = G+ , (T + ∂0 − δV )F− = G+ , respectively, where T + f = 21 (f + fb) and T − f = 12 (f − fb) denote the projections onto ∧ev V and ∧od V respectively. Proof. For example, we see that dW F = G is equivalent to ( −e0 ∧ ∂0 F+ + d0 F+ = e0 G− , −e0 ∧ ∂0 (e0 F− ) + d0 (e0 F− ) = G+ , Pn where d0 F := j=1 ej ∧ ∂j F . To relate the exterior product to the Clifford algebra isomorphism 4ev W ≈ 4V , we use the Riesz formula (3.4). We also note that w 7→ −e0 we0 yields an automorphism of 4ev W that negates e0 V . Therefore −e0 we0 corresponds to w b under the isomorphism 4ev W ↔ 4V . This yields for F ∈ ev 4 W ↔ 4V , e0 (e0 ∧ F ) = 21 (−F + e0 F e0 ) ↔ 12 (−F − Fb) = −T + F, e0 ∧ (e0 F ) = 21 (−F − e0 F e0 ) ↔ 12 (−F + Fb) = −T − F, Pn and with nabla calculus using ∇0 = j=1 ej ∂j that e0 (∇0 ∧ F ) = 12 ((e0 ∇0 )F − e0 F e0 (e0 ∇0 )) ↔ 12 (∇F + Fb∇) = dV F, ∇0 ∧ (e0 F ) = 1 (−(e0 ∇0 )F + e0 F0 (e0 ∇0 )) ↔ 1 (−∇F − Fb∇) = −dV F. 2

2

9.2. Dirac Equations in Physics

291

Similar calculations using the Riesz formula (3.3) prove the 4V representation for δW F = G. The 4-Dirac / operator in a general Euclidean or real inner product space, cannot be written as in (9.1). However, in the case of an even-dimensional Euclidean space with a complex structure given as in Example 5.1.5(i), we do have an invariant meaning of such exterior and interior derivative operators. Given a Euclidean space V of dimension n = 2m with an isometric complex structure J, consider the complex exterior algebra ∧V for the complex vector space V = (V, J), which comes with the corresponding Hermitian inner product (·, ·i. As in Example 5.1.5(i), the real linear map V → L(∧V) : v 7→ v y∗ (·) + v ∧ (·) gives a representation of the complex spinor space 4V / = ∧V. But the two terms induce separately the nabla operators Γ1 ψ := ∇ ∧ ψ

and

Γ2 ψ := ∇ y∗ ψ

/ = Γ1 + Γ2 . Fixing a complex ON-basis acting on spinor fields ψ : V → 4V / and D {ej }m for V and writing x for the real coordinates along ej and yj for the real j j=1 coordinates along Jej , we have, since {ej } ∪ {Jej } form a real ON-basis for V from Proposition 7.1.3, that Γ1 ψ =

m X

ej

∧

∂xj ψ +

j=1

and Γ2 ψ =

m X j=1

ej y∗ ∂xj ψ +

m X

iej

∧

∂yj ψ =

j=1

m X

ej

∧

∂zjc ψ

j=1

m m X X (iej ) y∗ ∂yj ψ = ej y∗ ∂zj ψ, j=1

j=1

since Jej = iej in V and (iej )y∗ w = −i(ej y∗ w) by sesquilinearity. Here we used the classical complex analysis operators ∂zj := ∂xj − i∂yj and ∂zjc := ∂xj + i∂yj . Since Γ∗1 = −Γ2 , one can develop a complex version of the theory of Hodge decomposition similar to the real theory in Chapter 10.

9.2

Dirac Equations in Physics

The aim of this section is to briefly review how Dirac equations appear in electromagnetic theory and quantum mechanics in physics. We model our universe by spacetime W with three space dimensions, as in special relativity. See Section 1.3. The unit of length is the meter [m]. Fixing a future-pointing time-like vector e0 with e20 = −1, we write V for the three-dimensional Euclidean space [e0 ]⊥ . We write the coordinate along e0 as x0 = ct,

Chapter 9. Dirac Wave Equations

292

where t is time measured in seconds [s] and c ≈ 2.998 · 108 [m/s] is the speed of light. Our discussion uses SI units. Out of the seven SI base units, we need the meter [m] for length, kilogram [kg] for mass, second [s] for time and ampere [A] for electric current. From this we have the SI derived units newton [N= kg·m/s2 ] for force, coulomb [C=A·s] for electric charge, volt [V= N·m/C] for electric potential, joule [J= Nm] for energy. We consider first Maxwell’s equations, which describe the time evolution of the electric and magnetic fields, which mediate the forces that electric charges in motion exert on each other. The charges that generate the electric and magnetic fields are described by a charge density and electric current density ρ(t, x) ∈ ∧0 V

and J(t, x) ∈ ∧1 V,

measured in units [C/m3 ] andR[A/m2 ] respectively. This means that a given domain the charge D ρdx, and the electric current through a 2-surface D ⊂ V contains R S in V is S hJ, ∗dyi, at time t. Here [dy] is the tangent plane to S and ∗dy is an infinitesimal vector normal to S, in the direction in which we measure the current. Maxwell’s four equations, which we discuss below, describe how ρ and J generate a vector field E(t, x) ∈ ∧1 V,

measured in units [N/C = V/m],

which is called the electric field, and a bivector field B(t, x) ∈ ∧2 V,

measured in units of tesla [T = N/(A · m)],

which we refer to as the magnetic field. The way we measure these fields is by placing a test charge with charge q0 at the point moving with velocity v0 ∈ ∧1 V . The electric and magnetic fields will then exert a force on this test charge given by the Lorentz force F = q0 E + q0 B x v0 . (9.2) Experiments show that the magnetic force is orthogonal to the velocity, and thus is described by a skew symmetric map. Recalling Proposition 4.2.3, this demonstrates that the magnetic field is a bivector field rather than a vector field. In classical vector notation, the magnetic field is described by the Hodge dual vector field ∗B, in which case the magnetic force is given by the vector product q0 v0 × (∗B). The three-dimensional exterior algebra ∧V = ∧0 V ⊕ ∧1 V ⊕ ∧2 V ⊕ ∧3 V provides a natural framework for expressing the four Maxwell equations, or more precisely eight scalar equations, for determining E ∈ ∧1 V and B ∈ ∧2 V from ρ ∈ ∧0 V and J ∈ ∧1 V . The constants of proportionality appearing are the permittivity of free space 0 ≈ 8.854 · 10−12 [C/(V· m)] and permeability of free space µ0 = 4π · 10−7 [V· s/(A· m)].

293

9.2. Dirac Equations in Physics

∧0 Gauss’s law for the electric field states that the flow of the electric field out through the boundary of a domain D is proportional to the charge Q = R ρdx contained in the domain: D Z hE, ∗dyi = −1 0 Q. ∂D

By Stokes’s theorem, Gauss’s law is equivalent to the ∧0 V -valued differential equation 0 ∇ y E = ρ. In classical vector notation this reads 0 h∇, Ei = ρ. ∧1 The Amp`ere–Maxwell law states that Z Z Z µ−1 hB, ∗dyi = hJ, ∗dxi + ∂ hE, ∗dxi 0 t 0 ∂S

S

S

for every 2-surface S. In the stationary case that ρ, J, E, and B are timeindependent, R it reduces to Amp`ere’s law, which shows that an electric current I := hJ, ∗dxi through S produces a magnetic field with circulation S R hB, ∗dyi = µ0 I. In the ∂S R nonstationary case, Maxwell added the necessary additional term 0 µ0 ∂t S hE, ∗dxi to the equation. By Stokes’s theorem, Amp`ere–Maxwell’s law is equivalent to the ∧1 V valued differential equation 0 ∂t E + µ−1 0 ∇ y B = −J. In classical vector notation this reads 0 ∂t E − µ−1 0 ∇ × (∗B) = −J. ∧2 Faraday’s law of induction states that a change of the integral of the magnetic field B over a 2-surface S induces an electric field around the boundary curve: Z Z hE, dyi = −∂t hB, dxi. ∂S

S

By Stokes’s theorem, Faraday’s law is equivalent to the ∧2 V -valued differential equation ∂t B + ∇ ∧ E = 0. In classical vector notation this reads ∂t (∗B) + ∇ × E = 0. ∧3 Gauss’s law for magnetic fields states that the integral of a magnetic field over the boundary of a domain D vanishes: Z hB, dyi = 0. ∂D

By Stokes’s theorem, the magnetic Gauss’s law is equivalent to the ∧3 V valued differential equation ∇ ∧ B = 0. In classical vector notation this reads h∇, ∗Bi = 0.

294

Chapter 9. Dirac Wave Equations

Figure 9.1: Maxwell’s equations for the electric vector field E and the magnetic bivector field B in ∧R3 . Since the electric and magnetic fields take values in the two different subspaces ∧1 V and ∧2 V of the exterior algebra, we can add them to obtain a sixdimensional total electromagnetic multivector field F . The most natural scaling is such that |F |2 is an energy density, with dimension [J/m3 ]. We set 1/2

−1/2

F := 0 E + µ0

B ∈ ∧1 V ⊕ ∧2 V.

Collecting and rescaling Maxwell’s equations, we have −1/2

∇ ∧ (µ0

B) = 0,

−1/2 1/2 c ∂t (µ0 B) + ∇ ∧ (0 E) 1/2 −1/2 c−1 ∂t (0 E) + ∇ y (µ0 B) 1/2 ∇ y (0 E) −1

= 0, 1/2

= −µ0 J, −1/2

= 0

ρ,

(9.3)

9.2. Dirac Equations in Physics

295

where c = (0 µ0 )−1/2 . Adding these four equations, we see that Maxwell’s equations are equivalent to the Dirac equation c−1 ∂t F + ∇ 4 F = G,

(9.4)

since Maxwell’s equations take values in the different homogeneous subspaces of −1/2 1/2 ∧V . Here G := 0 ρ − µ0 J is a ∧0 V ⊕ ∧1 V -valued multivector field, which we refer to as the electric four-current. From (9.4) it is clear that Maxwell’s equation is a wave Dirac equation for the ∧1 V ⊕ ∧2 V -valued electromagnetic field F . Example 9.2.1 (Static electromagnetic field). Assume that the sources ρ and J and the electromagnetic field are constant with respect to time, and that J is divergence-free. Then Maxwell’s equations reduce to the inhomogeneous Dirac equation ∇ 4 F = G, which by the Cauchy–Pompeiu formula from Theorem 8.1.8 has solution F (x) = Ψ(x) ∗ G(x) if G decays as x → ∞. This amounts to Z 1 y E(x) = ρ(x − y) 3 dy, 4π0 V |y| Z µ0 y B(x) = J(x − y) ∧ 3 dy. 4π V |y| Thus E is the Coulomb field from charge density ρ, and B is determined from J by the law of Biot–Savart. Exercise 9.2.2 (Pauli representation). Using an ON-basis {e1 , e2 , e3 } for V , write 1/2 ˜ 1 e1 + E ˜ 2 e2 + E ˜3 e3 , µ−1/2 B = B ˜1 e23 + B ˜2 e31 + B ˜3 e12 , µ1/2 J = J˜1 e1 + 0 E = E 0 0 −1/2 J˜2 e2 + J˜3 e3 , and 0 ρ = ρ˜. Represent the basis vectors {e1 , e2 , e3 } by the Pauli matrices from Example 3.4.19 and show that Maxwell’s equations become −1 ˜3 + iB ˜3 ˜ 1 − iE ˜2 + iB ˜1 + B ˜2 c ∂t + ∂3 ∂1 − i∂2 E E ˜1 + iE ˜2 + iB ˜1 − B ˜2 ˜ 3 − iB ˜3 ∂1 + i∂2 c−1 ∂t − ∂3 E −E ρ˜ − J˜3 −J˜1 + iJ˜2 = . ˜ ˜ −J1 − iJ2 ρ˜ + J˜3 Note that this representation requires that the components of the fields be realvalued, since we use a real algebra isomorphism 4R V ↔ C(2). For time-dependent electromagnetic fields we can obtain a spacetime Dirac formulation of Maxwell’s equation from Proposition 9.1.5. Namely, the electromagnetic field is really the spacetime bivector field 1/2

−1/2

FW := 0 e0 ∧ E + µ0

B ∈ ∧2 W,

solving the spacetime Dirac equation DW FW = −GW ,

(9.5)

Chapter 9. Dirac Wave Equations

296 −1/2

1/2

where GW := 0 ρe0 + µ0 J ∈ ∧1 W is the spacetime representation of the electric four-current. Since GW is a spacetime vector field and FW is a spacetime bivector field, Maxwell’s equations can equivalently be written as the system ( dW FW = 0, δW FW = −GW , by the mapping properties of DW = dW + δW . The difference between Maxwell’s equations and the Dirac equation is a constraint similar to the one described in Proposition 8.1.5. −1/2

1/2

Proposition 9.2.3 (Maxwell = Dirac + constraint). Let GW = 0 ρe0 + µ0 J ∈ ∧1 W . If FW is a ∧2 W -valued solution to (9.5), then ρ and J satisfy the continuity equation ∂t ρ + div J = 0. Conversely, if this continuity equation holds, then the multivector field FW solving the wave Dirac equation (9.5) described in Proposition 9.1.2 is ∧2 W -valued at all times, provided it is so at t = 0 with div E = ρ/0 and ∇ ∧ B = 0. Recall that the continuity equation ∂t ρ + div J = 0 expresses the fact that total charge is conserved. By Gauss’s theorem it shows that Z Z ρdx = − ∂t hJ, ∗dyi D

∂D

for every domain D ⊂ V . Proof. The necessity of the continuity equation follows from 2 δW GW = −δW FW = 0,

by the nilpotence of the spacetime interior derivative. For the converse, we investigate the proof of Proposition 9.1.2 and compute the ∧0 W and ∧4 W parts of FˆW = Fˆ . The ∧4 W part is −1/2 sin(|ξ|ct)

µ0

|ξ|

ˆ0 = 0, iξ ∧ B

since ∇ ∧ B0 = 0. The ∧0 W part is sin(|ξ|ct) e0 y (ξ y Fˆ0 ) |ξ| Z t 1/2 sin(|ξ|c(t − s)) −1/2 ρs − µ0 iξ y Jˆs ds +c 0 cos(|ξ|c(t − s))ˆ |ξ| 0 sin(|ξ|ct) 1/2 ˆ0 − −1/2 ρˆ0 ) = (0 iξ y E 0 |ξ| Z t sin(|ξ|c(t − s)) −1 −1/2 1/2 −c (c 0 ∂s ρˆs + µ0 iξ y Jˆs )ds, |ξ| 0

i

9.2. Dirac Equations in Physics

297

which vanishes, since div E0 = ρ0 /0 and ∂t ρ + div J = 0. This shows that FW is a homogeneous spacetime bivector field for all times. Example 9.2.4 (Lorentz transformation of E and B). From the spacetime representation FW of the electromagnetic field, we can find how the electric and magnetic fields transform under a change of inertial system. Consider two inertial observers O and O0 , with ON-bases {e0 , e1 , e2 , e3 } and {e00 , e10 , e20 , e03 } respectively. Assume that O sees O0 traveling in direction e3 at speed v. As in Example 4.4.1, the Lorentz boost that maps {ei } to {ei0 } is T x = exp(φe03 /2)x exp(−φe03 /2),

tanh φ = v/c.

In ∧2 W we have the electromagnetic field 1/2

µ0 F = c−1 e0 4 E + B = c−1 e00 4 E 0 + B 0 , where E = E1 e1 + E2 e2 + E3 e3 and B = B1 e23 + B2 e31 + B3 e12 are the fields 0 0 0 are the +B30 e12 measured by O and E 0 = E10 e10 +E20 e02 +E30 e30 and B 0 = B10 e23 +B20 e31 0 fields measured by O . We now compare the two measurements by identifying the ˜ = E 0 e1 +E 0 e2 +E 0 e3 two bases as in the discussion above Example 4.6.5, letting E 1 2 3 0 0 0 ˜ and B = B1 e23 + B2 e31 + B3 e12 . Then ˜ + B) ˜ exp(−φe03 /2). c−1 e0 4 E + B = exp(φe03 /2)(c−1 e0 4 E Applying the isomorphism 4ev W ≈ 4V from Proposition 3.3.5, we have equivalently ˜ = exp(−φe3 /2)(c−1 E + B) exp(φe3 /2). ˜+B c−1 E Computing the action of x 7→ exp(−φe3 /2)x exp(φe3 /2) on e1 , e2 , e3 , e23 , e31 , e12 , we get p E10 = (E1 − vB2 )/ 1 − v 2 /c2 , p 0 E2 = (E2 + vB1 )/ 1 − v 2 /c2 , E 0 = E , 3 3 p 0 B1 = (B1 + (v/c2 )B2 )/ 1 − v 2 /c2 , p 0 2 2 2 B2 = (B2 − (v/c )E1 )/ 1 − v /c , 0 B3 = B3 . From this we see that for speeds v comparable to the speed of light c, there is a significant mixing of E and B, which shows that indeed it is correct to speak of the electromagnetic field rather than electric and magnetic fields only, since the latter two depend on the inertial frame. We have seen that Maxwell’s equations can be written as a 4-Dirac wave equation DW FW = 0. However, the electromagnetic field FW is not a general spacetime multivector field, but a bivector field. This means that Maxwell’s equations are not identical to the 4-Dirac equation, but rather that we can embed

298

Chapter 9. Dirac Wave Equations

Maxwell’s equations in a Dirac equation. We show in the remaining sections of this chapter that this is a very useful technique, since in some respects, Dirac equations are better behaved than the Maxwell equations. An equation from physics that truly is a Dirac equation is Dirac’s original equation for the relativistic motion of spin-1/2 particles in quantum mechanics, such as electrons and quarks. With our notation this is a wave 4-Dirac equation in / physical spacetime with a lower-order mass term. Without any external potential, the free Dirac equation reads / = mcψ. (9.6) ~Dψ Here c is the speed of light and ~ = 6.626 · 10−34 [Js] is Planck’s constant. The parameter m is the mass of the particle, which in the case of the electron is m ≈ 9.109 · 10−31 [kg]. Dirac’s original approach was to look for a first-order differential equation that is a square root of the Klein–Gordon equation, that is, the wave equation with a mass term ~2 ψ = m2 c2 ψ,

(9.7)

which is obtained from the relativistic energy–momentum relation E 2 c−2 − p2 = m2 c2 by substituting E → i~∂t and p → −i~∇. Such a scalar first-order differential equation does not exist, but Dirac succeeded by allowing matrix coefficients. Having multivectors and spinors at our disposal, we already know that the 4/ has an invariant Dirac equation (9.6) for spacetime spinor fields ψ : W → 4W / geometric meaning. Exercise 9.2.5 (Matrix representation). Fix an ON-basis {e0 , e1 , e2 , e3 } for spacetime, and represent the dual basis {−e0 , e1 , e2 , e3 } by the imaginary Dirac matrices {iγ 0 , iγ 1 , iγ 2 , iγ 3 }, where γ k are Dirac’s gamma matrices as in Example 5.1.9. Show that Dirac’s equation reads ψ1 ∂1 − i∂2 ψ1 0 ∂0 ∂3 i∂ −∂ ψ 0 ∂ ∂ + 2 1 0 3 2 ψ2 i~ ψ3 = mc ψ3 . −∂3 −∂1 + i∂2 −∂0 0 −∂1 − i∂2 ψ4 ∂3 −∂0 ψ4 0 The physical interpretation of complex-valued wave functions ψ in quantum mechanics is that |ψ|2 represents a probability density for the position of the / , we require an inner particle. For the spinor-valued wave function ψ : W → 4W product on the spinor space 4V / . The following is a version of Proposition 5.3.1 for physical spacetime. Proposition 9.2.6 (Inner product). Let W be four-dimensional spacetime, with chosen future time direction fixed and complex spinor space 4W / . Then there exists a complex inner product (·, ·i on 4W / such that (ψ1 , v.ψ2 i = −(v.ψ1 , ψ2 i

9.2. Dirac Equations in Physics

299

for all ψ1 , ψ2 ∈ 4W / and v ∈ W , and −i(ψ, v.ψi > 0 / \ {0} and v ∈ Wt+ . If (·, ·i0 is any other such inner product, then for all ψ ∈ 4W there is a constant λ > 0 such that (ψ1 , ψ2 i0 = λ(ψ1 , ψ2 i for all ψ1 , ψ2 ∈ 4W / . Proof. The proof is analogous to that of Proposition 5.3.1. We look for a matrix M such that M ρ(v) = −ρ(v)∗ M, which exists, unique up to complex nonzero multiples, by Theorem 5.2.3, since (−ρ(v)∗ )2 = ρ(v 2 )∗ = hvi2 I. Using the representation ρ from Example 5.1.9, we see that we have M = λρ(e0 ), λ ∈ C \ {0}, where e0 is a fixed future-pointing time-like unit vector. For the duality to be an inner product, that is, symmetric, we must choose Re λ = 0, and to have −i(ψ, e0 .ψi > 0, we must have Im λ < 0. This shows uniqueness. Choosing λ = −i and v = e0 + v 0 , he0 , v 0 i = 0, we have −i(ψ, v.ψi = ψ ∗ (1 − ρ(e0 v 0 ))ψ > 0 if |v 0 | < 1, since ρ(e0 v 0 ) is |v 0 | times a C4 isometry. This completes the existence proof. This spacetime spinor inner product is used as follows. Given a wave function solving Dirac’s equation, a spinor field / ψ : W → 4W in spacetime, we define uniquely a vector field jp : W → ∧1 W by demanding hjp , vi = i(ψ, v.ψi for all v ∈ W . This exists by Proposition 1.2.3 and is referred to as the probability four-current. Fixing a future time direction e0 and writing jp = ρp e0 + c−1 Jp , Jp ∈ [e0 ]⊥ , it follows from the properties of (·, ·i that jp is a real vector field with time component ρp ≥ 0. This represents the probability density for the position of the particle. That Jp defines a probability current is clear from the continuity equation ∂t ρp + divV Jp = 0. This holds whenever ψ solves Dirac’s equation (9.6), since c−1 (∂t ρp + divV Jp ) = δW jp = −∂0 hjp , e0 i +

3 X

∂k hjp , ek i

1

=i

3 X 0

/ ψi = 0. / − i(Dψ, ∂k (ψ, ek∗ .ψi = i(ψ, Dψi

Chapter 9. Dirac Wave Equations

300

Recall the main reflector from Definition 5.2.1, which for physical spacetime we choose as w4 = ie0123 ∈ 44 W. In physics ρ(w4 ) is referred to as the chiral operator, and spinors in its eigenspaces 4± W are called right- and left-handed spinors respectively. To obtain a Euclidean formulation of Dirac’s equation, we fix a future time direction e0 and rewrite (9.6) as a coupled system of Euclidean 4-Dirac / equations for the right- and left-handed components of the wave function. As in the discussion after Proposition 9.1.5, we obtain ( / + = −mψ (c−1 ∂t + D)ψ ˜ −, − −1 / = mψ ˜ +, (c ∂t − D)ψ +

/ =D / V is the 4-Dirac / W , k = 1, 2, m / / ↔4 ˜ := mc/~, and D where ψ ± (t, x) ∈ 4V operator for the Euclidean three-dimensional space V = [e0 ]⊥ . Exercise 9.2.7. Under the algebra isomorphism +

−

/ / 2 3 (ψ + , ψ − ) ↔ ψ + + e0 ψ − ∈ 4 4V / W ⊕4 / W = 4W, show by uniqueness, with suitable normalization λ > 0, that the spacetime spinor inner product of ψ1 = ψ1+ + e0 ψ1− and ψ2 = ψ2+ + e0 ψ2− from Proposition 9.2.6 corresponds to i((ψ1+ , ψ2− i − (ψ1− , ψ2+ i), where (·, ·i denotes the Hermitian spinor inner product on 4V / . / 2 , we have Using the Hermitian inner product (ψ1+ , ψ2+ i + (ψ1− , ψ2− i on 4V the following Hilbert space result. Proposition 9.2.8 (Antiparticles and time evolution). Write Dirac’s equation as i~∂t ψ = H0 ψ, where / D m ˜ / 2 ). H0 := −i~c : L2 (V ; 4V / 2 ) → L2 (V ; 4V / −m ˜ −D Then the free Dirac Hamiltonian H0 has spectrum σ(H0 ) = (−∞, −mc2 ]∪[mc2 , ∞). We have an orthogonal splitting of L2 into spectral subspaces / 2 ) ⊕ L2− (V ; 4V / 2 ), L2 (V ; 4V / 2 ) = L+ 2 (V ; 4V where / 2) = L± 2 (V ; 4V

n −imψ ˜

/ ± iDψ

o t √ / m ˜ 2 − ∆ψ ; ψ ∈ 4V

are the spectral subspaces for the energy intervals [mc2 , ∞) and (−∞, −mc2 ] respectively. The solution to the inital value problem for the wave Dirac equation is m ˜ ψ(t, x) = (c−1 ∂t + H0 /(~c))Rct ψ(0, x),

ψ(t, ·) ∈ L2 (V ; 4V / 2 ),

9.2. Dirac Equations in Physics

301

˜ m where Rct denotes the Klein–Gordon Riemann function from Corollary 6.2.3, acting component-wise.

/ 2 ), the parts Splitting the wave function ψ = ψ + +ψ − , where ψ ± ∈ L2± (V ; 4V − ψ and ψ of positive and negative energy describe a particle and an antiparticle respectively. Note that time evolution by H0 preserves the subspaces L2± (V ; 4V / 2 ). It follows from Corollary 6.2.3 that Dirac’s equation has finite propagation speed ≤ c. However, unlike the massless case in Proposition 9.1.2, the Huygens principle is not valid for Dirac’s equation in three spatial dimensions. Compare also this time evolution for Dirac’s equation to that for Schr¨odinger’s equation in Example 6.3.6, where instead of finite propagation speed, we have the evolution given by an oscillatory quadratic exponential. +

Proof. Applying the Fourier transform in V , Dirac’s equation is turned into the ordinary differential equation ˜ iρ(ξ) m ∂t ψ = −c ψ, −m ˜ −iρ(ξ) p where ρ(ξ) ∈ L(4V / ). For the matrix we obtain eigenvalues ±i |ξ|2 + m ˜ 2 and p t eigenvectors −imψ ˜ −ρ(ξ)ψ ± |ξ|2 + m ˜ 2 ψ . Applying the inverse Fourier transform, this translates to the stated splitting. iρ(ξ) m ˜ To calculate the time evolution, we write j := √ 21 2 and ˜ |ξ| +m −m ˜ −iρ(ξ) note that j 2 = −I. It follows from Exercise 1.1.5 that p p p exp(−c |ξ|2 + m ˜ 2 j) = cos(ct |ξ|2 + m ˜ 2 ) − j sin(ct |ξ|2 + m ˜ 2 ), which under the Fourier transform is equivalent to the stated evolution formula. Example 9.2.9 (Foldy–Wouthuysen transformation). The particle and antiparticle splitting of a solution ψ : W → 4W / to Dirac’s equation (9.6) is independent of the inertial frame for W . Indeed, since H0 ψ = i~∂t ψ, we have p ic−1 ∂t ψ = ± m ˜ 2 − ∆ψ,

(9.8)

with sign +1 for particles, that is, ψ − = 0, and sign −1 for antiparticles, that is, ψ + = 0. Note that (9.6) is a differential equation that is a square root of the Klein–Gordon equation (9.7), and that (9.8) are also square roots of (9.7) although not differential equations. Using the spacetime Fourier transform, we see that the Fourier transforms of wave functions, in the distributional sense, for particles and antiparticles are supported on the two branches of the hyperboloid hξi2 + m ˜ 2 = 0. In particular, this shows the claimed relativistic invariance.

Chapter 9. Dirac Wave Equations

302

Exercise 9.2.10 (Charge conjugation). Consider the spinor space 4W / of physical spacetime, with spinor inner product as in Proposition 9.2.6. Show by generalizing the Euclidean theory from Section 5.3 that there exists an antilinear spinor / / → 4W conjugation 4W : ψ 7→ ψ † such that †

(v.ψ) = v.ψ † ,

v ∈ W, ψ ∈ 4W, /

†

and (ψ † ) = ψ, ψ ∈ 4W / , and that this is unique modulo a complex factor |λ| = 1. Show further that in the representation from Example 5.1.9, we can c choose ψ † = (ρ(e2 )ψ) and that this spinor conjugation is compatible with the spinor inner product as in Lemma 5.3.4. For Dirac’s equation, the operation ψ 7→ ψ † represents charge conjugation in physics, an operation that switches particles and antiparticles, which is readily † seen from (9.8). Mathematically, note that since (ψ † ) = ψ, the spinor conjugation yields a real structure on the spinor space of physical spacetime. This agrees with the fact that with our sign convention, the Clifford algebra 4W is isomorphic to R(4) by Theorem 3.4.13. Recall that a classical particle in an electromagnetic field is acted upon by the Lorentz force (9.2). For a quantum spin-1/2 particle in an electromagnetic field, the Dirac equation is modified by adding a source term and reads / = mcψ + iqAW .ψ. ~Dψ

(9.9)

The vector field AW : W → ∧1 W is a four-potential of the electromagnetic field FW = dW AW and q is the charge of the particle, which in case of the electron is q ≈ −1.602 · 10−19 [C]. A geometric interpretation of (9.9) is that AW provides Christoffel symbols for a covariant derivative as in Definition 11.1.5. The Faraday and magnetic Gauss laws show that the electromagnetic field FW is a closed spacetime bivector field, that is, dW FW = 0. Poincar´e’s theorem (Theorem 7.5.2) shows that locally this is equivalent to the existence of a spacetime vector field AW such that FW = dW AW . As we have seen, at least in the Euclidean setting, in Section 7.6, globally there can be topological obstructions preventing every closed field from being exact. And indeed, the famous Aharonov–Bohm experiment shows that in fact, FW being an exact bivector field is the correct −1/2 1/2 physical law, and not dW FW = 0. Writing AW = 0 Φe0 + µ0 A to obtain a Euclidean expression for potential, where Φ : W → R and A : W → V are scalar and vector potentials of the electromagnetic field, we have ( E = −∇Φ − ∂t A, B = ∇ ∧ A. Returning to (9.9), we note that a solution ψ still yields a probability four-current jp satisfying the continuity equation, as a consequence of AW being a real spacetime vector field. As in the free case, (9.9) describes the time evolution of the wave

9.3. Time-Harmonic Waves

303

functions for a particle and antiparticle pair. What is, however, not immediately clear is how the nonuniqueness of AW influences the solution ψ. To explain this, consider an exact spacetime bivector field FW : W → 42 W representing the electromagnetic field, and let AW , A˜W : W → 41 W be two different vector potentials, so that FW = dW AW = dW A˜W . Another application of Poincar´e’s theorem (Theorem 7.5.2) shows that locally, the closed vector field A˜W − AW is exact, so that A˜W = AW + ∇U, for some scalar potential U : W → 40 W = R. From the product rule, we deduce / = mcψ + iqAW .ψ if and only if that ~Dψ / iqU/~ ψ) = (mc + iq A˜W .)(eiqU/~ ψ). ~D(e Therefore ψ˜ := eiqU/~ ψ is the wave function of the particle in the electromagnetic ˜ = (ψ, v.ψi by sesquilinearity, the ˜ v.ψi field with potential A˜W . However, since (ψ, wave functions for the two choices of electromagnetic four-potential yield the same probability four-current jp . Therefore the physical effects are independent of the choice of electromagnetic four-potential AW .

9.3

Time-Harmonic Waves

Let W be a spacetime and fix a future pointing time-like unit vector e0 , and let V = [e0 ]⊥ . For the remainder of this chapter, we study time-harmonic solutions to the wave 4-Dirac equation DW F = DV F −e0 c−1 ∂t F = 0. We use the complexified spacetime Clifford algebra 4Wc , where the component functions Fs (x) belong to C. With a representation of the time-harmonic field as in Example 1.5.2, the Dirac equation reads (D + ike0 )F (x) = 0, with a wave number k := ω/c ∈ C. This is now an elliptic equation with a zeroorder term ike0 added to D = DV , rather than a hyperbolic equation. Since even the inner product on the real algebra 4W is indefinite, we require the following modified Hermitian inner product for the analysis and estimates to come. Definition 9.3.1 (Hermitian inner product). With V = [e0 ]⊥ ⊂ W as above, define the auxiliary inner product b1 e0−1 , w2 i, hw1 , w2 iV := he0 w

w1 , w2 ∈ 4W.

We complexify both the standard indefinite inner product h·, ·i on 4W and h·, ·iV to sesquilinear inner products (·, ·i and (·, ·iV on 4Wc respectively.

Chapter 9. Dirac Wave Equations

304

= u − e0 v. It b −1 We note that if w = u + e0 v, with u, v ∈ 4V , then e0 we 0 follows that (·, ·iV is a Hermitian inner product in which the induced basis {es } is an ON-basis for 4Wc whenever {ej }nj=1 is an ON-basis for V . We use the L2 R norm kf k2L2 = (f (x), f (x)iV dx of complex spacetime multivector fields f . The aim of this section is to generalize Section 8.3 from the static case k = 0 to k ∈ C. Note that (9.10) (D ± ike0 )2 = ∆ + k 2 . Definition 9.3.2 (Fundamental solution). Let Φk be the fundamental solution to the Helmholtz equation from Corollary 6.2.4 for Im k ≥ 0. Define fundamental solutions Ψk± = (D ± ike0 )Φk to the Dirac operators D ± ike0 . Note the relation Ψk− (x) = −Ψk+ (−x) − between these two families of fundamental solutions, and that Ψ+ 0 = Ψ0 equals ± Ψ from Definition 8.1.6. It is clear from Corollary 6.2.4 that Ψk in general can (1) be expressed in terms of Hankel functions Hν , which in odd dimensions are elementary functions involving the exponential function eik|x| .

Exercise 9.3.3 (Asymptotics). Show that in three dimensions, eik|x| x ik x ± Ψk (x) = − ± e0 . |x|3 |x| |x| 4π ± Note that Ψ± k ≈ Ψ near x = 0, while Ψk ∈ 4Wc is almost in the direction of the x light-like vector |x| ± e0 near x = ∞. Show that in dimension dim V = n ≥ 2 we have −(n−2) Ψ± ), k (x) − Ψ(x) = O(|x|

as x → 0,

−(n−1) as well as ∇ ⊗ (Ψ± ) as x → 0, and that k (x) − Ψ(x)) = O(|x| −ik|x| Ψ± k (x)e

−

π n−1 k (n−1)/2 1 −i 2 2 2e 2π

x |x| ± e0 |x|(n−1)/2

= O(|x|−(n+1)/2 ),

as x → ∞.

Theorem 9.3.4. Let D ⊂ V be a bounded C 1 -domain. If F : D → 4Wc solves (D + ike0 )F = 0 in D and is continuous up to ∂D, then Z F (x) = Ψ− for all x ∈ D. (9.11) k (y − x)ν(y)F (y) dy, ∂D

Note that since Ψ (x) = −Ψ+ k (−x), we can write the reproducing formula equivalently as Z F (x) = − Ψ+ k (x − y)ν(y)F (y) dy. −

∂D

9.3. Time-Harmonic Waves

305

Proof. The proof is analogous to that of Theorem 8.1.8. We define the linear 1-form D \ {x} × V → 4Wc : (y, v) 7→ θ(y, v) := Ψ− k (y − x)vF (y). For y 6= x, its exterior derivative is θ(y, ˙ ∇) = =

n X

∂yi Ψ− k (y − x) 4 ei 4 F (y)

i=1 (Ψ− k (y˙

− x) 4 ∇) 4 F (y) + Ψ− ˙ k (y − x) 4 (∇ 4 F (y)).

Since DF = −ike0 F and ˙ 4 ∇ = (∇ − ike0 ) 4 Φk− (x) ˙ 4∇ Ψ− k (x) ˙ = Φ− − ike0 )2 + (∇ − ike0 ) 4 ike0 ) = Ψ− k (x) 4 ike0 , k (x)((∇ we obtain θ(y, ˙ ∇) = 0. Applying the Stokes formula on the domain D := D \ B(x, ) and using the asymptotics of Ψ− near the origin from Exercise 9.3.3, the rest of the proof follows as for Theorem 8.1.8. It is essential in Theorem 9.3.4 that the domain D = D+ is bounded. In the exterior domain D− = V \ D, we need appropriate decay of F at ∞. When k 6= 0, this takes the form of a radiation condition as follows. Definition 9.3.5 (Radiating fields). Let F be a multivector field that solves (D + ike0 )F = 0 in D− . We say that F radiates at ∞ if Z lim Ψ− k (y − x)ν(y)F (y) dy = 0, R→∞

|y|=R

for every x ∈ D− . Note that by applying Theorem 9.3.4 to the annulus R1 < |x| < R2 , the limit is trivial since the integrals are constant for R > |x|. We need an explicit description of this radiation condition. Note that x ( |x| + e0 )2 = 0.

Proposition 9.3.6 (Radiation conditions). Let F be a multivector field that solves (D + ike0 )F = 0 in D− and is continuous up to ∂D, and assume that Im k ≥ 0 and k 6= 0. If x ( |x| + e0 )F = o(|x|−(n−1)/2 e(Im k)|x| ) as x → ∞, then F radiates. Conversely, if F radiates, then Z F (x) = Ψ+ k (x − y)ν(y)F (y) dy, ∂D

(9.12)

Chapter 9. Dirac Wave Equations

306 for all x ∈ D− . In particular, F = O(|x|−(n−1)/2 e−(Im k)|x| )

and

x ( |x| + e0 )F = O(|x|−(n+1)/2 e−(Im k)|x| )

as x → ∞. Not only does this give an explicit description of the radiation condition, but it also bootstraps it in that the necessary condition is stronger than the sufficient condition. x Proof. Assuming the decay condition on ( |x| + e0 )F , it suffices to prove that Z |F |2V dx = O(e2Im kR ) (9.13) |x|=R

as R → ∞. Indeed, the Cauchy–Schwarz inequality and the asymptotics for Ψ− k from Exercise 9.3.3 then show that F radiates. To estimate |F |V , we note that x x x x |( |x| + e0 )F |2V = (F, ( |x| − e0 )( |x| + e0 )F iV = 2|F |2V + 2(F, |x| e0 F iV . − := {x ∈ D− ; |x| < R}, Applying Stokes’s theorem for bodies on the domain DR we obtain Z Z x (F, |x| (F, νe0 F iV dx e0 F iV dx = |x|=R ∂D Z + (−ike0 F, e0 F iV + (F, −e0 (−ike0 F )iV dx. − DR

In total, this shows that Z Z Z Z 1 x |F |2V dx = (F, νe0 F iV dx−2Im k |F |2V dx, |( |x| +e0 )F |2V dx− − 2 |x|=R |x|=R DR ∂D from which (9.13) follows from the hypothesis, since ∂D is independent of R and since Im k ≥ 0. − The converse is an immediate consequence of Theorem 9.3.4 applied in DR , − and the asymptotics for Ψk from Exercise 9.3.3. There are two important applications of the Dirac equation (D + ike0 )F = 0 to classical differential equations, namely time-harmonic acoustic and electromagnetic waves. Example 9.3.7 (Helmholtz’s equation). For a scalar function u, define F = ∇u + ike0 u. Then u solves the Helmholtz equation ∆u + k 2 u = 0 from Example 6.3.4 if and only if F solves the Dirac equation DF + ike0 F = 0. However, note that F is not a general solution to this equation: it is a vector field F (x) ∈ 41 Wc .

9.3. Time-Harmonic Waves

307

To investigate the reproducing formula (9.11) for this vector field F , we evaluate the time-like and space-like parts of the equation, and get Z u(x) = ∂ν Φk (y − x)u(y) − Φk (y − x)∂ν u(y) dy, (9.14) Z∂D ∇u(x) = ∇Φk (y − x) 4 ν(y) 4 ∇u(y) + k 2 Φk (y − x)u(y)ν(y) dy, (9.15) ∂D

for x ∈ D, where ∂ν denotes the derivative in the normal direction. Equation (9.14) we recognise as the Green second identity for solutions to the Helmholtz equation, whereas (9.15) is an analogue of this for the gradient. This latter equation can be further refined by expanding the triple Clifford vector product as ∇Φk 4 ν 4 ∇u = (∂ν Φk )∇u + ∇Φk (∂ν u) − (∇Φk , ∇uiν + ∇Φk ∧ ν ∧ ∇u. Evaluating the vector part of (9.14), we obtain Z ∇u(x) = ∂ν Φk (y − x)∇u(y) + ∇Φk (y − x)∂ν u(y) ∂D − (∇Φk (y − x), ∇u(y)iν(y) + k 2 Φk (y − x)u(y)ν(y) dy,

x ∈ D.

For solutions to the Helmholtz equation ∆u + k 2 u = 0, the classical decay condition at ∞ is the Sommerfield radiation condition ∂r u − iku = o(|x|−(n−1)/2 e(Im k)|x| ), with ∂r denoting the radial derivative. To see its relation to the radiation condition for D + ike0 , we compute x x x x ( |x| + e0 )F = ( |x| + e0 )(∇u + ike0 u) = (1 + e0 ∧ |x| )(∂r u − iku) + ( |x| + e0 ) ∧ ∇S u, x x where ∇S u := |x| y ( |x| ∧ ∇u) is the angular derivative. By considering the scalar part of this identity, we see that the Dirac radiation condition entails the Sommerfield radiation condition. In fact the two conditions are equivalent. To see this, we can argue similarly to the proof of Proposition 9.3.6 to show that Green’s second identity (9.14) holds for the exterior domain D− . This will yield an estimate on ∇S u, given the Sommerfield radiation condition.

Example 9.3.8 (Time-harmonic Maxwell’s equations). Consider a time-harmonic electromagnetic wave F in a spacetime with three space dimensions. As in Sec1/2 −1/2 tion 9.2 we have F = 0 e0 ∧ E + µ0 B ∈ 42 Wc solving (D + ike0 )F = 0, √ with wave number k := ω/c = ω 0 µ0 . To investigate the reproducing formula (9.11) for this bivector field F , we evaluate the time-like and space-like bivector

Chapter 9. Dirac Wave Equations

308

parts of the equation, and obtain two classical equations known as the Stratton– Chu formulas: Z Φk (x − y) ν(y) × E(y) dy E(x) = ∇ × Z ∂U −∇ Φk (x − y) ν(y) · E(y) dy Z∂U + ikc Φk (x − y) ν(y) × (∗B)(y) dy, ∂U Z Φk (x − y) ν(y) × (∗B)(y) dy ∗B(x) = ∇ × Z ∂U −∇ Φk (x − y) ν(y) · (∗B)(y) dy ∂U Z − ik/c Φk (x − y) ν(y) × E(y) dy. ∂U

For solutions to the time-harmonic Maxwell’s equations, the classical decay conuller radiation condition dition at ∞ is the Silver–M¨ ( x −(n−1)/2 (Im k)|x| e ), |x| × E(x) − c(∗B)(x) = o(|x| x c |x| × (∗B)(x) + E(x) = o(|x|−(n−1)/2 e(Im k)|x| ).

Since x |x|

+ e0 (e0 E + cB) = e0 cB −

x |x| E

x + − E + c |x| B ,

we see that the Dirac radiation condition for the electromagnetic field is equivalent to the Silver–M¨ uller radiation condition. Note that both radiation conditions also give decay of the radial parts of the vector fields E and ∗B. Given the Cauchy reproducing formulas for D + ike0 , we can extend the theory of Hardy subspaces from Section 8.3 to the case k 6= 0. Acting on functions h : ∂D → 4Wc we define traces of Cauchy integrals Z + Ek h(x) := lim Ψ− k (y − z)ν(y)h(y)dy, z→x,z∈D + ∂D Z Ek− h(x) := − lim Ψ− x ∈ ∂D, k (y − z)ν(y)h(y)dy, z→x,z∈D −

∂D

and the principal value Cauchy integral Z Ek h(x) := lim+ Ψ− k (y − x)ν(y)h(y)dy, →0

x ∈ ∂D.

∂D\B(x;)

As in the static case k = 0, we limit ourselves to proving splittings of H¨older spaces of multivector fields into Hardy subspaces.

9.4. Boundary Value Problems

309

Theorem 9.3.9 (Hardy wave subspace splitting). Let D = D+ ⊂ V be a bounded C 1 domain, with exterior domain D− = V \D, and let Im k ≥ 0. Consider the function space C α = C α (∂D; 4Wc ), for fixed regularity 0 < α < 1. Then the operators Ek+ , Ek− , and Ek are well defined and bounded on C α (∂D). The operators Ek± are complementary projections, with Ek± = 12 (I ± Ek ), and they split C α (∂D) into Hardy subspaces C α (∂D) = Ek+ C α ⊕ Ek− C α . There is a one-to-one correspondence, furnished by the Cauchy integral (9.11) and the trace map, between fields in the interior Hardy subspace Ek+ C α and fields in D+ solving DF + ike0 F = 0 and H¨ older continuous up to ∂D. Likewise, there is a one-to-one correspondence, furnished by the Cauchy integral (9.12) and the trace map, between fields in the exterior Hardy subspace Ek− C α and fields in D− solving DF + ike0 F = 0, H¨ older continuous up to ∂D and radiating at ∞. Proof. Define the operator Z Rk h(x) = (Ψ− k (y − x) − Ψ(y − x))ν(y)h(y)dy,

x ∈ V.

∂D

By the asymptotics at x = 0 from Exercise 9.3.3, Rk h(x) is a well-defined convergent integral for all x ∈ V . Furthermore, by differentiating under the integral sign, we have Z dy |∂j Rk h(x)| . khk∞ (9.16) . khk∞ ln(1/dist (x, ∂D)), |y − x|n−1 ∂D for 0 < dist (x, ∂D) < 1/2 and j = 1, . . . , n, where khk∞ := sup∂D |h|. Integrating (9.16) and (8.6) as in Proposition 8.3.3, it follows that Ek± are bounded on C α , since Ek± = E ± ± Rk . Note that integrating (9.16) in fact shows that Rk h is H¨older continuous across ∂D. Therefore it follows from Proposition 8.3.4 that Ek± h = E ± h ± Rk h = 21 h ± 12 (Eh + 2Rk h) = 12 (h ± Ek h) and that Ek = E + 2Rk is a bounded operator on C α . As in Theorem 8.3.6, we conclude that Ek+ + Ek− = I and that E ± are projections by Theorem 9.3.4 and Proposition 9.3.6 respectively. This proves the splitting into Hardy subspaces for D + ike0 .

9.4

Boundary Value Problems

For the remainder of this chapter, we study boundary value problems (BVPs) for Dirac operators, where our problem is to find a solution F to DF (x) + ike0 F (x) = 0

Chapter 9. Dirac Wave Equations

310

in a domain D that satisfies a suitable condition on the trace F |∂D . To make the problem precise, one needs to state assumptions on ∂D: How smooth is it? Is it bounded or unbounded? We also need to specify the space of functions on ∂D in which we consider F |∂D , and in what sense the boundary trace F |∂D is meant. To start with, we postpone these details, and assume only given a Banach space H of functions on ∂D. A concrete example is H = C α (∂D) from Theorem 9.3.9. We assume that the Cauchy integral operator Ek acts as a bounded operator in H, and we recall that Ek is a reflection operator, Ek2 = I, and it induces a splitting of H into Hardy wave subspaces. Solutions F to DF + ike0 F = 0 in D = D+ are in one-to-one correspondence with f = F |∂D in Ek+ H. A formulation of a Dirac BVP is ( DF + ike0 F = 0, in D, T f = g, on ∂D. Here T : H → Y is a given bounded and linear operator onto an auxiliary Banach function space Y, which contains the boundary datum g to the BVP. In such an operator formulation, well-posedness of the BVP means that the restricted map T : Ek+ H → Y

(9.17)

is an isomorphism. Indeed, if so, then for every data g ∈ Y we have a unique solution f ∈ Ek+ H, or equivalently a solution F to DF + ike0 F = 0 in D, which depends continuously on g. The main goal in studying BVPs is to prove such well-posedness. Almost as good is to prove well-posedness in the Fredholm sense, meaning that T is a Fredholm map. In this case, g needs to satisfy a finite number of linear constaints for f to exist, and this is unique only modulo a finite-dimensional subspace. Proposition 9.4.1. Let T : H → Y be a surjective bounded linear operator. Then the restriction T : Ek+ H → Y is an isomorphism if and only if we have a splitting Ek+ H ⊕ N(T ) = H. Proof. If T : Ek+ H → Y is an isomorphism, denote its inverse by T0 : Y → Ek+ H. Then P := T0 T : H → H is a projection with null space N(T ) and range Ek+ H, which proves the splitting. Conversely, if we have a splitting Ek+ H ⊕ N(T ) = H, then clearly T : Ek+ H → Y is injective and surjective. Without much loss of generality, we assume from now on that T is a bounded projection on H with range Y ⊂ H. We consider the following abstract formulation of the BVP, in terms of two bounded reflection operators A and B on H: A2 = I

and B 2 = I.

The operator A plays the role of the Cauchy integral Ek , so that A+ = 21 (I + A) projects onto traces of solutions to the differential equation in D+ and A− =

9.4. Boundary Value Problems

311

1 − 2 (I − A) projects onto traces of solutions to the differential equation in D , with appropriate decay at infinity. The operator B encodes two complementary boundary conditions: either T = B + = 12 (I + B) or T = B − = 12 (I − B) can be used to define boundary conditions. Note that we have null spaces N(B + ) = B − H and N(B − ) = B + H. We note that the algebra for each of the operators A and B is similar to that of Ek in Theorem 9.3.9. We have two different splittings of H:

H = A+ H ⊕ A− H

and H = B + H ⊕ B − H,

and A = A+ − A− and B = B + − B − . The core problem in the study of BVPs is to understand the geometry between on the one hand the subspaces A± H related to the differential equation, and on the other hand the subspaces B ± H related to the boundary conditions. Example 9.4.2 (BVP = operator 4R2 ). The algebra of two reflection operators A and B can be viewed as an operator version of the Clifford algebra 4R2 for the Euclidean plane R2 . Indeed, consider two unit vectors a, b ∈ V . Since a2 = b2 = 1 in 4R2 , we have here a very simple example of an abstract BVP. The geometry of a and b is described by the angle φ between the vectors. We recall that this angle can be calculated from the anticommutator 1 2 (ab

+ ba) = cos φ,

or from the exponential ab = eφj , where j is the unit bivector with orientation of a ∧ b. Definition 9.4.3 (Well-posedness). Let A, B : H → H be two reflection operators on a Banach space H. Define the cosine operator 1 2 (AB

+ BA)

and the rotation operators AB

and BA = (AB)−1 .

We say that the AB boundary value problems are well posed (in the Fredholm sense) if the four restricted projections B ± : A± H → B ± H are all isomorphisms (Fredholm operators). Exercise 9.4.4 (Simplest abstract BVP). Let H = C2 and consider the two orthogonal reflection operators 1 0 cos(2α) sin(2α) A= and B = , 0 −1 sin(2α) − cos(2α) for some 0 ≤ α ≤ π/2. Compute the cosine and rotation operators and show that the AB BVPs are well posed if and only if 0 < α < π/2. Show that we have spectra σ( 12 (AB + BA)) = {cos(2α)} and σ(AB) = {ei2α , e−i2α }, and that the AB BVPs fail to be well posed exactly when these spectra hit {+1, −1}.

312

Chapter 9. Dirac Wave Equations

Figure 9.2: The two splittings encoding an abstract BVP, with associated reflection operators. For two general reflection operators A and B, the associated cosine and rotation operators each contain the necessary information to conclude well-posedness of the AB BVPs. Useful identities include the following, which are straightforward to verify: 1 2 (I 1 2 (I

+ BA) = B + A+ + B − A− , +

−

−

+

− BA) = B A + B A ,

(9.18) (9.19)

2(I + C) = (I + BA)B(I + BA)B,

(9.20)

2(I − C) = (I − BA)B(I − BA)B.

(9.21)

Proposition 9.4.5 (Well-posedness and spectra). Let A, B : H → H be two reflection operators on a Banach space H. Then the following are equivalent: (i) The AB BVPs are well posed. (ii) The spectrum of the rotation operator BA does not contain +1 or −1.

9.4. Boundary Value Problems

313

(iii) The spectrum of the cosine operator C = 21 (AB + BA) does not contain +1 or −1. Similarly, the AB BVPs are well posed in the Fredholm sense if and only if I ±BA are Fredholm operators, if and only if I ± C are Fredholm operators. Proof. We note that B + A+ + B − A− is invertible if and only if the BVPs B + : A+ H → B + H and B − : A− H → B − H are well posed, and similarly for B + A− + B − A+ . Also ((I + BA)B)2 is invertible if and only if I + BA is invertible, and similarly for I − BA. The equivalences follow. With this general setup, and Proposition 9.4.5 as our main tool for proving well-posedness of Dirac BVPs, we now consider the two main examples that we have in mind. The boundary condition B, unlike A, is typically a pointwise defined multiplier, derived from the orientation of the tangent space to ∂D, described by the normal vector ν. For the remainder of this section we assume that D is a bounded C 2 domain. In this case, we note that ν is a C 1 smooth vector field on ∂D. We will see below that the cosine operators for such smooth BVPs tend to be compact, leading directly to BVPs that are Fredholm well posed by Proposition 9.4.5. Indeed, by the general Fredholm theory outlined in Section 6.4, the operators I ± C will then be Fredholm operators with index zero. The cosine operators typically are generalizations of the following classical integral operator from potential theory. Exercise 9.4.6 (Double layer potential). Consider the integral operator Z Kf (x) := hΨ(y − x), ν(y)if (y)dy, x ∈ ∂D, ∂D

with kernel k(x, y) = hΨ(y − x), ν(y)i. In three dimensions, a physical interpretation of k(x, y) is that of the electric potential from a dipole at y, in the direction ν(y), and for this reason K is called the double layer potential operator. The operator K is weakly singular on smooth domains. More precisely, show that on a C 2 boundary ∂D of dimension n − 1, we have kernel estimates |k(x, y)| . |x − y|2−n and |∇0x k(x, y)| . |x − y|1−n , x 6= y, x, y ∈ ∂D, where ∇0x denotes the tangential gradient in the x-variable. Lemma 9.4.7 (Weakly singular = compact). Let Z T f (x) = k(x, y)f (y)dy,

x ∈ ∂D,

∂D

be a weakly singular integral operator with kernel estimates |k(x, y)| . |x − y|2−n and |∇0x k(x, y)| . |x − y|1−n , x, y ∈ ∂D. Here ∇0x denotes the tangential gradient along ∂D in the variable x. Then T is a compact operator on C α (∂D) for all 0 < α < 1.

Chapter 9. Dirac Wave Equations

314

Proof. Assume that x, x0 ∈ ∂D with |x − x0 | = . Write T f (x0 ) − T f (x) Z Z (k(x0 , y) − k(x, y))f (y)dy + = |y−x|≤2

(k(x0 , y) − k(x, y))f (y)dy

|y−x|>2

=: I0 + I1 . For I1 , we obtain from the mean value theorem the estimate |k(x0 , y) − k(x, y)| . /|y − x|1−n when |y − x| > 2. This yields |I1 | . ln −1 kf kL∞ . For I0 , we estimate |k(x0 , y) − k(x, y)| . |y − x|2−n and obtain |I0 | . kf kL∞ . It follows that T : C α (∂D) → C β (∂D) is bounded for all β < 1, and that T : C α (∂D) → C α (∂D) is a compact operator. Example 9.4.8 (Normal/tangential BVP). Our main example of a Dirac BVP occurs when the differential equation is DF + ike0 F = 0, that is, A = Ek , for a fixed wave number k ∈ C, and the boundary conditions are encoded by the reflection operator B = N given by N f (x) := ν(x) 4 fd (x) 4 ν(x),

x ∈ ∂D.

We know from Section 4.1 that N reflects the multivector f (x) across the tangent plane to ∂D at x, and assuming that ν ∈ C 1 , we have that N is bounded on C α (∂D). The projection N + in this case will yield a boundary condition that specifies the part of f (x) tangential to ∂D in the sense of Definition 2.8.6. This can be verified using the Riesz formula (3.4), as N + f = 12 (f + ν fbν) = ν 12 (νf + fbν) = ν(ν ∧ f ) = ν y (ν ∧ f ). The corresponding calculation using (3.3) shows that N − f = ν ∧ (ν y f ) yields the boundary condition that specifies the normal part of f (x). The four Ek N BVPs consist of two BVPs for solutions to DF + ike0 F = 0 in the interior domain D+ , where the tangential or normal part of F |∂D is specified, and two BVPs for solutions to DF + ike0 F = 0 in the exterior domain D− , where the tangential or normal part of F |∂D is specified. By Proposition 9.4.5, the wellposedness of these four BVPs may be studied via the associated cosine operator Ek N +N Ek = (Ek +N Ek N )N . When k = 0, we calculate using Ψν = 2hΨ, νi−νΨ

9.4. Boundary Value Problems

315

that 1 2 (E

Z + N EN )f (x) = p.v.

Ψ(y − x)ν(y)f (y) + ν(x)Ψ(y − x)f (y)ν(y)ν(x) dy

Z∂D = 2p.v. hΨ(y − x), ν(y)if (y)dy ∂D Z + p.v. (ν(x) − ν(y))Ψ(y − x)f (y)dy ∂D Z + ν(x) p.v. Ψ(y − x)f (y)(ν(y) − ν(x))dy ν(x). ∂D

Assume now that D is a bounded C 2 domain. We can then apply Lemma 9.4.7 to each of these three terms, showing that EN + N E is a compact operator on C α . Moreover, the compactness of Ek − E on C α follows by yet another application of Lemma 9.4.7. We conclude that the Ek N BVPs are well posed in the sense of Fredholm in C α (∂D) for C 2 domains D. Example 9.4.9 (Spin BVP). The second example of a Dirac BVP that we shall consider is that in which the boundary conditions are induced by left Clifford multiplication by the normal vector ν. For technical reasons we study boundary conditions encoded by the reflection operator B = S given by Sf (x) := e0 4 ν(x) 4 f (x),

x ∈ ∂D.

Note that (e0 n)2 = −e20 n2 = 1, so indeed S is a reflection operator, and it is bounded on C α , since ν is C 1 regular. The factor e0 is motivated by Proposition 3.3.5 as in Proposition 9.1.5, and makes 4W ev invariant under S. As before, we study the differential equation DF + ike0 F = 0 encoded by the reflection operator A = Ek . It would be more natural to consider the operators Ek and S acting on spinor fields ∂D → 4W / , though, since both operators use only left multiplication by multivectors. So the true nature of the Ek S BVPs are BVPs for the 4-Dirac / operator. However, we here consider the 4-Dirac operator, since we aim to combine the Ek S and the Ek N BVPs in Section 9.5. The ranges of the projections S + f = 12 (1 + e0 ν)f

and S − f = 12 (1 − e0 ν)f

are seen to be the subspaces of multivector fields containing left Clifford factors that are respectively the light-like vectors ν ±e0 . The advantage of the S boundary conditions is that in some sense, the Ek S BVPs are the best local BVPs possible for the differential equation DF + ike0 F = 0. We will see several indications of this below.

Chapter 9. Dirac Wave Equations

316

For the cosine operator 21 (ES + SE), we calculate Z 1 (x) = p.v. Ψ(y − x)ν(y)f (y) + ν(x)Ψ(y − x)f (y) dy (E + SES)f 2 ∂D Z = 2p.v. hΨ(y − x), ν(y)if (y)dy Z∂D + p.v. (ν(x) − ν(y))Ψ(y − x)f (y)dy, ∂D

since e0 anticommutes with the space-like vectors ν and Ψ. As in Example 9.4.8, we conclude from this, using Lemma 9.4.7, that the Ek S BVPs are well posed in the sense of Fredholm in C α on C 2 domains. Having established well-posedness in the Fredholm sense for the Ek N and Ek S BVPs, we know that the BVP maps (9.17) are Fredholm operators, so that the null spaces are finite-dimensional and the ranges are closed subspaces of finite codimension. It remains to prove injectivity and surjectivity, whenever possible. Proposition 9.4.10 (Injectivity). Let 0 < α < 1 and Im k ≥ 0. • For the Ek N BVPs we have Ek+ C α ∩N + C α = Ek+ C α ∩N − C α = Ek− C α ∩N + C α = Ek− C α ∩N − C α = {0} if Im k > 0. Moreover, if D− is a connected domain and k ∈ R \ {0}, then Ek− C α ∩ N + C α = Ek− C α ∩ N − C α = 0. • For the Ek S BVPs we have Ek+ C α ∩ S + C α = Ek− C α ∩ S − C α = {0} whenever Im k ≥ 0. Proof. For the estimates we require the Hermitian inner product (w1 , w2 iV := (e0 w b1 e−1 0 , w2 i on 4Wc from Definition 9.3.1. Consider first the interior BVPs. Given f = F |∂D ∈ Ek+ C α , we define the linear 1-form D × V → C : (y, v) → (e0 vF (y), F (y)iV , which has nabla derivative (e0 ∇F (y), ˙ F (y)i ˙ V = (e0 (∇F ), F iV − (e0 F, ∇F iV = (e0 (−ike0 F ), F iV − (e0 F, −ike0 F iV = −2Im k|F |2V . From the Stokes formula (7.4), it follows that Z Z (Sf, f iV dy = −2Im k ∂D

D+

|F |2V dx.

9.4. Boundary Value Problems

317

If f ∈ N ± C α , then (Sf, f iV = 0, and we conclude that F = 0 if Im k > 0. So in this case, Ek+ C α ∩ N ± C α = {0}. If f ∈ S + C α , then (Sf, f iV = |f |2V , and we conclude that f = 0 whenever Im k ≥ 0, so Ek+ C α ∩ S + C α = {0}. Consider next the exterior BVPs. Let f = F |∂D ∈ Ek− C α , and fix a large − := D− ∩ {|x| < R}, radius R. From Stokes’s theorem applied to the domain DR we have Z Z Z x (e0 |x| F, F iV dx − (Sf, f iV dy = −2Im k |F |2V dx. |x|=R

− DR

∂D

Furthermore, on the sphere |x| = R, we note that x x |( |x| + e0 )F |2V = 2|F |2V − 2(e0 |x| F, F iV ,

and obtain the identity Z Z x (|F |2V − 12 |( |x| + e0 )F |2V )dx − |x|=R

Z (Sf, f iV dy = −2Im k

∂D

− DR

|F |2V dx.

R x Using Proposition 9.3.6, we have limR→∞ |x|=R |( |x| +e0 )F |2V dx = 0 for all Im k ≥ − α 2 0. If f ∈ S C , then (Sf, f iV = −|f |V , and we again conclude that f = 0, so Ek− C α ∩ S − C α = {0}. If f ∈ N ± C α , then (Sf, f iV = 0, and we have Z Z Z x |F |2V dx + 2Im k |F |2V dx = 12 |( |x| + e0 )F |2V dx → 0, R → ∞. |x|=R

− DR

|x|=R

When Im k > 0, this shows that F = 0. When k ∈ R \ {0}, we have Z lim |F |2V dy = 0. R→∞

|x|=R

Applying Rellich’s lemma (Lemma 6.3.5) to the component functions Fs of F , which satisfy Helmholtz’s equation ∆Fs + k 2 Fs = 0, we also in this case conclude that F = 0, so in either case, Ek− C α ∩ N ± C α = {0}. Summarizing our findings, we have obtained the following well-posedness results. Theorem 9.4.11 (C α well-posedness). For the Dirac BVPs with boundary function space C α (∂D), 0 < α < 1, on domains with C 2 regular boundary ∂D, we have the following well-posedness results. The four BVPs N ± : Ek± C α → N ± C α are well posed when Im k > 0. If the exterior domain D− is connected, then the exterior BVPs N ± : Ek− C α → N ± C α are well posed for all nonzero Im k ≥ 0. The two spin-Dirac BVPs S − : Ek+ C α → S − C α and S + : Ek− C α → S + C α are well posed for all Im k ≥ 0.

318

Chapter 9. Dirac Wave Equations

We remark that by applying analytic Fredholm theory, one can prove that in fact, also the interior Ek N BVPs are well posed for k ∈ R, except for a discrete set of resonances. Proof. We make use of the Fredholm theory outlined in Section 6.4. By Example 9.4.8 and Proposition 9.4.5, the Ek N BVPs are well posed in the Fredholm sense for all k. By Proposition 9.4.10 the four maps N ± : Ek± C α → N ± C α are injective when Im k > 0. We conclude that I ± 12 (Ek N + N Ek ) are injective Fredholm operators with index zero, and therefore invertible. So the Ek N BVPs are well posed when Im k > 0. For k ∈ R\{0}, we have injective semi-Fredholm maps N ± : Ek− C α → N ± C α by Proposition 9.4.10. By perturbing Ek− to Im k > 0, Lemma 9.4.12 below proves that they are invertible. The well-posedness of S − : Ek+ C α → S − C α and S + : Ek− C α → S + C α follows from Example 9.4.8 and Proposition 9.4.10, using Proposition 9.4.5. Note that I − 12 (Ek S + SEk ) = ((S − Ek+ + S + Ek− )S)2 is an injective Fredholm operator with index zero, and hence invertible. The following two techniques for proving existence of solutions to BVPs turn out to be useful. Lemma 9.4.12 (Perturbation of domains). Let At , t ∈ [0, 1], and B be reflection operators on a Banach space H, and consider the family of BVPs described by B + : + A+ t H → B H. If these are all semi-Fredholm maps and if t 7→ At is continuous, + + + + then the indices of B + : A+ 0 H → B H and B : A1 H → B H are equal. + Proof. We parametrize the domains A+ t H by the fixed space A0 H. Considering + + + + ˜ At := At : A0 H → At H as one of the four abstract A0 At BVPs, we note that

I + At A0 = 2I + (At − A0 )A0 . If kAt −A0 k ≤ 1/kA0 k, it follows that I +At A0 is invertible, and from (9.18) we see + + + ˜+ in particular for 0 ≤ t ≤ , that A˜+ t is invertible. Let Bt := B : At H → B H. + + + + ˜t A˜t , we conclude that Ind(B ˜ A˜ ) = Applying the method of continuity to B + + ˜ + A˜+ ). Since A˜+ ˜ ˜ Ind(B are invertible, we obtain Ind( B ) = Ind( B ). Repeating t 0 0 0 ˜ + ) = Ind(B ˜ + ). this argument a finite number of times, we conclude that Ind(B 1 0 Lemma 9.4.13 (Subspace duality). Let A and B be two reflection operators on a Banach space H, and consider the BVP described by B + : A+ H → B + H. This map is surjective if and only if the dual BVP described by (B ∗ )− : (A∗ )− H∗ → (B ∗ )− H∗ is an injective map. Proof. Note that A∗ and B ∗ are reflection operators in H∗ . By duality as in Section 6.4, we have (A+ H)⊥ = R(A+ )⊥ = N((A∗ )+ ) = R((A∗ )− ) = (A∗ )− H∗ and similarly (B − H)⊥ = (B ∗ )+ H∗ . Similarly to Proposition 9.4.1, since (A+ H + B − H)⊥ = (A+ H)⊥ ∩ (B − H)⊥ , this translates to the claim.

9.5. Integral Equations

319

We end this section with two applications of the techniques in this section to Dirac’s equation. Example 9.4.14 (The MIT bag model). Consider Dirac’s equation i~∂t ψ = H0 ψ from Proposition 9.2.8 on a bounded domain D ⊂ V . The MIT bag model is used in physics to describe the quarks in a nucleon, that is, a proton or neutron. The bag D represents the nucleon, and the boundary condition is ν.ψ = ψ, 2

or in the 4V representation e0 ν.ψ = ψ. This boundary condition implies in particular that the probability current hjp , νi = i(ψ, ν.ψi = i(ψ, ψi across ∂D vanishes, since jp is a real spacetime vector field. We see that with suitable modifications, such BVPs for time-harmonic solutions to Dirac’s equation can be studied with the methods described in this section. Example 9.4.15 (Chirality of (anti-)particles). What we refer to here as abstract BVPs, namely the algebra of two reflection operators describing the geometry between two splittings of a function space, appear in many places independent of any BVPs. One of many such examples we saw in connection to Proposition 9.2.8. Consider the Hilbert space H := L / c2 ), where we saw two different split2 (V ; 4V I 0 tings. The reflection operator B = encodes the Chiral subspaces of right0 −I and left-handed spinors, whereas / −i D m ˜ A = sgn(H0 ) = √ / . ˜ −D m ˜ 2 − ∆ −m / by Pauli matrices, the Fourier mulUsing, for example, the representation of 4V tiplier of the rotation operator AB at frequency ξ ∈ V is seen to have the four eigenvalues p λ = (±|ξ| ± im)/ ˜ ˜ 2. |ξ|2 + m Therefore the spectrum of AB is precisely the unit circle |λ| = 1. We conclude that although the spectral subspaces L± / 2 ) do not intersect the chiral subspaces, 2 (V ; 4V the angle between them is zero. The problem occurs at high frequencies: particles or antiparticles of high energy may be almost right- or left-handed.

9.5

Integral Equations

The aim of this section is to use the somewhat abstract theory from Section 9.4 to derive integral equations for solving Dirac BVPs, with good numerical properties, that have recently been discovered.

320

Chapter 9. Dirac Wave Equations

• It is desirable to extend the theory to nonsmooth domains, which have boundaries that may have corners and edges, as is often the case in applications. Ideally, one would like to be able to handle general Lipschitz domains. • To solve a given BVP, we want to have an equivalent integral formulation Z k(x, y)f (y)dy = g(x), x ∈ ∂D, ∂D

where the boundary datum gives g and the integral equation is uniquely solvable for f if and only if the BVP to solve is well posed. Ideally we want to have a function space without any constraints, meaning a space of functions ∂D → L with values in a fixed linear space L and coordinate functions in some classical function space. In this section we let D be a bounded strongly Lipschitz domain. At this generality, the normal vector field ν is only a measurable function without any further smoothness. To extend the theory from Section 9.4 and keep the basic operators Ek , N and S bounded, we shall use L2 = L2 (∂D; 4Wc ), which is the most fundamental space to use for singular integral operators like Ek . Indeed, the singular integral operator Ek is bounded on L2 (∂D) for every Lipschitz domain D, by Theorem 8.3.2 and Exercises 9.3.3 and 6.4.3. We first consider Fredholm well-posedness of the Ek N BVPs in L2 on bounded strongly Lipschitz domains. On such nonsmooth domains, it is not true in general that Ek N + N Ek , or even the classical double layer potential from Exercise 9.4.6, is compact. However, we recall from Proposition 9.4.5 that it suffices to show that the spectrum of 12 (Ek N + N Ek ) does not contain ±1. Theorem 9.5.1 (Rellich estimates). Let D be a bounded strongly Lipschitz domain, and let θ be a smooth compactly supported field that is transversal to ∂D as in Exercise 6.1.8. Define the local Lipschitz constant L := sup∂D (|θ ∧ ν|/hθ, νi) for ∂D. Then λI + Ek N is a Fredholm operator on L2 (∂D) of index zero whenever λ = λ1 + iλ2 , |λ2 | < |λ1 |/L, λ1 , λ2 ∈ R. Note that since Ek N and (Ek N )−1 = N Ek are bounded, we also know that the spectrum of Ek N is contained in an annulus around 0. Furthermore, since ((λI + Ek N )Ek )2 = λ(λ + λ−1 + Ek N + N Ek ), the resolvent set of the cosine operator contains the hyperbolic regions onto which λ 7→ 12 (λ + λ−1 ) maps the double cone |λ2 | < |λ1 |/L. And for λ = ±1 it follows in particular that the Ek N BVPs are well posed in the Fredholm sense in L2 (∂D).

9.5. Integral Equations

321

Proof. To motivate the calculations to come, we consider first the BVP described by N + : Ek+ L2 → N + L2 . To estimate kf kL2 in terms of kN + f kL2 , we insert the factor hθ, νi and express it with the Clifford product as Z Z Z 1 2 2 kf kL ≈ |f | hθ, νidy = (f, f (θν + νθ)i dy = Re (f ν, f θiV dy. V V 2 2 ∂D ∂D ∂D We next use the reversed twin of the Riesz formula (3.4) to write f ν = 2f ∧ ν − ν fb. We estimate the last term so obtained by applying Stokes’s theorem with the linear 1-form (y, v) 7→ (v fd (y), f (y)θ(y)iV , giving Z (ν fb, f θiV dy = ∂D

Z

(−ike0 fb, f θiV + (fb, (−ike0 f )θiV +

D

n X (fb, ej f (∂j θ)iV dy j=1

(9.22) Combining and estimating, we get kf k2L2 (∂D) . kf

∧

νkL2 (∂D) kf kL2 (∂D) + kF k2L2 (Dθ ) ,

where Dθ := D ∩ supp θ. The Cauchy integral L2 (∂D) → L2 (Dθ ) : f 7→ F can be shown to be a bounded operator by generalizing the Schur estimates from Exercise 6.4.3 to integral operators from ∂D to Dθ . Moreover, such estimates show by truncation of the kernel that this Cauchy integral is the norm limit of Hilbert–Schmidt operators, and hence compact. On the first term we can use the 1 absorption inequality kN + f kkf k ≤ 2 kN + f k2 + 2 kf k2 . Choosing small leads to + + a lower bound, showing that N : Ek L2 → N + L2 is a semi-Fredholm operator. Next consider the integral equation λh + Ek N h = g, where we need to estimate khkL2 in terms of kgkL2 . To this end, we note that Ek N h = g − λh, so that Ek± L2 3 f ± := 2Ek± N h = N h ± (g − λh). Applying (9.22) to f + and the corresponding application of Stokes’s theorem to f − , we obtain estimates Z ± ± b (ν f , f θiV dy . kF k2L2 (supp θ) . ∂D

We now expand the bilinear expressions on the left, writing f ± = N h ∓ λh ± g, and observe that the integrals Z Z (ν b h, hθiV dy and (νN b h, N hθiV dy ∂D

∂D

are bad in the sense that we have only an upper estimate by khk2L2 , whereas the terms Z Z (νN b h, λhθiV dy = λ (h, hθνiV dy ∂D

∂D

Chapter 9. Dirac Wave Equations

322

c R R and ∂D (νλb h, N hθiV dy = λ ∂D (h, hθνiV dy are good in the sense that they are comparable to khk2L2 . To avoid the bad terms, we subtract identities and obtain Z 2 b+ , f + θiV − (ν fb− , f − θiV dy f kF kL & (ν 2 (supp θ) ∂D Z & 2 Re λ (h, hθνiV dy − khkL2 kgkL2 − kgk2L2 . ∂D

Writing θν = hθ, νi + that |θ2 ∧ ν| ≤ Lhθ, νi for some L < ∞. It R θ ∧ ν, we know follows that 2 Re λ ∂D (h, hθνiV dy & khkL if |λ2 | < |λ1 |/L, and we conclude 2 that in this case, λI + Ek N is a semi-Fredholm operator. That it is a Fredholm operator with index zero follows from the method of continuity, by perturbing λ, for example, to 0, where Ek N is an invertible operator. Theorem 9.5.2 (L2 well-posedness for Ek N ). For the Ek N Dirac BVPs with boundary regularity L2 (∂D) on bounded strongly Lipschitz domains D, we have the following well-posedness results. The four BVPs N ± : Ek± L2 → N ± L2 are well posed for all Im k ≥ 0 except for a discrete set of real k ≥ 0. If the exterior domain D− is connected, then the exterior BVPs N ± : Ek− L2 → N ± L2 are well posed for all nonzero Im k ≥ 0. Proof. By Theorem 9.5.1 and Proposition 9.4.5, the Ek N BVPs are well posed in the Fredholm sense for all k. Proposition 9.4.10 can be verified when the C α topology is replaced by L2 . The proof can now be completed as in Theorem 9.4.11. For the remainder of this section we consider the second problem posed above, namely how to formulate a given Dirac BVP as an integral equation that is good for numerical applications. As a concrete example, we take the exterior BVP with prescribed tangential part, that is, N + : Ek− L2 → N + L2 .

(9.23)

This BVP has important applications, since a solution of it yields an algorithm for computing, for example, how acoustic and electromagnetic waves are scattered by an object D. Assuming that the exterior domain D− is connected and Im k ≥ 0, k 6= 0, we know that this BVP is well posed. Although it is an invertible linear equation by which we can solve the BVP, it is not useful for numerical applications. The reason is that the solution space Ek− L2 is defined by a nonlocal constraint on f ∈ L2 . What we need is an ansatz, meaning some operator U : Y → Ek− L2 , where Y is a function space that is good for numerical purposes and U has good invertibility properties. Using such a U , we can solve the BVP (9.23) by solving N +U h = g

9.5. Integral Equations

323

for h ∈ Y. This gives the solution f = U h ∈ Ek− L2 . As a first try, we swap the roles of Ek and N and consider U = Ek− : N + L2 → Ek− L2 . This leads to the operator N + Ek− |N + L2 , which can be shown to be closely related to the double layer potential operator from Exercise 9.4.6. The function space Y = N + L2 is good, but although this U is a Fredholm operator, it fails to be invertible for a discrete set of real k. Indeed, N(U ) = N + L2 ∩ Ek+ L2 will contain the eigenvalues of the self-adjoint operator −ie0 D with tangential boundary conditions on the bounded domain D+ . This explains a well-known problem in the numerical solution of BVPs by integral equations: the existence of spurious interior resonances k, where the integral equation fails to be invertible, even though the BVP it is used to solve is itself well posed. A better try, which should be more or less optimal, comes from the Ek S BVPs. Swapping the roles of Ek and S, we consider U = Ek− : S + L2 → Ek− L2 . Similarly, a good ansatz for an interior Dirac BVP is U = Ek+ : S − L2 → Ek+ L2 . It is important not to swap S + and S − in these ansatzes. Maybe the best way to see that the Ek S BVPs have well-posedness properties superior to those for the Ek N BVPs on L2 , even in the Fredholm sense and in particular on Lipschitz domains, is to consider the rotation operator Z Ek Sf (x) = 2p.v. x ∈ ∂D. Ψ− k (y − x)f (y)dy, ∂D

2

Note that we used that ν = 1. Since Ek −E is a weakly singular integral operator, it is compact on L2 (∂D), and when k = 0, we note that ES is a skew-symmetric operator, since Ψ is a space-like vector depending skew-symmetrically on x and y. In particular, this means that the spectrum of ES is on the imaginary axis and the operators I ± ES are invertible with k(I ± ES)−1 k ≤ 1. By the identities from Proposition 9.4.5, this means, for example, that kE − hk = 21 k(I − Ek S)hk ≥ 12 khk,

h ∈ S + L2 .

For general k, we note that there is still a major difference in well-posedness properties of the Ek S BVPs as compared to those for the Ek N BVPs. The operator λI + Ek S can fail to be Fredholm only when Re λ = 0, whereas λI + Ek N can fail to be Fredholm whenever | Re λ| ≤ L|Im λ|, not far away from λ = ±1 for large L. So, as compared to the Ek N BVPs, the well-posedness properties for the Ek S BVPs do not essentially depend on the Lipschitz geometry of ∂D. Theorem 9.5.3 (L2 well-posedness for Ek S). For the Ek S Dirac BVPs with boundary regularity L2 (∂D) on bounded strongly Lipschitz domains D, we have the following well-posedness results. The two spin-Dirac BVPs S − : Ek+ L2 → S − L2 and

Chapter 9. Dirac Wave Equations

324

S + : Ek− L2 → S + L2 are well posed for all Im k ≥ 0. Equivalently, the ansatzes Ek− : S + L2 → Ek− L2 and Ek+ : S − L2 → Ek+ L2 are invertible for all Im k ≥ 0. Proof. As before, we note the identity 21 (I − Ek S) = Ek+ S − + Ek− S + and its twin S + Ek− + S − Ek+ = 12 (I − SEk ) = 12 (Ek S − I)SEk . From the discussion above it follows that I − Ek S is a Fredholm operator of index 0, which directly shows that the two BVPs and the two ansatzes are Fredholm maps. By Proposition 9.4.10 adapted to L2 , the four maps are injective for all Im k ≥ 0. Therefore I − Ek N is injective, hence surjective. We conclude that the two BVPs and the two ansatzes are invertible. Example 9.5.4 (Asymptotic APS BVPs). Consider the Cauchy reflection operator A = Ek encoding the Dirac equation DF + ike0 F = 0, together with the abstract boundary conditions B = El , where k, l ∈ C. Clearly not all four Ek El BVPs are well posed, since El − Ek is a compact operator. However, since 1 2 (I

+ El Ek ) = El+ Ek+ + El− Ek−

clearly is a Fredholm operator with index zero, the two BVPs El+ : Ek+ L2 → El+ L2 and El− : Ek− L2 → El− L2 are Fredholm operators. Such BVPs with nonlocal boundary conditions defined by the differential equation itself are essentially the boundary conditions employed by Atiyah, Patodi, and Singer (APS) in their work on index theory for manifolds with boundary. We next let l → ∞ along the upper imaginary axis. The operators El are not norm convergent, but for a fixed function h, one can show that El h → −Sh. Note from the formula for Ψ− l how the singular integral operators El localize to the pointwise multiplier −S. This shows that indeed, the operator S is related to the differential equation as a local asymptotic Cauchy singular integral, and to some extent explains why the Ek S BVPs are so remarkably well posed. Example 9.5.5 (Spin integral equation). We now return to the exterior Dirac BVP (9.23) with prescribed tangential parts, which we know is well posed whenever Im k ≥ 0, k 6= 0, and D− is connected. Using the invertible ansatz Ek− : S + L2 → Ek− L2 from Theorem 9.5.3, we can solve the BVP (9.23), given datum g ∈ N + L2 , by solving N + Ek− h = g (9.24) for h ∈ S + L2 , giving the solution f = Ek− h ∈ Ek− L2 and Z F (x) = Ψ− x ∈ D− , k (y − x)(−ν(y))h(y)dy, ∂D

solving DF + ike0 F = 0 in D− with N + F |∂D = g. This is certainly numerically doable, since both spaces N + L2 and S + L2 are defined by a simple pointwise

9.5. Integral Equations

325

constraint determined by the normal ν. However, we can enhance the integral equation somewhat as follows. Consider the reflection operator T given by T f = −e0 fbe0 . We note that, similarly to N , replacing ν by the time-like vector e0 , indeed T 2 = I and T reflects time-like multivectors in the subspace of space-like multivectors. Computing relevant cosine operators, we have b (T S + ST )f = −e0 (e[ 0 νf )e0 + e0 ν(−e0 f e0 ) = 0, (T N + N T )f = −e0 νf νe0 − νe0 f e0 ν 6= 0, b (N S + SN )f = ν(e[ 0 νf )ν + e0 ν(ν f ν) = 0. By Proposition 9.4.5, this means that we have optimally well posed abstract BVPs T S and N S. In particular, this allows us to parametrize the domain space S + L2 of the integral equation (9.24) for example by T + L2 = L2 (∂D; 4Vc ), is an ideal space for applications. the space of space-like multivector fields, which √ In fact, we verify that S + : T + L2 → S + L2 is 1/ 2 times an isometry. Since T N + N T 6= 0, we cannot directly parametrize the range space N + L2 of (9.24) by T + L2 . However, we can go via the splitting L2 = S + L2 ⊕ S − L2 , since for example, T + S + : N + L2 → T + L2 √ is invertible. In fact, both S + : N + L2 → S + L2 and T + : S + L2 → T + L2 are 1/ 2 times isometries. To summarize, we propose that the exterior BVP (9.23) with prescribed tangential part is best solved using the integral equation T + S + N + Ek− S + h = T + S + g, for h ∈ T + L2 . Indeed, the derivation above shows that this integral equation is uniquely solvable, and the function space for the variable h and the datum T + S + g is simply T + L2 = L2 (∂D; 4Vc ). To write out this equation more explicitly, we compute that T + S + g = 12 (g0 + νg1 ), when g = g0 + e0 g1 and g1 , g2 ∈ N + L2 ∩ T + L2 , so the time-like part is mapped onto a normal part when the original multivector is tangential. We also compute that T + S + N + S + T + = 14 T + . Writing Ek− = 12 (I − Ek ), the integral equation for h ∈ L2 (∂D; 4Vc ) becomes Z 1 + M (x)p.v. Ψ− x ∈ ∂D. h(x) k (y − x)(ν(y) − e0 )h(y)dy = 2M (x)g(x), 2 ∂D

Chapter 9. Dirac Wave Equations

326

Here M denotes the multiplier that projects onto tangential multivectors and maps tangential time-like multivectors onto normal space-like multivectors by replacing a left factor e0 into ν. We refer to this integral equation as a spin integral equation for solving the BVP (9.23), since the key feature is that it uses an ansatz derived from the Ek S BVPs, which, as we have discussed in Example 9.4.9, really are / + ike0 ψ = 0. equation Dψ / BVPs for the 4-Dirac Example 9.5.6 (Transmission problems). Transmission problems generalize boundary value problems in that we look for a pair of fields F + : D+ → 4Wc and F − : D− → 4Wc such that DF + + ik2 e0 F + = 0, in D+ , (9.25) DF − + ik1 e0 F − = 0, in D− , M f + = f − + g, on ∂D. Here the wave numbers k1 , k2 ∈ C are different in the two domains, with Im k1 ≥ 0 and Im k2 ≥ 0. The relation between the traces f + = F + |∂D and f − = F − |∂D on ∂D is described by a multiplier M ∈ L(L2 ) and a given source g ∈ L2 . For solving the transmission problem (9.25), unlike in the case of BVPs, we have a good ansatz directly available, namely U : L2 → Ek+2 L2 ⊕ Ek−1 L2 : h 7→ (Ek+2 h, Ek−1 h). In the case k1 = k2 , it is clear from the L2 analogue of Theorem 9.3.9 that U is invertible. What is somewhat surprising is that U is invertible for all Im k1 ≥ 0 and Im k2 ≥ 0. To prove this, it suffices by the method of continuity to show that U is injective. To this end, note that U h = 0 means that h = F + |∂D = F − |∂D , − − + − where DF + + ik1 e0 F + = 0 in R D and DF + ik2 e0 F = 0 in D . Applying Stokes’s theorem twice to ∂D (e0 νh, hiV dy, computations as in the proof of Proposition 9.4.10 give Z Z Z 2Im k1 |F − |2V dx + |F + |2V dx + 2Im k2 |F − |V2 dx − DR

D+

=

1 2

Z |x|=R

|x|=R

x |( |x| + e0 )F − |2V dx.

Using radiation conditions and jumps, this shows that F + = F − = 0 and therefore h = 0. Using this invertible ansatz U , we can now solve the transmission problem (9.25) by solving the integral equation (M Ek+2 − Ek−1 )h = g

9.6. Boundary Hodge Decompositions

327

for h ∈ L2 . Note that this is an integral equation in L2 (∂D; 4Wc ) without any constraints. From the solution h, we finally compute the field Z Ψk−2 (y − x)ν(y)h(y)dy, F + (x) = ∂D Z − F (x) = − Ψk−1 (y − x)ν(y)h(y)dy, ∂D

solving the transmission problem. In Section 9.7, we apply this integral equation for Dirac transmission problems to solve scattering problems for electromagnetic waves.

9.6

Boundary Hodge Decompositions

We have considered Dirac BVPs in the previous sections and how to solve them by integral equations. Returning to Examples 9.3.7 and 9.3.8, one important issue remains. We saw there that both the Helmholtz equation and Maxwell’s equations can be viewed as special cases of the Dirac equation DF + ike0 F = 0. However, in these examples F is a vector field and a bivector field respectively, and not a general multivector field. If we intend, for example, to solve BVPs for Helmholtz’s or Maxwell’s equations by a spin integral equation as in Example 9.5.5 or a transmission problem with a Dirac integral equation as in Example 9.5.6, then we need a tool to ensure that the solution multivector field F is in fact a vector or bivector field. It turns out that there exists an exterior/interior derivative operator acting on multivector fields ∂D → 4Wc , which we shall denote by Γk , which is the tool needed. Applications to Maxwell scattering are found in Section 9.7. The point of departure for our explanations is Proposition 8.1.5, where we noted that for a monogenic field ∇ 4 F = 0, each of its homogeneous component functions Fj is monogenic if and only if ∇ ∧ F = 0 = ∇ y F . Generalizing this to time-harmonic waves with wave number k ∈ C, we have the following. Lemma 9.6.1 (Two-sided k-monogenic fields). Assume that F : D → 4Wc solves DF + ike0 F = 0 in some open set D ⊂ V . Write F = F0 + F1 + · · · + Fn+1 , where Fj : D → 4j Wc . Then DFj + ike0 Fj = 0 in D for all 0 ≤ j ≤ n + 1 if and only if ( dF + ike0 ∧ F = 0, δF + ike0 y F = 0. The way we use this result is that if we construct F solving DF + ike0 and some BVP, and if dF + ike0 ∧ F = 0, then we can conclude, for example, that F2 is a bivector field solving the Dirac equation, since the homogeneous parts of F decouple, and thus F2 is an electromagnetic field satisfying Maxwell’s equations. Proof. If (∇ + ike0 ) ∧ F = 0 = (∇ + ike0 ) y F , then (∇ + ike0 ) ∧ Fj = ((∇ + ike0 ) ∧ F )j+1 = 0

Chapter 9. Dirac Wave Equations

328

and (∇ + ike0 ) y Fj = ((∇ + ike0 ) y F )j−1 = 0, and so (∇ + ike0 ) 4 Fj = (∇ + ike0 ) y Fj + (∇ + ike0 ) ∧ Fj = 0 for all j. Conversely, if (∇ + ike0 ) 4 Fj = 0 for all j, then (∇ + ike0 ) ∧ Fj = ((∇ + ike0 ) 4 Fj )j+1 = 0 and (∇ + ike0 ) y Fj = ((∇ + ike0 ) 4 Fj )j−1 = 0. Summing over j, we obtain (∇ + ike0 ) ∧ F = 0 = (∇ + ike0 ) y F . To proceed with the analysis, we need to choose a function space. Since our theory for Hodge decompositions as well as for spin integral equations is set in Hilbert spaces, we choose L2 (∂D). Definition 9.6.2 (Boundary Γk operator). Consider the Hardy space splitting L2 (∂D) = Ek+ L2 ⊕ Ek− L2 on a strongly Lipschitz domain. Define the operator Γk by Γk f := g + + g − , where f = Ek+ f + Ek− f , F ± denote the Cauchy integrals of f in D± so that Ek± f = F ± |∂D , and g ± = G± |∂D ∈ Ek± L2 are such that their Cauchy integrals equal G± = (∇ + ike0 ) ∧ F ± in D± . The domain of Γk is the set of f for which such g ± exist. In a series of lemmas, we derive below a more concrete expression for this unbounded operator Γk as a tangential differential operator on L2 (∂D). It turns out that Γk acts by exterior differentiation along ∂D on tangential fields and by interior differentiation along ∂D on normal fields, modulo zero order terms determined by k. Definition 9.6.3 (Tangential derivatives). Consider the Lipschitz boundary M = ∂D, which is a Lipschitz manifold in the sense that the transition maps, as in Section 6.1, are Lipschitz regular. As in Definitions 11.2, 11.2.6, 12.1.1 and extending to Lipschitz regularity as in Section 10.2, we define tangential exterior and interior derivative operators d0 and δ 0 in L2 (M ; ∧M ), such that (d0 )∗ = −δ 0 . In the notation of this chapter, complexifying the bundle ∧M to ∧Mc , we have N + L2 = {f1 + e0 ∧ f2 ; f1 , f2 ∈ L2 (M ; ∧Mc )}, and extending d0 and δ 0 to operators in N + L2 acting as d0 f := d0 f1 − e0 ∧ (d0 f2 ), δ 0 f := δ 0 f1 − e0 ∧ (δ 0 f2 ), on f = f1 + e0 ∧ f2 , with f1 , f2 ∈ L2 (M ; ∧M ). The reader is kindly advised to consult the relevant sections of the following chapters, as indicated in Definition 9.6.3, for further details. Note that the minus sign in the actions on N + L2 occurs because the time-like e0 and the formally space-like tangential ∇0 anticommute.

9.6. Boundary Hodge Decompositions

329

Lemma 9.6.4 (Ek± L2 to N + L2 ). If f ∈ Ek+ L2 ∩ D(Γk ), then N + f ∈ D(d0 ) and d0 (N + f ) + ike0 ∧ (N + f ) = N + (Γk f ). The same holds for f ∈ Ek− L2 ∩ D(Γk ). Proof. Let f = F |∂D , where F is the Cauchy extension of f . Write f = f1 + e0 ∧ f2 and F = F1 + e0 ∧ F2 , where fj and Fj are space-like fields, j = 1, 2. Generalizing Exercise 11.2.3, with methods as in Lemma 10.2.4, to Lipschitz regular hypersurfaces, we have N + fj = ρ∗ Fj , where ρ : ∂D → V denotes the embedding of ∂D into V . The commutation theorem shows that d0 ρ∗ Fj = ρ∗ (dFj ), giving d0 (N + f ) = d0 ρ∗ F1 − e0 ∧ d0 ρ∗ F2 = N + (dF ). This proves the first statement, since e0 ∧ N + f = N + (e0 ∧ f ), and the proof for Ek− L2 is similar. Using Hodge star dualities, we next derive the corresponding result for the normal part. This uses left Clifford multiplication by ν, which is an isometry between N − L2 and N + L2 . Lemma 9.6.5 (Ek± L2 to N − L2 ). If f ∈ Ek+ L2 ∩ D(Γk ), then ν y f = νN − f ∈ D(δ 0 ) and δ 0 (ν y f ) + ike0 y (ν y f ) = ν y (Γk f ). The same holds for f ∈ Ek− L2 ∩ D(Γk ). Proof. Using nabla calculus with ∇k := ∇ + ike0 , given f = F |∂D ∈ Ek+ L2 ∩ D(Γk ), we write for example DF + ike0 F = 0 as ∇k F = ∇k 4 F = 0. Extending Proposition 8.1.13, such solutions form a right Clifford module, so G = F w = F ∗ = F y w, writing w = e012···n for the spacetime volume element, with dual volume element w∗ = −w ∈ 4n W ∗ , is also a solution to ∇k G = 0 in D+ . Moreover, ∗(∇k ∧ G) = −w x (∇k ∧ G) = (−w x G) x ∇k = F x ∇k = ∇k y F , making use of the algebra from Section 2.6. By Lemma 9.6.4 applied to G, we have N + (∇k ∧ G)|∂D = ∇0k ∧ (N + g) with g = G|∂D , writing d0 formally with nabla calculus using ∇0k = ∇0 + ike0 along ∂D. The spacetime Hodge dual of the left-hand side is ∗(N + (∇k ∧ G)|∂D ) = N − (∇k y F )|∂D . For the right hand side, we note for h := ∇0k ∧(N + g) ∈ N + L2 that ∗h = −(νw0 )h = −ν(w0 h) = −ν(w0 x h), where w0 := ν y w. We used here Corollary 3.1.10 and hwi2 = −1. We get ∗(∇0k ∧ (N + g)) = −ν((w0 x N + g) x ∇0k ) = ν((ν y f ) x ∇0k ) = ν∇0k y (f x ν).

Chapter 9. Dirac Wave Equations

330

Note that the first step uses a nonsmooth extension of Exercise 11.2.7. Reversing these two equations, multiplying them from the left by ν, and equating them yields ν y (∇k y F )|∂D = ν(∇k0 y (f x ν))ν = ∇k0 y (fb x ν) = −∇0k y (ν y f ). In the second step we used that νh = b hν whenever h ∈ N + L2 , and in the last step we applied the commutation relation from Proposition 2.6.3. This proves the lemma for Ek+ L2 , since Γk f = −(∇k0 y F )|∂D . The proof for Ek− L2 is similar. We next show the converses of Lemmas 9.6.4 and 9.6.5. Lemma 9.6.6 (N ± L2 to Ek± L2 ). If f ∈ N + L2 and f ∈ D(d0 ), then f ∈ D(Γk ) with Γk f = d0 f + ike0 ∧ f. Similarly, if f ∈ N − L2 and ν y f ∈ D(δ 0 ), then f ∈ D(Γk ) with Γk f = ν ike0 y)(ν y f ). Proof. Let f ∈ N + L2 and define Cauchy extensions Z 0 F ± (x) = ± Ψ− k (y − x)ν(y)f (y)dy,

∧

(δ 0 +

x ∈ D± .

Differentiating under the integral sign, with notation as in the proof of Lemma 9.6.5, we have Z ∇k ∧ F (x) = ∓ ˙ Ψk− (y − x)(∇ k ∧ ν(y) ∧ f (y))dy ∂D Z =± Ψ− k (y˙ − x)(∇−k ∧ ν(y) ∧ f (y))dy, ∂D

where we have used the algebraic anticommutation relation − − ∇k ∧ (Ψ− k h) = (∇k , Ψk ih − Ψk (∇k ∧ h) − and the first term vanishes, since Ψ+ k (·−y) = −Ψk (y −·) is a fundamental solution to D + ike0 . Aiming to apply a nonsmooth extension of Exercise 11.2.7, we form the inner product with a fixed multivector w ∈ 4Wc , and obtain Z hw, ∇k ∧ F (x)i = ± hν(y) y (∇k y (Ψk+ (y˙ − x)w)), f (y)idy. ∂D

We choose to use the complex bilinear pairing on 4Wc , but this is not important. By Lemma 9.6.5, we have ν(y)y(∇k y(Ψk+ (y˙ −x)w)) = −(δ 0 +ike0 y)(ν(y)y(Ψk+ (y − x)w)). Note that F in the proof of Lemma 9.6.5 need not solve a Dirac equation for such a trace result to be true. Duality yields Z 0 hw, Ψ− hw, ∇k ∧ F (x)i = ± k (y − x)ν(y)(d f + ike0 ∧ f )idy. ∂D

9.6. Boundary Hodge Decompositions

331

Since w is arbitrary, this proves the lemma for N + L2 . The proof for f ∈ N − L2 is similar. We calculate Z ∇k ∧ F (x) = −∇k y F (x) = ∓ Ψk− (y˙ − x)(∇−k y (ν(y) y f (y)))dy. ∂D

Pairing with w gives Z hw, ∇k ∧ F (x)i = ∓ h∇k ∧ (Ψ+ k (y˙ − x)w), ν(y) y f (y)idy ∂D Z h(d0 + ike0 ∧)N + (Ψk+ (y˙ − x)w), ν(y) y f (y)idy =∓ Z∂D hw, Ψk− (y − x)ν(y)(ν(y) ∧ ((δ 0 + ike0 y)(ν(y) y f (y))))idy. =± ∂D

The second equality uses that ν yf ∈ N + L2 and Lemma 9.6.4. Since w is arbitrary, this proves the lemma for N − L2 . Summarizing the above results, we obtain the following concrete expression for Γk . Given Lemmas 9.6.4, 9.6.5, and 9.6.6, the proof is straightforward. Proposition 9.6.7. The operator Γk is a nilpotent operator in L2 (∂D) in the sense of Definition 10.1.1. Its domain equals D(Γk ) = {f ∈ L2 (∂D) ; N + f ∈ D(d0 ) and ν y f ∈ D(δ 0 )} and Γk f = (d0 + ike0 ∧)N + f + ν ∧ (δ 0 + ike0 y)(ν y f ),

f ∈ D(Γk ).

The operator Γk commutes with Ek and with N . Having uncovered this nilpotent operator Γk , we now investigate the Hodge splitting of L2 (∂D) that it induces. We need a Hermitian inner product on L2 (∂D), and we choose Z (f (x), g(x)iV dx.

(f, gi = ∂D

Proposition 9.6.8 (Boundary Hodge decomposition). When k 6= 0, the nilpotent operator Γk induces an exact Hodge splitting L2 (∂D) = R(Γk ) ⊕ R(Γ∗k ), where the ranges R(Γk ) = N(Γk ) and R(Γk∗ ) = N(Γk∗ ) are closed. When k = 0, the ranges are still closed, but the finite-dimensional cohomology space H(Γk ) = N(Γk ) ∩ N(Γ∗k ) will be nontrivial.

Chapter 9. Dirac Wave Equations

332

Proof. Proposition 10.1.2 shows that Γ = Γk induces an orthogonal splitting L2 (∂D) = R(Γk ) ⊕ H(Γk ) ⊕ R(Γ∗k ). When D is smooth and k = 0, it follows from Propositions 12.1.3 and 10.1.6 that the ranges are closed and that the cohomology space is finite-dimensional for Γ = Γk . Adapting the methods from Theorem 10.3.1 to the manifold setting, this result can be extended to the case that D is merely a Lipschitz domain. However, on nonsmooth boundaries ∂D we do not have D(d0 ) ∩ D(δ 0 ) = H 1 (∂D), but still D(d0 ) ∩ D(δ 0 ) is compactly embedded in L2 (∂D). Assume next that k 6= 0 and define the nilpotent operator µf = ike0 ∧ N + f + ν ∧ (ike0 y (ν y f )) so that Γk = Γ + µ. We compute µ∗ f = ik c e0 y N + f + ν ∧ (ik c e0 ∧ (ν y f )). As in Example 10.1.7, we note that N(µ) ∩ N(µ∗ ) = {0}. Consider the abstract Dirac operators Γk + Γ∗k = (Γ + Γ∗ ) + (µ + µ∗ ) : D(Γ) ∩ D(Γ∗ ) → L2 . Since Γ + Γ∗ : D(Γ) ∩ D(Γ∗ ) → L2 is a Fredholm operator and µ + µ∗ : D(Γ) ∩ D(Γ∗ ) → L2 is a compact operator, it follows from Proposition 10.1.6 that Γk +Γ∗k : D(Γ) ∩ D(Γ∗ ) → L2 is a Fredholm operator. Thus the ranges are closed. To prove that in fact the cohomology space N(Γk ) ∩ N(Γ∗k ) in fact is trivial, we note that Γµ∗ + µ∗ Γ = 0. Thus, if Γf + µf = 0 = Γ∗ f + µ∗ f , then 0 = (f, (Γµ∗ + µ∗ Γ)f iV = (Γ∗ f, µ∗ f iV + (µf, Γf iV = −kµ∗ f k2 − kµf k2 . This shows that f = 0 and completes the proof.

Exercise 9.6.9. Roughly speaking, Γ∗k acts as interior derivative on N + L2 and as exterior derivative on N − L2 . Write down the details of this, and show that Γ∗k commutes with N , but not with Ek in general.

9.7

Maxwell Scattering

In this section, we demonstrate how classical Helmholtz and Maxwell boundary value and transmission problems can be solved using the operators Ek , N , and Γk . Recall that Ek is the reflection operator for the Hardy space splitting from Theorem 9.3.9, that N is the reflection operator for the splitting into normal and

333

9.7. Maxwell Scattering

tangential fields from Example 9.4.8, and that Γk is the nilpotent operator for the boundary Hodge decomposition from Proposition 9.6.8. The basic operator algebra is that Ek2 = N 2 = I, Γ2k

= 0,

6 N Ek , Ek N =

Γk Ek = Ek Γk ,

Γk N = N Γk .

The Ek N BVPs are essentially well posed, so by Proposition 9.4.5, roughly speaking Ek and N are closer to anticommuting than to commuting.

Figure 9.3: A rough sketch of the splittings involved in a Dirac BVP. The splitting into Ek± H encodes the Dirac equation. The splitting into N ± H encodes the boundary conditions. The circle indicates the boundary Hodge splitting, with the interior of the circle illustrating N(Γk ) where the Maxwell BVP takes place. We note that the operator S from Example 9.4.9 does not commute with Γk , but we will not need this, since we use only S as a computational tool for solving an Ek N BVP.

Chapter 9. Dirac Wave Equations

334

We consider as an example the exterior Dirac BVP (9.23) with prescribed tangential part. The other three Ek N BVPs can be analyzed similarly. Using the operator Γk , we have three relevant L2 (∂D; 4Wc )-based function spaces, namely H = L2 ,

H = D(Γ),

and H = N(Γk ).

Note that D(Γ) = D(Γk ) is a dense subspace of L2 , which does not depend on k, although the equivalent norms kf kD(Γk ) = (kf k2L2 + kΓk f k2L2 )1/2 do depend on k. Further note that N(Γk ) is a closed subspace of L2 , as well as of D(Γ), which is roughly speaking half of the latter spaces by Hodge decomposition. Since Ek and N commute with Γk , they act as bounded linear operators in each of the three function spaces H, and in each case we see that Ek2 = N 2 = I. Therefore we can consider the BVP (9.23), expressed as the restricted projection N + : Ek− H → N + H, in each of the three function spaces H. Our aim in this section is to solve BVPs in H = N(Γk ). This, however, is a function space defined by a differential constraint, which we may want to avoid numerically. For this reason, we prefer to enlarge the function space to either L2 or to the function space D(Γ) in which roughly speaking half of the functions have Sobolev regularity H 1 , since Γk is nilpotent, and to solve the integral equation in such a space. Proposition 9.7.1 (Constrained Dirac BVPs). Consider the exterior Dirac BVP N + : Ek− L2 → N + L2 with prescribed tangential part at ∂D, and assume that Im k ≥ 0 and k 6= 0, so we have L2 well-posedness of this BVP by Theorem 9.5.2. Then the restricted map N + : Ek− H → N + H is also invertible for each of the function spaces H = D(Γ) and H = N(Γk ). For the solution f ∈ Ek− L2 to the BVP with datum g = N + f ∈ L2 , the following holds. If g ∈ D(Γ), then f ∈ D(Γ). If Γk g = 0, then Γk f = 0. If Γk g = 0 and g is a j-vector field, then f is a j-vector field. Note that if g ∈ N + L2 is a j-vector field, then in general the solution f ∈ to the BVP will not be a homogeneous j-vector field. The constraint Γk g = 0 is crucial. Ek− L2

Proof. (i) Lower bounds for N + : Ek− D(Γ) → N + D(Γ)

(9.26)

hold, since kf kD(Γ) ≈ kf kL2 + kΓk f kL2 . kN + f kL2 + kN + Γk f kL2 = kN + f kL2 + kΓk (N + f )kL2 ≈ kN + f kD(Γ) .

9.7. Maxwell Scattering

335

To show surjectivity, we can proceed as follows. First apply Lemma 9.4.12 with A = Ek , B = N , H = D(Γ) and perturb k into Im k > 0. This shows that it suffices to show surjectivity for Im k > 0. Then we use that N and Ek commute with Γk , and similarly to the above, we derive lower bounds for λI + Ek N : D(Γ) → D(Γ), when |λ1 | > L|λ2 |, from Theorem 9.5.1. Therefore the method of continuity shows that I ± Ek N are Fredholm operators of index zero on D(Γ). The argument in Proposition 9.4.10 shows that all four Ek N BVPs are injective when Im k > 0, and so it follows from (9.19) that (9.26) is surjective. If f ∈ Ek− L2 solves the BVP with datum g ∈ D(Γ), then let f˜ ∈ Ek− D(Γ) be the solution to the well-posed BVP described by (9.26). By uniqueness of the solutions to the L2 BVP, we conclude that f = f˜ ∈ D(Γ). (ii) Next consider N + : Ek− N(Γk ) → N + N(Γk ). This map is clearly bounded and injective with a lower bound. To show surjectivity, let g ∈ N + N(Γ) ⊂ N + D(Γ). By (i) there exists f ∈ Ek− D(Γ) such that N + f = g. Since N + (Γk f ) = Γk (N + f ) = Γk g = 0, it follows from L2 well-posedness that f ∈ Ek− N(Γk ). If furthermore g ∈ N(Γk ) is a j-vector field, then the solution f satisfies Γk f = 0, and we conclude from Lemma 9.6.1 that each homogeneous component function fm belongs to Ek+ L2 . Since N + fm = gm = 0 if m 6= j, it follows in this case that fm = 0 by uniqueness of solutions to the BVP. Therefore f = fj is a j-vector field. Example 9.7.2 (Helmholtz BVPs). In Example 9.3.7 we saw how the Helmholtz equation for a scalar acoustic wave u is equivalent to the vector field F = ∇u + ike0 u solving the Dirac equation DF + ike0 F = 0. (i) The Neumann BVP for u amounts to specifying the normal part N − f = (∂ν u)ν of f = F |∂D . In this case, by Proposition 9.6.7 the condition Γk (N − f ) = 0 is automatic for a vector field f , since ∧−1 W = {0}. Therefore, solving the Dirac BVP for F with this prescribed datum on ∂D will produce a vector field F according to Proposition 9.7.1. From Proposition 9.6.8 it follows that F ∈ R(Γk ), which means that there exists a scalar function u such that F = ∇u + ike0 u. In particular, u solves ∆u + k 2 u = 0 with prescribed Neumann datum. (ii) The Dirichlet BVP for u amounts to specifying the tangential part N + f = ∇0 u + ike0 u of f = F |∂D . For a given tangential vector field g = g1 + e0 g0 , where g1 is a space-like vector field and g0 is a scalar function, we note that Γk g = ∇0 ∧ g1 + e0 ∧ (−∇0 g0 + ikg1 ), so g ∈ N(Γk ) amounts to ikg1 = ∇0 g0 .

336

Chapter 9. Dirac Wave Equations

Therefore, solving the Dirac BVP for F with such a tangential vector field g ∈ N(Γk ) on ∂D as datum will produce a vector field of the form F = ∇u + ike0 u by Proposition 9.7.1, where u solves the Helmholtz Dirichlet problem. Example 9.7.3 (Maxwell BVPs). In Example 9.3.8 we saw how Maxwell’s equations 1/2 for an electromagnetic wave F are equivalent to the bivector field F = 0 e0 ∧ −1/2 E + µ0 B solving the Dirac equation DF + ike0 F = 0. We now assume that the interior domain D+ ⊂ R3 is a perfect electric conductor, so that E = B = 0 in D+ . If Maxwell’s equations are to hold in the distributional sense in all R3 , by the vanishing right-hand sides in the Faraday and magnetic Gauss laws, we need N + f = 0 for f = F |∂D . If the electromagnetic wave in D− is the superposition of an incoming wave f0 and a reflected wave f1 ∈ Ek− L2 , then f1 needs to solve the BVP where N + f1 is specified to cancel the datum N + f0 . Note that for the classical vector fields E and ∗B, the tangential part N + f corresponds to the tangential part of E and the normal part of ∗B. For a given tangential bivector field g = e0 ∧ g1 + g2 , where g1 is a space-like tangential vector field and g2 is a space-like tangential bivector field, we note that Γk g = e0 ∧ (−∇0 ∧ g1 + ikg2 ) + ∇0 ∧ g2 , so g ∈ N(Γk ) amounts to ikg2 = ∇0 ∧ g1 . In terms of the electric and magnetic fields, the tangential part of B is given by the tangential curl of the tangential part of E. From Proposition 9.7.1 it follows that if we solve the Dirac BVP (9.23) with such a tangential bivector field g ∈ N(Γk ) on ∂D as datum, then the solution f will indeed be a bivector field representing an electromagnetic field. Example 9.7.4 (Maxwell transmission problems). When an electromagnetic wave propagates in a material and not in vacuum, we account for the material’s response to the field by replacing 0 and µ0 by permittivity and permeability constants and µ depending on the material properties. These may in general be variable as well as matrices, but we limit ourselves to homogeneous and isotropic materials for which and µ are constant complex numbers. Similar to (9.3), we define the electromagnetic field F := 1/2 e0 ∧ E + µ−1/2 B. √ Maxwell’s equations in such a material read DF + ike0 F = 0, with k = ω µ. Consider the following transmission problem. We assume that the exterior domain D− consists of a material with electromagnetic properties described by 1 , √ µ1 , giving a wave number k1 := ω 1 µ1 , and that the interior domain D+ consists of a material with electromagnetic properties described by 2 , µ2 , giving a wave √ number k2 := ω 2 µ2 . We obtain a transmission problem of the form (9.25) for a pair of electromagnetic fields F ± : D± → 42 Wc .

337

9.7. Maxwell Scattering 1.5

0.5 0.4

1

0.3 0.5

0.1

1

0.2

0.5

0.1

0

0

0

-0.1

0

-0.1 -0.5

-0.2 -0.3

-0.5

-0.2 -0.3

-1

-0.4 -0.5 0

0.2

0.4

0.6

0.8

1

-1.5

0.4

1

0.3

-1

-0.4 -0.5 0

(b) 1.5

0.5

0.2

0.4

0.6

0.8

1

-1.5

1.5

0.5 0.4

1

0.3

0.2

0.5

0.1

0.2

0.5

0.1

0

0

0

-0.1

0

-0.1 -0.5

-0.2 -0.3

-0.5

-0.2 -0.3

-1

-0.4 -0.5

(c)

0.4 0.3

0.2

(a)

1.5

0.5

0

0.2

0.4

0.6

0.8

1

-1.5

-1

-0.4 -0.5

(d)

0

0.2

0.4

0.6

0.8

1

-1.5

Figure 9.4: TM magnetic waves U = B12 . ∂Ω parametrized√by sin(πs) exp(i(s−1/2) π/2), 0 ≤ s ≤ 1. (a) Incoming wave U0 = exp(18i(x + y)/ 2) from south-west. (b) Wave reflected by a perfect electric conductor, computed with the spin integral equation in Example 9.5.5. (c) Waves reflected into Ω− and transmitted into a dielectric object Ω+ , computed with a tweaked version of the Dirac integral equation in Example 9.5.6. Wave numbers k1 = 18 and k2 = 27 as in Example 9.7.4. (d) As in (c), but Ω+ is now a conducting √ object described by the Drude model and an imaginary wave number k2 = i18 1.1838. Here the wave decays exponentially into Ω+ and surface plasmon waves, excited by the corner singularity, appear near ∂Ω. The jump condition M f + = f − + g is found by returning to the original formulation of Maxwell’s equations for E and B. For these to hold in the distributional sense across ∂D, Faraday’s law and the magnetic Gauss laws dictate that ν ∧ E and ν ∧ B do not jump across ∂D. Furthermore, assuming that we do not have any electric charges and current except for those induced in the material described by and µ, the Amp`ere and Gauss laws require that ν y(µ−1 B) and ν y(E) not jump across ∂D. If we translate this to spacetime multivector algebra, this specifies the multiplier q q q q M = µµ21 N + T + + 12 N + T − + µµ12 N − T + + 21 N − T − ,

Chapter 9. Dirac Wave Equations

338 0.5

3

0.5

0.4 2

0.3

-13

0.3

0.2

1

0.1

-13.5

0.2 0.1

0

0

-0.1 -1

-0.2

-14

0 -0.1

-14.5

-0.2

-0.3

-2

-0.4

-0.3

-15

-0.4 -3

-0.5 0

(d)

0.4

0.2

0.4

0.6

0.8

1

-15.5

-0.5 0

(b)

0.2

0.4

0.6

0.8

1

0.5

0.5 0.4

-13

-13

-13.5

-13.5

0.3 0.2 0.1 -14

0 -0.1

-14

0

-14.5

-14.5

-15

-15

-0.2 -0.3 -0.4 -15.5

-0.5

(c)

0

0.2

0.4

0.6

0.8

1

-15.5

-0.5

(d)

0

0.2

0.4

0.6

0.8

1

Figure 9.5: Upper left (d) is same as Figure 9.4(d), but scaled so the peaks of the plasmon wave are visible. (b), (c) and (d) show log10 of the estimated absolute error for the three scattering computations. (d) indicates the numerical challenge in computing surface plasmon waves. Here the parameters hit the essential spectrum, where the integral equation fails to be Fredholm. using the normal reflection operator N and the time reflection operator T . Note how the two commuting reflection operators N and T split the electromagnetic field into these four parts. With this formulation, and with the datum g being the boundary trace g = F0 |∂D of an incoming electromagnetic wave F0 in D− , we can use the Dirac integral equation proposed in Example 9.5.6 to compute the transmitted wave F + in D+ and the reflected wave F − in D− . We end this chapter with some examples of how the integral equations from Examples 9.5.5 and 9.5.6 perform numerically when applied to scattering problems for electromagnetic fields as in Examples 9.7.3 and 9.7.4. Results are shown in Figures 9.4 and 9.5. For simplicity, we consider a two-dimensional scattering problem in which the object represented by the domain D+ ⊂ R3 is a cylinder D+ = Ω+ × R

9.8. Comments and References

339

along the z-axis over the base Ω+ ⊂ R2 = [e12 ] in the xy-plane, and the field is transversal magnetic. This means that we assume that √ F = F (x, y) = e0 ∧ (E1 (x, y)e1 + E2 (x, y)e2 ) + √1µ B12 (x, y)e12 . In classical vector calculus notation this means that E is parallel to R2 and the 2 vector field ∗B is orthogonal √ to R , explaining the terminology. Maxwell’s equations, after dividing F by , read (∇ + ike0 )(e0 E + cB) = (c∇B − ikE) + e0 (−∇E + iωB) = 0, where ∇ = e1 ∂1 + e2 ∂2 is the nabla symbol for R2 . From the space- and time-like parts of this equation, we get ∆B = ∇(∇B) = (ik/c)∇E = (ik/c)iωB = −k 2 B, that is, U := B12 solves the Helmholtz equation, and E = (c/ik)∇B = (c/ik)(∇U )e12 . This means that Maxwell’s equations for transversal magnetic fields F are equivalent to the Helmholtz equation for U = B12 and that E is obtained from the gradient ∇U by rotation and scaling. In particular, it follows that for transversal magnetic fields F , the tangential boundary datum N + f corresponds to the Neumann data ∂ν U for U .

9.8

Comments and References

9.2 Building on the work of Michael Faraday and Andr´e-Marie Amp`ere, James Clerk Maxwell (1831–1879) collected and completed the system of equations governing electromagnetic theory in the early 1860s. His Treatise on Electricity and Magnetism was published in 1873. The equations that he obtained showed that electric and magnetic fields propagate at the speed of light, and they were relativistically correct decades before Einstein formulated relativity theory. The fundamental equation of quantum mechanics, the Schr¨odinger equation from Example 6.3.6, was first discovered in 1925 and describes physics at small scales. The famous Stern–Gerlach experiment from 1922 showed that the intrinsic angular momentum of particles is quantized. The Pauli equation from 1927 is a modification of the Schr¨odinger equation that takes this spin phenomenon into account, but neither of these is the correct equation at high speeds, that is, they are not relativistically correct. The Klein–Gordon equation from 1926 is a relativistically correct version of the Schr¨odinger equation, but it does not incorporate spin. Paul Dirac finally succeeded in

340

Chapter 9. Dirac Wave Equations 1928 in finding the equation that is correct from the point of view of both quantum mechanics and relativity theory, as well as correctly describing spin1/2 particles, which include all the elementary particles constituting ordinary matter. The classical derivation of the Dirac equation is to seek matrices γ0 , γ1 , γ2 , γ3 by which one can factorize the Klein–Gordon equation into a first-order wave equation. This amounts to using a matrix representation of the spacetime Clifford algebra, something that the pioneers of quantum mechanics were unaware of. Starting from the 1960s there has been a renewed interest in Clifford’s geometric algebra, where in particular, David Hestenes [55], Hestenes and Sobczyk [57], and Hestenes [56] have advocated geometric algebra as the preferred mathematical framework for physics. In particular, [55] is a reference for using Clifford algebra to study Maxwell’s and Dirac’s equations. The formulations (9.4) and (9.5) of Maxwell’s equations as wave 4-Dirac equations go back to M. Riesz. A further reference for the use of multivectors in electromagnetic theory is Jancewicz [60]. A standard mathematics reference for the analysis of Dirac’s equation is Thaller [93]. Further references on Dirac operators and spinors in physics include Benn and Tucker [19] and Hitchin [58].

9.3-9.6 The material covered in these sections, which aim to solve Maxwell BVPs using multivector calculus, builds on the author’s PhD thesis and publications [8, 9, 7, 14, 10]. The first basic idea for solving boundary value problems for Maxwell’s equations is to embed it into a Dirac equation as in Example 9.3.8. This was first used by McIntosh and M. Mitrea in [67] in connection with BVPs on Lipschitz domains. The second basic idea is to formulate Dirac boundary value problems in terms of Hardy projections Ek± and projections N ± encoding boundary conditions, and to show that these subspaces are transversal. This was first worked out by Axelsson, Grognard, Hogan, and McIntosh [11]. The third main idea is to extract a Maxwell solution from the Dirac solution as in Proposition 9.7.1, using the Hodge decomposition on the boundary defined by the operator Γk from Section 9.6. This was worked out in detail in [9]. We have chosen to use the spacetime formulation, but as in Propositions 9.1.5 and 9.1.6, we can equally well use a 4V formulation in which the Dirac equation reads DF = ikF for F : D → 4Vc . The main reason for our choice is that the operator Γk in Section 9.6 is difficult, although not impossible, to handle using the latter formalism. To minimize the algebra, the 4Vc formulation was used in [84, 80], where the spin integral equation from Example 9.5.5 was first introduced.

9.8. Comments and References

341

A main philosophy in [9] and associated publications is to handle the boundary value problems by first-order operators. It is clear what this means for the differential operators: in (9.10) the second-order Helmholtz operator is factored by the first-order Dirac operator. But we also have corresponding factorizations of the boundary integral operators. In the abstract formulation with Proposition 9.4.5, the second-order cosine operator is factored by the first-order rotation operator in (9.20)–(9.21). We think of the rotation operators being as of first order, since they essentially are direct sums of two restricted projections as in (9.18)–(9.19). Similarly, the cosine operator can be seen to be essentially the direct sum of compositions of two restricted projections, hence of second order. A reference for Bessel functions and Exercise 9.3.3 is Watson [95]. Standard references for the classical double and single layer potential integral equations are Colton and Kress [29, 30] and Kress [62]. The method to prove semi-Fredholm estimates of singular integral operators on Lipschitz domains as in Theorem 9.5.1 using Stokes’s theorem and a smooth transversal vector field as in Exercise 6.1.8 goes back to Verchota [94]. The spectral estimates in Theorem 9.5.1 are from [7]. 9.7 Figures 9.4 and 9.5 have been produced by Johan Helsing using the spin and tweaked Dirac integral equations. The state-of-the-art numerical algorithm RCIP, recursively compressed inverse preconditioning that he uses is described in [51], with applications to Helmholtz scattering in [52] and [53]. Since the Dirac equation is more general than the Helmholtz and Maxwell equations that it embeds, the spin and Dirac integral equations cannot quite compete with the most efficient Kleinman–Martin type integral equation [53, eq. 45] in terms of computational economy. In terms of achievable numerical accuracy in the solution, however, the two systems of integral equations perform almost on par with each other. Moreover, the spin and Dirac integral equations apply equally well to Maxwell scattering in three dimensions, where the present understanding of integral formulations for Maxwell’s equations is incomplete.

Chapter 10

Hodge Decompositions Prerequisites: The reader is assumed to have read Sections 7.5 and 7.6, which this chapter develops further. A good understanding of unbounded Hilbert space operators and the material in Section 6.4 is desirable. Some exposure to distribution theory and algebraic topology helps, but is not necessary. Road map: We saw in Section 7.6 that every multivector field F on a domain D can be decomposed into three canonical parts F = ∇ ∧ U + H + ∇ y V, where ∇ ∧ H = 0 = ∇ y H, and H and the potential V are tangential on ∂D. This is the Hodge decomposition of the multivector field F , which amounts to a splitting of the space of all multivector fields F into two subspaces R(d) and R(δ ) of exact and coexact fields respectively, and a small subspace Ck (D) of closed and coclosed fields, all with appropriate boundary conditions. Alternatively, we can instead demand that H and the potential U be normal on ∂D. At least four types of questions arise. (i) Are the subspaces R(d) and R(δ) transversal, that is do they intersect only at 0 and at a positive angle? This would mean that these subspaces give a splitting of the function space H that we consider, modulo Ck (D). In the case of H = L2 (D) which we only consider here, these subspaces are in fact orthogonal, but more generally this problem amounts to estimating singular integral operators realizing the Hodge projections onto these subspaces. We touch on this problem in Proposition 10.1.5 and Example 10.1.8. (ii) Are the ranges R(d) and R(δ) closed subspaces? This is a main problem that we address in this chapter, and we show that this is indeed the case for © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_10

343

344

Chapter 10. Hodge Decompositions

bounded domains D. See Section 10.3 and Example 10.1.8. We saw in Section 7.6 that such closedness yields well-posedness results for boundary value problems. (iii) What properties, in particular regularity, of the potentials U and V do we have? Note that the parts ∇ ∧ U and ∇ y V are uniquely determined by F , but not so for the potentials U and V . We show in Section 10.4 that the most obvious choice, the Hodge potentials of minimal L2 norm, are not always the best choices. Even more surprising is the fact that there exist Bogovski˘ı potentials V , for which we have full Dirichlet boundary conditions V |∂D = 0. (iv) Is the cohomology space Ck (D) finite-dimensional? More exactly, how do we go about calculating the dimension of this subspace for a given domain D? As compared to the first three questions, which belong to analysis, this fourth question belongs to algebraic topology and is addressed in Section 10.6. In the analysis of Hodge decompositions on domains, the regularity and curvature of the boundary play an important role through Weitzenb¨ock formulas. Hodge decompositions can also be considered on manifolds, and in this case also the curvature of the manifold in the interior of the domain enters the picture. This will be a central idea in Chapters 11 and 12. In the present chapter we avoid the technicalities of vector bundles and limit the discussion to domains in affine spaces. Highlights: • Compactness and Hodge decomposition: 10.1.6 • Natural boundary conditions for d and δ: 10.2.3 • Weitzenb¨ock boundary curvature: 10.3.6 • Bogovski˘ı and Poincar´e potentials: 10.4.3 ˇ • Cech computation of Betti numbers: 10.6.5

10.1 Nilpotent operators In terms of operators, a splitting of a function space corresponds to a projection P , along with its complementary projection I − P . Somewhat similarly, we show in this section how Hilbert space operators Γ with the property that Γ2 = 0 induce splittings of the function space in a natural way, generalizing Hodge decompositions. Usually the condition Γk = 0 for some k ∈ Z+ defines nilpotence, but we shall always assume index k = 2. Definition 10.1.1 (Nilpotent). A linear, possibly unbounded, operator Γ : H → H in a Hilbert space H is said to be nilpotent (with index 2) if it is densely defined, closed, and if R(Γ) ⊂ N(Γ). In particular, Γ2 f = 0 for all f ∈ D(Γ). We say that a nilpotent operator Γ is exact if R(Γ) = N(Γ).

345

10.1. Nilpotent operators

Recall that the null space N(Γ) is always closed if Γ is closed but that in general, the range R(Γ) is not a closed subspace. If Γ is nilpotent, then we have inclusions R(Γ) ⊂ R(Γ) ⊂ N(Γ) ⊂ D(Γ) ⊂ H. Let H0 denote any closed subspace complementary to N(Γ), for example H0 = N(Γ)⊥ , so that H = N(Γ) ⊕ H0 . Then the restricted map Γ : H0 → R(Γ) ⊂ N(Γ) is injective, which roughly speaking means that N(Γ) is at least half of H. For this reason it is natural to combine a nilpotent operator Γ1 with a “complementary” nilpotent operator Γ2 . Ideally one would like to have a splitting of the Hilbert space H = R(Γ1 ) ⊕ R(Γ2 ), where R(Γ1 ) = N(Γ1 ) and R(Γ2 ) = N(Γ2 ). Since N(Γ∗1 ) = R(Γ1 )⊥ , the natural choice in a Hilbert space is Γ2 = Γ∗1 . Proposition 10.1.2 (Abstract Hodge decomposition). Let Γ be a nilpotent operator in a Hilbert space H. Then so is Γ∗ , and there is an orthogonal splitting into closed subspaces H = R(Γ) ⊕ C(Γ) ⊕ R(Γ∗ ), (10.1) where C(Γ) := N(Γ) ∩ N(Γ∗ ), N(Γ) = R(Γ) ⊕ C(Γ),

and

∗

N(Γ ) = C(Γ) ⊕ R(Γ∗ ). Note that C(Γ) = {0} if and only if Γ is exact. Proof. If T is a densely defined and closed operator in H, then R(T )⊥ = N(T ∗ ) and therefore R(T ) = N(T ∗ )⊥ . This proves that R(Γ∗ ) = N(Γ)⊥ ⊂ R(Γ)⊥ = N(Γ∗ ), showing that Γ∗ is nilpotent and that we have orthogonal splittings H = N(Γ) ⊕ R(Γ∗ ) = R(Γ) ⊕ N(Γ∗ ). But R(Γ) ⊂ N(Γ), since Γ is nilpotent, so using the second splitting in the first, we get ⊥ N(Γ) = R(Γ) ⊕ N(Γ) ∩ R(Γ) = R(Γ) ⊕ N(Γ) ∩ N(Γ∗ ) , which proves the stated splitting.

The mapping properties of Γ and Γ∗ are as follows. In the Hodge decomposition (10.1), the operator Γ is zero on R(Γ) ⊕ N(Γ) ∩ N(Γ∗ ) = N(Γ), and Γ∗ is zero on N(Γ∗ ) ∩ N(Γ) ⊕ R(Γ∗ ) = N(Γ∗ ). On the other hand, we see that the

Chapter 10. Hodge Decompositions

346

restrictions Γ : R(Γ∗ ) → R(Γ) and Γ∗ : R(Γ) → R(Γ∗ ) are injective and have dense ranges: H

H

=

=

⊕

R(Γ)

R(Γ)

s

N(Γ) ∩ N(Γ∗ )

Γ∗

Γ

⊕

∗

N(Γ) ∩ N(Γ )

R(Γ∗ )

⊕

⊕

+

(10.2)

R(Γ∗ )

We have been using the formally skew-adjoint Dirac operator D = d + δ in Chapters 8 and 9. Using instead the anti-Euclidean Clifford product leads to a formally self-adjoint Dirac operator d − δ. For the following results we can use either the abstract Dirac operator Γ − Γ∗ or its self-adjoint analogue Γ + Γ∗ . To be able to use resolvents without complexifying the space, we choose to work with Γ − Γ∗ . Note from the mapping properties of Γ and Γ∗ that such operators swap the subspaces R(Γ) and R(Γ∗ ). Proposition 10.1.3 (Abstract Hodge–Dirac operators). Let Γ be a nilpotent operator in a Hilbert space H. Consider the operator Π := Γ − Γ∗ with domain D(Π) := D(Γ) ∩ D(Γ∗ ). Then Π is skew-adjoint, that is, Π∗ = −Π in the sense of unbounded operators, with N(Π) = C(Γ) and R(Π) = R(Γ) + R(Γ∗ ). We refer to operators Π = Γ − Γ∗ , derived from a nilpotent operator Γ, as an abstract Hodge–Dirac operator. Note that in Euclidean spaces, the 4-Dirac operator D from Definition 9.1.1 is an example of a Hodge–Dirac operator, whereas / from Definition 9.1.3 as a Hodge–Dirac operator to have the 4-Dirac / operator D requires a complex structure on our Euclidean space, as discussed at the end of Section 9.1. Proof. We use the Hodge decomposition from Proposition 10.1.2. If Γu + Γ∗ u = 0, then Γu = 0 = Γ∗ u by orthogonality, from which N(Π) = C(Γ) follows. If f = Γu1 + Γ∗ u2 , then f = Π(PΓ∗ u1 + PΓ u2 ), from which R(Π) = R(Γ) + R(Γ∗ ) follows. Note that u1 − PΓ∗ u1 ∈ N(Γ) ⊂ D(Γ) and similarly for u2 . It is clear that −Π is the formal adjoint of Π. It remains to prove that if hf, Πgi + hf 0 , gi = 0 for all g ∈ D(Π), then f ∈ D(Π) and f 0 = Πf . Writing f = f1 + f2 + f3 in the Hodge splitting, and similarly for f 0 , we have hf1 , Γgi + hf30 , gi = 0, 0 + hf20 , gi = 0, hf3 , −Γ∗ gi + hf10 , gi = 0, by choosing g ∈ R(Γ∗ ) ∩ D(Γ), g ∈ C(Γ) and g ∈ R(Γ) ∩ D(Γ∗ ) respectively. Since Γ and Γ∗ are adjoint in the sense of unbounded operators, we conclude that f1 ∈ D(Γ∗ ), f30 = −Γ∗ f1 , f20 = 0, f3 ∈ D(Γ) and f10 = Γf3 . This shows that f ∈ D(Π) and f 0 = Πf .

347

10.1. Nilpotent operators

Definition 10.1.4 (Hodge projections). Let Γ be a nilpotent operator in a Hilbert space H. The associated Hodge projections are the orthogonal projections PΓ and PΓ∗ onto the subspaces R(Γ) and R(Γ∗ ) respectively. The orthogonal projection PC(Γ) onto the Γ-cohomology space C(Γ) is PC(Γ) = I − PΓ − PΓ∗ . Proposition 10.1.5 (Formulas for Hodge projections). Let Γ be a nilpotent operator in a Hilbert space H. If Γ is exact, then PΓ f = ΓΠ−1 f = −Π−1 Γ∗ f = −ΓΠ−2 Γ∗ f, PΓ∗ f = −Γ∗ Π−1 f = Π−1 Γf = −Γ∗ Π−2 Γf, for f ∈ D(Π) ∩ R(Π). If Γ is not exact, let ∈ R\{0}. Then we have PC(Γ) f = lim→0 (I +Π)−1 f , and the Hodge projections are PΓ f = lim Γ(I + Π)−1 f, →0

PΓ∗ f = − lim Γ∗ (I + Π)−1 f, →0

with convergence in H, for f ∈ H. We also have PΓ f = − lim(I + Π)−1 Γ∗ f for f ∈ D(Γ∗ ) and PΓ∗ f = lim(I + Π)−1 Γf for f ∈ D(Γ). Proof. The formulas for exact operators Γ involving Π−1 are immediate from (10.11), and the final second-order formulas follow since PΓ = PΓ2 and PΓ∗ = PΓ2∗ . For nonexact Γ, consider first PC(Γ) f . If f ∈ C(Γ), then (I + Π)−1 f = f . If f = Πu ∈ R(Π), then (I + Π)−1 Πu = u − 2 (I + Π)−1 u → 0 as → 0. We have used the skew-adjointness of Π, which implies that k(I + Π)−1 k ≤ 1. These uniform bounds also allow us to conclude that (I +Π)−1 f → 0 also for all f ∈ R(Π). This proves the formula for PC(Γ) f , from which it immediately follows that Γ(I + Π)−1 f = PΓ Π(I + Π)−1 f → PΓ (f − PC(Γ) f ) = PΓ f, and similarly for PΓ∗ . Alternatively, for f ∈ D(Γ∗ ), we have −(I + Π)−1 Γ∗ f = (I + Π)−1 ΠPΓ f → (I − PC(Γ) )PΓ f = PΓ f, and similarly for PΓ∗ .

The following result describes an important property that a nilpotent operator may have, which we will establish for d and δ on bounded Lipschitz domains. Proposition 10.1.6 (Compact potential maps). For a nilpotent operator Γ in a Hilbert space H, the following are equivalent.

Chapter 10. Hodge Decompositions

348

(i) The subspaces R(Γ) and R(Γ∗ ) are closed and C(Γ) is finite-dimensional, and the inverses of Γ : R(Γ∗ ) → R(Γ) and Γ∗ : R(Γ) → R(Γ∗ ) are compact. (ii) There exist compact operators K0 , K1 : H → H, with R(K1 ) ⊂ D(Γ), such that the homotopy relation ΓK1 f + K1 Γf + K0 f = f holds for all f ∈ D(Γ). (iii) The Hilbert space D(Γ) ∩ D(Γ∗ ), equipped with the norm (kf k2 + kΓf k2 + kΓ∗ f k2 )1/2 , is compactly embedded in H. Carefully note that unlike (i) and (iii), property (ii) does not involve the adjoint Γ∗ . We exploit this in Theorem 10.3.1 below to reduce the problem of existence of potentials, from Lipschitz domains to smooth domains. Also note for (i) that when the ranges are closed, Γ : R(Γ∗ ) → R(Γ) has a compact inverse if and only if there exists a compact operator KΓ : R(Γ) → H such that ΓKΓ = IR(Γ) . Indeed, if we have such KΓ , then PΓ KΓ is a compact operator giving a potential u ∈ R(Γ∗ ). Proof. Assume (i). Define compact operators K0 := PC(Γ) , and ( Γ−1 f ∈ R(Γ∗ ), f ∈ R(Γ), K1 f := 0, f ∈ N(Γ∗ ). It is straightforward to verify that ΓK1 = PΓ and K1 Γ = PΓ∗ , from which (ii) follows. ∞ be a sequence such that fj , Γfj and Γ∗ fj all are Assume (ii). Let (fj )j=1 bounded sequences in H. We have (I −PΓ )fj = (I −PΓ )(ΓK1 fj +K1 (Γfj )+K0 fj ) = (I −PΓ )K1 (Γfj )+(I −PΓ )K0 fj . By duality, we also obtain from the homotopy relation that PΓ fj = PΓ (Γ∗ K1∗ fj + K1∗ (Γ∗ fj ) + K0∗ fj ) = PΓ K1∗ (Γ∗ fj ) + PΓ K0∗ fj . ∞ ∞ This shows that (PΓ∗ fj )∞ j=1 , (PC(Γ) fj )j=1 and (PΓ∗ fj )j=1 have subsequences that converge in H, and (iii) follows. Assume (iii). The operator I + Π is an isometry between the Hilbert spaces D(Γ)∩D(Γ∗ ) and H, since Π is a skew-adjoint operator. Since I is compact between these spaces, perturbation theory shows that

Π : D(Γ) ∩ D(Γ∗ ) → H is a Fredholm operator, and (i) follows.

349

10.1. Nilpotent operators

Nilpotent operators appear naturally from the exterior and interior products, since v ∧ v ∧ w = 0 and v y (v y w) = 0. Example 10.1.7 (Algebraic Hodge decomposition). Fix a unit vector v ∈ V in an n-dimensional Euclidean space and define nilpotent linear maps µ(w) := v ∧ w,

µ∗ (w) := v y w,

w ∈ ∧V.

We apply the abstract theory above to Γ = µ and H, the finite-dimensional Hilbert space ∧V . Lemma 2.2.7 shows that R(µ) = N(µ), so in this case µ is exact and the Hodge decomposition reads ∧V = R(µ) ⊕ R(µ∗ ), where R(µ) are the multivectors normal to and R(µ∗ ) are the multivectors tangential to the hyperplane [v]⊥ , in the sense of Definition 2.8.6. We have (µ−µ∗ )2 = −1, and the Hodge projections are µµ∗ onto normal multivectors, and µ∗ µ onto tangential multivectors. Note that R(µ) and R(µ∗ ), for the full algebra ∧V , both have dimension 2n−1 . However, this is not true in general for the restrictions to ∧k V . For example, the space R(µ) ∩ ∧1 V of vectors normal to [v]⊥ is one-dimensional, whereas the space R(µ∗ ) ∩ ∧1 V of vectors tangential to [v]⊥ has dimension n − 1. The smaller k is, the more tangential k-vectors exist as compared to normal k-vectors. At the ends, all scalars are tangential and all n-vectors are normal. Example 10.1.8 (Rn Hodge decomposition). Consider the exterior and interior derivative operators dF (x) = ∇ ∧ F (x) and δF (x) = ∇ y F (x) in the Hilbert space H = L2 (V ; ∧Vc ) on the whole Euclidean space X = V , where we complexify the exterior algebra in order to use the Fourier transform. These two nilpotent operators are the Fourier multipliers Z n X c ei ∧ dF (ξ) = ∂i F (x)e−ihξ,xi dx = iξ ∧ Fˆ (ξ), c (ξ) = δF

k=1 n X k=1

X

Z ei y

∂i F (x)e−ihξ,xi dx = iξ y Fˆ (ξ) = (−iξ) y F (ξ),

X

defining the interior product as the sesquilinear adjoint of the exterior product. Define the pointwise multiplication operators µξ (Fˆ (ξ)) := ξ ∧ Fˆ (ξ),

µ∗ξ (Fˆ (ξ)) := ξ y Fˆ (ξ).

We view µξ , µ∗ξ : L2 (X; ∧Vc ) → L2 (X; ∧Vc ) as multiplication operators by the Pn radial vector field X 3 ξ 7→ ξ = k=1 ξk ek ∈ V . Thus we have F(dF ) = iµξ (F(F )),

F(δF ) = iµ∗ξ (F(F )).

Chapter 10. Hodge Decompositions

350

In particular, F is closed if and only if Fˆ is a radial multivector field, that is, ξ ∧ Fˆ = 0, and F is coclosed if and only if Fˆ is an angular multivector field, that is, ξ y Fˆ = 0. From Plancherel’s theorem it is clear that Γ = d and Γ∗ = −δ, with domains D(Γ) := {F ∈ L2 ; ξ ∧ Fˆ ∈ L2 } and D(Γ∗ ) := {F ∈ L2 ; ξ y Fˆ ∈ L2 }, are nilpotent operators in H, and that d = −δ ∗ in the sense of unbounded operators. In this case, the Hodge decomposition reads L2 (V ; ∧Vc ) = R(d) ⊕ R(δ). That d is exact is a consequence of µξ being exact for each ξ ∈ V \ {0}. By considering Fˆ near ξ = 0, we see that the ranges are not closed, which is a consequence of the domain X not being bounded. Using the formulas from Proposition 10.1.5, we see that the Hodge projections are the singular integrals Z Pd F (x) = ∇ ∧ Ψ(x − y) y F (y)dy X Z k = F (x) + p.v. ∇ ∧ (Ψ(x˙ − y) y F (y))dy, n X Z Pδ F (x) = ∇ y Ψ(x − y) ∧ F (y)dy X Z n−k = F (x) + p.v. ∇ y (Ψ(x˙ − y) ∧ F (y))dy, n X for k-vector fields F ∈ L2 (X; ∧k Vc ). We have used the distributional derivative ∂i Ψ(x) = ei δ(x)/n + p.v.∂i Ψ(x).

10.2

Half-Elliptic Boundary Conditions

For the remainder of this chapter, we study the nilpotent operators d and δ on bounded domains D, at least Lipschitz regular, in Euclidean space X. The main idea in this section is to use the commutation theorem (Theorem 7.2.9) and reduce the problems to smooth domains. Realizing the operators that are implicit in Definition 7.6.1 as unbounded nilpotent operators, we have the following. Definition 10.2.1 (d and δ on domains). Let D be a bounded Lipschitz domain in a Euclidean space (X, V ). Define unbounded linear operators d, d, δ, δ in L2 (D) = L2 (D; ∧V ) as follows. Assume that F, F 0 ∈ L2 (D) and consider the equation Z hF 0 (x), φ(x)i + hF (x), ∇ y φ(x)i dx = 0. D

If this holds for all φ ∈ C0∞ (D), then we define F ∈ D(d) and dF := F 0 . If this holds for all φ ∈ C ∞ (D), then we define F ∈ D(d) and dF := F 0 .

351

10.2. Half-Elliptic Boundary Conditions Assume that F, F 0 ∈ L2 (D) and consider the equation Z hF 0 (x), φ(x)i + hF (x), ∇ ∧ φ(x)i dx = 0. D

If this holds for all φ ∈ C0∞ (D), then we define F ∈ D(δ) and δF := F 0 . If this holds for all φ ∈ C ∞ (D), then we define F ∈ D(δ) and δF := F 0 . We recall from Section 7.6 that by Stokes’s theorem we interpret F ∈ D(d) as being normal at ∂D in a weak sense, and F ∈ D(δ) as being tangential at ∂D in a weak sense. Basic properties of these operators are the following. Proposition 10.2.2 (Nilpotence). Let D be a bounded Lipschitz domain in a Euclidean space. Then the operators d, d, δ, δ are well-defined nilpotent operators on L2 (D). In particular, they are linear, closed, and densely defined. With the pointwise Hodge star and involution maps, we have δ(F ∗) = (dFb)∗, d(F ∗) = (δ Fb)∗,

F ∈ D(d), F ∈ D(δ).

Proof. Consider d. The other proofs are similar. That d is defined on C0∞ (D), linear, and closed is clear. It is well defined, since F = 0 implies F 0 = 0, since C ∞ (D) is dense in L2 (D). To show nilpotence, assume F ∈ D(d). Then Z Z 0 + hdF (x), ∇ y φ(x)i dx = − hF (x), ∇ y (∇ y φ(x))idx = 0 D

D

for all φ ∈ C ∞ (D), which shows that d(dF ) = 0. The relation between d, δ and δ, d follows from Proposition 7.1.7(i).

The goal of this section is to prove the following duality. Recall the definition (6.4) of adjointness in the sense of unbounded operators. Proposition 10.2.3 (Duality). Let D be a bounded Lipschitz domain in a Euclidean space. Then d and −δ are adjoint in the sense of unbounded operators. Similarly, d and −δ are adjoint in the sense of unbounded operators. From Propositions 10.1.2 and 10.2.3 we obtain a Hodge decomposition with tangential boundary conditions L2 (D) = R(d) ⊕ Ck (D) ⊕ R(δ), where Ck (D) := N(d) ∩ N(δ), and a Hodge decomposition with normal boundary conditions L2 (D) = R(d) ⊕ C⊥ (D) ⊕ R(δ), where C⊥ (D) := N(d) ∩ N(δ). We will prove in Section 10.3 that the ranges of the four operators are closed, so the closures here are redundant. For the proof of Proposition 10.2.3, we need the following results.

352

Chapter 10. Hodge Decompositions

Lemma 10.2.4 (Local nonsmooth commutation theorem). Let ρ : D1 → D2 be a Lipschitz diffeomorphism between domains D1 and D2 in Euclidean space. If F ∈ D(d) in D2 with supp F ⊂ D2 , then ρ∗ F ∈ D(d) with d(ρ∗ F ) = ρ∗ (dF ) in D1 . Similarly, if F ∈ D(δ) in D1 with supp F ⊂ D1 , then ρ˜∗ F ∈ D(δ) with δ(˜ ρ∗ F ) = ρ˜∗ (δF ) in D2 . We recall, for example, that supp F ⊂ D2 means that F = 0 in a neighborhood of ∂D2 . Note that for general Lipschitz changes of variables, ρ∗ F and ρ˜∗ F are defined almost everywhere by Rademacher’s theorem. Proof. By Proposition 7.2.7 it suffices to prove the first statement. Consider first F ∈ C0∞ (D2 ). We mollify and approximate ρ by ρt (x) := ηt ∗ ρ(x), where η ∈ C0∞ (X; R) with η = 1 and ηt (x) := t−n η(x/t). Note that ρt is well defined on every compact subset of D1 for small t. It follows that ρt ∈ C ∞ and R

d(ρ∗t F ) = ρ∗t (dF ) holds by Theorem 7.2.9. From the dominated convergence theorem we conclude that ρ∗t F → ρ∗ F in L2 (D1 ). Since for the same reason ρ∗t (dF ) → ρ∗ (dF ), and d is a closed operator, it follows that ρ∗ F ∈ D(d) and d(ρ∗ F ) = ρ∗ (dF ). Next consider general F ∈ D(d) with compact support in D2 . Similarly to above, we now mollify and approximate F by Fn ∈ C0∞ (D2 ), with Fn → F and dFn → dF in L2 (D2 ). We have shown above that d(ρ∗ Fn ) = ρ∗ (dFn ). Using that ρ∗ : L2 (D2 ) → L2 (D1 ) is bounded and that d is closed, it follows that ρ∗ F ∈ D(d) and d(ρ∗ F ) = ρ∗ (dF ). The following shows that the normal and tangential boundary conditions for d and δ are obtained by closure from C0∞ . Proposition 10.2.5 (Half Dirichlet conditions). Let D be a bounded Lipschitz domain in a Euclidean space. If F ∈ D(d), then there exists Ft ∈ C0∞ (D) such that Ft → F

and

dFt → dF

in L2 (D) as t → 0. Similarly, if F ∈ D(δ), then there exists Ft ∈ C0∞ (D) such that Ft → F and δFt → δF in L2 (D) as t → 0. Proof. By Hodge star duality it suffices to consider d. By the compactness of D, we can localize and assume that supp F ⊂ Dp ∩ D near p ∈ ∂D as in Definition 6.1.4. We note from Definition 10.2.1 that extending F by 0 outside D, we have F ∈ D(d) on X as in Example 10.1.8. Pulling back by the local parametrization ρ, Lemma 10.2.4 shows that ρ∗ F ∈ D(d) on Rn . We translate ρ∗ F up into Ωp and

10.2. Half-Elliptic Boundary Conditions

353

pull back by ρ−1 to define F˜t := (ρ∗ )−1 (ρ∗ F (x0 , xn − t)). This yields F˜t ∈ D(d) with supp F˜t ⊂ D. Finally, we mollify and approximate F˜t by Ft (x) := ηt (x) ∗ F˜t (x), x ∈ D, R where η ∈ C0∞ (X; R) with η = 1, supp η ⊂ B(0, r), and ηt (x) := t−n η(x/t). If 0 < r < t is chosen small enough depending on the Lipschitz geometry, we obtain Ft ∈ C0∞ (D) and can verify that Ft and dFt converge to F and dF respectively. Proof of Proposition 10.2.3. Consider the equation Z hF 0 (x), φ(x)i + hF (x), ∇ y φ(x)i dx = 0. D

This holds for all F ∈ D(d) with F 0 = dF , and all φ ∈ C0∞ (D) by Definition 10.2.1. By Proposition 10.2.5 and a limiting argument, this continues to hold for φ ∈ D(δ). This shows that d and −δ are formally adjoint. Furthermore, assume that the equation holds for some F and F 0 ∈ L2 (D) and all φ ∈ D(δ). In particular, it holds for all φ ∈ C0∞ (D), and it follows by definition that F ∈ D(d) and F 0 = dF . This shows that d and −δ are adjoint in the sense of unbounded operators. The proof that d and −δ are adjoint in the sense of unbounded operators is similar. We next remove the assumption of compact support in Lemma 10.2.4. Lemma 10.2.6 (Nonsmooth commutation theorem). Let ρ : D1 → D2 be a Lipschitz diffeomorphism between bounded Lipschitz domains D1 and D2 in Euclidean space. If F ∈ D(d) on D2 , then ρ∗ F ∈ D(d) on D(D1 ) with d(ρ∗ F ) = ρ∗ (dF ) in D1 . Similarly, if F ∈ D(δ) on D1 , then ρ˜∗ F ∈ D(δ) with δ(˜ ρ∗ F ) = ρ˜∗ (δF ) on D2 . Proof. By Proposition 10.2.2, it suffices to consider d. In this case, we must show that Z hρ∗ (dF ), φi + hρ∗ F, ∇ y φi dx = 0 D1

C0∞ (D1 ).

for φ ∈ By the Lipschitz change of variables formula (6.2), see Section 6.5, and Lemma 10.2.4, this is equivalent to Z hdF, ρ˜∗ φi + hF, ∇ y (˜ ρ∗ φ)i dx = 0, D2

which holds by Proposition 10.2.3.

It is clear from the definition that D(d) on D can be viewed as a subspace of D(d) on X, by extending F on D by zero to all X. The following existence of extension maps shows that D(d) on D can be identified with the quotient space D(dX )/D(dX\D ).

Chapter 10. Hodge Decompositions

354

Proposition 10.2.7 (Extensions for d and δ). Let D be a bounded Lipschitz domain in a Euclidean space X. Assume that F ∈ D(d) on D. Then there exists F˜ ∈ D(d) on X such that F˜ |D = F . Furthermore, there exists Ft ∈ C ∞ (D) such that Ft → F and dFt → dF in L2 (D) as t → 0. Similarly, assume that F ∈ D(δ) on D. Then there exists F˜ ∈ D(δ) on X such that F˜ |D = F . Furthermore, there exists Ft ∈ C ∞ (D) such that Ft → F and δFt → δF in L2 (D) as t → 0. Proof. As in the proof of Proposition 10.2.5, it suffices to consider d, and we may assume that supp F ⊂ Dp ∩ D, a small neighborhood of p ∈ ∂D. By Lemma 10.2.6 we have ρ∗ F ∈ D(d) on Ωp ∩ {xn > 0}. Define ( ρ∗ F (x), xn > 0, G(x) := ∗ ∗ R ρ F (x), xn < 0, where R(x0 , xn ) := (x0 , −xn ) denotes reflection in Rn−1 . We claim that G ∈ D(d) on all Ωp across Rn−1 . To see this, for φ ∈ C0∞ (Ωp ), we calculate Z Z ∗ ∗ hdρ F, φi + hρ F, ∇ y φi dx + hdR∗ ρ∗ F, φi + hR∗ ρ∗ F, ∇ y φi dx xn >0 xn 0

Since φ + R∗ φ is tangential on Rn−1 , we have φ + R∗ φ ∈ D(δ) on Ωp ∩ Rn+ , so by Proposition 10.2.3, the integral vanishes. By Lemma 10.2.4, the field F˜ := (ρ∗ )−1 G ∈ D(d) on X is an extension of F , and if we mollify and approximate F˜ by Ft (x) := ηt ∗ F˜ (x),

x ∈ D,

∞

as above, we obtain Ft ∈ C (D) and can verify that Ft and dFt converge to F and dF respectively.

10.3

Hodge Potentials

Our main result on Hodge decompositions is the following. Theorem 10.3.1 (Hodge decompositions on Lipschitz domains). Let D be a bounded Lipschitz domain in a Euclidean space X. Then the operators d, δ, d, δ in L2 (D; ∧V ) all have closed ranges, the cohomology spaces Ck (D) = N(d) ∩ N(δ) and C⊥ (D) = N(d) ∩ N(δ) are finite-dimensional, and we have Hodge decompositions L2 (D; ∧V ) = R(d) ⊕ Ck (D) ⊕ R(δ) = R(d) ⊕ C⊥ (D) ⊕ R(δ). Moreover, the inverses of d : R(δ) → R(d), δ : R(d) → R(δ), d : R(δ) → R(d), and δ : R(d) → R(δ) are all L2 compact.

355

10.3. Hodge Potentials

The proof follows from the following reduction and Theorem 10.3.3 below. Reduction of Theorem 10.3.1 to a ball. We prove that there are compact operators K0 and K1 on L2 (D) such that dK1 F + K1 dF + K0 F = F for all F ∈ D(d). By Propositions 10.1.6 and 10.2.2, this will prove TheoremS10.3.1. By Definition 6.1.4 we have a finite covering D = α Dα , with Lipschitz diffeomorphisms ρα : B → Dα from the unit ball B. Moreover, we have a partition of unity ηα ∈ C ∞ (D) subordinate to this covering. By Theorem 10.3.3 for the ball B, we have compact maps K1B and K0B on L2 (B) such that dK1B F + K1B dF + K0B F = F . Note that we need only part (i) in the proof of Theorem 10.3.3 for this. Define X K1 F := ηα (ρ∗α )−1 K1B (ρ∗α F |Dα ), α

which is seen to be compact on L2 (D). We calculate X X dK1 F = ηα (ρ∗α )−1 (I − K1B d − K0B )(ρ∗α F |Dα )+ ∇ηα ∧ (ρ∗α )−1 K1B (ρ∗α F |Dα ) α

α

= F − K1 dF − K0 F, where K0 F :=

X

ηα (ρ∗α )−1 K0B (ρ∗α F |Dα ) −

X

α

∇ηα ∧ (ρ∗α )−1 K1B (ρ∗α F |Dα )

α

is seen to be compact on L2 (D). Note the critical use of Theorem 7.2.9. This proves Theorem 10.3.1 for Lipschitz domains D. In the proof of Theorem 10.3.1 we used Proposition 10.1.6(ii). As for the characterization (iii), it is natural to ask whether D(d) ∩ D(δ) ⊂ H 1 (D), that is, whether the total derivative ∇⊗F belongs to L2 (D), whenever F, dF, δF ∈ L2 (D). This is not true for general Lipschitz domains, where the irregularities of ∂D may prevent F ∈ D(d) ∩ D(δ) from having full Sobolev H 1 regularity, but it does hold for smooth domains. Example 10.3.2 (Nonconvex corner). Let Dα ⊂ R2 be a bounded domain that is smooth except at 0, in a neighborhood of which Dα coincides with the sector {reiφ ; r > 0, 0 < φ < α}. Define a scalar function u : Dα → R such that u = rπ/α sin(πφ/α)η, where η ∈ C0∞ (R2 ), η = 1 in a neighborhood of 0, and η = 0 where Dα differs from the sector. Consider the gradient vector field F := ∇u ∈ R(d). Using the estimate |F | . rπ/α−1 , we verify that F ∈ D(d) ∩ D(δ). However, Z Z 1 |∇ ⊗ F |2 dxdy & (rπ/α−2 )2 rdr. D

0

Chapter 10. Hodge Decompositions

356

Therefore, when Dα is not convex, that is, when α > π, then F ∈ / H 1 (D). Figure 10.1 shows the case α = 3π/2.

Figure 10.1: The harmonic function r2/3 sin(2φ/3) in quadrants 1–3 in the unit circle, with Dirichlet boundary conditions but infinite gradient at the origin.

Theorem 10.3.3 (Full regularity of Hodge potentials). Let D be a bounded C 2 domain. Then D(d) ∩ D(δ) = Hk1 (D) := {F ∈ H 1 (D) ; ν y F |∂D = 0} 1

and

1

D(d) ∩ D(δ) = H⊥ (D) := {F ∈ H (D) ; ν ∧ F |∂D = 0}. For the proof of Theorem 10.3.3 we shall prove a Weitzenb¨ock identity for d and δ on D, involving a boundary curvature term. This requires the following definitions from differential geometry and uses that the boundary ∂D is C 2 regular. In this case, the unit normal vector field ν on ∂D is C 1 , and the curvature of the boundary is a continuous function. Proposition 10.3.4 (Derivative of normal). Let D be a bounded C 2 domain, with outward-pointing unit normal vector field ν on ∂D. At p ∈ ∂D, let Tp (∂D) denote the tangent hyperplane. Then the map p S∂D : Tp (∂D) → Tp (∂D) : v 7→ ∂v ν,

is linear and symmetric. Moreover, for any tangential C 1 vector fields u and v on ∂D, at each p ∈ ∂D we have p p hu, S∂D vi = −h∂u v, νi = hS∂D u, vi.

357

10.3. Hodge Potentials

Proof. We have 0 = ∂v |ν|2 = 2h∂v ν, νi, since |ν| = 1 on ∂D, which shows that p p at p ∈ ∂D, we note (v) is a tangential vector. To show the symmetry of S∂D S∂D that 0 = ∂u hv, νi = h∂u v, νi + hv, ∂u νi and 0 = ∂v hu, νi = h∂v u, νi + hu, ∂v νi, p since u and v are tangential on ∂D. The symmetry of S∂D now follows, since the Lie bracket ∂u v − ∂v u = [u, v] = Lu v is tangential.

Definition 10.3.5 (Second fundamental form). Let D be a bounded C 2 domain. The symmetric bilinear form p B∂D : Tp (∂D) × Tp (∂D) → R : (u, v) 7→ −h∂u v, νi

from Proposition 10.3.4 is called the second fundamental form for ∂D. The associated symmetric map p S∂D : Tp (∂D) → Tp (∂D) : v 7→ ∂v ν

from Proposition 10.3.4 is called the Weingarten map, or shape operator, for ∂D. p The eigenvalues {κ1 , . . . , κn−1 } of S∂D are called the principal curvatures of ∂D 0 at p, and a corresponding ON-basis {e10 , . . . , en−1 } for Tp (∂D) of eigenvectors to p is referred to as the principal directions of curvatures at p. S∂D Note that if D is a convex domain, then κj ≥ 0. Theorem 10.3.6 (Weitzenb¨ock identities). Let D be a bounded C 2 domain, and let e0j denote the principal directions of curvatures κj , j = 1, . . . , n − 1. Then Z

Z

2

2

(|dF | + |δF | )dx −

|∇ ⊗ F | dx = D

Z

2

D

|∇ ⊗ F |2 dx =

Z

j=1

(|dF |2 + |δF |2 )dx −

D

D

where |∇ ⊗ F |2 =

Pn

n−1 XZ

j=1

n−1 XZ j=1

κj |ej0

F |2 dy,

F ∈ H⊥1 (D),

κj |ej0 y F |2 dy,

F ∈ Hk1 (D),

∧

∂D

∂D

|∂j F |2 .

Example 10.3.7 (Kadlec’s formula). Consider a scalar function U : D → R satisfying Poisson’s equation ∆U = f in D, with Dirichlet boundary conditions U |∂D = 0. This means that its gradient vector field F = ∇U is normal at the boundary. Assuming that F ∈ H⊥1 (D), we have Kadlec’s formula n Z X i,j=1

D

2

Z

2

Z

|f | dx − (n − 1)

|∂i ∂j U | dx = D

∂D

H(y) |∇U |2 dy,

Chapter 10. Hodge Decompositions

358

y where H(y) := Tr(S∂D )/(n − 1) is the mean curvature of the boundary. Note that Lagrange’s identity Proposition 3.1.1 shows that |e0j ∧ F |2 = |e0j |2 |F |2 −|he0j , F i|2 = |F |2 . If instead U satisfies Neumann boundary conditions hν, ∇U i = 0, then we get a similar identity n Z X i,j=1

|∂i ∂j U |2 dx =

Z

|f |2 dx −

D

D

n Z X j=1

κj |hej0 , ∇U i|2 dy,

∂D

but where all the principal curvatures appear and not only the mean curvature. For convex domains, the Weitzenb¨ock identities imply that Z Z |∇ ⊗ F |2 dx ≤ (|dF |2 + |δF |2 )dx, for all F ∈ H⊥1 (D) ∪ Hk1 (D), D

D

since in this case all κj ≥ 0. In general, we have the following estimates. Corollary 10.3.8 (Gaffney’s inequality). Let D be a bounded C 2 domain. Then Z Z |∇ ⊗ F |2 dx . (|dF |2 + |δF |2 + |F |2 )dx, for all F ∈ H⊥1 (D) ∪ Hk1 (D). D

D

Proof. For a C 2 domain, we note that the principal curvatures κj are bounded 2 . To functions, which shows that the boundary integral terms are . kF kL 2 (∂D) 2 replace this by a term kF kL2 (D) , we apply Stokes’s theorem to obtain a standard trace estimate as follows. Let θ ∈ C0∞ (X; V ) be a vector field such that inf y∈∂D hθ(y), ν(y)i > 0, that is, θ is uniformly outward pointing on ∂D. Stokes’s theorem gives Z Z |F |2 hθ, νidy = 2h∂θ F, F i + |F |2 div θ dx. ∂D

D

Estimating, this shows that Z Z 2 2 kF kL2 (∂D) . |∇ ⊗ F ||F | + |F |2 dx. |F | hθ, νidy . ∂D

D

It follows from the Weitzenb¨ock identities that Z Z Z Z |∇ ⊗ F ||F |dx + C (|dF |2 + |δF |2 )dx + C |∇ ⊗ F |2 dx ≤ |F |2 dx, (10.3) D

D

D

D

for some constant C < ∞. We next use an estimate technique called the absorption inequality, which is 1 2 b . ab ≤ 2 a2 + 2

10.3. Hodge Potentials

359

√ √ This is, of course, nothing deeper than ( a − b/ )2 ≥ 0. To use this, we take a = |∇ ⊗ F (x)|, b = |F (x)|, and = C −1 . This shows that the second term on the right-hand side in (10.3) is Z |∇ ⊗ F ||F |dx ≤

C D

1 2

Z

2

|∇ ⊗ F | dx + D

C2 2

Z

|F |2 dx,

D

where the first term can be moved to the left-hand side in (10.3) and be absorbed there. Gaffney’s inequality follows. Proof of Theorem 10.3.6. (i) Let first F ∈ C 2 (D) and consider the 1-form θ(x, v) :=

n X hv, ej ihF (x), ∂j F (x)i − hv ∧ F (x), dF (x)i − hv y F (x), δF (x)i, j=1

for x ∈ D, v ∈ V . We calculate its exterior derivative θ(x, ˙ ∇) = (|∇ ⊗ F |2 + hF, ∆F i) − (|dF |2 + hF, δdF i) − (|δF |2 + hF, dδF i) = |∇ ⊗ F |2 − |dF |2 − |δF |2 , since ∆ = δd + dδ. The Stokes’ formula (7.4) gives Z

Z |∇ ⊗ F |2 − |dF |2 − |δF |2 dx =

D

hF, ∂ν F i − hν ∧ F, dF i − hν y F, δF i dy.

∂D

We continue and rewrite the right-hand side with nabla calculus as hF, hν, ∇iF i − hF, ν ∧ (∇ y F )i − hν ∧ F, ∇ ∧ F i = hF, ∇ y (n ∧ F˙ )i − hν ∧ F, ∇ ∧ F i = −hF, ∇ y (n˙ ∧ F )i + hF, ∇ y (n ∧ F )i − hν ∧ F, ∇ ∧ F i,

(10.4) (10.5)

where n ∈ C 1 (X; V ) denotes an extension of ν. The first step uses the algebraic anticommutation relation ν ∧ (∇ y F ) = hν, ∇iF − ∇ y (n ∧ F˙ ), and the second step uses the analytic product rule ∇ y (n ∧ F ) = ∇ y (n˙ ∧ F ) + ∇ y (n ∧ F˙ ). At p ∈ ∂D, we calculate the first term in the ON-basis {e01 , . . . , e0n−1 , ν} and get hF, ∇ y (n˙ ∧ F )i =

n−1 X

κj |e0j

∧

F |2 + hν ∧ F, (∂ν n) ∧ F i.

j=1

On the other hand, the normal derivatives in the last two terms in (10.4) are hF, ν y ∂ν (n ∧ F )i − hν ∧ F, ν ∧ ∂ν F i = hν ∧ F, (∂ν n) ∧ F i.

Chapter 10. Hodge Decompositions

360

Therefore these three terms involving the normal derivatives cancel, and we obtain the identity Z (|∇ ⊗ F |2 − |dF |2 − |δF |2 )dx D

=−

n−1 XZ j=1

κj |e0j ∧

2

Z

F | dy +

∂D

hF, ∇0 y (ν ∧ F )i − hν ∧ F, ∇0 ∧ F i dy, (10.6)

∂D

Pn−1 where ∇0 := ν y (ν ∧ ∇) = j=1 e0j ∂e0j . (ii) Next consider F ∈ H⊥1 (D). To obtain the first Weitzenb¨ock identity, we use the fact that C 2 (D) is dense in H 1 (D) and take Fj ∈ C 2 (D) such that Fj → F and ∇ ⊗ Fj → ∇ ⊗ F in L2 (D). On the C 2 manifold ∂D, we use the Sobolev spaces H 1/2 (∂D) and H −1/2 (∂D), as discussed in Example 6.4.1, where H 1/2 ⊂ L2 ⊂ H −1/2 . As usual, we allow the functions to be multivector fields, and require that each component function be in such Sobolev space. We need the following well-known facts. • The trace map H 1 (D) → H 1/2 (∂D) : F 7→ F |∂D is a bounded linear operator. • The tangential derivative ∇0 defines a bounded linear operator ∇0 : H 1/2 (∂D) → H −1/2 (∂D). • Multiplication by a C 1 function like ν is a bounded operation on H 1/2 (∂D). • The spaces H 1/2 (∂D) and H −1/2 (∂D) are dual; in particular, we have the estimate Z . kF kH 1/2 (∂D) kGkH −1/2 (∂D) . hF, Gidx ∂D

Given this, we apply (10.6) to Fj and take the limit as j → ∞. Since ν ∧ Fj → ν ∧ F = 0 in H 1/2 (∂D), we obtain the Weitzenb¨ock identity for F ∈ H⊥1 (D). (iii) To obtain the Weitzenb¨ock identity for F ∈ Hk1 (D), we instead rewrite θ as θ(x, ν) = hF, hν, ∇iF i − hF, ν y (∇ ∧ F )i − hν y F, ∇ y F i = hF, ∇ ∧ (n y F˙ )i − hν y F, ∇ y F i = −hF, ∇ ∧ (n˙ y F )i + hF, ∇ ∧ (n y F )i − hν y F, ∇ y F i, and proceed as in (i) and (ii).

We finally prove Theorem 10.3.3. From the Weitzenb¨ock identities, the Gaffney inequalities show that on C 2 domains we have Hk1 (D) ⊂ D(d) ∩ D(δ) and H⊥1 (D) ⊂ D(d) ∩ D(δ) and that k∇ ⊗ F k2 + kF k2 ≈ k∇ ∧ F k2 + k∇ y F k2 + kF k2

361

10.3. Hodge Potentials

in L2 (D) norm, for all F ∈ Hk1 (D) and all F ∈ H⊥1 (D). It is important to note that this equivalence of norms, without further work, does not imply that D(d)∩D(δ ) ⊂ Hk1 (D) and D(d) ∩ D(δ) ⊂ H⊥1 (D). It particular, the absorption technique in the proof of Corollary 10.3.8 fails to prove this. Proof of Theorem 10.3.3. It remains to prove D(d) ∩ D(δ) ⊂ H⊥1 (D). By Hodge duality as in Proposition 10.2.2, this will imply the corresponding result for normal boundary conditions. (i) Consider first the case that D is the unit ball B := {x ∈ V ; |x| < 1} and let F ∈ D(d) ∩ D(δ). Using a partition of unity, we write F = F0 + F1 , F0 , F1 ∈ D(d) ∩ D(δ), where F0 (x) = 0 when |x| > 1/2 and F1 (x) = 0 when |x| < 1/3. We use inversion R(x) = 1/x in the unit sphere, with derivative Rx h = −x−1 hx−1 to extend F1 to ( F1 (x), |x| < 1, ˜ F1 (x) := ∗ R F1 (x), |x| > 1. Arguing as in the proof of Proposition 10.2.7, replacing Rn−1 by the sphere |x| = 1, we conclude that F˜ ∈ D(d) on X. Moreover, R is a conformal map and R∗ = ˜ ∗−1 . From this it follows that R∗ F1 ∈ D(δ) on X, with |x|2(n−1) R ˜ ∗−1 F1 (x) + |x|2(n−1) R ˜ ∗−1 (∇ y F1 )(x) ∇ y R∗ F1 (x) = |x|2(n−2) x y R for |x| > 1. Recall that F1 , extended by 0 for |x| > 1 belongs to D(δ). We obtain an extension F˜ := F0 + F˜1 of F to X, with F˜ = 0 for |x| > 3 and ˜ ˜ F , dF , and δ F˜ all belonging to L2 (X). By Plancherel’s theorem and Langrange’s identity, we get Z Z Z (2π)n |∇ ⊗ F˜ |2 dx = |F(F˜ )|2 |ξ|2 dξ = |ξ ∧ F(F˜ )|2 + |ξ y F(F˜ )|2 dξ < ∞. X

X

X

Recall from Example 10.1.8 that d and δ on X are the Fourier multipliers iµξ and iµ∗ξ . This shows that F ∈ H 1 (D). (ii) Next consider a general bounded C 2 domain D. Localizing the problem with a partition of unity, we may assume that D is C 2 diffeomorphic to B. Moreover, we may assume that we have a C 2 map ρ : [0, 1] × B → X such that ρt = ρ(t, ·) defines a C 2 diffeomorphism B → ρt (B) =: Dt , with D0 = B and D1 = D. For fixed t ∈ [0, 1], we consider the inclusion Hk1 (Dt ) ⊂ D(d)∩D(δ) on the C 2 domain Dt . We note from Proposition 10.1.3 that I + d + δ : D(d) ∩ D(δ) → L2 (Dt ) is an invertible isometry, so the inclusion amounts to I + d + δ : Hk1 (Dt ) → L2 (Dt ) being an injective semi-Fredholm operator. See Definition 6.4.9. From (i), we know that it is surjective for the ball at t = 0. To apply the method of continuity and

Chapter 10. Hodge Decompositions

362

conclude that it is surjective for all t, and in particular for D at t = 1, we note that the normalized pushforward ρ˜t∗ defines invertible maps Hk1 (B) → Hk1 (Dt ) and L2 (B) → L2 (Dt ). The method of continuity therefore applies to the family of semi-Fredholm operators ρt∗ )−1 (I + d + δ)˜ (˜ ρt∗ : Hk1 (B) → L2 (B). We conclude that I + d + δ : Hk1 (Dt ) → L2 (Dt ) is invertible, which shows that D(d) ∩ D(δ ) = Hk1 (D) and completes the proof of Theorem 10.3.3.

10.4

Bogovski˘ı and Poincar´e Potentials

Recall that exterior and interior potentials in general are highly nonunique. In this section we prove the following surprising results about potentials on strongly Lipschitz domains D. • We have seen in Example 10.3.2 that in contrast to smooth domains, the potential U in the subspace R(δ) to F = dU ∈ R(d) may not belong to H 1 (D). We refer to this potential U as the Hodge potential for F , which is characterized by its minimal L2 norm. It follows from Theorem 10.4.3 below that every exact field F ∈ R(d) on any bounded and strongly Lipschitz domain D nevertheless has a potential ˜ , in general different from the Hodge potential, such that U ˜ ∈ H 1 (D) and U dU = F . We refer to such potentials as (regularized) Poincar´e potentials for F. • We have seen that the Hodge potential U ∈ R(d) to F = δU ∈ R(δ) is tangential on ∂D, meaning that half of the component functions of U vanish there. Theorem 10.4.3 below show that every field F ∈ R(δ) on any bounded ˜ , in general different and strongly Lipschitz domain D in fact has a potential U ˜ ∈ H 1 (D) and δU = F . This means from the Hodge potential, such that U 0 ˜ vanish on ∂D, and we note that this is that all component functions of U a nontrivial result also for smooth domains. We refer to such potentials as Bogovski˘ı potentials for F . ˜∈ Similarly, and related by the Hodge star, there exist Poincar´e potentials U ˜ ∈ H 1 for F ∈ R(d). We will H 1 (D) for F ∈ R(δ), and Bogovski˘ı potentials U 0 formulate the results only for d and δ, and leave it to the reader to translate the results in this section to d and δ. First consider a star-shaped domain D. In what follows, we extend the operators initially defined on k-vector fields, by linearity to act on general multivector fields. The method we use to construct a Poincar´e potential U to a given field F ∈ R(d) on D builds on Poincar´e’s Theorem 7.5.2. If D is shar-shaped with respect to p0 ∈ D, then this gives the potential Z 1 Tp0 (F )(x) := (x − p0 ) y F (p0 + t(x − p0 )) tk−1 dt, x ∈ D, (10.7) 0

10.4. Bogovski˘ı and Poincar´e Potentials

363

provided k ≥ 1 and F is a smooth k-vector field. For a scalar function F : D → ∧0 R, we let Tp0 F = 0. We would like to extend (10.7) to fields that are square integrable, without any assumption on regularity. To obtain a bounded operator, we need to average the formula (10.7) over base points p around p0 . In what follows, we assume that D is star-shaped not only with respect to a point, but to a whole ball. We fix θ ∈ C0∞ (B(p0 ; )) andR assume that D is star-shaped with respect to each p ∈ B(p0 ; ), where > 0 and θdx = 1. Then define the averaged operator Z TD F (x) :=

θ(p)Tp F (x)dp,

x ∈ D.

(10.8)

|p−p0 |≤

We rewrite this formula by changing the variables p and t to y := p + t(x − p) and s = 1/(1 − t) − 1. This gives Z TD F (x) = (x − y) y F (y) kθ (x, y)dy, (10.9) D

where Z

∞

kθ (x, y) :=

θ(y + s(y − x))sk−1 (1 + s)n−k ds.

0

This operator TD constructs the regularized Poincar´e potential for an exact kvector field on a bounded domain that is star-shaped with respect to B(p0 ; ). Exercise 10.4.1 (Kernel support). Show that kθ (x, y) 6= 0 is possible only when y lies on the straight line between x and a point p ∈ supp η, and that we have estimates 1 |kθ (x, y)| . , x, y ∈ D, |x − y|n so that TD is a weakly singular integral operator. Note how by averaging with θ we have replaced the line integral for Tp0 by a volume integral over a conical region for TD . The adjoint operator ∗ TD F (x) =

Z (y − x) ∧ F (y) kθ (y, x)dy

(10.10)

D

constructs the Bogovski˘ı potential for a (k − 1)-vector field F ∈ R(δ) on the star∗ ∗ shaped domain D. We see from Exercise 10.4.1 that TD F |∂D = 0, since for TD we integrate over a cone starting at x, away from B(p0 , ). For domains D that are C 2 diffeomorphic to a domain that is star-shaped with respect to a ball, we can pull back and push forward these operators TD and ∗ TD to obtain Bogovski˘ı and regularized Poincar´e potentials. Next we extend these constructions to general strongly Lipschitz domains, and provide the necessary analysis.

Chapter 10. Hodge Decompositions

364

Definition 10.4.2 (Bogovski˘ı and Poincar´e maps). Let D be a bounded and strongly S Lipschitz domain. Fix a finite cover D = α Dα by domains Dα that Rare starshaped with respect to balls B(pα ; ). Further fix θα ∈ C0∞ (B(pα ; )) with θα dx = 1 and a partition of unity ηα ∈ C ∞ (D) subordinate to the covering Dα . We assume that ηα = 1 on a neighborhood of supp θα . The regularized Poincar´e map with these choices Dα , θα , ηα , for d on D is X TD F (x) = ηα (x)TDα (F |Dα )(x), x ∈ D. α

The Bogovski˘ı map, with these choices Dα , θα , ηα , for δ on D is X ∗ ∗ TD (ηα F |Dα )(x), x ∈ D. TD F (x) = α α ∗ Here TDα and TD are the Poincar´e and Bogovski˘ı maps on the star-shaped doα mains Dα , constructed as above.

Unlike the star-shaped case, these Bogovski˘ı and regularized Poincar´e maps on general strongly Lipschitz domains do not straight away give potentials for (co-)exact fields. We proceed as follows. Theorem 10.4.3 (Bogovski˘ı and Poincar´e homotopies). Let D be a bounded and strongly Lipschitz domain. The regularized Poincar´e potential map from Definition 10.4.2, maps TD : C ∞ (D) → C ∞ (D) and extends by continuity to a bounded operator TD : L2 (D) → H 1 (D). ∗ The Bogovski˘ı potential map from Definition 10.4.2 maps TD : C0∞ (D) → C0∞ (D) and extends by continuity to a bounded operator ∗ TD : L2 (D) → H01 (D).

We have homotopy relations d(TD F ) + TD (dF ) + KD F = F, ∗ ∗ −δ(TD F = F, F ) − T ∗ (δF ) + KD

F ∈ D(d), F ∈ D(δ),

with perturbation terms KD F (x) =

X

Z ηα (x)

θα F0 dy +

α ∗ KD F (x) =

X α

X

∇ηα (x) ∧ TDα (F |Dα )(x),

α

Z θα (x)

ηα F0 dy +

X

∗ TD (∇ηα y F |Dα )(x), α

α

∗ which are bounded, KD : L2 (D) → H 1 (D) and KD : L2 (D) → H01 (D). Here F0 0 denotes the ∧ V part of F .

365

10.4. Bogovski˘ı and Poincar´e Potentials

To see how Theorem 10.4.3 implies the existence of Bogovski˘ı and Poincar´e potentials, we consider the following Hodge decomposition: L2 (D)

L2 (D)

⊕

R(d)

=

=

R(d)

t

Ck (D)

δ

d

⊕

⊕

Ck (D)

⊕

*

R(δ)

(10.11)

R(δ)

Given F ∈ R(d), we apply the homotopy relation to the Hodge potential U ∈ R(δ), with dU = F , to obtain U = dTD U + TD dU + KD U , and in particular, F = dU = d(TD F + KD U ). ˜ := TD F + KD U ∈ H 1 (D) is a regularized Poincar´e potential Therefore the field U for F . Similarly, for F ∈ R(δ) we apply the homotopy relation to the Hodge potential U ∈ R(d), with δU = F , to obtain ∗ ∗ F = δU = δ(−TD F + KD U ),

˜ := −T ∗ F + K ∗ U ∈ H 1 (D) is a Bogovski˘ı potential for F . where the field U 0 D D Proof of Theorem 10.4.3. (i) Let F ∈ C ∞ (D). Then F |Dα ∈ C ∞ (Dα ), and we see from (10.8) for the star-shaped domain Dα that TDα (F |Dα ) ∈ C ∞ (Dα ). Note that TDα acts on C ∞ (X), but the values TDα F (x), for x ∈ Dα , depend only on F |D . With the partition of unity ηα , we obtain TD F ∈ C ∞ (D). Let F ∈ C0∞ (D). Then ηα F |Dα ∈ C0∞ (Dα ), and we see from Exercise 10.4.1 ∗ that supp TD (ηα F |Dα ) is compactly contained in Dα . To verify smoothness, we α write Z ∞ Z ∗ k−1 n−k TDα G(x) = − z ∧ G(x − z) θα (x + sz)s (1 + s) ds dz. Dα

0

∗ Differentiation with respect to x shows that TD (ηα F |Dα ) ∈ C0∞ (Dα ), and thereα ∗ ∞ fore that TD F ∈ C0 (D). Averaging the homotopy relation in Exercise 7.5.6, we obtain

d(TDα F ) + TDα (dF ) + KDα F = F on Dα , with

Z KDα F :=

θα F0 dx.

As in the proof of Theorem 10.3.1, the product rule yields d(TD F ) + TD (dF ) + KD F = F on D, and duality yields the stated formulas for δ. ∗ . To this end, assume (ii) It remains to establish H 1 bounds for TDα and TD α that D = Dα is star-shaped with respect to a ball and consider the operators (10.9)

Chapter 10. Hodge Decompositions

366

and (10.10). By Exercise 10.4.1, these are weakly singular integral operators, and Schur estimates as in Exercise 6.4.3 show that TD is bounded on L2 (D). Expanding (1 + s)n−k with the binomial theorem, we may further replace kθ (x, y) by Z ∞ θ(y + s(y − x))sn−1 ds. 0

Indeed, in estimating k∇⊗TD F kL2 the difference will be a weakly singular operator ∗ that is bounded as above, and similarly for k∇ ⊗ TD F kL2 . Make the change of variables t = s|y − x|, fix a coordinate 1 ≤ i ≤ n, and define Z ∞ z n−1 zi η x+t k(x, z) := t dt . n |z| |z| 0 Estimating the multivector fields componentwise, we see that it is enough to consider the operators Z Z ∗ Sf (x) := k(y, y − x)f (y) dy and S f (x) := k(x, x − y)f (y) dy, D

D ∗

and prove bounds on k∇Sf kL2 and k∇S f kL2 . We note that k(x, z) is homogeneous of degree −n + 1 with respect to z. For fixed x, we expand k(x, z/|z|) in a series of spherical harmonics on the unit sphere S. We get k(x, z) =

1 |z|n−1

hj ∞ X X

kjm (x)Yjm (z/|z|) =

j=0 m=1

hj ∞ X X

kjm (x)

j=0 m=1

Yjm (z) , |z|n−1+j

h

j denotes an ON-basis for the space Pjsh (S) of scalar-valued spherwhere {Yjm }m=1 ical harmonics, for j ∈ N. See Section 8.2. In particular the coefficients are R kjm (x) := S k(x, z)Yjm (z) dz. Define weakly singular convolution integral operators Z Yjm (x − y) f (y) dy. Sjm (x) := |x − y|n−1+j D

With kjm as multipliers we have Sf (x) =

hj ∞ X X (−1)j Sjm (kjm f )(x), j=0

m=1

S ∗ f (x) =

hj ∞ X X

kjm (x)Sjm f (x).

j=0 m=1

The main estimate we need is kSjm kL2 (D)→H 1 (D) . (1 + j)n−2 .

(10.12)

To see this, we use zonal harmonics as in Section 8.2 to estimate Z |Yjm (z)| = Zj (z, y)Yjm (y)dy ≤ kZj (z, ·)kL2 (S) kYjm kL2 (S) . (1 + j)n−2 |z|j , S

ˇ Cohomology 10.5. Cech

367

which yields the L2 estimate. To bound ∇Sjm f on L2 (X), we use Proposition 6.2.1 to see that ∇Sjm is a Fourier multiplier with estimates ξ2c Γ((1 + j)/2) Yjm (ξ)/|ξ|1+j . (1 + j)n−2 , ξ ∈ X, Γ((n − 1 + j)/2) of the symbol. This proves (10.12). To estimate the multipliers Z kjm (x) =

k(x, z)Yjm (z)dz, S

we use that k(x, ·) is smooth on S, while Yjm becomes more oscillatory as j grows, to show that kjm decays with j as follows. By Proposition 8.2.15, the spherical Laplace operator ∆S is a self-adjoint operator in L2 (S) with ∆S Yjm = (2 − n − j)jYjm . Using self-adjointness N times shows that Z 1 kjm (x) = (∆N k(x, z))Yjm (z)dz. (2 − n − j)N j N S S Since ∆N S k(x, ·), for any fixed N , is bounded, we get the estimate |kjm (x)| . (1 + j)−2N . Similarly, we bound Z ∇kjm (x) =

∇x k(x, z)Yjm (z)dz S

uniformly by (1 + j)−2N . Collecting our estimates, we obtain kSf kH 1 (D) .

∞ X

hj (1 + j)n−2 (1 + j)−N kf kL2 (D) . kf kL2 (D) ,

j=0

kS ∗ f kH 1 (D) .

∞ X

hj (1 + j)−N (1 + j)n−2 kf kL2 (D) . kf kL2 (D) ,

j=0

provided we fix large enough N . This completes the proof.

ˇ 10.5 Cech Cohomology In this section we collect some tools from algebraic topology that we use in Section 10.6 to calculate the dimensions of the finite-dimensional cohomology space

Chapter 10. Hodge Decompositions

368

N(d) ∩ N(δ ), more precisely the Betti numbers bk (D), from Definition 7.6.3. We also use these tools in Chapters 11 and 12. Our starting point is the notion of a sheaf, where we only use the following simplified version of this concept. We consider some set D and some fixed finite covering of it by subsets D1 , . . . , DN , so that D = D1 ∪ · · · ∪ DN . By a sheaf F on D we mean a collection of linear spaces F(D0 ), one for each intersection D0 of the subsets Dj . In fact, it is only the additive structure of ˇ sheaves that is relevant, and in Chapter 11 we shall use Cech cohomology, where the spaces F(D0 ) are the smallest additive group Z2 = {0, 1}. The linear spaces that we use in this chapter are supposed to behave like spaces of functions defined on D0 in the sense that we require that there exist linear restriction maps F(D0 ) → F(D00 ) : f 7→ f |D00 whenever D00 ⊂ D0 ⊂ D. If an intersection D0 is empty, then we require that the linear space F(D0 ) be trivial, that is, F(D0 ) = {0}. The intersections Ds = Ds1 ∩ · · · ∩ Dsk , of distinct subsets Dsj that we consider, are indexed by the 2N subsets s = ˇ {s1 , . . . , sk } ⊂ {1, . . . , N }. Since the Cech algebra that we are about to construct is alternating, we choose below to index the intersections not by s, but by auxiliary basis multivectors es in ∧RN . This is only a formal notation, which turns out to be useful, since it allows us to recycle some, by now well known to us, exterior algebra. Definition 10.5.1 (k-cochains). Let F be a sheaf on D as above, with covering ˇ D = {D1 , . . . , DN }. A Cech k-cochain f associates to each (k + 1)-fold intersection Ds , |s| = k + 1, an element in the linear space F(D0 ), which we denote by hf, es i ∈ F(Ds ). This is not an inner product, but only a convenient notation for the value of f on Ds . We also extend the definition of f homogeneously by letting hf, αes i := αhf, es i, for α ∈ R. ˇ The space of all Cech k-cochains f on D with values in F is denoted by k k C (D; F). Viewing C (D; F) as ⊕s:|s|=k+1 F(Ds ), it is clear that this is a linear space. For k < 0 and k ≥ N we let C k (D; F) := {0}. ˇ The Cech coboundary operator ∂k : C k (D; F) → C k+1 (D; F) is the linear map defined by h∂k f, es i :=

N X hf, ej y es i|Ds ,

|s| = k + 2, f ∈ C k (D; F).

j=1

For k < 0 and k ≥ N − 1, we let ∂k = 0. ˇ We will see that Cech k-cochains and ∂k behave in many ways like k-covector fields and the exterior derivative d. We need some terminology.

ˇ 10.5. Cech Cohomology

369

Definition 10.5.2 (Complex of spaces). A complex of linear spaces is a sequence of linear maps between linear spaces ∂j−2

∂j−3

∂j−1

∂j+1

∂j

∂j+2

→ Vj−2 → Vj−1 → Vj → Vj+1 → Vj+2 →

such that R(∂j−1 ) ⊂ N(∂j ) in Vj . The complex is said to be exact at Vj if R(∂j−1 ) = N(∂j ). If it is exact at all Vj , we say that the complex is exact. More generally, the cohomology of the complex at Vj is the quotient space H j (V ) := N(∂j )/R(∂j−1 ). An important special case occurs when Vj = {0} for some j, so that ∂j = ∂j−1 = 0. In this case, exactness at Vj+1 means that ∂j+1 is injective, and exactness at Vj−1 means that ∂j−2 is surjective. Lemma 10.5.3. If (Vj , ∂j ) is a an exact complex of finite-dimensional linear spaces and Vj1 = Vj2 = 0, then X (−1)j dim Vj = 0. j1 <j<j2

Proof. The dimension theorem for linear maps shows that dim Vj = dim N(∂j ) + dim R(∂j ). Since R(∂j ) = N(∂j+1 ), we get a telescoping sum X X (−1)j (dim N(∂j ) + dim N(∂j+1 )) (−1)j dim Vj = j1 <j<j2

j1 <j<j2

= (−1)j1 +1 dim R(∂j1 ) + (−1)j2 −1 dim N(∂j2 ) = 0.

ˇ ˇ Lemma 10.5.4 (Cech sequence complex). The Cech ∂k−2

∂k−1

∂k+1

∂

∂k+2

k → C k−1 (D; F) → C k (D; F) → C k+1 (D; F) → C k+2 (D; F) →

is a complex of linear spaces. Proof. Let f ∈ C k (D; F) and |s| = k + 3. Then X XX h∂k+1 ∂k f, es i = h∂k f, ej y es i|Ds = hf, ei y (ej y es )i|Ds\{j} j

=

X

j

hf, (ej

∧ ei )

i

Ds

y es i|Ds = 0,

i,j

since ei ∧ ej = −ej

∧ ei

on performing the sum.

ˇ We denote the Cech cohomology spaces associated with this complex by k H (D; F). A key result that we now prove is, roughly speaking, that sheaves of functions defined without any constraints will have trivial cohomology spaces. A constraint here could mean that we consider functions that are constant or that satisfy some differential equation. More precisely, a sheaf F is defined to be a fine sheaf if every sufficiently smooth cutoff function η : D → R gives well-defined multiplication operators f 7→ ηf on each of the linear spaces F(D0 ). In particular, if supp η ⊂ D0 , then ηf is supposed to be extendable by zero to a function ηf ∈ F(D) on all D. When restricted to some D00 , this defines ηf ∈ F(D00 ).

Chapter 10. Hodge Decompositions

370

Proposition 10.5.5 (Cohomology of fine sheaves). If F is a fine sheaf on D, then H k (D; F) = {0} when k ≥ 1. For any sheaf F, the restriction map gives an invertible map F(D) → H 0 (D; F). Proof. First consider the second claim. If f ∈ C 0 (D; F) and ∂0 f = 0, then for all 1 ≤ i < j ≤ N , we have 0 = h∂0 f, e{i,j} i = hf, ej i|D{i,j} − hf, ei i|D{i,j} . Thus there is a unique function f ∈ F(D) such that f |Dk = hf, ek i. Since ∂−1 = 0, this proves the statement. Now let F be a fine sheaf, k ≥ 1, and f ∈ C k (D; F) with ∂k f =P 0. Pick a partition of unity {ηj }N subordinate to D, so that supp η ⊂ D and j j j=1 j ηj = 1 on D. Define a (k − 1)-cochain hg, et i :=

X

ηi hf, ei ∧ et i,

|t| = k.

i

Note that hf, ei ∧ et i is defined only on Dt ∩ Di , but that after multiplication by ηi , the product can be extended by zero across (∂Di ) ∩ Dt to all Dt . The anticommutation relation from Theorem 2.8.1 yields X XX h∂k−1 g, es i = hg, ej y es i|Ds = ηi hf, ei ∧ (ej y es )i j

=

j

XX j

Ds

i

ηi (δi,j hf, es i − hf, ej y (ei ∧ es )i)

i

Ds

X X X = ηi hf, es i − ηi hf, ej y (ei ∧ es )i|Di ∩Ds |Ds i

i

j

= hf, es i − 0 = hf, es i, where δi,j = 1 if i = j and otherwise 0. This shows that N(∂k ) = R(∂k−1 ), as desired. We finish this section with two algebraic techniques that are useful in studying complexes. By a diagram of maps being commutative, we mean that whenever we have two different compositions of maps A → D, A

f1

/B

f4

/D

f2

C then we have f3 ◦ f1 = f4 ◦ f2 .

f3

ˇ 10.5. Cech Cohomology

371

Lemma 10.5.6 (Snake lemma). Let (Uj , ∂ju ), (Vj , ∂jv ) and (Wj , ∂jw ) be complexes of linear spaces, and for each j, let fj

gj

0 → Uj → Vj → Wj → 0 be a short exact sequence such that ∂jv fj = fj+1 ∂ju and ∂jw gj = gj+1 ∂jv for all j. This hypothesis is summarized in the following commutative diagram with exact columns: 0 ···

···

0

u ∂j−3

/ Uj−2

v ∂j−3

u ∂j−2

/ Uj−1

v ∂j−2

fj−2

/ Vj−2

/ Vj−1

···

/ Wj−2

/ Uj

v ∂j−1

/ Vj

/ Wj−1

0

/ Uj+1

∂jv

/ Vj+1

0

/ Wj

∂jw

u ∂j+1

/ Uj+2

v ∂j+1

fj+1

gj w ∂j−1

0

∂ju

fj

gj−1 w ∂j−2

0

u ∂j−1

fj−1

gj−2 w ∂j−3

0

0

/

··· /

··· /

···

fj+2 v ∂j+2

/ Vj+2

gj+1

/ Wj+1

u ∂j+2

gj+2 w ∂j+1

w ∂j+2

/ Wj+2

0

0

˜ j : H j (W ) → H j+1 (U ) and induced linear Then there are connecting linear maps h j j ˜ maps fj : H (U ) → H (V ) and g˜j : H j (V ) → H j (W ), for all j, such that the cohomology sequence H j−2 C (U )

˜ j−3 h

f˜j−2

H

j−2

(V )

g ˜j−2

H

j−2

˜ j−2 h

H j−1 A (U ) f˜j−1

H

j−1

(V )

g ˜j−1

(W )

H

j−1

˜ j−1 h

H jB (U )

j

f˜j

H (V ) g ˜j

(W )

j

H (W )

˜j h

H j+1 B (U ) f˜j+1

H

j+1

(V )

g ˜j+1

H

j+1

˜ j+1 h

H j+2 A (U ) f˜j+2

H

j+2

g ˜j+2

(W )

C ˜ j+2 h

(V )

H j+2 (W )

is an exact complex. Exercise 10.5.7 (Diagram chasing). Prove the snake lemma through diagram chasing. To see an example of this, consider the definition of the connecting map ˜ j : H j (W ) = N(∂ w )/R(∂ w ) → H j+1 (U ) = N(∂ u )/R(∂ u ). Take w ∈ N(∂ w ). h j j−1 j+1 j j Surjectivity of gj gives v ∈ Vj , which maps to a v 0 ∈ Vj+1 . Commutativity ∂jw gj = gj+1 ∂jv shows that v 0 maps to 0 in Wj+1 , and so exactness at Vj+1 gives u ∈ Uj+1 such that fj+1 (u) = v 0 . Since v 0 maps to 0 in Vj+2 , commutativity

Chapter 10. Hodge Decompositions

372

v u fj+1 = fj+2 ∂j+1 ∂j+1 shows that u ∈ N(∂j+1 ), since fj+2 is injective.

u

/0

v

/ v0

/ 0

w

/ 0

˜ j ([w]) := [u]. Further diagram chasing shows The connecting map is defined as h u w ˜ j is a well-defined map between cohomolthat u ∈ R(∂j ) if w ∈ R(∂j−1 ), so that h ogy spaces. Through diagram chasing, one can similarly prove the following useful result. Lemma 10.5.8 (Five lemma). Consider the following commutative diagram of linear spaces and linear maps, where the two rows form complexes: U1

f1

h1

V1

/ U2

f2

h2

g1

/ V2

/ U3

f3

h3

g2

/ V3

/ U4

f4

h4

g3

/ V4

/ U5 h5

g4

/ V5

(i) If the row complexes are exact at U3 and V2 , and if h2 , h4 are injective and h1 is surjective, then the middle map h3 is injective. (ii) If the row complexes are exact at V3 and U4 , and if h5 is injective and h2 , h4 are surjective, then the middle map h3 is surjective. In particular, if the row complexes are exact, and if h1 , h2 , h4 , and h5 are invertible, then h3 is also invertible.

10.6

De Rham Cohomology

Let D be a bounded Lipschitz domain in an n-dimensional Euclidean space X, and consider the operators d and δ = −d∗ in L2 (D). In this section, we study the finite-dimensional subspace Ck (D) in the Hodge decomposition L2 (D) = R(d) ⊕ Ck (D) ⊕ R(δ). Definition 10.6.1 (De Rham cohomology spaces). Let D be a bounded Lipschitz domain. The De Rham cohomology spaces are the quotient spaces H k (D) := N(d; ∧k )/R(d; ∧k ). We identify their direct sum with H(D) := N(d)/R(d).

10.6. De Rham Cohomology

373

ˇ These cohomology spaces H k (D) should not be confused with the Cech cohomology spaces H k (D; F) from Section 10.5. We shall, however, show in this section that for the sheaf F = R they are indeed closely related. We note that the following spaces essentially are the same. The last three are indeed equal. • The De Rham space H(D). • The dual quotient space N(δ)/R(δ). • The intersection of the Hodge subspaces Ck (D) = N(d) ∩ N(δ). • The null space of the Hodge–Dirac operator d + δ from Proposition 10.1.3. • The null space of the Hodge–Laplace operator (d + δ)2 = dδ + δd. Note that the orthogonal complement Ck (D) of R(d) in N(d) is different from, but can be identified with the quotient space H(D). One can show that for smooth domains, Ck (D) is a subspace of C ∞ (D). For Lipschitz domains this is not true, but using the potential maps from Section 10.4, one can show, at least for strongly Lipschitz domains, that there is another complement of R(d) in N(d) that is contained in C ∞ (D). This means that the de Rham cohomology space H(D) can be represented by C ∞ (D) fields. The cohomology space Ck (D) splits into its homogeneous k-vector parts Ck (D; ∧k ), and it is our aim in this section to calculate the Betti numbers bk (D) = dim Ck (D; ∧k ) = dim H k (D) from Definition 7.6.3 for a given domain D. On a domain D with boundary, we can similarly consider the cohomology space C⊥ (D) with normal boundary conditions. But by Hodge star duality as in Proposition 10.2.2, we have dim C⊥ (D; ∧k ) = dim Ck (D; ∧n−k ) = bn−k (D). For the remainder of this section, we therefore consider only tangential boundary conditions. The following observation shows that the Betti numbers do not depend on the geometry, but only on the topology of the domain. Proposition 10.6.2 (Topological invariance). Let ρ : D1 → D2 be a Lipschitz diffeomorphism between bounded Lipschitz domains in Euclidean space. Then the pullback induces an invertible linear map ρ∗ : H(D2 ) → H(D1 ). In particular, bk (D1 ) = bk (D2 ) for all k = 0, 1, 2, . . . , n. Note that the pullback will not in general map between R(d)⊥ = N(δ) for the two domains, and that consequently the spaces Ck (D, ∧k ) depend on the geometry of D. It is only the dimensions that are topological invariants.

374

Chapter 10. Hodge Decompositions

Proof. The result is immediate from Lemma 10.2.6, which shows that ρ∗ yields invertible maps between the ranges R(d) as well as between the null spaces N(d) for the two domains. The Betti numbers give information about the topology of the domain, the simplest example being b0 (D) = number of connected components of D, which is clear, since dF = 0 for a scalar function means that F locally constant. Note also that by imposing Dirichlet boundary conditions dF = 0 for a scalar function forces F = 0. By Hodge star duality, this translates to bn (D) = 0 for every bounded Lipschitz domain in Euclidean space. The heuristic is that for general k, the Betti number bk (D) measures the number of k-dimensional obstructions in D. For the topologically trivial domains, those that are Lipschitz diffeomorphic to the ball, the following result is clear from Poincar´e’s theorem (Theorem 7.5.2) and its L2 extension in Section 10.4. Proposition 10.6.3 (Cohomology for balls). If D is Lipschitz diffeomorphic to a ball, then b0 (D) = 1 and bk (D) = 0 for k = 1, 2, . . . , n. To calculate the Betti numbers for a general domain, we cover D by subsets Dj , all diffeomorphic to balls, such that D = D1 ∪ · · · ∪ DN . We also require that all intersections be topologically trivial as follows. Definition 10.6.4 (Good cover). Let D be a bounded Lipschitz domain, and assume that D = D1 ∪ · · · ∪ DN is a finite cover of D by open subsets. We say that D = {Dj }N j=1 is a good cover of D if all nonempty intersections Ds = Ds1 ∩· · ·∩Dsk are Lipschitz diffeomorphic to balls. We use the algebra from Section 10.5 and three simple examples of sheaves F on D. • The sheaf D(d; ∧k ), where the Hilbert space associated with an open set D0 ⊂ D consists of k-vector fields F ∈ L2 (D0 ) such that dF ∈ L2 (D0 ). • The sheaf N(d; ∧k ), where the Hilbert space associated with an open set D0 ⊂ D consists of k-vector fields F ∈ L2 (D0 ) such that dF = 0. • The sheaf R, where a real line is associated with an open set D0 ⊂ D, and restriction is the identity map.

375

10.6. De Rham Cohomology

Note that D(d; ∧k ) is a fine sheaf, but not N(d; ∧k ) or R. Note also that we cannot use δ here to define sheaves, since restriction does not preserve the boundary conditions. The main result of this section is the following characterization of the Betti numbers. ˇ Theorem 10.6.5 (De Rham = Cech). Let D be a bounded Lipschitz domain in Euclidean space, and let D be a good cover as in Definition 10.6.4. Then bk (D) = dim H k (D; R),

k = 0, 1, . . . , n,

(10.13)

ˇ where H k (D; R) is the Cech cohomology space for the constant sheaf R. This shows in particular that the Betti numbers do not depend on the exterior and interior derivative operators, since the right-hand side in (10.13) does not. Conversely, dim H k (D; R) does not depend on the choice of good cover, since the left-hand side in (10.13) does not. Proof. Consider the following sequence of maps on Ds : i

d

0 → N(dDs ; ∧k ) → D(dDs ; ∧k ) → N(dDs ; ∧k+1 ) → 0, where i denotes inclusion. Assuming that the intersection Ds is Lipschitz diffeomorphic to a ball, we know that this is an exact complex for k ≥ 0. Acting componentwise, this induces an exact complex i

d

0 → C j (D; N(d; ∧k )) → C j (D; D(d; ∧k )) → C j (D; N(d; ∧k+1 )) → 0. Consider the following commutative diagram: 0

···

···

···

0

∂j−2

/ C j−1 (D; N(d; ∧k ))

∂j−2

∂j−2

/

∂j−1

/ C j (D; N(d; ∧k ))

∂j−1

i

/ C j−1 (D; D(d; ∧k ))

0

∂j−1

/

∂j

/ C j+1 (D; N(d; ∧k ))

∂j

i

/ C j (D; D(d; ∧k ))

d

C j−1 (D; N(d; ∧k+1 ))

0

0

∂j

/ C j+1 (D; D(d; ∧k )) /

/

···

∂j+1

/

···

∂j+1

/

···

i

d

C j (D; N(d; ∧k+1 ))

∂j+1

d

C j+1 (D; N(d; ∧k+1 ))

0

According to Proposition 10.5.5, the cohomology spaces H j (D; D(d; ∧k )) for the second row vanish when j ≥ 1. The exact cohomology complex provided by the snake lemma (Lemma 10.5.6), thus splits into exact sequences 0 → H j (D; N(d; ∧k+1 )) → H j+1 (D; N(d; ∧k )) → 0

376

Chapter 10. Hodge Decompositions

for k ≥ 0, j ≥ 1, and for j = 0 the exact complex 0 → N(d; ∧k ) → D(d; ∧k ) → N(d; ∧k+1 ) → H 1 (N(d; ∧k )) → 0. This shows that dim H j+1 (D; N(d; ∧k )) = dim H j (D; N(d; ∧k+1 )) for j ≥ 1, and dim H 1 (N(d; ∧k )) = dim(N(d; ∧k+1 )/dD(d; ∧k )) = bk+1 (D). Thus for k ≥ 1, we get dim H k (D; R) = dim H k (D; N(d; ∧0 )) = dim H k−1 (D; N(d; ∧1 )) = dim H k−2 (D; N(d; ∧2 )) = · · · = dim H 1 (D; N(d; ∧k−1 )) = bk (D), since the sheaves N(d; ∧0 ) and R coincide. Proposition 10.5.5 shows that H 0 (D; R) = R(D), which equals the space Ck (D; ∧0 ) of locally constant functions on D. Thus dim H 0 (D; R) = b0 (D), which completes the proof. Theorem 10.6.5 reduces the computation of the Betti numbers to a finite problem, although the construction of a good cover can be nontrivial. Note that we started by defining the Betti numbers as the dimension of the finite-dimensional space N(d; ∧k )/R(d; ∧k ). However, note that both the numerator and denominator are infinite-dimensional Hilbert spaces in general. On the other hand, we have now characterized the Betti numbers as the dimensions of the spaces N(∂k ; C k (D; R))/R(∂k−1 ; C k (D; R)), where all spaces involved are finite-dimensional. Example 10.6.6 (Annulus). The simplest domain with nontrivial topology is the two-dimensional annulus D = {x = (x1 , x2 ) ; r < |x| < R}. We see that a good cover of D requires three subsets D1 , D2 , D3 . For example, D1 := {x ∈ D ; x2 > 0}, D2 := {x ∈ D ; x1 > x2 }, and D3 := {x ∈ D ; x1 + x2 < 0} give a good cover. The nonempty intersections are D1 , D2 , D3 , D12 , D13 , and D23 . We see that C 0 (D; R) is a three-dimensional space, with a basis (ω1 , ω2 , ω3 ), where hωi , ei i = 1, and 0 on the other subsets. Similarly, C 1 (D; R) is a three-dimensional space, with a basis (ω12 , ω13 , ω23 ), where hωij , eij i = 1, and 0 on the other subsets. By Definition 10.5.1, the matrix for ∂0 is −1 1 0 −1 0 1 . 0 −1 1 This has a one-dimensional null space, so that b0 (D) = dim N(∂0 ) = 1. Since ∂0 has two-dimensional range and ∂1 = 0, since D123 = ∅, we get b1 (D) = dim N(∂1 ) − dim R(∂0 ) = 3 − 2 = 1. We have shown that the Betti numbers for D are bi (D) = (1, 1, 0).

377

10.6. De Rham Cohomology

Figure 10.2: (a) Vector field F to be decomposed. (b) The cohomology part H1 of F in the two-dimensional space Ck (D). By Exercise 10.6.8, the cohomology part of any F is a linear combination of this H1 and H2 appearing in Figure 10.3. (c) The curl-free part ∇U of F . (d) The divergence-free part ∇ y (V j) of F , with tangential boundary conditions.

Exercise 10.6.7 (3D spherical shell). Show by constructing a good cover of the three dimensional spherical shell D := {x = (x1 , x2 , x3 ) ; r < |x| < R} with four subsets that bi (D) = (1, 0, 1, 0). Exercise 10.6.8 (General plane domain). For a plane domain consisting of m disks with ni smaller interior disks removed in their respective disks, i = 1, . . . , m,

378

Chapter 10. Hodge Decompositions

Figure 10.3: (a) Vector field F to be decomposed. (b) The cohomology part H2 of F in the two-dimensional space Ck (D). By Exercise 10.6.8, the cohomology part of any F is a linear combination of this H2 and H1 appearing in Figure 10.2. (c) The curl-free part ∇U of F . (d) The divergence-free part ∇ y (V j) of F , with tangential boundary conditions. construct a good cover and show that the Betti numbers are m X bi (D) = m, ni , 0 . i=1

The case m = 1, n = n1 = 2 is illustrated in Figures 10.2 and 10.3. The disadvantage with Theorem 10.6.5 is that the construction of good covers soon gets complicated in higher dimensions. We therefore discuss two complemen-

379

10.6. De Rham Cohomology

tary techniques for computing Betti numbers: The Mayer–Vietoris sequence and unneth formula. These enable us to calculate Betti numbers for unions and the K¨ Cartesian products of domains. Theorem 10.6.9 (Mayer–Vietoris sequence). Let D1 , D2 be bounded Lipschitz domains such that D1 ∪ D2 and D1 ∩ D2 also are Lipschitz domains. Then we have the following exact complex, the Mayer–Vietoris sequences: H k−17 (D1 ∪ D2 )

H k (D : 1 ∪ D2 )

H k−1 (D1 ) ⊕ H k−1 (D2 )

H k+1 (D : 1 ∪ D2 )

H k (D1 ) ⊕ H k (D2 )

H k+1 (D1 ) ⊕ H k+1 (D2 )

H k−1 (D1 ∩ D2 )

7

H k (D1 ∩ D2 )

H k+1 (D1 ∩ D2 )

Proof. Consider the complex j

i

0 → D(dD1 ∪D2 ; ∧k ) → D(dD1 ; ∧k ) ⊕ D(dD2 ; ∧k ) → D(dD1 ∩D2 ; ∧k ) → 0, where i(f ) := (f |D1 , f |D2 ) is restriction, and j is the map j(g1 , g2 ) := g1 |D1 ∩D2 − g2 |D1 ∩D2 . We see that i is injective and that the sequence is exact at the middle space. To verify that j is surjective, take h ∈ D(dD1 ∩D2 ; ∧k ). Proposition 10.2.7 shows that we can extend h to g ∈ D(dD1 ; ∧k ) such that g|D1 ∩D2 = h. Then j(g, 0) = h. The stated Mayer–Vietoris sequence is obtained by applying the snake lemma (Lemma 10.5.6) with ∂ = d to this complex. Theorem 10.6.10 (K¨ unneth formula). Let D1 and D2 be bounded Lipschitz domains in Euclidean spaces X1 and X2 respectively, not necessarily of the same dimension. Then the Betti numbers of the Cartesian product D1 × D2 ⊂ X1 × X2 are given by the K¨ unneth formula bk (D1 × D2 ) =

k X

bj (D1 ) bk−j (D2 ).

j=0

Proof of Theorem 10.6.10. Let pi : D1 × D2 → Di , i = 1, 2, denote the coordinate projections and use pullbacks to form the bilinear map (f1 (x1 ), f2 (x2 )) 7→ (p∗1 (f1 ) ∧ p∗2 (f2 ))(x1 , x2 ) = f1 (x1 ) ∧ f2 (x2 ), where fi : Di → ∧ki Vi . Note that d(p∗1 f1 ∧ p∗2 f2 ) = p∗1 (df1 ) ∧ p∗2 f2 + (−1)k1 p∗1 f1 ∧ p∗2 (df2 ). This shows that p∗1 f1 ∧ p∗2 f2 is closed if f1 and f2 are closed, and that it is exact if one of the factors is exact and the other factor is closed. Using the universal property for tensor products, we obtain a well-defined linear map H k1 (D1 ) ⊗ H k2 (D2 ) → H k1 +k2 (D1 ×D2 ) : [f1 (x1 )] ⊗ [f2 (x2 )] 7→ [f1 (x1 ) ∧ f2 (x2 )].

380

Chapter 10. Hodge Decompositions

Assume that Di have a good cover by Ni subsets, i = 1, 2. We prove the K¨ unneth formula by induction on N1 + N2 , the induction hypothesis being that the map M (H i (D1 ) ⊗ H k−i (D2 )) → H k (D1 × D2 ), i

defined as the direct sum of the maps above, is invertible for all fixed k. Evaluating the dimensions will then give the K¨ unneth formula. If N1 + N2 = 2, then D1 , D2 , and D1 × D2 are all diffeomorphic to balls in the respective spaces, and the induction hypothesis is readily verified. For the induction step, write, for example, D1 = U ∪ V , where U is one of the sets in a good cover of D1 and V is the union of the rest. The induction hypothesis applies to U × D2 , V × D2 , and (U ∩ V ) × D2 , since both V and U ∩ V have good covers by at most N1 − 1 sets. Consider the diagram / H k−1 (U × D2 ) ⊕ H k−1 (V × D2 ) (H i−1 (U ) ⊗ H k−i (D2 )) ⊕ (H i−1 (V ) ⊗ H k−i (D2 ))

/ H k−1 ((U ∩ V ) × D2 )

/ H k ((U ∪ V ) × D2 )

H i−1 (U ∩ V ) ⊗ H k−i (D2 )

H i (U ∪ V ) ⊗ H k−i (D2 )

(H i (U ) ⊗ H k−i (D2 )) ⊕ (H i (V ) ⊗ H k−i (D2 ))

/ H k (U × D2 ) ⊕ H k (V

H i (U ∩ V ) ⊗ H k−i (D2 )

× D2 )

/ H k ((U ∩ V ) × D2 ).

The horizontal maps are defined as above in the natural way, and the vertical maps come from the Mayer–Vietoris sequence. On the left, we have taken the tensor product by H k−i (D2 ), and the maps act trivially in this factor. For the first and fourth rows we have a distributional rule for tensor products and direct sums. It is straightforward to verify that this diagram commutes. Taking the direct sum over i, we can apply the induction hypothesis and the five lemma (Lemma 10.5.8), which proves the theorem. Example 10.6.11 (Rn annulus). Consider an n-dimensional annulus An = {x = (x1 , . . . , xn ) ; 2 < |x| < 3}, and let An+ := An ∩ {xn > −1} and An− := An ∩ {xn < 1}. Then An± are both diffeomorphic to the n-dimensional unit ball B n , and An+ ∩ An− is diffeomorphic to An−1 × (0, 1). We prove by induction that bi (An ) = (1, 0, 0, . . . , 0, 0, 1, 0), so that bn−1 (An ) = 1. The cases n = 2, 3 are known from Example 10.6.6 and Exercise 10.6.7. For n > 3, the Mayer–Vietoris sequence and K¨ unneth formula give exact sequences 0 → H k−1 (An−1 ) → H k (An ) → 0

10.7. Comments and References

381

for k ≥ 2 and using that b1 (An ) = 1, we have for k = 1 that 0 → R → R2 → R → H1 (An ) → 0 is exact. This proves the stated formula for the Betti numbers. Note that x/|x|n is a divergence and curl-free vector field that is normal on the boundary of the annulus. Applying the Hodge star map, it follows that Ck (An ; ∧n−1 ) is spanned by the tangential (n − 1)-vector field ∗x/|x|n . Exercise 10.6.12 (3D cohomology). Let D be a three-dimensional ball with m smaller balls and n cylinders removed, all of these disjoint. Show by inductively applying a Mayer–Vietoris sequence that bi (D) = (1, n, m, 0).

10.7

Comments and References

10.1 The formulation of Hodge decompositions presented here is based on the survey paper by Axelsson and McIntosh [14]. The notation Γ for an abstract nilpotent operator, generalizing d, was introduced there, along with Π = Γ ± Γ∗ for an abstract Hodge–Dirac operator. Playing with symbols, Γ∗ can be viewed as a backward Γ, and together they combine to a Π. As in Sections 9.3–9.6, a main philosophy in this chapter is to handle Hodge decomposition by first-order operators. Concretely, this means that we study Hodge decompositions as far as possible using Γ, Γ∗ , and Π, which in applications are always first-order differential operators, rather than involving the abstract Laplace operator Π2 . The latter is standard in the literature, but may sometimes complicate the problems. A concrete example is the proof of Theorem 10.3.1. The reason that this works is that we work with the firstorder operator Γ = d for which the commutation theorem is available. A reference for the classical variational second-order approach to Hodge decompositions is Morrey [72]. 10.2 We have two natural choices of boundary conditions leading to skew-adjoint Hodge–Dirac operators d + δ and d + δ respectively. It is standard to consider the second-order Laplace operators. Here the generalized Dirichlet Laplacian (d + δ)2 = dδ + δd is said to have relative boundary conditions, and the generalized Neumann Laplacian (d + δ)2 = dδ + δd is said to have absolute boundary conditions. See, for example, [91]. For geometrical reasons, we use the terminology normal and tangential, rather than relative and absolute, in this book. 10.3 The beautiful and simple reduction from Lipschitz domains to smooth domains using the commutation theorem in Theorem 10.3.1 appears in the work by R. Picard [74].

382

Chapter 10. Hodge Decompositions For strongly Lipschitz domains there is a singular integral proof, based on Theorems 8.3.2 and 9.5.1, showing that D(d+δ ) and D(d+δ) are contained in H 1/2 (D). This exponent 1/2 is sharp for the class of strongly Lipschitz domains. See [14].

10.4 These regularity and support properties of potentials where proved by Mitrea, Mitrea, and Monniaux [70] and Costabel and McIntosh [31]. In fact, [31] ∗ shows the stronger result that the constructed operators TD and TD are pseudodifferential operators of order −1. By well-known estimates for such operators, bounds on a large number of scales of function spaces follow immediately. Rather than relying on the theory of pseudodifferential operators, however, we follow the proof from [70]. ˇ 10.5-10.6 Two references for Cech cohomology theory are Bott and Tu [21] and Fulton [40]. The author’s original inspiration for this section comes from lecture notes by Ben Andrews at the Australian National University. The snake and five lemmas are usually used in an algebraic context. However, Pryde [76] shows that such techniques also are useful in the context of Fredholm operators on Banach spaces in analysis. Two references for de Rham cohomology are [21] and Madsen and Tornehave [65].

Chapter 11

Multivector and Spinor Bundles Prerequisites: The reader should be familiar with the basic ideas of differential geometry. Section 11.1 gives a short survey of the required material and fixes notation. Section 11.2 builds on Chapter 7, and Section 11.6 builds on Chapter 5. Section 11.4 uses Section 4.5. The material from Section 10.5 is used in Section 11.6 and some in Section 11.2. Road map: In many situations the appropriate model of space is not that of an affine space, but rather that of a manifold as defined in Section 6.1. A manifold behaves only locally like affine space, whereas the global behavior can be quite different. So far we have considered only manifolds embedded as k-dimensional surfaces in some surrounding affine space. A general abstract C ∞ manifold M can always be embedded in some finite-dimensional affine space X. However, when the geometry of the manifold is prescribed by a Riemannian metric, most manifolds do not appear as k-surfaces embedded in some affine space. In the remaining chapters we consider general compact Riemannian manifolds M , from an intrinsic point of view. We also change focus: instead of being mainly concerned with local nonsmooth analysis as in the previous two chapters, we now focus on global smooth analysis in the remaining two chapters. Extending the concepts of multivectors and spinors to such manifolds presents new problems, although their relation is fairly straightforward after our having developed the affine theory thoroughly in the previous chapters. The first step is to construct the space Tp M of tangent vectors at a point p ∈ M . These form a family of vector spaces {Tp M }p∈M indexed by M , that is, a vector bundle. In contrast to the case in which M is © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_11

383

384

Chapter 11. Multivector and Spinor Bundles

embedded as a k-surface in an affine space (X, V ), in which case all tangent spaces Tp M are subspaces of V , in general two tangent spaces Tp M and Tq M , p 6= q, are incomparable. As a consequence, some care is needed to define directional derivatives, referred to as covariant derivatives in this context, of tangential vector fields along M . It is straightforward to construct a bundle of multivectors {∧(Tp M )}p∈M and to extend the multivector calculus from Chapter 7, which we do in Section 11.2. Particularly useful are the pullback and pushforward operations from Section 7.2, which allow us to pass between different charts on the manifold and give an invariant meaning to multivectors on the manifold. For spinors we also have induced maps from Proposition 5.3.5. However, these are defined only up to a sign, which makes the definition and even the existence of spinor bundles {4(T / p M )}p∈M a ˇ Z2 cohomology in Section 11.6 to investigate when delicate matter. We use Cech there are topological obstructions for spinor bundles to exist, and if they exist, how many different such spinor bundles there are globally over M . The most central concept for Riemannian manifolds is curvature, which in some sense measures how much the manifold locally differs from flat Euclidean space. The curvature operator in vector notation can be a confusing beast. Having access to multivectors, though, we show in Section 11.3 that this is a symmetric operator on bivectors ∧2 (Tp M ) → ∧2 (Tp M ) : b 7→ R(b). The input bivector b represents the oriented measure of an infinitesimal 2-surface Σ at p, and the output bivector R(b) represents, as in Proposition 4.2.3, an infinitesimal rotation determined by the total variation that a vector undergoes as we move it around ∂Σ, keeping it constant as much as the curved manifold allows. / acting on mulIn Sections 11.5 and 11.6, we define Dirac operators D and D tivector and spinor fields respectively. In flat Euclidean space we saw in Section 9.1 that both these Dirac operators are first-order partial differential operators that are square roots of the Laplace operator ∆. This is no longer true on a curved man2 / differ from ∆ by a zeroth-order term determined ifold, but the squares D2 and D by the curvature R of the manifold. We prove such Weitzenb¨ock identities D2 = ∆ + R in preparation for Chapter 12, although there they have many other applications to geometry. We have seen an analogue for bounded domains in Euclidean space in Theorem 10.3.6, where the curvature of the boundary appears rather than the curvature of space itself. Section 11.4 contains a proof of Liouville’s theorem on conformal maps in dimension n ≥ 3. This completes Theorem 4.5.12, which is otherwise not used elsewhere.

11.1. Tangent Vectors and Derivatives

385

Highlights: • Curvature as infinitesimal rotations around infinitesimal bivectors:11.3.2 • Liouville’s theorem on conformal maps: 11.4.2 • 4M and 4M / Weitzenb¨ock identities: 11.5.9 and 11.6.10 ˇ • Cech cohomology of spinor bundles: 11.6.3 and 11.6.5

11.1

Tangent Vectors and Derivatives

We define general vector bundles E following standard terminology, although from the discussion in Section 1.1 we keep in mind that we in general are considering bundles of linear spaces. Let M be a manifold, and let L be an N -dimensional linear space, over R or possibly C. When considering nonembedded manifolds and vector bundles over them, we shall assume C ∞ regularity. space L, is a collection of linear spaces • A vector bundle E over M , with fiber S {Ep }p∈M together with a cover M = α∈I Uα of M by open sets Uα ⊂ M , and linear invertible maps µα (p) ∈ L(L; Ep ), p ∈ Uα , α ∈ I, such that each bundle transition map µβα (p) := (µβ (p))−1 µα (p) ∈ L(L) is a C ∞ function of p ∈ Uβα := Uβ ∩ Uα for all α, β ∈ I. The linear space Ep is referred to as the fiber of E over p ∈ M . We refer to µα as bundle charts and to {µα }α∈I as a bundle atlas. More generally, a bundle chart is a family µ(p) ∈ L(L; Ep ) of linear invertible maps defined in open sets U such that the bundle transition maps (µ(p))−1 µα (p) ∈ L(L) are C ∞ functions of p ∈ U ∩ Uα for all α ∈ I. • A section w of E is a map that, to each p ∈ M associates a vector w(p) ∈ Ep . We let C ∞ (M ; E) be the space of all sections w such that µ(p)−1 (w(p)) ∈ L is a C ∞ -regular function of p ∈ U for every bundle chart µ. Spaces of sections like C k (M ; E) and L2 (M ; E) are defined similarly. • A (local) frame for E in an open set U ⊂ M is a set of sections ei ∈ C ∞ (U ; E), i = 1, . . . , N , such that {ei (p)} is a basis for Ep for each p ∈ U . PN Thus, if w is a section of E in U , then w(p) = i=1 wi (p)ei (p), where wi are the scalar coordinate functions for w in the frame {ei }. Note that in general there may not exist a frame defined on all M .

Chapter 11. Multivector and Spinor Bundles

386

• A metric g on E is a family of inner products g(·, ·) = h·, ·ip , one on each fiber Ep , p ∈ M . It is assumed that the symmetric nondegenerate bilinear forms L × L → R : (u, v) 7→ hµ(p)u, µ(p)vip are C ∞ functions of p ∈ U for all bundle charts µ. Let {ei }N i=1 be a frame in U ⊂ M . The metric coordinates in this frame are the functions gij (p) := hei (p), ej (p)ip ,

p ∈ U, i, j = 1, . . . , N.

A vector bundle equipped with a metric is called a metric vector bundle. We use only Euclidean metrics on real vector bundles and only Hermitian metrics on complex vector bundles. The metric coordinates for the constant standard inner product on Rn are the Kronecker delta ( 1, i = j, δij := 0, i 6= j. Example 11.1.1 (E ∗ = E identification). If E is a vector bundle over M , with bundle charts µ(p) : L → Ep , then the dual vector bundle E ∗ is defined to be the bundle {Ep∗ }p∈M of dual spaces, with fiber space L∗ and bundle charts (µ(p)−1 )∗ : L∗ → Ep∗ . When a vector bundle E is equipped with a metric, the dual bundle E ∗ and E can be identified by applying Proposition 1.2.3 to each fiber. However, some care has to be taken with regard to charts. We now have twins of bundle charts µ(p) : L → Ep ,

p ∈ U,

and (µ(p)−1 )∗ : L∗ → Ep∗ , Ep∗

p ∈ U,

∗

for = Ep . That L 6= L is a technicality; what is important is that these spaces do not depend on p. Fixing an auxiliary inner product on L, we have L∗ = L as the fiber space for E ∗ = E. This gives a transition map µ(p)∗ µ(p) : L → L∗ between the twin bundle charts. The most fundamental vector bundle is the tangent vector bundle. To define the tangent space Tp M to a nonembedded manifold M at a point p ∈ M , we proceed as follows. We use a C 1 curve γ : (−, ) → M with γ(0) = p and > 0 to define a tangent vector v. If µα is any chart for M around p, then t 7→ µ−1 α (γ(t)) yields a curve through x in Rn , where µα (x) = p. We define two curves γ1 and γ2 0 −1 0 n through p to be equivalent, γ1 ∼ γ2 , if (µ−1 α ◦ γ1 ) (0) = (µα ◦ γ2 ) (0) in R . From the chain rule, we see that this equivalence relation is independent of the choice of chart µα . A tangent vector v at p ∈ M is defined to be an equivalence class [γ] of curves γ through p on M , under the equivalence relation ∼.

11.1. Tangent Vectors and Derivatives

387

Definition 11.1.2 (Tangent bundle). Let (M, {µα }α∈I ) be an n-dimensional manifold, and let p ∈ M . We define the tangent space Tp M to be the set of equivalence classes v = [γ] of C 1 curves γ through p on M as above. Defining the bijective map −1 (11.1) Tp M → Rn : v = [γ] 7→ (µα ◦ γ)0 (0) to be linear gives Tp M the structure of an n-dimensional linear space. The tangent bundle T M is the vector bundle {Tp M }p∈M , with fiber space Rn . Each manifold chart µα : Dα → Mα induces a bundle chart µα (p) ∈ L(Rn ; Tp M ), p ∈ Mα , being the inverse of the map (11.1). We say that M is a Riemannian manifold if T M is equipped with a metric. We refer to sections of T M as tangent vector fields. From the standard basis {ei } for Rn , we define the coordinate frame {µα (ei )} in Mα . Note that for T M , the bundle transition functions µβα (p) : Rn → Rn equal the total derivatives µβα : Rn → Rn , p = µα (x), of the manifold transition x functions µβα . Definition 11.1.3 (Directional derivatives). Let f ∈ C 1 (M ; R) be a scalar-valued function on a manifold M , and let v ∈ Tp M be a tangent vector at p ∈ M . We define the tangential directional derivative (∂v f )(p) := (f ◦ γ)0 (0) if v = [γ]. In a chart µα around p = µα (x), let v = µα (p)u, u ∈ Rn . Then the chain rule shows that ∂v f (p) = ∂u fα (x), so ∂v f (p) is independent of the choice of curve γ representing v. Next consider a vector bundle E over M and a section w ∈ C ∞ (M ; E). We ask whether there is a well-defined tangential directional derivative ∂v w(p). Proceeding similarly to the above, we consider wα (p) := µα (p)−1 w(p) ∈ L and wβ (p) := µβ (p)−1 w(p) ∈ L. In this case, wβ (p) = µβα (p)(wα (p)). However, because of the p-dependence of the linear maps µβα (p), the derivatives ∂v wα (p) and ∂v wβ (p) will in general differ by a zeroth-order term according to the product rule. More precisely, (∂v wβ )(p) = µβα (p)(∂v wα )(p) + (∂v µβα (p))wα (p), where the directional derivative is taken componentwise. Directional derivatives of sections of vector bundles in differential geometry are traditionally called covariant derivatives, denoted by ∇v rather than ∂v , and amount to a choice of zeroth-order terms, consistent with the transition maps as above.

388

Chapter 11. Multivector and Spinor Bundles

Definition 11.1.4 (Covariant derivative). Let E be a vector bundle over a manifold M . A covariant derivative on E is a map C ∞ (U ; T M ) × C ∞ (U ; E) 7→ C ∞ (U ; E) : (v, w) 7→ ∇v w such that • ∇v1 +v2 w = ∇v1 w + ∇v2 w, for all v1 , v2 ∈ C ∞ (U ; T M ) and w ∈ C ∞ (U ; E), • ∇f v w = f (∇v w), for all f ∈ C ∞ (U ; R), v ∈ C ∞ (U ; T M ) and w ∈ C ∞ (U ; E), • ∇v (w1 + w2 ) = ∇v w1 + ∇v w2 , for all v ∈ C ∞ (U ; T M ) and w1 , w2 ∈ C ∞ (U ; E), • ∇v (f w) = f (∇v w) + (∂v f )w, for all f ∈ C ∞ (U ; R), v ∈ C ∞ (U ; T M ), and w ∈ C ∞ (U ; E). N Let E be a vector bundle over a Riemannian manifold M , and let {ei (p)}i=1 be a frame P for E in an open subset U ⊂ M . By the Leibniz rule, a section w(p) = i fi (p)ei (p) of E in U has covariant derivative

∇v w =

X X (∂v fi )ei + fi (∇v ei ). i

i

Since v 7→ ∇v w is linear, this uniquely defines vector fields wji ∈ C ∞ (U ; T M ) in U such that ∇v ei =

N X hωji , viej ,

v ∈ C ∞ (U ; T M ), i = 1, . . . , N.

j=1

This shows in particular that ∇v w is well defined at p ∈ M whenever v ∈ Tp M is given and w is a section differentiable at p. Definition 11.1.5 (Christoffel symbols). Let E be an N -dimensional vector bundle N over a manifold M , with a covariant derivative ∇v . Let e = {ei (p)}i=1 be a frame 2 for E in an open subset U ⊂ M . Then the N vector fields wij ∈ C ∞ (U ; T M ) specifying ∇v in the frame e are called the Christoffel symbols of ∇ in e. We write ωe ∈ C ∞ (U ; L(T M, L(E))) for the section such that the matrix for the map ωe (v) ∈ L(E) in the basis {ei } is {hωij , vi}ij . With this notation, we have X X X (∂v fi )ei + ωe (v) ∇v fi ei = fi ei i

i

i

in the frame e = {ei (p)}N i=1 . To single out a canonical covariant derivative, we demand that it be compatible with natural structures of the vector bundle. In

11.1. Tangent Vectors and Derivatives

389

particular, if E is equipped with a metric {h·, ·ip }p∈M , then we demand that the product rule ∂v hw1 , w2 i = h∇v w1 , w2 i + hw1 , ∇v w2 i hold for all w1 , w2 ∈ C ∞ (U ; E) and v ∈ C ∞ (U ; T M ) and say that ∇v is a metric covariant derivative. If ωij are the Christoffel symbols in an ON-basis, this happens if and only if ωji = −ωij holds for all i, j = 1, . . . , N . To obtain the unique existence of a covariant derivative on the tangent bundle T M , one imposes, besides it being metric, a condition in terms of Lie brackets of vector fields. Exercise 11.1.6 (Lie brackets on manifolds). Let µα : Dα → M and µα : Dβ → M be two charts on a manifold M . Push forward two vector fields u and v in Dα ⊂ Rn , by the transition map ρ := µβα , to u0 := ρ∗ u and v 0 := ρ∗ v in Dβ . Consider the Lie brackets [u, v] := ∂u v − ∂v u and [u0 , v 0 ] := ∂u0 v 0 − ∂v0 u0 . Show that [u0 , v 0 ] = ρ∗ [u, v], and deduce that the Lie bracket of tangent vector fields on M is well defined. Proposition 11.1.7 (Levi-Civita covariant derivative). Let M be a Riemannian manifold. Then there exists a unique covariant derivative on T M , the Levi-Civita covariant derivative, which is both metric and torsion-free, in the sense that ∂v hv1 , v2 i = h∇v v1 , v2 i + hv1 , ∇v v2 i, ∇v1 v2 − ∇v2 v1 = [v1 , v2 ], for all v, v1 , v2 ∈ C ∞ (M ; T M ). Proof. We use an ON-frame {ei (p)}N i=1 in some open set U ⊂ M . If a covariant derivative is metric and torsion-free, then its Christoffel symbols satisfy ωij = −ωji and hωkj , ei i − hωki , ej i = h[ei , ej ], ek i, for all i, j, k. Permuting (i, j, k) to (j, k, i), (k, i, j) and subtracting and adding these two obtained equations gives the formula 2hωkj , ei i = h[ei , ej ], ek i + h[ek , ei ], ej i − h[ej , ek ], ei i. Conversely, the covariant derivative defined through this formula is seen to be metric and torsion-free.

Chapter 11. Multivector and Spinor Bundles

390

Exercise 11.1.8. Generalize the above argument, and show that in a general frame {ei } for T M over a Riemannian manifold, the Levi-Civita covariant derivative is given by 2h∇ei ej , ek i = ∂ei hej , ek i + ∂ej hek , ei i − ∂ek hei , ej i + h[ei , ej ], ek i + h[ek , ei ], ej i − h[ej , ek ], ei i. Note that in a coordinate frame the last three terms vanish, whereas in an ONframe the first three terms vanish. The Christoffel symbols for a metric covariant derivative in an ON-frame give a skew-symmetric map on the fibers Ep . This is true in particular for the Levi-Civita covariant derivative on T M , and in this case we will identify these skew-symmetric maps and the corresponding bivector using Proposition 4.2.3. Definition 11.1.9 (T M Christoffel bivectors). Let M be a Riemannian manifold and let e = {ei (p)} be an ON-frame for the tangent bundle T M in an open set U ⊂ M . Denote the Christoffel symbols for the Levi-Civita covariant derivative by Γe ∈ C ∞ (U ; L(T M ; ∧2 M )), so that ∇v ei = Γe (v) x ei . For the definition of ∧2 M , see Section 11.2.

11.2

Multivector Calculus on Manifolds

In this section we show how the basic affine multivector calculus from Chapter 7 generalizes to manifolds. To simplify the presentation, we consider only compact Riemannian manifolds. Definition 11.2.1 (Multivector bundle). Let M be an n-dimensional Riemannian manifold with atlas {µα }α∈I . The multivector bundle ∧M over M is the vector bundle {∧(Tp M )}p∈M , with fiber space L = ∧Rn and a bundle atlas of bundle charts µα : ∧Rn → ∧(Tp M ) comprising the linear maps induced by the bundle charts of T M as in Definition 2.3.1. We refer to sections of ∧M as (tangent) multivector fields on M . Given a frame {ei }ni=1 for T M , we obtain an induced frame {es }s⊂n for ∧M . The bundle ∧M is a metric bundle equipped with the metric induced from T M , as in Definition 2.5.2, on each fiber. Using the standard inner product on the fiber space ∧Rn , in Mα = µα (Dα ) the manifold chart µα gives rise to the following three bundle charts for ∧M . • The pushforward chart µα (p) : ∧Rn → ∧Tp M . • The pullback chart (µ∗α )−1 (p) : ∧Rn → ∧Tp M .

11.2. Multivector Calculus on Manifolds

391

• The normalized pushforward chart |Jµα |−1 µα (p) : ∧Rn → ∧Tp M . Denote by ∧k M the subbundle of tangential k-vectors, k = 0, 1 . . . , n, so that ∧M = ∧0 M ⊕ ∧1 M ⊕ ∧2 M ⊕ · · · ⊕ ∧n M, where ∧1 M = T M and ∧0 M = M × R. Note that we shall not use the dual multicovector bundle (∧M )∗ , since we identify it with ∧M as in Example 11.1.1. Exercise 11.2.2 (Bundle transition maps). In a fixed manifold chart Mα = µα (Dα ), show that the transition map for ∧M from the pushforward chart to the pullback chart is G(p) : ∧Rn → ∧Rn , the ∧-extension of the linear map of Rn corresponding to the metric gij (x) = hµα (ei ), µα (ej )ip , p = µα (x). Also show that the transition map from the normalized pushforward chart to the √ pullback chart is G/ det g. If µβ : Dβ → Mβ is a second manifold chart and the manifold transition map from Dα to Dβ is µβα , show that the bundle transition maps between the pushforward, pullback, and normalized pushforward charts are µβα∗ , µ∗βα , and µ ˜βα∗ respectively. Note for ∧M that the pushforward charts µα are not only linear, but also ∧homomorphisms, as are the transition maps µβα . The same is true for the pullback charts, but not for the normalized pushforward charts. We have more precisely constructed not only a bundle of linear spaces but a bundle of associative algebras equipped with the exterior product. Applying the affine theory on each fiber, we also have interior products and Hodge star operations defined on ∧M . The latter require a choice of orientation of M , that is, a choice of frame en for the line bundle, that is, a vector bundle with one-dimensional fibers, ∧n M with |en | = 1. The local existence of such en in a chart Mα presents no problem. The global existence is discussed below. A C 1 map ρ : M → M 0 between two different manifolds M and M 0 induces, by working with appropriate types of charts, maps between the respective multivector bundles ∧M and ∧M 0 . Let µα : Dα → M be a chart around p ∈ M , let µ0β : Dβ0 → M 0 be a chart around q = ρ(p) ∈ M 0 , and define ρβα (x) := (µ0β )−1 (ρ(µα (x))). By Exercise 11.2.2, the following constructions are independent of choices of charts. • We define the pullback of a multivector field F 0 on M 0 to be the multivector ∗ field ρ∗ F 0 on M such that µ∗α (ρ∗ F 0 (p)) = ρ∗βα (µ0β F 0 (q)). • If ρ is a diffeomorphism, then we define the pushforward of a multivector field F on M to be the multivector field ρ∗ F on M 0 such that (µ0β∗ )−1 (ρ∗ F (q)) = ρβα∗ (µ−1 ˜∗ F on M 0 is defined similarly α∗ F (p)). The normalized pushforward ρ 0 −1 by demanding that (˜ µβ∗ ) (˜ ρ∗ F (q)) = ρ˜βα∗ (˜ µ−1 α∗ F (p)).

Chapter 11. Multivector and Spinor Bundles

392

Exercise 11.2.3 (Tangential restriction). Let M be a k-surface in a Riemannian manifold N with embedding ρ : M → N , and let p ∈ M ⊂ N . This means in particular that we can regard the tangent space Tp M as a subspace of Tp N . Generalize Exercise 7.2.2 and show that ρ∗ F (p), where F is a multivector field on N , equals the part of F (p) ∈ ∧Tp N tangential to Tp M . We next consider how directional, exterior, and interior derivatives generalize to manifolds. Proposition 11.2.4 (∧M covariant derivative). Let M be a Riemannian manifold. Then there exists a unique covariant derivative ∇v on ∧M that • equals the Levi-Civita covariant derivative from Proposition 11.1.7 on ∧1 M = TM, • equals the tangential directional derivative on ∧0 M = M × R, and • satisfies the product rule ∇v (w1 ∧ w2 ) = (∇v w1 ) ∧ w2 + w1 ∧ (∇v w2 ) for all C 1 multivector fields w1 and w2 and vectors v. Note that this induced covariant derivative, which we refer to as the LeviCivita covariant derivative on ∧M , preserves the subbundles ∧k M , k = 0, 1, . . . , n. Proof. Consider an induced ON-frame {es }s⊂n for ∧M . It is clear that a covariant derivative with the stated properties is unique, since ∇v (es1

∧

· · · ∧ esk ) =

k X

es1

∧

· · · ∧ ∇v esi

∧

· · · ∧ esk

(11.2)

i=1

and ∇v 1 = 0. To show existence, we define a covariant derivative by these identities. It suffices to verify the product rule for w1 = es1 ∧ · · · ∧ esk and w2 = et1 ∧ · · · ∧ etl . We note that (11.2) continues to hold for arbitrary indices 1 ≤ si ≤ n, not necessarily distinct or in increasing order, as a consequence of ∇v esi

∧

esi + esi

∧

∇v esi = 0.

This shows that ∇v (es ∧ et ) = (∇v es ) ∧ et + es ∧ (∇et ).

By its definition above, the Levi-Civita covariant derivative on ∧M satisfies the product rule with respect to exterior multiplication. The following shows that it also satisfies the natural product rules with respect to other products. This is also true for the Clifford product, but we postpone this discussion to Section 11.5. Proposition 11.2.5 (Covariant product rules). Let M be a Riemannian manifold, and let ∇v denote the Levi-Civita covariant derivative on ∧M . This is a metric covariant derivative, that is, ∂v hF, Gi = h∇v F, Gi + hF, ∇v Gi. Moreover, we have ∇v (F y G) = (∇v F ) y G + F y (∇v G), for all v ∈ C(M ; T M ) and F, G ∈ C 1 (M ; ∧M ). In particular, ∇v (F ∗) = (∇v F )∗. The analogous product rules for the right interior product also hold.

11.2. Multivector Calculus on Manifolds

393

Proof. It is clear from the definition that ∇v is metric on ∧0 M and ∧1 M . By bilinearity, it suffices to consider simple k-vector fields F = F1 ∧ · · · ∧ Fk and G = G1 ∧ · · · ∧ Gk . In this case, we have hF1 , G1 i · · · hF1 , Gk i X Y .. .. ... ∂ hFi , Gj i = ∂v hF, Gi = ∂v (±) v . . hFk , G1 i · · · hFk , Gk i X Y hFi , Gj i = · · · = h∇v F, Gi + hF, ∇v Gi, (±)∂v = Q since each vector field Fi and Gj appears exactly once in each product hFi , Gj i. The covariant derivative being metric, the identity for the left interior product follows by dualilty from the product rule for the exterior derivative, since h∇v (F y G), Hi = ∂v hF y G, Hi − hF y G, ∇v Hi = ∂v hG, F ∧ Hi − hG, F ∧ ∇v Hi = h∇v G, F ∧ Hi + hG, ∇v (F ∧ H) − F ∧ ∇v Hi = hF y ∇v G, Hi + h∇v F y G, Hi for all H ∈ C 1 (M ; ∧M ). From this, the identity for ∇v (F ∗) will follow if we prove that ∇v en = 0 for en ∈ C ∞ (M ; ∧n M ) such that |en | = 1. This follows from the fact that ∧n M has one-dimensional fibers and 0 = ∂v |en |2 = 2h∇v en , en i. The proof of the identities for the right interior product and Hodge star are similar. Using the pullback and normalized pushforward charts respectively, Exercise 11.2.2 and the commutation theorem (Theorem 7.2.9) show that the following gives well-defined exterior and interior derivatives of multivector fields on a manifold. Definition 11.2.6 (d and δ on manifolds). Consider the exterior bundle ∧M over a Riemannian manifold M . We define the exterior and interior derivatives dF and δF of a multivector field F ∈ C 1 (M ; ∧M ) as follows. In a chart µα : Dα → Mα , we define dF (p) := (µ∗α )−1 d(µ∗α F (x)), δF (p) := |Jµα |−1 µα δ (|Jµα |(µ−1 )F (x)), α at p = µα (x), x ∈ Dα , where d and δ on the right-hand sides denote the Rn derivatives from Definition 7.1.5. Exercise 11.2.7 (d and δ duality). Let M be a Riemannian manifold, with measure dp. Generalize Proposition 7.1.7 to Riemannian manifolds and prove the following duality relations.

Chapter 11. Multivector and Spinor Bundles

394

For all F ∈ C 1 (M ; ∧k M ), we have δ(F ∗) = (−1)k (dF )∗ at p ∈ M , where the Hodge star uses any fixed orientation of the manifold locally around p. For all F ∈ C 1 (M ; ∧M ) and G ∈ C 1 (M ; ∧M ) that vanish on ∂M , we have Z Z hF, δGidp. hdF, Gidp = − M

M

Hint: Use a partition of unity to localize to charts. The following main result in this section shows that the exterior and interior derivatives on a Riemannian manifold are nabla operators in the natural covariant sense. Proposition 11.2.8 (Covariant nabla). Let M be a Riemannian manifold and let ∇v be the Levi-Civita covariant derivative on ∧M . If {ei } is a frame for T M , with dual frame {ei∗ }, then X dF = e∗i ∧ ∇ei F, F ∈ C 1 (M ; ∧M ), i

δF =

X

ei∗ y ∇ei F,

F ∈ C 1 (M ; ∧M ).

i

Proof. The proof for the exterior derivative uses the expansion rule for d from Exercise 7.4.9. Applying pullbacks, pushforwards, and Exercise 11.1.6, this is seen to hold on M and shows that hdF, v0 ∧ · · · ∧ vk i =

k X (−1)i ∂vi hF, v0 ∧ · · · vˇi · · · ∧ vk i i=0

+

X

(−1)i+j hF, [vi , vj ] ∧ v0 ∧ · · · vˇi · · · vˇj · · · ∧ vk i,

0≤i<j≤k

for F ∈ C 1 (M ; ∧k M ) and vj ∈ C 1 (M ; ∧1 M ). Using that the Levi-Civita covariant derivative is metric, the first sum on the right-hand side is seen to equal k X (−1)i h∇vi F, v0 ∧ · · · vˇi · · · ∧ vk i i=0

−

X

(−1)i+j hF, (∇vi vj − ∇vj vi ) ∧ v0 ∧ · · · vˇi · · · vˇj · · · ∧ vk i

0≤i<j≤k

Pk Since ∇v is also torsion-free, it follows that hdF, v0 ∧ · · · ∧ vk i = i=0 h∇vi F, vi y (v0 ∧ · · · ∧ vk )i. This proves the stated formula, since both sides equal the gradient in the case k = 0 of scalar functions.

11.2. Multivector Calculus on Manifolds

395

To transfer this result to δ, we use Hodge duality. Write F = G∗, where we may assume that G is a k-vector section. Then X δF = (−1)k (dG)∗ = (−1)k e∗i ∧ ∇ei G ∗ i

=

X

∇ei G

∗ ∧ ei

∗=

i

X

e∗i y (∇ei G)∗ =

X

i

e∗i y ∇ei F,

i

where we have used Exercise 11.2.7 in the first step and Proposition 11.2.5 in the last step. In the case of an affine space (X, V ), the two one-dimensional linear spaces ∧0 V = R and ∧n V are isomorphic. For manifolds, the vector bundles ∧0 M and ∧n M need not be globally isomorphic: this happens precisely when M is orientable, that is, when there exists a globally defined section en ∈ C ∞ (M ; ∧n M ) with |en | = ˇ cohomology. 1. We can characterize orientability of the manifold M using Cech Example S 11.2.9 (The sheaf Z2 ). Let M be an n-dimensional manifold, with a cover M = α∈I Uα by open sets. For simplicity, we shall always assume that this is a good cover, although we will not always need all of this hypothesis. Analogous to Definition 10.6.4, this means that each nonempty intersection Us := Uα1 ∩ Uα2 ∩ · · · ∩ Uαk , s = {α1 , . . . , αk } is diffeomorphic to a ball in Rn . Consider the sheaf Z2 , which to each Us associates ˇ the additive group Z2 = {0, 1}. The Cech algebra from Section 10.5 generalizes to Z2 , since it really only requires the spaces to be abelian groups and not vector spaces. In this simplest group Z2 we have −1 = 1, which means that we do not need an alternating algebra. Therefore we write f (s) := hf, es i = −hf, es i ˇ for the value of the Cech cochain at the intersection Us . Also the coboundary operator simplifies to X ∂k f (s) = f (s \ {α}), |s| = k + 2, f ∈ C k (U ; Z2 ). α∈s

Proposition 11.2.10 (First Stiefel–Whitney class). Let M be a manifold with a bundle atlas of bundle charts µα (p) ∈ L(Rn ; Tp M ), p ∈ Uα , α ∈ I, for T M . Consider the sheaf Z2 on the good cover {Uα } from Example 11.2.9. Let f ∈ C 1 (U ; Z2 ) be the 1-cochain, for which the value on the intersection Uα ∩ Uβ is f ({α, β}) :=

( 0,

if det µβα > 0,

1,

if det µβα < 0.

Chapter 11. Multivector and Spinor Bundles

396

ˇ Then ∂1 f = 0, and the Cech cohomology class [f ] does not depend on the choice of local orientation in each Uα specified by the bundle charts µα . The manifold M is orientable if and only if there exists g ∈ C 0 (U ; Z2 ) such that ∂0 g = f , that is, if [f ] = [0]. ˇ The Cech cohomology class [f ] ∈ H 1 (U ; Z2 ) is called the first Stiefel–Whitney class w1 (M ) of M . The two orientations of M correspond to H 0 (U ; Z2 ) = Z2 . Proof. That ∂1 f ({α, β, γ}) = f ({α, β}) + f ({α, γ}) + f ({β, γ}) = 0

mod 2

follows from the transitivity µγβ µβα = µγα of the transition maps. Another choice of local orientation in each Uα corresponds to g ∈ C 0 (U ; Z2 ), where g({α}) = 1 if we change the orientation in Uα , and g({α}) = 0 otherwise. If f 0 ∈ C 1 (U ; Z2 ) is the 1-cochain for this new choice, then f 0 ({α, β}) = f ({α, β}) + g({α}) + g({β})

mod 2,

that is, f 0 = f + ∂0 g. Since M is orientable, we can choose f 0 = 0; this completes the proof. The integral calculus from Section 7.3 carries over with minor changes from affine spaces to general manifolds M . In the absence of an affine space X in which M is embedded, the main change is that we no longer can consider integrals like R oriented measure ∧k M or integrals M F (p)dp of tangent vector fields F . Such notions are not defined for general manifolds, for the simple reason that if E is a general vector bundle over a manifold M , then two fibers Ep and Eq , p 6= q, are incomparable. Two consequences are the following. • The vector sum vp + vq ,

vp ∈ Ep , vq ∈ Eq ,

is not defined, except as a direct sum vp ⊕ vq = (vp , vq ), which is not what we want. The continuous analogue of this, the integral, is therefore not possible either. • It is not well defined what is meant by a constant section of a general vector bundle. When the form we integrate takes values in a fixed linear space L for all p ∈ M , then we can proceed as before. We outline the main steps of the extension of Section 7.3 to an integral calculus of forms on k-surfaces M embedded in an n-dimensional manifold N . • Extending Definition 7.3.1, a general k-form on N is now a map that to each p ∈ N , associates a homogeneous function Θ : ∧k Tp N → L : w 7→ Θ(p, w). If these functions are linear, we refer to Θ as a linear k-form.

11.2. Multivector Calculus on Manifolds

397

• Using charts and a partition of unity, we define the integral Z Θ(p, dp) M

of a k-form Θ in N over an oriented C 1 -regular k-surface M . When R Θ is an even form, for example when we are integrating a scalar function M f (p)dp in a Riemannian manifold, where dp = |dp|, we do not need M to be oriented. • A linear k-form Θ on N can be uniquely written Θ(p, w) = hF1 (p), wiv1 + · · · + hFm (p), wivm , if {vj }m j=1 is a basis for L, where Fj and w are k-vector fields on N . By Proposition 11.2.8, its exterior derivative is Θ(p, ˙ ∇ y w) =

m m X n X X hdFj , wi = h∇ei Fj , e∗i y wivj j=1

j=1 i=1

in a frame {ei } for T M . Note that we cannot regard w as a constant section of ∧k M , and the product rule formula in Definition 7.3.7 needs to be adjusted by a term involving the covariant derivative of e∗i y w if we want to use this on a manifold N . • If the k-surface M in N is oriented and C 1 -regular, and if ∂M has the orientation induced by M , then for every C 1 -regular k-vector field F in a neighborhood of M in N , we have the Stokes formula Z Z hdF (p), dpi = hF (q), dqi. M

∂M

Here dF denotes the exterior derivative. Extending the nabla notation to manifolds, we may write ∇ ∧ F to avoid notational confusion between the exterior derivative dF and the oriented measure dp. For L-valued linear kforms, the Stokes formula reads Z Z Θ(p, ˙ ∇ y dp) = Θ(q, dq). M

∂M

Also the Hodge dual Stokes formula (7.4) extends to manifolds, by inspection of the affine proof, using Proposition 11.2.5. In particular, the following special case is useful. Exercise 11.2.11 (Divergence theorem). Let D be a domain in a Riemannian manifold N , with outward pointing unit normal vector field ν on ∂D. Show that Z Z div F (p) dp = hF (q), ν(q)i dq, D 1

∂D

for a C vector field F on D. Here div F = δF = tion 11.2.8.

P

i hei , ∇ei F i,

using Proposi-

Chapter 11. Multivector and Spinor Bundles

398

11.3 Curvature and Bivectors Consider a vector bundle E over a manifold M , with a given covariant derivative ∇v as in Definition 11.1.4 that allows us to differentiate sections of E. Given a n for E, we have Christoffel symbols ωe defined by frame {ei }i=1 X hωji , viej ∇v ei = j

as in Definition 11.1.5. Although in general we cannot give a definition of what it means for a section to be constant, using the covariant derivative, we can define what is meant by a section being constant along a given curve on M as follows. Definition 11.3.1 (Parallel sections). Let E be a vector bundle over a manifold M , with covariant derivative ∇v , and consider a curve γ, that is, a 1-surface, in M . A section F of E over γ is said to be parallel if ∇v F = 0 at each p ∈ γ and tangent vector v to γ at p. The following calculation shows that in general we cannot even extend this notion of parallel sections to 2-surfaces, which leads us to the fundamental notion of curvature. Example 11.3.2 (Parallel transport around triangles). Let E be a vector bundle over a manifold M , with covariant derivative ∇v . Consider a 2-surface Σ in M , at a given point p ∈ M , defined by a chart µ : D → Σ ⊂ M. Here D is the triangle D := {(x, y) ; x, y ≥ 0, x + y ≤ 1} ⊂ R2 and µ(0, 0) = p. For 0 < < 1 we also consider the subsurfaces Σ = µ(D ), where D := {(x, y) ; x, y ≥ 0, x + y ≤ }. Assume that F is a section of E over the curve ∂Σ that is parallel along the three sides and continuous at µ(, 0) and µ(0, ). We do not assume continuity at p. This defines a map Ep → Ep : lim F (µ(x, 0)) 7→ lim F (µ(0, y)), x→0+

y7→0+

which is referred to as the parallel transport around ∂Σ . We want to calculate this map to orderP 2 . To this end, fix a frame {ei }N i=1 for E in a neighborhood of Σ and write F = i Fi ei . The equation ∇v F = 0 becomes X X ∂v Fi + hωij , viFj ei = 0. i

j

11.3. Curvature and Bivectors

399

Figure 11.1: Parallel transport of a vector around a triangle Σ. Since the sphere has positive curvature, the resulting rotation will be in the direction corresponding to the orientation of Σ. 0 Write ωij := µ∗ (ωij ) for the pulled back vector field on D . Then the equation becomes N scalar differential equations X 0 ∂ u fi + hωij , uifj = 0, i = 1, . . . , N, j

along the triangle curve ∂D , where fi := µ∗ Fi = Fi ◦ µ and u is a tangent vector to ∂D . 0 Let zi := fi (0+ , 0) be given, and write aij (x, y) := hωij , e1 i and bij (x, y) := 0 hωij , e2 i. Also write aij := aij (0, 0), ∂x aij := ∂x aij (0, 0), ∂y aij := ∂y aij (0, 0), and similarly for bij . (i) We first solve the parallel transport equations ∂x fi (x, 0) +

X

aij (x, 0)fj (x, 0) = 0

j

along {(x, 0) ; 0 < x < }. With the ansatzes fi (x, 0) = φ0i + φ1i x + φ2i x2 + O(x3 )

400

Chapter 11. Multivector and Spinor Bundles

0 1 + αij x + O(x2 ), we obtain and aij (x, 0) = αij

fi (, 0) = zi −

X

aij zj +

j

2 X X aik akj − (∂x aij ) zj + O(3 ) 2 j k

under the inital condition φ0i = zi . (ii) We next solve the parallel transport equations X ∂x fi ( − x, x) + (−aij ( − x, x) + bij ( − x, x))fj ( − x, x) = 0 j

along {( − x, x) ; 0 < x < }. We approximate − aij ( − x, x) + bij ( − x, x) = (−aij + bij − ∂x aij + ∂x bij ) + (∂x aij − ∂y aij − ∂x bij + ∂y bij )x + O(2 ). 0 1 Letting φ0i = fi (, 0), αij = −aij + bij − ∂x aij + ∂x bij , and αij = ∂x aij − ∂y aij − ∂x bij + ∂y bij in (i), we get X fi (0, ) = zi − bij zj j 2

XX + (−aik bkj + bik akj + bik bkj ) − (∂y aij − ∂x bij − ∂y bij ) zj + O(3 ). 2 j k

(iii) We finally solve the parallel transport equations X ∂x fi (0, − x) − bij (0, − x)fj (0, − x) = 0 j

along {(0, − x) ; 0 < x < }. We approximate −bij (0, − x) = (−bij − ∂y bij ) + (∂y bij )x + O(2 ). 0 1 Letting φ0i = fi (0, ), αij = −bij − ∂y bij , and αij = ∂y bij in (i), we get

fi (0, 0+ ) = zi +

2 X X (−aik bkj + bik akj ) + (∂y aij − ∂x bij ) zj + O(3 ). 2 j k

To interpret this result, we push forward the oriented measure D to 2 ∧2 (Σ ) := µ(0,0) ( 2 e1 ∧ e2 ) ∈ ∧2 (Tp M ).

2 2 e1 ∧

e2 of

Although there is no notion of oriented measure of k-surfaces in manifolds in general, for ≈ 0 the bivector ∧2 (Σ ) approximately describes an oriented measure

401

11.3. Curvature and Bivectors

0 of the small 2-surface Σ . For the vector fields ωij = aij e1 + bij e2 in R2 , we note that 0 dωij = (∂x bij − ∂y aij )e12 , 0 0 ∧ ωkj = (aik bkj − bik akj )e12 . ωik

Therefore our calculation shows that X X 0 0 2 0 3 fi (0, 0+ ) = zi − ∧ ωkj , 2 e12 izj + O( ). ωik hdωij + j

k

Back on M we conclude that parallel transport around Σ equals X Ep → Ep : F 7→ F − hΩij , ∧2 (Σ )iFj ei + O(3 ), j

where Ω is the curvature operator defined below. The following encodes the curvature of a general manifold, as compared to Definition 10.3.5 which encodes the curvature of an n − 1 surface embedded in n-dimensional Euclidean space. Definition 11.3.3 (Curvature operator). Consider a vector bundle E with a covariant derivative ∇v , over a manifold M . Let {ei }N i=1 be a frame for E, and let ωij denote the Christoffel symbols of ∇v in this frame. Assuming that M is a Riemannian manifold, these are vector fields on M , defined in the domain of the frame. The curvature operator Ω for ∇v is the linear map that, at each p, to each bivector b ∈ ∧2 (Tp M ) associates the linear map Ω(b) ∈ L(Ep ), which in the given frame has matrix {hΩij , bi}N i,j=1 , where the bivector fields are Ωij := dωij +

N X

ωik ∧ ωkj ,

i, j = 1, . . . , N.

k=1

Note that we do not write Ωe as for ωe . As Example 11.3.2 indicates, unlike the Christoffel symbols, the operator Ω(b) is independent of the choice of frame, as the following shows. Proposition 11.3.4. Let E be a vector bundle with a covariant derivative ∇v , over a manifold M . Consider two frames {ei } and {˜ ei }, defined on the same open e set, with associated Christoffel symbols ω, ω ˜ and P curvature operators Ω and Ω ˜i = j ej αji , then respectively. If the relation between frames is e X X ω ˜ ij = αik ωkl αlj + αik (dαkj ), kl

e ij = Ω

X

k ik

α Ωkl αlj ,

kl

for all i, j = 1, . . . , N , where {αij }ij denotes the inverse matrix to A = {αij }ij .

402

Chapter 11. Multivector and Spinor Bundles

These transformation properties of Ωij mean that the curvature operator is a well-defined section Ω ∈ C ∞ (M ; L(∧2 M ; L(E))), globally on all M , that does not depend on the choice of frame. Note that the corresponding statement is not true for the Christoffel symbols, because of term A−1 dA. It is only when the change of frame matrix A is constant that the Christoffel symbols transform as an operator. ˜i = Proof. On the one hand, ∇v e ˜i = ∇v e

X

P

ej . ωji , vi˜ j h˜

∇v (ej αji ) =

j

On the other hand,

X X hωkj , viek αji + ej (∂v αji ), j

jk

and ∂v αji = hdαji , vi. Comparing these expressions, we obtain X X X hωkj , viek αji + h˜ ωji , viαkj ek = hdαki , viek , jk

jk

k

from which the relation for the Christoffel symbols follows. Furthermore, e ij = Ω

X

dαik ∧

k

+

X

X X αik ωkl αlj + dαkj + (dωkl )αlj − ωkl ∧ dαlj

l

X kmk0

α

ik

X

l

k

ωkl αlm + dαkm

∧

α

mk0

X

ωk0 l0 αl0 j + dαk0 j .

l0

l

to the stated relation, and the Here the terms containing dω and ω ∧ ω combine P remaining terms cancel pairwise, using 0 = k d(αik αkj ) = (dαik )αkj +αik (dαkj ). A fundamental fact about second-order partial derivatives in affine space is that ∂i ∂j = ∂j ∂i . The quantity that describes the failure of this for covariant derivatives on vector bundles turns out to be the curvature operator, as the following shows. Proposition 11.3.5 (Curvature as commutator). Let E be a vector bundle with a covariant derivative ∇v , over a manifold M . Then ∇u (∇v F ) − ∇v (∇u F ) = ∇[u,v] F + Ω(u ∧ v)F, for every section F ∈ C 2 (M ; E).

403

11.3. Curvature and Bivectors Proof. Calculating in a frame {ei }, we have X X ∇u (∇v F ) = ∇u hωki , viFi ek (∂v Fi )ei + i

ki

X X = (∂u ∂v Fi )ei + (∂u hωki , vi)Fi + hωki , vi(∂u Fi ) ek + (∂v Fi )hωki , uiek i

ik

X hωji , viFi hωkj , uiek . + ijk

Subtracting the corresponding formula for ∇v (∇u F ), it suffices to show that ∂u ∂v Fi − ∂v ∂u Fi = ∂[u,v] Fi , ∂u hωki , vi − ∂v hωki , ui = hdωki , u ∧ vi + hωki , [u, v]i, hωji , vihωkj , ui − hωji , uihωkj , vi = hωkj ∧ ωji , u ∧ vi. The first and last equations are straightforward to verify, and the second follows from Exercise 7.4.9, generalized to manifolds. This proves the proposition. Our interest is in metric vector bundles and covariant derivatives, in which case it is clear from Definition 11.3.3 and the skew-symmetry of the Christoffel symbols and the exterior product of vectors that Ωji = −Ωij in an ON-frame {ei }. The map Ω(b) ∈ L(E) being skew-symmetric, recalling Proposition 4.2.3 we can equivalently view it as the bivector X hΩij , biei ∧ ej ∈ ∧2 E, Ω(b) = i<j

so that Ω ∈ C ∞ (M ; L(∧2 M ; ∧2 E)). In this setup, the action of Ω(b) on a section F ∈ C ∞ (M ; E) is Ω(b)F = Ω(b) x F, using the right interior product in exterior bundle ∧E. Although possible for any metric vector bundle, we shall use only this bivector representation of the curvature operator for the tangent bundle E = T M . Definition 11.3.6 (Riemann curvature operator). Let M be a Riemannian manifold. By the Riemann curvature operator we mean the curvature operator R ∈ C ∞ (M ; L(∧2 M )) for the Levi-Civita covariant derivative on the tangent bundle T M . Using the relation ∧2 M = SO(T M ) ⊂ L(T M ) provided by Proposition 4.2.3, we have R ∈ C ∞ (M ; L(T M ) ⊗ L(T M )). Using the contractions L(T M ) ⊗ L(T M ) → L(T M ) = T M ⊗ T M → M × R,

Chapter 11. Multivector and Spinor Bundles

404

obtained by lifting the bilinear products A1 A∗2 and hv1 , v2 i of matrices and vectors respectively to the tensor products, we define from R the Ricci curvature Ric ∈ C ∞ (M ; L(T M )) and the scalar curvature S ∈ C ∞ (M ; R). If {ei } is an ON-frame for T M , then we define the Riemann curvature coefficients Rijkl := hei ∧ ej , R(ek ∧ el )i. Then the matrix for the Ricci curvature operator in this frame is Ricij := hei , Ric(ej )i =

X

Rikjk ,

k

and the scalar curvature is S =

P

i

Ricii =

P

ik

Rikik .

Proposition 11.3.7 (Symmetries of R). Let R ∈ C ∞ (M ; L(∧2 M )) be the Riemann curvature operator on a Riemannian manifold M . Then at each p ∈ M , the operator R : (∧2 M )p → (∧2 M )p is symmetric, and the Bianchi identities R(u ∧ v) x w + R(v ∧ w) x u + R(w ∧ u) x v = 0 hold for all vectors u, v, w ∈ Tp M . Proof. To prove the Bianchi identities, by trilinearity it suffices to consider a coordinate frame {ei }. In this case, [ei , ej ] = 0, so by Proposition 11.3.5 the identities follow from the computation ∇ei ∇ej ek − ∇ej ∇ei ek + ∇ej ∇ek ei − ∇ek ∇ej ei + ∇ek ∇ei ej − ∇ei ∇ek ej = ∇ei (∇ej ek − ∇ek ej ) + ∇ej (∇ek ei − ∇ei ek ) + ∇ek (∇ei ej − ∇ej ei ) = 0, since ∇u v − ∇v u = [u, v] for the Levi-civita covariant derivative. Using the Bianchi identities, we compute hu2 ∧ v2 , R(u1 ∧ v1 )i = hu2 , R(u1 ∧ v1 ) x v2 i = −hu2 , R(v1 ∧ v2 ) x u1 i − hu2 , R(v2 ∧ u1 ) x v1 i = hu1 , R(v1 ∧ v2 ) x u2 i + hv1 , R(v2 ∧ u1 ) x u2 i = −hu1 , R(v2 ∧ u2 ) x v1 i − hu1 , R(u2 ∧ v1 ) x v2 i − hv1 , R(u1 ∧ u2 ) x v2 i − hv1 , R(u2 ∧ v2 ) x u1 i = 2hu1 ∧ v1 , R(u2 ∧ v2 )i + hv2 , R(u2 ∧ v1 ) x u1 i + hv2 , R(u1 ∧ u2 ) x v1 i = 2hu1 ∧ v1 , R(u2 ∧ v2 )i − hv2 , R(v1 ∧ u1 ) x u2 i, from which the stated symmetry follows.

405

11.4. Conformal Maps and ON-Frames

11.4 Conformal Maps and ON-Frames The main result in this section, Liouville’s theorem on conformal maps, is difficult to place. It concerns the fractional linear maps from Section 4.5 and completes the proof of Theorem 4.5.12. It is also relevant for the hypercomplex analysis from Chapter 8. We have nevertheless placed it here, since the proof makes use of ON-frames and curvature. We start with the following method of Cartan for calculating the Christoffel symbols in an ON-frame. Proposition 11.4.1 (Cartan). Let {ei } be an ON-frame for T M on a Riemannian manifold M . Then the Christoffel symbols for the Levi-Civita covariant derivative in {ei } are the unique vector fields ωij that satisfy ωji = −ωij and dei +

n X

ωij

∧

ej = 0,

j = 1, . . . , n.

(11.3)

j=1

Note that the uniqueness part of the proposition means that any method of writing the exterior derivatives of the ON-frame vectors as in (11.3) must give the Christoffel symbols. Proof. Using Proposition 11.2.8, we have X XX X X X ωki ∧ ek dei = ej ∧ ∇ej ei = ej ∧ hωki , ej iek = hωki , ej iej ∧ ek = j

j

k

k

j

k

0 is any skew-symmetric for the Christoffel symbols. For uniqueness, assume that ωij ˜ ij := family of vector fields that satisfies (11.3), and consider the differences ω 0 . We have ωij − ωij X ω ˜ ij ∧ ej = 0, i = 1, . . . , n. j

Forming the inner product with frame bivectors gives X X 0= hek ∧ el , ω ˜ ij ∧ ej i = ˜ ij ihel , ej i − hek , ej ihel , ω ˜ ij i hek , ω j

j

˜ ik i. = hek , ω ˜ il i − hel , ω ωji for all i, j. This yields ˜ ij = −˜ So hek , ω ˜ il i = hel , ω ˜ ik i for all i, k, l, as well as ω ˜ ik i = hek , ω ˜ ki i = hej , ω ˜ jk i = −hei , ω hek , ω ˜ ij i = −hek , ω ˜ ji i ˜ ji i = hei , ω ˜ kj i = −hej , ω 0 for all k, i, j. This proves that the ωij coincide with the Christoffel symbols.

A two-dimensional Euclidean space abounds with conformal maps: all analytic functions are conformal where the derivative is nonzero. The goal of this

Chapter 11. Multivector and Spinor Bundles

406

section is to prove Liouville’s theorem on conformal maps, which shows that in n-dimensional Euclidean space, n ≥ 3, the situation is drastically different: the only conformal maps are the fractional linear maps / M¨obius maps from Theorem 4.5.12. These are very few indeed, since they depend only on as many parameters as the Lorentz isometries in dimension n + 2, that is, n+2 real parameters. 2 Theorem 11.4.2 (Liouville). Let D be a connected open subset of a Euclidean space (X, V ) of dimension dim X ≥ 3. Assume that f : D → V is a conformal map of class C 3 . Then there exists a fractional linear map g : V → V such that g|D = f . Here X = V with origin fixed, and extended V ⊂ V with notation as in Section 4.5. Even though the conclusion is global in the sense that we obtain a conformal diffeomorphism of the extended space V, the problem is local. Indeed, since fractional linear maps are real analytic, it suffices to prove that f coincides with a fractional linear map in D, when D is a ball. A concrete proof of this is as follows. Assume that we have proved that f = f1 on a ball B1 , and that f = f2 on a ball B2 , where B1 ∩ B2 6= ∅ and f1 and f2 are fractional linear maps. To prove that f1 = f2 on all V, write f0 (x) := f2−1 (f1 (x)) = (ax + b)(cx + d)−1 as in Theorem 4.5.16. Then x = (ax + b)(cx + d)−1 ,

for all x ∈ B1 ∩ B2 ,

or equivalently xcx + xd − ax − b = 0. If this second-order polynomial vanishes on an open set, then c = b = 0 and xd = ax for all vectors x. This proves that f0 (x) = axd−1 = x for all x ∈ V . Proof of Theorem 11.4.2. Without loss of generality, we assume that D is a ball with small enough radius that f : D → f (D) is a C 3 -diffeomorphism by the inverse function theorem. Since f is conformal, there exists a unique scalar function λ(x) > 0 in D such that λ(x)f x

is an isometry at each x ∈ D,

and by assumption λ is C 2 -regular. Fix an ON-basis {ei } for V and define the ON-frame ei (y) := λ(x)f x (ei ), y = f (x), in the open set f (D). By definition of λ, f∗ λ is an isometry, so f ∗ f∗ = λ−2 I when acting on vector fields in D. Therefore, pulling back {ei (y)} to D gives f ∗ (ei )(x) = 1/λ(x)ei , x ∈ D. Applying the exterior derivative and using Theorem 7.2.9, we get X f ∗ (∇ ∧ ei ) = −λ−2 (∇λ) ∧ ei = −λ−2 λj ej ∧ ei , j

since ei is a constant vector field, where λi := ∂i λ and ∂i is the partial derivative along ei . Pushing forward this equation to f (D), noting that f∗ f ∗ = µ−4 I

407

11.4. Conformal Maps and ON-Frames

when P acting on bivector fields in f (D), where µ := λ ◦ f −1 , gives µ−4 ∇ ∧ ei = −µ−2 j µi (µ−1 ej ) ∧ (µ−1 ei ), where µj := f∗ (λj ) = λj ◦ f −1 . We obtain X X ∇ ∧ ei = − µi ej − µj ei ∧ ej . µj ej ∧ ei = − j

j

Uniqueness of the Christoffel symbols in Proposition 11.4.1 shows that ωij = µi ej − µj ei . This gives X − µi ek ∧ ωkj + µj ek ∧ ωki ∇ ∧ ωij = (∇µi ) ∧ ej − (∇µj ) ∧ ei + k

= (∇µi ) ∧ ej − (∇µj ) ∧ ei +

X

− µi µk ek ∧ ej + µj µk ek ∧ ei ,

k

X

ωik ∧ ωkj

k

X = (µi ek − µk ei ) ∧ (µk ej − µj ek ) k

X = (µi µk ek ∧ ej − µ2k ei ∧ ej + µk µj ei ∧ ek ). k

Since Euclidean space has zero P curvature, it follows from Definition 11.3.3 that (∇µi ) ∧ ej − (∇µj ) ∧ ei = ( k µ2k )ei ∧ eP j . Pulling back this equation to D gives (∇λi ) ∧ (λ−1 ej ) − (∇λj ) ∧ (λ−1 ei ) = ( k λ2k )(λ−1 ei ) ∧ (λ−1 ej ). We obtain the following nonlinear system of second-order partial differential equations λ

n X

n X λ2k ei ∧ ej , λik ek ∧ ej − λjk ek ∧ ei =

k=1

i, j = 1, . . . , n,

(11.4)

k=1

where λij := ∂j ∂i λ. If dim X ≥ 3, then the system (11.4) is overdetermined, which we exploit as follows. Evaluating the em ∧ ej component of the equation (11.4), where i, j, m are distinct, we deduce that λim = 0 whenever i 6= m. On the other hand, evaluating the ei ∧ ej component of the same equation, it follows that λii + λjj = λ−1 |∇λ|2 . Since this equation holds for all i 6= j, we get λii = |∇λ|2 /(2λ), i = 1, . . . , n. Since λim = 0 whenever i 6= m, λi is a function of xi only, and so is λii . This holds for all i, so |∇λ|2 /λ =: c must be a constant independent of all xi . We have shown that λii (x) = |∇λ(x)|2 /(2λ(x)) = c, i = 1, . . . , n, x ∈ D. If c = 0, then ∇λ = 0 and λ(x) =: b is constant. Since ei = (bf∗ )(ei ) = (f −1 )∗ (b−1 ei ), it follows from Theorem 7.2.9 that ∇ ∧ ei = 0, since b−1 ei is constant. Proposition 11.4.1 shows that the Christoffel symbols vanish. Therefore ei must be constant in D since the covariant derivatives for the Euclidean space are the standard partial derivatives. Hence the total derivative f x is a constant matrix, and we conclude that the conformal map is of the form X X X f xi ei = b xi ei + d, xi ei ∈ D, i

i

i

Chapter 11. Multivector and Spinor Bundles

408

for some constant vector d ∈ V . Thus f is a restriction of a composition of an isometry, a dilation, and a translation, and thus a fractional linear map. On the other hand, if c 6= 0, then integration gives ∇λ = c(x − a) for some a ∈ V , since λi depends only on xi . This yields λ(x) =

1 c |c(x − a)|2 = |x − a|2 . 2c 2

This can be reduced to the case that λ(x) is constant, by composing y = f (x) with the fractional linear map x = g(z) = 1/z + a. Since λ > 0 on D, it follows that a ∈ / D, and therefore g maps a domain D0 ⊂ V bijectively onto D. From Exercise 4.5.18 we have that g z (h) = |z|−2 zhz −1 . Thus f ◦ g : D → f (D) has derivative f ◦ g z (h) = f x (|z|−2 zhz −1 ) = |x − a|2 f x (zhz −1 ) =

2 λf (zhz −1 ), c x

which is a constant times an isometry, so the scale factor λ associated to f ◦ g is constant. Thus the calculation above for constant λ applies to f ◦ g and shows that f ◦ g, and hence f , is a fractional linear map. This completes the proof. Exercise 11.4.3. In dimension n = 2, show that (11.4) reduces to the single equation λ∆λ = |∇λ|2 , that is, ∆(ln λ) = 0, saying that ln |f 0 (z)| is harmonic when f (z) is an analytic function.

11.5

Weitzenb¨ock Identities

Proceeding as in Section 3.1, we define on each fiber ∧(Tp M ) of the multivector bundle over a Riemannian manifold M a Clifford product 4. Definition 11.5.1 (Clifford bundle). Let M be a Riemannian manifold. The Clifford bundle 4M over M equals the multivector bundle ∧M as a bundle of linear spaces, but equipped with the Clifford product 4 on the fibers ∧(Tp M ), making it into a bundle of associative algebras over M . Just as in the case of a single affine space, we write 4M instead of ∧M when we use the Clifford product rather than the exterior and interior products, but we are not strict in this, since sometimes we use all these products and it may not be clear whether 4M or ∧M is the appropriate notation. Translating from the multivector bundle, the Clifford bundle is the direct sum of subbundles 4M = 40 M ⊕ 41 M ⊕ 42 M ⊕ · · · ⊕ 4n M. Note that since the charts µα : Rn → T M in general are not isometries, they do not induce isomorphisms between Clifford algebras. However, given an ON-frame {ei }ni=1 for T M , we obtain an induced ON-frame {es }s⊂n for 4M .

409

11.5. Weitzenb¨ock Identities

Exercise 11.5.2 (Clifford product rule). Show for the Levi-Civita covariant derivative on 4M = ∧M that the product rule ∇v (F

4

G) = (∇v F ) 4 G + F

4

(∇v G)

holds for all multivector fields F, G ∈ C 1 (M ; 4M ) on M . A first use of the Clifford product on 4M is to express the Christoffel symbols and curvature of ∧M = 4M in terms of the bivectors Γe and R for T M through Clifford commutators. Proposition 11.5.3 (Clifford bivector commutators). Let M be a Riemannian manifold, and let e = {ei (p)} be an ON-frame for the tangent bundle T M in an open set U ⊂ M . Denote by Γe ∈ C ∞ (U ; L(∧1 M ; ∧2 M )) and R ∈ C ∞ (M ; L(∧2 M ; ∧2 M )) the Christoffel symbols and curvature operator for T M . ˜ = {es (p)} be the induced ON-frame for 4M in U . Denote by ωe˜ ∈ Let e C ∞ (U ; L(∧1 M ; L(4M )) and Ω ∈ C ∞ (M ; L(∧2 M ; L(4M ))) the Christoffel symbols and curvature operator for 4M . Then ωe˜ (v)F = 21 [Γe (v), F ], Ω(b)F = 12 [R(b), F ], for v ∈ C(M ; T M ), b ∈ C(M ; ∧2 M ), and F ∈ C(M ; 4M ), using the Clifford commutator [w1 , w2 ] := w1 4 w2 − w2 4 w1 . Proof. Consider first the Christoffel symbols. For a basis vector field ei ∈ C(M ; T M ), we have ωe˜ (v)ei = ∇v ei = Γe (v) x ei = 12 (Γe (v) 4 ei − ei 4 Γe (v)) = 12 [Γe (v), ei ], using the analogue of (3.3) for the right interior product, and Definitions 11.1.5 and 11.1.9. For scalar functions F ∈ C(M ; 40 M ) both sides of the identity vanish, and for F ∈ C(M ; 4k M ), k ≥ 2, the result follows from the vector case k = 1 and the derivation property D(F1 4 F2 ) = (DF1 ) 4 F2 + F1 4 (DF2 ),

F1 , F2 ∈ C 1 (M ; 4M ),

which holds for D = Γe (v), as a consequence of Exercise 11.5.2, as well as for D : F 7→ 12 [Γe (v), F ], by general properties of commutators. For the curvature operator, we can argue similarly. Since Ω(b)ei = R(b) x ei = 12 (R(b) 4 ei − ei 4 R(b)) = 12 [R(b), ei ], it suffices to show that Ω(b)(F1 4 F2 ) = (Ω(b)F1 ) 4 F2 + F1 4 (Ω(b)F2 )

(11.5)

410

Chapter 11. Multivector and Spinor Bundles

for all F1 , F2 ∈ C ∞ (M ; 4M ). By Exercise 11.5.2, we have ∇u ∇v (F1 4 F2 ) = (∇u ∇v F1 ) 4 F2 + (∇v F1 ) 4 (∇u F2 ) + (∇u F1 ) 4 (∇v F2 ) + F1 4 (∇u ∇v F2 ), as well as ∇[u,v] (F1 4 F2 ) = (∇[u,v] F1 ) 4 F2 + F1 property (11.5) follows from Proposition 11.3.5.

4

(∇[u,v] F2 ), so the derivation

The fundamental first-order differential operator on M acting on sections of 4M is the following natural generalization of the Euclidean nabla operator from Definition 9.1.1. Definition 11.5.4 (The Dirac operator on 4M ). Let M be a Riemannian manifold. The 4-Dirac operator on 4M is the operator X DF := dF + δF = e∗i 4 ∇ei F, F ∈ C 1 (M ; 4M ), i

where {ei } is a frame for T M , with dual frame {e∗i }. Note that the Dirac operator D = d+δ is not locally similar to the Euclidean Dirac operator. This in contrast to d and δ, where by definition µ∗ d(µ∗ )−1 and (˜ µ∗ )−1 δ µ ˜∗ are the Euclidean exterior and interior derivatives, locally in any chart µ for M . As in Euclidean space, d2 = 0 and δ 2 = 0, and as a consequence, D2 = dδ + δd for every Riemannian manifold. We saw in Section 8.1 that D2 equals the Laplace operator ∆, acting componentwise on multivector fields, in Euclidean space. The situation on a manifold is more subtle, where D2 differs from the following Laplace operator, on the bundle E = 4M , by a curvature term. Proposition 11.5.5. Let E be a metric vector bundle with a metric covariant derivative, over a Riemannian manifold M . If {ei } is an ON-frame for T M in an open set U ⊂ M , then Z X Z X h∇ei F, ∇ei Gidp = − h∇ei ∇ei F − ∇∇ei ei F, Gidp (11.6) M

i

M

i

for all F, G ∈ C02 (U ; E). Here ∇ei ei denotes the Levi-Civita covariant derivative on T M . Proof. Using that the covariant derivative on E is metric, we get the pointwise identity ∂ei h∇ei F, Gi = h∇ei ∇ei F, Gi + h∇ei F, ∇ei Gi.

411

11.5. Weitzenb¨ock Identities Define the vector field v :=

X h∇ei F, Giei ∈ C01 (U ; T M ). i

From Proposition 11.2.8 we have X X hej , ∇ej (h∇ei F, Giei )i = hej , (∂ej h∇ei F, Gi)ei + h∇ei F, Gi(∇ej ei )i div v = ij

ij

=

X

∂ei h∇ei F, Gi + h∇ei F, Gi div ei .

i

Using the identity X X X X div(ei )ei = hej , ∇ej ei iei = − ∇ej ej , h∇ej ej , ei iei = − i

ij

ij

j

the stated identityR now follows from the divergence theorem in Exercise 11.2.11, which shows that M vdp = 0. Definition 11.5.6 (Laplace–Beltrami operator). Let E be a metric vector bundle with a metric covariant derivative, over a Riemannian manifold M . The Laplace– Beltrami operator is the second-order differential operator that in an ON-frame {ei } for T M is given by X ∆F := (∇ei ∇ei F − ∇∇ei ei F ). i

The left-hand side in (11.6), for G = F , can be viewed as the H 1 Sobolev (semi-)norm k∇ ⊗ F k2L2 (M ;E) , and −∆ is the L2 operator corresponding to this quadratic form. Exercise 11.5.7 (Second covariant derivative). Show that 2 F := ∇u ∇v F − ∇∇u v F ∇u,v

is bilinear in u and v. Deduce that the Laplace–Beltrami operator is well defined, that is, independent of choice of ON-basis. Exercise 11.5.8 (Scalar Laplace–Beltrami). Show that for the trivial vector bundle E = R, the Laplace–Beltrami operator acting on scalar functions f ∈ C 2 (M ; R) is ∆f = δdf. Verify also that this follows from Proposition 11.5.9 below. With a chart for M , and inverse metric (g ij ) = (gij )−1 and determinant g := det(gij ), generalize Example 7.2.12 and show that X √ √1 ∂i ( gg ij ∂j f ). ∆f = g ij

Chapter 11. Multivector and Spinor Bundles

412

The following main result in this section is the key identity used in the proof of the Chern–Gauss–Bonnet theorem in Section 12.3. Proposition 11.5.9 (Weitzenb¨ock identity for D). Let M be a Riemannian manifold M , and let {ei } be an ON-frame for T M . Then X 1 1X ei ej Ω(ei ∧ ej )F = −∆F + SF + −D2 F = −∆F − Rijkl ei ej F ek el , 4 8 i<j ijkl

∞

for F ∈ C (M ; 4M ), where ∆ and Ω denote the Laplace–Beltrami and curvature operators on 4M , and R denotes the Riemann curvature operator on T M with coefficients Rijkl and scalar curvature S. Proof. With notation as in Exercise 11.5.7 and suppressing 4, Proposition 11.3.5 yields X X D2 F = ei ∇ei (ej ∇ej F ) = ei (∇ei ej )∇ej F + ej ∇ei ∇ej F ij

=

X

=

X

ij

ei ej ∇2ei ,ej F

+ ej ∇∇ei ej F + (∇ei ej )∇ej F

ij

e2i ∇2ei ,ei F +

i

+

X

ei ej (∇2ei ,ej F − ∇2ej ,ei F )

i<j

X

ei ej ∇∇ei ej F + (∇ei ej )∇ej F

ij

= ∆F +

X

ei ej Ω(ei ∧ ej )F +

X

i<j

ei h∇ei ej , ek iej ∇ek F + h∇ei ej , ek iek ∇ej F .

ijk

Since h∇ei ej , ek i + hej , ∇ei ek i = ∂ei hej , ek i = 0, this proves the first identity. To express the identity in terms of the Riemann curvature operator, we use Proposition 11.5.3 to write X 1X ei ej Ω(ei ∧ ej )F = ei ej [R(ei ∧ ej ), F ] 2 i<j i<j 1 XX 1X Rijkl ei ej [ek el , F ] = Rijkl (ei ej ek el F − ei ej F ek el ). 4 i<j 8 kl ijkl P It remains to simplify ijkl Rijkl ei ej ek el . From the Bianchi identities in Proposition 11.3.7 and multivector algebra, we obtain X X X Rijkl ei ej ek = − (Rjkil + Rkijl )ei ej ek = − Rijkl (ek ei ej + ej ek ei ) =

ijk

=−

ijk

X

ijk

Rijkl (ei ej ek + 2ek y (ei ej ) + ei ej ek + 2(ej ek ) x ei )

ijk

= −2

X ijk

Rijkl ei ej ek − 4

X ij

Rijil ej + 2

X ij

Rijjl ei + 2

X ik

Riikl ek .

413

11.6. Spinor Bundles

P P Since Riikl = 0 and Rijil = −Rjiil , this yields − ijk Rijkl ei ej ek = 2 i Ricil ei , and in particular X X − Rijkl ei ej ek el = 2 Ricil ei el = 2S, (11.7) i

ijkl

since the Ricci curvature is symmetric. This proves the second form of the Weitzenb¨ock identity.

11.6 Spinor Bundles Let M be a Riemannian manifold. The multivector bundle ∧M = 4M over M is a globally well-defined vector bundle over M , built fiberwise from the tangent vector bundle T M . In this section we investigate when and how it is possible to / globally over M . What we want to achieve construct a bundle of spinor spaces 4M is the following. Definition 11.6.1 (Spinor bundle). Let M be an oriented Riemannian manifold of real dimension n. A (complex) spinor bundle over M is a complex vector bundle 4M / , of complex dimension 2bn/2c , together with linear maps ρ = ρp : Tp M → / )p ) depending smoothly on p ∈ M such that L((4M ρ(v)2 ψ = |v|2 ψ,

v ∈ C(M ; T M ), ψ ∈ C(M ; 4M / ).

/ is normed in the sense of Definition 5.3.3, that is, that each We assume that 4M / )p is equipped with a spinor inner product (·, ·i and a spinor conjugation fiber (4M ·† , which vary smoothly with p. Complexifying each fiber of the multivector bundle as in Section 1.5, we obtain a globally well-defined complex vector bundle 4Mc . Applying Lemma 5.1.4 fiberwise, we see that for a spinor bundle we have homomorphisms 4Mc → / ) of complex algebras at each p ∈ M . Assuming M to be oriented, we L(4M have globally defined main reflectors wn as in Definition 5.2.1. Since ρ(wn ) = ±I in odd dimension, we see that orientability is necessary for the existence of spinor bundles in this case. In even dimension, ρ(wn ) will yield globally defined subbun± dles 4 / M , which are fundamental in Chapter 12.4. For the construction of spinor bundles over an orientable Riemannian manifold M , we fix a bundle atlas for T M of bundle charts µα (p) ∈ L(Rn ; Tp M ), p ∈ Uα , α ∈ I. We assume that each µα (p) is an isometry and that {Uα } is a good cover of M . We also assume that all the charts µα (p) are orientation-preserving. In particular, we have orientation-preserving transition maps µβα (p) ∈ SO(Rn ). Locally over Uα , the existence of a spinor bundle is clear. Indeed, let ρ0 : Rn → L(4R / n)

414

Chapter 11. Multivector and Spinor Bundles

be the complex spinor space from Definition 5.2.4. Define (4M / )p := 4R / n and ρα (v) := ρ0 (µ−1 for v ∈ T M , p ∈ U . This clearly yields a spinor bundle (p)v), p α α locally over Uα . To construct a spinor bundle globally over M , we study the transition maps µβα (p) = µβ (p)−1 µα (p) ∈ SO(Rn ). These clearly satisfy the symmetry condition µαβ = µ−1 βα on Uβα , and the transitivity condition µγβ µβα = µγα on Uγβα := Uγ ∩ Uβ ∩ Uα . Definition 11.6.2 (Spin structure). Let M be an oriented Riemannian manifold, with a bundle atlas for T M with transition maps µβα (p) ∈ SO(Rn ) as above. A spin structure for M is a choice of smooth functions qβα (p) ∈ Spin(Rn ) ⊂ 4ev Rn such that µβα (p)v = qβα (p)vqβα (p)−1 ,

v ∈ Tp M, p ∈ Uβα , α, β ∈ I,

−1 on Uβα , and the transitivity and that satisfy the symmetry condition qαβ = qβα condition qγβ qβα = qγα , on Uγβα .

Since the map Spin(Rn ) → SO(Rn ) from Proposition 4.1.9 is a covering map, there are lifts qβα such that µβα v = −1 qβα vqβα , and there are two possible such lifts, differing only in sign. Choosing −1 appropriate signs, we can always satisfy the symmetry condition qαβ = qβα . However, Proposition 4.1.9 shows only that qγβ qβα = ±qγα . To investigate when it is ˇ possible to choose signs so that the transitivity condition holds, there is a Cech cohomology criterion for existence of spin structures analogous to that for orientability in Proposition 11.2.10.

Proposition 11.6.3 (Second Stiefel–Whitney class). Let M be an oriented Riemannian manifold, with a bundle atlas for T M with transition maps µβα (p) ∈ SO(Rn ), and let qβα (p) ∈ Spin(Rn ), p ∈ Uβα as above. Consider the sheaf Z2 on the good ˇ cover {Uα } as in Example 11.2.9. Define the Cech 2-cochain f ∈ C 2 (U ; Z2 ) by ( 0, qαγ qγβ qβα = 1, f ({α, β, γ}) := 1, qαγ qγβ qβα = −1. ˇ Then ∂2 f = 0, and the Cech cohomology class [f ] does not depend on the choice of lifts qβα . There exists a spin structure for M if and only if there exists g ∈ C 1 (U ; Z2 ) such that ∂1 g = f , that is, if [f ] = [0].

415

11.6. Spinor Bundles

ˇ cohomology class w2 (M ) := [f ] is called the second Stiefel–Whitney The Cech class of M . Proof. A straightforward verification shows that f is a well-defined element of C 2 (U ; Z2 ), that is, symmetric with respect to permutations of α, β, γ. To show that ∂2 f ({α, β, γ, δ}) = f ({α, β, γ}) + f ({α, β, δ}) + f ({α, γ, δ}) + f ({β, γ, δ}) =0

mod 2,

it suffices to see that (qαγ qγβ qβα )(qαβ qβδ qδα )(qαδ qδγ qγα )(qβγ qγδ qδβ ) = 1, by the symmetry of f . Since the left-hand side is ±1, it suffices to show that the scalar part w0 of this multivector w is 1. Using that (u 4 v)0 = (v 4 u)0 , this follows −1 from the assumed symmetry condition qβα = qαβ . 0 If qβα are other choices of lifts, define ( g({α, β}) :=

0, 1,

0 qαβ = qαβ , 0 qαβ = −qαβ .

This is a well-defined element of C 1 (U ; Z2 ) because of the assumed symmetry 0 condition on qβα and qβα , and we see that f 0 ({α, β, γ}) = f ({α, β, γ}) + g({α, β} + g({α, γ}) + g({β, γ})

mod 2,

where f 0 ∈ C 2 (U ; Z2 ) is the 2-cochain for this new choice. Therefore f 0 = f + ∂1 g, 0 so that [f 0 ] = [f ]. Since there exists a spin structure for M if and only if lifts qβα 0 can be chosen such that f = 0, this completes the proof. We now define a normed spinor bundle 4M / , using a given spin structure {qβα }. Locally in each Uα , by fixing an ON-frame, we have a spinor bundle over Uα with fiber 4R / n and representation ρα (v)ψ = ρ0 (µ−1 (p)v) α over each p ∈ Uα , α ∈ I. In terms of the spin structure, the bundle transition maps for T M are µβα (p)v = qβα (p)vqβα (p)−1 ,

αβ ∈ I.

To define a fiber of a global spinor bundle at p ∈ M , let I p := {α ∈ I ; p ∈ Uα }

416

Chapter 11. Multivector and Spinor Bundles

and consider the set I p × 4R / n . The symmetry and transitivity conditions on qβα show that the relation (α, ψ) ∼ (β, φ), defined to hold if φ = qβα (p).ψ is an equivalence relation on I p × 4R / n . We define the fiber of the global spinor bundle at p ∈ M to be the set of equivalence classes (4M / )p := {[(α, ψ)] ; α ∈ I p , ψ ∈ 4R / n }. With fiber space 4R / n and bundle charts ψ ∈ 4R / n , p ∈ Uα , α ∈ I,

qα (p)ψ := [(α, ψ)],

/ globally over M . Furthermore, defining we obtain a well-defined vector bundle 4M the Tp M representation on (4M / )p to be ρ(v)[(α, ψ)] := [(α, ρα (v)ψ)],

v ∈ Tp M, ψ ∈ 4R / n , p ∈ Uα , α ∈ I,

/ as a spinor bundle over M . we obtain 4M To norm 4M / , we fix spinor inner products and conjugations on 4R / n as in Definition 5.3.3. Define a spinor inner product and conjugation on each fiber (4M / )p by ([(α, ψ)], [(α, φ)]i := (ψ, φi, †

[(α, ψ)] := [(α, ψ † )], for ψ, φ ∈ 4R / n , p ∈ Uα , α ∈ I. / Exercise 11.6.4. Verify the details of the above construction and show that 4M indeed is a well-defined normed spinor bundle over M . Concretely, show • that ∼ is an equivalence relation and that (4M / )p is a linear space in a natural way, • that qα (p) defines bundle charts for a vector bundle 4M / , • that ρ is well defined independent of α as a consequence of qβα being the rotors representing the rotations µβα , • that the spinor inner product and conjugation are well defined independent of α as a consequence of Rn rotors qβα being real multivectors acting isometrically on 4R / n. ˇ Each spin structure on M , represented by a Cech 1-cochain g ∈ C 1 (U ; Z2 ),

∂1 g = [f ] = [0],

417

11.6. Spinor Bundles

yields a normed spinor bundle 4M over M as above. Consider two spinor bundles / / 1 M and 4 / 2 M corresponding to g1 , g2 ∈ C 1 (U ; Z2 ) with ∂1 g1 = ∂1 g2 = 0. We 4 say that 4 / 1 M and 4 / 2 M are isomorphic as normed spinor bundles over M if there exists a section F ∈ C ∞ (M ; L(4 / 1M ; 4 / 2 M )) such that at each p ∈ M , the linear map F = F (p) : (4 / 1 M )p → (4 / 2 M )p is an isometry that commutes with spinor conjugation and F (ρ1 (v)ψ) = ρ2 (v)(F ψ),

v ∈ Tp M, ψ ∈ (4 / 1 M )p .

Proposition 11.6.5 (Isomorphic spinor bundles). Let M be an oriented Riemannian manifold, and consider two spinor bundles 4 / 1 M and 4 / 2 M corresponding to g1 , g2 ∈ C 1 (U ; Z2 ) with ∂1 g1 = ∂1 g2 = 0 as above. Then 4 / 1 M and 4 / 2 M are isomorphic as normed spinor bundles over M if and only if there exists h ∈ C 0 (U ; Z2 ) such that g1 = g2 + ∂0 h. ˇ This means that there is a one-to-one correspondence between the Cech co1 homology classes in H (U ; Z2 ) and isomorphism classes of spinor bundles over M , provided that w2 (M ) = [0]. Analogously to the remark following Theorem 10.6.5, we note that this shows in particular that the number of elements in H 1 (U ; Z2 ) does not depend on the choice of good cover, and that the number of isomorphism classes of spinor bundles over M does not depend on the choice of Riemannian metric, but only the global topology of M . Proof. Assume that g1 = g2 + ∂0 h. This means that 1 2 qβα = (−1)hα +hβ qβα , j writing qβα for the two spin structures, j = 1, 2, and hα := h({α}). Consider the linear maps 4R / n → 4R / n : ψ 7→ (−1)hα ψ

in Uα . These are seen to be compatible, that is, give well-defined maps (4 / 1 M )p → (4 / 2 M )p by inspection of the equivalence relations ∼1 and ∼2 , and yield an isomorphism of normed spinor spaces. Conversely, assume that there exists an isomorphism F ∈ C ∞ (M ; L(4 / 1M ; 4 / 2 M )). In Uα , consider Fα : 4R / n → 4R / n defined by [(α, Fα (ψ))] = F ([(α, ψ)]),

ψ ∈ 4R / n.

We verify that Fα is a spinor map induced by the identity on Rn , as in Proposition 5.3.5. By uniqueness, there exists hα ∈ Z2 such that Fα = (−1)hα I. At

418

Chapter 11. Multivector and Spinor Bundles

p ∈ Uβα , we have 2 [(β, (−1)hα qβα ψ)] = [(α, (−1)hα ψ)] = [(α, Fα ψ)] = F ([(α, ψ)]) 1 1 1 = F ([(β, qβα ψ)]) = [(β, Fβ qβα ψ)] = [(β, (−1)hβ qβα ψ)],

where the equivalence classes on the first line are with respect to ∼2 , and those 1 2 on the second line are with respect to ∼1 . It follows that qβα = (−1)hα +hβ qβα , so that g1 = g2 + ∂0 h. We next proceed to analysis on a fixed normed spinor bundle 4M / over a spin manifold M . Definition 11.6.6 (Spinor fields and frames). Sections of a normed spinor bundle 4M / over an oriented Riemannian manifold M are referred to as spinor fields on M . We mainly use Ψ, Φ to denote spinor fields. We use induced ON-frames for 4M / , constructed as follows. Fix the standard representation of Rn from Example 5.1.5. This means that we index the standard basis {ei } for Rn by −m ≤ i ≤ m, including i = 0 if n = 2m + 1 and excluding i = 0 if n = 2m. The spinor space 4R / n equals ∧Cm as a complex linear space, with basis {es } indexed by s ⊂ {1, . . . , m}. When viewing ∧Cm as 4R / n , we write /es := es . Given an ON-frame {ei } for T M , indexed as the Rn standard basis above, consider the isometry Rn → Tp M that maps the standard basis onto the frame. By Proposition 5.3.5, there is an isometry 4R / n → (4M / )p compatible with the representation and spinor conjugation, unique up to sign. We refer to either of these two ON-bases {/ es } corresponding to {/es } under these isometries as a spinor frame for 4M / induced by {ei }. By construction, frame es multivector fields act on induced frame e /t spinor fields on M , in the same way as the es act on /et in Rn . Note that s and t are subsets of different index sets as above. To avoid technicalities, we shall not write out these index sets {−m, . . . , m}, modulo 0, and {1, . . . , m}. Note that if 4M / is constructed from local ON bundle charts µα (p) for T M in Uα as above, then the spinor frame induced by the vector ON-frame {µα (p)ei } equals {[(α, /es )]}. The first step in setting up the calculus on 4M / is to identify a canonical covariant derivative on 4M / . As in Chapter 5, we shall abbreviate the action of multivectors on spinors by writing w.ψ := ρ(w)ψ,

w ∈ 4p Mc , ψ ∈ 4 / p M, p ∈ M.

Recall from Lemma 5.1.4 that the representation ρ extends to a complex algebra representation of 4Mc on 4M / . Proposition 11.6.7 (4M / covariant derivative). Let 4M / be a normed spinor bundle over a Riemannian manifold M . Then there exists a unique covariant derivative ∇v on 4M / that is

419

11.6. Spinor Bundles • compatible with the representation in the sense that ∇v (F.Ψ) = (∇v F ).Ψ + F.(∇v Ψ), where ∇v F denotes the Levi-Civita covariant derivative on 4Mc , • metric in the sense that ∂v (Ψ, Φi = (∇v Ψ, Φi + (Ψ, ∇v Φi, and †

• compatible with spinor conjugation in the sense that (∇v Ψ) = ∇v (Ψ† ), for all vectors v ∈ C(M ; T M ), multivector fields F ∈ C 1 (M ; 4Mc ), and all spinor / ). fields Ψ, Φ ∈ C 1 (M ; 4M e v are two covariant derivatives Proof. To prove uniqueness, assume that ∇v and ∇ e v − ∇v . Since with the desired properties, and let Lv := ∇ e v Ψ) − ((∂v f )Ψ + f ∇v Ψ) = f Lv Ψ, Lv (f Ψ) = ((∂v f )Ψ + f ∇ for scalar functions f and spinor fields Ψ, it follows that Lv is a linear map on each fiber (4M / )p . A similar subtraction of the identities assumed to hold for ∇ e yields the following. From Lv (F.Ψ) = F.(Lv Ψ) and the uniqueness result and ∇ in Theorem 5.2.3, it follows that Lv Ψ = λv Ψ for some λv ∈ C ∞ (M ; C). From † (λv Ψ, Φi + (Ψ, λv Φi = 0 it follows that Re λv = 0, and from (λv Ψ) = λv (Ψ† ) it e v = ∇v . follows that Im λ = 0. This proves that ∇ To show the existence of a covariant derivative with these properties, it suffices to consider the problem locally in a frame. Fix a vector ON-frame e in U ⊂ M , with induced spinor frame e /. Let Γe ∈ C ∞ (U ; L(T M ; ∧2 M )) be the Christoffel symbols for T M and define a covariant derivative X es + 21 Γe (v).Ψ ∇v Ψ := (∂v Ψs )/ s

P

of spinor fields Ψ = s Ψs e /s . To prove ∇v (F.Ψ) = (∇v F ).Ψ + F.(∇v Ψ), it suffices to consider multivector and spinor frame fields F = es and Ψ = e et } is /t . Since {/ an induced spinor frame, it is clear that es ./ et is ±1 times a spinor frame element. Therefore, by Proposition 11.5.3 we need to prove 1 et ) 2 Γe (v).(es ./

= 12 [Γe (v), es ]./ et + es .( 12 Γe (v)./ et ),

which is clear. Further we note that ∇v is a metric covariant derivative, since Γe (v) = −Γe (v), because Γe (v) ∈ 42 M , and it is compatible with spinor conjugation since Γe (v) is a real bivector field. This completes the proof. Proposition 11.6.8 (Bivector derivations). Let 4M / be a normed spinor bundle over a Riemannian manifold M , with the covariant derivative from Proposition 11.6.7, and let e = {ei (p)} be an ON-frame for T M in an open set U ⊂ M . Denote by Γe ∈ C ∞ (U ; L(∧1 M ; ∧2 M )) and R ∈ C ∞ (M ; L(∧2 M ; ∧2 M )) the Christoffel symbols and curvature operator for T M .

Chapter 11. Multivector and Spinor Bundles

420

/ es (p)} be the induced ON-frame for 4M Let e in U . Denote by ωe/ ∈ / = {/ / )) and Ω ∈ C ∞ (M ; L(∧2 M ; L(4M / ))) the Christoffel symC ∞ (U ; L(∧1 M ; L(4M / . Then bols and curvature operator for 4M ωe/ (v)Ψ = 21 Γe (v).Ψ, Ω(b)Ψ = 12 R(b).Ψ, / ). for v ∈ C(M ; T M ), b ∈ C(M ; ∧2 M ), and Ψ ∈ C(M ; 4M Proof. The result for the Christoffel symbols is contained in the proof of Proposition 11.6.7. Consider therefore the curvature operator. Defining e Ω(b)Ψ := Ω(b)Ψ − 12 R(b).Ψ, we have e e Ω(b)(F.Ψ) = 12 [R(b), F ].Ψ + F.(Ω(b)Ψ) − 12 R(b).(F.Ψ) = F.(Ω(b)Ψ), by a computation similar to the proof of (11.5). The uniqueness result in Thee e orem 5.2.3 shows that Ω(b)Ψ = λb Ψ. Since Ω(b) is skew-symmetric, we have Re λb = 0, and since it commutes with spinor conjugation, we have Im λb = 0. The fundamental first-order differential operator on M acting on sections of 4M / is the following natural generalization of the Euclidean nabla operator from Definition 9.1.3. Definition 11.6.9 (The Atiyah–Singer Dirac operator). Let 4M / be a normed spinor bundle over a Riemannian manifold M . The 4-Dirac / operator on 4M / is the operator X / := DΨ e∗i .(∇ei Ψ), Ψ ∈ C 1 (M ; 4M / ), i

where {ei } is a frame for T M , with dual frame {e∗i }. 2

/ to the The analogue of Proposition 11.5.9 is the following result relating D Laplace–Beltrami operator ∆ on 4M / . This is the key identity used in the proof of the Atiyah–Singer index theorem in Section 12.4. / Let 4M Proposition 11.6.10 (Weitzenb¨ock identity for D). / be a normed spinor bundle over a Riemannian manifold M , and let {ei } be an ON-frame for T M . Then X 1 2 / Ψ = −∆Ψ − −D ei ej .Ω(ei ∧ ej )Ψ = −∆Ψ + SΨ, 4 i<j / as in Deffor Ψ ∈ C 2 (M ; 4M / ), where ∆ is the Laplace–Beltrami operator on 4M inition 11.5.6, Ω is the curvature operator on 4M / , and S is the scalar curvature S of T M as in Definition 11.3.6.

11.7. Comments and References

421

Proof. Calculations identical to those in Proposition 11.5.9 lead to the first identity. From Propositions 11.6.8, and (11.7) in the proof of Proposition 11.5.9, we then obtain X i<j

11.7

ei ej .Ω(ei ∧ ej )Ψ =

1X Rijkl ei ej ek el .Ψ = − 14 SΨ. 8

ijkl

Comments and References

11.1–11.3 A reference for the differential geometry needed in this book, and more, is Taubes [90]. It is standard in differential geometry to identify tangent vectors v and directional derivatives ∂v . In coordinates {x1 , . . . , xn }, it is standard to denote the coordinate basis vectors {e1 , . . . , en } by {∂/∂x1 , . . . , ∂/∂xn }, and to denote the dual basis {e∗1 , . . . , e∗n } by {dx1 , . . . , dxn }. There is of course a canonical one-to-one correspondence v ↔ ∂v , but to identify v and ∂v as objects is not natural, and leads to serious notational problems. We therefore refrain from doing so. A main example is the nabla operators, whereP we define exterior and interior derivatives, and Dirac operators using ∇ = j ei ∂i in an ON-basis. It leads to obvious notational problems if we write the basis vector ei as a derivative ∂i , and also writing the dual basis covector e∗i as dxi causes problems for the inexperienced reader. We use the terminology Christoffel symbols in a more general sense than the standard usage in the literature. Normally, Christoffel symbols Γijk refer to the tangent bundle and a coordinate frame. We more generally refer to the zero-order part of the covariant derivative, in a fixed frame for a general vector bundle, as Christoffel symbols. A reference for the use of differential forms, or multicovector fields in our terminology, in differential geometry is Darling [32]. For the curvature operator, one should note the two different formalisms. Following the Cartan tradition, using differential form, we have Definition 11.3.3. Without using differential forms, the standard definition of the curvature operator is by Proposition 11.3.5. 11.4 Using an ON-frame, equations (11.3), and the equations defining curvature in Definition 11.3.3 provide a way to compute curvature. This is called the orthonormal moving frame method, and the equations are referred to as the structure equations. Some references are [96], [32], and [57]. The proof of Liouville’s theorem presented in Section 11.4 follows Flanders [39]. The methods we use require the C 3 hypothesis. However, it has been shown by Reˇsetnjak [77] that this can be weakened to C 1 only, or even local integrability of |∂i f |n , i = 1, . . . , n.

422

Chapter 11. Multivector and Spinor Bundles The conformal maps in dimension n ≥ 3 are indeed very scarce. However, by relaxing the conformal requirement and only requiring the ratio between the largest and smallest singular value of the total derivative to be uniformly bounded, one obtains a much richer class of maps: the quasiconformal maps.

11.5–11.6 Identities like those in Propositions 11.5.9 and 11.6.10, with integral versions as formulated in Section 12.1, or versions for domains with a boundary as in Theorem 10.3.6, are named after Roland Weitzenb¨ock. In the literature they are sometimes referred to as Lichnerowicz identities, in particular that / which Andr´e Lichnerowicz derived. Also related are Bochner identities, for D which are second-order analogues of these identities. In the literature, the standard approach to defining spinor bundles and / operator is via principal bundles. These are similar to vector the 4-Dirac bundles, but instead of having a family of vector spaces indexed by M , we now have a family of copies {Px }x∈M of a given group G. But in contrast to vector bundles, we do not have a fixed identity element in the fibers Px , but the group G acts freely and transitively from the right on each Px . The typical way that this construction is used in connection to vector bundles E is that the fiber Px collects all bases for Ex of a certain type, described by G, and the coordinates for vectors in Ex are described by a representation of G. For example, to construct spinor bundles starting from T M , we collect all positively oriented bases into a principal bundle P with G = SO(Rn ), / similar to Proposition 4.1.9, then lift this to a principal Spin(Rn ) bundle P by combining the frames implicit in and finally obtain a spinor bundle 4M / / with coordinates coming from the representation of Spin(Rn ) on 4R / n. P In this book, we have avoided principal bundles to minimize technicalities. However, one can show that the spinor bundles obtained from a / as above are precisely the normed spinor bundles principal spin bundle P considered in Section 11.6. In particular, they come equipped with a spinor inner product and a spinor conjugation, and the induced spinor frames from /. Definition 11.6.6 correspond to P Our discussion of spin structures follows Gilkey [43].

Chapter 12

Local Index Theorems Prerequisites: Chapter 11 should contain the material from differential geometry needed to read the present chapter. Section 12.1 builds on part of Chapter 10. Road map: Let M be a two-dimensional closed Riemannian manifold. The famous Gauss– Bonnet theorem states that Z 1 χ(M ) = S(p)dp, 4π M where S = 2R1212 denotes the scalar curvature at p ∈ M and χ(M ) = b0 (M ) − b1 (M ) + b2 (M ) is the Euler characteristic for M . Here bj (M ) are the Betti numbers for M , with the obvious generalization of Definition 7.6.3 to the compact manifold M without boundary. Hiding behind this result is the 4-Dirac operator D and the splitting L2 (M ; 4M ) = L2 (M ; 4ev M ) ⊕ L2 (M ; 4od M ). Indeed, D is a skew-adjoint operator on L2 (M ; 4M ) that swaps these two subspaces, and χ(M ) equals the index of the restriction D : L2 (M ; 4ev M ) → L2 (M ; 4od M ). To appreciate the power of the Gauss–Bonnet theorem, one must note that it relates three fundamentally different quantities, S, χ(M ), and D, where S is a local geometric quantity, χ is a global topological quantity, and D is an analytic object. © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_12

423

Chapter 12. Local Index Theorems

424

In this chapter, we study two generalizations of the Gauss–Bonnet theorem. The first is the Chern–Gauss–Bonnet theorem, which is the direct generalization to higher-dimensional manifolds. In odd dimension the Euler characteristic vanishes due to Poincar´e/Hodge duality. In the interesting case of even dimension, the scalar curvature in the integrand is replaced by the Pfaffian Pf(R), a quantity pointwise derived from the Riemann curvature operator R. The second generalization is the Atiyah–Singer index theorem for the 4/ / For this, we consider a spinor bundle 4M / over a closed oriented Dirac operator D. even-dimensional Riemannian manifold M , and consider the splitting +

+

/ M ) ⊕ L2 (M ; 4 L2 (M ; 4M / ) = L2 (M ; 4 / M) into the chiral subspaces of right- and left-handed spinor fields, which are swapped / The Atiyah–Singer index theorem, which ranks among the very top achieveby D. ments in twentieth-century mathematics, states in particular that the index of the restriction + − / : L2 (M ; 4 D / M ) → L2 (M ; 4 / M) ˆ obtained pointwise from the Riemann equals an integral over M of a quantity A(R) curvature operator. It turns out that this index is nonzero only for manifolds of ˆ does not depend on the dimension divisible by four, and the integral of A(R) choice of spinor bundle. It may happen that the integral over some manifolds M is not be integer-valued, which is for example the case for complex projective plane CP 2 , which shows that this four-dimensional real manifold does not possess a spin structure. It is not, however, the purpose of this chapter to pursue the large amount of interesting applications of these results, which can be found in the extensive existing literature. Rather our goal is to demonstrate how our systematic buildup of the multivector and spinor theory makes advanced results in modern mathematics rather easily accessible. For both proofs, we use the well known heat equation method. For a self/ adjoint elliptic differential operator D, with applications D = iD and D = iD − in mind, the strategy is as follows. Write L2 = L+ 2 ⊕ L2 for the splitting in this abstract formulation. Then by definition, the index we consider is dim N(D|L+ ) − dim N(D|L− ). 2

2

Aiming at curvature and having the Weitzenb¨ock identities in mind, we note that N(D|L± ) = N(D2 |L± ) and consider the eigenvalues of the nonnegative self-adjoint 2 2 ± 2 operator D2 . Write λ± j for the eigenvalues of D on the invariant subspaces L2 . − + Here the nonzero eigenvalues λj are the same as the nonzero eigenvalues λj , as a consequence of D+ : L2+ → L2− and D− : L2− → L2+ being adjoint operators. We now apply a suitable function f : R → R by Borel functional calculus to the self-adjoint operator D2 = D− D+ ⊕D+ D− . Choosing f (0) = 1 and f to decay fast enough toward ∞, we obtain a trace-class operator f (D2 ), and the index of

425

12.1. Fredholm Dirac Operators D+ equals the difference of traces X X f (λ− f (λj+ ) − j ) = Trf (D− D+ ) − Trf (D+ D− ). j

j

2

In the heat equation method, we choose f (λ) = e−tλ with a parameter t > 0, which yields trace-class operators by the Weyl asymptotics λj± ≈ j 2/n on an n2

dimensional manifold. The solution operators e−tD to the heat equation ∂t f + D2 f = 0 become local and converge to the identity in a suitable sense as t → 0+ . To prove the index theorems, we need to work with multivector and spinor calculus on M to identify the limit of trace differences lim+ Tr(e−tD− D+ ) − Tr(e−tD+ D− ) . t→0

Note that the existence of the limit is trivial, since the trace difference is independent of t. Sections 12.1 and 12.2 contain material on L2 Dirac operators and charts, preliminary to the proofs of the index theorems. Highlights: • The Chern–Gauss–Bonnet theorem: 12.3.1 • The Atiyah–Singer index theorem: 12.4.1

12.1 Fredholm Dirac Operators Throughout this chapter, we consider a closed compact Riemannian manifold M without boundary. In this section we study the L2 properties of the Dirac oper/ on M , analogously to our study of d and δ on affine domains in ators D and D Chapter 10. However, now our main work concerns the geometry of the manifold M rather than the boundary ∂M = ∅. In particular, no boundary conditions are needed for the operators. Consider first the multivector operators D, d, and δ. By transferring the Euclidean result to M using pullbacks and pushforwards as in Exercise 11.2.7, it follows that d and −δ defined on C ∞ (M ; ∧M ) are formally adjoint operators. From this it follows that D = d + δ, defined on C ∞ (M ; 4M ), is formally skew-adjoint. Using the covariant nabla expression for D from Definition 11.5.4, we can also verify this directly on M as follows. Localizing with a partition of unity, we may assume that F, G ∈ C0∞ (U ; 4M ) in an open set U ⊂ M , in which we have an ON-frame {ei } for T M . Define the vector field X v := hei 4 F, Giei . i

Chapter 12. Local Index Theorems

426 We compute div v =

X

h(∇ei ei ) 4 F, Gi + hei 4 ∇ei F, Gi + hei 4 F, ∇ei Gi

i

+

X hei 4 F, Gihej , ∇ej ei i.

(12.1)

ij

P P Since i ei hej , ∇ej ei i = − i ei hei , ∇ej ej i = −∇ej ej in the last term, this canRcels the first term. The skew-adjointness now follows from the divergence theorem div v dp = 0 from Exercise 11.2.11. M A similar calculation, replacing 4 by ∧, gives a second proof of the fact that −δ is a formal adjoint of d on M . The following extension of the domains of operators in the natural distributional sense yields closed and densely defined linear operators in L2 (M ; 4M ). Definition 12.1.1 (L2 operators). Let M be a closed Riemannian manifold. Consider the equation Z hF 0 , Gi + hF, DGi dp = 0. (12.2) M

The domain D(d) of the exterior derivative in L2 (M ; ∧M ) is the set of F ∈ L2 (M ; ∧M ) for which there exists F 0 ∈ L2 (M ; ∧M ) such that (12.2), with D = δ, holds for all G ∈ C ∞ (M ; ∧M ). For F ∈ D(d), we define dF := F 0 . The domain D(δ) of the interior derivative in L2 (M ; ∧M ) is the set of F ∈ L2 (M ; ∧M ) for which there exists F 0 ∈ L2 (M ; ∧M ) such that (12.2), with D = d, holds for all G ∈ C ∞ (M ; ∧M ). For F ∈ D(δ), we define δF := F 0 . The domain D(D) of the Dirac operator in L2 (M ; 4M ) is the set of F ∈ L2 (M ; 4M ) for which there exists F 0 ∈ L2 (M ; 4M ) such that (12.2), with D = D, holds for all G ∈ C ∞ (M ; 4M ). For F ∈ D(D), we define DF := F 0 . In the absence of boundary ∂M , we can prove full Sobolev H 1 regularity of fields F ∈ D(D) following the same route as in Section 10.3. We start with the integral form of the Weitzenb¨ock identity. Proposition 12.1.2 (Integral Weitzenb¨ock identity). Let M be a closed Riemannian manifold. Then Z Z 2 2 |DF |2 dp (|dF | + |δF | )dp = M M Z X Z Z X 1 1 hRijkl ei ej F ek el , F idp, (12.3) = |∇ei F |2 dp + S|F |2 dp + 4 M 8 M M i ijkl

for all F ∈ H 1 (M ; ∧M ). Here {ei } is an ON-frame for T M , and R and S denote the Riemann curvature operator and scalar curvature for T M .

427

12.1. Fredholm Dirac Operators

Proof. Assume first that F ∈ C 2 (M ; 4M ). Then the result follows from the nilpotence and formal adjointness of d and −δ, and duality from Propositions 11.5.5 and 11.5.9. Since all terms are continuous in H 1 norm, a limiting argument finishes the proof. Proposition 12.1.3 (Regularity). Let M be a closed Riemannian manifold. Then D = d + δ with D(D) = D(d) ∩ D(δ) = H 1 (M ; 4M ) and equivalences of norms. Proof. Clearly H 1 (M ; 4M ) ⊂ D(d) ∩ D(δ) ⊂ D(D), so it suffices to show that D(D) ⊂ H 1 (M ; 4M ). To this end, we argue as in the proof of Theorem 10.3.3, but replacing the ball B by the n-torus T n . (i) Consider first the case M = T n with the flat metric given by the constant standard inner product on Rn . This is the manifold T n = Rn /Zn , with charts obtained from quotient maps Rn → T n in the natural way. A multivector field on T n corresponds to a Zn -periodic field on Rn , and D(D) = H 1 (T n ; ∧T n ) follows from Plancherel’s theorem and the Fourier series analogue of Example 10.1.8. (ii) Next consider a general closed Riemannian manifold M , and F ∈ D(D). Using a partition of unity, we may assume that supp F is contained in the range Mα ⊂ M of a coordinate chart µα : Dα → Mα . Assuming that Dα ⊂ Rn is small, 1 (x) on T n such we identify Dα with a subset of T n , and define a C ∞ metric gij that µα : Dα → Mα is an isometry. In this way, we may regard Mα as an open subset of the n-torus 1 T n with geometry determined by the metric gij . 1 To show regularity of F , we perturb the metric gij on T n continuously to the flat Euclidean metric δij , by letting t 1 gij (x) := (1 − t)δij + tgij (x),

0 ≤ t ≤ 1.

t It is clear that gij defines a Riemannian metric on T n for each 0 ≤ t ≤ 1. Write Ttn for this Riemannian manifold, and note that all Ttn are the same as C ∞ manifolds, but their Riemannian geometries are distinct. Consider the family of bounded linear operators I + Dt : H 1 (Ttn ; 4Ttn ) → L2 (Ttn ; 4Ttn ), (12.4)

where Dt is the Dirac operator on 4Ttn . By Proposition 12.1.2, these are all injective semi-Fredholm maps with k∇ ⊗ F k2 + kF k2 ≈ kDF k2 + kF k2 = kF + DF k2 , using the formal skew-adjointness of Dt for the last equality. We want to show that these are all invertible maps. This is clear for I + D0 from (i). To apply the

428

Chapter 12. Local Index Theorems

method of continuity for semi-Fredholm operators, as explained in Section 6.4, we define auxiliary maps of multivector fields At : C(T0n ; 4T0n ) → C(Ttn ; 4Ttn ). Since the Ttn are all equal to T n as C ∞ manifolds, we can realize the multivector bundle 4Ttn as T n × ∧Rn , where ∧Rn is defined independently of any metric as in Section 2.1, before equipping each fiber with the inner product int . We therefore let At be the identity map on T n × ∧Rn , but with duced by gij different but equivalent metrics in the domain and range. In particular we obtain bounded and invertible linear maps At : L2 (T0n ; 4T0n ) → L2 (Ttn ; 4Ttn ) and At : H 1 (T0n ; 4T0n ) → H 1 (Ttn ; 4Ttn ), and the method of continuity applies to n n 1 n n A−1 t (I + Dt )At : H (T0 ; 4T0 ) → L2 (T0 ; 4T0 ).

It follows that I + D1 : H 1 (T1n ; 4T1n ) → L2 (T1n ; 4T1n ) is invertible. Since H 1 ⊂ D(D) on T1n and I +D is injective on D(D) by formal skew-adjointness, this shows that H 1 = D(D). This completes the proof. Recall the definition (6.4) of adjointness in the sense of unbounded operators. Proposition 12.1.4 (Duality). Let M be a closed Riemannian manifold. Then d and −δ, with domains as in Definition 12.1.1, are adjoint operators in L2 (M ; ∧M ) in the sense of unbounded operators. The Dirac operator D, with domain as in Definition 12.1.1, is a skew-adjoint operator in L2 (M ; 4M ) in the sense of unbounded operators. Proof. Consider first d∗ = −δ. As in Proposition 10.2.3, it is clear from Definition 12.1.1 that it suffices to show that for every F ∈ D(d), there exists a family of fields Ft ∈ C ∞ (M ; ∧M ) such that Ft → F and dFt → dF in L2 (M ; ∧M ) as t → 0+ . By localizing with a partition of unity, we may assume that F is supported in a chart Mα . Since d commutes with the pullback µ∗α , the result follows from the Euclidean case in Section 10.2. To prove D∗ = −D, we similarly note that it suffices to show that for every F ∈ D(D), there exists a family of fields Ft ∈ C ∞ (M ; 4M ) such that Ft → F and DFt → DF in L2 (M ; 4M ) as t → 0+ . In this case, we obtain from Proposition 12.1.3 that F ∈ H 1 (M ; 4M ). This completes the proof, since C ∞ is dense in H 1 . From Propositions 10.1.2 and 10.1.6, we now obtain the Hodge decomposition L2 (M ; ∧M ) = R(d) ⊕ C(D) ⊕ R(δ) of L2 multivector fields on M , with finite-dimensional cohomology space C(D) = N(d) ∩ N(δ), closed subspaces R(d) and R(δ) of exact and coexact fields, and compact potential maps.

429

12.1. Fredholm Dirac Operators In terms of the Dirac operator, this means that D : H 1 (M ; 4M ) → L2 (M ; 4M )

is a Fredholm operator. Since it is a skew-adjoint L2 operator, its index is zero. Concretely, N(D) = R(D)⊥ = C(D). Splitting the fields further into homogeneous k-vector fields, we write C(D; ∧k ) := C(D) ∩ L2 (M ; ∧k M ) and define the Betti numbers bk (D) := dim C(D; ∧k ),

k = 0, 1, 2, . . . , n.

As explained in the introduction, we are particularly interested in the following integer. Definition 12.1.5 (Euler characteristic). Let M be a closed Riemannian manifold. The Euler characteristic of M is the alternating sum X (−1)k bk (M ) χ(M ) := k

of Betti numbers, or equivalently the index of the restricted Dirac operator D : H 1 (M ; 4ev M ) 7→ L2 (M ; 4od M ). Exercise 12.1.6. Compute the three Betti numbers b0 (M ), b1 (M ), and b2 (M ) for the two-dimensional sphere S 2 as well as the two-dimensional torus T 2 = S 1 × S 1 , using Hodge star maps and the Gauss–Bonnet theorem stated in the introduction. Note that there exists a flat metric on T 2 , that is, a metric for which the curvature operator vanishes. Show also that there exists no flat metric on S 2 . / acting on sections of a given Next consider the spinor Dirac operator D normed spinor bundle 4M / over an oriented closed Riemannian manifold M . For a general real Riemannian manifold, without any further complex structure, we / in terms of some invariantly defined nilpotent operators Γ and Γ∗ , cannot write D analogously to D = d + δ for the 4-Dirac operator. But besides this, we have an / with similar properties to those of D above. Some details are as L2 operator D follows. / is formally • A calculation like (12.1), replacing ei 4 F by ei .Ψ, shows that D / ). skew-adjoint on C ∞ (M ; 4M / from C ∞ to D(D) / • Similar to Definition 12.1.1, we extend the domain of D / ) for which there exists Ψ0 ∈ consisting of spinor fields Ψ ∈ L2 (M ; 4M L2 (M ; ∧M ) such that Z / (Ψ0 , Φi + (Ψ, DΦi dp = 0 M

/ := Ψ0 . / we define DΨ / ). For Ψ ∈ D(D), holds for all Φ ∈ C (M ; 4M ∞

Chapter 12. Local Index Theorems

430

• From Propositions 11.5.5 and 11.6.10, we readily obtain an integral Weitzenb¨ock identity Z X Z Z 1 |∇ei Ψ|2 dp + |DΨ|2 dp = S|Ψ|2 dp, 4 M M i M / ). Here {ei } is an ON-frame and S is the scalar valid for Ψ ∈ H 1 (M ; 4M curvature for T M . / = H 1 (M ; 4M • Also for the 4-Dirac / operator, we have D(D) / ). However, some care concerning the map At is needed in adapting the perturbation argument in the proof of Proposition 12.1.3. On each n-torus Ttn we have an ON-frame {ei } for the tangent bundle globally defined on Ttn , for example by polar decomposition as in Definition 12.2.4. Upon mapping this onto the standard basis for Rn , we use the trivial spinor bundle 4T / tn = Ttn × 4R / n over Ttn . Note that there is no problem with topological obstructions here, since the problem is local. Defining At as the identity map on T n × 4R / n, but with different metrics on the domain and range, the proof proceeds as for D. / is a skew-adjoint operator in • From H 1 regularity it follows that D L2 (M ; 4M / ) in the sense of unbounded operators, and that / : H 1 (M ; 4M D / ) → L2 (M ; 4M / ) / = R(D) / ⊥, is a Fredholm operator. Since it is skew-adjoint, we have N(D) and in particular, its index is zero. • In contrast to the 4-Dirac operator, here we do not have access to a finer splitting into subspaces like L2 (M ; 4k M ) and a notion of Betti numbers, but assuming that the dimension n = 2m of M is even, we have the pointwise splitting into chiral subspaces. Let en be the unit n-vector field on M describing the orientation of M , and define the main reflector wn := i−m en as in Definition 5.2.1. Consider the pointwise orthogonal splitting +

−

L2 (M ; 4M / ) = L2 (M ; 4 / M ) ⊕ L2 (M ; 4 / M ), ±

where (4 / M )p are the ranges of the projections wn± = 12 (1 ± wn ), at each p ∈ M. ± / swaps the subspaces L2 (M ; 4 • It follows from Definition 11.6.9 that D / M ), since the ei swap them, while the ∇v preserve them, since the Γe (v) do so. This is so because the vectors ei anticommute with wn , while the bivectors Γe (v) commute with wn . Our goal in Section 12.4 is to calculate the index of the restricted 4-Dirac / operator + − / : H 1 (M ; 4 D / M ) → L2 (M ; 4 / M ).

12.2. Normal Coordinates

12.2

431

Normal Coordinates

Let M be a closed Riemannian manifold, with tangent bundle T M , Levi-Civita covariant derivative ∇v , and Riemann curvature operator R. Fix a base point q ∈ M . To do computations on M near q it is useful to choose a chart for M around that q is as good as possible. We use the exponential map expq : Tq M → M , which is the map taking a tangent vector v ∈ Tq M to the point p ∈ M at distance d(p, q) := |v| from q along the geodesic, that is, the length-minimizing curve γ, starting at γ(0) = q with tangent vector γ 0 (0) = v/|v| and parametrized by arc length. Definition 12.2.1. A normal chart for M around q is a chart µ : D → M, with D = B(0, r) ⊂ Rn , obtained by fixing an ON-basis for Tq M and identifying it with Rn , and applying the exponential map Tq M → M . The supremum over δ such that r ≥ δ can be chosen at each q ∈ M is called the injectivity radius of M . We start by formulating a condition for a chart to be normal, in terms of the metric gij := hei , ej i, where we write ei for the coordinate frame vector fields µ(ei ), by slight abuse of notation. In such a normal chart, we write ωij for the Christoffel symbols. Proposition 12.2.2 (Normal chart equations). Let µ : D → M be a chart such that µ(0) = q, with metric coordinates gij (x). Then µ is a normal chart if and only if X xi = gij (x)xj , x ∈ D, i = 1, . . . , n. j

Proof. That radial lines are geodesic is equivalent to ∇µ

x

(x) (µx (x/|x|))

x ∈ D,

= 0,

and that they are parametrized by arc length means that hµx (x/|x|), µx (x/|x|)iµ(x) = 1. In coordinates, the second equation reads X X gij (x)xi xj = x2i , (12.5) ij

i

P whereasP the first equation becomes ∂x (x/|x|) + ij ei hωij , xixj /|x| = 0, or equivalently j hωij , xixj = 0 for all i. Using Exercise 11.1.8, this reads X (2∂i gjk (x) − ∂k gij (x))xi xj = 0. (12.6) ij

Chapter 12. Local Index Theorems

432

P Under (12.5), equation (12.6) is seen to be equivalent to ij xi ∂i (gjk (x)xj −xk ) = 0. Interpreting this as the radial derivatives of gjk (x)xj − xk vanishing, the result follows. We next show that for normal coordinates, all second derivatives of the metric at q are given by the curvature coefficients. Proposition 12.2.3 (Metric Taylor expansion). Let µ : D → M be a normal chart for M at q, with metric gij (x). Then 1X gij (x) = δij − Rikjl (q)xk xl + O(|x|3 ), i, j = 1, . . . , n, 3 kl

where Rijkl (q) are the Riemann curvature coefficients in the coordinate basis {ei }, which is ON at q. In particular, gij = δij , ∂i gjk = 0, ∂k hωij , el i = −∂j ∂l gik , ∂i ∂j gkl + ∂i ∂k gjl + ∂j ∂k gil = 0, ∂i ∂j gkl = ∂k ∂l gij , Rijkl = ∂i ∂l gjk − ∂j ∂l gik ,

∂k ∂l gij = − 13 (Rikjl + Riljk ),

hold at q, for all i, j, k, l. Here the ωij denote the Christoffel symbols in the coordinate frame {ei }. Proof. We prove the identities at q, from which the stated Taylor expansion follows. Differentiating xi = gim (x)xm three times gives δij = (∂j gim )xm + gij , 0 = (∂k ∂j gim )xm + ∂j gik + ∂k gij , 0 = (∂l ∂k ∂j gim )xm + ∂k ∂j gil + ∂l ∂j gik + ∂l ∂k gij . At q, the first equation gives gij = δij , using the second equation three times gives ∂i gjk = −∂j gik = ∂k gij = −∂i gkj , thus ∂i gjk = 0, and the third equation gives ∂k ∂j gil + ∂l ∂j gik + ∂l ∂k gij = 0. Using this last equation three times gives ∂i ∂j gkl = −∂i ∂k gjl − ∂k ∂j gil = (∂k ∂l gji + ∂i ∂l gjk ) + (∂j ∂l gik + ∂k ∂l gij ) = 2∂k ∂l gij − ∂i ∂j gkl , thus ∂i ∂j gkl = ∂k ∂l gij . For the Christoffel symbols, we get from Exercise 11.1.8 that X 2h∇ei ej , ek i = 2 gαk hωαj , ei i = ∂i gjk + ∂j gik − ∂k gij . α 1 2 ∂k (∂l gji

At q, this gives ∂k hωij , el i = + ∂j gli − ∂i glj ) = 12 (−∂j ∂l gki − ∂k ∂i glj ) = −∂j ∂l gik . This gives curvature coefficients Rijkl = hdωij , ek ∧ el i = ∂k hωij , el i − ∂l hωij , ek i = −∂j ∂l gik + ∂i ∂l gjk ,

12.2. Normal Coordinates

433

using Exercise 7.4.9. Finally, we have Rikjl +Riljk = (∂i ∂l gkj −∂k ∂l gij )+(∂i ∂k glj − ∂l ∂k gij ) = −∂k ∂l gij − 2∂k ∂l gij = −3∂k ∂l gij . This completes the proof. Besides the coordinate frame {ei }, we require an ON-frame {ei } in which to do multivector calculus. We shall use the following construction. Definition 12.2.4 (Polar ON-frame). Let µ be a normal chart for M around q, with coordinate frame {ei }. Let G = (gij ) be the metric, with gij = hei , ej i, and consider its positive inverse square root G−1/2 = (αij )ij . The polar ON-frame for µ is the ON-frame {ei }, where X ei := αki ek = µ(G−1/2 ei ). k

This is a pointwise construction of {ei } from {ei } based on polar factorization as in Proposition 1.4.4. Note that {ei } is ON at q, whereas {ei } is an ON-frame in all of the chart. Indeed, hei , ej i = hµ(G−1/2 ei ), µ(G−1/2 ej )i = hG−1/2 ei , GG−1/2 ej )iRn = δij . Proposition 12.2.5. Let µ : D → M be a normal chart for M at q, with metric gij (x). Denote the Christoffel symbols and Riemann curvature coefficients in the associated polar ON-frame {ei } by Γij and Rijkl respectively. Then the identities Γij = 0,

∂k hΓij , el i = 12 Rijkl ,

hold at q for all i, j, k, l. Proof. It follows from Proposition 12.2.3 that G = (gij ) = I + O(|x|2 ). Thus the change-of-basis matrix from {ei } to {ei } is A := G−1/2 = I − 12 (G − I) + O(|x|4 ). By Proposition 11.3.4, Γij = (A−1 )ik (ωkl Alj + dAkj ) = ωij − 12 dgij + O(|x|2 ), where the {ωij } denote the Christoffel symbols in the coordinate frame {ei }. Using Proposition 12.2.3, we get at q that ∂k hΓij , el i = ∂k hΓij , el i = −∂j ∂l gik − 12 ∂k ∂l gij = −∂j ∂l gik − 12 (−∂j ∂l gik − ∂j ∂k gil ) = 12 (−∂j ∂l gik + ∂j ∂k gil ) = 12 Rijkl .

Note that at q, the curvature coefficients Rijkl are the same in the frames {ei } and {ei }, since these coincide there.

434

12.3

Chapter 12. Local Index Theorems

The Chern–Gauss–Bonnet Theorem

In this section, we prove the following local index theorem for the 4-Dirac operator. Recall from Definition 12.1.5 the relation between the Euler characteristic χ(M ), the Betti numbers bk (M ), and the index of D restricted to even multivector fields. Theorem 12.3.1 (Chern–Gauss–Bonnet). Let M be an n-dimensional compact and closed Riemannian manifold. If n = 2m is even, then 1 m Z χ(M ) = hPf(R), dpi, 2π M where Pf(R) denotes the Pfaffian of the Riemann curvature operator R. If n is odd, then χ(M ) = 0. We begin by explaining Pf(R). First we replace the antisymmetric matrix R = (Rij ) of bivectors by a scalar antisymmetric matrix A = (Aij ), where Aij ∈ R. Here Rij := R(ei ∧ ej ) ∈ ∧2 M in an ON-frame {ei }, or equivalently, by symmetry of R, the Rij are the bivectors from Definition 11.3.3. Definition 12.3.2 (Pfaffian). Let the dimension n = 2m be even, let A ∈ SO(Rn ) be an antisymmetric matrix, and let b ∈ ∧2 Rn be the bivector that represents A as in Proposition 4.2.3. Then the Pfaffian of A is Pf(A) := hb ∧ · · · ∧ b, e1 ∧ · · · ∧ en i/(n/2)!, where the first exterior product for b is m-fold. The Pfaffian behaves like a square root of the determinant for skew-symmetric matrices, as the following shows. Proposition 12.3.3 (Pfaffian algebra). Let the dimension n = 2m be even. We have the following properties of the Pfaffian functional of a skew-symmetric matrix A. (i) For A in standard form as in Proposition 4.3.6(ii), we have

0 −x1 Pf ... 0 0

x1 0 .. .

··· ··· .. .

0 0 .. .

0 0

··· ···

0 −xm

0 0 .. .

= x1 · · · xm . xm 0

435

12.3. The Chern–Gauss–Bonnet Theorem (ii) For general skew-symmetric A = (Aij ), we have the formula X 1 Pf(A) = m (i1 , j1 , . . . , im , jm )Ai1 j1 · · · Aim jm , 2 m! i ,j ,...,i ,j 1

1

m

m

where (i1 , j1 , . . . , im , jm ) = hei1 ∧ ej1 ∧ · · · ∧ eim ∧ ejm , e1 ∧ · · · ∧ en i is the sign of the permutation (i1 , j1 , . . . , im , jm ) → (1, . . . , n). (iii) If A ∈ SO(Rn ) and T ∈ L(Rn ), then Pf(T AT ∗ ) = det(T )Pf(A). Proof.P(i) and (ii) follow from the facts that b = x1 e12 + · · · + xm e2m−1,2m and b = 12 ij Aij eij in these cases respectively. To prove (iii), we use Proposition 2.7.1 to obtain T AT ∗ v = T (b x (T ∗ v)) = (T b) x v. Thus T b represents T AT ∗ , and (T b) ∧ · · · ∧ (T b) = T (b ∧ · · · ∧ b) = det(T )(n/2)!Pf(A)e1 ∧ · · · ∧ en , which proves (iii).

Definition 12.3.4 (Exterior Pfaffian). Let M be a Riemannian manifold of dimension n = 2m, with Riemann curvature operator R. Define the Pfaffian X 1 Pf(R) = m (i1 , j1 , . . . , im , jm )Ri1 j1 ∧ · · · ∧ Rim jm ∈ ∧n M. 2 m! i ,j ,...,i ,j 1

1

m

m

The generalization from A to R amounts to replacing the real field R by the algebra (∧ev Rn , ∧), which is commutative by Proposition 2.1.14. Proposition 12.3.3(iii) generalizes to show that Pf(R) is independent of the choice of positively oriented ON-frame. Lemma 12.3.5. The integrand hPf(R), dpi in Theorem 12.3.1 does not depend on the choice of ON-frame or orientation. P ˜i = Proof. Consider two ON-frames {ei } and {˜ ei }, related as e j ej αji . By ˜ ij = Proposition 11.3.4, the corresponding curvature coefficients are related as R P k,l αki Rkl αlj . This gives X ˜i j ∧ · · · ∧ R ˜i j (i1 , j1 , . . . , im , jm )R 1 1 m m i1 ,j1 ,...,im ,jm

=

X

X

hei1 αk1 i1

∧ ej1 αl1 j1 ∧

e1 ∧ · · · ∧ en i Rk1 l1 =

X

hA∗ (ek1

∧ el1 ∧

· · · ∧ ekm

∧

∧ ejk αlm jm ,

X k1 ,l1 ,...,km ,lm

Rk1 l1

∧

· · · ∧ Rkm lm

∧ elm ), e1 ∧

k1 ,l1 ,...,km ,lm

= det(A)

· · · ∧ eim αkm im

i1 ,j1 ,...,im ,jm

k1 ,l1 ,...,km ,lm

· · · ∧ Rkm lm ,

· · · ∧ en i Rk1 l1

∧

· · · ∧ Rkm lm

Chapter 12. Local Index Theorems

436

where A = (αij ). This, together with the observation that the oriented measure dp also changes sign if det(A) = −1, proves the proposition. 0 R12 , so that Example 12.3.6 (Gauss–Bonnet). If n = 2, then R = −R12 0 hPf(R), e12 i = R1212 = 12 S. Thus the Gauss–Bonnet theorem stated in the introduction to this chapter is the two-dimensional case of the Chern–Gauss–Bonnet theorem (Theorem 12.3.1). Exercise 12.3.7. Write down explicitly the Chern–Gauss–Bonnet integrand in dimension n = 4. We now embark on the proof of the Chern–Gauss–Bonnet theorem, which covers the remainder of this section. When n is odd, consider the Hodge star map ∗ : L2 (M ; 4k M ) → L2 (M ; 4n−k M ). This gives an isomorphism between Hk (M ) and Hn−k (M ) by Exercise 11.2.7. When n is odd, this Poincar´e duality implies that χ(M ) = 0. We next consider the nontrivial case of even dimension n = 2m. Following the heat equation method described in the introduction, we calculate χ(M ) = dim N(D|L2 (M ;4ev M ) − dim N(D|L2 (M ;4od M ) ) = dim N(D2 |L2 (M ;4ev M ) − dim N(D2 |L2 (M ;4od M ) ) 2

2

= Tr(etD |L2 (M ;4ev M ) ) − Tr(etD |L2 (M ;4od M ) ), for all t > 0. The second identity is valid because D is a normal operator, and the last identity follows from the general fact that the nonzero eigenvalues of operators A∗ A and AA∗ are the same, in particular for A = D : L2 (M ; 4ev M ) → L2 (M ; 4od M ). The idea of the proof is to compute this trace difference in the limit as t → 0+ . According to Proposition 11.5.9, the square D2 of the Dirac operator differs from the Laplace–Beltrami operator only by zero-order terms. Thus it is 2 reasonable to expect the operator etD to be an integral operator resembling the solution operator Z 2 t∆ e f (x) = (4πt)−n/2 e−|x−y| /(4t) f (y)dy Rn n

for the heat equation on R from Example 6.3.3. With this in mind, we make an ansatz of the form Z N X 2 (Ht f )(p) := (4πt)−m e−d(p,q) /(4t) tk H k (p, q)f (q)dq, (12.7) M

k=0

where N < ∞ is to be chosen. Here d(p, q) denotes the shortest distance between points p and q on M , and below, we shall choose linear maps H k (p, q) ∈

437

12.3. The Chern–Gauss–Bonnet Theorem

L(4(Tq M ); 4(Tp M )) depending smoothly on p, q ∈ M . We want to choose H k 2 such that Ht f approximates etD f well for small t, and in particular, H 0 (q, q) = I, 2

for all q ∈ M . This will ensure that limt→0+ Ht f = f = limt→0+ etD f for all 2 2 f . Secondly, since ∂t etD f = D2 etD f , we want (∂t − D2 )Ht f to be as small as possible. Lemma 12.3.8. Let µ : D → M be a normal chart for M around q = µ(0). Push forward the radial vector field x in D to the vector field rq (p) := µx (x), p = µ(x), and let dq (p) := |x| denote the shortest distance from p to q on M . Then ∇(d2q ) = 2rq

∆(d2q ) = 2n + ∂rq ln g,

and

where g = det(gij ) and ∆ is the scalar Laplace–Beltrami operator on M , and 2 2 (∂t − D2 )(tk−m e−dq /(4t) h) = tk−m e−dq /(4t) t−1 ∇rq + 14 ∂rq (ln g) + k h − D2 h , for all h ∈ C 2 (M ; 4M ). P Proof. Define the frame e∗i = j g ij ej dual to the coodinate frame {ei }, where (g ij ) denotes the inverse of the metric (gij ). Proposition 12.2.2 shows that X X X ∇(d2q ) = e∗i ∂i |x|2 = g ij ej 2xi = 2 xj ej = 2rq . i

ij

j

To compute ∆(d2q ), we use Exercise 11.5.8 to get ∆(d2q ) =

√1 g

X

√ ∂i ( gg ij 2xj ) =

√2 g

ij

X

√ ∂i ( gxi ) = 2n + ∂rq ln g.

i

For the last formula, clearly 2

2

∂t (tk−m e−dq /(4t) ) = (d2q /(4t2 ) + (k − m)/t)tk−m e−dq /(4t) . By Proposition 11.5.9, it suffices to prove the identity with D2 replaced by the Laplace–Beltrami operator on 4M . We compute X 2 2 2 2 ∆(e−dq /(4t) h) = (∆e−dq /(4t) )h + 2 (∂ei e−dq /(4t) )∇ei h + e−dq /(4t) ∆h, i

in an ON-frame {ei }. Here 2

∆e−dq /(4t) =

√1 g

X ij

d2 2n + ∂rq ln g −d2 /(4t) 2 √ q ∂i ( gg ij ∂j e−dq /(4t) ) = − e q 4t2 4t

Chapter 12. Local Index Theorems

438 and 2

X

2

ei (∂ei e−dq /(4t) ) = −

i

2 2 1 X ei (∂ei d2q )e−dq /(4t) = −t−1 e−dq /(4t) rq . 2t i

Combining these calculations proves the stated formula.

Applying ∂t − D2 to (12.7), we obtain (∂t − D2 )Ht f Z N X 1 −d2q /(4t) = ˙ q)f (q) e tk−m−1 ∇rq + 14 ∂rq (ln g) + k H k (p, m (4π) M k=0

−

N +1 X

tk−m−1 D2 H k−1 (p, ˙ q)f (q) dq.

k=1

This leads us to the following recursive definition of H k (p, q). For p, q ∈ M e k (p, q) such such that d(p, q) < δ, where δ is the injectivity radius of M , define H 0 k e e that H (q, q) = I and H (p, q) solves k e (p, e k−1 (p, ∇rq + 14 ∂rq (ln g) + k H ˙ q)f (q) = D2 H ˙ q)f (q) (12.8) e −1 (p, q) = 0. Note that for each k, for k = 0, . . . , N and f (q) ∈ (4M )q . Here H (12.8) is an ordinary differential equation along the geodesic from q to p, and that e k (q, q)f (q) = k −1 D2 H e k−1 (q, q)f (q) is specified, since for k ≥ 1, the initial value H rq (q) = 0. Existence theory for ordinary differential equations shows that this uniquely determines maps H k (p, q) ∈ L(4(Tq M ); 4(Tp M )) depending smoothly on p, q ∈ M , at distance < δ. To extend this construction to general p, q ∈ M , we make a smooth cutoff as follows. Let η ∈ C ∞ (R) be such that η(x) = 1 for x < δ/3 and η(x) = 0 for x > 2δ/3. Define e k (p, q), H k (p, q) := η(d(p, q))H where we understand that H k (p, q) = 0 if d(p, q) ≥ 2δ/3. 2 We have constructed Ht and next compare this ansatz to etD . Let Z Kt f (p) = Kt (p, q)f (q)dq := (∂t − D2 )Ht f (p). M

Here the kernel Kt (p, q) is a smooth function of p, q ∈ M and t > 0, and by construction we have 2

Kt (p, q) = −(4π)−m tN −m e−d(p,q)

/(4t)

D2 H N (p, ˙ q) 2

when d(p, q) < δ/3. Consider the difference Ht f − etD f , which satisfies 2

(∂t − D2 )(Ht f − etD f ) = Kt f

12.3. The Chern–Gauss–Bonnet Theorem

439

2

and initial conditions limt→0+ (Ht f − etD f ) = 0. Integration gives Ht f − e

tD2

t

Z

2

e(t−s)D Ks f ds,

f=

(12.9)

0

from which we deduce the following. Proposition 12.3.9 (Trace formula). We have the formula Z 1 m m ev M (H χ(M ) = (q, q)) − Tr (q, q)) dq Tr od M (H 4 4 (4π)m M for the index of D : H 1 (M ; 4ev M ) 7→ L2 (M ; 4od M ). Proof. We estimate the trace norm of (12.9). See Example 6.4.8. To estimate the trace norm kKs kL1 (L2 (M ) , we factorize into Hilbert–Schmidt operators Ks = (I + D)−j (I + D)j Ks . We use that the eigenvalues of D grow in size as |λk | = k 1/n , and as a consequence (I + D)−j is a Hilbert–Schmidt operator if j > m = n/2. For the second operator and d(p, q) < δ/3, we have |(I +D)j Ks (p, ˙ q)| . sN −m−j , whereas for d(p, q) ≥ δ/3, 2 we have |(I + D)j Ks (p, ˙ q)| . e−δ /(36s) s−m−j−2 . This shows that the Hilbert– Schmidt norm of (I + D)j Ks is bounded by sN −m−j for 0 < s < 1. Choosing N > n therefore shows that kKs kL1 (L2 (M ) . 1, and in particular, 2

kHt − etD kL1 (L2 (M )) ≤

Z

t

2

ke(t−s)D kL(L2 (M )) kKs kL1 (L2 (M )) ds . t → 0,

0

as t → 0+ , since D is a skew-adjoint operator, and trace-class operators form an ideal in L(L2 ). Since the trace functional is continuous in the trace norm, we get 2 2 χ(M ) = lim Tr(etD |L2 (M ;4ev M ) ) − Tr(etD |L2 (M ;4od M ) ) t→0+ = lim+ Tr(Ht |L2 (M ;4ev M ) ) − Tr(Ht |L2 (M ;4od M ) ) t→0

N

X 1 = lim tk−m (4π)m t→0+ k=0

Z

Tr4ev M (H k (q, q)) − Tr4od M (H k (q, q)) dq.

M

Since we know that the limit exists, all the terms 0 ≤ k < m must be zero, and we have proved the stated formula for χ(M ). It remains to compute Tr4ev M (H m (q, q)) − Tr4od M (H m (q, q)).

Chapter 12. Local Index Theorems

440

To this end, we fix a normal chart µ : x 7→ µ(x) around q and let {ei } be the associated polar ON-frame from Definition 12.2.4. To handle linear operators in L(∧M ), we use the frame − − {es+1 · · · e+ sk et1 · · · etl }s1 0

Equating the tm terms yields H m (0, 0) = im p0 (a1 , . . . , am ), where p0 is the polynomial from Definition 12.4.2. This coincides with the result for H0m from the k−1 k−1 recurrence (β + k)Hβk = D2 Hβ+2 + D0 Hβ−2 . (ii) By Definition 12.4.2, we have H m (0, 0) = im p(TrA2 , TrA4 , . . .).

448

Chapter 12. Local Index Theorems

By analytic continuation, this continues to hold for all aj ∈ C. Furthermore, it continues to hold for general A ∈ SO(Rn ), by changing basis to an orthogonal splitting for A as in Proposition 4.3.6 and using the invariance of the trace functional. Similarly, using instead the commutative algebra (∧ev Rn , ∧), it is clear that the recurrence (12.12) with D1 = 0 will result in H m (q, q) = im p(TrR2 , TrR4 , . . .). Using Exercise 12.4.5, we obtain that the index of + − / : L2 (M ; 4 D / M ) → L2 (M ; 4 / M)

is

1 (4π)m

Z

(2i)m im hp(TrR2 , TrR4 , . . .), dqi.

M

Since we have shown in Exercise 12.4.7 that the index vanishes when m is odd, this completes the proof of the Atiyah–Singer index theorem. To quote a famous mathematician: I think I’ll stop here.

12.5

Comments and References

12.3 The Gauss–Bonnet theorem goes back to Gauss, who, however, never published it. Bonnet first published a special case of the theorem in 1848. The full Chern–Gauss–Bonnet theorem for a general compact manifold was first proved by Chern in 1945. 12.4 The index theorem for general elliptic differential operators was proved by Atiyah and Singer in 1963, and is regarded as one of the great landmarks of twentieth-century mathematics. A main special case of this general index theorem is the index theorem for the 4-Dirac / operator. In fact, this operator was rediscovered by Atiyah and Singer in their work on the index theorem. The early proofs of the index theorem used different methods. The now dominant heat equation method, which we use, originates in the works of Atiyah, Bott, and Patodi [3] and Gilkey. Standard references for the index theory of Dirac operators include the books by Gilkey [43], Berline, Getzler, and Vergne [20] and Lawson and Michelsohn [63]. Further treatments of the index theory can be found in Taylor [92] and Gilbert and Murray [42]. The proofs of the index theorems given here follow the book [96] by Yanlin Yu, and they do not rely on the theory of pseudodifferential operators. Some minor variations on the setup from [96] include our use of the pointwise constructed polar ON-frame from Definition 12.2.4 and our use of skew-adjoint Dirac operators rather than self-adjoint Dirac operators. By a Wick-type rotation argument it is straightforward to see that this does not

12.5. Comments and References

449

affect the index of the operator. A reference for results on differential equations that we use, including Weyl’s law and existence results for ordinary ˆ differential equations, is Taylor [91, 92]. The local indices Pf(R) and A(R), the integrands appearing in the Chern–Gauss–Bonnet and Atiyah–Singer index theorems, are examples of what are called characteristic classes, which are certain polynomials in the curvature bivectors. A reference for the theory of symmetric polynomials, used in Definition 12.4.2, is Nicholson [73].

Bibliography [1] Ahlfors, L. V. M¨obius transformations in Rn expressed through 2 × 2 matrices of clifford numbers. Complex Variables Theory Appl. 5, 2-4 (1986), 215–224. [2] Arnold, D., Falk, R., and Winther, R. Finite element exterior calculus, homological techniques, and applications. Acta Numer. 255, 15 (2006), 1–155. [3] Atiyah, M., Bott, R., and Patodi, V. K. On the heat equation and the index theorem. Invent. Math. 19 (1973), 279–330. [4] Auscher, P., Axelsson, A., and Hofmann, S. Functional calculus of Dirac operators and complex perturbations of Neumann and Dirichlet problems. J. Funct. Anal. 255, 2 (2008), 374–448. [5] Auscher, P., Axelsson, A., and McIntosh, A. Solvability of elliptic systems with square integrable boundary data. Ark. Mat. 48 (2010), 253– 287. [6] Auscher, P., Hofmann, S., Lacey, M., McIntosh, A., and Tchamitchian, P. The solution of the Kato square root problem for second order elliptic operators on Rn . Ann. of Math. (2) 156, 2 (2002), 633–654. [7] Axelsson, A. Oblique and normal transmission problems for Dirac operators with strongly Lipschitz interfaces. Comm. Partial Differential Equations 28, 11-12 (2003), 1911–1941. [8] Axelsson, A. Transmission problems for Dirac’s and Maxwell’s equations with Lipschitz interfaces. PhD thesis, The Australian National University, 2003. Available at https://openresearch-repository.anu.edu.au/handle/1885/46056. [9] Axelsson, A. Transmission problems and boundary operator algebras. Integral Equations Operator Theory 50, 2 (2004), 147–164. [10] Axelsson, A. Transmission problems for Maxwell’s equations with weakly Lipschitz interfaces. Math. Methods Appl. Sci. 29, 6 (2006), 665–714. © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8

451

452

Bibliography

[11] Axelsson, A., Grognard, R., Hogan, J., and McIntosh, A. Harmonic analysis of Dirac operators on Lipschitz domains. In Clifford analysis and its applications (Prague, 2000), vol. 25 of NATO Sci. Ser. II Math. Phys. Chem. Kluwer Acad. Publ., Dordrecht, 2001, pp. 231–246. [12] Axelsson, A., Keith, S., and McIntosh, A. Quadratic estimates and functional calculi of perturbed Dirac operators. Invent. Math. 163, 3 (2006), 455–497. [13] Axelsson, A., Kou, K., and Qian, T. Hilbert transforms and the Cauchy integral in euclidean spaces. Studia Math. 193, 2 (2009), 161–187. [14] Axelsson, A., and McIntosh, A. Hodge decompositions on weakly Lipschitz domains. In Advances in analysis and geometry, Trends Math. Birkh´’auser, Basel, 2004, pp. 3–29. [15] Axler, S. Down with determinants! Amer. Math. Monthly 102, 2 (1995), 139–154. [16] Axler, S., Bourdon, P., and Ramey, W. Harmonic function theory. No. 137 in Graduate Texts in Mathematics. Springer-verlag, 1992. ´n, A. Riesz continuity of the [17] Bandara, L., McIntosh, A., and Rose Atiyah-Singer Dirac operator under perturbations of the metric. Math. Ann. 370, 1-2 (2018), 863–915. ´n, A. Riesz continuity of the Atiyah-Singer [18] Bandara, L., and Rose Dirac operator under perturbations of local boundary conditions. To appear in Communications in Partial Differential Equations, 2019, DOI: 10.1080/03605302.2019.1611847. [19] Benn, I., and Tucker, R. An introduction to spinors and geometry with applications in physics. Adam Hilger, Ltd., 1987. [20] Berline, N., Getzler, E., and Vergne, M. Heat kernels and Dirac operators. No. 298 in Grundlehren der Mathematischen Wissenschaften. Springerverlag, 1992. [21] Bott, R., and Tu, L. Differential forms in algebraic topology. No. 82 in Graduate Texts in Mathematics. Springer-Verlag, 1982. [22] Bourguignon, J.-P. Spinors, Dirac operators, and changes of metrics. In Differential geometry: geometry in mathematical physics and related topics (Los Angeles, CA, 1990), vol. 54 of Proc. Sympos. Pure Math. Amer. Math. Soc., 1993, pp. 41–44. [23] Brackx, F., Delanghe, R., and Sommen, F. Clifford Analysis. No. 76 in Research Notes in Mathematics. Pitman, 1982. ¨ cker, T., and tom Dieck, T. Representations of compact Lie groups. [24] Bro Graduate Texts in Mathematics. Springer-Verlag, 1985.

Bibliography

453

[25] Cartan, E. Sur certaines expressions diff´erentielles et le probl`eme de Pfaff. Ann. Sci. Ecole Norm. Sup. 3, 16 (1899), 239–332. [26] Cartan, E. The theory of spinors. The M.I.T. Press, Cambridge, Mass., 1967. [27] Clifford, W. Applications of Grassmann’s Extensive Algebra. Amer. J. Math. 1, 4 (1878), 350–358. [28] Coifman, R. R., McIntosh, A., and Meyer, Y. L’int´egrale de Cauchy d´efinit un op´erateur born´e sur L2 pour les courbes lipschitziennes. Ann. of Math. (2) 116, 2 (1982), 361–387. [29] Colton, D., and Kress, R. Integral equation methods in scattering theory, first ed. John Wiley & Sons, New York, 1983. [30] Colton, D., and Kress, R. Inverse acoustic and electromagnetic scattering theory, second edition ed. Springer-Verlag, Berlin, 1998. [31] Costabel, M., and McIntosh, A. On Bogovskiˇi and regularized Poincar´e integral operators for the de Rham complexes on Lipschitz domains. Math. Z. 265, 2 (2010), 297–320. [32] Darling, R. Differential forms and connections. Cambridge University Press, 1994. [33] Delanghe, R., Sommen, F., and Souvcek, V. Clifford algebra and spinorvalued functions. A function theory for the Dirac operator. Mathematics and its Applications. Kluwer Academic Publishers Group, 1992. ´, J. The tragedy of Grassmann. Linear and multilinear algebra [34] Dieudonne 8, 1 (1979/80), 1–14. [35] Evans, L. Partial differential equations, vol. 19 of Graduate Studies in Mathematics. American Mathematical Society, 1998. [36] Evans, L., and Gariepy, R. Measure theory and fine properties of functions. Studies in Advanced Mathematics. CRC Press, 1992. [37] Fearnley-Sander, D. Hermann Grassmann and the creation of linear algebra. Amer. Math. Monthly 86, 10 (1979), 809–817. [38] Federer, H. Geometric measure theory. Die Grundlehren der mathematischen Wissenschaften, Band 153. Springer-Verlag, 1969. [39] Flanders, H. Liouville’s theorem on conformal mapping. J. Math. Mech. 15 (1966), 157–161. [40] Fulton, W. Algebraic topology. No. 153 in Graduate Texts in Mathematics. Springer-verlag, 1985. [41] Fulton, W., and Harris, J. Representation theory. A first course. No. 129 in Graduate Texts in Mathematics. Springer-verlag, 1991.

454

Bibliography

[42] Gilbert, J., and Murray, M. Clifford algebras and Dirac operators in harmonic analysis. Cambridge Studies in Advanced Mathematics. Cambridge University Press, 1991. [43] Gilkey, P. Invariance theory, the heat equation, and the Atiyah-Singer index theorem. No. 11 in Mathematics Lecture Series. Publish or Perish, Inc., 1984. [44] Grassmann, H. Die Lineale Ausdehnungslehre, ein neuer Zweig der Mathematik. 1844. andig und in strenger Form [45] Grassmann, H. Die Ausdehnungslehre: Vollst¨ bearbeitet. 1864. [46] Greub, W. Multilinear algebra. Second edition. Universitext. SpringerVerlag, 1864. [47] Griffiths, P., and Harris, J. Principles of algebraic geometry. Pure and Applied Mathematics. Wiley-Interscience, 1978. [48] Grisvard, P. Elliptic problems in nonsmooth domains. Monographs and Studies in Mathematics. Pitman, 1985. [49] Grove, L. Classical groups and geometric algebra. No. 39 in Graduate Studies in Mathematics. American Mathematical Society, 2002. [50] Harvey, F. R. Spinors and calibrations. No. 9 in Perspectives in Mathematics. Academic Press, Inc., 1990. [51] Helsing, J. Solving integral equations on piecewise smooth boundaries using the RCIP method: a tutorial. Available at https://arxiv.org/abs/1207.6737v9. [52] Helsing, J., and Karlsson, A. On a Helmholtz transmission problem in planar domains with corners. J. Comput. Phys. 371 (2018), 315–332. [53] Helsing, J., and Karlsson, A. Physical-density integral equation methods for scattering from multi-dielectric cylinders. J. Comput. Phys. 387 (2019), 14–29. obius differential geometry. [54] Hertrich-Jeromin, U. Introduction to M¨ No. 300 in London mathematical society lecture notes series. Cambridge university press, 2003. [55] Hestenes, D. Space-time algebra. Gordon and Breach, 1966. [56] Hestenes, D. New foundations for classical mechanics. No. 99 in Fundamental Theories of Physics. Kluwer Academic Publishers Group, 1999. [57] Hestenes, D., and Sobczyk, G. Clifford algebra to geometric calculus. A unified language for mathematics and physics. Fundamental Theories of Physics. D. Reidel Publishing Co., 1984.

Bibliography

455

[58] Hitchin, N. The Dirac operator. In Invitations to geometry and topology, Oxf. Grad. Texts Math. Oxford Univ. Press, Oxford, 2002, pp. 208–232. [59] Hladik, J. Spinors in physics. Graduate Texts in Contemporary Physics. Springer-Verlag, 1999. [60] Jancewicz, B. Multivectors and Clifford algebra in electrodynamic. No. 11 in Mathematics Lecture Series. World Scientific Publishing Co., Inc., 1988. [61] Kato, T. Perturbation theory for linear operators, second ed. SpringerVerlag, Berlin, 1976. Grundlehren der Mathematischen Wissenschaften, Band 132. [62] Kress, R. Linear integral equations. No. 82 in Applied Mathematical Sciences. Springer-Verlag, New York, 1999. [63] Lawson, H. B., J., and Michelsohn, M.-L. Spin geometry. No. 38 in Princeton Mathematical Series. Princeton University Press, 1989. [64] Lounesto, P. Clifford algebras and spinors. London Mathematical Society Lecture Note Series. Cambridge University Press, 2001. [65] Madsen, I., and Tornehave, J. From calculus to cohomology. de Rham cohomology and characteristic classes. No. 11 in Mathematics Lecture Series. Cambridge University Press, 1997. [66] McIntosh, A. Clifford algebras and the higher-dimensional Cauchy integral. In Approximation and function spaces (Warsaw, 1986), vol. 22 of Banach Center Publ. PWN, Warsaw, 1989, pp. 253–267. [67] McIntosh, A., and Mitrea, M. Clifford algebras and Maxwell’s equations in Lipschitz domains. Math. Methods Appl. Sci. 22, 18 (1999), 1599–1620. [68] McIntosh, A., and Monniaux, S. Hodge-Dirac, Hodge-Laplacian and Hodge-Stokes operators in Lp spaces on Lipschitz domains. Rev. Mat. Iberoam. 34, 4 (2018), 1711–1753. [69] Meyer, Y. Wavelets and operators. No. 37 in Cambridge Studies in Advanced Mathematics. Cambridge University Press, 1992. [70] Mitrea, D., Mitrea, M., and Monniaux, S. The Poisson problem for the exterior derivative operator with Dirichlet boundary condition in nonsmooth domains. Commun. Pure Appl. Anal. 7, 6 (2008), 1295–1333. [71] Mitrea, M. Clifford Wavelets, Singular Integrals and Hardy Spaces. No. 1575 in Lecture Notes in Mathematics. Springer, 1994. [72] Morrey, C.B., J. Multiple integrals in the calculus of variations. No. 130 in Die Grundlehren der mathematischen Wissenschaften. Springer-Verlag, 1966. [73] Nicholson, W. K. Introduction to abstract algebra. Wiley-Interscience. John Wiley & Sons, 2007.

456

Bibliography

[74] Picard, R. An elementary proof for a compact imbedding result in generalized electromagnetic theory. Math. Z. 187, 2 (1984), 151–164. [75] Porteous, I. Clifford algebras and the classical groups. No. 50 in Cambridge Studies in Advanced Mathematics. Cambridge University Press, 1995. [76] Pryde, A. J. The five lemma for Banach spaces. Proc. Amer. Math. Soc. 65, 1 (1977), 37–43. [77] Resetnjak, J. G. Liouville’s conformal mapping theorem under minimal regularity hypotheses. Sibirsk. Mat. 8 (1967), 835–840. [78] Riesz, M. Clifford numbers and spinors. With the author’s private lectures to E. Folke Bolinder. No. 54 in Fundamental Theories of Physics. Kluwer Academic Publishers Group, 1993. [79] Rindler, W. Relativity. Special, general, and cosmological. Oxford University Press, 2006. ´n, A. Boosting the Maxwell double layer potential using a right spin [80] Rose factor. To appear in Integral Equations and Operator Theory. ´n, A. Fredholm theory, singular integrals and T b theorems. Unpublished [81] Rose lecture notes from 2011, available at http://www.math.chalmers.se/∼rosenan/FST.html. ´n, A. Layer potentials beyond singular integral operators. Publ. Mat. [82] Rose 57, 2 (2013), 429–454. ´n, A. Square function and maximal function estimates for operators [83] Rose beyond divergence form equations. J. Evol. Equ. 13, 3 (2013), 651–674. ´n, A. A spin integral equation for electromagnetic and acoustic scat[84] Rose tering. Appl. Anal. 96, 13 (2017), 2250–2266. [85] Schwarz, G. Hodge decomposition - a method for solving boundary value problems. No. 1607 in Lecture Notes in Mathematics. Springer-Verlag, 1995. [86] Sommen, F. Spingroups and spherical means. In Clifford algebras and their applications in mathematical physics (Canterbury, 1985), vol. 183 of NATO Adv. Sci. Inst. Ser. C Math. Phys. Sci. Reidel, Dordrecht, 1986, pp. 149–158. [87] Sommen, F. Spingroups and spherical means. II. In Proceedings of the 14th winter school on abstract analysis (Srni, 1986), no. 14. Rend. Circ. Mat. Palermo (2) Suppl., 1987, pp. 157–177. [88] Sommen, F. Spingroups and spherical means. III. In Proceedings of the Winter School on Geometry and Physics (Srni, 1988), no. 21. Rend. Circ. Mat. Palermo (2) Suppl., 1989, pp. 295–323. [89] Stein, E., and Weiss, G. On the theory of harmonic functions of several variables. I. The theory of H p -spaces. Acta Math. 103 (1960), 25–62.

Bibliography

457

[90] Taubes, C. Differential geometry. Bundles, connections, metrics and curvature. No. 23 in Oxford Graduate Texts in Mathematics. Oxford University Press, 2011. [91] Taylor, M. Partial differential equations. I. Basic theory. No. 115 in Applied Mathematical Sciences. Springer-Verlag, 1996. [92] Taylor, M. Partial differential equations. II. Qualitative studies of linear equations. No. 116 in Applied Mathematical Sciences. Springer-Verlag, 1996. [93] Thaller, B. The Dirac equation. Springer-Verlag, 1992.

Texts and Monographs in Physics.

[94] Verchota, G. Layer potentials and regularity for the Dirichlet problem for Laplace’s equation in Lipschitz domains. J. Funct. Anal. 59, 3 (1984), 572–611. [95] Watson, G. N. A Treatise on the Theory of Bessel Functions. Cambridge University Press, New York, 1944. [96] Yu, Y. The index theorem and the heat equation method. No. 2 in Nankai Tracts in Mathematics. World Scientific Publishing Co., 2001.

Index ∗w, 54 :=, =:, 2 A∗ , 17 At , 17 E ± , 277 Ek N BVP, 314 Ek S BVP, 315 Ek± , 308 H k (p, q), 436 Jρ , 188 R, 403 Tp M , 387 V ∗, 6 V∞ , 135 Wl , 10 Ws , 10 Wt , 10 [v1 , v2 ], 236 , 197 ∆, 197, 411 Γ(z), 192 Γe , 390 Γk , 328 H, 84 Ω, 401 Psh , 266 Psm , 266 Ψ fundamental solution, 260 spinor field, 289 Ψ± k , 304 Θ∗, 54 ≈, 186 4M , 408 4V , 79

z c , 17 χ(M ), 429 b , 109 4V b 4V (2), 139 4, 76 ∂k , 368 curl F , 214 δΘ = ∇ y Θ, 212 DF = ∇ 4 F , 257 D, 410 div F , 214 D(T ), 202 x, ˙ 215 V, 135 ei , 385 (s, t), 32 grad f , 214 G(T ), 202 b k V , 34 ∧ ˆ 442 A, w, b 33 [w], 35 bwc, 36 y, 52 ↔, 2 ., 186 µ, 385 ρx , 187 C⊥ (D), 244 Ck (D), 244 Fvt , 235 L(V1 ; V2 ), 13, 205 Lv F , 236 | · |, 5, 201

© Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8

459

460 ∇, 211 ∇ ⊗ F , 211 ∇v , 388 n, 27 N(T ), 202 ωe , 388 O(V ), 110 ⊕, 4 dwe, 36 ⊗, 14 k · k, 201 ∂i , 187 ∂v , 387 Pf, 434 π1 (M ), 118 Pin(V ), 110 R(T ), 202 x, 52 dx, 45 w, 33 ρ∗ Θ, 217 ρ∗ F , 217 /es , 418 4M / , 413 4V / , 163 h·, ·i, 5, 18 h·, ·iV , 304 / 420 D, / = ∇.Ψ, 289 DΨ e /s , 418 σn−1 , 194 SO(3), 86 SO(V ), 110 SO+ (W ), 128 ψ † , 168 Spin(V ), 110 Spin+ (W ), 128 hvi2 , 5 (·, ·i, 18 SU(2), 175 ρ˜∗ F , 217 ×, 58 Tr, 205 d, 351

Index δ, 351 SO(V ), 113 Spin(V ), 113 ∧, 27 ∧M , 390 ∧V , 30 ∧ev V , 30 ∧od V , 30 ∧k V , 27 ∧k (M ), 46 bk (D), 244 dF = ∇ ∧ F , 212 e∞ , 135 e∗i , 7 es , 27 f (x) ∗ g(x), 192 gij , 386 k-covector, 47 k-form, 225 linear, 225 k-vector, 27 v ± , 95 w.ψ := ρ(w)ψ, 163 adjoint formal, 203 operator, 16 unbounded operators, 203 affine space, 2 algebra associative, 4 completeness, 165 alternating, 24 alternating forms, 50 Amp`ere-Maxwell law, 293 anticommutation, 63, 76 antiparticles, 300 Atiyah–Singer–Patodi BVP, 324 Banach space, 200 Betti numbers, 244 Bianchi identities, 404 bilinear, 14 boundary orientation, 228

461

Index bundle Clifford, 408 multivector, 390 spinor, 413 vector, 385 BVP=boundary value problem, 309 Cartan formula, 238 lemma, 33 method, 405 Cartan–Dieudonn´e theorem, 107 Cauchy extension, 279 principal value integral, 277 singular integral, 277 Cauchy–Pompeiu formula, 260 Cauchy–Riemann equations, 258 celestial sphere, 143 character, 174 charge conjugation, 302 chart, 385 normal, 431 normalized pushforward, 391 pullback, 391 pushforward, 391 chiral, 300, 319 Christoffel bivectors, 390 symbols, 388 class function, 175 Clifford algebra, 90 algebra standard, 79 cone, 109 product, 75 trace, 87 closed, 240 closed range theorem, 206 coboundary operator, 368 cochain, 368 cofactor formula, 62 cohomology space, 373 Coifman–McIntosh–Meyer theorem, 279

commutation theorem, 221 commutative diagram, 370 complementary subspace, 203 complex, 369 complex eigenvector, 21 complex spinor space, 163 complex structure, 19 complexification, 20 composite, 34 conformal map, 134 cosine operator, 311 covariant derivative, 388 Cramer rule, 42 curvature operator, 401 coefficients, 404 commutator, 402 Ricci, 404 Riemann, 403 curvilinear coordinates, 222 d’Alembertian, 197 degenerate, 9 diagram chasing, 371 differential form, 253 Dirac equation, 298 Dirac matrix, 160 Dirac operator, 257 4, 287 4 on manifold, 410 4, / 289 Atiyah–Singer, 420 harmonic, 287 spherical, 274 wave, 287 Dirichlet problem, 249 domain C k , 188 Lipschitz, 188 starshaped, 190 strongly Lipschitz, 189 double covering, 112 double layer potential, 313 dual basis, 7

462 operator, 16 space, 6 duality, 5 elliptic rotation, 129 Euclidean, 5 Euler characteristic, 429 even Clifford subalgebra, 92 exact, 240, 344 expansion, 64 exponential function, 5 extension maps, 354 exterior algebra, 30 derivative, 212 product, 28 Faraday law, 293 fine sheaf, 369 five lemma, 372 flow, 235 Fourier transform, 192 fractional linear map, 135 frame, 385 polar ON, 433 fundamental group, 118 Gaffney inequality, 358 Gauss law, 293 magnetic, 293 Gegenbauer polynomial, 273 good cover, 374 Gram–Schmidt, 65 Grassmann cone, 34 Haar measure, 174 Hankel function, 196 Hardy projection, 278 subspace, 278 harmonic oscillator, 200 heat equation, 198 Helmholtz decomposition, 247 Helmholtz equation, 199, 306

Index Hermitian, 18 Hilbert space, 200 Hodge decomposition, 245 Rn , 349 abstract, 345 algebraic, 349 boundary, 331 Hodge projection, 347 Hodge star, 54 Hodge-Dirac operator, 346 homogeneous, 30 homotopy relations, 242, 364 hyperbolic rotation, 129 ideal, 164 induced basis, 27 induced map, 40 multivectors, 40 spinors, 171 inner product, 5 complex bilinear, 18 sesquilinear, 18 inner space, 36 interior derivative, 212 interior product left, 52 right, 52 inversion, 75, 134 involution, 33 isometry, 106 Jacobian, 188 K¨ unneth formula, 379 Kadlec formula, 357 Kelvin transform, 264 Lagrange identity, 74 Laplace operator, 197 spherical, 274 Laplace–Beltrami operator, 411 Legendre polynomial, 273 Levi-Civita covariant derivative, 389 Lie derivative, 236

463

Index bracket, 236, 389 light cone, 10 light-like, 10 linear space, 2 Liouville theorem conformal map, 406 monogenic field, 262 Lorentz boost, 125 M¨obius map, 137 main reflector, 161 manifold atlas, 190 chart, 190 closed, 191 oriented, 190 transition map, 190 with boundary, 191 Maxwell equations, 295 spacetime, 296 time-harmonic, 307 Mayer–Vietoris sequence, 379 Mehler formula, 200 method of continuity, 206 metric, 386 MIT bag model, 319 monogenic field, 257 two-sided, 259, 327 multicovector, 47 multilinear, 14 nabla symbol, 211 Neumann problem, 251 neutral Clifford algebra, 94 nilpotent, 213, 344 normal multivector, 66 normalized pushforward, 217 normed spinor space, 169 ON-basis, 8 operator compact, 204 Fredholm, 205

Hilbert–Schmidt, 204 semi-Fredholm, 205 trace class, 205 orientation, 44 oriented manifold, 395 measure, 46, 226 volume element, 45 orthochronous rotation, 128 rotor, 128 orthogonal complement, 7 outer space, 36 parabolic rotation, 126 parallel section, 398 transport, 398 partition of unity, 191 Pauli matrix, 101, 159 Peter–Weyl theorem, 174 pfaffian, 434 ucker relations, 67 Pl¨ Poincar´e theorem, 240 polar decomposition, 16 potential, 240 Bogovski˘ı, 362 Hodge, 246, 362 maps, 347 Poincar´e, 362 principal curvature, 357 principle of abstract algebra, 1 projection, 204 projective space, 143 pullback, 217 pushforward, 217 quaternion, 84 Rademacher theorem, 189 radiation condition Dirac, 305 Silver–M¨ uller, 308 Sommerfield, 307

464 real structure, 21 rectangular determinant, 28 reflection operator, 310 Rellich estimates, 320 Rellich lemma, 199 representation algebra, 156 group, 154, 173 irreducible, 173 minimal, 158 real Clifford algebras, 98 standard, 157 uniqueness, 162 vector, 156 reversion, 33 Riemann function, 193 Riemannian manifold, 387 Riesz formulas, 76 rotation operator, 311 rotor, 110 scalar curvature, 404 Schr¨ odinger equation, 200 Schur estimate, 202 Schur lemma, 173 second fundamental form, 357 section, 385 sheaf, 368 signature, 9 simple, 34 simply connected, 118 singular, 5 skew symmetric map, 114 snake lemma, 370 Sokhotski–Plemelj jump, 281 space-like, 10 spacetime, 10 spectral theorem, 203 sphere paradox, 146 spherical harmonics, 266 spherical monogenics, 266 spin, 86 group, 110 integral equation, 324

Index structure, 414 spinor abstract space, 154 conjugation, 168, 302 inner product, 168, 298 splitting, 185, 203 Hardy subspace, 282 Stein–Weiss vector field, 259 stereographic projection, 135 Stiefel–Whitney class first, 396 second, 415 Stokes theorem, 228 Cauchy formula, 232 classical, 230 Gauss formula, 231 Green formula, 231 Stratton–Chu formulas, 308 symmetric polynomial, 449 tangent bundle, 387 tangential multivector, 66 tensor abstract space, 154 product, 14 time-harmonic function, 18 time-like, 10 total derivative, 187 trace, 16, 205 transition map, 385 transmission problem, 326, 336 transversal magnetic, 339 ultrahyperbolic space, 117 universal property, 14, 26, 90 universality, 91 Vahlen cone, 139 Vahlen matrix, 139 vector product, 58 vector space, 2 volume-norm, 44 wave equation, 198

Index weakly singular integral, 202, 313 Weingarten map, 357 Weitzenb¨ock identity C 2 domain, 357 4M , 412 4M / , 420 integral 4M , 426 integral 4M / , 430 well-posedness, 311 Weyl representation, 161 zonal harmonic, 271 zonal monogenic, 271

465

Andreas Rosén

Geometric Multivector Analysis From Grassmann to Dirac

Birkhäuser Advanced Texts Basler Lehrbücher

Series editors Steven G. Krantz, Washington University, St. Louis, USA Shrawan Kumar, University of North Carolina at Chapel Hill, Chapel Hill, USA Jan Nekováˇr, Sorbonne Université, Paris, France

More information about this series at http://www.springer.com/series/4842

Andreas Rosén

Geometric Multivector Analysis From Grassmann to Dirac

Andreas Rosén Department of Mathematical Sciences Chalmers University of Technology and the University of Gothenburg Gothenburg, Sweden

ISSN 1019-6242 ISSN 2296-4894 (electronic) Birkhäuser Advanced Texts Basler Lehrbücher ISBN 978-3-030-31411-8 (eBook) ISBN 978-3-030-31410-1 https://doi.org/10.1007/978-3-030-31411-8 Mathematics Subject Classification (2010): 15-01, 15A72, 15A66, 35-01, 35F45, 45E05, 53-01, 58A10, 58A12, 58J20 © Springer Nature Switzerland AG 2019 This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmission or information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar methodology now known or hereafter developed. The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication does not imply, even in the absence of a specific statement, that such names are exempt from the relevant protective laws and regulations and therefore free for general use. The publisher, the authors, and the editors are safe to assume that the advice and information in this book are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or the editors give a warranty, express or implied, with respect to the material contained herein or for any errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. This book is published under the imprint Birkhäuser, www.birkhauser-science.com by the registered company Springer Nature Switzerland AG The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland

Contents Preface 1

2

3

4

ix

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

. . . . . .

1 2 5 9 13 17 21

Exterior Algebra 2.1 Multivectors . . . . . . . . . . . . . 2.2 The Grassmann Cone . . . . . . . 2.3 Mapping Multivectors . . . . . . . 2.4 Oriented Measure . . . . . . . . . . 2.5 Multicovectors . . . . . . . . . . . 2.6 Interior Products and Hodge Stars 2.7 Mappings of Interior Products . . . 2.8 Anticommutation Relations . . . . 2.9 The Pl¨ ucker Relations . . . . . . . 2.10 Comments and References . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

. . . . . . . . . .

23 24 34 40 44 47 51 61 63 67 69

Clifford Algebra 3.1 The Clifford Product . . . . . . . . . 3.2 Complex Numbers and Quaternions 3.3 Abstract Clifford Algebras . . . . . . 3.4 Matrix Representations . . . . . . . 3.5 Comments and References . . . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

73 . 74 . 82 . 89 . 93 . 102

Prelude: Linear Algebra 1.1 Vector Spaces . . . . . . . . . . 1.2 Duality . . . . . . . . . . . . . 1.3 Inner Products and Spacetime . 1.4 Linear Maps and Tensors . . . 1.5 Complex Linear Spaces . . . . 1.6 Comments and References . . .

. . . . . .

Rotations and M¨obius Maps 105 4.1 Isometries and the Clifford Cone . . . . . . . . . . . . . . . . . . . 106 4.2 Infinitesimal Rotations and Bivectors . . . . . . . . . . . . . . . . . 113 v

vi

Contents 4.3 4.4 4.5 4.6 4.7

5

6

7

8

9

Euclidean Rotations . . . . . . . Spacetime Rotations . . . . . . . Fractional Linear Maps . . . . . Mappings of the Celestial Sphere Comments and References . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

117 125 134 142 150

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

153 154 161 167 172 183

Interlude: Analysis 6.1 Domains and Manifolds . . . 6.2 Fourier Transforms . . . . . . 6.3 Partial Differential Equations 6.4 Operator Theory . . . . . . . 6.5 Comments and References . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

185 186 191 197 200 206

Multivector Calculus 7.1 Exterior and Interior Derivatives . . 7.2 Pullbacks and Pushforwards . . . . . 7.3 Integration of Forms . . . . . . . . . 7.4 Vector Fields and Cartan’s Formula 7.5 Poincar´e’s Theorem . . . . . . . . . 7.6 Hodge Decompositions . . . . . . . . 7.7 Comments and References . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

209 211 216 224 235 239 242 253

Hypercomplex Analysis 8.1 Monogenic Multivector Fields 8.2 Spherical monogenics . . . . . 8.3 Hardy Space Splittings . . . . 8.4 Comments and References . .

Spinors in Inner Product Spaces 5.1 Complex Representations . 5.2 The Complex Spinor Space 5.3 Mapping Spinors . . . . . . 5.4 Abstract Spinor Spaces . . 5.5 Comments and References .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

. . . .

255 257 265 277 283

Dirac Wave Equations 9.1 Wave and Spin Equations . . . . 9.2 Dirac Equations in Physics . . . 9.3 Time-Harmonic Waves . . . . . . 9.4 Boundary Value Problems . . . . 9.5 Integral Equations . . . . . . . . 9.6 Boundary Hodge Decompositions 9.7 Maxwell Scattering . . . . . . . . 9.8 Comments and References . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

. . . . . . . .

285 287 291 303 309 319 327 332 339

. . . .

vii

Contents 10 Hodge Decompositions 10.1 Nilpotent operators . . . . . . . . . 10.2 Half-Elliptic Boundary Conditions 10.3 Hodge Potentials . . . . . . . . . . 10.4 Bogovski˘ı and Poincar´e Potentials ˇ 10.5 Cech Cohomology . . . . . . . . . 10.6 De Rham Cohomology . . . . . . . 10.7 Comments and References . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

343 344 350 354 362 367 372 381

11 Multivector and Spinor Bundles 11.1 Tangent Vectors and Derivatives . 11.2 Multivector Calculus on Manifolds 11.3 Curvature and Bivectors . . . . . . 11.4 Conformal Maps and ON-Frames . 11.5 Weitzenb¨ock Identities . . . . . . . 11.6 Spinor Bundles . . . . . . . . . . . 11.7 Comments and References . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

. . . . . . .

383 385 390 398 405 408 413 421

12 Local Index Theorems 12.1 Fredholm Dirac Operators . . . . . . 12.2 Normal Coordinates . . . . . . . . . 12.3 The Chern–Gauss–Bonnet Theorem 12.4 The Atiyah–Singer Index Theorem . 12.5 Comments and References . . . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

. . . . .

423 425 431 434 441 448

Bibliography

451

Index

459

Preface I guess all mathematicians have had their defining moments, some events that led them to devote much of their lives and energy to mathematics. Myself, I vividly recall the spring and summer of 1997, spending my days reading about Clifford algebras in David Hestenes’s inspirational books and listening to the Beatles. Don’t misunderstand me. To a Swede, there is nothing that beats ABBA, but that summer it happened that the Clifford algebras were enjoyed in this particular way. I was a fourth-year undergraduate student at Link¨oping University studying the civil engineering program of applied physics and electrical engineering, and the very last course I took there came to change my life in a way that no one could have anticipated. The course was on “applied mathematics”, and we were supposed to pursue a math project of our choice, typically to solve some differential equation. One odd topic proposed was learning Clifford algebras, and it appealed to me. I fell deeply in love with the beauty of it all, and I read and I read. I found the biographies [34, 37] about Hermann Grassmann, and I learned what an unfortunate turn mathematics had taken since the 1800s. During my university studies I had had a sense of something missing in the vector calculus that we were taught. I remember students asking me in the linear algebra sessions that I taught how the vector product could have area as dimension while at the same time being a vector. I discovered that Grassmann had figured it all out more than 150 years ago, and now it was all strangely hidden to us students of mathematics, all but the one-dimensional vectors. No one had told me anything about vector products in dimensions other than three, or about determinants of rectangular matrices. My personal relations with the vector product had in fact begun some five years earlier, when I borrowed a telescope from my high school for a science project on satellites. Using Kepler’s laws, I calculated a formula for the altitude of a satellite’s orbit, using as input two observations of the satellite’s position and the time elapsed between the two observations. Of course you don’t need a telescope for this, it’s just to look for a slow falling star, but I did other things as well. As you may guess, I stumbled upon a curious expression involving three mixed products, for the plane of rotation of the satellite. It was only the following year, when I had started my university studies, that I learned in the linear algebra lectures that this intriguing formula was called a vector product. ix

x

Preface

A second defining moment occurred two years later, around May 1999. I was spending a Saturday or Sunday in the library at the mathematics department in Lund, and stumbled upon a friend. We started a discussion that led to a search on this rather new thing called the internet, where I found the perfect PhD supervisor, Alan McIntosh, from Australia, one of the giants in harmonic analysis and operator theory. It was a perfect match, since he was doing real analysis, singular integrals and operator theory, as well as mixing in the algebras of Clifford and Grassmann when needed. And so I ended up down under in Canberra, and spent three years applying singular integrals and Clifford algebra to solve Maxwell boundary value problems on Lipschitz domains with Alan McIntosh. The publications [11, 8, 9, 7, 14, 10] related to my thesis work are perhaps the real starting point for this book. To shed light on the confusion: Axelsson = Ros´en before 2011. The reason for telling this story is not that I think the reader is more interested in my personal story than in the subject of the book. I certainly hope not. But nothing is without context, and it may help to know the background to understand this book. The basic algebra is not new; it goes back to the pioneering works of Hermann Grassmann, first published in 1843, whose exterior algebra of multivectors is the topic of Chapter 2, and of William Kingdon Clifford from 1878, whose geometric algebra is the topic of Chapter 3. Although these algebras are geometric and useful enough that one would expect them to fit into the mainstream mathematics curriculum at a not too advanced level, this has not really happened. But over the last century, they have been rediscovered over and over ´ Cartan developed his calculus of again. Inspired by the Grassmann algebra, Elie differential forms in the early 1900s. He was also the first to discover spinors in general in 1913, which is the topic of Chapter 5. In 1928, Paul Dirac formulated his famous equation that describes massive spin 1/2 particles in relativistic quantum mechanics, which we discuss in Section 9.2, and which makes use of spacetime spinors and matrix representations of Clifford’s algebra. In 1963, Michael Atiyah and Isadore Singer rediscovered and generalized the Dirac operator to Riemannian manifolds in connection with their celebrated index theorem, which is the topic of Chapter 12. There are also works by Marcel Riesz from 1958 on spacetime isometries and by Lars Ahlfors from 1985 on M¨obius maps, using Clifford algebra, which is the topic of Chapter 4. Mentioned above, David Hestenes has been advocating the use of Clifford algebra, in particular in mathematical physics, since the 1960s. There is also a research field of Clifford analysis, where a higher-dimensional complex analysis using Clifford algebras has been developed, starting from around 1980 and which is the topic of Chapter 8. Included in this book are also some more recent results related to my own research. The material in Sections 9.3 to 10.4 on Dirac integral equations and Hodge decompositions originates with my early thesis work with Alan McIntosh in 2000– 2002, and most of the key ideas there are an inheritance from him. Since then, the material covered in this book has been a continued source of inspiration for my research. The following publications of mine in particular make use, explicitly

Preface

xi

or implicitly, of the algebras of Grassmann and Clifford in real analysis: Axelsson, Keith, and McIntosh [12]; Auscher, Axelsson, and Hofmann [4]; Auscher, Axelsson, and McIntosh [5]; Axelsson, Kou, and Qian [13]; Ros´en [82, 83]; Bandara, McIntosh, and Ros´en [17]; Bandara and Ros´en [18]; and Ros´en [84, 80]. This book was written in four stages. The first part, on the algebras of Grassmann and Clifford, was written around 2008 at Stockholm University and was used as material for a graduate course given there. In the second stage I wrote basically Chapters 7, 8, and 10 for a graduate course given in Link¨oping in 2010. In the third stage I wrote Chapters 11 and 12 for a graduate course in Gothenburg 2014. In between and after these writing periods, the manuscript was collecting dust until I decided, upon returning to mathematics after an extended period of parental leave in 2018, to prepare this final version for publication. Having been away from math for a while gave me new perspectives on things, and this final preparation turned into a major rewriting of the whole book, which I hope will benefit the reader. A number of mathematicians and friends deserve a sincere thanks for being helpful, directly or indirectly, in the creation of this book. Those who untimely have passed away by now, Peetre, McIntosh, and Passare, will always be remembered fondly by me. In mainly chronological order the following people come to mind. Hans Lundmark, who was my mentor for that very first Clifford algebra project in Link¨ oping. I wonder whether and where I would have discovered this mathematics had he not proposed this project to me. Mats Aigner in Link¨oping, whom I first met in Lund and with whom I have had uncountably many interesting discussions about the algebras of Clifford and Grassmann. Jaak Peetre, who encouraged me and provided interesting discussions on the subject. Wulf Staubach at Uppsala University, that friend from the library who changed my life by being well read and knowing about Alan McIntosh. Alan Mcintosh at the Australian National University, my mathematical father from whom I have learned so much. I doubt very much that I will ever again meet someone with as deep an understanding of life and mathematics as he possessed. Mikael Passare at Stockholm University, who supported me at a critical stage. Erik Duse, who was a student attending that first course that I gave in Stockholm, and who more recently himself gave a course based on the third version of this book in Helsinki, and who has given me valuable feedback, including some exercises contained in this book. The book is organized so that the reader finds in the introduction to each chapter a description of and a road map to the material in that chapter. Comments and references are collected in the final section of each chapter. There are two parts of the book. In the first part, the affine multivector and spinor algebra and geometry are explained. A key idea here is the principle of abstract algebra, as explained in the introduction to Chapter 1. In the second part, we use multivectors and spinors in analysis, first in affine space and later on manifolds. A key idea here is that of splittings of function spaces, as explained in the introduction to Chapter 6. My intention is that the material covered should be accessible to basically anyone with mathematical maturity corresponding to that of an advanced undergraduate

Preface

xii

student, with a solid understanding of standard linear algebra, multi-variable and vector calculus, and complex analysis. My hope is that you will find this beautiful mathematics as useful and inspiring as I have.

Andreas Ros´en G¨ oteborg, August 2019

...the horrible “Vector analysis”, which we now see as a complete perversion of Grassmann’s best ideas. (It is limited to 3 dimensions, replaces bivectors by the awful “vector product” and trivectors by the no less awful “mixed product”, notions linked to the euclidean structure and which have no decent algebraic properties!) / J. Dieudonn´e

Chapter 1

Prelude: Linear Algebra Road map: This chapter is not where to start reading this book, which rather is Chapter 2. The material in the present chapter is meant to be used as a reference for some background material and ideas from linear algebra, which are essential to this book, in particular to the first part of it on algebra and geometry consisting of Chapters 2 through 5. The main idea in this part of the book is what may be called the principle of abstract algebra: It is not important what you calculate with, it is only important how you calculate. Let us explain by examples. Consider √ for example the complex numbers x + iy, where you of course ask what is i = −1 when you first encounter this mathematical construction. But that uncomfortable feeling of what this strange imaginary unit really is fades away as you get more experienced and learn that C is a field of numbers that is extremely useful, to say the least. You no longer care what kind of object i is but are satisfied only to know that i2 = −1, which is how you calculate with i. It is this principle of abstract algebra that one needs to bear in mind for all our algebraic constructions in this book: the exterior algebra of multivectors in Chapter 2, Clifford algebras in Chapter 3, and spinors in Chapter 5. In all cases the construction starts by specifying how we want to calculate. Then we prove that there exist objects that obey these rules of calculation, and that any two constructions are isomorphic. Whenever we know the existence and uniqueness up to isomorphism of the objects, we can regard them as geometric objects with an invariant meaning. Which concrete representation of the objects we have becomes irrelevant. In this chapter, Sections 1.1, 1.2, and 1.4 contain background material for Chapter 2, whereas Sections 1.3 and 1.5 are mainly relevant for Chapters 4 and 5 respectively. © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_1

1

Chapter 1. Prelude: Linear Algebra

2

1.1

Vector Spaces

Two general notations which we use throughout this book are the following. By X := Y or Y =: X we mean that X is defined to be / is assigned the value of Y . By A ↔ B we denote a one-to-one correspondence, or an isomorphism between A and B, depending on context. We shall distinguish the concept of a vector space from the more general concept of a linear space. Except for function spaces, which we use later in part two of the book, we shall assume that our linear spaces are finite-dimensional. The difference between linear spaces and vector spaces is only a conceptual one, though. Indeed, any linear space V is naturally an affine space (V, V ), where V acts on itself through the addition in V ; see below. Thus, strictly mathematically speaking, a linear space is the same thing as a vector space. The difference between linear and vector spaces lies in the geometric interpretation of their objects, and we want to make this distinction clear to start with, since we are going to work with linear spaces whose objects are not to be interpreted as geometric vectors. Definition 1.1.1 (Linear space). A real linear space (L, +, ·) is an abelian group (L, +) together with a scalar multiplication R × L → L that is bilinear with respect to addition and a group action of the multiplicative group R∗ = R \ {0} on L. We recall that a group is a set equipped with a binary associative multiplication, containing an identity element and an inverse to each element. For an abelian group, we assume commutativity and write the binary operation as addition. In a linear space, we sometimes write a product xv of x ∈ R and v ∈ L as vx. Since the product of real numbers is commutative, this presents no problem. On the other hand, by a vector space V we mean a linear space consisting of geometric vectors, that is, “one-dimensional directed objects”, which we refer to as vectors. More precisely, this means that V is the space of translations in some affine space X as follows. Definition 1.1.2 (Vector space). An affine space (X, V ) is a set X on which a real linear space V , the space of translations/vectors in X, acts freely and transitively by addition, that is, there exists an addition by vectors map X × V → X that is a (left or right) action of (V, +) on X such that for all x, y ∈ X there exists a unique v ∈ V , the vector denoted by y − x, for which x + v = y. If x, y ∈ X, then the vector v = y − x has the interpretation of a onedimensional arrow starting at x and ending at y. Starting at a different point x0 ∈ X, the same vector v also appears as the arrow from x0 to x0 + v. Thus a vector v is characterized by its orientation and length, but not its position in X. In general affine spaces, the notion of lengths, and more generally k-volumes, have only a relative meaning when we do not have access to an inner product on the space to measure angles and absolute lengths. Thus in general affine spaces, only

1.1. Vector Spaces

3

the relative lengths of two parallel vectors v1 and v2 can be compared: if v1 = λv2 , then v1 is λ times longer than v2 . In practice, one often identifies the affine space X and its vector space V . The difference is the origin 0: X is V , but where we have “forgotten” the origin. Given an origin point x0 ∈ X, we can identify the vector v ∈ V with the point x0 + v ∈ X. In particular, x0 ∈ X is identified with 0 ∈ V . The reader will notice that in Chapters 2 and 7 we carefully distinguish between X and its vector space V , but that in the later chapters, we become more pragmatic and often identify X =V. Definition 1.1.3 (Rn ). The vector space Rn is the set of n-tuples Rn := {(x1 , . . . , xn ) ; xi ∈ R}, with the usual addition and multiplication by scalars. This linear space has a distinguished basis, the standard basis {ei }, where ei := (0, . . . , 0, 1, 0, . . . , 0), with coordinate 1 at the ith position. We adopt the practical convention that we identify row vectors with column vectors, as is often done in doing analysis in Rn . More precisely, Rn should be the space of column vectors, since matrix multiplication is adapted to this convention. However, whenever no matrix multiplication is involved, it is more convenient x1 t to write x1 . . . xn than ... = x1 . . . xn , where ·t denotes matrix xn transpose. We will not distinguish between parentheses (·) and brackets · . Note the decreasing generality of the notions: an affine space is homogeneous and isotropic, that is, without any distinguished points or directions. A linear space is isotropic, but has a distinguished point: the origin 0. The linear space Rn is neither homogeneous or isotropic: it has an origin and a distinguished basis, the standard basis. Whenever we have fixed a basis {ei } in a vectorP space V , there is a xi ei corresponds natural identification between V and Rn , where a vector v = to the coordinate tuple x = (x1 , . . . , xn ) ∈ Rn . Recall the notion of direct sums of linear spaces. Define the sum of subspaces V1 + V2 := {v1 + v2 ; v1 ∈ V1 , v2 ∈ V2 } when V1 and V2 are two subspaces of a linear space V . When V1 ∩ V2 = {0}, we write V1 ⊕ V2 and call the sum a direct sum. This is an intrinsic direct sum. In contrast, suppose that we are given two linear spaces V1 and V2 , without any common embedding space V . In this case we define the (extrinsic) direct sum of these spaces as V1 ⊕ V2 := {(v1 , v2 ) ∈ V1 × V2 ; v1 ∈ V1 , v2 ∈ V2 }.

4

Chapter 1. Prelude: Linear Algebra

In a natural way, V1 ⊕ V2 is a linear space that contains both spaces V1 , V2 , under suitable identifications. As an example, Rn is the exterior direct sum of n copies of the one-dimensional linear space R. Recall the notions of linear independence of a set S ⊂ V and its linear span span(S) ⊂ V . For concrete calculations in a given linear space V , it is often needed to fix a basis {e1 , . . . , en } ⊂ V, with n = dim V being the dimension of V . It is conceptually important to understand that a basis in general is an unordered set. But often bases for vector spaces are linearly ordered e1 , e2 , e3 , . . . by the positive integers and considered as ordered sets. In particular, this is needed in order to represent v ∈ V , x1 n X . xi ei = e1 . . . en .. , v = x1 e1 + · · · + xn en = i=1 xn by its coordinates (x1 , . . . , xn ) ∈ Rn , and in order to represent a linear map T : V1 → V2 between linear spaces V1 , V2 , a1,1 · · · a1,n x1 m X n X 0 . .. . 0 . 0 . . . ei ai,j xj = e1 · · · em . T (x1 e1 +· · ·+xn en ) = . . . , i=1 j=1 xn am,1 · · · am,n by its matrix A = (ai,j ) relative to the bases {ej } for V1 and {e0i } for V2 . However, many fundamental types of bases used in mathematics do not come with any natural linear order. Indeed, this will be the usual situation in this book, where the basic linear spaces of multivectors, tensors, and spinors have standard bases that are not linearly ordered but rather have some sort of lattice ordering, meaning that the basis elements naturally are indexed by subsets of integers or tuples of integers. Another central theme in this book is that many basic linear spaces that appear are not only linear spaces, but associative algebras in the sense that they come equipped with an associative, but in general noncommutative, product. Definition 1.1.4 (Associative algebra). A real associative algebra (A, +, ∗, 1), with identity, is a linear space over R equipped with a bilinear and associative product ∗, with identity element 1. Scalars λ ∈ R are identified with multiples λ1 ∈ A of the identity, and it is assumed that (λ1) ∗ v = λv = v ∗ (λ1) for all v ∈ A. Let (A1 , +1 , ∗1 , 11 ) and (A2 , +2 , ∗2 , 12 ) be two algebras. Then a map T : A1 → A2 is said to be an algebra homomorphism if it is linear and satisfies T (v1 ∗1 v2 ) = T (v1 ) ∗2 T (v2 ) for all v1 , v2 ∈ A1 and if T (11 ) = 12 . An invertible homomorphism is called an algebra isomorphism.

1.2. Duality

5

Exercise 1.1.5. Let A be an associative algebra. Define the exponential function exp(x) :=

∞ X 1 k x , k!

x ∈ A.

k=0

Show that exp(x + y) = exp(x) exp(y), provided that x and y commute, that is, if xy = yx. If φ ∈ R, show that if j 2 = −1, cos φ + j sin φ, exp(φj) = cosh φ + j sinh φ, if j 2 = 1, 1 + φj, if j 2 = 0.

1.2

Duality

There are several reasons for us to consider inner products and dualities more general than Euclidean ones. A first reason is that we want to study the geometry of multivectors in Minkowski spacetimes, the closest relative to Euclidean spaces among inner product spaces, which are modeled by an indefinite inner product as in Section 1.3. A second reason is that we want to study real Clifford algebras where the fundamental representation Theorem 3.4.2 involves inner product spaces of signature zero. A third reason is that we want to study spinor spaces, where more general nonsymmetric dualities may appear. Definition 1.2.1 (Duality and inner product). A duality of two linear spaces V1 and V2 is a bilinear map V1 × V2 → R : (v1 , v2 ) 7→ hv1 , v2 i that is non-degenerate in the sense that hv1 , v2 i = 0 for all v1 ∈ V1 only if v2 = 0, and hv1 , v2 i = 0 for all v2 ∈ V2 only if v1 = 0. In the case V1 = V2 = V , we speak of a duality on V . If a duality on V is symmetric in the sense that hv1 , v2 i = hv2 , v1 i for all v1 , v2 ∈ V , then we call the duality an inner product and V an inner product space. We use the notation hvi2 := hv, vi ∈ R. A vector v such that hvi2 = 0 is called singular. If an inner product has the additional property that hvi2 > 0 for all 0 6= v ∈ V , then we call it a Euclidean inner product, and V is called a Euclidean space. In this case, we define the norm p |v| := hvi2 ≥ 0, so that hvi2 = |v|2 . If a duality on V is skew-symmetric in the sense that hv1 , v2 i = −hv2 , v1 i for all v1 , v2 ∈ V , then we call the duality a symplectic form and V a symplectic space. Note carefully that in general, hvi2 may be negative, as compared to the square of a real number. We do not define any quantity hvi, and the square in the notation hvi2 is only formal.

Chapter 1. Prelude: Linear Algebra

6

Exercise 1.2.2. Show that an inner product is Euclidean if hvi2 ≥ 0 for all v ∈ V . Let V be a linear space. There is a canonical linear space V ∗ and duality hV , V i, namely the dual space of V defined as ∗

V ∗ := {linear functionals θ : V → R}. Given such a scalar-valued linear function θ ∈ V ∗ , its value θ(v) ∈ R at v ∈ V will be denoted by hθ, vi := θ(v) ∈ R. Note that this is indeed a duality: if θ(v) = 0 for all v ∈ V , then θ = 0 by definition. On the other hand, if θ(v) = 0 for all θ, then it follows that v = 0, since otherwise, we can take a complementary subspace V 0 ⊂ V so that V = span{v} ⊕ V 0 and define the linear functional θ(αv + v 0 ) := α, α ∈ R, v 0 ∈ V 0 for which θ(v) 6= 0. If V is a vector space with a geometric interpretation of v ∈ V as in Section 1.1, then θ ∈ V ∗ , which we refer to as a covector, is best described in V by its level sets {v ∈ V ; hθ, vi = C}, for different fixed values of C ∈ R. Since θ is linear, these level sets are parallel hyperplanes. The following observation is fundamental in understanding dualities. Proposition 1.2.3 (Representation of dual space). Fix a linear space V . Then there is a one-to-one correspondence between dualities hV 0 , V i and invertible linear maps g : V 0 → V ∗ : v 7→ θ, given by hg(v 0 ), vi := hv 0 , vi,

v ∈ V.

Here the pairing on the left is the functional value g(v 0 )v, whereas the pairing on the right is as in Definition 1.2.1. If V 0 = V , then V is an inner product/symplectic space if and only if g : V → V ∗ is a symmetric/antisymmetric linear map. With Proposition 1.2.3 in mind, we write a duality between two linear spaces as hV ∗ , V i, where V ∗ not necessarily is the dual space of V , but rather a linear space dual to V in the sense of Definition 1.2.1. By Proposition 1.2.3 this abuse of notation presents no problem. In particular, when we have a duality or inner product on V , we shall write θ=v to mean θ = g(v).

7

1.2. Duality

Definition 1.2.4 (Orthogonal complement). Consider a linear space V and a duality hV ∗ , V i. If hv 0 , vi = 0, then we say that v 0 ∈ V ∗ and v ∈ V are orthogonal. The orthogonal complement of a set S 0 ⊂ V ∗ is the subspace (S 0 )⊥ := {v ∈ V ; hv 0 , vi = 0, for all v 0 ∈ S 0 } ⊂ V. For S ⊂ V we similarly define the orthogonal complement S ⊥ := {v 0 ∈ V ∗ ; hv 0 , vi = 0, for all v ∈ S} ⊂ V ∗ . Definition 1.2.5 (Dual basis).P Let {e1 , . . . , en } be a basis for V . Then each v ∈ V can be uniquely written v = j xj ej , and we define covectors e∗j by he∗j , vi := xj = the jth coordinate of v. We call {e∗1 , . . . , en∗ } ⊂ V ∗ the dual basis of {e1 , . . . , en } ⊂ V . Note that the dual basis {e∗1 , . . . , e∗n } is indeed a basis for V ∗ whenever {e1 , . . . , en } is a basis for V , and is characterized by the property ( 1, i = j, he∗i , ej i = 0, i 6= j. When we have a duality on V , then the dual basis is another basis for V . Exercise 1.2.6. Consider V = R2 , the Euclidean plane with its standard inner product. Find the dual basis to {(3/2, 0), (1/4, 1/2)} and draw the two bases. Example 1.2.7 (Crystal lattices). Let {e1 , e2 , e3 } be the standard basis for R3 . In solid-state physics one studies crystal structures. These have the atoms arranged/packed in a regular pattern that repeats itself, a lattice, which may be different for different crystals. Mathematically a crystal lattice is described by a basis {v1 , v2 , v3 }, which is such that the atoms in the crystal are located at the lattice points {n1 v1 + n2 v2 + n3 v3 ; n1 , n2 , n3 ∈ Z}. Two commonly occurring crystal structures are the body-centered cubic lattice, which has basis { 21 (−e1 + e2 + e3 ), 12 (e1 − e2 + e3 ), 21 (e1 + e2 − e3 )}, and the face-centered cubic lattice, which has basis { 12 (e2 + e3 ), 12 (e1 + e3 ), 21 (e1 + e2 )}. Except for a factor 2, these two bases are seen to be dual bases: one speaks of reciprocal lattices for crystal lattices. The names of these lattices are clear if one draws the basis vectors in relation to the unit cube {0 ≤ x1 , x2 , x3 ≤ 1} and its integer translates.

Chapter 1. Prelude: Linear Algebra

8

Example 1.2.8 (Basis FEM functions). When solving partial differential equations numerically using the finite element method (FEM), the following problem appears. For a three-dimensional computation we consider simplices D, the closed convex hull of four points. Using one corner as the origin 0, and vectors {v1 , v2 , v3 } along the edges to the other three corners, we wish to construct linear functions fk : D → R such that fk (vk ) = 1 and fk = 0 on the opposite face of D, for k = 1, 2, 3. Using the dual basis {v1∗ , v2∗ , v3∗ }, we immediately obtain fk (x) = hvk∗ , xi. For practical calculations in an inner product space, we prefer to use the simplest bases: the ON-bases. Definition 1.2.9 (ON-bases). Let h·, ·i be a duality on V . Then {ei } is called an ON-basis if hei , ej i = 0 when i 6= j and if hei i2 = ±1 for all i. In terms of dual bases, a basis {ei } is an ON-basis if and only if e∗i = ±ei ,

i = 1, . . . , n.

In particular, for a Euclidean space, a basis is an ON-basis if and only if it coincides with its dual basis. Proposition 1.2.10 (Existence of ON-bases). Consider a linear space V with a duality hV, V i. Then V is an inner product space if and only if there exists an ON-basis for V . Proof. Clearly V is an inner product space if an ON-basis exists. Conversely, fix any basis {vi } for V , and define the matrix A = (ai,j ) of hV, V i in this basis by ai,j := hvi , vj i. If V is an inner product space, then A is a symmetric matrix. Using the spectral theorem, we can write D = M ∗ AM , for some invertible matrix diagonal matrix D with ±1 as diagonal elements. The basis {ei } M = (mi,j ) and P defined by ei := j vj mj,i is seen to be an ON-basis. For symplectic spaces, the following is the analogue of ON-bases. Let h·, ·i be k k a duality on V , with dim V = 2k. Then {ei }i=1 is called a Darboux basis ∪ {e0i }i=1 if 0 0 1 ≤ i, j ≤ k, hei , ej i = 0 = hei , ej i, 0 0 i 6= j, 1 ≤ i, j ≤ k, hei , ej i = 0 = hei , ej i, 0 hei , ei i = 1 = −hei , e0i i, 1 ≤ i ≤ k. In terms of dual bases, a basis is clearly a Darboux basis if and only if ei∗ = e0i ,

(ei0 )∗ = −ei ,

for each i = 1, . . . , n.

Exercise 1.2.11 (Existence of Darboux bases). Consider a linear space V with a duality hV, V i. Adapt the proof of Proposition 1.2.10 and prove that V is a symplectic space if and only if there exists a Darboux basis for V . Hint: The spectral theorem for normal complex linear operators applies.

1.3. Inner Products and Spacetime

9

1.3 Inner Products and Spacetime In this section we consider non-Euclidean inner product spaces, and in particular Minkowski spacetimes, the mathematical model for special relativity theory. Definition 1.3.1. Let V be an inner product space. Let n+ be the maximal dimension of a subspace V+ ⊂ V such that hvi2 > 0 for all v ∈ V+ \ {0}, and let n− be the maximal dimension of a subspace V− ⊂ V such that hvi2 < 0 for all v ∈ V− \ {0}. The signature of V is the integer n+ − n− . We say that a subspace V1 ⊂ V is degenerate if there exists 0 6= v1 ∈ V1 such that hv1 , vi = 0 for all v ∈ V1 . Otherwise, V1 is called nondegenerate. If hu, vi = 0 for all u, v ∈ V1 , then V1 is called totally degenerate. Note that a subspace of an inner product space is itself an inner product space if and only if the subspace is nondegenerate. Also, a subspace of an inner product space is totally degenerate if and only if all its vectors are singular, as is seen through polarization, that is, the identity hu + vi2 − hu − vi2 = 4hu, vi. A nonzero singular vector spans a one-dimensional totally degenerate subspace. Proposition 1.3.2 (Sylvester’s law of inertia). Let h·, ·i be an inner product on an n-dimensional vector space V , and let n+ and n− be as in Definition 1.3.1. For every ON basis {ei } for V , the number of basis vectors with hei i2 = 1 equals n+ , and the number of basis vectors with hei i2 = −1 equals n− . If n0 denotes the maximal dimension of a totally degenerate subspace V0 ⊂ V , then n+ + n− = n,

min(n+ , n− ) = n0 .

Proof. Let V+ , V− , and V0 be any Euclidean, anti-Euclidean and totally degenerate subspaces, respectively. Then clearly V+ ∩ V− = V+ ∩ V0 = V− ∩ V0 = {0}, and it follows that n+ + n− ≤ n, n+ + n0 ≤ n, and n− + n0 ≤ n. Fix an ON-basis {ei } for V and choose V± := span{ei ; hei i2 = ±1}. Then dim V+ +dim V− = n and dim V± ≤ n± . It follows that n± = dim V± and n+ +n− = n. From n+ + n− = n, it follows that n0 ≤ min(n − n+ , n − n− ) = min(n− , n+ ) =: m. To see that equality is attained, let V0 := {ei1 − ej1 , . . . , eim − ejm }, where heik i2 = 1 and hejk i2 = −1. Then V0 is seen to be totally degenerate. Exercise 1.3.3. Generalize Proposition 1.3.2 to degenerate bilinear and symmetric forms B(·, ·). Let Rad(V ) := {v ∈ V ; B(v, v 0 ) = 0 for all v 0 ∈ V } be the radical of V , and let n00 := dim Rad(V ). Show that n+ + n− + n00 = n and n0 = n00 + min(n+ , n− ).

10

Chapter 1. Prelude: Linear Algebra

Geometrically, the most important difference between a general inner product space and Euclidean spaces concerns orthogonal complements. For any subspace V1 of a Euclidean space V , we always have a direct sum decomposition V = V1 ⊕ V1⊥ , since V1 ∩ V1⊥ = {0}, because there are no singular vectors. This is not always true in general inner product spaces, but we have the following general result. Proposition 1.3.4 (Orthogonal sums). Let V1 be a k-dimensional subspace in an n-dimensional inner product space V . Then dim V1⊥ = n − k and (V1⊥ )⊥ = V1 , and V1 is a nondegenerate subspace if and only if V1 ∩ V1⊥ = {0}, or equivalently, V = V1 ⊕ V1⊥ . In particular, if V1 is one-dimensional and is spanned by a vector v, then V = span{v} ⊕ span{v}⊥ if and only if v is a nonsingular vector. For the remainder of this section, we study the following non-Euclidean inner product spaces. Definition 1.3.5 (Spacetime). An inner product space (W, h·, ·i) is said to be a Minkowski spacetime, or spacetime for short, with n space dimensions if dim W = 1 + n and the signature is n − 1. We always index spacetime ON-bases as {e0 , e1 , . . . , en }, where he0 i2 = −1. Note that in spacetime coordinates, hx0 e0 + x1 e1 + · · · + xn en i2 = −x20 + x21 + · · · + x2n . To describe the geometry given by such an inner product, we use the following terminology. See Figure 1.1. • The double cone Wl := {v ∈ W ; hvi2 = 0} consisting of all singular vectors v is referred to as the light cone in spacetime. Vectors v ∈ Wl are called light-like. We make a choice and declare one of these two cones to be the future light cone Wl+ , and the other cone Wl− is the past light cone. Thus Wl = Wl+ ∪ Wl− and Wl+ ∩ Wl− = {0}. • We denote the interior of the light cone by Wt := {v ∈ W ; hvi2 < 0}, and it contains the time-like vectors. Since Wt is disconnected, we write it as the disjoint union of the future time-like vectors Wt+ , which is the interior of the future light cone, and the past time-like vectors Wt− , which is the interior of the past light cone. We always assume that e0 ∈ Wt+ , that is, that e0 is a future-pointing time-like vector. • We denote the exterior of the light cone by Ws := {v ∈ W ; hvi2 > 0}, and it contains the space-like vectors. Except when the space dimension is n = 1,

1.3. Inner Products and Spacetime

11

Ws is connected. The whole spacetime thus can be written as the disjoint union W = Wt+ ∪ Wt− ∪ Ws ∪ Wl+ ∪ Wl− , except for the origin. • The analogue of the Euclidean unit sphere is the spacetime unit hyperboloid H(W ) := {v ∈ W ; hvi2 = ±1}. Except for space dimension n = 1, this hyperboloid has three connected components: the future time-like part H(Wt+ ) := H(W ) ∩ Wt+ , the past time-like part H(Wt− ) := H(W ) ∩ Wt− , and the space-like part H(Ws ) := H(W ) ∩ Ws = {v ∈ W ; hvi2 = +1}.

Figure 1.1: The lightcone partition of spacetime, and the straight line representing an inertial observer. Exercise 1.3.6. Let {e0 , e1 , e2 } be an ON-basis for a Minkowski spacetime W . Calculate the dual basis {v1 , v2 , v3 } ⊂ W to {e0 +e1 , e2 , e0 −e1 }. If instead {e0 , e1 , e2 } were an ON-basis for a Euclidean space V , what would this dual basis be?

12

Chapter 1. Prelude: Linear Algebra

A main reason for considering Minkowski spacetime is that it is the mathematical model for Einstein’s special relativity theory, when n = 3. Fix an ON-basis {e0 , e1 , e2 , e3 } with he0 i2 = −1. Once an origin is fixed, points in W are identified with vectors x0 e 0 + x1 e 1 + x2 e 2 + x3 e 3 . The coordinates xi are lengths, and we shall use the meter [m] as the unit of length. We shall write the time coordinate x0 as x0 = ct, where t is time measured in seconds [s] and c = 299792458 [m/s] is the exact speed of light. In relativity theory, the points in spacetime are referred to as events, at time t and position x. The entire life of an observer forms a curve γ(s) ∈ W , s ∈ R, containing all the events that he is present at, at least if he has lived and will live forever. For each s ∈ R, the tangent vector γ 0 (s) ∈ Wt+ will be future-pointing and time-like, since the observer always moves at a speed less than that of light. An observer moving without acceleration is called an inertial observer, and is described p by a straight line in spacetime W spanned by a time-like vector. The quantity −hvi2 /c for a time-like vector v has the meaning of time elapsed as measured by an inertial observer present at two events separated by v in spacetime. We refer to the physics literature for further details on relativity theory. See Section 1.6. In the literature, one often models spacetime as an inner product space with signature 1 − 3, as opposed to the signature convention 3 − 1 used here. An advantage is that the important time-like vectors then have hvi2 > 0. A disadvantage is that in this case, spacetimes are close relatives to the anti-Euclidean space, rather than the Euclidean spaces. Of course, these differences are minor technical ones rather than real geometrical or physical ones. A geometric result about spacetime subspaces that we need is the following. Proposition 1.3.7. Let W be a spacetime and let V ⊂ W be a subspace. Then V is of exactly one of the following types. (i) A space-like subspace. In this case V is nondegenerate and is a Euclidean space, whereas V ⊥ is a spacetime. (ii) A time-like subspace. In this case V is nondegenerate and is a spacetime, whereas V ⊥ is a Euclidean space. (iii) A light-like subspace. In this case V is a degenerate subspace and contains a unique one-dimensional subspace V0 spanned by a light-like vector. The hyperplane V0⊥ in W is the tangent space to the light cone Wl along the line V0 and V0 ⊂ V ⊂ V0⊥ . If V 0 is a complement of V0 in V , so that V = V0 ⊕V 0 , then V 0 is space-like.

1.4. Linear Maps and Tensors

13

0 Proof. Consider first the case that V is nondegenerate, and let n± be the signature indices for V as in Proposition 1.3.2. If n+ = n and n− = 1 are the indices for W , then clearly n0− ≤ n− = 1 and n0+ ≤ n+ . Thus two cases are possible. Either 0 n0− = 0, in which case V is a Euclidean space, or n− = 1, in which case V is a 00 00 = n− , which spacetime. Furthermore, if n± are the indices for V ⊥ , then n0− + n− proves the statement about V ⊥ . On the other hand, if V is a degenerate subspace, write n000 and n00 for the dimensions of the radical and a maximal totally degenerate subspace in V as in Exercise 1.3.3. Then

1 ≤ n000 ≤ n00 ≤ n0 = min(n− , n+ ) = 1. 0 0 0 ) = n00 − n00 = 1 − 1 = 0, and also n0− ≤ n− = 1. We claim Therefore min(n+ , n− 0 0 = 0, that n− = 0. To prove this, assume on the contrary that n0− = 1. Then n+ 0 0 0 so that dim V = n00 + n+ + n− = 1 + 0 + 1 = 2. Let v− ∈ V be a time-like vector, and consider the splitting W = span{v− } ⊕ span{v− }⊥ . If v0 ∈ Rad(V ) \ {0}, then v0 = αv− + v+ , which shows that V contains a space-like vector v+ = v0 − αv− by (ii). This contradicts n0+ = 0. We have proved that 0 0 = dim V − 1. = n00 = 1, n+ n0− = 0, n00

Write V0 := Rad(V ). Then V0 ⊂ V ⊂ V0⊥ . Let t 7→ v(t) ∈ Wl be a curve on the light cone such that v(0) ∈ V0 \ {0}. Then 0 = ∂t hv(t), v(t)i|t=0 = 2hv 0 (0), v(0)i. This shows that the hyperplane V0⊥ must contain the tangent space to Wl along V0 . Since the dimensions are equal, this proves the proposition.

1.4

Linear Maps and Tensors

We denote the set of linear operators between two given linear spaces V1 and V2 by L(V1 ; V2 ) := {T : V1 → V2 ; T is linear}, which itself forms a linear space of dimension dim V1 × dim V2 . For V1 = V2 = V , we write L(V ). The null space of a linear map T is denoted by N(T ), and its range is denoted by R(T ) = T V1 . In this section we discuss a less well known generalization that is essential to this book: the tensor product of linear spaces. Just as a linear operator can be represented by its matrix, a two-dimensional rectangular scheme of numbers, general tensor products can be represented by k-dimensional schemes of numbers. However, we shall restrict ourselves to k = 2 and the relation between operators and tensors. The construction of tensors uses the following maps. Definition 1.4.1 (Multilinearity). A map M : V1 × · · · × Vk → V , where V1 , . . . , Vk and V are linear spaces, is called multilinear, or more precisely k-linear, if for each 1 ≤ j ≤ k, the restricted map Vj 3 vj 7→ M (v1 , . . . , vj , . . . , vk ) ∈ V

Chapter 1. Prelude: Linear Algebra

14

is linear for every fixed vi ∈ Vi , i 6= j. When k = 2, we use the name bilinear. The construction of tensors is very similar to that of multivectors in Section 2.1, but is less geometrically transparent. Following the principle of abstract algebra, we proceed as follows to construct the tensor product V ⊗ V 0 of two given linear spaces V and V 0 . • We first note that there exist a linear space VM and a bilinear map M : V × V 0 → VM such that for two given bases {ei }1≤i≤n and {e0j }1≤j≤n0 for V and V 0 respectively, the set {M (ei , e0j )}1≤i≤n,1≤j≤n0 forms a basis for VM . To see this, just let VM be any linear space of dimension nn0 and define M (ei , ej ) to be some basis for VM . Then extend M to a bilinear map. • We next note that if {M (ei , ej0 )}ij is a basis, then {M (fi , fj0 )}ij is also a basis for VM , for any other choice of bases {fi }i and {fj0 }j for V and V 0 respectively. Indeed, using the bilinearity one checks that {M (fi , fj0 )}ij is a linearly independent set in VM . • If M : V × V 0 → VM maps bases onto bases as above, we note the following. If N : V × V 0 → VN is any other bilinear map, then since {M (ei , e0j )}ij is a basis, setting T (M (ei , ej0 )) := N (ei , ej0 ),

1 ≤ i ≤ n, 1 ≤ j ≤ n0 ,

we have the existence of a unique linear map T : VM → VN such that N = T ◦ M . If M has the property that every other bilinear map factors through it in this way, we say that M has the universal property (U). We shall encounter universal properties for other constructions, so more precisely, this is the universal property for tensor products. Conversely, if a given bilinear map M satisfies (U), then it must map bases onto bases as above. Indeed, take any bilinear map N : V × V 0 → VN such that {N (ei , ej0 )}ij is a basis. We now have a unique linear map T : VM → VN mapping {M (ei , ej0 )}ij onto a basis. This is possible only if {M (ei , e0j )}ij is a basis. Definition 1.4.2 (Tensor product). Let V and V 0 be linear spaces. Fix any bilinear map M : V × V 0 → VM satisfying (U). The tensor product of V and V 0 is the linear space V ⊗ V 0 := VM . We call elements in V ⊗ V 0 tensors and we write u ⊗ v := M (u, v).

1.4. Linear Maps and Tensors

15

Note that if some other bilinear map N : V × V 0 → VN satisfies (U), then the linear map T : VM → VN given by the universal property for M has inverse T −1 : VN → VM given by the universal property for N . Therefore, T provides a unique identification of VM and VN . By the principle of abstract algebra, our definition of V ⊗ V 0 makes sense. If {ei } and {e0j } are bases for V and V 0 , then a general tensor in V ⊗ V 0 is of the form XX αij ei ⊗ e0j , j

i

for some αij ∈ R. Proposition 1.4.3 (Operator=tensor). Let V1 and V2 be linear spaces and consider a duality hV1∗ , V1 i. Then there is a unique invertible linear map V2 ⊗ V1∗ → L(V1 ; V2 ) such that v ⊗ θ 7→ T , where T x := hθ, xiv, x ∈ V1 . Proof. Consider the bilinear map V2 × V1∗ → L(V1 ; V2 ) : (v, θ) 7→ T, where T (x) := hθ, xiv for all x ∈ V1 . According to the universal property for V2 ⊗ V1∗ , there exist a unique linear map V2 ⊗ V1∗ → L(V1 ; V2 ) such that v ⊗ θ 7→ T . Let {ei0 } be a basis for V2 , and let {ej } be a basis for V1 with dual basis {e∗i } for V1∗ . Then we see that the tensor X αij e0i ⊗ ej∗ ij

maps onto the linear operator with matrix {αij }ij . This proves the invertibility. The following shows how this translation between tensors and linear operators works. • If T = v ⊗ θ : V1 → V2 and T 0 = v 0 ⊗ θ0 : V2 → V3 , then the composed operator T 0 ◦ T : V1 → V3 corresponds to the tensor (v 0 ⊗ θ0 )(v ⊗ θ) = hθ0 , viv 0 ⊗ θ. This yields a multiplication of tensors, which is referred to as a contraction. • Let T : V → V be a linear operator on a linear space V . Applying the universal property to the pairing V × V ∗ → R : (v, θ) 7→ hθ, vi,

Chapter 1. Prelude: Linear Algebra

16 we get a canonical linear map

Tr : L(V ) = V ⊗ V ∗ → R. R is called the trace of the operator T . If {ei } The obtained number Tr(T ) ∈ P is a basis for V , then Tr(T ) = i αii if {αij } is the matrix for T . • If V1 and V2 are two linear spaces, then there is a natural swapping map S : V2 ⊗ V1∗ → V1∗ ⊗ V2 : v ⊗ θ 7→ θ ⊗ v, defined using the universal property. Identifying V2 ⊗ V1∗ = L(V1 ; V2 ) and V1∗ ⊗ V2 = L(V2∗ ; V1∗ ), this map S of tensors corresponds to the operation of taking adjoints of linear operators. Recall that the adjoint, or dual, of a linear operator T ∈ L(V1 ; V2 ) is T ∗ ∈ L(V2∗ ; V1∗ ) given by hT ∗ θ, vi = hθ, T vi,

θ ∈ V2∗ , v ∈ V1 .

• Let V be a Euclidean space, and let T = T ∗ be a symmetric operator. By the spectral theorem, there exists an ON-basis {ei } for V in which T has a diagonal matrix. Translated to tensors, this result means that if a tensor w ∈ V ⊗ V is fixed by the above swapping map S, then there is an ON-basis in which X αi ei ⊗ ei , w= i

where we as usual identify V and V ∗ through the inner product. • Let V and V 0 be two Euclidean spaces, and w ∈ V ⊗ V 0 . Then there exist ON-bases {ej } for V and {ej0 } for V 0 , and µj ∈ R such that w=

n X

µj ej

0 ⊗ ej .

j=1

This follows, by translation to tensors, from the spectral theorem and Proposition 1.4.4 for operators, where µj are the singular values of the corresponding operator. Proposition 1.4.4 (Polar decomposition). Let V1 , V2 be Euclidean spaces, and consider an invertible linear map T ∈ L(V1 , V2 ). Then there exists a unique symmetric map S ∈ L(V1 ) such that hSu, ui > 0 for all u ∈ V1 \ {0}, and a unique isometric map U ∈ L(V1 , V2 ) such that T = U S. Similarly, there exists a unique factorization T = S 0 U 0 of T , where S 0 is positive symmetric on V2 and U 0 : V1 → V2 is isometric. We have U 0 = U and S 0 = U SU ∗ .

1.5. Complex Linear Spaces

17

Proof. For such S, U we have T ∗ T = S(U ∗ U )S = S 2 . Thus S = (T ∗ T )1/2 , so S and U are uniquely determined by T . To show existence, define S := (T ∗ T )1/2 and U := T (T ∗ T )−1/2 . Then S is positive, T = U S, and hU x, U yi = hT S −1 x, T S −1 yi = hS 2 S −1 x, S −1 yi = hx, yi. Similarly, U 0 = (T T ∗ )−1/2 T = T (T ∗ T )−1/2 = = (V AV −1 )−1/2 for every positive A and invertible V .

1.5

U , since V A−1/2 V −1

Complex Linear Spaces

The fundamental constructions of exterior algebras and Clifford algebras in this book can be made for linear spaces over more general fields than the real numbers R. We will consider only the field of complex numbers C besides R, which is particularly useful in analysis. We write the complex conjugate of z ∈ C as z c . Given a complex matrix A = (aij )ij , its conjugate transpose is A∗ := (acji )ij , as compared to its transpose At := (aji )ij . Definition 1.5.1. A complex linear space (V, +, ·) is an abelian group (V, +) together with a scalar multiplication C × V → V that is bilinear with respect to the addition operations and a group action of the multiplicative group C∗ = C \ {0} on V. By a complex vector space we shall mean simply a complex linear space, without any interpretation like that in Definition 1.1.2, since this concerns the additive structure of the vector space. Before proceeding with the algebra, an example is in order, to show why complex linear spaces are natural and very useful in analysis. Example 1.5.2 (Time-harmonic oscillations). Consider a quantity f (t, x) that depends on time t ∈ R and position x in some space X. We assume that f takes values in some real linear space. Fixing a basis there, we can assume that f (t, x) ∈ RN . One example is the electromagnetic field in which case N = 6, since it consists of a three-dimensional electric field and a three-dimensional magnetic field. The most convenient way to represent f oscillating at a fixed frequency ω ∈ R is to write f (t, x) = Re(F (x)e−iωt ), for a function F : X → CN , where the real part is taken componentwise. In this way, each component fk (t, x), k = 1, . . . , N , at each point x will oscillate at frequency ω. The complex-valued function F has a very concrete meaning: the absolute value |Fk (x)| is the amplitude of the oscillation of component k at the point x, and the argument arg Fk (x) is the phase of this oscillation. Note that we

Chapter 1. Prelude: Linear Algebra

18

do not assume that the oscillations at different points have the same phase; this happens only for standing waves. Since the complex field has two automorphisms, the identity and complex conjugation, there are two types of dualities that are natural to consider. These correspond to linear and antilinear identification of V 0 and the dual space V ∗ = {θ : V → C ; θ is complex linear} of V. • A complex bilinear duality of two complex linear spaces V 0 and V is a complex bilinear map V 0 × V → C : (v 0 , v) 7→ hv 0 , vi that is nondegenerate. When V 0 = V, we refer to a bilinear duality as a complex bilinear inner product if it is symmetric, that is, if hx, yi = hy, xi. A main difference is that notions like signature are not present in the complex bilinear case since we can normalize −hx, xi = hix, ixi. • A complex sesquilinear duality of V 0 and V, is a nondegenerate pairing (·, ·i such that (v 0 , ·i is complex linear for each v 0 ∈ V 0 and (·, vi is complex antilinear for each v ∈ V . Note the difference in left and right parantheses, which we use to indicate the sesquilinearity. When V 0 = V, we refer to a sesquilinear duality as a complex inner c product if it is symmetric, that is, if (x, yi = (y, xi. A complex inner product is called Hermitian if it is positive definite, that is (u, ui > 0 for all p u ∈ V \{0}. The norm associated with a Hermitian inner product is |u| := (u, ui. The existence of the following types of canonical bases can be derived from the spectral theorem for normal complex linear operators. Proposition 1.5.3 (Complex ON-bases). Let V be a complex linear space. (i) A sesquilinear duality (·, ·i is symmetric if and only if there exists a basis {ei } that is ON in the sense that (ei , ej i = 0 when i 6= j and (ei , ei i = ±1. (ii) A bilinear duality h·, ·i is symmetric in the sense that hv1 , v2 i = hv2 , v1 i if and only if there exists a basis {ei } that is ON in the sense that (ei , ej i = 0 when i 6= j and (ei , ei i = 1. Exercise 1.5.4. (i) Prove that a sesquilinear duality (x, yi is skew-symmetric, that c is, (x, yi = −(y, xi, if and only if i(x, yi is an inner product. (ii) Prove that a bilinear duality h·, ·i is skew-symmetric in the sense that hv1 , v2 i = −hv2 , v1 i if and only if dim V = 2k and there exists a Darboux basis, that is, a basis {ei }ki=1 ∪{e0i }ki=1 in which the only nonzero pairings are he0i , ei i = 1, hei , e0i i = −1, i = 1, . . . , k.

1.5. Complex Linear Spaces

19

We next consider the relation between real and complex linear spaces. We first consider how any complex linear space can be turned into a real linear space, and how to reverse this process. • Let V be a complex linear space. Simply forgetting about the possibility of scalar multiplication by nonreal numbers, V becomes a real linear space, which we denote by V = V. Note that dimC V = 2 dimR V . Besides this real linear structure, V also is equipped with the real linear operator J : V → V : v 7→ iv, which has the property that J 2 = −I. A complex linear map T : V1 → V2 is the same as a real linear map T : V1 → V2 between these spaces regarded as real linear spaces, for which T J1 = J2 T . Given a complex functional θ ∈ V ∗ , the real linear functional V 3 v 7→ Re θ(v) ∈ R belongs to V ∗ . This gives a real linear one-to-one correspondence between V ∗ and V ∗ . In particular, if (·, ·i is a complex inner product on V, taking the real part of the antilinear identification V → V ∗ , we obtain a real inner product hv 0 , viR := Re(v 0 , vi on V , and h·, ·iR is a Euclidean inner product if and only if (·, ·i is a Hermitian inner product. It is possible but less useful to start with a complex bilinear inner product, since this always leads to a real inner product with signature zero. • We can reverse the above argument. Let V be a real linear space equipped with a complex structure, that is, a real linear operator J : V → V such that J 2 = −I. Then (α + βi)v := αv + βJ(v),

v ∈ V, α, β ∈ R,

defines a complex scalar multiplication, which turns V into a complex linear space V. If dim V is odd, then no such J exists, since we would then have (det J)2 = det(−I) = (−1)n = −1, which is unsolvable over R. If dim V is even, there are infinitely many complex structures among which to choose. Indeed, if {e1 , . . . , e2k } is any basis, then J

k k X X (α2j−1 e2j−1 + α2j e2j ) = (−α2j e2j−1 + α2j−1 e2j ) j=1

is one such complex structure.

j=1

Chapter 1. Prelude: Linear Algebra

20

If furthermore the complex structure J on V is an isometry J ∗ J = I, or equivalently skew-adjoint, then polarizing hv 0 , viR = Re(v 0 , vi recovers the sesquilinear duality (v 0 , vi = hv 0 , viR − ihv 0 , JviR . We next consider how any real linear space can be embedded in a complex linear space, and how to reverse this process. • Let V be a real linear space. Define the real linear space V ⊕ V , and consider V as a subspace of V ⊕ V by identifying v ∈ V and (v, 0) ∈ V ⊕ V . Define the standard complex structure J(v1 , v2 ) := (−v2 , v1 ),

(v1 , v2 ) ∈ V ⊕ V.

Then the complex linear space Vc := (V ⊕ V, J) is called the complexification of V . The complex vector (v1 , v2 ) is usually written as the formal sum v1 +iv2 , so that complex scalar multiplication becomes (α + βi)(v1 + iv2 ) = (αv1 − βv2 ) + i(αv2 + βv1 ). The complexification Vc of a real linear space V is a complex linear space V, with dimC Vc = dimR V , which comes with two canonical real linear c subspaces. Defining a complex conjugation operator (x + iy) := x−iy, this is a complex antilinear operation that fixes V ⊂ Vc and squares to the identity. A real linear map T : V → V 0 extends to a complex linear map Tc : Vc → Vc0 by complexification: Tc (v1 + iv2 ) := T v1 + iT v2 . The complexification (V ∗ )c of the real dual can in a natural way be identified with the complex dual (Vc )∗ of the complexification, through the complex linear invertible map given by hθ1 +iθ2 , v1 +iv2 i := hθ1 , v1 i−hθ2 , v2 i+ i(hθ1 , v2 i + hθ2 , v1 i). In particular, if h·, ·i is a duality on V , by complexifying the linear identification V → V ∗ , we obtain a complex bilinear inner product h·, ·iC on Vc , described by Vc 7→ (V ∗ )c = (Vc )∗ . Concretely, hu0 + iv 0 , u + iviC := hu0 , ui − hv 0 , vi + i(hv 0 , ui + hu0 , vi). Alternatively, we may equip Vc with the complex (sesquilinear) inner product (u0 + iv 0 , u + iviC := hu0 , ui + hv 0 , vi + i(−hv 0 , ui + hu0 , vi), which is Hermitian if h·, ·i is Euclidean. We can also complexify a real associative algebra (A, +, ∗, 1), by complexifying the linear space A as well as the bilinear product ∗, to obtain an associative algebra Ac over the complex field.

1.6. Comments and References

21

• We can reverse the above argument. Let V be any complex linear space equipped with a real structure, that is, a complex antilinear operator V → V : z 7→ z c c

such that (z c ) = z. Then V is isomorphic to the complexification Vc of the real subspace V := {z ∈ V ; z c = z} through 1 V 3 z = x + iy ←→ (x, y) = 21 (z + z c ), 2i (z − z c ) ∈ Vc . Clearly, on any complex linear space there are infinitely many real structures. A important advantage over the real theory is that every complex linear operator has an eigenvector, by the fundamental theorem of algebra. For a normal operator, that is, if T ∗ T = T T ∗ on a Hermitian space, we can iterate this result on the orthogonal complement, yielding an ON-basis of eigenvectors. If we apply these results to the complexification of a real linear operator, we obtain the following real result. • Every real linear map T : V → V has either an eigenvector or an invariant two-dimensional subspace. More precisely, in the latter case there exist α, β ∈ R, with β 6= 0, and linearly independent vectors v1 , v2 ∈ V such that T (v1 ) = αv1 − βv2 ,

T (v2 ) = βv1 + αv2 .

• Let T : V → V be a real linear normal operator, that is, T ∗ T = T T ∗ , on a Euclidean space. Then, there exists an ON-basis in which the matrix for T is block diagonal, with 2 × 2 and 1 × 1 blocks along the diagonal. Examples include isometries and skew-symmetric maps.

1.6

Comments and References

1.1 A reference for basic algebraic structures such as groups, rings, fields, vector spaces, and algebras is Nicholson [73]. 1.2 I thank Mats Aigner, Link¨oping University, for suggesting the notation for dualities used in this book, which incorporates the dual space of linear functionals as a special case. 1.3 Spacetime in the sense of Definition 1.3.5 was first constructed by Hermann Minkowski (1864–1909), for Maxwell’s equations. He had Albert Einstein as a student and realized later when Einstein created his special theory of relativity that this could be modeled mathematically by a four-dimensional spacetime. A reference for the theory of relativity is Rindler [79]. The most common sign convention for spacetime in the literature is + − −−, that is, opposite to the sign − + ++ used in this book.

22

Chapter 1. Prelude: Linear Algebra

1.4 Tensors and tensor products appear in the work of J.W. Gibbs (1839–1903), although some specific examples of tensors such as the Cauchy stress tensor and the Riemann curvature tensor had been found earlier. A reference for our construction of tensor products, using the universal property, is Greub [46]. 1.5 We use of the word Hermitian as the complex analogue of Euclidean, with a meaning of positivity. However, in many contexts in the literature, Hermitian refers to the conjugate-symmetry, without any implied positivity. The proof of Proposition 1.5.3(ii) uses a variant of the spectral theorem known as the Autonne–Takagi factorization. An equivalent way to define the complexification Vc of a real linear space V , which is standard but not used in this book, is as the tensor product Vc := V ⊗ C of real linear spaces.

Chapter 2

Exterior Algebra Prerequisites: This chapter is where this book starts, and everything else in the book depends on it, except for Section 2.9, which is not needed elsewhere. Chapter 1 is meant to be used as a reference while reading this and later chapters. Otherwise, a solid background in linear algebra should suffice. Section 2.4 requires a small amount of analysis. Road map: We all know the algebra of vectors, the one-dimensional oriented/directed arrows. Here we construct and develop the algebra for bivectors, the two-dimensional oriented objects, and 3-vectors, the three-dimensional oriented objects, and so on, which live in n-dimensional affine space. In total we obtain a linear space of dimension 2n containing all the multivectors in the space, referred to as the exterior algebra. Algebraically, multivectors are in some sense nothing but rectangular determinants, but it is important to understand the geometry to be able to use the theory. Sections 2.2 and 2.4 aim to convey the geometric meaning of multivectors to the reader. Most applications use Euclidean space, but for a number of practical reasons, including applications to Minkowski spacetime, we allow for more general inner product spaces and dualities. The exterior product u ∧ v can be seen as a higherdimensional generalization of the vector product, but in a more fundamental way, so that it corresponds to the direct sum [u] ⊕ [v] of subspaces [u] and [v]. Since ∧ is noncommutative, two different but closely related dual products come into play, the right and left interior products v x u and u y v, which geometrically correspond to the orthogonal complement [u]⊥ ∩ [v] of subspace [u] in a larger subspace [v]. When the larger space is the whole space, we have the Hodge star map, which corresponds to taking orthogonal complements of subspaces. © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_2

23

Chapter 2. Exterior Algebra

24

Developing the algebra of these products of multivectors, we obtain a geometric birds-eye view on various algebraic results in linear algebra such as identities for the vector product, Cramer’s rule, and the cofactor formula for inverses of linear maps, and expansion rules for determinants. Highlights: • Simple k-vectors ↔ k-dimensional subspaces: 2.2.3 • Factorization algorithm for k-vectors: 2.2.8 • Geometry of Cramer’s rule: 2.3.6 • Algebra for interior product: 2.6.3 • Geometry of cofactor formula: 2.7.1 • Anticommutation relation between exterior and interior products: 2.8.1

2.1

Multivectors

Let us fix an affine space (X, V ) of dimension 1 ≤ n < ∞. The letter n will be the standard notation for the dimension of the vector space V . We set out to construct, for any 0 ≤ k ≤ n, a linear space ∧k V of k-vectors in X. A k-vector w ∈ ∧k V is to be interpreted as an affine k-dimensional object in X determined by its orientation and k-volume. When k = 1, then ∧1 V := V and 1-vectors are simply vectors in X, or oriented 1-volumes. We build k-vectors from vectors using certain multilinear maps. See Definition 1.4.1. Lemma 2.1.1. For a multilinear map M : V × · · · × V → L, the following are equivalent: (i) M (v1 , . . . , vk ) = 0 whenever {v1 , . . . , vk } are linearly dependent. (ii) M (v1 , . . . , vk ) = 0 whenever vi = vj for some i 6= j. (iii) M is alternating, that is, for all 1 ≤ i < j ≤ k and vectors {vm }, we have M (v1 , . . . , vi , . . . , vj , . . . , vk ) = −M (v1 , . . . , vj , . . . , vi , . . . , vk ). Proof. That (i) implies (ii) is clear, as is (iii) implies (ii). P For (ii) implies (i), recall that if {v1 , . . . , vk } are linearly dependent, then vj = i6=j xi vi for some j. Doing this substitution and expanding with multilinearity shows that all terms have two identical factors. This proves (i), using (ii). Finally, to prove (ii) implies (iii), note that 0 = M (v1 , . . . , vi + vj , . . . , vi + vj , . . . , vk ) = M (v1 , . . . , vi , . . . , vj , . . . , vk ) + M (v1 , . . . , vj , . . . , vi , . . . , vk ), from which (iii) follows.

2.1. Multivectors

25

The theory of k-vectors can be thought of as a theory of rectangular determinants. Let us start with a definition of the usual concept of a (quadratic) determinant from linear algebra. Proposition 2.1.2 (Determinant). There exists a unique multilinear map det : Rn × · · · × Rn → R, where the number of copies of Rn is n, with the following properties. (A) If the vectors {v1 , . . . , vn } are linearly dependent, then det(v1 , . . . , vn ) = 0. (B) If {ei } is the standard basis, then det(e1 , . . . , en ) = 1. Let us sketch the proof of this well-known fact. P If det exists, then (A), (B), and multilinearity show that for any vectors vj = i αi,j ei , we must have det(v1 , . . . , vn ) =

n X s1 =1

···

n X

αs1 ,1 · · · αsn ,n (s1 , . . . , sn ),

(2.1)

sn =1

where (s1 , . . . , sn ) is zero if an index is repeated and otherwise denote the sign of the permutation (s1 , . . . , sn ) 7→ (1, . . . , n). Hence uniqueness is clear. Note now that if such det exists, then necessarily it must satisfy (2.1). Thus all that remains is to take (2.1) as the definition and verify properties (A) and (B). Note carefully this frequently useful technique to prove existence, using inspiration from a uniqueness proof. P If vj = i αi,j ei and A = (αi,j ), then we use the standard notation α1,1 .. det(v1 , . . . , vn ) = det(A) = . αn,1

··· .. . ···

α1,n .. . . αn,n

We now generalize this construction to fewer than n vectors, replacing the range R by a more general linear space L. Proposition 2.1.3. Let 2 ≤ k ≤ n and let {e1 , . . . , en } be a basis for V . Then there exist a linear space L and a multilinear map ∧k : V × · · · × V → L, where the number of copies of V is k, that satisfy the following properties. (A) If the {v1 , . . . , vk } are linearly dependent, then ∧k (v1 , . . . , vk ) = 0. (B) The set {∧k (es1 , . . . , esk )}s1 0, [b] = W1 , and T v = exp(b/2)v exp(−b/2) for all v ∈ W . (iii) A parabolic rotation: T has one eigenvector along the light cone. In this case W0 ⊂ W0⊥ ⊂ W , where W0 is a line spanned by a singular vector fixed by T , and W0⊥ is the plane tangent to the light cone along the b 2 W such line W0 , and W0⊥ is invariant under T . In this case there is b ∈ 4 2 ⊥ that b = 0, [b] = W0 , and T v = exp(b/2)v exp(−b/2) for all v ∈ W .

130

Chapter 4. Rotations and M¨obius Maps

Proof. Let T ∈ SO+ (W ) \ {I}. By the Cartan–Dieudonn´e theorem (Theorem 4.1.3), and since T 6= I is a rotation, there are linearly independent vectors v1 , v2 such that T x = v1 v2 x(v1 v2 )−1 . Note that since v1 v2 is orthochronous, we have hv1 , v2 i2 − (v1 ∧ v2 )2 = hv1 , v2 i2 + hv1 ∧ v2 i2 > 0. Three cases are possible: (i) The space [v1 ∧ v2 ] is a space-like plane. In this case, let {e0 , e1 , e2 } be an ON-basis such that {e1 , e2 } is an ON-basis for [v1 ∧ v2 ]. Then there exist α, φ ∈ R such that exp(φe12 ) = cos φ + e12 sin φ = αv1 v2 . This gives conclusion (i) with b := φe12 , W1 := [e0 ], and W2 := [e12 ]. (ii) The space [v1 ∧ v2 ] is a time-like plane. In this case, let {e0 , e1 , e2 } be an ON-basis such that {e0 , e1 } is an ON-basis for [v1 ∧ v2 ]. Then there exist α, φ ∈ R such that exp(φe01 ) = cosh φ + e01 sinh φ = αv1 v2 . This gives conclusion (ii) with b := φe01 , W1 := [e01 ], and W2 := [e2 ]. (iii) The space [v1 ∧ v2 ] is a light-like plane. In this case, let {e0 , e1 , e2 } be an ON-basis such that {e0 + e1 , e2 } is a basis for [v1 ∧ v2 ]. Then there exist α, φ ∈ R such that exp(φ(e0 + e1 )e2 ) = 1 + φ(e0 + e1 )e2 = αv1 v2 . This gives conclusion (iii) with b := φ(e0 + e1 )e2 , W0 := [e0 + e1 ], and W0 = [(e0 + e1 ) ∧ e2 ]. Our next objective is to show that every orthochronous rotation decomposes into two-dimensional elliptic and hyperbolic rotations and three-dimensional parabolic rotations. Proposition 4.4.7 (Spacetime rotations). Let T ∈ SO+ (W ) be an orthochronous rotation in a spacetime W . Then there exist a time-like subspace W−1 of dimension two or three, two-dimensional space-like subspaces W1 , . . . , Wk , and a space-like subspace W0 such that W splits as a direct sum of orthogonal subspaces W = W−1 ⊕ W1 ⊕ · · · ⊕ Wk ⊕ W0 , where all the subspaces are invariant under T , T |W−1 is a Lorentz boost if dim W−1 = 2 or a parabolic rotation if dim W−1 = 3, T |Wj is a Euclidean rotation for j ≥ 1, and T |W0 = I. ˜ := N(T − I). If W ˜ is space- or Proof. We first split off the subspace W0 . Define W ˜ ˜ of ˜ time-like, let W0 := W . If W is light-like, take any space-like subspace W0 ⊂ W ˜ one dimension less than W , which is possible by Proposition 1.3.7. Looking into W0⊥ , replacing W by W0⊥ , we assume from now on that N(T − I) is either {0} or is a line along the light cone. Complexification as in Section 1.5 shows that there exists a one- or twodimensional subspace W 0 that is invariant, and we have the following possible cases:

4.4. Spacetime Rotations

131

1. A space-like plane W 0 , where det T |W 0 = 1. 2. A space-like line W 0 , where T = −I due to the assumption on N(T − I). 3. A time-like plane W 0 , where det T |W 0 = 1. A time-like line is not possible because T |W 0 6= I, by the assumption on N(T − I), and T 6= −I, since T is orthochronous. 4. A light-like plane or line. However, in a plane, any vector along the intersection with the light cone must be an eigenvector, since T is an isometry. Assume, therefore, that T v1 = λv1 , where v1 ∈ Wl . Since T is orthochronous, λ > 0, and we may assume that λ = 1. If this is not the case, since 0 = det(T − λ) = det(T ∗ − λ) = det(T −1 − λ); we see that also λ−1 must be an eigenvalue to T , giving another eigenvector not parallel to v1 . But λ−1 6= 1 can be an eigenvalue of an isometry only if the eigenvector belongs to Wl , which would give us two eigenvectors on Wl , and hence an invariant time-like plane, and we are in the previous case. To summarize, in this last case, we may assume that there exists a line along Wl spanned by v1 such that T v1 = v1 . In all but the last case, the invariant subspace is nondegenerate, and we can split W orthogonally into a direct sum W = W 0 ⊕ (W 0 )⊥ , as in Proposition 1.3.4, with both subspaces invariant under T . In the last case, we need to find a larger invariant and nondegenerate subspace W 0 containing v1 . Let E1 := span{v1 } = N(T − I), and note as in Proposition 1.3.7 that E1⊥ is the tangent hyperplane to the light cone Wl along E1 . We have R(T − I) = N(T ∗ − I)⊥ = N(T −1 − I)⊥ = E1⊥ 3 v1 . Therefore we have v2 such that (T − I)v2 = v1 . Let E2 := span{v1 , v2 }, and note that v2 is not parallel to v1 , since T v1 = v1 . We note that 0 = hv1 i2 = hT v2 − v2 i2 = 2hv2 i2 − 2hT v2 , v2 i = −2hv1 , v2 i, so v2 ∈ E1⊥ . By Proposition 1.3.7, E2 is a degenerate subspace, and we proceed by taking v3 such that (T −I)v3 = v2 , which is possible, since v2 ∈ E1⊥ = R(T −I). Let E3 := span{v1 , v2 , v3 }. We note that T −I : E2 → E1 , so that T ∗ −I = (I −T )T −1 : / E1⊥ . Since / E2⊥ because hv2 i2 6= 0, it follows that T v3 ∈ E1⊥ → E2⊥ . Since v2 ∈ ⊥ ⊥ ⊥ T : E1 → E1 , we have v3 ∈ / E1 . Thus we have constructed an invariant and nondegenerate subspace W 0 = E3 by Proposition 1.3.7. We proceed recursively and split the invariant complements (W 0 )⊥ . Collecting all these orthogonal subspaces produced, we note that the only subspace for which det TW 0 = −1 is that of the space-like lines from 2. Since det T = 1, the number of such lines must be even, and we can evenly collect them in invariant

Chapter 4. Rotations and M¨obius Maps

132

2-planes. Note that these planes are nondegenerate and cannot be time-like, since T is orthochronous. Hence they are space-like, and T acts in them as Euclidean rotation through an angle π. This proves the orthogonal splitting of W . As for Euclidean rotations, this orthogonal decomposition implies the following representation theorem for rotations of spacetime. Theorem 4.4.8 (SO+ surjectivity). Let W be a spacetime. If T ∈ SO+ (W ), then there exist A ∈ SO(W ) and b ∈ Spin(W ) such that T v = exp(A)v = exp(b/2)v exp(−b/2),

v ∈ W.

Thus each T ∈ SO+ (W ) belongs to a one-parameter group of rotations T (φ)v := exp(φA)v = exp(φb/2)v exp(−φb/2), and the exponential map exp : SO(W ) → SO+ (W ) is surjective. Proof. Let T ∈ SO+ (W ), and split spacetime W = W−1 ⊕W1 ⊕· · ·⊕Wk ⊕W0 as in Proposition 4.4.7. By Proposition 4.4.6 there are bj ∈ Spin(Wj ) such that T |Wj v = exp(bj /2)v exp(−bj /2), v ∈ Wj , j = −1, 1, 2, . . . , k. The bivectors commute, so letting b := b−1 + b1 + · · · + bk , we have T v = exp(b−1 /2) exp(b1 /2) · · · exp(bk /2)v exp(−bk /2) · · · exp(−b1 /2) exp(−b−1 /2) = exp(b/2)v exp(−b/2). Also, if Av := b x v, then exp(A) = T .

We end this section with a number of exercises, which completes the extension of the Euclidean results to spacetime. Exercise 4.4.9. Prove the following by modifying the proofs of Proposition 4.4.7, letting N(A) play the role of N(T − I). Let A ∈ SO(W ) be a skew-symmetric map in a spacetime W . Then there exist a time-like subspace W−1 of dimension two or three, two-dimensional spacelike subspaces W1 , . . . , Wk , and a space-like subspace W0 such that W splits as a direct sum of orthogonal subspaces W = W−1 ⊕ W1 ⊕ · · · ⊕ Wk ⊕ W0 , where all the subspaces are invariant under T . If dim W−1 = 2, then the matrix 0 φ of A in an ON-basis {e0 , e1 } is , for some φ ∈ R, that is, Av = φe0,1 x v. If φ 0 dim W−1 = 3, then there is an ON-basis {e0 , e1 , e2 } in which the matrix of A is 0 0 φ 0 0 φ φ −φ 0 for some φ ∈ R, that is, Av = φ(e0 + e1 )e2 x v. The restrictions T |Wj , j ≥ 1, are all Euclidean skew maps, and T |W0 = I.

4.4. Spacetime Rotations

133

Exercise 4.4.10. Deduce the following from the above results for spacetime rotations and skew maps, similar to what was done in the Euclidean case in Proposition 4.3.9. Let W be a spacetime. (i) If b ∈ 42 W is a bivector, then there exists an ON-basis {ei } in which b has the form b = φ0 e0 e1 + φ1 e2 e3 + φ2 e4 e5 + · · · + φk e2k e2k+1 or b = φ0 (e0 + e1 )e2 + φ1 e3 e4 + φ2 e5 e6 + · · · φk e2k+1 e2k+2 . (ii) If q ∈ Spin+ (W ) is an orthochronous rotor, then there exist an ON-basis {ei }, a sign = ±1, and angles 0 < φj ≤ π such that q has the form q = cosh(φ0 /2) + e0,1 sinh(φ0 /2) cos(φ1 /2) + e2,3 sin(φ1 /2) · · · cos(φk /2) + e2k,2k+1 sin(φk /2) , or q = 1 + φ20 (e0 + e1 )e2 cos(φ1 /2) + e3,4 sin(φ1 /2) · · · cos(φk /2) + e2k+1,2k+2 sin(φk /2) . In contrast to Theorem 4.4.8, the exponential map exp : Spin(W ) → Spin+ (W ) is surjective in spacetime only when dim W ≥ 5. In dimensions two and three, exp is far from being surjective, whereas in dimension 4, it is only half of the orthochronous rotors representing parabolic rotations that cannot be represented by a bivector, as the following clarifies. Exercise 4.4.11. Let W be a two-dimensional spacetime, and let j ∈ 42 W be such that j 2 = 1. Show that n o p Spin+ (W ) = α + βj ; α = ± 1 + β 2 n o p % exp(Spin(W )) = α + βj ; α = 1 + β 2 . Exercise 4.4.12. Let W be a three-dimensional spacetime. Then Spin+ (W ) = {α + β1 j1 + β2 j2 + β3 j3 ; α2 + β32 = β12 + β22 + 1} is a connected Lie group by Proposition 4.4.4, where {e0 , e1 , e2 } is an ON-basis with e20 = −1 and j1 := e01 , j2 := e02 , and j3 := e12 . Show that exp(Spin(W )) contains all orthochronous rotors except those α + β1 j1 + β2 j2 + β3 j3 ∈ Spin+ (W ) for which β32 ≤ β12 + β22 and α ≤ −1.

Chapter 4. Rotations and M¨obius Maps

134

Exercise 4.4.13. Let W be a four-dimensional spacetime as in relativity. Show that all q ∈ Spin+ (W ) can be written q = exp(b/2) for some b ∈ Spin(W ), except those q that are of the form q = −(1 + b), where b ∈ 42 W satisfies b2 = 0. Theorem 4.4.14 (Spin+ surjectivity). Let W be a spacetime with dim W ≥ 5. If q ∈ Spin+ (W ), then there exists b ∈ Spin(W ) such that q = exp(b/2). Thus each q ∈ Spin+ (W ) belongs to a one-parameter group of rotors q(φ) := exp(φb/2), and the exponential map exp : Spin(W ) → Spin+ (W ) is surjective. Proof. If q ∈ Spin+ (W ), consider the corresponding rotation T v := qvq −1 . Theorem 4.4.8 shows that T v = exp(b/2)v exp(−b/2), where b = b−1 + b1 + · · · + bk for some commuting simple bivectors bj . Thus q = ± exp(b/2). In case of a minus psign, we can eliminate this as follows. If k ≥ 1, then b2k < 0 and we let φ := 2π/ −b2k . This gives exp((b − φbk )/2) = (−q)(−1) = q, since b and bk commute. If b = b−1 , then since dim W ≥ 5, there exists b1 ∈ 42 W that commutes with b and satisfies b21 = −π 2 . In this case exp((b − b1 )/2) = (−q)(−1) = q also.

4.5

Fractional Linear Maps

In this section we study a fundamental class of maps of a Euclidean space (X, V ) that are closely related to isometries. Definition 4.5.1 (Conformal map). Let (X, V ) be a Euclidean space and let D ⊂ X be an open set. A differentiable map f : D → X is called conformal in D if at each point x ∈ D, the derivative f x : V → V is a nonzero multiple of an isometry, that is, if there exists λ : D → R \ {0} such that λ(x)f x is an isometry for all x ∈ D. Example 4.5.2. We have the following four classes of basic conformal maps in a Euclidean space X. (i) Fix an origin in X, and identify X and V . Then the isometry x 7→ qxq −1 , b . x ∈ X, is conformal for every q ∈ 4V (ii) For every v ∈ V , translation x 7→ x + v, x ∈ X, by the vector v is a conformal map. (iii) Fix an origin in X, and identify X and V . For every c ∈ R \ {0}, dilation x 7→ cx, x ∈ X, with scale factor c is a conformal map. (iv) Fix an origin in X, and identify X and V . Then inversion in the unit sphere x 7→ is a conformal map.

1 x = , x |x|2

x ∈ X \ {0},

4.5. Fractional Linear Maps

135

Exercise 4.5.3. Let f (x) = x−1 . Show that |x|2 f x is reflection in the hyperplane orthogonal to x, and in particular that f is conformal with derivative f x (h) = −x−1 hx−1 ,

h ∈ V, x 6= 0.

Clearly compositions of conformal maps are conformal, on appropriate domains of definition. Thus we have the following group consisting of conformal maps. Definition 4.5.4 (Fractional linear map). Let (X, V ) be a Euclidean space. A map f : X → X is said to be a fractional linear map if it is a finite composition of isometries, translations, dilations, and inversions. In studying fractional linear maps it is convenient to extend the Euclidean space V by adding the point at infinity ∞. More precisely, we embed V in a Euclidean space V∞ that is one dimension larger. A vector e∞ ∈ V∞ with |e∞ | = 1 is fixed and V is identified with the hyperplane orthogonal to e∞ . The extended Euclidean space V is the unit sphere in V∞ , where the north pole e∞ is identified with the point at infinity ∞. The remaining points V \ {e∞ } are identified with V through stereographic projection. Definition 4.5.5 (Stereographic projection). Stereographic projection is the map V \ {e∞ } 3 x = x0 + x∞ e∞ 7→ x ∈ V, for x0 ∈ V and x∞ ∈ [−1, 1), where the projection x is defined as the intersection of V and the straight line in V∞ that intersects V at e∞ and x. See Figure 4.4. We may also refer to the inverse map x 7→ x as stereographic projection. Exercise 4.5.6. Prove the following explicit formulas for the stereographic projection and its inverse. x=

x0 , 1 − x∞

x=

2 |x|2 − 1 x+ 2 e∞ . +1 |x| + 1

|x|2

In particular, the south pole −e∞ projects onto 0 ∈ V and the equator x∞ = 0 projects onto the unit sphere |x| = 1. After adding the point at infinity ∞ to V , the inversion map x 7→ x−1 becomes a smooth map V → V. Indeed, it is seen to correspond to the equatorial reflection x = x0 + x∞ e∞ 7→ x0 − x∞ e∞ = eb∞ xe∞ , where 0 ↔ −e∞ is mapped onto ∞ ↔ e∞ and vice versa. We also extend isometries, translations, and dilation to continuous maps of the extended space V, identified by V through stereographic projection, that fixes the point at infinity e∞ . Thus all fractional linear maps are homeomorphisms as maps V → V.

136

Chapter 4. Rotations and M¨obius Maps

Figure 4.4: Stereographic projection. Proposition 4.5.7 (Mapping of spheres). Let f : V ∪{∞} → V ∪{∞} be a fractional linear map, and let S ⊂ V ∪ {∞} be either a hypersphere, that is, of the form S = {x ∈ V ; |x − a|2 = r2 }, or a hyperplane, that is, of the form S = {x ∈ V ; hx − a, bi = 0} ∪ {∞}. Then the image f (S) is a hypersphere if f −1 (∞) ∈ / S, and f (S) is a hyperplane if f −1 (∞) ∈ S. Proof. It suffices to show that a hypersphere or hyperplane is mapped onto either a hypersphere or a hyperplane, since hyperspheres cannot be unbounded and hyperplanes cannot be bounded. It also suffices to prove this for isometries, translations, dilations, and inversions, and only the last case needs proof. Consider x 7→ x−1 = y and a hypersphere |x − a|2 = r2 . This is mapped onto the points satisfying |1 − ay|2 = r2 |y|2 , where we have used Lagrange’s identity for the Clifford product. This yields 1 + |a|2 |y|2 − 2ha, yi = r2 |y|2 . If |a| = r, that is, ∞−1 = 0 ∈ S, then this is a hyperplane ha, yi = 1/2. If |a| 6= r, then this is the hypersphere |y − a/(|a|2 − r2 )|2 = r2 /(|a|2 − r2 )2 . On the other hand, consider a hyperplane ha, xi = b. This is mapped onto the points satisfying ha, yi = b|y|2 . If b = 0, this is a hyperplane, and if b 6= 0, this is the hypersphere |y − a/(2b)|2 = |a|2 /(4b2 ). This proves the proposition. This result can also be used in the larger space V∞ , where it turns out that the stereographic projection is a restriction of a fractional linear map. Proposition 4.5.8. The map V∞ \ {e∞ } → V∞ \ {e∞ } : y 7→ (e∞ y + 1)(y − e∞ )−1

(4.4)

is a self-inverse fractional linear map. Its restriction to V coincides with the stereographic projection V → V , and its restriction to V coincides with the inverse V → V of the stereographic projection.

4.5. Fractional Linear Maps

137

Under the identification V ∪ {∞} ↔ V, hyperspheres in V correspond to hyperspheres in V not passing through e∞ , and hyperplanes in V correspond to hyperspheres in V passing through e∞ . By a hypersphere in V we mean a nontangential intersection of V and a hyperplane in V∞ . Proof. The map (4.4) can be written y 7→ (e∞ (y − e∞ ) + 2)(y − e∞ )−1 = e∞ + 2(y − e∞ )−1 , from which it is seen that it is a self-inverse fractional linear map. If y = x ∈ V \ {e∞ }, then |x − e∞ |2 = 2 − 2hx, e∞ i = 2(1 − x∞ ), and thus (e∞ x + 1)(x − e∞ ) e∞ − e∞ xe∞ + x − e∞ = 2(1 − x∞ ) 2(1 − x∞ ) 0 x = . 1 − x∞

(e∞ x + 1)(x − e∞ )−1 =

On the other hand, if y = x ∈ V , then |x − e∞ |2 = |x|2 + 1, and thus (e∞ x + 1)(x − e∞ ) |x|2 + 1 2 e∞ |x| + x + x − e∞ = |x|2 + 1 2 |x|2 − 1 = x + e∞ . |x|2 + 1 |x|2 + 1

(e∞ x + 1)(x − e∞ )−1 =

To prove the mapping properties of the stereographic projection, we use Proposition 4.5.7 on the map (4.4). Given a hypersphere S : |x − a|2 = r2 in V , we view this as the intersection of V and the hypersphere S : |y − (a + te∞ )|2 = r2 + t2 passing through e∞ , where t := 12 (|a|2 + 1 − r2 ). Then (4.4) maps S onto a hyperplane not passing through e∞ . Similarly, given a hyperplane S : ha, xi = b in V , we view this as the intersection of V and the hyperplane S : ha + be∞ , yi = b passing through e∞ . Then (4.4) maps S onto a hyperplane passing through e∞ . This proves the proposition. Exercise 4.5.9. Find a fractional linear map of three-dimensional space that maps the unit sphere onto itself and maps (0, 0, 1/2) to the origin. Definition 4.5.10 (M¨obius map). A diffeomorphism f : V → V is said to be a M¨ obius map if it maps hyperspheres in V onto hyperspheres in V. A map f : V → V is called a M¨ obius map if it extends to a M¨obius map of V. We denote the group of M¨ obius maps by M¨ob(V), or equivalently M¨ob(V ). Definition 4.5.11 (Global conformal map). A differentiable map f : V → V, or the corresponding map f : V ∪{∞} → V ∪{∞}, is said to be a global conformal map if f is conformal at each p ∈ V. That f : V ∪ {∞} → V ∪ {∞} is globally conformal

138

Chapter 4. Rotations and M¨obius Maps

means that each point in V has a neighborhood, in which either f (x) or 1/f (x) is a well-defined conformal map, and that 0 ∈ V has a neighborhood where either f (1/x) or 1/f (1/x) is a well-defined conformal map. We here let 1/0 = ∞ and 1/∞ = 0, and well defined at x means that f (x) 6= ∞. Theorem 4.5.12. Assume dim V ≥ 2. For a diffeomorphism f : V → V, the following are equivalent: (i) f is a fractional linear map. (ii) f is a M¨ obius map. (iii) f is a global conformal map. Note that this is a somewhat remarkable result, in that (i) is an algebraic statement, (ii) is a geometric statement, and (iii) is an analytic statement. Proof. We have seen that fractional linear maps are M¨obius maps as well as global conformal maps. To show that all M¨obius maps are conformal, assume that f : V → V maps hyperspheres to hyperspheres. Let a ∈ V. Composing f by a suitable fractional linear map, we may assume that a, f (a) 6= ∞. We have f (a + h) − f (a) → f a (h),

→ 0,

uniformly for |h| = 1. Therefore the linear image f a (S) of the unit sphere S ⊂ V is the uniform limit of spheres, and is there itself a sphere. This is possible only if f a is a multiple of an isometry, proving that f is conformal. To show that all global conformal maps are fractional linear, assume that f : V → V is conformal. For dim V ≥ 3 it follows from Liouville’s theorem (Theorem 11.4.2) on conformal maps that f is a fractional linear map. If dim V = 2, by composing f with a suitable fractional linear map, we may assume that f (∞) = ∞ and that f is orientation preserving. Identifying V and C, we have in this case a bijective entire analytic function f : C → C, and 1/f (1/z) is analytic around z = 0. It is well known from complex analysis that this implies that f (z) = az + b, z ∈ C, for some a, b ∈ C, a 6= 0. Our next objective is to develop the algebra of fractional linear maps. Exercise 4.5.13. Generalizing the algebra of fractional linear maps in complex analysis, we represent a map 4V 3 w 7→ (aw + b)(cw + d)−1 ∈ 4V, a where a, b, c, d ∈ 4V are constants, by the matrix c composition of matrices corresponds to composition of Be careful: the algebra is not commutative!

w ∈ V, b ∈ 4V (2). Show that d the corresponding maps.

4.5. Fractional Linear Maps

139

The following definition gives conditions on a, b, c, d such that (aw + b)(cw + d)−1 preserves the vectors V ⊂ 4V , analogous to the Clifford cone for isometries. In Section 4.6 we will show that this rather ad hoc looking definition is very natural indeed and closely related to the Clifford group for spacetime. Definition 4.5.14 (Vahlen matrices). Let V be a Euclidean space. A matrix M = a b ∈ 4V (2) is called a Vahlen matrix if c d b ∪ {0}, (i) a, b, c, d ∈ 4V (ii) ab, cd ∈ V = 41 V , (iii) ∆(M ) := ad − bc ∈ R \ {0} = 40 V \ {0}. b (2) to be the set of Vahlen matrices in 4V (2). Define the Vahlen cone 4V b (2) form a multiplicative group in 4V (2), Lemma 4.5.15. The Vahlen matrices 4V b and the determinant satisfies ∆(M1 M2 ) = ∆(M1 )∆(M 2 ) for M1 , M2 ∈ 4V (2). If d −b b (2) is a Vahlen matrix, then M −1 = 1 M ∈ 4V is a Vahlen matrix ∆(M ) −c a and ab = ba ∈ V, ab = ba ∈ V,

cd = dc ∈ V, cd = dc ∈ V,

ac = ca ∈ V, ac = ca ∈ V,

bd = db ∈ V, bd = db ∈ V,

ad − bc = da − cb = da − bc = ad − cb ∈ R \ {0}. Thus either a, d ∈ 4ev V and b, c ∈ 4od V , or a, d ∈ 4od V and b, c ∈ 4ev V . Proof. (i) For the inverse of a Vahlen matrix M , the stated matrix is seen to be a right inverse, hence a left inverse also, which shows that ∆(M ) 0 d −b a b da − bc db − bd = = . 0 ∆(M ) c d −c a −ca + ac −cb + ad The diagonal entries show that ∆(M ) = da − bc. Applying the reversion proves the remaining two determinant formulas. Clearly ba = ab ∈ V . To prove ab ∈ V , we may assume that a 6= 0. In this b and a−1 = a/|a|2 = a/(aa). case V 3 a−1 (ba)a = a−1 b(aa) = ab, since a ∈ 4V Repeating this argument, it suffices to prove that ac ∈ V and bd ∈ V . For the first we may assume that a, c 6= 0. Consider the equation ad − bc = λ ∈ R \ {0}. Multiplying by a−1 from the left and c−1 from the right, we get c−1 d − a−1 b = λ(ca)−1 , from which ca ∈ V follows and thus ac ∈ V . The proof of bd ∈ V is similar. b (2) is closed under multiplication. Consider a (ii) Next we show that 4V product a1 b1 a2 b2 a1 a2 + b1 c2 a1 b2 + b1 d2 M1 M2 = = . c1 d1 c2 d2 c1 a2 + d1 c2 c1 b2 + d1 d2

Chapter 4. Rotations and M¨ obius Maps

140

b ∪ {0}, we may assume that a1 , c2 6= 0. In this case To show that a1 a2 + b1 c2 ∈ 4V −1 a1 a2 + b1 c2 = a1 (a2 c−1 2 + a1 b1 )c2 ,

b ∪ {0}. The proofs for the other three entries are where all factors belong to 4V similar. To show that (b2 a1 +d2 b1 )(a1 a2 +b1 c2 ) ∈ V , it suffices to show that b2 a1 b1 c2 + d2 b1 a1 a2 ∈ V , where we may assume that a2 6= 0, for otherwise c2 is parallel to b2 , by the determinant condition on M2 , and the result follows. We need to show that a2 (b2 a1 b1 c2 + d2 b1 a1 a2 )a2 = (a2 b2 )(a1 b1 )(c2 a2 ) + (a2 d2 )(b1 a1 )(a2 a2 ) = (a2 b2 )(2ha1 b1 , c2 a2 i − (c2 a2 )(a1 b1 )) + (a2 d2 )(a2 a2 )(b1 a1 ) = 2ha1 b1 , c2 a2 i(a2 b2 ) + a2 (−b2 c2 + d2 a2 )a2 (a1 b1 ) is a vector, which is clear. A similar calculation shows that (c1 b2 + d1 d2 )(c1 a2 + d1 c2 ) ∈ V . Finally, we calculate the determinant ∆(M1 M2 ) = (a1 a2 + b1 c2 )(b2 c1 + d2 d1 ) − (a1 b2 + b1 d2 )(a2 c1 + c2 d1 ) = a1 (a2 b2 − b2 a2 )c1 + b1 (c2 d2 − d2 c2 )d1 + a1 (a2 d2 − b2 c2 )d1 + b1 (c2 b2 − d2 a2 )c1 = a1 ∆(M2 )d1 − b1 ∆(M2 )c1 = ∆(M1 )∆(M2 ). Theorem 4.5.16 (Representation by Vahlen matrices). Let V be an n-dimensional a b Euclidean space. If M = is a Vahlen matrix, then c d T : V → V : x 7→ (ax + b)(cx + d)−1 is a well-defined fractional linear map. The map M 7→ T is a surjective homob (2) to the group of fractional linear maps. Its morphism from the Vahlen cone 4V kernel consists of the Vahlen matrices 1 0 (−1)n+1 en 0 , α0 , α1 0 en 0 1 where 0 6= α0 ∈ R, 0 6= α1 ∈ R. Proof. To show that T is a fractional linear map, assume first that c 6= 0. Then (ax + b)(cx + d)−1 = ac−1 (cx + d) + (b − ac−1 d) (cx + d)−1 = ac−1 + (b − acd|c|−2 )(x + c−1 d)−1 c−1 = ac−1 + (bc − ad)c|c|−2 (x + c−1 d)−1 c−1 = ac−1 − (∆(M )/|c|2 ) c(x + c−1 d)−1 c−1 ,

4.5. Fractional Linear Maps

141

using that cd = cd = dc, since cd ∈ V . Thus T is a composition of a translation, an inversion, an isometry, a dilation, and finally a translation. If c = 0, then d−1 = |d|−2 d = |d|−2 ∆(M )a−1 , and (ax + b)(cx + d)−1 = (∆(M )/|d|2 )axa−1 + bd−1 , which is an isometry, a dilation, and a translation. In each case we have a fractional linear map. It follows from Exercise 4.5.13 that M 7→ T is a homomorphism. Surjectivity is clear, since the Vahlen matrices qˆ 0 1 v α 0 0 1 , , , , 0 q 0 1 0 1 1 0 represent isometries, translations, dilations, and inversions respectively. To find the kernel, assume that T x = x for all x ∈ V , that is, (ax+b) = x(cx+ d). Letting x = 0 shows that b = 0, and x = ∞ shows that c = 0. The equation ax = xd can hold for all vectors only if a = α0 + α1 en and d = α0 + (−1)n−1 α1 en , as is seen by expressing a and d in an induced ON-basis and using the fact that if s 6= ∅ and s 6= n, then there exist i ∈ s and j ∈ / s such that es has different commutation relations with ei and ej . Checking the conditions ad ∈ R \ {0} and b in the four possible different dimensions modulo 4 shows that either a, d ∈ 4V α0 = 0 or α1 = 0. This proves the theorem. Example 4.5.17. To find a fractional linear map that has certain desired mapping properties in an n-dimensional Euclidean space, one proceeds as in complex analysis, the only difference being that circles and lines are replaced by hyperspheres and hyperplanes. For example, to find the map (e∞ y + 1)(y − e∞ )−1 = e∞ + 2(y − e∞ )−1 from the properties that it should map the sphere V to the hyperplane V in V∞ , one composes the following maps. First the translation x 7→ x − e∞ , which maps e∞ to 0, followed by inversion x 7→ 1/x, which further maps this point to ∞, and also the translated sphere to the hyperplane x∞ = −1/2. Finally, one dilates and translates by x 7→ 2x + e∞ to obtain the image V of the sphere. In total, this gives 2(y − e∞ )−1 + e∞ = 2 + e∞ (y − e∞ ) (y − e∞ )−1 = (e∞ y + 1)(y − e∞ )−1 . Exercise 4.5.18. Extend Exercise 4.5.3 to cover general fractional linear maps. Use the factorization T in the proof of Theorem 4.5.16 to show that every fractional linear map T is conformal, with derivative T x (h) =

∆(M ) (cx + d)h(cx + d)−1 . |cx + d|2

142

Chapter 4. Rotations and M¨obius Maps

The general theory for fractional linear maps of Euclidean spaces can be expressed with complex numbers in the plane and with quaternions in threedimensional space using the standard geometric representations of C and H from Section 3.2. Example 4.5.19 (2D M¨obius maps). Let dim V = 2 and fix an ON-basis {e1 , e2 }. This gives an identification V ∈ x = e1 z ←→ e1 x = z ∈ C = 4ev V between vectors x and complex numbers z. On the one hand, if (ax + b)(cx + d)−1 is a fractional linear map such that a, d ∈ C and b, c ∈ V , then in the complex representation of vectors this corresponds to z 7→ e1 (ae1 z + b)(ce1 z + d)−1 = (w1 z + w2 )(w3 z + w4 )−1 , where w1 := e1 ae1 = a, w2 := e1 b, w3 := ce1 , and w4 := d. The conditions on a, b, c, d translate to wi ∈ C and da − cb = w4 w1 − w3 e1 w2 e1 = w1 w4 − w2 w3 ∈ R \ {0}. On the other hand, if a, d ∈ V and b, c ∈ C, then multiplying the two parentheses by e1 from the right shows that the map is z 7→ e1 (a(e1 ze1 ) + be1 )(c(e1 ze1 ) + de1 )−1 = (w1 z + w2 )(w3 z + w4 )−1 , where w1 := e1 a, w2 := b, w3 := c, and w4 := de1 , and it follows that wi ∈ C and w1 w4 − w2 w3 ∈ R \ {0}. These two cases correspond to orientation-preserving and orientation-reversing M¨obius maps respectively. Example 4.5.20 (3D M¨obius maps). Let dim V = 3 and fix a volume element J ∈ 43 V such that J 2 = −1. This gives an identification V 3 x = −Jz = −zJ ←→ z = Jx = xJ ∈ H ∩ 42 V between vectors and pure quaternions, with the Hodge star map. Let (ax + b)(cx + d)−1 be a fractional linear map. In this case we may assume that a, d ∈ 4ev V and b, c ∈ 4od ∈ V , after possibly multiplying all coefficients by J. In quaternion representation, the fractional linear map is z 7→ J(−aJz + b)(−cJz + d)−1 = (q1 z + q2 )(q3 z + q4 )−1 , where q1 := a, q2 := Jb, q3 := −cJ, and q4 := d. The conditions on a, b, c, d translate to qi ∈ H and q1 q 4 + q2 q 3 ∈ R \ {0}. Orientation-preserving maps arise when q1 q 4 + q2 q 3 > 0, and orientation-reversing maps when q1 q 4 + q2 q 3 < 0.

4.6

Mappings of the Celestial Sphere

In this section we prove that there is a 2−1 homomorphism taking spacetime isometries to Euclidean M¨obius maps. To see the connection with Lorentz isometries,

4.6. Mappings of the Celestial Sphere

143

we identify the higher-dimensional Riemann sphere V with the celestial sphere in a spacetime W . Definition 4.6.1 (Celestial sphere). Given a vector space W , the projective space P (W ) is P (W ) := {[x] ; 0 = 6 x ∈ W }, so that an object in P (W ) is a one-dimensional line through the origin in W . Given a subspace V∞ ⊂ W of codimension one and a vector e0 ∈ W \ V∞ , we identify V∞ and the subset P (V∞ − e0 ) ⊂ P (W ) with the injective map V∞ → P (V∞ − e0 ) : v 7→ [v − e0 ]. The subset P (V∞ ) := {[v] ; 0 6= v ∈ V∞ } is the complement of V∞ in P (W ), and is referred to as the hyperplane at infinity. If W is a spacetime, we assume that e0 ∈ Wt+ and that V∞ = [e0 ]⊥ is a spacelike hyperplane. The celestial sphere is the image P (Wl ) of the light cone Wl , which coincides with the unit sphere V in V∞ under the identification P (V∞ −e0 ) ↔ V∞ . See Figure 4.5. The relativity theory interpretation of this is as follows. Consider an inertial observer O with world line spanned by the future-pointing vector e0 . At the event in spacetime represented by the origin, what O can observe is all light sent out by past events, traveling exactly at the speed of light and reaching the eyes of O at the origin. For this to happen, the past events must lie on the past light cone Wl− . If we write the past event as v − te0 ∈ Wl− , v ∈ V∞ , then v represents the direction from which the light ray reaches O, and t = |v| means that the time it took the light to reach O is proportional to the space distance to the past event. Thus the point [v −te0 ] ∈ P (Wl ) represents all superimposed past events, the light from which O sees in direction v. In this way, the celestial sphere P (Wl ) ⊂ V∞ represents what O observes of the universe W at the origin event. Definition 4.6.2 (Induced map of the sphere). Let W be a spacetime. Fix a futurepointing vector e0 ∈ Wt+ and the Euclidean subspace V∞ := [e0 ]⊥ , and identify the celestial sphere P (Wl ) and the unit sphere V in V∞ . Let T ∈ O(W ) be an isometry. Then T maps the light cone Wl onto itself, and thus induces a diffeomorphism TV : V → V. We define this induced map of the celestial sphere by [TV (v) − e0 ] = [T (v − e0 )],

v ∈ V.

As in Section 4.5, we fix a zenith direction e∞ ∈ V, write V := V∞ ∩ [e∞ ]⊥ , and identify V and V ∪ {∞} under stereographic projection. Write TV for the map in V corresponding to TV . Proposition 4.6.3. If T : W → W is an isometry, then the induced map TV : V → obius map, or equivalently, TV V of the celestial sphere is well defined and is a M¨ is a fractional linear map.

144

Chapter 4. Rotations and M¨obius Maps

Figure 4.5: The celestial sphere and the past spacetime light cone.

Proof. Take any hyperplane P ⊂ V∞ , and let PW be the unique hyperplane in W passing through the origin in W and intersecting V∞ − e0 along P . Then the intersection of PW and Wl corresponds to the hypersphere in V, the intersection of P and V. Since T is linear, it maps PW onto another hyperplane T (PW ) passing through the origin in W and intersecting V∞ − e0 along a hyperplane P 0 . Then by the definition of TV , the intersection P 0 ∩ V is the image TV (P ∩ V). This proves that TV maps hyperspheres to hyperspheres. We wish to obtain an algebraic expression for TV , given the spacetime rotor representing T . Proposition 4.6.4 (M¨obius reflection). Fix an ON-basis {e0 , e1 , . . . , en , e∞ } for spacetime W , and consider the reflection T v = −(a + a∞ e∞ + a0 e0 )v(a + a∞ e∞ + a0 e0 )−1 ,

v∈W

in the hyperplane orthogonal to a + a∞ e∞ + a0 e0 ∈ W , where a ∈ V , V = [e∞ ∧

4.6. Mappings of the Celestial Sphere

145

e0 ]⊥ ⊂ V∞ , and a20 6= |a|2 + a2∞ . Then TV (x) = (−ax + a− )(a+ x + a)−1 ,

x ∈ V,

where a+ := a∞ + a0 and a− := a∞ − a0 . Proof. Consider first the induced map on TV and write a0 := a + a∞ e∞ . Let x ∈ V and consider the point −e0 + x ∈ Wl− . This is mapped by T to − (a0 e0 + a0 )(−e0 + x)(a0 e0 + a0 )−1 = λ(a0 + a0 e0 x − a0 e0 + a0 x)(a0 e0 + a0 ) = λ(a20 e0 + a20 x + a0 a0 + a0 a0 xe0 + a0 a0 + a0 e0 xa0 + |a0 |2 e0 + a0 xa0 ) = λ((a20 + 2a0 ha0 , xi + |a0 |2 )e0 + (a20 x + 2a0 a0 + a0 xa0 )) ∈ Wl , where λ ∈ R \ {0}. By normalizing the e0 coordinate to −1, this means that TV (x) = −

a20 x + 2a0 a0 + a0 xa0 (a0 x + a0 )(a0 x + a0 ) =− 2 0 0 2 a0 + 2a0 ha , xi + |a | |a0 x + a0 |2

= −(a0 x + a0 )(a0 x + a0 )−1 . Using matrix representation as in Exercise 4.5.13, through stereographic projection, in V this corresponds to the map 0 e∞ 1 a a0 e∞ 1 1 −e∞ −a0 −a0 1 −e∞ 0 0 e ae −a e∞ a0 + a0 e∞ − 2a0 −2a 2(a∞ − a0 ) = 0 ∞ ∞ 0 = . a e∞ + e∞ a + 2a0 a0 − e∞ a0 e∞ 2(a∞ + a0 ) 2a This proves that the map in V is x 7→ (−ax+a− )(a+ x+a)−1 , which is a fractional linear map by Theorem 4.5.16. We next consider some applications to special relativity theory. Let {e0 , e1 , . . . , en , e∞ } be an ON-basis for an inertial observer O. This means that the e0 coordinate is the time of an event that O measures and that the e∞ coordinate measures how far in the zenith direction in space the event lies as measured by O, and similarly for the other space coordinates. Assume that O0 is another inertial observer passing by O at the origin event at relativistic speed, having an ON-basis {e00 , e01 , . . . , e0n , e0∞ } relative to which he measures time and space. Denote by T the isometry that maps T (ei ) = e0i , andPlet A = (αi,j ) denote the matrix of T relative to the basis {ei }, that is, T (ei ) = j ej αj,i . Consider an event with coordinates X = {xi } in the basis {ei } as observed by O. The same event observed by O0 has coordinates Y = {yi } in the basis {e0i }, where Y = A−1 X. If O and O0 compare their observations by identifying ei = e0i , then passing from O’s observation to O0 ’s observation defines the experienced map X 7→ A−1 X,

146

Chapter 4. Rotations and M¨obius Maps

in the basis {ei }, that is, T −1 . In particular, the map taking O’s observation of the celestial sphere to O0 ’s observation of the celestial sphere is TV−1 . Compare this result to how one experiences a usual rotation T in threedimensional Euclidean space: if one does not realize that one has rotated by T , it looks as if space has been rotated by T −1 . Sometimes these two points of view are referred to as active and passive transformations. The above result is the analogue for the hyperbolic rotations of spacetime. Example 4.6.5 (Watching a Lorentz boost). Consider four dimensional spacetime in relativity, and an inertial observer O with ON-basis {e0 , e1 , e2 , e∞ }. Consider another observer O’, viewed by O as travelling with speed tanh(φ) > 0 toward the north pole e∞ . As in Example 4.4.1, the Lorentz boost taking O:s ON-basis to O’:s ON-basis is T v = exp(φe0∞ /2)v exp(−φe0∞ /2) −1 = (cosh(φ/2)e∞ + sinh(φ/2)e0 )e∞ ve−1 . ∞ (cosh(φ/2)e∞ + sinh(φ/2)e0 )

Computing the induced fractional linear map on V = span{e1 , e2 } by two applications of Proposition 4.6.4, we get x 7→ x−1 7→ (cosh(φ/2) − sinh(φ/2))/(cosh(φ/2) + sinh(φ/2))x = exp(−φ)x. The experienced M¨obius map of the celestial sphere V, going from O’s to O0 ’s observation, is the inverse of this map corresponding to dilation x 7→ exp(φ)x on V . This means that the faster an observer travels in the direction e∞ , the more he will see the stars move in this direction! See Figure 4.6. Note that the two fixed points of the celestial sphere, ±e∞ , correspond to the two eigenvectors of the Lorentz boost on the light cone. Example 4.6.6 (Relativistic sphere paradox). Consider a Lorentz boost from O to O0 as in Example 4.6.5, and consider a spherical object at rest relative to O0 described by the equation (x0∞ )2 + (x01 − a1 )2 + (x02 − a2 )2 = 1. As the observer O0 passes O, O will see the object passing by at the same speed v as O0 in the direction e∞ . However, according to O’s measurements, it will not be a spherical object but rather an ellipsoid. Indeed, with the Lorentz boost in Example 4.4.1 in the {e0 , e∞ } plane, O will at time t describe the object by the equation p ((x∞ − vt)/ 1 − v 2 )2 + (x1 − a1 )2 + (x2 + a2 )2 = 1, which is an ellipsoid that is shorter in the e∞ direction. However, an amazing phenomenon occurs due to the finite speed of light. Even though O measures the object to be an ellipsoid after taking into account the finite propagation speed of light, the image he sees of the object is a circular shape, just like O0 , although any pattern on the surface of the sphere would be distorted. This is clear from Proposition 4.6.3, since the circular shape O0 sees of the object is mapped by

4.6. Mappings of the Celestial Sphere

147

Figure 4.6: A selection of stars and constellations on the northern celestial sphere, with the Polar Star at (0, 0) and scale 1 corresponding to angle π/4. (a) Our view from Earth showing Cygnus, Cepheus, Cassiopeia, Perseus, Ursa Minor, Auriga, and Ursa Major. (b) The same view for an inertial observer passing Earth at 2/3 the speed of light towards the Polar Star, showing also Aquila, Pegasus, Andromeda, Pisces, Taurus, Orion, Gemini, Leo, Virgo, Bo¨otes, and Hercules. Note that since the constellations have changed by a conformal map, there is little distortion between the views. Note also that at speeds close enough to that of light, even the Southern Cross will move up from behind and be visible next to the Polar Star. a M¨ obius map to another circular shape of the object on O’s celestial sphere! This phenomenon occurs only for spherical objects. Differently shaped objects will become distorted by the M¨obius map. Example 4.6.7 (Watching a parabolic rotation). Consider four-dimensional spacetime in relativity with ON-basis {e0 , e1 , e2 , e∞ } and the parabolic rotation T v = exp(φ(e0 − e∞ )a/2)v exp(−φ(e0 − e∞ )a/2), where a = a1 e1 + a2 e2 is a unit vector in V . The rotor here is q = exp(φ(e0 − e∞ )a/2) = (a+ φ2 (e0 −e∞ ))a. Proposition 4.6.4 applied twice shows that T induces a map of the celestial sphere that corresponds to the fractional linear map x 7→ −axa−1 7→ (−a(−axa−1 ) − φ)a−1 = x − φa of V . Thus a parabolic spacetime rotation translates the celestial sphere, fixing only the north pole e∞ , which corresponds to the eigenvector of the rotation on the light cone. We next turn to a more detailed study of the fractional linear maps appearing in Proposition 4.6.4. Collecting the four coefficients in a matrix, we have a map of

148

Chapter 4. Rotations and M¨obius Maps

spacetime vectors a: C : a + a∞ e∞ + a0 e0 7→

−a a∞ + a0

a∞ − a0 . a

(4.5)

In fact, we already used such a map in the proof of Lemma 3.4.14, so the following should not come as a surprise. Proposition 4.6.8 (Vahlen spacetime algebra). Let W be a spacetime, and fix an orthogonal splitting W = V ⊕ [e∞ ] ⊕ [e0 ] as above. Let 4V (2) denote the algebra of 2 × 2 matrices with coefficients in 4V . Then (4V (2), C), where C is defined in (4.5), is a Clifford algebra for W . Thus there is a unique algebra isomorphism 4W → 4V (2) that identifies a + a∞ e∞ + a0 e0 ∈ W ⊂ 4W and C(a + a∞ e∞ + a0 e0 ) ∈ 4V (2). Proof. The basis vectors map to 0 1 0 −1 C(e∞ ) = , C(e0 ) = , 1 0 1 0

−ej C(ej ) = 0

0 , ej

j = 1, . . . , n.

The Clifford condition (C) is straightforward to verify. To verify (U), which is needed only when n ≡ 3 (mod 4), it suffices by Proposition 3.3.3 to show that the volume element is not scalar. We have (−1)n+1 en 0 C(e0 )C(e∞ )C(e1 ) · · · C(en ) = , 0 en which is not scalar.

To understand the connection between spacetime isometries and fractional linear operators, we need to identify the Clifford cone in 4V (2). We use the following. a b Lemma 4.6.9. Let M = ∈ 4V (2). Then the involution and reversion in c d the Clifford algebra 4V (2) = 4W are # \ " "b b# a b b a −bb a b = , = d b . c d c d bc b −b c db a Proof. Recall that the involution and reversion act by (−1)k and (−1)k(k−1)/2 on 4k W . The corresponding subspace in 4V (2) is spanned by eb 0 0 et 0 −et0 −b et00 0 es = s , e∞ et = , e0 et0 = , e0 e∞ et00 = , 0 es ebt 0 ebt0 0 0 et00 where |s| = k, |t| = |t0 | = k − 1, and |t00 | = k − 2, and the result follows by inspection.

4.6. Mappings of the Celestial Sphere

149

b (2) is isomorProposition 4.6.10 (Vahlen = Clifford cone). The Vahlen cone 4V b phic to the Clifford cone 4W under the isomorphism of Clifford algebras 4V (2) = b (2) can be writ4W determined by (4.5). In particular, every Vahlen matrix in 4V ten as a product of at most dim V + 2 matrices of the form (4.5), with a ∈ V , a0 , a∞ ∈ R, and a02 6= a2∞ + |a|2 . b b (2), note that spacetime vectors belong to 4V b (2) ⊂ 4V Proof. To show that 4W b (2) is closed under multiplication by Lemma 4.5.15. Next consider and that 4V a b b b the converse inclusion 4V (2) ⊂ 4W . Let M = be a Vahlen matrix, and c d thus invertible. By Proposition 4.1.5 it suffices to prove, for all v ∈ V , α, β ∈ R, that # " b a −bb −v α d −b β v −c a c db −b is a spacetime vector, that is, the off-diagonal entries are scalars and the diagonals are ± a vector. By linearity, it suffices to check the cases in which onlyone of v, α, −b ac b aa and β is nonzero. If v = β = 0, α = 1, then the product is , which is a b cc −b ca spacetime vector since b ac = −b ac ∈ V . The case v = α = 0, β = 1 abc = −b ca, since b is similar. For α = β = 0, we get " # −b avd + bbvc b avb − bbva . b b b cvd − dvc −b cvb + dva

To show that the lower left entry is a scalar, we may assume that d 6= 0. In this case, the question is whether b db = (db b − (dd)v(c b b = (dd)((db b b = (dd)((db b d(b cvd − dvc) c)v(dd) d) c)v − v(cd)) c)v + v(db c)) b c, since is scalar. This is clear, since db c ∈ V and ddb ∈ R. Note that cdb = −cdb = −db cdb ∈ V . To show that the upper left entry is a vector, we may assume that b 6= 0, for otherwise, a and d are parallel, and the result follows. We need to show that b(−b avd + bbvc)bb = −(bb a)v(dbb) + (bbb)v(cbb) = (v(bb a) − 2hv, bb ai)(dbb) + v(bbb)cbb = v(bb a)(dbb) + v(bbb)cbb − 2hv, bb ai(dbb) = −vbb(ad − bc)bb − 2hv, bb ai(dbb) is a vector, which is clear. The right entries are shown to be scalars and vectors similarly. Finally, d c b −b avd + bbvc = b avd − bbvc = b cvb − dva. b (2) = 4W b , from which the factorization result for This completes the proof of 4V Vahlen matrices follows by Proposition 4.1.5.

150

Chapter 4. Rotations and M¨obius Maps

We end by summarizing the relation between spacetime isometries and M¨obius maps. Let W be an n-dimensional spacetime, and let V ↔ V ⊂ V∞ ⊂ W represent the celestial sphere as above. Then we have group homomorphisms b q ∈ 4W ρ0 y

pW

−−−−→ O(W ) 3 T ρ1 y

b (2) −−p−V−→ M¨ob(V ) 3 f M ∈ 4V such that ρ1 ◦ pW = pV ◦ ρ0 . • The map ρ0 from Proposition 4.6.10 is an isomorphism. • The map pW as in Proposition 4.1.9 is surjective and has kernel 40 W \ {0}. • The map pV from Theorem 4.5.16 is surjective and has kernel ρ0 (40 W \ {0}) ∪ ρ0 (4n W \ {0}). • The map ρ1 from Proposition 4.6.3 is surjective and has kernel ±1. That ρ1 ◦pW = pV ◦ρ0 is straightforward to verify. We have seen the mapping properties for ρ0 , pW , and pV , and those for ρ1 follow from this. Note that if we b b (2) correspondingly, then pW normalize 4W to Pin(W ), and also normalize 4V and ρ1 are 2 − 1 maps, pV is a 4 − 1 map, and ρ0 is a 1 − 1 map.

4.7

Comments and References

4.1 Our proof of the Cartan–Dieudonn´e theorem (Theorem 4.1.3) is from Grove [49], with a minor simplification using a continuity argument. b is usually referred to as the Clifford What we here call the Clifford cone 4V group in the literature. Our terminology has been chosen to go together with b V . The orthogonal and special orthogonal the notion of the Grassmann cone ∧ groups and their abbreviations are standard, and the name of the spin group is due to its connection to physics. J.-P. Serre introduced the name Pin by removing the S in Spin, in analogy with the orthogonal group. The terminology rotor for an object in the spin group is not used in the literature. 4.2 Example 4.2.6, showing that the exponential map is not surjective for an ultrahyperbolic inner product space, is taken from [78]. 4.4 The source of inspiration for the treatment of rotations in space and spacetime with Clifford algebra is M. Riesz [78], where most of the results are found. 4.5 Our treatment of M¨obius maps with Clifford algebra follows L. V. Ahlfors [1]. Theorem 4.5.12 is from Hertrich–Jeromin [54].

4.7. Comments and References

151

4.6 I want to thank Malcolm Ludvigsen, whom I had as a teacher in a course in cosmology at Link¨oping University as an undergraduate student and who shared with us students many fascinating insights into relativity theory, including the sphere paradox described in Example 4.6.6.

Chapter 5

Spinors in Inner Product Spaces Prerequisites: This chapter builds on Chapters 3 and 4, and uses the material in Sections 1.4 and 1.5. Any knowledge of representation theory is helpful, but the presentation is self-contained and should be accessible to anyone with a solid background in linear algebra. Road map: In a certain sense, one can form a square root p 4V / = 4V of the Clifford algebra of a given inner product space V . Indeed, we have seen in Section 3.4 that 4V is isomorphic to a matrix algebra. For example, the Clifford algebra of spacetime W with three space dimensions is isomorphic to L(R4 ) = R4 ⊗ R4 , √ so in this sense, R4 = 4W . Such spaces are referred to as spinor spaces, and have they deep applications in both physics and mathematics. Two problems with this construction need to be addressed, though. The first is that depending on the dimension and signature of the inner product space, by Theorem 3.4.13 the coefficients in the matrices may belong to R, R2 , C, H, or H2 . However, the standard construction of spinor spaces is over the complex field, and indeed, complex Clifford algebras are always isomorphic to matrix algebras over C or C2 depending on the parity of the dimension, which simplifies matters. The second and more fundamental problem is that isomorphisms 4Vc ↔ L(S) are a priori not unique. To show that a spinor space S has an invariant geometric meaning, we need to show that different choices of matrices used in setting up the isomorphism amount to only a renaming of the elements in S. This © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_5

153

Chapter 5. Spinors in Inner Product Spaces

154

is the purpose of Section 5.2, which, through the principle of abstract algebra, completes the construction of the complex spinor space S = 4V / of an inner product space V . In Section 5.3 we show how to map spinors between different spaces, something that becomes important, for example, in considering spinors on manifolds. / is a fundamental construction for the celebrated Atiyah– The spinor space 4V Singer index theorem for Dirac operators, which we look at in Chapter 12. In physics, spinors are famous for describing a certain intrinsic state for some elementary particles such as electrons in quantum mechanics, which we look at briefly in Section 9.2. Spinor spaces generalize to higher dimensions, a topological feature of the complex square root √ z 7→ z. As we know, the √ complex square root has two possible values differing by sign. If we √ start at 1 = 1 √ and move continuously around the unit circle, then we have e2πi = −1 and e4πi = 1. This means that it takes two full rotations for the square root to return to its original value. In higher dimensions, this is the characteristic behavior of spinor spaces: two full rotations of the physical, or vector, space are needed to return the spinors to their original positions. Beyond the standard linear representation 4V / , there are countably infinitely many nonisomorphic spinor spaces possible for a given inner product space. This is the topic of Section 5.4. Highlights: • The standard representation of spinors: 5.1.5 • The main invariance theorem for spinors: 5.2.3 • Induced spinor maps: 5.3.5 • Finding all abstract spinor spaces in three and four dimensions: 5.4.2, 5.4.8

5.1

Complex Representations

We start with the following general idea of a spinor space. Definition 5.1.1 (Group representation). Let V be a real inner product space, and consider the spin group Spin(V ) ⊂ 4ev V . A complex Spin(V ) representation is a smooth group homomorphism ρ : Spin(V ) → L(S) into the space of linear operators on a complex linear space S. Consider −1 ∈ Spin(V ). If ρ(−1) = −I, then S is called an abstract spinor space for V . If ρ(−1) = I, then S is called an abstract tensor space for V .

5.1. Complex Representations

155

As spaces, abstract tensor and spinor spaces are nothing more than linear spaces in general. The point is that there is a coupled action of rotations in V and its various abstract tensor and spinor spaces. To explain the topological idea, let T (t) ∈ SO(V ), t ∈ [0, 1], be a loop of rotations of V . Lift T (t) to a path q(t) in Spin(V ) such that p(q(t)) = T (t) with the covering map p from Proposition 4.1.9. Assuming that T (0) = T (1) = I and that q(0) = 1, it may happen that q(1) = 1 or that q(1) = −1. See Theorem 4.3.4. Assume that T (t) is such that q(1) = −1, that is, that T (t) is a loop of rotations that is not homotopic to the constant loop. If S is an abstract tensor space for V , then ρ(−q) = ρ(q). This means that ρ is just a representation of SO(V ), and as we rotate V by T (t), the associated linear operators ρ(q(t)) rotate the abstract tensor space S a full loop. If S is instead an abstract spinor space for V , then ρ(−q) = −ρ(q). This means that as we rotate V by T (t), the associated linear operators ρ(q(t)) rotate the abstract tensor space S. But when we have completed one full turn of V at T (1) = I, the space S is not in its initial position, since ρ(−1) = −I. See Figure 5.1. On performing the same rotation of V by T (t) one more time, however, the associated linear operators ρ(q(t)) will return the abstract spinor space S to its original position.

Figure 5.1: The coupled rotation of vector space V (left) and a spinor space S (right), where one full rotation of V corresponds to half a rotation of S. In this concrete example V = R3 and S = 4R / 3 ↔ C2 , as in Definition 5.2.4, where 2 2 only the real subspace R ⊂ C is shown to the right, and the spinor rotation is furnished by the two real Pauli matrices as in Examples 5.1.5 and 5.1.7.

156

Chapter 5. Spinors in Inner Product Spaces

Example 5.1.2. We list some abstract tensor and spinor spaces for a given Euclidean space V that we encounter elsewhere in this book. (i) The exterior powers ∧k V of V are (real) abstract tensor spaces for V , with the action of rotations T being the induced rotations ρ(T )w = T (w) as in Definition 2.3.1. In terms of the rotor q representing T , we have ρ(q)w = qwq −1 as in Proposition 4.1.10. (ii) Similarly, the tensor product V ⊗ V from Definition 1.4.2 is a (real) abstract tensor space for V , with the action of a rotation T being the linear map induced by the bilinear map (v1 , v2 ) 7→ T (v1 ) ⊗ T (vk ) with the universal property of tensor algebras. (iii) The Clifford algebra 4V , as well as the even subalgebra 4ev V , is an abstract spinor space for V , with the action of a rotor q being ρ(q)w = qw. (iv) The spaces Pksh of scalar-valued k-homogeneous harmonic polynomials from Definition 8.2.1 are (real) abstract tensor spaces for V with the action of a rotation T being P (x) 7→ P (T −1 (x)). Note that the rotation invariance of the Laplace equation shows that the rotated polynomial P ◦ T −1 will be harmonic. (v) The spaces Pkem of k-homogeneous monogenic polynomials with values in the even subalgebra from Definition 8.2.1 are (real) abstract spinor spaces for V with the action of a rotor q being P (x) 7→ qP (q −1 xq). Note that Proposition 8.1.14 shows that the so obtained polynomial will be monogenic, and that −1 acts by −I. There are infinitely many nonisomorphic abstract tensor and spinor spaces for a given inner product space V , as we demonstrate in Section 5.4 in low dimension. We will mainly be concerned with one particular example of a spinor space, which builds on the following related notion. Definition 5.1.3 (Vector representation). Let V be a real inner product space. A complex V representation is a real linear map ρ : V → L(S) into the space of complex linear operators on a complex linear space S such that ρ(v)2 = hvi2 I,

v ∈ V.

We note that such V representations are nothing but algebra homomorphisms.

5.1. Complex Representations

157

Lemma 5.1.4 (Vector=algebra representation). Let V be a real inner product space, and consider its real Clifford algebra 4V with complexification 4Vc . Then every complex V representation ρ : V → L(S) extends in a unique way to a homomorphism 4Vc → L(S) of complex algebras. Conversely, every such homomorphism restricts to a complex V representation. Proof. Clearly, for any algebra homomorphism we have ρ(v)2 = ρ(v 2 ) = ρ(hvi2 1) = hvi2 ρ(1) = hvi2 I,

v ∈ V.

Given a complex V representation, it follows from the universal property of Clifford algebras as in Section 3.3 that it extends uniquely to a real algebra homomorphism 4V → L(S). As in Section 1.5, this complexifies in a unique way to a complex algebra homomorphism. A complex V representation is a special case of a complex Spin(V ) representation. Indeed, if we restrict an algebra homomorphism 4Vc → L(S) to the embedded Lie group Spin(V ) ⊂ 4ev V ⊂ 4Vc , we obtain a smooth group homomorphism. Moreover, since ρ(−1) = −ρ(1) = −I for every algebra homomorphism, S will be an abstract spinor space for V . We also note that in general, a complex Spin(V ) representation does not arise in this way from a complex V representation, not even for abstract spinor spaces. Indeed, in general, a complex Spin(V ) representation is not the restriction to Spin(V ) of a linear map on 4ev V . In Example 5.1.2(v) above, this happens only for k = 0. The spinor spaces that we mainly will use are the following. Example 5.1.5 (The standard representation). We set out to construct a complex V representation, for a given Euclidean space V , which will be the most important example for us of a spinor space. The basic idea of this construction is Theorem 3.4.2. (i) Consider first the case that dim V = n = 2m is even. Fix a complex structure J on V , as in Section 1.5, which is isometric. This turns V into a complex linear space V = (V, J). In fact, this is a Hermitian inner product space with complex inner product (·, ·i such that Re(·, ·i is the original Euclidean inner product. Define the complex exterior algebra S := ∧V, which has complex dimension 2m . Generalizing the real theory from Chapter 2 to complex linear spaces, we obtain a complex bilinear exterior product w1 ∧ w2

158

Chapter 5. Spinors in Inner Product Spaces

and a dual complex sesquilinear product w1 y w2 on ∧V. Define a real linear map ρ : V → L(S) by ρ(v)w := v y w + v ∧ w, w ∈ S. Since v y (v y w + v ∧ w) + v ∧ (v y w + v ∧ w) = v y (v ∧ w) + v ∧ (v y w) = (v, viw, by the complex analogue of Theorem 2.8.1, where (v, vi = hv, vi = hvi2 , this ρ is a complex V representation. To make the construction above more concrete, choose a complex ON-basis m for V. This means that this together with the vectors e−k := Jek , k = {ek }k=1 1, . . . , m, forms a real ON-basis for V . Then ρ(ek )ψ = ek y ψ + ek ∧ ψ, ρ(e−k )ψ = i(−ek y ψ + ek ∧ ψ),

ψ ∈ S.

(ii) Consider now the case that dim V = n = 2m + 1 is odd. In this case, we fix a unit vector e0 and consider V 0 := [e0 ]⊥ . Proceeding as in (i) with the evendimensional space V 0 , fixing an isometric complex structure J on V 0 , we obtain a complex V 0 representation on S = ∧V 0 . We extend this to a real linear map of V by defining b ψ ∈ S, ρ(e0 )ψ := ψ, where ψb denotes the complex analogue of the involution from Definition 2.1.18. Note that we define this as a complex linear map. This yields a complex V representation. Indeed, ρ(e0 ) anticommutes with all ρ(v), v ∈ V 0 , since these latter operators swap ∧ev V 0 and ∧od V 0 . Proposition 5.1.6 (Minimal vector representations). Let V be a real inner product space of dimension n = 2m or n = 2m + 1. Then dimC S ≥ 2m for every complex V representation ρ : V → L(S). There exists such a complex V representation with dimC S = 2m . If n is even, such a minimal representation is an isomorphism of complex algebras 4Vc → L(S). If n is odd, then 4ev Vc → L(S) is an algebra isomorphism. Proof. Consider a complex V representation ρ, and let {ek } be an ON-basis for V . As in the proof of Proposition 3.3.3, it follows that {ρ(es )}s⊂n are linearly independent operators when n is even, and that {ρ(es )}|s| is even are linearly independent operators when n is odd. This shows that dimC L(S) ≥ 22m and therefore that dimC S ≥ 2m . It also proves the statements about algebra isomorphisms. The existence of minimal V representations follows from Example 5.1.5 when V is a Euclidean space. For a non-Euclidean inner product space (V, h·, ·i) we write V = V+ ⊕ V− , where V+ and V− are orthogonal subspaces on which the inner

5.1. Complex Representations

159

product is positive and negative definite respectively. Write Ve for V made into a Euclidean space by changing the sign on the inner product on V− , and let ρ be a standard Ve representation as in Example 5.1.5. Then it is straightforward to verify that V = V+ ⊕ V− → L(S) : v+ + v− 7→ ρ(v+ ) + iρ(v− ) is a complex V representation.

We end this section by showing what these matrix representations look like in low-dimensional spaces. Example 5.1.7 (2D and 3D matrices). (i) Let V be a two-dimensional Euclidean space. Fix an ON-basis {e−1 , e1 } and a complex structure J such that Je1 = e−1 . We take e1 as an ON-basis for the one-dimensional Hermitian space V, and for S = ∧V we fix the ON-basis {1, e1 }. In this basis, the basis multivectors act as follows: 0 1 0 −i , ρ(e1 ) = , ρ(e−1 ) = 1 0 i 0 and thus ρ(1) =

1 0 , 0 1

i 0 . 0 −i

ρ(e1 e−1 ) =

(ii) Next add a basis vector e0 and consider three-dimensional Euclidean space with ON-basis {e−1 , e0 , e1 }. With the same spinor space and basis as in two dimensions, we have the action 1 0 . ρ(e0 ) = 0 −1 Note that {ρ(e1 ), ρ(e −1 ),ρ(e0 )} are the Pauli matrices from Example 3.4.19, and i 0 . that ρ(e1 e−1 e0 ) = 0 i (iii) Let W be a spacetime with one space dimension. Fix an ON-basis {e0 , e1 } and write {ie0 , e1 } for the associated Euclidean ON-basis inside Wc . Identifying {ie0 , e1 } and {e−1 , e1 }, we obtain from (i) the representation 0 −1 0 1 , , ρ(e1 ) = ρ(e0 ) = 1 0 1 0 of W . This coincides with the representation we obtained in Example 3.4.18. (iv) Consider next a spacetime with two space dimensions. Adding a Euclidean ON-basis vector e2 , with 1 0 ρ(e2 ) = , 0 −1 we obtain a representation of this spacetime.

160

Chapter 5. Spinors in Inner Product Spaces

Example 5.1.8 (4D and 5D matrices). (i) Let V be a four-dimensional Euclidean space. Fix an ON-basis {e−2 , e−1 , e1 , e2 } and a complex structure J such that Je1 = e−1 and Je2 = e−2 . We take {e1 , e2 } as ON-basis for the two-dimensional Hermitian space V, and for S = ∧V we fix the ON-basis {1, e12 , e1 , e2 }. In this basis, the basis vectors {e1 , e2 , e−1 , e−2 } act by matrices 0 0 0 1 0 0 0 −i 0 0 1 0 0 0 −i 0 0 i , 0 0 −i 0 0 0 0 1, 0 0 −1 0, 0 0 1 0 0 0 0 −1 0 0 i 0 0 i 0 0 0 0 0 1 0 0 0 −i 0 0 0 1 0 0 0 i 0 0 respectively. (ii) Next add a basis vector e0 and consider five-dimensional Euclidean space with ON-basis {e−2 , e−1 , e0 , e1 , e2 }. With the basis for S from (i), we have the action 0 1 0 0 0 1 0 0 ρ(e0 ) = 0 0 −1 0 . 0 0 0 −1 Example 5.1.9 (Dirac’s γ-matrices). Let W be a spacetime with three space dimensions. Fix an ON-basis {e0 , e1 , e2 , e3 } and write {ie0 , e1 , e2 , e3 } for the associated Euclidean ON-basis inside Wc . Identifying {ie0 , e1 , e2 , e3 } and {e1 , e−2 , −e2 , e−1 }, we obtain from Example 5.1.8(i) the representation 0 −i 0 0 0 0 0 −i 0 0 −i , ρ(e1 ) = 0 0 −i 0 , 0 ρ(e0 ) = 0 i 0 0 0 0 −i 0 0 −i 0 0 i 0 0 0 0 0 −i 0 0 0 0 −1 0 0 0 0 1 0 0 i ρ(e2 ) = 0 1 0 0 , ρ(e3 ) = i 0 0 0 0 −i 0 0 −1 0 0 0 for W . The relation to standard representations used in quantum mechanics is the f as an inner product space of signature following. Consider instead spacetime W 1 − 3. Here the Dirac gamma matrices 0 σ3 I 0 σ1 σ2 0 0 3 0 1 2 , , γ = , γ = γ = , γ = 0 −I −σ2 0 −σ3 0 −σ1 0 f , where {σ1 , σ2 , σ3 } are the Pauli matrices from Exrepresent an ON-basis for W ample 3.4.19. Another important matrix is the so-called fifth gamma matrix 0 I 5 0 1 2 3 . γ := iγ γ γ γ = I 0

5.2. The Complex Spinor Space

161

f. In quantum mechanics this represents chirality, since γ 5 ∈ 44 W However, matrix representations of the ON-basis vectors of W are quite arf, bitrary. Another representation is the Weyl representation of an ON-basis for W namely {γ 5 , γ 1 , γ 2 , γ 3 }. f representation, since (γ 5 )2 = I and γ 5 anticomThis gives another complex W 2 3 1 mutes with γ , γ , γ . Comparing to our basis {e0 , e1 , e2 , e3 } for spacetime with signature 3 − 1, we see that e0 = −iγ 5 ,

e1 = −iγ 1 ,

e2 = −iγ 2 ,

e3 = −iγ 3 .

5.2 The Complex Spinor Space To formulate the main theorem for spinor spaces, we need the following terminology. Definition 5.2.1 (Main reflectors). Let V be a real inner product space of dimension n, and consider its complexified Clifford algebra 4Vc . The main reflectors in 4Vc are the two n-vectors wn ∈ 4n Vc satisfying wn2 = 1. For a given choice of main reflector wn , define wn+ := 21 (1 + wn ) and wn− := 12 (1 − wn ). Note that if we fix 0 6= w0 ∈ 4n V , then any w ∈ 4n Vc can be written w = λw0 for some λ 6= 0. The equation 1 = λ2 w02 has exactly two solutions over the complex field, which yield the two main reflectors. Lemma 5.2.2. Fix a main reflector wn ∈ 4n Vc in a real inner product space V , and let ρ : V → L(S) be a complex V representation. Then ρ(wn+ ) and ρ(wn− ) are complementary projections in the sense that ρ(wn± )2 = ρ(wn± ) and ρ(wn+ )+ρ(wn− ) = I. If dim V is even, then the dimensions of the ranges of the two projections ρ(wn± ) are equal. If dim V = 2m + 1 is odd and if dimC S = 2m , then either ρ(wn+ ) = 0 or ρ(wn− ) = 0. Equivalently, ρ(wn ) = I or ρ(wn ) = −I. Proof. That ρ(wn± ) are complementary projections is a consequence of ρ being an algebra homomorphism. In even dimension, we fix a unit vector v ∈ V and note that ρ(v)2 = I and therefore ρ(v) is self-inverse. Since vw = −wv, we deduce that ρ(wn− ) = ρ(v)ρ(wn+ )ρ(v)−1 , from which it follows in particular that ρ(wn± ) projects onto subspaces of S of equal dimension. In odd dimension, it follows as in the proof of Proposition 3.3.3 for a minimal complex V representation that ρ(wn ) and ρ(1) = I must be linearly dependent. Since ρ(wn )2 = I, we have in fact ρ(wn ) = ±I.

162

Chapter 5. Spinors in Inner Product Spaces

The main result needed for the geometric construction of spinors is the following uniqueness result. This is similar to the universal property for the exterior algebra from Proposition 2.1.4, and the universal property for Clifford algebras from Definition 3.3.1, in that it provides a useful way to construct mappings of spinors. See Section 5.3. Theorem 5.2.3 (Uniqueness of minimal representations). Let V be a real inner product space of dimension n = 2m or n = 2m + 1. Assume that ρj : V → L(Sj ) are two complex V representations, both with dimension dimC Sj = 2m , j = 1, 2. (i) If n = 2m, then there exists an invertible linear map T : S1 → S2 such that w ∈ 4Vc .

ρ2 (w)T = T ρ1 (w),

If Te is a second such map, then Te = λT for some λ ∈ C \ {0}. (ii) If n = 2m + 1, then there exists an invertible linear map T : S1 → S2 such that either ρ2 (w)T = T ρ1 (w), ρ2 (w)T = T ρ1 (w), b

w ∈ 4Vc , w ∈ 4Vc .

or

The first case occurs when ρ1 (wn ) = ρ2 (wn ) for the main reflectors wn , and the second case occurs when ρ1 (wn ) = −ρ2 (wn ). In particular, ρ2 (w)T = T ρ1 (w) for all w ∈ 4ev Vc . If T ∈ L(S1 ; S2 ) is a second map with this property, then Te = λT for some λ ∈ C \ {0}. This section is devoted to the proof of Theorem 5.2.3, but before embarking on this, we make use of this result to give a proper invariant geometric definition of spinors and the complex spinor space of V . To this end, recall the principle of abstract algebra, as discussed in the introduction of Chapter 1. At this stage the reader hopefully is so comfortable using multivectors that he or she has forgotten that we never actually defined a k-vector to be a specific object, but rather as an element in the range space of some arbitrarily chosen multilinear map with properties (A) and (U) in Definition 2.1.6. Proposition 5.1.6 and Theorem 5.2.3 now allow us to define spinors in a similar spirit, as elements in some arbitrarily chosen representation space S of minimal dimension 2m . Indeed, Theorem 5.2.3 shows that any other choice of ρ and S amounts only to a renaming of the objects in the space S, and does not affect how 4Vc acts on S. Definition 5.2.4 (The complex spinor space). Let (X, V ) be a real inner product space of dimension n = 2m or n = 2m + 1. Fix one complex V representation / , and refer to (4V, / ρ) ρ : V → L(S) with dimC S = 2m . We denote this S by 4V as the complex spinor space for V . We also use the shorthand notation w.ψ := ρ(w)ψ

163

5.2. The Complex Spinor Space

for the action of w ∈ 4Vc on ψ ∈ 4V / . + Fix a main reflector wn ∈ 4n Vc . When dim V is even, we denote by 4 / V − and 4 / V the ranges of ρ(wn± ), so that +

−

/ =4 / V ⊕4 / V. 4V When dim V is odd, we write ρ− (w)ψ := w.ψ ˆ for the second nonisomorphic action / . of 4Vc on 4V

Figure 5.2: Multivector action on spinors in a four-dimensional inner product space. We have constructed the complex spinor space 4V / for V in a way very similar in spirit to the construction of the exterior algebra ∧V for V . There is one big difference between the constructions of multivectors and spinors, though: with the map ∧k , a set of k vectors single out a certain k-vector, but vectors are not factors in spinors in this way. Rather vectors, and more generally multivectors, act as linear operators on spinors. +

−

Exercise 5.2.5. Assume that dim V is even. Show that ρ(v) maps 4 / V →4 / V − + and 4 / V , for every vector v ∈ 41 V = V . Using the standard represen/ V →4 + − tation from Example 5.1.5(i), show that 4 / V and 4 / V coincide with ∧ev V and od ∧ V.

164

Chapter 5. Spinors in Inner Product Spaces

We now turn to the proof of Theorem 5.2.3. The idea is roughly that two minimal complex V representations yield an isomorphism L(S1 ) → L(S2 ) of complex algebras. Given an invertible map T ∈ L(S1 ; S2 ), L(S1 ) → L(S2 ) : X → T XT −1 is such an algebra isomorphism. We prove below that every isomorphism L(S1 ) → L(S2 ) arises in this way, and we deduce from this fact Theorem 5.2.3. For this proof, it does not matter whether we work over the real or complex field. Definition 5.2.6 (Ideals). Let A be a complex associative algebra, as in Definition 1.1.4 but replacing R by C. A linear subspace I ⊂ A is called a left ideal if xy ∈ I whenever x ∈ A and y ∈ I. If yx ∈ I whenever x ∈ A and y ∈ I, then I is called a right ideal. A linear subspace I that is both a left and right ideal is called a two-sided ideal. The notion of ideal is important in identifying the T representing a given algebra isomorphism as above. We also recall from Proposition 1.4.3 that there is a natural isomorphism S ⊗ S ∗ ↔ L(S), which identifies the simple tensor v ⊗ θ ∈ S ⊗ S ∗ and the rank-one linear operator x 7→ hθ, xiv. We use both these views on linear operators below. Proposition 5.2.7 (Matrix ideals). Consider the algebra L(S) of all linear operators on a linear space S. There is a one-to-one correspondence between linear subspaces of S ∗ and left ideals in L(S) that identifies a subspace U ⊂ S ∗ and the left ideal o nX vk ⊗ θk ; vk ∈ S, θk ∈ U . IUl := {T ∈ L(S) ; R(T ∗ ) ⊂ U } = k

Similarly there is a one-to-one correspondence between linear subspaces of S and right ideals in L(S) that identifies a subspace U ⊂ S and the right ideal nX o IUr := {T ∈ L(S) ; R(T ) ⊂ U } = vk ⊗ θk ; vk ∈ U, θk ∈ S ∗ . k

The only two-sided ideals in L(S) are the two trivial ones, namely {0} and L(S) itself. Note that in terms of matrices, a left ideal consists of all matrices with all row vectors in a given subspace, whereas a right ideal consists of all matrices with all column vectors in a given subspace.

5.2. The Complex Spinor Space

165

Proof. Clearly IUl is a left ideal for any subspace U ⊂ S ∗ . Note that R(T ∗ ) ⊂ U is equivalent to U ⊥ ⊂ N(T ). To see that all left ideals are of the form IUl , consider any left ideal I ⊂ L(S). Let U ⊂ S ∗ be such that \ N(T ). U ⊥ := T ∈I

Then I ⊂ IUl . For the converse we claim that there exists T0 ∈ I such that N(T0 ) = U ⊥ . From this it follows that IUl ⊂ I. Indeed, if N(T0 ) ⊂ N(T ), then T = AT0 ∈ I for some operator A. One way to prove the claim is to observe that N(T1 ) ∩ N(T2 ) = N(T1∗ T1 + T2∗ T2 ), and that T1∗ T1 + T2∗ T2 ∈ I whenever T1 , T2 ∈ I. Here the adjoints are with respect to any auxiliary Euclidean inner product. Using this observation a finite number of times, it follows that there exists T0 ∈ I with minimal null space. The proof for right ideals is similar. Finally, if I is a two-sided ideal, then I = IUl = IVr for some subspaces U ⊂ S ∗ and V ⊂ S. This can happen only if U = {0} = V or if U = S ∗ and V = S. Thus I = {0} or I = L(S). The following is the key result in the proof of uniqueness of minimal spinor representations. Proposition 5.2.8 (Completeness of matrix algebras). The algebra L(S) of all linear operators on a linear space S is complete in the sense that every algebra automorphism φ : L(S) → L(S) is inner, that is, there exists a linear invertible map T ∈ L(S) such that φ(X) = T XT −1 for all X ∈ L(S). Such T are unique up to scalar multiples. Proof. By Proposition 5.2.7, every minimal left ideal, that is, a left ideal I such that no left ideal J such that {0} $ J $ I exists, must be of the form Lθ := {v ⊗ θ ; v ∈ S},

for some θ ∈ S ∗ \ {0}.

Similarly, all minimal right ideals are of the form Rv := {v ⊗ θ ; θ ∈ S ∗ },

for some v ∈ S \ {0}.

Let φ : L(S) → L(S) be an algebra automorphism, and fix θ0 ∈ S ∗ and v 0 ∈ S such that hθ0 , v 0 i = 1. Since φ only relabels the objects in L(S) without changing the algebraic structure, it is clear that φ(Lθ0 ) = Lθ00 for some θ00 ∈ S ∗ \ {0}. It follows that there is an invertible linear operator T1 ∈ L(S) such that φ(v ⊗ θ0 ) = T1 (v) ⊗ θ00 ,

for all v ∈ S.

(5.1)

Similarly, by considering the mapping of minimal right ideals, there exist an invertible linear operator T2 ∈ L(S ∗ ) and v 00 ∈ S such that φ(v 0 ⊗ θ) = v 00 ⊗ T2 (θ),

for all θ ∈ S ∗ .

166

Chapter 5. Spinors in Inner Product Spaces

Since φ is an automorphism, it follows that φ(v ⊗ θ) = φ((v ⊗ θ0 )(v 0 ⊗ θ)) = (T1 (v) ⊗ θ00 )(v 00 ⊗ T2 (θ)) = hθ00 , v 00 iT1 (v) ⊗ T2 (θ), for all v ∈ S, θ ∈ S ∗ . For example, by inversely rescaling T1 and θ00 , we may assume that hθ00 , v 00 i = 1. Furthermore, for all v1 , v2 ∈ S and θ1 , θ2 ∈ S ∗ we have hθ1 , v2 iφ(v1 ⊗ θ2 ) = φ((v1 ⊗ θ1 )(v2 ⊗ θ2 )) = (T1 (v1 ) ⊗ T2 (θ1 ))(T1 (v2 ) ⊗ T2 (θ2 )) = hT2 (θ1 ), T1 (v2 )iT1 (v1 ) ⊗ T2 (θ2 ), so (T2 )−1 = (T1 )∗ . Thus, with T = T1 , we get that φ(v ⊗ θ) = T (v) ⊗ (T ∗ )−1 (θ), and therefore φ(X) = T XT −1 by linearity. The uniqueness result is a consequence of the fact that Z(L(S)) = span{I}. Indeed, if TeX Te−1 = T XT −1 for all X, then T −1 Te ∈ Z(L(S)). Note that formula (5.1) can be used to calculate T for a given automorphism ∗ φ. Fixing a basis for S and dual00 basis for S , we work with matrices and may 0 assume θ = 1 0 . . . 0 and θ = a1 a2 . . . ak . Then we have for vectors v ∈ S the matrix identity φ v 0 . . . 0 = a1 T (v) a2 T (v) . . . ak T (v) . Since at least one aj is nonzero, we find that T (v) can be defined as a nonzero column of the matrix φ(v ⊗ θ0 ). Proof of Theorem 5.2.3. (i) Assume dim V = 2m. Consider two representations ρ1 and ρ2 with dim S1 = 2m = dim S2 , which by Proposition 5.1.6 are algebra isomorphisms 4Vc → L(Si ). Then ρ2 ρ−1 1 : L(S1 ) → L(S2 ) is an algebra isomorphism. Take any linear invertible map T0 : S1 → S2 and consider the induced −1 algebra isomorphism ρ0 : L(S1 ) → L(S2 ) : X 7→ T0 XT0−1 . Then ρ−1 is an 0 ρ2 ρ 1 automorphism of L(S1 ), and Proposition 5.2.8 shows the existence of T1 ∈ L(S1 ), unique up to scalar multiples, such that −1 −1 ρ−1 0 ρ2 ρ1 (X) = T1 XT1 ,

for all X ∈ L(S1 ).

Letting T := T0 T1 : S1 → S2 , this means that ρ2 (w) = T ρ1 (w)T −1 ,

for all w ∈ 4Vc .

(ii) Assume dim V = 2m + 1. Consider two representations ρ1 and ρ2 with dim S1 = 2m = dim S2 . Then ρi : 4ev V → L(S) are both isomorphisms, so as in (i), we get from the algebra isomorphism (ρ2 |4ev V )(ρ1 |4ev V )−1 : L(S1 ) → L(S2 ) the existence of T ∈ L(S1 ; S2 ) such that ρ2 (w) = T ρ1 (w)T −1 ,

for all w ∈ 4ev Vc ,

5.3. Mapping Spinors

167

unique up to multiples. Consider next an arbitrary multivector w ∈ 4Vc . This can be uniquely written as w = w1 + wn w2 , with w1 , w2 ∈ 4ev Vc , if we fix a main reflection wn ∈ 4n Vc . If ρ1 (wn ) = ρ2 (wn ), then ρ2 (w)T = (ρ2 (w1 ) + ρ2 (wn )ρ2 (w2 ))T = T (ρ1 (w1 ) + ρ1 (wn )ρ1 (w2 )) = T ρ1 (w), since w1 , w2 ∈ 4ev Vc and ρi (wn ) = ±I. If ρ1 (wn ) = −ρ2 (wn ), then ρ2 (w)T = (ρ2 (w1 ) + ρ2 (wn )ρ2 (w2 ))T = T (ρ1 (w1 ) − ρ1 (wn )ρ1 (w2 )) = T ρ1 (w). b

5.3

Mapping Spinors

Consider two vector spaces V1 and V2 and a linear map T : V1 → V2 . In Section 2.3, we saw that this induces a unique linear map T = T∧ : ∧V1 → ∧V2 of multivectors, which is in fact a homomorphism with respect to the exterior product. When V1 and V2 are inner product spaces, we saw in Proposition 4.1.10 that T∧ : 4V1 → 4V2 will be a homomorphism with respect to Clifford products if and only if T is an isometry. In this section, we study in what sense a linear map T : V1 → V2 induces a linear map of spinors / 2. T4 / 1 → 4V / : 4V To avoid extra technicalities, we consider only Euclidean spinors in this section. Consider first an invertible isometry T between Euclidean spaces. We have fixed complex Vi representations ρi : Vi → L(4V / i ), i = 1, 2. This means that on V = V1 , we have the two complex V representations ρ1 and ρ2 T . It follows from Theo/ 2 ) such that rem 5.2.3 that there exists an invertible map T4 / 1 ; 4V / ∈ L(4V T4 / (w.ψ) = (T w).(T4 / ψ),

/ 1. w ∈ 4V1 , ψ ∈ 4V

In odd dimension, some care about how the main reflectors map is needed. See Proposition 5.3.5. This construction of induced maps T4 / of spinors leads to the following two questions, which we address in this section. • How unique can we make the spinor map T4 / induced by the vector map T ? • Is there a natural way to define a spinor map T4 / for more general invertible vector maps T that are not isometries? / 2 ) is Concerning the first question, the problem is that T4 / 1 ; 4V / ∈ L(4V unique only as a projective map T4 / 1 /C → 4V / 2 /C, / : 4V

Chapter 5. Spinors in Inner Product Spaces

168

that is, T4 / , λ ∈ C \ {0}. When, for / is unique only up to complex multiples λT4 example, we are constructing and working with spinors over manifolds as we do in Chapter 12, this presents problems. To this end, we next define two additional natural structures, an inner product and a conjugation, on spinor spaces that allow us to obtain induced maps of spinors that are unique only up to sign ±T4 / . Such sign ambiguity will always be present, but this discrete nonuniqueness will not cause any problems. Proposition 5.3.1 (Spinor inner product). Let V be a Euclidean space, with complex spinor space 4V / . Then there exists a Hermitian complex inner product (·, ·i on / such that 4V (ψ1 , v.ψ2 i = (v.ψ1 , ψ2 i, ψ1 , ψ2 ∈ 4V, (5.2) / for all vectors v ∈ V . If (·, ·i0 is another Hermitian inner product for which (5.2) holds, then there is a constant λ > 0 such that (ψ1 , ψ2 i0 = λ(ψ1 , ψ2 i for all ψ1 , ψ2 ∈ 4V / . Proof. Fix a basis for 4V / , view ψi as column vectors and ρ(v) as matrices. Then a sesquilinear duality is uniquely represented by an invertible matrix M such that (ψ1 , ψ2 i = ψ1∗ M ψ2 . Condition (5.2) translated to M is that M ρ(v) = ρ(v)∗ M for all vectors v ∈ V . We note that v 7→ ρ(v)∗ is a second complex V representation, which extends to the algebra homomorphism w 7→ ρ(wc )∗ . Therefore the existence of M , unique up to complex nonzero multiples, follows from Theorem 5.2.3. Note that in odd dimensions, these two representations coincide on 4n Vc . It remains to see that M can be chosen as a self-adjoint positive definite matrix. We note that when we are using the standard representation from Example 5.1.5, all matrices ρ(v) are self-adjoint. Hence M = I can be used in this basis. Proposition 5.3.2 (Spinor conjugation). Let V be a Euclidean space of dimension n, with spinor space 4V / . Then there exists an antilinear map 4V / → 4V / : ψ 7→ ψ † such that ( † (v.ψ) = v.ψ † , n 6≡ 3 mod 4, (5.3) † (v.ψ) = −v.ψ † , n ≡ 3 mod 4, for all vectors v ∈ V and spinors ψ ∈ 4V / , and satisfying ( † (ψ † ) = ψ, n ≡ 0, 1, 2, 7 mod 8, †

(ψ † ) = −ψ, 0

n ≡ 3, 4, 5, 6

mod 8.

If ψ 7→ ψ † is another such map for which this holds, then there is λ ∈ C, |λ| = 1, 0 such that ψ † = λψ † for all ψ ∈ 4V / .

5.3. Mapping Spinors

169

Note that in even dimensions n ≡ 0, 2 mod 8, this spinor conjugation provides a real structure on the complex spinor space, as in Section 1.5, in accordance with Theorem 3.4.13, which shows that in these dimensions the real-Euclidean Clifford algebras are isomorphic to real matrix algebras. The technicalities about signs and dimension in the statement of Proposition 5.3.2 are best understood from its proof. Proof. Fix a basis for 4V / , view ψ as column vectors and ρ(w) as matrices. Write N c for the componentwise complex conjugation of a matrix N . Every antilinear / : ψ 7→ ψ † can be written / → 4V map 4V c

ψ † = (N ψ) for some matrix N . The condition †

(v.ψ) = v.(ψ † )

(5.4) c

c

is equivalent to N ρ(v) = (ρ(v)) N , for all vectors v. We note that that v 7→ (ρ(v)) is a second complex V representation, which extends to the algebra homomorphism c w 7→ (ρ(wc )) , where wc denotes the real structure on 4Vc . Existence of antilinear maps satisfying (5.4), unique up to λ ∈ C \ {0}, follows from Theorem 5.2.3, provided n is even, or if c

c

ρ(e1 ) · · · ρ(en ) = ρ(e1 ) · · · ρ(en ) when n is odd. Using a standard complex representation from Example 5.1.5, we see that this holds unless n ≡ 3 mod 4, since all but m of the matrices representing the basis vectors are real. When n ≡ 3 mod 4, by Theorem 5.2.3 the correct c b = (ρ(wc )) N . relation is N ρ(w) To complete the proof, we claim that using a standard representation, we may choose ρ(e−1 · · · e−m ), n = 2m ≡ 0 mod 4, n = 2m + 1 ≡ 1 mod 4, ρ(e−1 · · · e−m ), N := n = 2m ≡ 2 mod 4, ρ(e1 · · · em ), n = 2m + 1 ≡ 3 mod 4. ρ(e1 · · · em e0 ), c

c

Indeed, we note that ρ(ek ) = ρ(ek ), k ≥ 0, and ρ(ek ) = −ρ(ek ), k < 0, and we verify that N c N = I when n ≡ 0, 1, 2, 7 mod 8 and N c N = −I when n ≡ 3, 4, 5, 6 † † mod 8. This completes the proof, since λ(λψ † ) = |λ|2 (ψ † ) . Definition 5.3.3 (Normed spinor space). Let V be a Euclidean space, with spinor space 4V / . Fix a spinor inner product, by which we mean a Hermitian inner product on 4V / such that all vectors act as self-adjoint maps as in Proposition 5.3.1. This amounts to a choice of the parameter λ > 0.

170

Chapter 5. Spinors in Inner Product Spaces

Fix also a spinor conjugation, by which we mean an antilinear map on 4V / with properties as in Proposition 5.3.2. This amounts to a choice of the parameter |λ| = 1. We refer to the triple (4V, / (·, ·i, ·† ) as a normed spinor space. Lemma 5.3.4 (Compatibility). A spinor inner product and a spinor conjugation are compatible in the sense that c

(ψ1 , ψ2 i = (ψ1 † , ψ2 † i,

ψ1 , ψ2 ∈ 4V. /

Proof. Note that (ψ1 , ψ2 i0 := (ψ1 † , ψ2 † i

c

defines a second Hermitian spinor inner product. By uniqueness in Proposition c 5.3.1, we have (ψ1 , ψ2 i = λ(ψ1 † , ψ2 † i for some λ > 0. In particular, †

†

(ψ, ψi = λ(ψ † , ψ † i = λ2 ((ψ † ) , (ψ † ) i = λ2 (ψ, ψi, so λ = 1.

We can now answer the first question posed above, concerning uniqueness of induced spinor maps. For simplicity we write the inner product and conjugation in both spaces below with the same symbols. Proposition 5.3.5 (Uniqueness of spinor maps). Let V1 , V2 be Euclidean spaces, with normed spinor spaces 4V / 1 and 4V / 2 respectively. Assume that T : V1 → V2 is an invertible isometry. When dim V1 is odd, we assume that the main reflectors wn and T wn in V1 and V2 respectively both act as +I or as −I. Then there exists † † an isometric complex linear map T4 / 1 → 4V / 2 such that (T4 / : 4V / ψ) = T4 / (ψ ) and T4 ψ ∈ 4V, / v ∈ V. (5.5) / (v.ψ) = (T v).(T4 / ψ), 0 0 If T4 /. / is another such map, then T4 / = ±T4

Proof. We saw at the beginning of this section how Theorem 5.2.3 implies the existence of T4 / satisfying (5.5), and every other such map is of the form λT4 / for some λ ∈ C \ {0}. To see that T4 / can be chosen to be isometric, consider the Hermitian inner product (ψ1 , ψ2 i0 := (T4 / 1 . We calculate / ψ1 , T4 / ψ2 i on 4V 0 (ψ1 , v.ψ2 i0 = (T4 / ψ1 , T v.T4 / ψ2 i = (T v.T4 / ψ1 , T4 / ψ2 i = (v.ψ1 , ψ2 i ,

for v ∈ V . Proposition 5.3.1 shows that (T4 / ψ1 , T4 / ψ2 i = µ(ψ1 , ψ2 i for some µ > 0, −1/2 so λT4 will be isometric if |λ| = µ . / To see that T4 / can be chosen to be compatible with spinor conjugation, † −1 consider the antilinear map ψ 7→ T4 / 1 . We calculate / ψ) on 4V / (T4 †

†

†

†

−1 −1 −1 −1 T4 / (v.ψ)) = T4 / ψ)) = T4 / ψ) ) = v.T4 / ψ) , / (T4 / ((T v).(T4 / ((T v).(T4 / (T4

5.3. Mapping Spinors

171

for v ∈ V , where = ±1 depending on the dimension as in Proposition 5.3.2. Since also the square of this antilinear map coincides with the square of spinor † −1 † conjugation on 4V / 1 , we conclude from Proposition 5.3.2 that T4 / ψ) = σψ , / (T4 iα for some σ ∈ C, |σ| = 1. Therefore e T4 / will be compatible with spinor conjugation if e−2iα = σ. These two equations for the modulus and argument of λ have exactly two solutions differing by sign, which completes the proof. We next consider the second question posed above, concerning how to define a map of spinors T4 / , when the map of vectors T : V1 → V2 is not an isometry. Recall that when T is an isometry, writing ρk : Vk → L(4V / k ), k = 1, 2, for the complex Vk representations defining the spinor spaces, we used that ρ2 T was a second complex V1 representation. Comparing this to ρ1 , the existence of T4 / followed from Theorem 5.2.3. When T is not an isometry, this argument breaks down, since ρ2 T is not a complex V1 representation. Indeed, ρ2 (T v)2 = hT vi2 I 6= hvi2 I. What we do in this case is to produce an isometry U : V1 → V2 from T by polar factorization. Proposition 1.4.4 shows that there is a unique isometry U : V1 → V2 such that T = U S1 = S2 U, for some positive symmetric maps S1 ∈ L(V1 ) and S2 ∈ L(V2 ). The formula for this U , which we refer to as the polar isometric factor of T , is U := T (T ∗ T )−1/2 = (T T ∗ )−1/2 T. Definition 5.3.6 (Induced spinor map). Let V1 , V2 be Euclidean spaces, with normed spinor spaces 4V / 1 and 4V / 2 respectively. Assume that T : V1 → V2 is an invertible linear map. Denote by U : V1 → V2 the polar isometric factor of T . If dim V1 is odd, we assume that the main reflectors wn and U wn in V1 and V2 respectively both act as +I or as −I. Then we refer to the two maps / 2 / 1 → 4V T4 / := U4 / : 4V constructed from U as in Proposition 5.3.5 as the spinor maps induced by T . Exercise 5.3.7. Consider the Euclidean plane V , with ON-basis {e1 , e−1 }, and con/ from Examsider the standard representation of the complex spinor space 4V ple 5.1.5(i), equipped with the spinor duality and conjugation from the proofs of Propositions 5.3.1 and 5.3.2. Calculate the two spinor maps T4 / / → 4V / : 4V induced by the linear map T : V → V with matrix 1 0 T = 3 2 in the basis {e1 , e−1 }.

Chapter 5. Spinors in Inner Product Spaces

172

Recall from Section 2.3 that if vector maps T induce multivector maps T∧ , then (T −1 )∧ = (T∧ )−1

and

(T2 ◦ T1 )∧ = (T2 )∧ ◦ (T1 )∧ ,

but (λT )∧ 6= λT∧ and (T1 + T2 )∧ 6= (T1 )∧ + (T2 )∧ . The corresponding result for induced spinor maps holds for isometries, but in general the composition rule fails for non-isometries. Exercise 5.3.8 (Failure of transitivity). (i) Let T : V1 → V2 be an invertible linear map with polar isometric factor U : V1 → V2 . Show that T −1 has polar isometric factor U −1 . Conclude that −1 (T4 /) are the spinor maps induced by T −1 . (ii) Let T1 = T be as in (i), and let T2 : V2 → V3 be an invertible linear map with polar isometric factor U2 : V2 → V3 . Construct T1 and T2 such that U2 U1 is not the polar isometric factor of T2 T1 . Conclude that (T2 )4 / ◦ (T1 )4 / are not in general the spinor maps induced by T2 ◦ T1 . Show, however, that this is the case when at least one of the maps T1 and T2 is an isometry. For the action of general multivectors on spinors, we note the following somewhat surprising result. Proposition 5.3.9 (Polar factorization of induced maps). Let T : V1 → V2 be an invertible linear map between Euclidean spaces, and define its polar isometric factor U : V1 → V2 as above. Let T∧ : ∧V1 → ∧V2 be the ∧ homomorphism induced by T from Proposition 2.3.2, and let U∧ = U4 : 4V1 → 4V2 be the induced 4 (as well as ∧) homomorphism induced by U . Then the polar isometric factor of T∧ equals U∧ . Proof. We have T∧ ((T∧ )∗ T∧ )−1/2 = T∧ ((T ∗ )∧ T∧ )−1/2 = T∧ ((T ∗ T )∧ )−1/2 = T∧ ((T ∗ T )−1/2 )∧ = (T (T ∗ T )−1/2 )∧ = U∧ . For the first equality, see Section 2.7. The third equality uses (A2 )∧ = ((A−1 )∧ )−2 for A = (T ∗ T )1/2 , which is true.

5.4

Abstract Spinor Spaces

In Section 3.3 we introduced the notion of abstract Clifford algebras, among which we treat 4V = (∧V, 4) as the standard Clifford algebra. Similarly, we introduced the concept of abstract spinor spaces in Definition 5.1.1, for which it takes two full

5.4. Abstract Spinor Spaces

173

V -rotations two complete one full rotation of the spinor space. Among these we / from Section 5.2 as the standard spinor space. However, treat the spinor space 4V there is one important difference: all abstract Clifford algebras are isomorphic to 4V , but there are infinitely many nonisomorphic abstract spinor spaces, as we shall see. The goal of the present section is to identify all possible abstract complex spinor and tensor spaces of three and four-dimensional Euclidean space, up to isomorphism. We start by collecting the basic tools from the theory of representations of compact Lie groups, which we need. The only groups we use are G = SO(V ) and G = Spin(V ), where V is a given Euclidean space. • Let G denote a compact Lie group. A complex representation (S, ρ) of G is a complex linear space S together with a smooth homomorphism ρ : G → L(S). If (S, ρ) is a representation of G, and if S 0 ⊂ S is a subspace such that ρ(g)w ∈ S 0 ,

for all w ∈ S 0 , g ∈ G,

then (S 0 , ρ0 ) is said to be a subrepresentation of (S, ρ), where ρ0 (g) denotes the restriction of ρ(g) to the invariant subspace S 0 . If (S, ρ) has no nontrivial subrepresentations, that is none besides S 0 = {0} and S 0 = S, then we say that (S, ρ) is an irreducible representation of G. • Let (S1 , ρ1 ) and (S2 , ρ2 ) be two representations of G. We write LG (S1 , S2 ) := {T ∈ L(S1 , S2 ) ; ρ2 (g)T = T ρ1 (g), g ∈ G} and call T ∈ LG (S1 , S2 ) a G-intertwining map. The representations (S1 , ρ1 ) and (S2 , ρ2 ) are isomorphic if there exists a bijective map T ∈ LG (S1 , S2 ). b be the set of equivalence classes of mutually isomorphic irreducible Let G representations of G. Schur’s lemma shows that if (S1 , ρ1 ), (S2 , ρ2 ) are irreducible representations of G, then (i) LG (S1 , S2 ) = {0} if they are nonisomorphic, and (ii) LG (S1 , S2 ) = {λT ; λ ∈ C \ {0}} if T ∈ LG (S1 , S2 ) is an isomorphism. The proof follows from the observation that if T ∈ LG (S1 , S2 ), then N(T ) is a subrepresentation of S1 and R(T ) is a subrepresentation of S2 . Irreducibility allows us to conclude. • Given a representation (S, ρ) of G, we can write S = S1 ⊕ S2 ⊕ · · · ⊕ Sk ,

(5.6)

where each Sj is an invariant subspace of ρ(g) for all g ∈ G, and each Sj is an irreducible representation of G. For the proof, we construct an auxiliary Hermitian inner product (·, ·i such that G acts isometrically on S, that is, (ρ(g)wi2 = (wi2 ,

for all w ∈ S, g ∈ G.

174

Chapter 5. Spinors in Inner Product Spaces Such invariant inner products are not unique. The existence follows from the well-known result in measure theory that there exists a Haar measure on G, that is, a Borel measure dµ with total measure µ(G) = 1 that is left and right invariant in the sense that Z Z Z f (gx)dµ(x) = f (x)dµ(x) = f (xg)dµ(x), G

G

G

for all Borel measurable functions f : G → C and g ∈ G. Starting from any Hermitian inner product (·, ·i0 on S, we see that Z (w1 , w2 i := (ρ(g)w1 , ρ(g)w2 i0 dµ(g) G

defines a Hermitian inner product that is invariant under G. To obtain a decomposition of the form (5.6), we simply note that if S1 is a subrepresentation, then so is S1⊥ , using a complex analogue of Lemma 4.1.2. We continue to split S1 and S1⊥ further until irreducible subrepresentations have been obtained. • Given a representation (S, ρ) of G, the number of irreducible subrepresentab that it contains is well defined. To tions from each equivalence class α ∈ G see this, assume that S = S1 ⊕ S2 ⊕ · · · ⊕ Sk = S10 ⊕ S20 ⊕ · · · ⊕ Sl0 b and are two decompositions of S into irreducible representations. Fix α ∈ G consider any Si ∈ α appearing in the first decomposition, and p0j : Si → Sj0 : w 7→ p0j (w), where p0j denotes orthogonal projection onto Sj0 . Schur’s lemma implies that Si ⊂ ⊕Sj0 ∈α Sj0 . With this and the reverse result obtained by swapping the roles of Si and Sj0 , we get k = l and ⊕Si ∈α Si = ⊕Sj0 ∈α Sj0 ,

b α ∈ G.

b are present in • A tool for identifying which irreducible representations α ∈ G a given representation (S, ρ), and their multiplicities, is the character χS of (S, ρ). This is the function χS : G → C : g 7→ χS (g) := Tr(ρS (g)), where Tr denotes the trace functional as in Section 1.4. Since Tr(T ρ(g)T −1 ) = Tr(ρ(g)), isomorphic representations have the same character. The Peter– b Weyl theorem in representation theory shows that the characters χα , α ∈ G, of the irreducible representations form an ON-basis for −1 Lcl ) = f (x), for all x, g ∈ G}. 2 (G) := {f ∈ L2 (G) ; f (gxg

5.4. Abstract Spinor Spaces

175

Such functions f are referred to as class functions on G. We shall not use the general fact that the class functions span L2cl (G). To see the orthogonality of χS1 and χS2 for two nonisomorphic irreducible representations, we consider the auxiliary representation of G on S := L(S1 ; S2 ) given by ρ(g)T := ρS2 (g) ◦ T ◦ ρS1 (g −1 ), g ∈ G. c

This is seenR to have character χ(g) = χS1 (g) χS2 (g). Define the linear operator P := G ρ(g)dµ(g) ∈ L(S), using componentwise integration, and note that Z Z Z Z 2 ρ(gh)dµ(g)dµ(h) ρ(h)dµ(h) = P = ρ(g)dµ(g) G G G G Z Z = ρ(g)dµ(g) dµ(h) = P. G

G

Thus P is a projection, and we check that its range is LG (S1 ; S2 ), which gives Z c χ1 (g) χS2 (g)dµ(g) = TrP = dim LG (S1 ; S2 ). G

The orthonormality of characters therefore follows from Schur’s lemma. • As a corollary of the Peter–Weyl theorem, it follows that there are at most countably many nonisomorphic irreducible representations of G. Moreover, the number of R irreducible subrepresentations that a given representation contains equals G |χS |2 . We now apply these tools from Lie group representation theory, starting with the three-dimensional spin group. Fixing an ON-basis for V , we write Spin(n) := Spin(Rn ),

and SO(n) := SO(Rn ).

Recall from Theorem 3.4.13 that 4R3 is isomorphic to C(2). By restricting such an isomorphism to Spin(3) ⊂ 4R3 , this Lie group is seen to be isomorphic to the complex isometries on C2 . Definition 5.4.1. Let the special unitary group SU(2) in two complex dimensions be n a o b 2 2 SU(2) := c c ; a, b ∈ C, |a| + |b| = 1 . −b a By the standard isomorphism SU(2) ↔ Spin(3), we mean a b ↔ a1 − a2 j1 − b1 j2 − b2 j3 ∈ Spin(3) = S 3 ⊂ H, SU(2) 3 −bc ac where j1 = e23 , j2 = e31 , j3 = e12 , and a = a1 + ia2 and b = b1 + ib2 . Here i ∈ C is the algebraic imaginary unit, not related to H.

Chapter 5. Spinors in Inner Product Spaces

176

We set out to find the characters of all irreducible representations of SU(2) = Spin(3) = S 3 . Note that if f is a class function on this group, then it is uniquely determined by its values at points exp(−it) 0 ↔ cos t + j1 sin t = exp(tj1 ) ∈ Spin(3), SU(2) 3 exp(it) 0 for t ∈ [0, π]. Indeed, as noted at the end of Section 3.2, all rotations of pure quaternions ∧2 V can be represented x 7→ qxq −1 , q ∈ Spin(3). In particular, we can rotate any pure quaternion to the line [j1 ] in this way. From hqxi2 = hxi2 = hxqi2 , for all q ∈ S 3 , x ∈ H, it is clear that Lebesgue surface measure on Spin(V ) is invariant under R Spin(3), and therefore equals the Haar measure, modulo the normalizing factor S 3 |dˆ x| = 2π 2 . Consider representations (Vk , ρk ) of SU(2), where Vk := {polynomials P : C2 → C ; P (λz, λw) = λk P (z, w), λ ∈ R, z, w ∈ C}, and ρk (T )P := P ◦ T −1 . Concretely, Vk is spanned by {z k−j wj ; 0 ≤ j ≤ k} and a b c c . (ρk (T )P )(z, w) = P (a z − bw, b z + aw) if T = −bc ac Proposition 5.4.2 (Finding all 3D spinor spaces). The irreducible representations d of the Lie group SU(2) = Spin(3) are indexed by the natural numbers Spin(3) = d {0, 1, 2, 3, . . .}. The representations in the equivalence class k ∈ Spin(3) are (k+1)dimensional and are uniquely determined by the character values k/2 X 1 + 2 cos(2mt), k even, m=1 χk (exp(tj1 )) = (k−1)/2 2 X cos((2m + 1)t), k odd. m=0

With the terminology of Definition 5.1.1, these irreducible representations are abstract spinor spaces if k is odd and abstract tensor spaces if k is even. Proof. Note that z k−m wm is an eigenvector to ρk (exp(tj1 )) with eigenvalue eit(k−2m) . Summing these gives the stated characters. We calculate Z π 1 |χk (exp(tj1 ))|2 4π sin2 tdt |χk (g)| dµ(g) = 2 2π 3 S 0 2 Z π X Z k 2 2 π 1 it(k−2m) it −it = e (e − e ) dt = | sin((k + 1)t)|2 dt = 1, π 0 2i m=0 π 0

Z

2

5.4. Abstract Spinor Spaces

177

so these representations are irreducible. Moreover, it is known from Fourier theory ∞ is dense L2 , so these are all the irreducible representations. For that {cos(jx)}j=0 the last statement, we note that −1 ∈ Spin(V ) acts by multiplication with (−1)k on Vk . Example 5.4.3 (4R3 ). The standard representation of Spin(V ) on the three-dimensional vector space Vc itself is given by ρ = p from Proposition 4.1.9. The rotor q = exp(tj1 ) acts by rotation the angle 2t and with plane of rotation [j1 ]. Summing the diagonal elements of the matrix for this rotation in the standard basis yields d ) the character χ(exp(tj1 )) = 1 + 2 cos(2t). Thus the tensor space 2 ∈ Spin(V is the standard representation Vc = ∧1 Vc . Through the Hodge star map and Proposition 4.1.10, ∧2 Vc is an isomorphic representation. Even more trivial: the d ), where representations on ∧0 Vc and ∧3 Vc are both representatives of 0 ∈ Spin(V all rotors act as the identity on a one-dimensional space. / for three-dimensional Eu/ 3 ). Consider the spinor space 4V Example 5.4.4 (4R clidean space V from Definition 5.2.4, and restrict this to a complex Spin(V ) representation. Using the matrices from Example 5.1.7(ii), we have 0 1 0 −i i 0 , ρ(j1 ) = ρ(e1 e−1 ) = = 1 0 i 0 0 −i it e 0 tj1 = and therefore the character for this representation is χ(e ) = Tr 0 e−it d ) is the 2 cos t. We conclude that the smallest abstract V -spinor space 1 ∈ Spin(V standard spinor space 4V / . The following two examples build on Section 8.2, but are not used elsewhere and may be omitted and returned to after Section 8.2 has been read. Example 5.4.5 (Harmonic polynomials). Consider the abstract V -tensor space of scalar (complexified) k-homogeneous harmonic polynomials Pksh from Example 5.1.2(iii). To avoid dealing with bases for these spaces, we recall Proposition 8.2.3, which amounts to a statement about decomposition of the representation on all k-homogeneous polynomials into subrepresentations. We have the representation ρ(q)P (x) = P (q −1 xq) on s Pks = Pksh ⊕ |x|2 Pk−2 , s s where both the terms are subrepresentations, and |x|2 ∈ LSpin(3) (Pk−2 , |x|2 Pk−2 ) sh is an isomorphism. It follows that the character for the representation Pk is the s difference between the characters of the two representations Pks and Pk−2 . To avoid unnecessarily technical trace computations, we choose the convenient basis β γ {x1α z β z γ }α+β+γ=k = {xα 1 (x2 + ix3 ) (x2 − ix3 ) }α+β+γ=k

for Pks . Here {x1 , x2 , x3 } are the coordinates in the ON-basis {e1 , e2 , e3 } for V , that is, the dual basis for V ∗ = P1s . From ρ(etj1 )P (x) = P (e−tj1 xetj1 ) we see that

Chapter 5. Spinors in Inner Product Spaces

178

{x1α z β z γ } is an eigenbasis of the operator ρ(etj1 ) with eigenvalues e2it(β−γ) . This gives the character X X χ(etj1 ) = e2it(β−γ) = e2it(β−γ) α+β+γ=k

β+γ≤k

for the representation Pks , and therefore the character χ(etj1 ) =

X

e2it(β−γ) = 1 +

k X

cos(2mt)

m=0

β+γ=k−1,k

for the representation Pksh . We conclude that Pksh is the irreducible V -tensor space d ). 2k ∈ Spin(V Example 5.4.6 (Monogenic polynomials). Consider the abstract V -spinor space of k-homogeneous monogenic polynomials Pkm from Example 5.1.2(iv), for a threedimensional Euclidean space V . On this real linear space of dimension 8(k + 1), there is a rather natural complex structure, namely multiplication by the orientation e123 ∈ 43 V , which belongs to the center of the algebra. In this way, we consider Pkm as a 4(k + 1)-dimensional complex spinor space with rotors acting as ρ(q)P (x) = qP (q −1 xq). Similar to Example 5.4.5, having Proposition 8.2.3 in mind, we first compute the character of the spinor space Pk (also considered as a complex linear space) with same same action. To handle the multivector-valued polynomials, we note that we have an isomorphism of representations Pk ↔ H ⊗ Pks , where q ∈ Spin(V ) acts on H as x 7→ qx, and on Pks as P (x) 7→ P (q −1 xq). It follows that the character is X χPk (etj1 ) = χH (etj1 )χPks (etj1 ) = 4 cos t e2it(β−γ) , β+γ≤k

from which we get χPkm (etj1 ) = χPk (etj1 ) − χPk−1 (etj1 ) = 4 cos t

k X j=0

e2it(k−2j) = 4

k X

cos((2j + 1)t).

j=0

Proposition 5.4.2 now shows that the V -spinor space Pkm is not irreducible but d contains two copies of the irreducible spinor space 2k + 1 ∈ Spin(3). We can find such a subrepresentation as a minimal left ideal S ⊂ 4V . Then the subspace of monogenic polynomials Pkm with values in S will be such an irreducible spinor subrepresentation. Note that such subrepresentations are not unique. And indeed, there are infinitely many left ideals S that can be used.

5.4. Abstract Spinor Spaces

179

We next study representations of the group Spin(4) of rotors in four-dimensional Euclidean space V . It is a fortunate fact that we have an isomorphism Spin(4) ↔ Spin(3) × Spin(3), as we shall see. This should not come as a surprise, since 4ev V , by Proposition 3.3.5, is isomorphic to the Clifford algebra for R3 with negative definite inner product, which in turn is isomorphic to H ⊕ H according to Theorem 3.4.13. To make these isomorphisms explicit, fix an ON-basis {e1 , e2 , e3 , e4 } and an orientation e1234 for V , and define p± := 21 (1 ± e1234 ). We have p2± = p± ,

p+ + p− = 1,

ev

and p± ∈ Z(4 V ). Therefore p± split the even subalgebra ev 4ev V = 4ev + V ⊕ 4− V, ev where 4ev ± V := p± 4 V are the two two-sided ideals in the even subalgebra. The subspaces 4ev V are real algebras in themselves, each being isomorphic to H, but ± should not be considered subalgebras of 4ev V , since p± , and not 1, is the identity ev element in 4ev ± V . To make the isomorphism 4 V = H ⊕ H explicit, we identify basis elements as 1 4ev + V 3 p+ = p+ e1234 = 2 (1 + e1234 ) ↔ e ∈ H, 1 4ev + V 3 p+ e41 = p+ e23 = 2 (e41 + e23 ) ↔ j1 = −i ∈ H, 1 4ev + V 3 p+ e42 = −p+ e13 = 2 (e42 − e13 ) ↔ j2 = −j ∈ H, 1 4ev + V 3 p+ e43 = p+ e12 = 2 (e43 + e12 ) ↔ j3 = −k ∈ H,

with notation as in Section 3.2, but writing e ∈ H for the identity in H to avoid confusion with 1 ∈ 4ev V . We then use the self-inverse automorphism 4ev V → 4ev V : w 7→ e4 we4 ev 0 0 to identify 4ev − V = 4+ V , and write e := e4 ee4 = p− and jk := e4 jk e4 , k = 1, 2, 3.

Proposition 5.4.7 (Spin(4) = Spin(3) × Spin(3)). Let V be a four-dimensional Euclidean space. We have an algebra isomorphism 4ev V → H ⊕ H : w 7→ (p+ w, e4 p− we4 ), where we identify H = 4ev , + V as above. Letting h·, ·iH = 2h·, ·i4ev + V this isomorphism is an isometry. This algebra isomorphism restricts to a group isomorphism Spin(V ) → S 3 ×S 3 , where S 3 = Spin(3) is the unit quaternion 3-sphere, as in the following diagram: / H⊕H 4evO V (5.7) O ? Spin(V ) p

SO(V )

? / S3 × S3 p˜

/ SO(H)

180

Chapter 5. Spinors in Inner Product Spaces

The standard covering map p : Spin(V ) → SO(V ) : q 7→ q(·)q −1 corresponds to the covering map p˜ : S 3 × S 3 → SO(H) : (q1 , q2 ) 7→ q1 (·)q2−1 if we use the linear identification V → H : v 7→ p+ ve4 . Proof. From the above discussion it is clear that 4ev V → H ⊕ H : w 7→ (p+ w, e4 p− we4 ) is an algebra isomorphism. To check the stated mapping of the spin group, write w ∈ 4ev V as w = (w0 1 + w1234 e1234 ) + (w12 e12 + w34 e34 ) + (w13 e13 + w24 e24 ) + (w14 e14 + w23 e23 ) = (w0 + w1234 )e + (w0 − w1234 )e0 + (−w14 + w23 )j1 + (w14 + w23 )j10 + (−w13 − w24 )j2 + (−w13 + w24 )j20 + (w12 − w34 )j3 + (w12 + w34 )j30 = (x0 e + x1 j1 + x2 j2 + x3 j3 ) + (y0 e0 + y1 j10 + y2 j20 + y3 j30 ). It follows that x20 + x21 + x22 + x32 = 1 = y02 + y12 + y22 + y32 if and only if |w| = 1 and w0 w1234 − w14 w23 + w13 w24 − w12 w34 . By Example 4.1.8 this is equivalent to w ∈ Spin(V ). Next consider the action of Spin(V ) on V with p. We have qvq −1 = qv q = (p+ q + p− q)v(p+ q + p− q) = q1 ve4 q 2 e4 + e4 q2 e4 vq 1 , where q1 := p+ q, q2 := e4 p− qe4 ∈ H = 4ev + V . Note that two terms vanish, since vp+ = p− v and p+ p− = 0. To write this action entirely in terms of quaternions, we need to identify V = H. Multiplying v ∈ V by e4 , we have ve4 ∈ 4ev V . Projecting onto the subspace 4ev + V , we verify that p+ ve4 = v1 i + v2 j + v3 k + v4 e ∈ H = 4ev +V if v = v1 e1 + v2 e2 + v3 e3 + v4 e4 ∈ V . We obtain p+ (qvq −1 )e4 = q1 (p+ ve4 )q 2 + (p+ e4 q2 e4 )vq 1 e4 = q1 (p+ ve4 )q2−1 , since p+ e4 q2 = 0, which completes the proof.

We now construct all irreducible representations of Spin(V ) for a four-dimensional Euclidean space V . Fix an ON-basis {e1 , e2 , e3 , e4 } and consider the subgroup T := {eθ1 e12 +θ2 e34 ; θ1 , θ2 ∈ R} ⊂ Spin(V ). In representation theory, such a subgroup is referred to as a maximal torus. This subgroup plays the same role for Spin(4) as did the circle, the one-dimensional

5.4. Abstract Spinor Spaces

181

torus, {etj1 ; t ∈ R} for Spin(3). Using the isomorphism from Proposition 5.4.7, we have e12 ↔ (j3 , j3 ) and e34 ↔ (−j3 , j3 ), and thus eθ1 e12 +θ2 e34 ↔ (et1 j3 , et2 j3 ),

where t1 = θ1 − θ2 , t2 = θ1 + θ2 .

The importance of the maximal torus T ⊂ Spin(V ) is that a class function f : Spin(V ) → C is uniquely determined by its values on T , in the sense that for each element q ∈ Spin(V ) there is q1 ∈ Spin(V ) such that q1 qq1−1 ∈ T . In fact, only the values on part of T are needed. Using the S 3 × S 3 characterization, this result carries over from three dimensions, but we can also prove it directly for Spin(V ). According to Proposition 4.3.9(i), which we note is a result on maximal tori in the Lie groups Spin(V ), each q ∈ Spin(V ) belongs to one such maximal torus (with e10 e20 , e03 e40 instead of e12 , e34 ). We can then find q1 ∈ Spin(V ) such that the induced rotation w 7→ q1 wq1−1 of 4V maps e10 e02 , e03 e04 to e12 , e34 , possibly after first having adjusted the angles so that the two bases have the same orientation. Proposition 5.4.8 (Finding all 4D spinor spaces). For the Lie group Spin(4) = Spin(3)×Spin(3), the irreducible representations are indexed by the pairs of natural d numbers Spin(4) = N2 = {(k, l) ; k, l = 0, 1, 2, 3, . . .}. The representations in the d are (k + 1)(l + 1)-dimensional and are uniquely equivalence class (k, l) ∈ Spin(4) determined by the character values χ(k,l) (eθ1 e12 +θ2 e34 ) = χk (e(θ1 −θ2 )j3 )χl (e(θ1 +θ2 )j3 ), d of where χk denotes the character for the irreducible representation k ∈ Spin(3) Spin(3) from Proposition 5.4.2. The irreducible abstract V -tensor spaces correspond to those pairs for which k + l is even, and the irreducible abstract V -spinor spaces correspond Rto those for which k + l is odd. The integral Spin(4) f (x)dµ(x) of a class function f on Spin(4) with respect to Haar measure equals Z πZ π 4 f (exp( 12 (t1 + t2 )e12 + 12 (−t1 + t2 )e34 )) sin2 t1 sin2 t2 dt1 dt2 . π2 0 0 Proof. It is clear that the Haar measure on Spin(3) × Spin(3) is the product measure dµ(x1 )dµ(x2 ). We see that a class function is determined by its values on the quarter {(et1 j3 , et2 j3 ) ; 0 < t1 , t2 < π} of the maximal torus, and the stated integral formula follows by translating back to Spin(V ). Now let (Vk , ρk ) be the irreducible representations of SU(2) = Spin(3) used in the proof of Proposition 5.4.2. For (k, l) ∈ N2 , define the representation ρ(k,l) of Spin(3) × Spin(3) on the tensor product space Vk ⊗ Vl by applying the universal property to the bilinear map (P1 , P2 ) 7→ (ρk (q1 )P1 ) ⊗ (ρl (q2 )P2 ) to obtain a linear map ρ(k,l) (q1 , q2 ) ∈ L(Vk ⊗ Vl ) for each (q1 , q2 ) ∈ Spin(3) × Spin(3). The character of this representation is seen to be χ(k,l) (q1 , q2 ) =

182

Chapter 5. Spinors in Inner Product Spaces

χk (q1 )χl (q2 ), and Fubini’s theorem shows that

R Spin(4)

|χ(k,l) (q1 , q2 )|2 = 1, so this is

an irreducible representation of Spin(3)× Spin(3). In particular, for q1 = e(θ1 −θ2 )j3 and q2 = e(θ1 +θ2 )j3 we obtain the stated character values. 2 2 Since {χk (etj3 )}∞ k=0 is an ON-basis for L2 ((0, π); π sin tdt), the functions t1 j3 t2 j3 ∞ {χk (e )χl (e )}k,l=0 form an ON-basis for L2 ((0, π) × (0, π); π42 sin2 t1 sin2 t2 dt1 dt2 ). Hence the representations ρ(k,l) constitute all the possible irreducible representations of Spin(3) × Spin(3) = Spin(4). Example 5.4.9 (4R4 ). The standard representation of Spin(V ) on the four-dimensional space Vc itself is given by ρ = p from Proposition 4.1.9. The rotor q = eθ1 e12 +θ2 e34 acts by rotation in the plane [e12 ] through an angle 2θ1 and rotation in the plane [e34 ] through an angle 2θ2 . This gives the character χ(eθ1 e12 +θ2 e34 ) = 2 cos(2θ1 ) + 2 cos(2θ2 ) = (2 cos(θ1 − θ2 ))(2 cos(θ1 + θ2 )), d so (1, 1) ∈ Spin(4) is the standard V -tensor space. Turning to the induced representation of Spin(V ) on ∧2 V , this is not irreducible for a four-dimensional space V . Indeed, it splits into two subrepresentations 2 ev on ∧2± V := 4ev ± V ∩ ∧ V , where 4± V are the two two-sided ideals of the even subalgebra as above. Using the basis {e41 + e23 , e42 − e13 , e43 + e12 } for ∧2+ V , we obtain the character χ∧2+ V (eθ1 e12 +θ2 e34 ) = 1 + 2 cos(2(θ1 − θ2 )), d so this is the V -tensor space (2, 0) ∈ Spin(4). On the other hand, a trace calculation with the basis {−e41 +e23 , e42 +e13 , −e43 +e12 } for ∧2− V gives χ∧2− V (eθ1 e12 +θ2 e34 ) = 1 + 2 cos(2(θ1 + θ2 )), and we conclude that ∧2− V is the V -tensor space (0, 2) ∈ d Spin(4). Note the somewhat surprising result that the standard four-dimensional V -tensor space (1, 1) is not the smallest: there are two nonisomorphic threedimensional V -tensor spaces (2, 0) and (0, 2)! Example 5.4.10 (4R / 4 ). Consider the spinor space 4V / from Definition 5.2.4 for a four-dimensional space V and the restriction of ρ to Spin(V ) ⊂ 4ev V . We pick bases {1, e12 , e1 , e2 } for 4V / and {e1 , e−1 , e2 , e−2 } for V . Using the matrices from Example 5.1.8(i), we obtain cos(θ1 − θ2 ) − sin(θ1 − θ2 ) 0 0 sin(θ1 − θ2 ) cos(θ1 − θ2 ) 0 0 . ρ(eθ1 e12 +θ2 e34 ) = 0 0 cos(θ1 + θ2 ) sin(θ1 + θ2 ) 0 0 − sin(θ1 + θ2 ) cos(θ1 + θ2 )

5.5. Comments and References

183 +

The block structure of this matrix is due to the two subrepresentations 4 / V − and 4 / V . For these we obtain the characters χS + = 2 cos(θ1 − θ2 ) and χS − = 2 cos(θ1 + θ2 ) respectively, so these two representations are the two irreducible d respectively. V -spinor spaces (1, 0) and (0, 1) ∈ Spin(4) Exercise 5.4.11. Extend Examples 5.4.5 and 5.4.6 to four-dimensional Euclidean space, and find the irreducible subrepresentations contained in (Pksh )c and (Pkem )c . Note that in this case, (Pkem )c 6= Pkm .

5.5

Comments and References

´ Cartan. Here is a quotation from 5.1 Spinors in general were discovered by Elie his book [26]: “Spinors were first used under that name, by physicists, in the field of Quantum Mechanics. In their most general form, spinors were discovered in 1913 by the author of this work, in his investigations on the linear representations of simple groups.” There exist many variations of the construction of spinors. The complex spinors that we construct here are usually referred to as Dirac spinors in physics. The method of imaginary rotations used in the proof of Proposition 5.1.6 to reduce to the case of Euclidean space is, in the case of spacetime, referred to as Wick rotation. An inspirational book on Clifford algebras and spinors is Lounesto [64]. 5.2 In Chapter 3, we recycled Grassmann’s exterior algebra ∧V , and identified the Clifford algebra 4V and the exterior algebra ∧V . This space of multivectors is the same as a linear space, but we have two different associative products 4 and ∧ on it, leading to two different algebras. In view of Ex/ as some ample 5.1.5, it is tempting to try to realize the spinor space 4V suitable subspace of a Clifford algebra. One obvious way is to view 4V / as a minimal left ideal in 4V , as first proposed by M. Riesz. However, there is no canonical choice of minimal left ideal, and there is no geometric reason for such an identification. The ideal point of view is also problematic in considering spinor bundles as in Section 11.6. The approach in this book is that the spinor space 4V / is a new independent universe, without any relation to other spaces other than that multivectors act as linear operators on spinors, and that it is unnatural to try to set up any identification between the spinor space and any other space. As before we follow the principle of abstract algebra, as explained in the introduction of Chapter 1, to construct / . However, to calculate with spinors one may choose any favorite repre4V sentation of them, which may consist in using some ad hoc identification. In this book, we choose the identification with a certain exterior algebra in Example 5.1.5.

184

Chapter 5. Spinors in Inner Product Spaces / was introduced by Feynman in physics to denote The slash notation D the representation of a vector by gamma matrices. In this book we use the notation differently. We have chosen to use the slash as a symbol for spinor objects, to distiguish them from the related Clifford algebra objects. For / for spinor space as compared to 4V for Clifford algebra, and example 4V / will denote Dirac operators acting on spinor fields as compared to D for D Dirac operators acting on multivector fields. We mainly denote objects in spinor spaces by ψ, following the tradition from quantum mechanics, where they represent the wave functions of particles.

5.3 The notion of spinor conjugation is related to charge conjugation in physics, which is a transformation that switches particles and antiparticles. See Hladik [59]. A reference for spinor inner products is Harvey [50]. Spinor inner products and conjugation are well-defined natural structures on spinor spaces also for general inner product spaces. In Section 9.2 we study the case of spacetime with three space dimensions, and explain how spinors are used in the Dirac theory in quantum mechanics. The construction of induced spinor maps using polar decomposition of non-isometries is due to Bourgignon [22]. A motivating application is to construct a map of spinor fields between nonisometric manifolds, and this was / operators in Bandara, McIntosh, used in studying perturbations of 4-Dirac and Ros´en [17]. 5.4 The representation theory for compact Lie groups used in this section can be found in many textbooks on the subject, including results for higherdimensional spin groups beyond the three- and four-dimensional examples that we limit ourselves to in this book. Our discussion of small spin groups gathers inspiration from Br¨ocker and tom Dieck [24]. See also Fulton and Harris [41] and Gilbert and Murray [42]. The spin representation from Example 5.4.6 was introduced by Sommen [86, 87, 88].

Chapter 6

Interlude: Analysis Road map: This chapter is not where to start reading the second part of this book on multivector analysis, which rather is Chapter 7. The material in the present chapter is meant to be used as a reference for some background material from analysis, which we use in the remaining chapters. A main idea in this second part is that of splittings of function spaces: H = H1 ⊕ H 2 .

(6.1)

When H1 is a k-dimensional subspace of a linear space H of finite dimension n, then every subspace H2 of dimension n − k that intersects H1 only at 0 is a subspace complementary to H1 . When H is an infinite-dimensional linear space, which is typically the case for the linear spaces of functions that we use in analysis, it is seldom of any use to have only an algebraic splitting (6.1), meaning that every x ∈ H can be written x = x1 + x2 , for unique x1 ∈ H1 and x2 ∈ H2 . Using the axiom of choice, one can show that every subspace H1 has an algebraic complement H2 . Instead, we restrict attention to topological splittings of Hilbert and Banach spaces H, meaning that we assume that H1 and H2 are closed subspaces and that we have an estimate kx1 k + kx2 k ≤ Ckxk for some C < ∞. See Definition 6.4.5. The Hilbert space interpretation of this latter reverse of the triangle inequality is that the angle between H1 and H2 is positive. We consider splittings of Banach spaces also, and it should be noted that closed subspaces H1 of a Banach space H in general do not have any topological complement H2 . Such a complementary subspace exists precisely when there exists a bounded projection onto H1 . A well-known example is the subspace c0 of the sequence space `∞ , which does not have any topological complement at all. In © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_6

185

Chapter 6. Interlude: Analysis

186

⊥ Hilbert space, though, we can always use the orthogonal √ complement H2 = H1 . In this case we have kx1 k2 + kx2 k2 = kxk2 , and C = 2 suffices. When doing analysis, not only with scalar valued functions but with more general multivector fields, we can replace the Laplace operator ∆ by more fundamental first-order partial differential operators d, δ, and D. When working with these first-order operators on domains D, splittings of function spaces occur frequently. In Section 7.6, Chapter 10, and Section 9.6 we encounter the Hodge splittings associated with the exterior and interior derivatives d and δ. These are orthogonal splittings of L2 (D), involving also a finite-dimensional third subspace. In Section 8.3 we encounter the Hardy splittings associated to the Dirac operator D. These are in general nonorthogonal, but topological, splittings of L2 (∂D). In Section 9.3 we consider generalizations of the Hardy splittings for time-harmonic waves, and in Sections 9.4 and 9.5 we show that the fundamental structure behind elliptic boundary value problems consists not only of one splitting, but two independent splittings: one related to the differential equation and one related to the boundary conditions. For Dirac’s original equation for the time evolution of spin-1/2 particles in relativistic quantum mechanics, we also see in Section 9.2 how splittings appear, for example, in the description of antiparticles and chirality. Splittings also appear in index theory for Dirac operators in Chapter 12, where the Dirac operators are considered in the splittings L2 (M ; 4M ) = + − L2 (M ; 4ev M )⊕L2 (M ; 4od M ) and L2 (M ; 4M / ) = L2 (M ; 4 / M )⊕L2 (M ; 4 / M ), respectively, on our manifold M . In the present chapter, the material in Section 6.4 should not be needed before Section 9.4, with the exception of splittings of function spaces.

6.1

Domains and Manifolds

An extremely useful notation for estimates in analysis is the following, which we use in the remainder of this book. Definition 6.1.1 (The analyst’s (in)equality). By the notation X . Y , we mean that there exists C < ∞ such that X ≤ CY for all relevant values of the quantities X and Y . By X & Y we mean Y . X, and X ≈ Y means X . Y and X & Y . We use the following standard terminology and notation concerning the regularity of functions. Consider a function f : D → L, with D an open subset of some affine space X, and L a linear space, or possibly affine. Denote by B(x, r) := {y ∈ X ; |y − x| < r} the ball with center x and radius r > 0 when X is Euclidean space. • The function f is C k -regular in D, k = 0, 1, 2, . . ., if all directional/partial derivatives ∂v1 · · · ∂vm f (x) of order m ≤ k exist as continuous functions of x ∈ D, for all directions vi ∈ V . Here ∂v f (x) := lim (f (x + hv) − f (x))/h. h→0

6.1. Domains and Manifolds

187

Given a basis {ei }, we write ∂i := ∂ei . We say that f is C ∞ -regular if it is C k -regular for all k < ∞. older regular of order 0 < α < 1 in D if • The function f is H¨ |f (x) − f (y)| . |x − y|α ,

for all x, y ∈ D,

and we write f ∈ C α (D; L). For α = 0, f ∈ C 0 (D; L) = C(D; L) means that f is continuous on D. When α = 1, we say that f is Lipschitz regular and write f ∈ C 0,1 (D; L). Note that the precise value of the implicit constant C as in Definition 6.1.1, but not the H¨older or Lipschitz property of f , depends on the choice of Euclidean norm | · | on X. • A bijective function f : D → D0 , with an open set D0 ⊂ L, is a homeomorphism if f ∈ C 0 (D; D0 ) and f −1 ∈ C 0 (D0 ; D). Lipschitz diffeomorphisms and C k -diffeomorphisms are defined similarly. A diffeomorphism refers to a C ∞ -diffeomorphism. • The support of a function defined in X is the closed set supp f := {x ∈ X ; f (x) 6= 0}. If f ∈ C ∞ (D), D an open set, then we write f ∈ C0∞ (D) if supp f is a compact subset of D. • Write C k (D) := {F |D ; F ∈ C k (X)}, and similarly for C α and C ∞ . When the range L of the function is clear from the context, we suppress L in the notation and abbreviate C k (D; L) to C k (D). Definition 6.1.2 (Total derivative). Let D ⊂ X be an open set in an affine space X with vectors V , and let (X 0 , V 0 ) be a second affine space. If ρ : D → X 0 is differentiable at x ∈ D, then we define its total derivative at x to be the unique linear map ρx : V → V 0 such that ρ(x + v) − ρ(x) = ρx (v) + o(v), where o(v) denotes a function λ(v) such that λ(v)/|v| → 0 when v → 0. With respect to bases {ei } and {e0i }, ρx has matrix ∂1 ρ1 (x) · · · ∂k ρ1 (x) .. .. .. , . . . ∂1 ρn (x) · · ·

∂k ρn (x)

P 0 where ρ = ρi ei and partial derivatives ∂i = ∂ei are with respect to ei . Equivalently, the total derivative is ρx (v) = ∂v ρ(x). Note that when ρ maps between affine spaces, then the total derivative ρx maps between the vector spaces since differences of points are vectors. To simplify notation, we shall often drop subscripts x and write ρ.

Chapter 6. Interlude: Analysis

188

The total derivative of a differential map between affine spaces extends from a map of vectors to a map of multivectors as in Section 2.3. With our notation, for example, the chain rule takes the form ρ2 ◦ ρ1 x (w) = ρ2 ρ

1 (x)

ρ1 x (w),

w ∈ ∧V1 ,

for the composition of maps ρ1 : X1 → X2 and ρ2 : X2 → X3 . Definition 6.1.3 (Jacobian). Let ρ : D → X 0 be as in Definition 6.1.2, with total derivative ρx : V → V 0 . Denote by ρx : ∧V → ∧V 0 the induced linear map. Assume that X, X 0 are oriented n-dimensional affine spaces with orientations en and e0n respectively. Then its Jacobian Jρ (x) is the scalar function representing ρx |∧n V , that is, the determinant Jρ (x) := he0∗ n , ρx (en )i of ρx . The main use of Jacobians is in the change of variables formula Z Z f (y)dy = f (ρ(x))Jρ (x)dx ρ(D)

(6.2)

D

for integrals. For Lipschitz change of variables ρ, this continues to hold. Note that in this case Jρ is well defined almost everywhere, since Lipschitz maps ρ are differentiable almost everywhere by Rademacher’s theorem. We use the following standard terminology for domains D ⊂ X. Definition 6.1.4 (Domains). Let D be a domain, that is, an open subset, in an ndimensional affine space (X, V ). We say that D is a C k -domain, k = 1, 2, . . ., if its boundary D is C k -smooth in the following sense. At each p ∈ ∂D, we assume that there exists a C k diffeomorphism ρ : Ωp → Dp between a neighborhood Ωp ⊂ Rn of 0 and a neighborhood Dp ⊂ X such that ρ({x ∈ Ωp ; xn > 0}) = Dp ∩ D, ρ({x ∈ Ωp ; xn = 0}) = Dp ∩ ∂D, and ρ({x ∈ Ωp ; xn < 0}) = Dp \ D. Lipschitz domains are defined similarly, by requiring that the local parametrizations ρ be C 0,1 diffeomorpisms. In a Euclidean space X, we denote by ν the outward-pointing unit normal vector field on ∂D. For a C k -domain, ν is a C k−1 -regular vector field defined on

6.1. Domains and Manifolds

189

all ∂D. For a Lipschitz domain, by Rademacher’s theorem, ν is well defined at almost every point p ∈ ∂D. In many cases it is important to consider domains beyond C 1 , such as Lipschitz domains. For example, the intersection and union of two C 1 domains is much more likely to be Lipschitz than C 1 . However, as the following example indicates, Lipschitz domains constitute a far wider class than domains with a finite number of corners, edges, etc. Example 6.1.5 (Lipschitz scale invariance). We consider how a function φ : R → R scales. Assume that φ(0) = 0 and let φn (x) := nφ(x/n). Thus the graph of φn represents what φ looks like around 0 through a magnifying glass that magnifies n times. If φ is C 1 regular, then |φn (x) − φ0 (0)x| ≤ n |x|, |x| < 1, where n → 0 when n → ∞. This means that φ “looks flat” on small enough scales, since it is well approximated by the straight line y = φ0 (0)x. On the other hand, if φ is a Lipschitz function, then φn is another Lipschitz function with the same Lipschitz constant C. In contrast to the C 1 case, φn will not converge to a linear function, as is seen, for example, from φ(x) = |x|, for which φn (x) = |x| for all n. However, this example is very atypical for Lipschitz functions. In general, each φn will give an entirely new function. This means that a Lipschitz function is nontrivial, that is, nonflat, on each scale, but still nondegenerate, that is, still a Lipschitz function. By the implicit function theorem, the boundary of a C k domain, k = 1, 2, . . ., is locally the graph of a C k function, in the sense that the local parametrization ρ can be written ρ(x0 , xn ) = (x0 , xn + φ(x0 )),

x0 ∈ Rn−1 ,

(6.3)

in a suitable basis for X = V , where φ : Rn−1 → R is a C k -regular function. In stark contrast, this is not true for Lipschitz domains. Example 6.1.6 (Bricks and spirals). (i) In R3 , let D1 := {(x, y, z) ; −1 < x < 0, −1 < y < 1, −1 < z < 0} and D2 := {(x, y, z) ; −1 < x < 1, −1 < y < 0, 0 < z < 1}. Placing the “brick” D2 on top of D1 , consider the two-brick domain D with D = D1 ∪ D2 . Then D is a Lipschitz domain, but at the origin ∂D is not the graph of a Lipschitz function. (ii) In polar coordinates (r, θ) in R2 , consider the logarithmic spiral D := {(r cos θ, r sin θ) ; e−(θ+a) < r < e−(θ+b) , θ > 0}, where b < a < b + 2π are two constants. Then D is a Lipschitz domain, but at the origin ∂D is not the graph of a Lipschitz function. If D is a Lipschitz domain in which all local parametrizations ρ of ∂D are of the form (6.3) with C 0,1 functions φ, then we say that D is a strongly Lipschitz domain.

190

Chapter 6. Interlude: Analysis

Exercise 6.1.7 (Star-shaped domains). We say that a domain D is star-shaped with respect to some point p ∈ D if for each x ∈ D, the line {p + t(x − p) ; t ∈ [0, 1]} in contained in D. Show that every bounded domain in a Euclidean space that is starshaped with respect to each point in some ball B(p; ) ⊂ X, > 0, is a strongly Lipschitz domain. Conversely, show that every bounded strongly Lipschitz domain is a finite union of such domains that are star-shaped with respect to some balls. Exercise 6.1.8 (Rellich fields). Let D be a bounded strongly Lipschitz domain in a Euclidean space (X, V ). Show that there exists a vector field θ ∈ C0∞ (X; V ) such that inf hν(x), θ(x)i > 0. x∈∂D

A partition of unity, see below, may be useful. Besides open subsets, that is, domains in affine space, we also make use of lower-dimensional curved surfaces. More generally, we require the notion of a manifold from differential geometry, which we now fix notation. We consider only compact manifolds, but both with and without boundary, and in many cases embedded in an affine space. For simplicity, we consider only regularity k ≥ 1. n Our notation is the following. Let H+ := {(x0 , xn ) ; x0 ∈ Rn−1 , xn ≥ 0} and n−1 0 0 n R+ := {(x , xn ) ; x ∈ R , xn > 0} denote the closed and open upper halfspaces, and identify Rn−1 and Rn−1 ×{0}. In general, let M be a compact (second countable Hausdorff) topological space, for example a compact subset of an affine space X. n , in the sense that we are • We assume that M is locally homeomorphic to H+ given a collection of charts, that is, homeomorphisms {µα : Dα → Mα }α∈I , n the atlas S for M , between open sets Dα ⊂ H+ and Mα ⊂ M such that M = α∈I Mα . By compactness, we may assume that the index set I is finite.

• Define open sets Dβα := µ−1 α (Mβ ) ⊂ Dα , and transition maps µβα : Dβα → Dαβ : x 7→ µβα (x) := µ−1 β (µα (x)) for α, β ∈ I. We say that M is a (compact) C k -manifold if µβα ∈ C k (Dβα ) for all α, β ∈ I. In this case, these transition maps are C k diffeomorphisms, ∞ since µ−1 βα = µαβ . A manifold refers to a C -manifold. If all these transition maps are orientation-preserving, then we say that M is oriented. When it is possible to find another atlas with all transition maps between its charts orientation-preserving, then we say that M is orientable. More generally, a chart for M refers to any homeomorphism µ0 : D0 → n M between open sets D0 ⊂ H+ and M 0 ⊂ M such that µ0−1 ◦ µα ∈ 0

0 C k (µ−1 α (M )) for all α ∈ I.

6.2. Fourier Transforms

191

n • If Dα ⊂ R+ for all α ∈ I, then we say that M is a closed manifold. This means that M is a compact manifold without boundary. If Dα ∩ Rn−1 6= ∅ for some α ∈ I, then we say that M is a manifold with boundary. In this case, the boundary of M , denoted by ∂M , is the closed manifold defined as follows. Let Dα0 := Aα ∩ Rn−1 , 0 := µα |Rn−1 , µα

∅, and we may and Mα0 := µ0α (Dα0 ). It suffices to consider α such that Dα0 6= S 0 assume that Dα0 ⊂ Rn−1 + . Then ∂M is the closed manifold α∈I Mα with 0 0 0 atlas {µα : Dα → Mα }α∈I . • When M is a compact n-dimensional C k -manifold that is also a subset of an affine space X, with the topology inherited from X, then we say that M is an n-surface in X if the derivative µα of µα : Dα → Mα ⊂ X is injective for all x ∈ Dα and all α ∈ I. If µα ∈ C k (Dα ; X), then we say that M is a C k -regular n-surface in X. By the inverse function theorem, an n-surface is locally the graph of a C k -regular function in n variables, in a suitably rotated coordinate system for X. As above, n-surfaces may be closed or may have a boundary. If D ⊂ X is a bounded C k -domain in an affine space X as in Definition 6.1.4, then we see that M = D is a compact C k -regular n-surface with boundary. More generally but similarly, we can consider n-surfaces M embedded in some, in general higher-dimensional, manifold N . • For a function f : M → L on a C k manifold M , with values in a linear space L, we define f ∈ C j (M ; L) to mean that f ◦ µα ∈ C j (Dα ; L) for all α ∈ I, when j ≤ k. k A partition S of unity for a C -manifold M , subordinate to a finite covering M = α∈I Mα by open sets MP α ⊂ M , is a collection {µα }α∈I of functions such that supp ηα ⊂ Mα and α∈I ηα (x) = 1 for all x ∈ M . There exists such a partition of unity with ηα ∈ C k (M ; [0, 1]) on every C k -manifold M.

The standard use of a partition of unity is to localize problems: Given a function f on M , we write X f= ηα f. α

Here supp ηα f ⊂ Mα , and by working locally in this chart, we can obtain results for ηα f , which then we can sum to a global result for f .

6.2

Fourier Transforms

This section collects computations of certain Fourier transforms that are fundamental to the theory of partial differential equations. Fix a point of origin in an

192

Chapter 6. Interlude: Analysis

oriented affine space X and identify it with its vector space V . In particular, V is an abelian group under addition, and as such it comes with a Fourier transform. This is the linear operator Z ˆ f (x)e−ihξ,xi dx, ξ ∈ V ∗ . F(f )(ξ) = f (ξ) := V

This Fourier transform maps a complex-valued function f on V to another complex-valued function on V ∗ . If instead f takes values in some complex linear space L, we let F act componentwise on f . Assuming that V is a Euclidean space and V ∗ = V , the fundamental theorem of Fourier analysis is Plancherel’s theorem, which states that F defines, modulo a constant, an L2 isometry: Z Z 1 |f (x)|2 dx = |fˆ(ξ)|2 dξ. n (2π) V V We recall that the inverse Fourier transform is given by Z F −1 (fˆ)(x) = f (x) = fˆ(ξ)eihξ,xi dξ,

x ∈ V,

V∗

and basic formulas F(∂k f (x)) = iξk fˆ(ξ), F(f (x) ∗ g(x)) = fˆ(ξ) · gˆ(ξ), where the convolution of f (x) and g(x) is the function Z (f ∗ g)(x) := f (x − y)g(y)dy. V

The most fundamental of Fourier transforms is 2

F{e−|x| 2

that is, the Gauss function e−|x|

/2

/2

} = (2π)n/2 e−|ξ|

2

/2

,

is an eigenfunction to F.

Proposition 6.2.1 (Gaussians and homogeneous functions). Let f (x) be a homogeneous polynomial of degree j that is harmonic on an n-dimensional Euclidean space V . Then for every constant s > 0, we have Fourier transforms 2

2

F{f (x)e−s|x| } = 2−j cs−(n/2+j) f (ξ)e−|ξ|

/(4s)

,

where c = π n/2 (−i)j . For every constant 0 < α < n, we have Fourier transforms F{f (x)/|x|n−α+j } = 2α c where Γ(z) :=

R∞ 0

Γ((α + j)/2) f (ξ)/|ξ|α+j , Γ((n − α + j)/2)

e−t tz−1 dt is the gamma function, with Γ(k) = (k − 1)!.

193

6.2. Fourier Transforms Proof. (i) Calculating the Fourier integral, we have Z Z 2 −s|x|2 −ihx,ξi −|ξ|2 /(4s) e f (x)e−shx+iξ/(2s)i dx f (x)e dx = e V V Z −|ξ|2 /(4s) −s|x|2 f (x − iξ/(2s))e =e dx V ! Z Z 2

= e−|ξ|

∞

2

e−sr rn−1

/(4s)

f (rω − iξ/(2s))dω dr, |ω|=1

0

where we have extended f to a polynomial of n complex variables. According to the mean value theorem for harmonic functions, Z f (rω + y)dω = σn−1 f (y), |ω|=1

for every y ∈ V , where σn−1 is the area of the unit sphere in V . By analytic continuation, this formula remains valid for all complex y ∈ Vc . Since Z ∞ Z ∞ 1 1 −sr 2 n−1 e dr = n/2 r e−u un/2−1 = n/2 Γ(n/2) 2s 2s 0 0 and σn−1 = 2π n/2 /Γ(n/2), the stated identity follows. (ii) To establish the second Fourier transform identity, we use the identity Z ∞ Z ∞ 2 1 Γ((n − α + j)/2) x(n−α+j)/2−1 e−x dx = . s(n−α+j)/2−1 e−sr ds = n−α+j r rn−α+j 0 0 2

Writing r−(n−α+j) as a continuous linear combination of functions e−sr in this way, we deduce that Z ∞ 2 1 n−α+j F{f (x)/|x| }= s(n−α+j)/2−1 F{f (x)e−s|x| }ds Γ((n − α + j)/2) 0 Z ∞ 2 1 = s(n−α+j)/2−1 2−j cs−(n/2+j) f (ξ)e−|ξ| /(4s) ds Γ((n − α + j)/2) 0 Z ∞ 2 2−j cf (ξ) s−(α+j)/2−1 e−(1/s)(|ξ|/2) ds = Γ((n − α + j)/2) 0 Γ((α + j)/2) f (ξ) = 2α c . Γ((n − α + j)/2) |ξ|α+j The following functions, or more precisely distributions in dimension ≥ 3, appear in solving the wave equation. Proposition 6.2.2 (Riemann functions). Let Rt , for t > 0, be the Fourier multiplier F(Rt f )(ξ) =

sin(t|ξ|) F(f )(ξ). |ξ|

194

Chapter 6. Interlude: Analysis

In low dimensions, the Riemann function Rt has the following expression for t > 0: Z 1 Rt f (x) = f (x − y)dy, dim V = 1, 2 |y| 0, x2 > 0}, then ρ : D1 → D2 is a diffeomorphism. Let F be the constant vector field F (y) = e1 parallel to the y1 -axis. To push forward and pull back F to the x1 x2 -plane, we calculate the derivative a cos y2 −ay1 sin y2 ρy = . b sin y2 by1 cos y2 This gives the pushed forward vector field 1 a cos y2 −ay1 sin y2 1 a cos y2 x1 ρ∗ F = = =p . 2 2 b sin y2 by1 cos y2 0 b sin y2 (x1 /a) + (x2 /b) x2 On the other hand, pulling back F by ρ−1 gives −1 −1 a cos y2 −a−1 y1−1 sin y2 1 a cos y2 (ρ−1 )∗ F = −1 = 0 b−1 sin y2 b sin y2 b−1 y1−1 cos y2 −2 1 a x1 =p . −2 (x1 /a)2 + (x2 /b)2 b x2

220

Chapter 7. Multivector Calculus

Note that ρ∗ F is tangent to the radial lines ρ({y2 = constant}), and that (ρ−1 )∗ F is normal to the ellipses ρ({y1 = constant}), in accordance with the discussion above. See Figure 7.1(b)–(c).

Figure 7.1: (a) Change of variables F ◦ ρ−1 . (b) Pushforward ρ∗ F . (c) Inverse pullback (ρ−1 )∗ F . (d) Normalized pushforward ρ˜∗ F . The field has been scaled by a factor 0.3, that is, the plots are for F = 0.3e1 . Since F is constant, it is of course divergence- and curl-free: ∇yF = 0 = ∇∧F . By direct calculation, we find that (ρ−1 )∗ F is curl-free. For the pushforward, we note that p div(ρ∗ F ) = 1/ (x1 /a)2 + (x2 /b)2 6= 0. However, the normalized pushforward

7.2. Pullbacks and Pushforwards

221

1 x1 ρ˜∗ F = ab((x1 /a)2 + (x2 /b)2 ) x2 is seen to be divergence-free. See Figure 7.1(d). This is in accordance with Theorem 7.2.9 below. We now show that in general, pullbacks commute with the exterior derivative, and dually that normalized pushforwards commute with the interior derivative. At first it seems that taking the exterior derivative of a pulled back multicovector field would give two terms, a first-order term when the derivatives hit Θ(ρ(x)) according to the chain rule and a zero-order term when the derivatives hit ρx according to the product rule. However, it turns out that the zero-order term vanishes miraculously due to the alternating property of the exterior product and the equality of mixed derivatives. Theorem 7.2.9 (The commutation theorem). Let ρ : D1 → D2 be a C 2 map between open sets D1 ⊂ X1 and D2 ⊂ X2 . (i) If Θ : D2 → ∧V2∗ is a C 1 multicovector field in D2 , then the pullback ρ∗ Θ : D1 → ∧V1∗ is C 1 and ∇ ∧ (ρ∗ Θ)(y) = ρ∗ (∇ ∧ Θ)(y) for y ∈ D1 , that is, d(ρ∗ Θ) = ρ∗ (dΘ). (ii) Further assume that ρ is a C 2 diffeomorphism. If F : D1 → ∧V1 is a C 1 multivector field in D1 , then the normalized pushforward ρ˜∗ F : D2 → ∧V2 is C 1 and ∇ y (˜ ρ∗ F )(x) = ρ˜∗ (∇ y F )(x) for x ∈ D2 , that is, δ(˜ ρ∗ F ) = ρ˜∗ (δF ). The proof uses the following lemma, the proof of which we leave as an exercise. Lemma 7.2.10. Let {ei } and {e0i } be bases for V1 and V2 , with dual {e∗i } and P bases 0∗ 0∗ {ei } respectively. Then the pullback of a covector field θ(x) = i θi (x)ei is X ρ∗ θ(y) = θi (x) ∂j ρi (y)e∗j , x = ρ(y) ∈ D2 , y ∈ D1 , i,j

P and the pushforward of a vector field v(y) = i vi (y)ei is X ρ∗ v(x) = vi (y) ∂i ρj (y)e0j , x = ρ(y) ∈ D2 , y ∈ D1 . i,j

Proof of Theorem 7.2.9. Since both y 7→ ρ∗y and y 7→ Θ(ρ(y)) are C 1 , so is ρ∗ Θ. (i) When Θ = f is a scalar field, then the formula is the chain rule. Indeed, changing variables x = ρ(y) in the scalar function f (x), for ρ∗ f = f ◦ ρ we have X ∇y (f (ρ(y))) = e∗i (∂i ρk (y))(∂xk f )(x) = ρ∗y (∇f )(x), i,k

222

Chapter 7. Multivector Calculus

using Lemma 7.2.10. P (ii) Next consider a vector field Θ = θ = i θi ei : D2 → ∧V2∗ . Fix bases {ei } and {e0i } for V1 and V2 respectively and write {e∗i } and {e0∗ i } for the dual bases and ∂i and ∂i0 for the partial derivatives. From Lemma 7.2.10 we have X ∇ ∧ (ρ∗ θ) = ∇y ∧ θi (ρ(y)) ∂j ρi (y) e∗j i,j

=

X

(∂k θi ∂j ρi + θi ∂k ∂j ρi )e∗k ∧ e∗j =

i,j,k

X

∂k θi ∂j ρi e∗k ∧ e∗j ,

i,j,k

since ∂k ∂j = ∂j ∂k and ek ∧ el = −el ∧ ek . This is the key point of the proof. On the other hand, we have X X X 0∗ 0∗ ρ∗ (∇∧θ) = ρ∗ ∂j0 θi e0∗ = ∂j0 θi ρ∗ (e0∗ ∂j0 θi ∂k ρj ∂l ρi e∗k ∧e∗l . j ∧ ei j ∧ ei ) = i,j

i,j

i,j,k,l

Note that j ∂j0 θi ∂k ρj = ∂k θi by the chain rule. Thus changing the dummy index l to j proves the formula for covector fields. (iii) Next consider a general multicovector field Θ. By linearity, we may assume that Θ(x) = θ1 (x) ∧ · · · ∧ θk (x) for C 1 covector fields θj . We need to prove P

X

e∗i ∧ ρ∗ θ1 ∧ · · · ∧ ∂i (ρ∗ θj ) ∧ · · · ∧ ρ∗ θk =

i,j

X

∗ 1 ∗ 0 j ∗ k ρ∗ e0∗ i ∧ ρ θ ∧ · · · ∧ ρ (∂i θ ) ∧ · · · ∧ ρ θ .

i,j

For this, it suffices to show that X X ∗ 0 e∗i ∧ ∂i (ρ∗ θ) = ρ∗ e0∗ j ∧ ρ (∂j θ) i

j

for all C 1 covector fields θ in D2 . But this follows from step (ii) of the proof. (iv) From the hypothesis it follows that x 7→ ρρ−1 (x) , x 7→ |Jρ (ρ−1 (x))|, and x 7→ F (ρ−1 (x)) are C 1 . Therefore the product rule shows that ρ˜∗ F is C 1 . Let Θ : D2 → ∧V2∗ be any compactly supported smooth multicovector field. Then Propositions 7.1.7(ii) and 7.2.5(ii) and step (iii) above show that Z Z Z hΘ, ∇ y ρ˜∗ F idx = − h∇ ∧ Θ, ρ˜∗ F idx = − hρ∗ (∇ ∧ Θ), F idy D2 D2 D1 Z Z Z =− h∇ ∧ ρ∗ Θ, F idy = hρ∗ Θ, ∇ y F idy = hΘ, ρ˜∗ (∇ y F )idx. D1

D1

D2

Since Θ is arbitrary, we must have ∇ y (˜ ρ∗ F ) = ρ˜∗ (∇ y F ).

Example 7.2.11 (Orthogonal curvilinear coordinates). Let ρ : R3 → X be curvilinear coordinates in three-dimensional Euclidean space X. Important examples

7.2. Pullbacks and Pushforwards

223

treated in the standard vector calculus curriculum are spherical and cylindrical coordinates. The pushforward of the standard basis vector fields are e˜i := ρ∗ ei = ∂yi ρ(y),

i = 1, 2, 3,

ei } is in general not where {ei } denotes the standard basis in R3 . The frame {˜ an ON-frame in X, but in important examples such as the two mentioned above, these frame vector fields are orthogonal at each point. Assuming that we have such orthogonal curvilinear coordinates, we define hi (y) := |ρ∗ ei | and ei := e˜i /hi , for y ∈ R3 . This gives us an ON-frame {ei (y)} in X. We now show how the well-known formulas for gradient, divergence, and curl follow from Theorem 7.2.9. Note that 0 h1 0 ρ = ρ∗ = 0 h2 0 0 0 h3 with respect to the ON-bases {ei } and {ei }. For the gradient, we have X X ∇u = (ρ∗ )−1 grad(ρ∗ u) = (ρ∗ )−1 (∂i u)ei = h−1 i (∂i u)ei . i

i

∗

Note that ρ acts on scalar functions just by changing variables, whereas ρ∗ acts on vectors by the above matrix. P For the curl of a vector field F = i Fi ei in X, we similarly obtain X ∇ ∧ F = (ρ∗ )−1 curl(ρ∗ F ) = (ρ∗ )−1 ∇ ∧ hi Fi ei i ∗ −1

= (ρ )

XX j

∂j (hi Fi )ej

∧ ei

=

i

h1 e1 1 h2 e2 = h1 h2 h3 h3 e3

XX (hi hj )−1 ∂j (hi Fi )ej j

∂1 ∂2 ∂3

∧

ei

i

h1 F1 h2 F2 . h3 F3

Note that ρ∗ acts on ∧2 V by two-by-two subdeterminants of the above matrix as in Example 2.3.4. P For the divergence of a vector field F = i Fi ei in X, we use instead the normalized pushforward to obtain X X ∇ y F = ρ˜∗ div(˜ ρ∗ )−1 F = ρ˜∗ ∇ y h1 h2 h3 h−1 F e = ρ˜∗ ∂i (h1 h2 h3 h−1 i i i i Fi ) i

i

= (h1 h2 h3 )−1 ∂1 (h2 h3 F1 ) + ∂2 (h1 h3 F2 ) + ∂3 (h1 h2 F3 ) . Note that ρ˜∗ = ρ∗ /(h1 h2 h3 ) and that ρ∗ acts on vectors by the above matrix and simply by change of variables on scalars.

Chapter 7. Multivector Calculus

224

Example 7.2.12 (Pullback of Laplace equation). To see how the Laplace operator ∆ transforms under a change of variables, let ρ : D1 → D2 be a C 2 -diffeomorphism of Euclidean domains and let u : D2 → R be harmonic, that is, ∆u = 0. Changing variables, u corresponds to a function v(y) = u(ρ(y)) = ρ∗ u(y) in D1 . According to the commutation theorem (Theorem 7.2.9), we have −1 ρ∗ )−1 (ρ∗ )−1 (∇v) v) = ∇ y (ρ∗ )−1 (∇v) = ρ˜∗ ∇ y (˜ 0 = ∇ y ∇ ∧ (ρ∗ = ρ˜∗ ∇ y (ρ∗ ρ˜∗ )−1 (∇v) = ρ˜∗ div(A−1 grad v). Since ρ˜∗ is invertible, we see that the Laplace equation transforms into the divergence-form equation div(A−1 grad v) = 0. Here the linear map A(y), for fixed y ∈ D1 , in an ON-basis {ei } has matrix elements Ai,j (y) = hei , ρ∗ ρ˜∗ ej i = |Jρ (y)|−1 hρy (ei ), ρy (ej )i = g(y)−1/2 gi,j (y), where gi,j is the metric in D1 representing the Euclidean metric in D2 and g(x) := 2 det(gi,j (x)) = det(ρ∗ ρP ∗ ) = |Jρ (x)| is the determinant of the metric matrix. Thus 2 the Laplace equation i ∂i u = 0 transforms to the divergence-form equation X √ ∂i ( g g i,j ∂j v) = 0, i,j

where g i,j denotes the inverse of the metric matrix. Example 7.2.12 is a special case of the following pullback formulas for exterior and interior derivatives. Proposition 7.2.13 (Pullback of interior derivatives). Let X1 , X2 be oriented Euclidean spaces, let ρ : D1 → D2 be a C 2 -diffeomorphism, and denote by G(y) : ∧V1 → ∧V1 , y ∈ D1 , the metric of D2 pulled back to D1 , that is, hGw1 , w2 i = hρ∗ w1 , ρ∗ w2 i for multivector fields w1 , w2 in D1 . Write g(y) = det G|∧1 V1 (y) = |Jρ (y)|2 . Then we have pullback formulas ρ∗ (dF ) = d(ρ∗ F ), ρ∗ (δF ) = (g −1/2 G)δ(g 1/2 G−1 )(ρ∗ F ).

7.3

Integration of Forms

In this section, we develop integration theory for forms over k-surfaces. To avoid technicalities concerning bundles, at this stage we limit ourselves to k-surfaces in affine spaces, but the integration theory we develop generalizes with minor changes to general manifolds.

7.3. Integration of Forms

225

Definition 7.3.1 (k-Form). A k-form defined on a subset D of an affine space (X, V ) is a function b k V → L, Θ:D×∧ with range in a finite-dimensional linear space L, such that Θ(x, λw) = λΘ(x, w),

k

b V, λ > 0. x ∈ M, w ∈ ∧

We say that Θ is a linear k-form if for each x ∈ D, w 7→ Θ(x, w) extends to a linear function of w ∈ ∧k V . The idea with k-forms is that in integrating at a point x ∈ D, the integrand also depends on the orientation at x ∈ M of the k-surface M that we integrate over. Definition 7.3.2 (Integral of form). Let M be a compact oriented C 1 -regular kb k V → L be a continuous surface in an affine space (X, V ), and let Θ : M × ∧ k-form. We define the integral of Θ over M to be Z Θ(x, dx) := M

XZ α∈I

Dα

ηα (µα (y))Θ(µα (y), µα y (e1 ∧ · · · ∧ ek ))dy,

where {ei } is the standard basis in Rk and dy is standard Lebesgue measure in Rk ⊃ Dα . Here {µα }α∈I is the atlas of M , and {ηα }α∈I denotes a partition of unity for M . Note as in Section 2.4 how the induced action of the derivative µα y : ∧k Rk → ∧k V maps the oriented volume element e1 ∧ · · · ∧ ek dy in Rk to the oriented volume element dx = µα y (e1 ∧ · · · ∧ ek dy) = µα y (e1 ∧ · · · ∧ ek )dy on M . Note that this infinitesimal k-vector is simple, and hence Θ in general, needs to be defined only on the Grassmann cone. However, it is only to linear forms that Stokes’s theorem applies, as we shall see. The following proposition shows that such integrals do not depend on the precise choice of atlas and partition of unity for M , but only on the orientation of M. Proposition 7.3.3 (Independence of atlas). Consider a compact oriented C 1 -regular k-surface M , with atlas {µα : Dα → Mα }α∈I , in an affine space (X, V ). Let {µβ : Dβ → Mβ }β∈I 0 , I 0 ∩ I = ∅, be a second atlas for M such that all transition maps between Dα and Dβ , α ∈ I, β ∈ I 0 , are C 1 -regular and orientation-preserving. Further assume that {ηα }α∈I is a partition of unity for {µα }α∈I and that {ηβ }β∈I 0

Chapter 7. Multivector Calculus

226

is a partition of unity for {µβ }β∈I 0 . Then for every continuous k-form Θ, we have XZ ηα (µα (y))Θ(µα (y), µα y (e1 ∧ · · · ∧ ek ))dy α∈I

=

Dα

XZ β∈I 0

ηβ (µβ (z))Θ(µβ (z), µβ (e1 ∧ · · · ∧ ek ))dz. z

Dβ

P Proof. Inserting 1 = β∈I 0 ηβ in the integral on the left-hand side and 1 = P α∈I ηα in the integral on the right-hand side, it suffices to show that Z Z Θαβ (µα (y), µα y (e1 ∧ · · · ∧ ek ))dy = Θαβ (µβ (z), µβ (e1 ∧ · · · ∧ ek ))dz, Dβα

Dαβ

z

where Θαβ (x, w) := ηα (x)ηβ (x)Θ(x, w), since supp ηα ηβ ⊂ Mα ∩ Mβ . Changing variables z = µβα (y) in the integral on the right-hand side, we get Z Θαβ (µβ (µβα (y)), µβ (e1 ∧ · · · ∧ ek ))Jµβα (y)dy. µβα (y)

Dβα

Since µβ ◦ µβα = µα , and therefore µβ (e1 ∧ · · · ∧ ek )Jµβα = µα (e1 ∧ · · · ∧ ek ), the stated formula follows from the homogeneity of w 7→ Θαβ (x, w). Example 7.3.4 (Oriented and scalar measure). The simplest linear k-form in an affine space is Θ(x, w) = w ∈ L := ∧k V. R R In this case, M Θ(x, dx) = M dx = ∧k (M ) is the oriented measure of M discussed in Section 2.4. In a Euclidean space, given a continuous function f : M → L, the integral of the k-form Θ(x, w) := f (x)|w| R is seen to be the standard surface integral M f (x)dx, where dx = |dx|. Note that these are R not linear k-forms. In particular, f = 1 yields the usual (scalar) measure |M | := M dx of M . Using that dx = |dx| and the usual triangle inequality for integrals, we obtain from the definitions that oriented and scalar measure satisfy the triangle inequality Z Z dx ≤ dx. M

M

We continue with Example 2.4.4, where we calculated the oriented area element dx = (e12 + 2y2 e13 − 2y1 e23 )dy1 dy2 . Hence dx = |dx| =

q 1 + 4y12 + 4y22 dy1 dy2 ,

7.3. Integration of Forms

227

giving an area of the paraboloid equal to Z Z q 1 + 4y12 + 4y22 dy1 dy2 = 2π |y| 0. In the limit of small velocities v, the Navier– Stokes equations reduce to the linear Stokes equations ( ∇p − µ∆v = f, div v = 0. We are given the external forces f , and we want to compute the velcity vector field v and the scalar pressure p. The equations express conservation of momentum and the incompressibility of the fluid respectively. Assuming that the flow takes place inside a domain D, it is natural to demand in particular the boundary condition that the velocity vector field v is tangential at ∂D. We claim that the tangential Hodge decomposition Theorem 7.6.6 contains the solvability of this boundary value problem. To see this, we recall that with Clifford algebra we have a2 v = a y (a ∧ v) + aha, vi for vectors a and v. Replacing a by ∇, we obtain by incompressibility that ∆v = ∇ y (∇ ∧ v). Given a force vector field f ∈ L2 (D; ∧1 ), we apply the tangential Hodge decomposition (7.9) and conclude when f is orthogonal to Ck (D) that there exist a unique scalar function p ∈ R(δ; ∧0 ) and a tangential and exact bivector field ω ∈ R(d; ∧2 ) such that ∇p − µ∇ y ω = f. Again as in Corollary 7.6.8, there further exists a unique tangential vector field v ∈ R(δ; ∧1 ) such that ω = ∇ ∧ v. Note in particular that v is divergence-free and tangential at ∂D, and that it is uniquely determined if we demand that it be orthogonal to Ck (D; ∧1 ). The pressure p is unique modulo Ck (D; ∧0 ). The curl ω = ∇ ∧ v is the vorticity of the flow, which in three-dimensional space can be represented by the Hodge dual vector field.

7.7. Comments and References

7.7

253

Comments and References

7.1 The standard terminology for k-covector fields is differential forms. The anal´ Cartan, and they ysis of such differential forms was first developed by Elie have become a standard tool in modern differential geometry. The notation d and δ for the exterior and interior derivatives, when considered as partial differential operators, is standard in the literature. Beware, though, that it is standard to include an implicit minus sign in the notation δ, so that δF = −∇ y F and δ = d∗ . In the literature it is also common to write df also for the total derivative f = ∇ ⊗ f . We reserve the notation df for the exterior derivative. The nabla notations ∇ · F and ∇ × F are common in vector calculus. The generalizations ∇ ∧ F , ∇ y F , ∇ ⊗ F of this notation, used in this book, are inspired by Hestenes and Sobczyk [57] and Hestenes [56]. Nilpotence of an object a usually means that ak = 0 for some positive integer k. In this book, we consider only nilpotence of order 2. 7.2 The normalized pushforward of multivector fields is rarely found in the literature. In the special case of vector fields, this operation goes under the name Piola transformation in continuum mechanics. Pullbacks and pushforwards by smooth maps act in a natural way on multivector-valued distributions. For maps that are not diffeomorphisms, the pullback acts on test functions and the normalized pushforward act on distributions. Using k-covector fields as test functions, following George de Rham one defines a k-current as a k-vector-valued distribution. A classical reference for the applications of currents to geometric measure theory is Federer [38]. 7.3 The classical Stokes theorem from Example 7.3.10 dates back to Lord Kelvin and George Stokes in 1850. The general higher-dimensional result (7.5) was ´ Cartan in 1945. Standard notation with differential forms formulated by Elie ω is Z Z dω = ω, M

∂M

where the oriented measure dx is implicit in the notation ω. The trivial extension to forms with values in a finite-dimensional linear space, as presented in Theorem 7.3.9, is rarely found in the literature. The numerous applications in this book, though, show the usefulness of having such a generalized Stokes formula ready to use. 7.4–7.5 The identity L∗v Θ = ∇ ∧ (v y Θ) + v y (∇ ∧ Θ) is often referred to as Cartan’s magic formula due to its usefulness. A reference for the material in these two sections is Taylor [91], where results on ordinary differential equations, flows of vector fields, and the proof of Poincar´e’s theorem presented here can be found.

254

Chapter 7. Multivector Calculus

7.6 The theory of Hodge decompositions was developed by William Vallance Douglas Hodge in the 1930s, as a method for studying the cohomology of smooth manifold using PDEs. For further references for the presentation given here, which builds on the survey paper by Axelsson and McIntosh [14], we refer to Section 10.7. We find it convenient to work in the full exterior algebra ∧V . However, since the exterior and interior products preserve homogeneity of multivector fields, we may rather state the results at each level ∧k V to obtain Hodge splittings of k-vector fields, which is standard in the literature. A reference for solving boundary value problems with Hodge decompositions, also on more general manifolds with boundary, but under the assumption that the boundary is smooth, can be found in Schwarz [85]. Two harmonic analysis works using multivector calculus, motivated by Example 7.6.14 and nonsmooth boundary value problems for the Stokes equations, are McIntosh and Monniaux [68] and Ros´en [83]. Techniques from multivector calculus have also been used successfully in numerical analysis. A seminal paper on finite element exterior calculus is Arnold, Falk, and Winther [2].

Chapter 8

Hypercomplex Analysis Prerequisites: A solid background in analysis of one complex variable is required for this chapter, but no knowledge of analysis in several complex variables is needed. We make use of real-variable calculus, and build some on Chapter 7. Road map: We saw in Chapter 3 that even though it is natural to view Clifford algebras as a kind of hypercomplex numbers, the analogy fails in some important aspects, and it may be more appropriate to view Clifford algebras as matrix algebras, but from a geometric point of view. Nevertheless, a great deal of one-dimensional complex analysis does generalize to a noncommutative hypercomplex analysis in n-dimensional Euclidean space, replacing complex-valued functions by multivector fields. This yields a generalization of one-variable complex analysis that is fundamentally different from the commutative theory of several complex variables. Recall from one-dimensional complex analysis the following equivalent characterizations of analytic/holomorphic functions f : C → C defined in a domain D ⊂ C, where we assume that the total derivative f z is injective. The analysis definition: f is analytic if the limit f 0 (z) = limw→0 (f (z + w) − f (z))/w exists at each z ∈ D. The partial differential equation definition: f is analytic if it satisfies the Cauchy–Riemann system of partial differential equations ∂1 f1 (z) − ∂2 f2 (z) = 0,

∂2 f1 (z) + ∂1 f2 (z) = 0,

in D, where f1 , f2 are the real component functions of f = f1 + if2 . The algebra definition: f isP analytic if around each point z ∈ D it is locally the sum of a power ∞ series f (w) = k=0 ak (w − z)k , convergent in {w ∈ D ; |w − z| < r(z)} for some r(z) > 0. The geometry definition: f is analytic if it is an (orientation-preserving) © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_8

255

Chapter 8. Hypercomplex Analysis

256

conformal map, that is if at each z ∈ D the derivative f z is of the form a −b fz = , b a where a = ∂1 f1 and b = ∂1 f2 . This means that f z is a nonzero multiple of a rotation matrix and can be expressed as complex multiplication by f 0 (z). In generalizing to a hypercomplex analysis in higher-dimensional Euclidean spaces, the partial differential equation definition turns out to be most successful, where the Cauchy–Riemann equations are replaced by a Dirac equation ∇ 4 F (x) = 0, using the nabla operator induced by the Clifford product. As in Example 7.3.13, we may express this Dirac equation in terms of an integral difference quotient. Behind the Dirac equation, a fundamental type of splitting of function spaces is lurking: splittings into Hardy subspaces. With a solid understanding of Clifford algebra, these are straightforward generalizations of the classical such splittings in the complex plane. Recall that in the complex plane, any function f : ∂D → C on the boundary of a bounded domain D is in a unique way the sum f = f + + f −, where f + is the restriction to ∂D of an analytic function in D, and f − is the restriction to ∂D of an analytic function in C \ D that vanishes at ∞. The two subspaces consisting of traces of analytic functions from the interior or the exterior domain are the Hardy subspaces, and the Cauchy integral formula Z 1 f (w) dw 2πi ∂D w − z provides the projection operators onto these subspaces. There is one important difference with the Hodge splitting from Section 7.6: the two Hardy spaces are in general not orthogonal subspaces of L2 (∂D), but the angle between them depends on the geometry of ∂D. We show in Section 8.2 that the algebraic definition can be generalized to give power series expansions in higher dimensions. This is closely related to the classical theory of spherical harmonics. Later in Section 11.4, we shall see that the geometry definition does not generalize well. The higher dimensional conformal maps are very scarce indeed: the only ones are the M¨obius maps! Highlights: • The higher dimensional Cauchy integral formula: 8.1.8 • M¨ obius pullbacks of monogenic fields: 8.1.14 • Splitting into spherical monogenics: 8.2.6 • Spherical Dirac operator: 8.2.15 • Splittings into Hardy subspaces: 8.3.6

8.1. Monogenic Multivector Fields

257

8.1 Monogenic Multivector Fields In this chapter we work in Euclidean space (X, V ), and we study the following generalization of the Cauchy–Riemann equations. Definition 8.1.1 (Monogenic fields). Let D be an open set in a Euclidean space X. If F : D → 4V is differentiable at x ∈ D, we define the Clifford derivative ∇ 4 F (x) =

n X

ei∗ 4 ∂i F (x)

i=1

{e∗i }

is the basis dual to {ei }, and ∂i is the partial as in Definition 7.1.2, where derivative with respect to the corresponding coordinate xi . The Dirac operator D : F 7→ ∇ 4 F is the nabla operator induced by Clifford multiplication. If F is a C 1 multivector field for which ∇ 4 F (x) = 0 in all of D, then F is said to be a monogenic field in D. Let {es } be an induced ON-basis for 4V and write X F (x) = Fs (x)es . s

If F is a monogenic field, then each scalar component function Fs is a harmonic function. To see this, we note that X XX XX ei ej ∂i ∂j F (x) = ∂i2 F (x) = 0 = D2 F (x) = ∂i2 Fs (x) es i

j

i

s

i

X = (∆Fs (x))es . s

This is a consequence of the defining property v 2 = |v|2 for the Clifford product, and it means that D is a first-order differential operator that is a square root of the componentwise Laplace operator. Similar to the situation for analytic functions, a monogenic multivector field consists of 2n scalar harmonic functions, which are coupled in a certain sense described by the Dirac equation DF = 0. In particular, monogenic fields are smooth, even real analytic. The Dirac derivative further combines the exterior and interior derivative. Indeed, since e∗i 4 w = ei∗ y w + e∗i ∧ w, it is clear that DF (x) = ∇ 4 F (x) = ∇ y F (x) + ∇ ∧ F (x) = δF (x) + dF (x). This means that D2 = (d + δ)2 = d2 + δ 2 + dδ + δd = dδ + δd, by nilpotence. Another way to see this is to put v = θ = ∇ in the anticommutation relation from Theorem 2.8.1.

Chapter 8. Hypercomplex Analysis

258

As in Chapter 3, in using the Dirac operator, it is in general necessary to work within the full Clifford algebra, since typically DF will not be a homogeneous multivector field even if F is. However, in some applications the fields have values in the even subalgebra, or even are homogeneous k-vector fields. Example 8.1.2 (Analytic functions). Let X be a two-dimensional Euclidean plane, and let C = 40 V ⊕ 42 V be the standard geometric representation of the complex plane as in Section 3.2. Consider the Dirac equation ∇ 4 f = 0 for a complex valued function f = u + vj : C → C, where we have fixed an origin and ON-basis {e1 , e2 } in X = V , giving the identification V ↔ C, e1 ↔ 1, e2 ↔ j = e12 . Writing out the equation, we have ∇ 4 f = (e1 ∂1 + e2 ∂2 ) 4 (u + e12 v) = (∂1 u − ∂2 v)e1 + (∂1 v + ∂2 u)e2 . Thus ∇ 4 f = 0 coincides with the Cauchy–Riemann equations, and f is monogenic if and only if it is analytic. Note also that the only functions f : C → C that satisfy ∇ y f = 0 = ∇ ∧ f are the locally constant functions, since ∇ ∧ f = grad u and ∇ y f = −j grad v. On the other hand, the complex function f corresponds to the vector field F (x) = e1 f (x) under the identification V ↔ C. Reversing this relation gives F (x) = F (x) = f (x)e1 . Since the Clifford product is associative, it follows that F (x) is a plane divergence and curl-free vector field if and only if 0 = ∇ 4 F (x) = ∇ 4 (f (x) 4 e1 ) = (∇ 4 f (x)) 4 e1 , that is, if f is antianalytic. Example 8.1.3 (3D monogenic fields). Let F, G be vector fields and let f, g be scalar functions defined in an open set D in three-dimensional oriented Euclidean space. Then the multivector field f (x) + F (x) + ∗G(x) + ∗g(x) is monogenic if and only if div F (x) = 0, ∇f (x) − ∇ × G(x) = 0, ∇ × F (x) + ∇g(x) = 0, div G(x) = 0. We note that there is no restriction to assume that a monogenic field F takes values in the even subalgebra 4ev V . Indeed, if F : D → 4V is monogenic, we write F = F ev + F od where F ev : D → 4ev V and F od : D → 4od V . Then 0 = ∇ 4 F (x) = ∇ 4 F ev + ∇ 4 F od , where ∇ 4 F ev : D → 4od V and ∇ 4 F od : D → 4ev V , so we conclude that F ev and F od each are monogenic.

8.1. Monogenic Multivector Fields

259

Example 8.1.4 (Stein–Weiss vector fields). If F : D → V = 41 V is a vector field in a general Euclidean space, then F is monogenic if and only if F is divergenceand curl-free. Thus, for vector fields the equation DF = 0 is equivalent to the first-order system ( div F (x) = 0, curl F (x) = 0. This is a consequence of the fact that ∇ y F : D → 40 V and ∇ ∧ F : D → 42 V , where 40 V ∩ 42 V = {0}. We note that Example 8.1.4 generalizes as follows. When F : D → 4k V is a homogeneous multivector field, then ∇ 4 F = 0 if and only if ( ∇ y F (x) = 0, ∇ ∧ F (x) = 0. The following proposition shows when all the homogeneous parts of a monogenic field are themselves monogenic. Proposition 8.1.5 (Two-sided monogenicity). Let F : D → 4V be a C 1 multivector field, and write F = F0 + · · · + Fn , where Fj : D → 4j V . Then the following are equivalent. (i) All the homogeneous parts Fj are monogenic fields. (ii) The field F satisfies both ∇ ∧ F = 0 and ∇ y F = 0. (iii) The field F is two-sided monogenic, that is, ∇4F = 0 and F 4∇ = ei∗ = 0.

P

∂i F (x)4

Proof. (i) implies (iii): If Fj is monogenic, then ∇ ∧ Fj = 0 = ∇ y Fj as above, and therefore ∇ 4 Fj as well as Fj 4 ∇ = (−1)j (∇ ∧ Fj − ∇ y Fj ) is zero. Adding up all Fj proves (iii). (iii) implies (ii): this is a consequence of the Riesz formulas, which show that and ∇ y F = 12 (∇ 4 F + F\ 4 ∇) 4 ∇). ∇ ∧ F = 21 (∇ 4 F − F\ (ii) implies (i): If ∇ ∧ F = 0, then 0 = (∇ ∧ F )j+1 = ∇ ∧ Fj for all j, since d maps j-vector fields to (j + 1)-vector fields. Similarly ∇ y Fj = 0 for all j. Thus ∇ 4 Fj = ∇ y Fj + ∇ ∧ Fj = 0. We next consider the fundamental solution for the Dirac operator. In order to apply the Fourier transform componentwise as in Section 6.2, we complexify the Clifford algebra 4V ⊂ 4Vc . We note that the exterior, interior, and Clifford derivatives are the Fourier multipliers \ dF (x) = iξ ∧ Fˆ (ξ), \ δF (x) = iξ y Fˆ (ξ), \ DF (x) = iξ 4 Fˆ (ξ),

and ξ ∈ V.

Chapter 8. Hypercomplex Analysis

260

From this it follows that unlike d and δ, the Dirac operator D is elliptic and has a fundamental solution Ψ(x) with Fourier transform ξ ˆ = (iξ)−1 = −i 2 . Ψ(ξ) |ξ| Using the formula for the fundamental solution Φ to the Laplace operator ∆ ˆ from Example 6.3.1, where Φ(ξ) = −1/|ξ|2 , we obtain the following formula for Ψ(x) = ∇Φ. Note that unlike the situation for Φ, the two-dimensional case does not use any logarithm. Definition 8.1.6 (Fundamental solution). The fundamental solution to the Dirac operator D in an n-dimensional Euclidean space with origin fixed, n ≥ 1, is the vector field 1 x Ψ(x) := , σn−1 |x|n R where σn−1 := |x|=1 dx = 2π n/2 /Γ(n/2) is the measure of the unit sphere in V , R ∞ −t z−1 and Γ(z) := 0 e t dt is the gamma function, with Γ(k) = (k − 1)!. Exercise 8.1.7. Show by direct calculation that ∇ ∧ Ψ(x) = 0 and ∇ y Ψ(x) = δ0 (x) in the distributional sense in V , where δ0 (x) is the Dirac delta distribution. The following application of the general Stokes theorem is central to hypercomplex analysis. Theorem 8.1.8 (Cauchy–Pompeiu formula for D). Let D be a bounded C 1 -domain in Euclidean space X. If F ∈ C 1 (D; 4V ), then Z Z F (x) + Ψ(y − x)(DF )(y)dy = Ψ(y − x)ν(y)F (y) dy, D

∂D

for all x ∈ D, where ν(y) denotes the outward-pointing unit normal vector field on ∂D. In particular, for monogenic multivector fields F , we have the Cauchy reproducing formula Z F (x) = Ψ(y − x)ν(y)F (y) dy, x ∈ D. (8.1) ∂D

Proof. For fixed x ∈ D, consider the linear 1-form θ(y, v) := Ψ(y − x) 4 v 4 F (y). For y ∈ D \ {x}, its nabla derivative is θ(y, ˙ ∇) =

n X

∂yi Ψ(y − x) 4 ei 4 F (y)

i=1

= (Ψ(y˙ − x) 4 ∇) 4 F (y) + Ψ(y − x) 4 (∇ 4 F (y)) ˙ = Ψ(y − x) 4 (DF )(y),

8.1. Monogenic Multivector Fields

261

by associativity of the Clifford product and since Ψ 4 ∇ = ∇ y Ψ − ∇ ∧ Ψ = 0 by Exercise 8.1.7. To avoid using distribution theory, we consider the domain D := D \ B(x, ), obtained by removing a small ball around x. On ∂B(x, ) the outward-pointing unit normal relative D is (x − y)/|x − y|. The Stokes formula (7.4) gives Z Ψ(y − x)(DF )(y)dy D Z Z x−y = Ψ(y − x) Ψ(y − x)ν(y)F (y) dy F (y) dy + |x − y| ∂B(x,) ∂D Z Z 1 y−x 1 =− F (y) dy + ν(y)F (y) dy. n−1 σ σ |y − x|n n−1 n−1 ∂B(x,) ∂D Upon taking limits → 0, the first term on the right-hand side will converge to −F (x), and the Cauchy–Pompeiu formula follows. Exercise 8.1.9 (Cauchy integral theorem). Apply Stokes’s theorem and prove the general Cauchy theorem Z G(y) 4 ν(y) 4 F (y) dy = 0 ∂D

for a left monogenic field F and a right monogenic G, that is,R ∇ 4 F = 0 = G(x) ˙ 4∇ = 0, in D. Deduce from this the classical Cauchy theorem ∂D f (w)dw = 0 for an analytic function f from complex analysis. See Example 7.3.12. Example 8.1.10 (Cauchy formula in C). The Cauchy formula for analytic functions in the complex plane is a special case of Theorem 8.1.8. To see this, consider an analytic function f (z) in a plane domain D. As in Example 8.1.2, we identify the vector x ∈ V and the complex number z = e1 x ∈ C = 4ev V , y ∈ V with w = e1 y ∈ C, and the normal vector ν with the complex number n = e1 ν. If f (z) : D → C = 4ev V is analytic, thus monogenic, and if x ∈ D, then Theorem 8.1.8 shows that Z 1 e1 (w − z) f (z) = (e1 n(w))f (w)|dw| σ1 ∂D |w − z|2 Z Z e1 (w − z)e1 1 f (w)dw 1 = = f (w)(jn(w)|dw|) . 2 2πj ∂D |w − z| 2πj ∂D w − z Here we have used that e1 (w − z)e1 = w − z, that complex numbers commute, and that jn is tangent to a positively oriented curve. We have written |dw| for the scalar length measure on ∂D. Note that unlike the situation for analytic functions, in higher dimensions the normal vector must be placed in the middle, between the fundamental solution and the monogenic field. This is because the Clifford product is noncommutative. For

Chapter 8. Hypercomplex Analysis

262

analytic functions, the normal infinitesimal element ν(y)dy corresponds to dw/j, which can be placed, for example, at the end of the expression, since complex numbers commute. As in the complex plane, the Cauchy formula for monogenic fields has a number of important corollaries, of which we next consider a few. Corollary 8.1.11 (Smoothness). Let F : D → 4V be a monogenic field in a domain D. Then F is real analytic, and in particular a C ∞ -regular field. Proof. Fix a ball B(x0 , ) such that B(x0 , 2) ⊂ D. Then Z 1 y−x F (x) = ν(y)F (y) dy, for all x ∈ B(x0 , ). σn−1 |y−x0 |= |y − x|n The stated regularity now follows from that of the fundamental solution x 7→ y−x |y−x|n . We also obtain a Liouville theorem for entire monogenic fields. Corollary 8.1.12 (Liouville). Let F : X → 4V be an entire monogenic field, that is monogenic on the whole Euclidean space X. If F is bounded, then F is a constant field. Proof. Let x0 ∈ X. For all R > 0 and 1 ≤ k ≤ n we have Z 1 y − x ∂xk ν(y)F (y) dy. ∂k F (0) = σn−1 |y−x0 |=R |y − x|n x=0 If F is bounded, the triangle inequality for integrals shows that Z |∂k F (0)| . R−n dy . 1/R. |y|=R

Taking limits as R → 0 shows that ∂k F (x0 ) = 0 for all k. Since x0 was arbitrary, F must be constant. Next we consider what further properties monogenic fields do and do not share with analytic functions. In contrast to analytic functions, monogenic fields do not form a multiplication algebra, that is, F (x) and G(x) being monogenic in D does not imply that x 7→ F (x) 4 G(x) is monogenic. The obstacle here is the noncommutativity of the Clifford product, which causes D(F G) = (DF )G + DF G˙ 6= (DF )G + F (DG). Although monogenic fields in general cannot be multiplied to form another monogenic field, we can do somewhat better than the real linear structure of monogenic fields. Recall that analytic functions form a complex linear space. This generalizes to monogenic fields as follows.

8.1. Monogenic Multivector Fields

263

Proposition 8.1.13 (Right Clifford module). Let D be an open set in a Euclidean space X. Then the monogenic fields in D form a right Clifford module, that is, if F (x) is monogenic, then so is x 7→ F (x) 4 w for every constant w ∈ 4V . Proof. This is a consequence of the associativity of the Clifford product, since ∇ 4 (F (x) 4 w) = (∇ 4 F (x)) 4 w = 0 4 w = 0.

In contrast to analytic functions, monogenic functions do not form a group under composition, that is, F (y) and G(x) being monogenic in appropriate domains does not imply that x 7→ F (G(x)) is monogenic. Indeed, in general the composition is not even well defined, since the range space 4V is not contained in V . Although it does not make sense to compose monogenic fields, the situation is not that different in higher dimensions. Recall that in the complex plane, analytic functions are the same as conformal maps, at least for functions with invertible derivative. In higher dimensions one should generalize so that the inner function G is conformal and the outer function F is monogenic. In this way, we can do the following type of conformal change of variables that preserves monogenicity. Sections 4.5 and 11.4 are relevant here. Proposition 8.1.14 (Conformal Kelvin transform). Let T x = (ax + b)(cx + d)−1 be a fractional linear map of a Euclidean space X = V , and let D ⊂ X be an open set such that ∞ ∈ / T (D). For a field F : T (D) → 4V , define a pulled back field KTm F : D → 4V : x 7→ Then D(KTm F )(x) =

cx + d |cx + d|n

4

F (T (x)).

det4 (T ) m K (DF )(x), |cx + d|2 T

x ∈ D,

where det4 (T ) := ad − bc. In particular, if F is monogenic, then so is KTm F . Proof. Applying the product rule as in Example 7.1.9 shows that ! n X x + c−1 d c cx + d ∇ 4 (KT F )(x) = ∇ + ei F (T (x)) ∂i (F (T (x))). |x + c−1 d|n |c|n |cx + d|n i=1 The first term is zero, since the fundamental solution is monogenic outside the origin. For the second term, we note that since T is conformal, the derivative T x will map the ON-basis {ei } onto a basis {e0i = T x ei } of orthogonal vectors of equal length. By Exercise 4.5.18, we have e0i = (det4 (T )/cx + d)ei (cx + d)−1 . The dual basis is seen to be e0∗ i = ((cx + d)/det4 (T ))ei cx + d,

Chapter 8. Hypercomplex Analysis

264 so that

ei cx + d = (det4 (T )/(cx + d))e0∗ i . According to the chain rule, the directional derivatives are ∂ei (F (T (x))) = (∂e0i F )(T (x)), so ∇ 4 (KTm F )(x) =

n det4 (T ) cx + d X 0∗ det4 (T ) m e (∂e0i F )(T (x)) = K (DF )(x). |cx + d|2 |cx + d|n i=1 i |cx + d|2 T

Specializing to the inversion change of variables T x = 1/x, we make the following definition. Definition 8.1.15 (Kelvin transform). The monogenic Kelvin transform of a field F : D → 4V is the field K m F (x) :=

x |x|n

4

F (1/x).

Similarly, using the fundamental solution for the Laplace operator, we define the harmonic Kelvin transform of a function u : D → R to be K h u(x) := |x|2−n u(1/x). For the monogenic Kelvin transform, we have shown that DK m = −|x|−2 K m D. We now use this to obtain a similar result for the harmonic Kelvin transform. Proposition 8.1.16. The harmonic Kelvin transform satisfies the commutation relation ∆(K h u)(x) = |x|−4 K h (∆u)(x). In particular, the Kelvin transform of an harmonic function is harmonic. Proof. We note that ∆ = D2 and K h u = xK m u. Thus ∆K h u = DD(xK m u) = D(nK m u + (2∂x − xD)K m u) = nDK m u + 2(DK m u + ∂x DK m u) − nDK m u − (2∂x − xD)DK m u = 2DK m u + xD2 K m u = −2|x|−2 K m Du − xD|x|−2 K m Du = −x−1 DK m Du = x−1 |x|−2 K m D2 u = |x|−4 K h ∆u. Pn Here ∂x f = j=1 xj ∂j f = hx, ∇if denotes the radial directional derivative, and Pn we have used that D∂x f = j=1 (∇xj )∂j f + ∂x Df = (1 + ∂x )Df by the product rule.

8.2. Spherical monogenics

265

Exercise 8.1.17. Consider the special case of Proposition 8.1.14 in which ρ(x) = T x = qbxq −1 is an isometry, where q ∈ Pin(V ). Investigate how the conformal Kelvin transform KTm F , the pullback ρ∗ F , the pushforward ρ∗−1 F and the normalized pushforward ρ˜∗−1 F are related. Show that all these four fields are monogenic whenever F is monogenic, and relate this result to Proposition 8.1.13.

8.2

Spherical monogenics

In our n-dimensional Euclidean space X = V with a fixed origin, we denote by S := {x ∈ X ; |x| = 1} the unit sphere. We generalize the well-known theory of Taylor series expansions of analytic functions in the plane. When n = 2, we know that a function analytic at 0 can be written as a convergent power series f (x + iy) =

∞ X

Pk (x, y),

x2 + y 2 < 2 ,

k=0

for some > 0, where Pk ∈ Pkm := {ak (x + iy)k ; ak ∈ C}. A harmonic function can be written in the same way if we allow terms Pk ∈ Pkh := {ak (x + iy)k + bk (x − iy)k ; ak , bk ∈ C}. Note that P0h and all Pkm are one-dimensional complex linear spaces, whereas Pkh are two-dimensional when k ≥ 1. The spaces Pkm and Pkh are subspaces of the space Pk of all polynomials of order k, which has dimension k + 1. A polynomial P ∈ Pk is in particular homogeneous of degree k in the sense that P (rx) = rk P (x),

for all r > 0, x ∈ R2 .

This shows that P is uniquely determined by its restriction to the unit circle |x| = 1 if the degree of homogeneity is known. In the power series for f , the term Pk describes the kth-order approximation of f around the origin. Next consider an n-dimensional space X, and the following generalization of the spaces above. Definition 8.2.1 (Spherical harmonics and monogenics). Let X = V be a Euclidean space, and let k ∈ N and s ∈ R. Define function spaces Pk := {P : X → 4V ; all component functions are homogeneous polynomials of degree k}, m Ps := {F : X \ {0} → 4V ; DF = 0, F (rx) = rs F (x), x 6= 0, r > 0}, Psh := {F : X \ {0} → 4V ; ∆F = 0, F (rx) = rs F (x), x 6= 0, r > 0}.

Chapter 8. Hypercomplex Analysis

266

Let Pks ⊂ Pk and Pssh ⊂ Psh be the subspaces of scalar functions F : X \ {0} → 40 V = R, and let Psem ⊂ Psm be the subspace of functions F : X \ {0} → 4ev V that take values in the even subalgebra. Denote by Psh (S) the space of restrictions of functions P ∈ Psh to the unit sphere S. Denote by Psm (S) the space of restrictions of functions P ∈ Psm to the unit sphere S. We refer to these functions as (multivector-valued) spherical harmonics and spherical monogenics respectively. Note that the spaces Pk and Psh essentially are spaces of scalar functions: Each function in these spaces has component functions that belong to the same space, since the conditions on the function do not involve any coupling between the component functions. Even if the definitions of Psm and Psh are quite liberal, these are essentially spaces of polynomials, as the following shows. Proposition 8.2.2. Let n := dim X. The monogenic space Psm contains nonzero functions only if s ∈ {. . . , −(n + 1), −n, −(n − 1), 0, 1, 2, . . .}. The harmonic space Psh contains nonzero functions only if s ∈ {. . . , −(n + 1), −n, −(n − 1), −(n − 2), 0, 1, 2, . . .}. If k ∈ N, then Pkm ⊂ Pkh ⊂ Pk . The Kelvin transforms give self-inverse one-to-one correspondences m , K m : Psm → P−(s+n−1)

h K h : Psh → P−(s+n−2) .

Proof. (i) First consider the monogenic spaces Psm . Apply the Cauchy formula (8.1) to P ∈ Psm in the domain D := B(0; 1) \ B(0; ) for fixed 0 < < 1. For x ∈ D, we have Z Z Ψ(y − x)yP (y)dy + Ψ(y − x)ν(y)P (y)dy. P (x) = S

|y|=

For fixed x 6= 0, the second integral is dominated by n−1 sup|y|= |P |. Letting → 0, this tends to zero if s > −(n − 1), and it follows that 0 is a removable singularity of P (x). If −(n − 1) < s < 0, Liouville’s Theorem 8.1.12 shows that P = 0. Furthermore, generalizing the proof of Liouville’s theorem by applying higher-order derivatives shows that if s ≥ 0, then P (x) must be a polynomial. Thus Psm 6= {0} m is bijective and self-inverse is straightonly if s ∈ N. That K m : Psm → P−(s+n−1) forward to verify. m (ii) Next consider the harmonic spaces Psh . If P ∈ Psh , then DP ∈ Ps−1 . If m s∈ / Z or −(n − 2) < s < 0, then (i) shows that DP = 0, so that P ∈ Ps . Again by (i), we conclude that P = 0. If s ∈ N, then the same argument shows that DP is a polynomial. Here we may assume that P is scalar-valued, so that DP = ∇P . h Integrating we find that P is a polynomial as well. That K h : Psh → P−(s+n−2) is bijective and self-inverse is straightforward to verify.

267

8.2. Spherical monogenics

We next examine the finite-dimensional linear spaces Pkm and Pkh for k ∈ N. m and As we have seen, this also gives information about the spaces P−(k+n−1) h P−(k+n−2) via the Kelvin transforms. Note that unlike the situation in the plane, there is a gap −(n − 1) < s < 0 and −(n − 2) < s < 0 respectively between the nonzero spaces, and that this gap grows with dimension. A polynomial P (x) ∈ Pk , can be written X X P (x) = Pαs xα es . s⊂n α∈Nk αk 1 Here we use multi-index notation xα = x(α1 ,...,αk ) := xα 1 · · · xk , and we shall write δi := (0, . . . , 0, 1, 0, . . . , 0), where 1 is the ith coordinate. We introduce an auxiliary inner product XX hP, Qip := α!Pαs Qαs , s

α

where α! = (α1 , . . . , αk )! := α1 ! · · · αk !. Proposition 8.2.3. With respect to the inner product h·, ·ip on Pk , we have orthogonal splittings Pk = Pkm ⊕ xPk−1 , Pk = Pkh ⊕ x2 Pk−2 , where xPk−1 := {x 4 P (x) ; P ∈ Pk−1 }, as well as m Pkh = Pkm ⊕ xPk−1 ,

k ≥ 1,

P0h = P0m .

Proof. (i) The key observation is that Pk → Pk−1 : P (x) 7→ ∇ 4 P (x)

and Pk−1 → Pk : P (x) 7→ x 4 P (x)

are adjoint maps with respect to h·, ·ip .P In fact, the inner product P is designed for this purpose. To see this, write P (x) = s,α Pα,s xα es and Q(x) = t,β Qβ,t xβ et . Pn P Then ∇ 4 P (x) = i=1 s,α Pα,s αi xα−δi (i, s)ei4s , so that X h∇ 4 P, Qip = Pα,s αi (i, s)Qβ,t hxα−δi ei4s , xβ et i i,s,α,t,β

=

X

Pα,s αi (i, s)Qα−δi ,i4s (α − δi )!

i,s,α

Pn P On the other hand, x 4 Q(x) = i=1 t,β Qβ,t xβ+δi (i, t)ei4t , so that X hP, x 4 Qip = Pα,s Qβ,t (i, t)hxα es , xβ+δi ei4t i i,s,α,t,β

=

X i,s,α

Pα,s Qα−δi ,i4s (i, i 4 s)α!.

Chapter 8. Hypercomplex Analysis

268

Since αi (α − δi )! = α! and (i, s) = (i, i 4 s), the duality follows. (ii) We note that Pkm = N(∇) and that xPk−1 = R(x). Since the maps are adjoint, these subspaces are orthogonal complements in Pk . Similarly, Pk = m Pkh ⊕x2 Pk−2 , since (∇2 )∗ = x2 . Finally, we consider the map Pkh → Pk−1 : P (x) 7→ ∇ 4 P (x). This is well defined, since ∇ 4 P is monogenic if P is harmonic. The m adjoint operator will be Pk−1 → Pkh : Q(x) 7→ x 4 Q(x), provided x 4 Q is harmonic whenever Q is monogenic. To verify that this is indeed the case, we calculate as in the proof of Proposition 8.1.16 that D2 (xQ) = D(nQ + (2∂x − xD)Q) = nDQ + 2(D + ∂x D)Q − (nDQ + (2∂x − D)DQ) = (2 + D)DQ. m This proves that Pkm = N(∇) is the orthogonal complement to xPk−1 = R(x) in h Pk .

Corollary 8.2.4 (Dimensions). Let X be an n-dimensional Euclidean space. Then k+n−1 , dim Pk = 2n dim Pks , dim Pks = n−1 s dim Pkem = 2n−1 (dim Pks − dim Pk−1 ), s , dim Pksh = dim Pks − dim Pk−2

dim Pkm = 2 dim Pkem ,

dim Pkh = 2n dim Pksh .

Proof. To find dim Pks , note that this is the number of monomials of degree k in n variables. The standard combinatorial argument is as follows. Choose n − 1 of the numbers 1, 2, 3, . . . , k + n − 1, ways. say 1 ≤ m1 < m2 < · · · < mn−1 ≤ k + n − 1. This can be done in k+n−1 n−1 Such choices {mi } are in one-to-one correspondence with monomials 3 −m2 −1 · · · xnk+n−1−mm−1 . x1m1 −1 x2m2 −m1 −1 xm 3

From Proposition 8.2.3 the remaining formulas follow.

Exercise 8.2.5 (Two and three dimensions). Let V be a two-dimensional Euclidean space. In this case dim Pksh = 2 = dim Pkem . Show that dim Pkem is a one-dimensional complex linear space with the geometric complex structure j = e12 ∈ 42 V . Find bases for these spaces using the complex powers z k = (x + jy)k . Identifying vectors and complex numbers as in Section 3.2, write the splitting m Pkh = Pkm ⊕ xPk−1 in complex notation. Let V be a three-dimensional Euclidean space. In this case, dim Pksh = 2k + 1 and dim Pkem = 4(k + 1). Find bases for the spherical harmonics Pksh and for the spherical monogenics Pkem . Note that Pkem is a right vector space over H of dimension k + 1.

8.2. Spherical monogenics

269

Recall from Fourier analysis that the trigonometric functions {eikθ }k∈Z , suitably normalized, form an ON-basis for L2 (S) on the unit circle S ⊂ C = R2 in the complex plane. Thus every f ∈ L2 (S) can be uniquely written ∞ X

f (eiθ ) =

ak eikθ .

k=−∞

For k ≥ 0, the function eikθ extends to the analytic function z k on the disk |z| < 1. For k < 0, the function eikθ extends to the analytic function z k on |z| > 1, which vanishes at ∞, or alternatively to the antianalytic and harmonic function z −k on |z| < 1. In higher dimensions, we have the following analogue. Theorem 8.2.6. Let S be the unit sphere in an n-dimensional Euclidean space. The subspaces Pkh (S), k = 0, 1, 2, . . ., of spherical harmonics are pairwise orthogonal with respect to the L2 (S) inner product Z hF, Gi := hF (x), G(x)idx. S m Moreover, within each Pkh (S), the two subspaces Pkm (S) and xPk−1 (S) are orm m thogonal, and xPk−1 (S) = P2−n−k (S). The Hilbert space L2 (S) splits into finitedimensional subspaces as

L2 (S) =

∞ M

Pkh (S) =

k=0

∞ M

−(n−1)

Pkm (S) ⊕

k=0

M

Pkm (S).

k=−∞

Proof. Let P ∈ Pkh (S) and Q ∈ Plh (S) with k 6= l. Green’s second theorem, as in Example 7.3.11, shows that Z Z h∂x P (x), Q(x)i − hP (x), ∂x Q(x)i dx = h∆P, Qi − hP, ∆Qi dx = 0. |x| 0. For the two spherical operators we have the following. Proposition 8.2.15. Let V be an n-dimensional Euclidean space, and consider the Hilbert space L2 (S) on the unit sphere S. Then DS defines a self-adjoint operator in L2 (S) with spectrum σ(DS ) = Z \ {−(n − 2), . . . , −1}. The spherical Laplace operator equals ∆S = DS (2 − n − DS ). h In the splitting into spherical harmonics, L2 (S) = ⊕∞ k=0 Pk (S), the spherical Laplace operator acts according to

∆S

∞ X k=0

∞ X fk = k(2 − n − k)fk , k=0

8.2. Spherical monogenics

275

whereas in the splitting into spherical monogenics, L2 (S) =

∞ M

−(n−1)

Pkm (S)

k=0

M

⊕

Pkm (S),

k=−∞

the spherical Dirac operator acts according to DS

∞ X k=0

−(n−1)

fk +

X

−(n−1) ∞ X X fk = kfk . kfk +

k=−∞

k=0

k=−∞

Proof. It remains to prove that ∆S = DS (2 − n − DS ). Using polar coordinates x = ry, y ∈ S, we note that D = r−1 yxD = r−1 y(∂x − DS ) = y∂r − r−1 yDS . Squaring this Euclidean Dirac operator, we get ∆ = D2 = (y∂r − r−1 yDS )2 = y∂r y∂r − y∂r r−1 yDS − r−1 yDS y∂r + r−1 yDS r−1 yDS = ∂r2 − ∂r r−1 DS − r−1 yDS y∂r + r−2 yDS yDS . Writing [A, B] = AB − BA for the commutator of operators, we have used that [∂r , y] = 0 and [DS , r] = 0. To simplify further, we compute that [∂r , r] = 1 and [∂x , DS ] = [∂x , ∂x − xD] = −[∂x , xD] = 0. Thus ∂r r−1 DS = −r−2 DS + r−1 DS ∂r , so that ∆ = ∂r2 − r−1 (DS + yDS y)∂r + r−2 (DS + yDS yDS ). Comparing this equation and (8.2), we see that n − 1 = −(DS + yDS y) and ∆S = DS + yDS yDS = DS + (1 − n − DS )DS = DS (2 − n − DS ), as claimed.

In three dimensions, it is standard to introduce spherical coordinates (r, θ, φ), and DS and ∆S can be expressed in terms of ∂θ and ∂φ . The classical expression for the spherical harmonics, obtained by separation of variables, is rk Pkm (cos θ)eimφ , where the Pkm (t) denote the associated Legendre polynomials, m = −k, . . . , −1, 0, 1, . . . , k. The optimal parametrization of the sphere S, though, uses stereographic projection, which is conformal and has only one singular point for the coordinate system.

Chapter 8. Hypercomplex Analysis

276

Proposition 8.2.16 (Stereographic projection of DS ). Fix an (n − 1)-dimensional subspace VS ⊂ V and a point p ∈ S orthogonal to VS , and consider the stereographic parametrization T : VS → S : y 7→ T (y) = (py + 1)(y − p)−1 , as in (4.4). The monogenic Kelvin transform associated to the stereographic projection T defines an isometry of Hilbert spaces 2(n−1)/2 KTm : L2 (S) → L2 (VS ), and the spherical Dirac operator corresponds to (KTm DS (KTm )−1 )G(y) = − 21 p (|y|2 + 1)Dy + (y − p) G(y),

y ∈ VS ,

where Dy denotes the Dirac operator in the Euclidean space VS . Proof. According to Proposition 8.1.14, the Kelvin transform y−p KTm F (y) = F ((py + 1)(y − p)−1 ) |y − p|n satisfies D(KTm F )(y) = −2|y − p|−2 KTm (DF )(y). From the definition of DS we get KTm (DS F ) = KTm (∂x F ) − KTm (xDF ), where KTm (xDF )(y) =

y−p (py + 1)(y − p)−1 (DF )(T (y)) |y − p|n

= (y − p)−1 (py + 1)KTm (DF )(y) = − 12 (y − p)(py + 1)D(KTm F )(y) = 12 (1 + |y|2 )pD(KTm F )(y). To rewrite KTm (∂x F ), we observe that the vertical derivative of the stereographic parametrization at y ∈ VS is T y (p) =

−2 2 (y − p)p(y − p)−1 = x 2 1 + |y| 1 + |y|2

(8.3)

according to Exercise 4.5.18. Thus the chain and product rules give y−p y − p 1 + |y|2 = (∂ F )(T (y)) ∂yn (F (T (y)) x |y − p|n |y − p|n 2 1 + |y|2 p m = ∂yn KT F (y) − F (T (y)) 2 |y − p|n 1 + |y|2 p = ∂yn KTm F (y) − (y − p)KTm F (y). 2 2 Here ∂yn is the partial derivative in the direction p. Since pD = ∂yn + pDy , we obtain the stated formula. To show that the stated map is a Hilbert space isometry, note that by (8.3) the Jacobian is JT (y) = (2/(1 + |y|2 ))n−1 , since T is conformal. Thus Z Z Z 2n−1 dy 2 2 n−1 |F (x)| dx = |F (T (y))| =2 |KTm F (y)|2 dy. (1 + |y|2 )n−1 S VS VS KTm (∂x F )(y) =

8.3. Hardy Space Splittings

277

8.3 Hardy Space Splittings Let D = D+ be a bounded Lipschitz domain in Euclidean space X, with boundary ∂D separating it from the exterior unbounded domain D− = X \ D. Let ν denote the unit normal vector field on ∂D pointing into D− . The main operator in this section is the principal value Cauchy integral Z Z Eh(x) := 2p.v. Ψ(y − x)ν(y)h(y) dy = 2 lim Ψ(y − x)ν(y)h(y) dy, →0

∂D

∂D\B(x;)

x ∈ ∂D, which appears when we let x ∈ ∂D, rather than x ∈ D, in the Cauchy reproducing formula from Theorem 8.1.8. Here we assume only suitable bounds on h, and in particular we do not assume that h is a restriction of a monogenic field. The factor 2 is a technicality that will ensure that E 2 = I. The singularity at y = x in the integral is of order |y − x|1−n on the (n − 1)-dimensional surface ∂D, which makes the definition and boundedness of E a nontrivial matter, and cancellations need to be taken into account. Due to the strong singularity at y = x, we also refer to E as the Cauchy singular integral. Ignoring these analytic problems for the moment, we first investigate by formal calculations how E is related to the two limits Z + E h(x) := lim Ψ(y − z)ν(y)h(y) dy, x ∈ ∂D, z∈D + ,z→x

and

∂D

Z

E − h(x) :=

lim −

z∈D ,z→x

Ψ(y − z)(−ν(y))h(y) dy,

x ∈ ∂D,

∂D

in the Cauchy reproducing formula (8.1) for D+ and D− respectively. Placing z = x infinitesimally close, but interior, to ∂D, we have for E + ! Z Z + E h(x) = lim Ψ(y − x)ν(y)h(x)dy + Ψ(y − x)ν(y)h(y)dy →0

=

1 2 h(x)

Σ0x

+

Σ1x

1 2 Eh(x).

where Σ0x := {y ∈ D− ; |y − x| = } and Σ1x := {y ∈ ∂D ; |y − x| > }. We have here approximated h(y) ≈ h(x), changed the integration surface from ∂D \ Σ1x to Σ0x using Stokes’s theorem, and used that Ψ(y − x)ν(y) = 1/(σn−1 n−1 ) on Σ0x in the first integral. Thus the first term h/2 appears when we integrate around the singularity y = x on an infinitesimal half-sphere. Since −ν is outward pointing from D− , a similar formal calculation indicates that E − h(x) = 12 h(x) − 12 Eh(x), and we deduce operator relations 1 2 (I 1 2 (I +

+ E) = E + , − E) = E − ,

E + E − = I, E + − E − = E.

278

Chapter 8. Hypercomplex Analysis Moreover, from Theorem 8.1.8 we conclude that E+E+ = E+, E−E− = E−,

since E ± h by definition is the restriction of a monogenic field to ∂D, no matter what h is. This shows that E + and E − are complementary projection operators. For a suitable space of multivector fields H on ∂D, these projections define a splitting H = E + H ⊕ E − H. This means that any given field h on ∂D can be uniquely written as a sum h = h+ +h− , where h+ is the restriction to ∂D of a monogenic field in D+ and h− is the restriction to ∂D of a monogenic field in D− that decays at ∞. We refer to E ± H as Hardy subspaces, and to E ± as Hardy projections. Note also the structure of the Cauchy singular integral operator E = E + − E − : it reflects the exterior Hardy subspace E − H across the interior Hardy subspace E + H. In particular, E 2 = I, as claimed.

Figure 8.1: (a) The piecewise constant vector field h : ∂D → 41 R2 which equals e1 in the second quadrant and vanishes on the rest of the curve ∂D. (b) The Hardy splitting of h as the sum of two traces of divergence- and curl-free vector fields.

Example 8.3.1 (Constant-curvature boundaries). The most natural space for the singular integral operator E is H = L2 (∂D). In the simplest case, in which D is the upper complex half-plane, with ∂D the real axis, then E is a convolution singular integral, which under the Fourier transform corresponds to multiplication by ( 1, ξ > 0, sgn(ξ) = −1, ξ < 0,

8.3. Hardy Space Splittings

279

at least if h takes values in the even subalgebra and we use the geometric imaginary unit j as in Example 8.1.10. The second simplest example is that in which L∞ D is the unit ball |x| < 1 as in Theorem 8.2.6. In this case, E + projects onto k=0 Pkm (S), whereas E − projects L−(n−1) onto k=−∞ Pkm (S). In these examples the Hardy subspaces are orthogonal and kE ± k = 1. However, unlike Hodge splittings, the splitting into Hardy subspaces is not orthogonal for more general domains D. When ∂D has some smoothness beyond Lipschitz, Fourier methods apply to prove that E is a bounded operator on L2 (∂D), which geometrically means that the angle between the Hardy subspaces, although not straight, is always positive. A breakthrough in modern harmonic analysis was the discovery that this continues to hold for general Lipschitz domains. Theorem 8.3.2 (Coifman–McIntosh–Meyer). Let D be a bounded strongly Lipschitz domain. Then the principal value Cauchy integral Eh(x) of any h ∈ L2 (∂D) is well defined for almost every x ∈ ∂D, and we have bounds Z Z |h(x)|2 dx. |Eh(x)|2 dx . ∂D

∂D

This is a deep result that is beyond the scope of this book. There exist many different proofs. A singular integral proof is to estimate the matrix for E in a wavelet basis for L2 (∂D) adapted to ∂D. A spectral proof is to identify E ± as spectral projections of a Dirac-type operator on ∂D, generalizing the spherical Dirac operator DS from Definition 8.2.13. The problem is that for general domains this operator is no longer self-adjoint, but rather has spectrum in a double sector around R, and it becomes a nontrivial matter involving Carleson measures to estimate the spectral projections corresponding to the two sectors. See Section 8.4 for references and further comments. We remark only that from Theorem 8.3.2 one can prove that for h ∈ L2 (∂D), the Cauchy extensions Z + (8.4) Ψ(y − x)ν(y)h(y) dy, x ∈ D+ , F (x) := ∂D

and −

Z Ψ(y − x)(−ν(y))h(y) dy,

F (x) :=

x ∈ D− ,

(8.5)

∂D

have limits as x → ∂D both in an L2 (∂D) sense and pointwise almost everywhere, provided that we approach ∂D in a nontangential way. In the remainder of this section, we perform a rigorous analysis of the splitting of the space of H¨older continuous multivector fields C α (∂D) = C α (∂D; 4V ),

0 < α < 1,

from Example 6.4.1, into Hardy subspaces on a bounded C 1 surface ∂D. This setup is a good starting point for studying Hardy splittings that only requires

Chapter 8. Hypercomplex Analysis

280

straightforward estimates. We exclude the endpoint cases α = 0, continuous functions, and α = 1, Lipschitz continuous functions, for the reason that typically singular integral operators like E are not bounded on these spaces. It is also not bounded on L∞ (∂D), something that could be seen in Figure 8.1 if we zoom in at the discontinuities of h. Proposition 8.3.3 (Hardy projection bounds). Let D be a bounded C 1 domain and 0 < α < 1, and assume that h ∈ C α (∂D). Define the Cauchy extensions F ± in D± as in (8.4) and (8.5). Then F + is a monogenic field in D+ , and F − is a monogenic field in D− with decay F − = O(|x|−(n−1) ) at ∞. At the boundary ∂D, the traces f + (y) :=

lim

x∈D + ,x→y

F + (x),

f − (y) :=

lim

x∈D − ,x→y

F − (x),

y ∈ ∂D,

exist, with estimates kf + kα . khkα and kf − kα . khkα . In terms of operators, this means that the Hardy projections E ± : h 7→ f ± are bounded on C α (∂D). Proof. (i) That F + and F − are monogenic is a consequence of the associativity of the Clifford product. Indeed, applying the partial derivatives under the integral sign shows that Z + ∇x 4 Ψ(y − x) 4 ν(y) 4 h(y) dy ∇ 4 F (x) = Z∂D ∇x 4 (Ψ(y − x) 4 ν(y) 4 h(y)dy = 0, = ∂D

/ ∂D. The decay at infinity follows from the fact that ∂D and h are when x ∈ bounded and the decay of the fundamental solution Ψ. (ii) We next consider the boundary trace of F + . A similar argument applies to the trace of F − . Note that in order to estimate kf + kα , it suffices to estimate |f + (x) − f + (y)| . |x − y|α khkα for |x − y| ≤ δ, provided that |f + (x)| . khkα for all x ∈ ∂D, since ∂D is bounded. Thus we may localize to a neighborhood of a point p ∈ ∂D, in which we can assume that ∂D coincides with the graph of a C 1 -function φ. We choose a coordinate system {xi } so that p is the origin and ∂D is given by xn = φ(x0 ), where x0 = (x1 , . . . , xn−1 ), in the cylinder |x0 | < r, |xn | < s. Let δ < min(r, s) and consider a point x = (x0 , xn ) ∈ D ∩ B(0, δ). We claim that (8.6) |∂j F + (x)| . khkα (xn − φ(x0 ))α−1 , j = 1, . . . , n. To show this, considerR the vertical projection z = (x0 , φ(x0 )) of R x onto ∂D, and note that F + (x) − h(z) = ∂D Ψ(y − x)ν(y)(h(y) − h(z))dy, since Ψ(y − x)ν(y)dy = 1, according to the Cauchy formula. Thus differentiation with respect to x, with z fixed, gives Z |∂j Ψ(y − x)| |y − z|α dy = khkα (I + II). |∂j F + (x)| . khkα ∂D

8.3. Hardy Space Splittings

281

Here I denotes the part of the integral inside the cylinder, and II is the part outside. Since |∂j Ψ(y − x)| . 1/|y − x|n , the term II is bounded. For the integral I, we change variable from y = (y 0 , φ(y 0 )) =: ρ(y 0 ) ∈ ∂D to y 0 ∈ Rn−1 . To find the change of (n − 1)-volume, we calculate ρy0 (e1 ∧ · · · ∧ en−1 ) = (e1 + (∂1 φ)en ) ∧ · · · ∧ (en−1 + (∂n−1 φ)en ) = e1···(n−1) + (∂1 φ)en2···(n−1) + (∂2 φ)e1n3···(n−1) + · · · + (∂n−1 φ)e123···(n−2) , p the norm of which is 1 + |∇φ|2 . Since the function φ is C 1 , we conclude that |∂j Ψ(y − x)| ≈ 1/(|y 0 − x0 | + t)n , |y − z|α ≈ |y 0 − x0 |α , and dy ≈ dy 0 , where t = xn − φ(x0 ). Therefore Z Z ∞ 0 0 −n 0 0 α 0 I. (|y − x | + t) |y − x | dy . (r + t)−n rα rn−2 dr . tα−1 . |y 0 | φ(x0 ) and consider first the vertical limit f + (y). Since Z r Z r−φ(y0 ) + 0 + 0 0 + 0 |F (y , r) − F (y , φ(y ) + t)| ≤ |∂n F (y , s)|ds . khkα sα−1 ds, φ(y 0 )+t

0

it is clear that this limit exists, since the integral is convergent. Moreover, we get the estimate |f + (y)| . khkα , since |F + (y 0 , r)| is bounded by khkα . Next we aim to show that {F + (x)} converges when x → y from D+ in general, and not only along the vertical direction. Let x1 = (x01 , t1 ), x2 = (x02 , t2 ) ∈ D ∩ B(y; ), and define t := max(t1 , t2 ) + 2(1 + k∇φk∞ ). Then Z F + (x2 ) − F + (x1 ) = hdx, ∇iF + (x), γ

where γ is the piecewise straight line from x1 to x2 viaR(x01 , t) and (x2 , t). The first and last vertical line integrals are dominated by khkα 0 tα−1 dt as above, whereas in the middle horizontal line integral, the integrand is dominated by khkα α−1 . In total we obtain the estimate |F + (x2 ) − F + (x1 )| . khkα α , when x1 , x2 ∈ D ∩ B(y, ). This shows the existence of the limit f + (y) as x → y from D+ . By taking x1 , x2 ∈ ∂D, it also shows that kf + kα . khkα , which completes the proof. Proposition 8.3.4 (Sokhotski–Plemelj jumps). Let D be a bounded C 1 domain and 0 < α < 1. Then the Cauchy principal value integral E : C α (∂D) → C α (∂D) is a well-defined and bounded linear operator. The Hardy projections E ± equal Z ± 1 E h(x) = 2 h(x) ± p.v. Ψ(y − x)ν(y)h(y) dy, x ∈ ∂D. ∂D

In terms of operators, this means that E ± = 12 (I ± E).

Chapter 8. Hypercomplex Analysis

282

Proof. We start by verifying the identity E + h(x) = 21 h(x) + 12 Eh(x) for x ∈ ∂D. As in the proof of Proposition 8.3.3, write x = (x0 , φ(x0 )) in a coordinate system in a cylinder around x. If h ∈ C α (∂D), the integrand of Z Ψ(y − (x + ten ))ν(y) h(y) − h(x) dy ∂D

is seen to be bounded by |y − x|α−(n−1) , uniformly for 0 < t ≤ t0 . Here we view h(x) as a constant function. Letting t → 0+ and applying the Lebesgue dominated convergence theorem, it follows that Z E + h(x) − h(x) = E + (h − h(x))(x) = Ψ(y − x)ν(y) h(y) − h(x) dy ∂D ! Z Z Ψ(y − x)ν(y)h(y)dy −

= lim

→0

∂D\B(x;)

Ψ(y − x)ν(y)dy h(x).

lim

→0

∂D\B(x;)

The first equality follows from the fact that the Cauchy integral of the constant field h(x) is the constant field h(x) in D+ . It suffices to show that Z lim Ψ(y − x)ν(y)dy = 12 . →0

∂D\B(x;)

formula for the domain D+ \ B(x; ) shows that it suffices Applying the Cauchy R to prove lim→0 ∂B(x;)∩D+ Ψ(y − x)ν(y)dy = 12 . But Z Ψ(y − x)

lim

→0

∂B(x;)∩D +

y−x |∂B(x; ) ∩ D+ | , dy = lim →0 |y − x| |∂B(x; )|

(8.7)

and on approximating ∂D by its tangent hyperplane at x, this limit is seen to be 1/2, since ∂D is assumed to be C 1 regular at x. To summarize, we have shown that E + = 12 (1 + E), and Proposition 8.3.3 shows that E = 2E + − I is a bounded and well-defined operator. Letting t → 0− instead, we get −E − h(x) − 0 = Eh(x) − 12 h(x), which shows that E − = 12 (I − E). Exercise 8.3.5. Generalize Proposition 8.3.3 to bounded Lipschitz domains. Show that Proposition 8.3.4 fails for bounded Lipschitz domains. Summarizing the H¨older estimates in this section, we have the following main result. Theorem 8.3.6 (Hardy subspace splitting). Let D be a bounded C 1 domain and let 0 < α < 1. Then we have a splitting of the H¨ older space C α (∂D) into Hardy subspaces C α (∂D) = E + C α ⊕ E − C α .

8.4. Comments and References

283

The Hardy subspaces are the ranges of the Hardy projections E ± : C α (∂D) → C α (∂D), which are the spectral projections E ± = 12 (I ± E) of the Cauchy singular integral operator E. α The interior Hardy subspace E + C+ consists of all traces F + |∂D of monogenic + + fields F in D that are H¨ older continuous up to ∂D. The exterior Hardy subspace E − C α consists of all traces F − |∂D of monogenic fields F − in D− that are H¨ older continuous up to ∂D and have limit limx→∞ F − (x) = 0. In fact, all such F − have decay O(1/|x|n−1 ) as x → ∞. Proof. Proposition 8.3.3 shows that E ± : C α (∂D) → C α (∂D) are bounded projection operators. Proposition 8.3.4 shows in particular that they are complementary: E + + E − = I. This shows that C α (∂D) splits into the two Hardy subspaces. It is clear from the definition and Proposition 8.3.3 that the Hardy subspaces consist of traces of H¨older continuous monogenic fields F ± in D± respectively. The decay of F − at ∞ follows from that of Ψ. Conversely, the fact that the trace of every H¨ older continuous monogenic field F + in D+ belongs to E + C α follows from Theorem 8.1.8. For the corresponding result for D− , we apply the Cauchy − reproducing formula to the bounded domain DR := D− ∩ B(0; R) for large R. We have Z Z − F − (x) = − Ψ(y − x)ν(y)F − (y)dy + Ψ(y − x)ν(y)F − (y)dy, x ∈ DR . ∂D

|y|=R

Since |∂B(0; R)| grows like Rn−1 and Ψ(x − y) decays like 1/Rn−1 as R → ∞, the last integral will vanish if limx→∞ F − (x) = 0, showing that F − is the Cauchy integral of F − |∂D , so that F − |∂D ∈ E − C α .

8.4

Comments and References

8.1 The higher-dimensional complex analysis obtained from the Dirac equation and Clifford algebra has been developed since the 1980s. This research field is referred to as Clifford analysis. The pioneering work is Brackx, Delanghe, and Sommen [23]. Further references include Gilbert and Murray [42] and Delanghe, Sommen, and Soucek [33]. Div/curl systems like those in Example 8.1.4 have been used to define higher-dimensional harmonic conjugate functions in harmonic analysis. The seminal work is Stein and Weiss [89] 8.2 This material builds on the treatment by Axler, Bourdon, and Ramey [16] of spherical harmonics. We have generalized mutatis mutandis the theory for spherical harmonics to spherical monogenics. 8.3 The classical Lp -based Hardy spaces, named after G.H. Hardy, on the real axis or the unit circle in the complex plane where introduced by F. Riesz in

284

Chapter 8. Hypercomplex Analysis 1923. The function space topologies for p ≤ 1 that they provide are fundamental in modern harmonic analysis. Theorem 8.3.2 was proved by R. Coifman, A. McIntosh, and Y. Meyer in [28] for general Lipschitz graphs in the complex plane. Earlier, A. Calder´on had obtained a proof in the case of small Lipschitz constants. The higherdimensional result in Theorem 8.3.2 is equivalent to the L2 boundedness of the Riesz transforms on Lipschitz surfaces, and this was known already in [28] to follow from the one-dimensional case by a technique called Calder´on’s method of rotations. A direct proof using Clifford algebra is in [66]. From Calder´on–Zygmund theory, also Lp boundedness for 1 < p < ∞ follows. A reference for wavelet theory, which is intimitely related to Theorem 8.3.2, is Meyer [69]. It is interesting to note that just like induced bases {es } for multivectors, wavelet bases for function spaces also do not come with a linear order of the basis functions, but these are rather ordered as a tree. For Clifford algebras and wavelets, see Mitrea [71]. Unpublished lecture notes by the author containing the wavelet proof of Theorem 8.3.2 are [81]. The basic idea behind estimating singular integrals like the Cauchy integral using wavelets is simple: the matrices of such operators in a wavelet basis are almost diagonal in a certain sense. However, the nonlinear ordering of the basis elements and the details of the estimates make the proof rather technical. There is also a much deeper extension to higher dimensions of the result in [28] that was known as the Kato square root problem. It was finally solved affirmatively by Auscher, Hofmann, Lacey, McIntosh, and Tchamitchian [6] 40 years after it was formulated by Kato, and 20 years after the one-dimensional case [28] was solved. As McIntosh used to tell the story, the works on linear operators by T. Kato and J.-L. Lions closed that field of research in the 1960s; only one problem remained open, and that was the Kato square root problem. A reference for a spectral/functional calculus approach to Theorem 8.3.2 is Axelsson, Keith, and McIntosh [12]. See in particular [12, Consequence 3.6] for a proof of Theorem 8.3.2, and [12, Consequence 3.7] for a proof of the Kato square root problem. This paper illustrates well how Dirac operators and Hodge- and Hardy-type splittings can be used in modern research in harmonic analysis.

Chapter 9

Dirac Wave Equations Prerequisites: Some familiarity with electromagnetism and quantum mechanics is useful for Section 9.2. A background in partial differential equations, see Section 6.3, and boundary value problems is useful but not necessary for the later sections. For the operator theory that we use, the reader is referred to Section 6.4. Ideally, we would have liked to place Section 9.6 after Chapter 10. But since it belongs to the present chapter, we ask the reader to consult Chapter 10 for more on Hodge decompositions when needed. Road map: Acting with the nabla symbol through the Clifford product ∇4F (x) on multivector fields, or through a representation ∇.ψ(x) on spinor fields, in Euclidean space we obtain first-order partial differential operators which are square roots of the Laplace operator ∆. However, Paul Dirac first discovered his original equation in 1928 for spin-1/2 massive particles, in the spacetime setting as a square root of the Klein–Gordon equation, that is, the wave equation with a zero-order term ∂x2 ψ + ∂y2 ψ + ∂z2 ψ − c−2 ∂t2 ψ =

m2 c2 ψ. ~2

The resulting Dirac wave equation ~∇.ψ = mcψ describing the free evolution of the wave function for the particle, a spinor field ψ : W → 4W / in physical spacetime, has been described as one of the most successful and beautiful equations ever. For example, it predicted the existence of antiparticles some years before these were experimentally found in 1932. In Section 9.2 we survey Dirac’s equation, as well as Maxwell’s equations from the early 1860s, which describes the evolution of the electrical and magnetic fields. We show how, in a very geometric way, the electromagnetic field is a multivector field and that the Maxwell equations, when written in terms of Clifford algebra, form a Dirac wave equation. The four classical © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_9

285

286

Chapter 9. Dirac Wave Equations

equations correspond to the four spaces ∧j V of homogeneous multivectors in threedimensional Euclidean space V . Motivated by applications to Maxwell’s equations, Sections 9.3 to 9.7 develop a theory for boundary value problems (BVPs) for Dirac equations, and they show how it applies to electromagnetic scattering. We consider only time-harmonic waves at a fixed frequency. Our abstract setup for a BVP is to consider two splittings of a space H of functions on the boundary ∂D of the domain D. The first splitting, H = A+ H ⊕ A− H, encodes the differential equation and generalizes the splitting into Hardy subspaces A+ H = E + H and A− H = E − H from Theorem 8.3.6. The second splitting, H = B + H ⊕ B − H, encodes the boundary conditions. Typically the projections B ± are pointwise and determined by the normal vector ν. Relating them to the classical boundary value problems for the Laplace operator, in that case B + would encode Dirichlet boundary conditions and B − would encode Neumann boundary conditions. From this point of view of functional analysis, studying BVPs amounts to studying the geometry between these two different splittings. Well-posedness of BVPs will mean that the subspaces A± H do not intersect the subspaces B ± H, and in the optimal case, the two reflection operators A and B, where A generalizes the Cauchy principal value integral, anticommute. In Section 9.5, we formulate integral equations for solving scattering problems for Dirac equations. The aim is to find singular but not hypersingular integral operators that are both bounded and invertible also on Lipschitz boundaries, whenever the scattering problem considered is well posed. A problem that we need to overcome to find an integral equation which is numerically useful is that we cannot easily discretize spaces like the Hardy spaces E ± H, which are defined by a nonlocal integral constraint. We obtain good integral equations on good function spaces for solving BVPs for the Dirac equation. To apply these to scattering problems for electromagnetic waves, we show in Sections 9.6 and 9.7 how we require a third splitting of the boundary function space H: a boundary Hodge decomposition H = R(Γk ) ⊕ R(Γ∗k ), where the Maxwell fields live in the Hodge component R(Γk ). Embedding Maxwell’s equations into the Dirac equation, and solving the BVP with a Dirac integral equation, we give examples in Section 9.7 of how this algorithm performs numerically. Highlights: • Boosting E and B using Clifford algebra: 9.2.4

9.1. Wave and Spin Equations

287

• Discovery of antiparticles: 9.2.8 • Stratton–Chu as a Clifford–Cauchy integral: 9.3.8 • Well-posedness via operator Clifford algebra 4R2 : 9.4.5 • Rellich spectral sector vs. Lipschitz geometry of boundary: 9.5.1 • Spin integral equation: 9.5.5 • Maxwell fields and boundary Hodge decompositions: 9.7.1

9.1

Wave and Spin Equations

The Dirac operator on a Euclidean space from Definition 8.1.1 generalizes in the obvious way to an inner product space of arbitrary signature. Definition 9.1.1 (4-Dirac operator). Let (X, V ) be an inner product space. The 4-Dirac operator D = DV acting on multivector fields F : D → 4V defined on some domain D ⊂ V is the nabla operator (DF )(x) := ∇ 4 F (x) =

n X

e∗j 4 ∂j F (x)

j=1

induced by the Clifford product V × 4V → 4V : (v, w) → 7 v 4 w as in Definition 7.1.2. Here ∂j are partial derivatives with respect to coordinates in a basis {ej } for V , with dual basis {e∗j }. Since 4-Dirac operators are our main type of Dirac operator, we sometimes omit 4 in the notation. When V is a Euclidean space, we speak of a harmonic Dirac operator, while if V = W is a spacetime, we speak of a wave Dirac operator. For a Euclidean space, we know from Chapter 8 that D2 = ∆. We have seen that this means that multivector fields F solving DF = 0 in particular have scalar component functions Fs that are harmonic. Turning to the wave Dirac operator, we now have D2 = , where is the d’Alembertian from Section 6.3. Indeed, in an ON-basis {ei } we have D2 F = (−e0 ∂0 + e1 ∂1 + · · · + en ∂n )2 F = (−∂02 + ∂12 + · · · + ∂n2 )F. Similar to the Euclidean case, it follows that multivector fields solving the wave Dirac equation DF = 0 have scalar component functions Fs that solve the wave equation Fs = 0.

Chapter 9. Dirac Wave Equations

288

The wave Dirac operator describes a wave propagation of multivector fields. For the harmonic Dirac operator we saw in Chapter 8 how the fundamental solution Φ to the Laplace operator yielded a fundamental solution Ψ = ∇Φ to D. Similarly, the fundamental solution to the wave equation encoded by the Riemann operators Rt from Proposition 6.2.2 and Example 6.3.2 now yield solution formulas for the wave Dirac operator. Proposition 9.1.2 (Propagation of Dirac waves). Fix a time-like unit vector e0 in a spacetime W and let V = [e0 ]⊥ be the space-like complement. Consider the initial value problem for the wave Dirac equation DW F = G for given initial data F |V = f and source G. We assume that f and Gx0 (·) = G(x0 , ·), for each fixed time x0 , belong to L2 (V ; 4W ). Then the solution Fx0 (x) = F (x0 , x) is given by Z x0 F (x0 , x) = Mx0 F0 (x) + Mx0 −s (e0 Gs )(x)ds, x0 > 0, 0

where the Fourier multiplier Mx0 on L2 (V ; 4W ) is Mx0 g := (∂0 − e0 D)Rx0 g. Proof. We apply the partial Fourier transform to DW F = G in the x-variables, ˆ x (ξ), ξ ∈ V , for each fixed x0 . We obtain the ODE −e0 ∂0 Fˆx0 (ξ) + iξ Fˆx0 (ξ) = G 0 with solution Z x0 ˆ s (ξ)ds. exp(−ie0 ξ(x0 − s))e0 G Fˆx (ξ) = exp(−ie0 ξx0 )fˆ(ξ) + 0

0

We have exp(−ie0 ξx0 ) = cos(|ξ|x0 )−ie0 ξ sin(|ξ|x0 )/|ξ| according to Exercise 1.1.5, which is seen to be the symbol of Mx0 . The inverse Fourier transformation yields the stated formula for F . It follows that the time evolution of the wave Dirac equation is quite similar to that for the scalar second-order wave equation: with our scaling, the propagation speed is 1, and in odd dimensions there is a Huygens principle. However, although evolution backward in time is well posed, the wave Dirac equation is not symmetric in the time variable, unlike the scalar wave equation. Another difference is that the only inital datum that we need is F (0, ·), and no normal derivative. The second type of Dirac operator that we consider are the following spinDirac operators. / Dirac operator). Let (X, V ) be an inner product space, with Definition 9.1.3 (4/ V acting on spinor fields / =D / . The 4-Dirac operator D / complex spinor space 4V Ψ : X → 4V / is the nabla operator / := ∇.Ψ(x) = (DΨ)(x)

n X j=1

ej∗ .∂j Ψ(x),

9.1. Wave and Spin Equations

289

which is induced by the bilinear map V × 4V → 4V / / : (θ, ψ) 7→ θ.ψ as in Definition 7.1.2. Here ∂j are partial derivatives with respect to coordinates in a basis {ej } for V , with dual basis {ej∗ }. / operator, When V is a Euclidean space, we speak of a harmonic 4-Dirac while if V = W is a spacetime, we speak of a wave 4-Dirac operator. The 4/ / Dirac operators are best known for their representations as matrix first-order partial differential operators. Such expressions are straightforward to derive using representations of Clifford algebras. See Section 5.1. 2 / = ∆, Analogous to the 4-Dirac operator, for a Euclidean space, we have D 2 / = 0 have harmonic / = , and spinor fields solving DΨ while in spacetime D functions and solutions to the wave equation as component functions, respectively. Exercise 9.1.4 (Hypercomplex spin analysis). Show how the theory from Chapter 8 for the Cauchy integral, the monogenic Kelvin transform, and spherical monogen/ operator. ics generalizes in a natural way for solutions to the harmonic 4-Dirac Explain why such solutions do not form a right Clifford module as in Proposition 8.1.13 and why the notion of two-sided monogenicity from Proposition 8.1.5 does not generalize. Show how Proposition 9.1.2 generalize to describe the free wave evolution for equations. wave 4-Dirac / We next consider how the wave Dirac equations are related to the harmonic Dirac equations. Proposition 9.1.5 (4V representation of DW ). Let W be a spacetime and fix a time-like unit vector e0 and its Euclidean orthogonal complement V = [e0 ]⊥ . Identify the even part 4ev W of the spacetime Clifford algebra, and the Euclidean Clifford algebra 4V via the isomorphism in Proposition 3.3.5. We write a general spacetime multivector w ∈ 4W as w = w+ + e0 w− , where w± ∈ 4ev W ↔ 4V . Identifying a multivector field F in W in this way with a pair (F + , F − ) of multivector fields in V , and similarly for G, the wave Dirac equation DW F = G corresponds to ( (∂0 + DV )F+ = −G− , (∂0 − DV )F− = G+ . Proof. Since DW swaps 4ev W and 4od W fields, we obtain in a spacetime ONbasis {ei } that DW F = G is equivalent to ( (−e0 ∂0 + D0 )F+ = e0 G− , (−e0 ∂0 + D0 )(e0 F− ) = G+ , Pn where D0 := j=1 ej ∂j . Multiplying the first equation by e0 and commuting e0 to the left in the second equation establishes the claim.

Chapter 9. Dirac Wave Equations

290

Changing notation, an argument as in the proof of Proposition 9.1.5 in the /WF = G / equation D case that dim W is even also shows that the wave 4-Dirac corresponds to ( / V )F+ = −G− , (∂0 + D / V )F− = G+ . (∂0 − D +

+

/ W : v 7→ Here we use that 4V / = 4 / W via the representation ρ : V → 4 e0 v.(·), and write the spacetime spinor field as F = F+ + e0 F− in the splitting − + + − 4 / W . Note that e0 : 4 / W is invertible. / W ⊕4 / W →4 We end this section by considering how these Dirac operators are related to exterior and interior derivative operators d and δ. For any inner product space, it is clear from the definitions that the 4-Dirac operators are D = d + δ.

(9.1)

This holds true also for the wave 4-Dirac operator. We note the following refinement of Proposition 9.1.5. Proposition 9.1.6 (∧V representation of dW , δW ). Let W be a spacetime and use notation as for D in Proposition 9.1.5. Then the differential equations dW F = G and δW F = G correspond to ( ( (T − ∂0 + δV )F+ = −G− , (T + ∂0 + dV )F+ = −G− , − (T ∂0 − dV )F− = G+ , (T + ∂0 − δV )F− = G+ , respectively, where T + f = 21 (f + fb) and T − f = 12 (f − fb) denote the projections onto ∧ev V and ∧od V respectively. Proof. For example, we see that dW F = G is equivalent to ( −e0 ∧ ∂0 F+ + d0 F+ = e0 G− , −e0 ∧ ∂0 (e0 F− ) + d0 (e0 F− ) = G+ , Pn where d0 F := j=1 ej ∧ ∂j F . To relate the exterior product to the Clifford algebra isomorphism 4ev W ≈ 4V , we use the Riesz formula (3.4). We also note that w 7→ −e0 we0 yields an automorphism of 4ev W that negates e0 V . Therefore −e0 we0 corresponds to w b under the isomorphism 4ev W ↔ 4V . This yields for F ∈ ev 4 W ↔ 4V , e0 (e0 ∧ F ) = 21 (−F + e0 F e0 ) ↔ 12 (−F − Fb) = −T + F, e0 ∧ (e0 F ) = 21 (−F − e0 F e0 ) ↔ 12 (−F + Fb) = −T − F, Pn and with nabla calculus using ∇0 = j=1 ej ∂j that e0 (∇0 ∧ F ) = 12 ((e0 ∇0 )F − e0 F e0 (e0 ∇0 )) ↔ 12 (∇F + Fb∇) = dV F, ∇0 ∧ (e0 F ) = 1 (−(e0 ∇0 )F + e0 F0 (e0 ∇0 )) ↔ 1 (−∇F − Fb∇) = −dV F. 2

2

9.2. Dirac Equations in Physics

291

Similar calculations using the Riesz formula (3.3) prove the 4V representation for δW F = G. The 4-Dirac / operator in a general Euclidean or real inner product space, cannot be written as in (9.1). However, in the case of an even-dimensional Euclidean space with a complex structure given as in Example 5.1.5(i), we do have an invariant meaning of such exterior and interior derivative operators. Given a Euclidean space V of dimension n = 2m with an isometric complex structure J, consider the complex exterior algebra ∧V for the complex vector space V = (V, J), which comes with the corresponding Hermitian inner product (·, ·i. As in Example 5.1.5(i), the real linear map V → L(∧V) : v 7→ v y∗ (·) + v ∧ (·) gives a representation of the complex spinor space 4V / = ∧V. But the two terms induce separately the nabla operators Γ1 ψ := ∇ ∧ ψ

and

Γ2 ψ := ∇ y∗ ψ

/ = Γ1 + Γ2 . Fixing a complex ON-basis acting on spinor fields ψ : V → 4V / and D {ej }m for V and writing x for the real coordinates along ej and yj for the real j j=1 coordinates along Jej , we have, since {ej } ∪ {Jej } form a real ON-basis for V from Proposition 7.1.3, that Γ1 ψ =

m X

ej

∧

∂xj ψ +

j=1

and Γ2 ψ =

m X j=1

ej y∗ ∂xj ψ +

m X

iej

∧

∂yj ψ =

j=1

m X

ej

∧

∂zjc ψ

j=1

m m X X (iej ) y∗ ∂yj ψ = ej y∗ ∂zj ψ, j=1

j=1

since Jej = iej in V and (iej )y∗ w = −i(ej y∗ w) by sesquilinearity. Here we used the classical complex analysis operators ∂zj := ∂xj − i∂yj and ∂zjc := ∂xj + i∂yj . Since Γ∗1 = −Γ2 , one can develop a complex version of the theory of Hodge decomposition similar to the real theory in Chapter 10.

9.2

Dirac Equations in Physics

The aim of this section is to briefly review how Dirac equations appear in electromagnetic theory and quantum mechanics in physics. We model our universe by spacetime W with three space dimensions, as in special relativity. See Section 1.3. The unit of length is the meter [m]. Fixing a future-pointing time-like vector e0 with e20 = −1, we write V for the three-dimensional Euclidean space [e0 ]⊥ . We write the coordinate along e0 as x0 = ct,

Chapter 9. Dirac Wave Equations

292

where t is time measured in seconds [s] and c ≈ 2.998 · 108 [m/s] is the speed of light. Our discussion uses SI units. Out of the seven SI base units, we need the meter [m] for length, kilogram [kg] for mass, second [s] for time and ampere [A] for electric current. From this we have the SI derived units newton [N= kg·m/s2 ] for force, coulomb [C=A·s] for electric charge, volt [V= N·m/C] for electric potential, joule [J= Nm] for energy. We consider first Maxwell’s equations, which describe the time evolution of the electric and magnetic fields, which mediate the forces that electric charges in motion exert on each other. The charges that generate the electric and magnetic fields are described by a charge density and electric current density ρ(t, x) ∈ ∧0 V

and J(t, x) ∈ ∧1 V,

measured in units [C/m3 ] andR[A/m2 ] respectively. This means that a given domain the charge D ρdx, and the electric current through a 2-surface D ⊂ V contains R S in V is S hJ, ∗dyi, at time t. Here [dy] is the tangent plane to S and ∗dy is an infinitesimal vector normal to S, in the direction in which we measure the current. Maxwell’s four equations, which we discuss below, describe how ρ and J generate a vector field E(t, x) ∈ ∧1 V,

measured in units [N/C = V/m],

which is called the electric field, and a bivector field B(t, x) ∈ ∧2 V,

measured in units of tesla [T = N/(A · m)],

which we refer to as the magnetic field. The way we measure these fields is by placing a test charge with charge q0 at the point moving with velocity v0 ∈ ∧1 V . The electric and magnetic fields will then exert a force on this test charge given by the Lorentz force F = q0 E + q0 B x v0 . (9.2) Experiments show that the magnetic force is orthogonal to the velocity, and thus is described by a skew symmetric map. Recalling Proposition 4.2.3, this demonstrates that the magnetic field is a bivector field rather than a vector field. In classical vector notation, the magnetic field is described by the Hodge dual vector field ∗B, in which case the magnetic force is given by the vector product q0 v0 × (∗B). The three-dimensional exterior algebra ∧V = ∧0 V ⊕ ∧1 V ⊕ ∧2 V ⊕ ∧3 V provides a natural framework for expressing the four Maxwell equations, or more precisely eight scalar equations, for determining E ∈ ∧1 V and B ∈ ∧2 V from ρ ∈ ∧0 V and J ∈ ∧1 V . The constants of proportionality appearing are the permittivity of free space 0 ≈ 8.854 · 10−12 [C/(V· m)] and permeability of free space µ0 = 4π · 10−7 [V· s/(A· m)].

293

9.2. Dirac Equations in Physics

∧0 Gauss’s law for the electric field states that the flow of the electric field out through the boundary of a domain D is proportional to the charge Q = R ρdx contained in the domain: D Z hE, ∗dyi = −1 0 Q. ∂D

By Stokes’s theorem, Gauss’s law is equivalent to the ∧0 V -valued differential equation 0 ∇ y E = ρ. In classical vector notation this reads 0 h∇, Ei = ρ. ∧1 The Amp`ere–Maxwell law states that Z Z Z µ−1 hB, ∗dyi = hJ, ∗dxi + ∂ hE, ∗dxi 0 t 0 ∂S

S

S

for every 2-surface S. In the stationary case that ρ, J, E, and B are timeindependent, R it reduces to Amp`ere’s law, which shows that an electric current I := hJ, ∗dxi through S produces a magnetic field with circulation S R hB, ∗dyi = µ0 I. In the ∂S R nonstationary case, Maxwell added the necessary additional term 0 µ0 ∂t S hE, ∗dxi to the equation. By Stokes’s theorem, Amp`ere–Maxwell’s law is equivalent to the ∧1 V valued differential equation 0 ∂t E + µ−1 0 ∇ y B = −J. In classical vector notation this reads 0 ∂t E − µ−1 0 ∇ × (∗B) = −J. ∧2 Faraday’s law of induction states that a change of the integral of the magnetic field B over a 2-surface S induces an electric field around the boundary curve: Z Z hE, dyi = −∂t hB, dxi. ∂S

S

By Stokes’s theorem, Faraday’s law is equivalent to the ∧2 V -valued differential equation ∂t B + ∇ ∧ E = 0. In classical vector notation this reads ∂t (∗B) + ∇ × E = 0. ∧3 Gauss’s law for magnetic fields states that the integral of a magnetic field over the boundary of a domain D vanishes: Z hB, dyi = 0. ∂D

By Stokes’s theorem, the magnetic Gauss’s law is equivalent to the ∧3 V valued differential equation ∇ ∧ B = 0. In classical vector notation this reads h∇, ∗Bi = 0.

294

Chapter 9. Dirac Wave Equations

Figure 9.1: Maxwell’s equations for the electric vector field E and the magnetic bivector field B in ∧R3 . Since the electric and magnetic fields take values in the two different subspaces ∧1 V and ∧2 V of the exterior algebra, we can add them to obtain a sixdimensional total electromagnetic multivector field F . The most natural scaling is such that |F |2 is an energy density, with dimension [J/m3 ]. We set 1/2

−1/2

F := 0 E + µ0

B ∈ ∧1 V ⊕ ∧2 V.

Collecting and rescaling Maxwell’s equations, we have −1/2

∇ ∧ (µ0

B) = 0,

−1/2 1/2 c ∂t (µ0 B) + ∇ ∧ (0 E) 1/2 −1/2 c−1 ∂t (0 E) + ∇ y (µ0 B) 1/2 ∇ y (0 E) −1

= 0, 1/2

= −µ0 J, −1/2

= 0

ρ,

(9.3)

9.2. Dirac Equations in Physics

295

where c = (0 µ0 )−1/2 . Adding these four equations, we see that Maxwell’s equations are equivalent to the Dirac equation c−1 ∂t F + ∇ 4 F = G,

(9.4)

since Maxwell’s equations take values in the different homogeneous subspaces of −1/2 1/2 ∧V . Here G := 0 ρ − µ0 J is a ∧0 V ⊕ ∧1 V -valued multivector field, which we refer to as the electric four-current. From (9.4) it is clear that Maxwell’s equation is a wave Dirac equation for the ∧1 V ⊕ ∧2 V -valued electromagnetic field F . Example 9.2.1 (Static electromagnetic field). Assume that the sources ρ and J and the electromagnetic field are constant with respect to time, and that J is divergence-free. Then Maxwell’s equations reduce to the inhomogeneous Dirac equation ∇ 4 F = G, which by the Cauchy–Pompeiu formula from Theorem 8.1.8 has solution F (x) = Ψ(x) ∗ G(x) if G decays as x → ∞. This amounts to Z 1 y E(x) = ρ(x − y) 3 dy, 4π0 V |y| Z µ0 y B(x) = J(x − y) ∧ 3 dy. 4π V |y| Thus E is the Coulomb field from charge density ρ, and B is determined from J by the law of Biot–Savart. Exercise 9.2.2 (Pauli representation). Using an ON-basis {e1 , e2 , e3 } for V , write 1/2 ˜ 1 e1 + E ˜ 2 e2 + E ˜3 e3 , µ−1/2 B = B ˜1 e23 + B ˜2 e31 + B ˜3 e12 , µ1/2 J = J˜1 e1 + 0 E = E 0 0 −1/2 J˜2 e2 + J˜3 e3 , and 0 ρ = ρ˜. Represent the basis vectors {e1 , e2 , e3 } by the Pauli matrices from Example 3.4.19 and show that Maxwell’s equations become −1 ˜3 + iB ˜3 ˜ 1 − iE ˜2 + iB ˜1 + B ˜2 c ∂t + ∂3 ∂1 − i∂2 E E ˜1 + iE ˜2 + iB ˜1 − B ˜2 ˜ 3 − iB ˜3 ∂1 + i∂2 c−1 ∂t − ∂3 E −E ρ˜ − J˜3 −J˜1 + iJ˜2 = . ˜ ˜ −J1 − iJ2 ρ˜ + J˜3 Note that this representation requires that the components of the fields be realvalued, since we use a real algebra isomorphism 4R V ↔ C(2). For time-dependent electromagnetic fields we can obtain a spacetime Dirac formulation of Maxwell’s equation from Proposition 9.1.5. Namely, the electromagnetic field is really the spacetime bivector field 1/2

−1/2

FW := 0 e0 ∧ E + µ0

B ∈ ∧2 W,

solving the spacetime Dirac equation DW FW = −GW ,

(9.5)

Chapter 9. Dirac Wave Equations

296 −1/2

1/2

where GW := 0 ρe0 + µ0 J ∈ ∧1 W is the spacetime representation of the electric four-current. Since GW is a spacetime vector field and FW is a spacetime bivector field, Maxwell’s equations can equivalently be written as the system ( dW FW = 0, δW FW = −GW , by the mapping properties of DW = dW + δW . The difference between Maxwell’s equations and the Dirac equation is a constraint similar to the one described in Proposition 8.1.5. −1/2

1/2

Proposition 9.2.3 (Maxwell = Dirac + constraint). Let GW = 0 ρe0 + µ0 J ∈ ∧1 W . If FW is a ∧2 W -valued solution to (9.5), then ρ and J satisfy the continuity equation ∂t ρ + div J = 0. Conversely, if this continuity equation holds, then the multivector field FW solving the wave Dirac equation (9.5) described in Proposition 9.1.2 is ∧2 W -valued at all times, provided it is so at t = 0 with div E = ρ/0 and ∇ ∧ B = 0. Recall that the continuity equation ∂t ρ + div J = 0 expresses the fact that total charge is conserved. By Gauss’s theorem it shows that Z Z ρdx = − ∂t hJ, ∗dyi D

∂D

for every domain D ⊂ V . Proof. The necessity of the continuity equation follows from 2 δW GW = −δW FW = 0,

by the nilpotence of the spacetime interior derivative. For the converse, we investigate the proof of Proposition 9.1.2 and compute the ∧0 W and ∧4 W parts of FˆW = Fˆ . The ∧4 W part is −1/2 sin(|ξ|ct)

µ0

|ξ|

ˆ0 = 0, iξ ∧ B

since ∇ ∧ B0 = 0. The ∧0 W part is sin(|ξ|ct) e0 y (ξ y Fˆ0 ) |ξ| Z t 1/2 sin(|ξ|c(t − s)) −1/2 ρs − µ0 iξ y Jˆs ds +c 0 cos(|ξ|c(t − s))ˆ |ξ| 0 sin(|ξ|ct) 1/2 ˆ0 − −1/2 ρˆ0 ) = (0 iξ y E 0 |ξ| Z t sin(|ξ|c(t − s)) −1 −1/2 1/2 −c (c 0 ∂s ρˆs + µ0 iξ y Jˆs )ds, |ξ| 0

i

9.2. Dirac Equations in Physics

297

which vanishes, since div E0 = ρ0 /0 and ∂t ρ + div J = 0. This shows that FW is a homogeneous spacetime bivector field for all times. Example 9.2.4 (Lorentz transformation of E and B). From the spacetime representation FW of the electromagnetic field, we can find how the electric and magnetic fields transform under a change of inertial system. Consider two inertial observers O and O0 , with ON-bases {e0 , e1 , e2 , e3 } and {e00 , e10 , e20 , e03 } respectively. Assume that O sees O0 traveling in direction e3 at speed v. As in Example 4.4.1, the Lorentz boost that maps {ei } to {ei0 } is T x = exp(φe03 /2)x exp(−φe03 /2),

tanh φ = v/c.

In ∧2 W we have the electromagnetic field 1/2

µ0 F = c−1 e0 4 E + B = c−1 e00 4 E 0 + B 0 , where E = E1 e1 + E2 e2 + E3 e3 and B = B1 e23 + B2 e31 + B3 e12 are the fields 0 0 0 are the +B30 e12 measured by O and E 0 = E10 e10 +E20 e02 +E30 e30 and B 0 = B10 e23 +B20 e31 0 fields measured by O . We now compare the two measurements by identifying the ˜ = E 0 e1 +E 0 e2 +E 0 e3 two bases as in the discussion above Example 4.6.5, letting E 1 2 3 0 0 0 ˜ and B = B1 e23 + B2 e31 + B3 e12 . Then ˜ + B) ˜ exp(−φe03 /2). c−1 e0 4 E + B = exp(φe03 /2)(c−1 e0 4 E Applying the isomorphism 4ev W ≈ 4V from Proposition 3.3.5, we have equivalently ˜ = exp(−φe3 /2)(c−1 E + B) exp(φe3 /2). ˜+B c−1 E Computing the action of x 7→ exp(−φe3 /2)x exp(φe3 /2) on e1 , e2 , e3 , e23 , e31 , e12 , we get p E10 = (E1 − vB2 )/ 1 − v 2 /c2 , p 0 E2 = (E2 + vB1 )/ 1 − v 2 /c2 , E 0 = E , 3 3 p 0 B1 = (B1 + (v/c2 )B2 )/ 1 − v 2 /c2 , p 0 2 2 2 B2 = (B2 − (v/c )E1 )/ 1 − v /c , 0 B3 = B3 . From this we see that for speeds v comparable to the speed of light c, there is a significant mixing of E and B, which shows that indeed it is correct to speak of the electromagnetic field rather than electric and magnetic fields only, since the latter two depend on the inertial frame. We have seen that Maxwell’s equations can be written as a 4-Dirac wave equation DW FW = 0. However, the electromagnetic field FW is not a general spacetime multivector field, but a bivector field. This means that Maxwell’s equations are not identical to the 4-Dirac equation, but rather that we can embed

298

Chapter 9. Dirac Wave Equations

Maxwell’s equations in a Dirac equation. We show in the remaining sections of this chapter that this is a very useful technique, since in some respects, Dirac equations are better behaved than the Maxwell equations. An equation from physics that truly is a Dirac equation is Dirac’s original equation for the relativistic motion of spin-1/2 particles in quantum mechanics, such as electrons and quarks. With our notation this is a wave 4-Dirac equation in / physical spacetime with a lower-order mass term. Without any external potential, the free Dirac equation reads / = mcψ. (9.6) ~Dψ Here c is the speed of light and ~ = 6.626 · 10−34 [Js] is Planck’s constant. The parameter m is the mass of the particle, which in the case of the electron is m ≈ 9.109 · 10−31 [kg]. Dirac’s original approach was to look for a first-order differential equation that is a square root of the Klein–Gordon equation, that is, the wave equation with a mass term ~2 ψ = m2 c2 ψ,

(9.7)

which is obtained from the relativistic energy–momentum relation E 2 c−2 − p2 = m2 c2 by substituting E → i~∂t and p → −i~∇. Such a scalar first-order differential equation does not exist, but Dirac succeeded by allowing matrix coefficients. Having multivectors and spinors at our disposal, we already know that the 4/ has an invariant Dirac equation (9.6) for spacetime spinor fields ψ : W → 4W / geometric meaning. Exercise 9.2.5 (Matrix representation). Fix an ON-basis {e0 , e1 , e2 , e3 } for spacetime, and represent the dual basis {−e0 , e1 , e2 , e3 } by the imaginary Dirac matrices {iγ 0 , iγ 1 , iγ 2 , iγ 3 }, where γ k are Dirac’s gamma matrices as in Example 5.1.9. Show that Dirac’s equation reads ψ1 ∂1 − i∂2 ψ1 0 ∂0 ∂3 i∂ −∂ ψ 0 ∂ ∂ + 2 1 0 3 2 ψ2 i~ ψ3 = mc ψ3 . −∂3 −∂1 + i∂2 −∂0 0 −∂1 − i∂2 ψ4 ∂3 −∂0 ψ4 0 The physical interpretation of complex-valued wave functions ψ in quantum mechanics is that |ψ|2 represents a probability density for the position of the / , we require an inner particle. For the spinor-valued wave function ψ : W → 4W product on the spinor space 4V / . The following is a version of Proposition 5.3.1 for physical spacetime. Proposition 9.2.6 (Inner product). Let W be four-dimensional spacetime, with chosen future time direction fixed and complex spinor space 4W / . Then there exists a complex inner product (·, ·i on 4W / such that (ψ1 , v.ψ2 i = −(v.ψ1 , ψ2 i

9.2. Dirac Equations in Physics

299

for all ψ1 , ψ2 ∈ 4W / and v ∈ W , and −i(ψ, v.ψi > 0 / \ {0} and v ∈ Wt+ . If (·, ·i0 is any other such inner product, then for all ψ ∈ 4W there is a constant λ > 0 such that (ψ1 , ψ2 i0 = λ(ψ1 , ψ2 i for all ψ1 , ψ2 ∈ 4W / . Proof. The proof is analogous to that of Proposition 5.3.1. We look for a matrix M such that M ρ(v) = −ρ(v)∗ M, which exists, unique up to complex nonzero multiples, by Theorem 5.2.3, since (−ρ(v)∗ )2 = ρ(v 2 )∗ = hvi2 I. Using the representation ρ from Example 5.1.9, we see that we have M = λρ(e0 ), λ ∈ C \ {0}, where e0 is a fixed future-pointing time-like unit vector. For the duality to be an inner product, that is, symmetric, we must choose Re λ = 0, and to have −i(ψ, e0 .ψi > 0, we must have Im λ < 0. This shows uniqueness. Choosing λ = −i and v = e0 + v 0 , he0 , v 0 i = 0, we have −i(ψ, v.ψi = ψ ∗ (1 − ρ(e0 v 0 ))ψ > 0 if |v 0 | < 1, since ρ(e0 v 0 ) is |v 0 | times a C4 isometry. This completes the existence proof. This spacetime spinor inner product is used as follows. Given a wave function solving Dirac’s equation, a spinor field / ψ : W → 4W in spacetime, we define uniquely a vector field jp : W → ∧1 W by demanding hjp , vi = i(ψ, v.ψi for all v ∈ W . This exists by Proposition 1.2.3 and is referred to as the probability four-current. Fixing a future time direction e0 and writing jp = ρp e0 + c−1 Jp , Jp ∈ [e0 ]⊥ , it follows from the properties of (·, ·i that jp is a real vector field with time component ρp ≥ 0. This represents the probability density for the position of the particle. That Jp defines a probability current is clear from the continuity equation ∂t ρp + divV Jp = 0. This holds whenever ψ solves Dirac’s equation (9.6), since c−1 (∂t ρp + divV Jp ) = δW jp = −∂0 hjp , e0 i +

3 X

∂k hjp , ek i

1

=i

3 X 0

/ ψi = 0. / − i(Dψ, ∂k (ψ, ek∗ .ψi = i(ψ, Dψi

Chapter 9. Dirac Wave Equations

300

Recall the main reflector from Definition 5.2.1, which for physical spacetime we choose as w4 = ie0123 ∈ 44 W. In physics ρ(w4 ) is referred to as the chiral operator, and spinors in its eigenspaces 4± W are called right- and left-handed spinors respectively. To obtain a Euclidean formulation of Dirac’s equation, we fix a future time direction e0 and rewrite (9.6) as a coupled system of Euclidean 4-Dirac / equations for the right- and left-handed components of the wave function. As in the discussion after Proposition 9.1.5, we obtain ( / + = −mψ (c−1 ∂t + D)ψ ˜ −, − −1 / = mψ ˜ +, (c ∂t − D)ψ +

/ =D / V is the 4-Dirac / W , k = 1, 2, m / / ↔4 ˜ := mc/~, and D where ψ ± (t, x) ∈ 4V operator for the Euclidean three-dimensional space V = [e0 ]⊥ . Exercise 9.2.7. Under the algebra isomorphism +

−

/ / 2 3 (ψ + , ψ − ) ↔ ψ + + e0 ψ − ∈ 4 4V / W ⊕4 / W = 4W, show by uniqueness, with suitable normalization λ > 0, that the spacetime spinor inner product of ψ1 = ψ1+ + e0 ψ1− and ψ2 = ψ2+ + e0 ψ2− from Proposition 9.2.6 corresponds to i((ψ1+ , ψ2− i − (ψ1− , ψ2+ i), where (·, ·i denotes the Hermitian spinor inner product on 4V / . / 2 , we have Using the Hermitian inner product (ψ1+ , ψ2+ i + (ψ1− , ψ2− i on 4V the following Hilbert space result. Proposition 9.2.8 (Antiparticles and time evolution). Write Dirac’s equation as i~∂t ψ = H0 ψ, where / D m ˜ / 2 ). H0 := −i~c : L2 (V ; 4V / 2 ) → L2 (V ; 4V / −m ˜ −D Then the free Dirac Hamiltonian H0 has spectrum σ(H0 ) = (−∞, −mc2 ]∪[mc2 , ∞). We have an orthogonal splitting of L2 into spectral subspaces / 2 ) ⊕ L2− (V ; 4V / 2 ), L2 (V ; 4V / 2 ) = L+ 2 (V ; 4V where / 2) = L± 2 (V ; 4V

n −imψ ˜

/ ± iDψ

o t √ / m ˜ 2 − ∆ψ ; ψ ∈ 4V

are the spectral subspaces for the energy intervals [mc2 , ∞) and (−∞, −mc2 ] respectively. The solution to the inital value problem for the wave Dirac equation is m ˜ ψ(t, x) = (c−1 ∂t + H0 /(~c))Rct ψ(0, x),

ψ(t, ·) ∈ L2 (V ; 4V / 2 ),

9.2. Dirac Equations in Physics

301

˜ m where Rct denotes the Klein–Gordon Riemann function from Corollary 6.2.3, acting component-wise.

/ 2 ), the parts Splitting the wave function ψ = ψ + +ψ − , where ψ ± ∈ L2± (V ; 4V − ψ and ψ of positive and negative energy describe a particle and an antiparticle respectively. Note that time evolution by H0 preserves the subspaces L2± (V ; 4V / 2 ). It follows from Corollary 6.2.3 that Dirac’s equation has finite propagation speed ≤ c. However, unlike the massless case in Proposition 9.1.2, the Huygens principle is not valid for Dirac’s equation in three spatial dimensions. Compare also this time evolution for Dirac’s equation to that for Schr¨odinger’s equation in Example 6.3.6, where instead of finite propagation speed, we have the evolution given by an oscillatory quadratic exponential. +

Proof. Applying the Fourier transform in V , Dirac’s equation is turned into the ordinary differential equation ˜ iρ(ξ) m ∂t ψ = −c ψ, −m ˜ −iρ(ξ) p where ρ(ξ) ∈ L(4V / ). For the matrix we obtain eigenvalues ±i |ξ|2 + m ˜ 2 and p t eigenvectors −imψ ˜ −ρ(ξ)ψ ± |ξ|2 + m ˜ 2 ψ . Applying the inverse Fourier transform, this translates to the stated splitting. iρ(ξ) m ˜ To calculate the time evolution, we write j := √ 21 2 and ˜ |ξ| +m −m ˜ −iρ(ξ) note that j 2 = −I. It follows from Exercise 1.1.5 that p p p exp(−c |ξ|2 + m ˜ 2 j) = cos(ct |ξ|2 + m ˜ 2 ) − j sin(ct |ξ|2 + m ˜ 2 ), which under the Fourier transform is equivalent to the stated evolution formula. Example 9.2.9 (Foldy–Wouthuysen transformation). The particle and antiparticle splitting of a solution ψ : W → 4W / to Dirac’s equation (9.6) is independent of the inertial frame for W . Indeed, since H0 ψ = i~∂t ψ, we have p ic−1 ∂t ψ = ± m ˜ 2 − ∆ψ,

(9.8)

with sign +1 for particles, that is, ψ − = 0, and sign −1 for antiparticles, that is, ψ + = 0. Note that (9.6) is a differential equation that is a square root of the Klein–Gordon equation (9.7), and that (9.8) are also square roots of (9.7) although not differential equations. Using the spacetime Fourier transform, we see that the Fourier transforms of wave functions, in the distributional sense, for particles and antiparticles are supported on the two branches of the hyperboloid hξi2 + m ˜ 2 = 0. In particular, this shows the claimed relativistic invariance.

Chapter 9. Dirac Wave Equations

302

Exercise 9.2.10 (Charge conjugation). Consider the spinor space 4W / of physical spacetime, with spinor inner product as in Proposition 9.2.6. Show by generalizing the Euclidean theory from Section 5.3 that there exists an antilinear spinor / / → 4W conjugation 4W : ψ 7→ ψ † such that †

(v.ψ) = v.ψ † ,

v ∈ W, ψ ∈ 4W, /

†

and (ψ † ) = ψ, ψ ∈ 4W / , and that this is unique modulo a complex factor |λ| = 1. Show further that in the representation from Example 5.1.9, we can c choose ψ † = (ρ(e2 )ψ) and that this spinor conjugation is compatible with the spinor inner product as in Lemma 5.3.4. For Dirac’s equation, the operation ψ 7→ ψ † represents charge conjugation in physics, an operation that switches particles and antiparticles, which is readily † seen from (9.8). Mathematically, note that since (ψ † ) = ψ, the spinor conjugation yields a real structure on the spinor space of physical spacetime. This agrees with the fact that with our sign convention, the Clifford algebra 4W is isomorphic to R(4) by Theorem 3.4.13. Recall that a classical particle in an electromagnetic field is acted upon by the Lorentz force (9.2). For a quantum spin-1/2 particle in an electromagnetic field, the Dirac equation is modified by adding a source term and reads / = mcψ + iqAW .ψ. ~Dψ

(9.9)

The vector field AW : W → ∧1 W is a four-potential of the electromagnetic field FW = dW AW and q is the charge of the particle, which in case of the electron is q ≈ −1.602 · 10−19 [C]. A geometric interpretation of (9.9) is that AW provides Christoffel symbols for a covariant derivative as in Definition 11.1.5. The Faraday and magnetic Gauss laws show that the electromagnetic field FW is a closed spacetime bivector field, that is, dW FW = 0. Poincar´e’s theorem (Theorem 7.5.2) shows that locally this is equivalent to the existence of a spacetime vector field AW such that FW = dW AW . As we have seen, at least in the Euclidean setting, in Section 7.6, globally there can be topological obstructions preventing every closed field from being exact. And indeed, the famous Aharonov–Bohm experiment shows that in fact, FW being an exact bivector field is the correct −1/2 1/2 physical law, and not dW FW = 0. Writing AW = 0 Φe0 + µ0 A to obtain a Euclidean expression for potential, where Φ : W → R and A : W → V are scalar and vector potentials of the electromagnetic field, we have ( E = −∇Φ − ∂t A, B = ∇ ∧ A. Returning to (9.9), we note that a solution ψ still yields a probability four-current jp satisfying the continuity equation, as a consequence of AW being a real spacetime vector field. As in the free case, (9.9) describes the time evolution of the wave

9.3. Time-Harmonic Waves

303

functions for a particle and antiparticle pair. What is, however, not immediately clear is how the nonuniqueness of AW influences the solution ψ. To explain this, consider an exact spacetime bivector field FW : W → 42 W representing the electromagnetic field, and let AW , A˜W : W → 41 W be two different vector potentials, so that FW = dW AW = dW A˜W . Another application of Poincar´e’s theorem (Theorem 7.5.2) shows that locally, the closed vector field A˜W − AW is exact, so that A˜W = AW + ∇U, for some scalar potential U : W → 40 W = R. From the product rule, we deduce / = mcψ + iqAW .ψ if and only if that ~Dψ / iqU/~ ψ) = (mc + iq A˜W .)(eiqU/~ ψ). ~D(e Therefore ψ˜ := eiqU/~ ψ is the wave function of the particle in the electromagnetic ˜ = (ψ, v.ψi by sesquilinearity, the ˜ v.ψi field with potential A˜W . However, since (ψ, wave functions for the two choices of electromagnetic four-potential yield the same probability four-current jp . Therefore the physical effects are independent of the choice of electromagnetic four-potential AW .

9.3

Time-Harmonic Waves

Let W be a spacetime and fix a future pointing time-like unit vector e0 , and let V = [e0 ]⊥ . For the remainder of this chapter, we study time-harmonic solutions to the wave 4-Dirac equation DW F = DV F −e0 c−1 ∂t F = 0. We use the complexified spacetime Clifford algebra 4Wc , where the component functions Fs (x) belong to C. With a representation of the time-harmonic field as in Example 1.5.2, the Dirac equation reads (D + ike0 )F (x) = 0, with a wave number k := ω/c ∈ C. This is now an elliptic equation with a zeroorder term ike0 added to D = DV , rather than a hyperbolic equation. Since even the inner product on the real algebra 4W is indefinite, we require the following modified Hermitian inner product for the analysis and estimates to come. Definition 9.3.1 (Hermitian inner product). With V = [e0 ]⊥ ⊂ W as above, define the auxiliary inner product b1 e0−1 , w2 i, hw1 , w2 iV := he0 w

w1 , w2 ∈ 4W.

We complexify both the standard indefinite inner product h·, ·i on 4W and h·, ·iV to sesquilinear inner products (·, ·i and (·, ·iV on 4Wc respectively.

Chapter 9. Dirac Wave Equations

304

= u − e0 v. It b −1 We note that if w = u + e0 v, with u, v ∈ 4V , then e0 we 0 follows that (·, ·iV is a Hermitian inner product in which the induced basis {es } is an ON-basis for 4Wc whenever {ej }nj=1 is an ON-basis for V . We use the L2 R norm kf k2L2 = (f (x), f (x)iV dx of complex spacetime multivector fields f . The aim of this section is to generalize Section 8.3 from the static case k = 0 to k ∈ C. Note that (9.10) (D ± ike0 )2 = ∆ + k 2 . Definition 9.3.2 (Fundamental solution). Let Φk be the fundamental solution to the Helmholtz equation from Corollary 6.2.4 for Im k ≥ 0. Define fundamental solutions Ψk± = (D ± ike0 )Φk to the Dirac operators D ± ike0 . Note the relation Ψk− (x) = −Ψk+ (−x) − between these two families of fundamental solutions, and that Ψ+ 0 = Ψ0 equals ± Ψ from Definition 8.1.6. It is clear from Corollary 6.2.4 that Ψk in general can (1) be expressed in terms of Hankel functions Hν , which in odd dimensions are elementary functions involving the exponential function eik|x| .

Exercise 9.3.3 (Asymptotics). Show that in three dimensions, eik|x| x ik x ± Ψk (x) = − ± e0 . |x|3 |x| |x| 4π ± Note that Ψ± k ≈ Ψ near x = 0, while Ψk ∈ 4Wc is almost in the direction of the x light-like vector |x| ± e0 near x = ∞. Show that in dimension dim V = n ≥ 2 we have −(n−2) Ψ± ), k (x) − Ψ(x) = O(|x|

as x → 0,

−(n−1) as well as ∇ ⊗ (Ψ± ) as x → 0, and that k (x) − Ψ(x)) = O(|x| −ik|x| Ψ± k (x)e

−

π n−1 k (n−1)/2 1 −i 2 2 2e 2π

x |x| ± e0 |x|(n−1)/2

= O(|x|−(n+1)/2 ),

as x → ∞.

Theorem 9.3.4. Let D ⊂ V be a bounded C 1 -domain. If F : D → 4Wc solves (D + ike0 )F = 0 in D and is continuous up to ∂D, then Z F (x) = Ψ− for all x ∈ D. (9.11) k (y − x)ν(y)F (y) dy, ∂D

Note that since Ψ (x) = −Ψ+ k (−x), we can write the reproducing formula equivalently as Z F (x) = − Ψ+ k (x − y)ν(y)F (y) dy. −

∂D

9.3. Time-Harmonic Waves

305

Proof. The proof is analogous to that of Theorem 8.1.8. We define the linear 1-form D \ {x} × V → 4Wc : (y, v) 7→ θ(y, v) := Ψ− k (y − x)vF (y). For y 6= x, its exterior derivative is θ(y, ˙ ∇) = =

n X

∂yi Ψ− k (y − x) 4 ei 4 F (y)

i=1 (Ψ− k (y˙

− x) 4 ∇) 4 F (y) + Ψ− ˙ k (y − x) 4 (∇ 4 F (y)).

Since DF = −ike0 F and ˙ 4 ∇ = (∇ − ike0 ) 4 Φk− (x) ˙ 4∇ Ψ− k (x) ˙ = Φ− − ike0 )2 + (∇ − ike0 ) 4 ike0 ) = Ψ− k (x) 4 ike0 , k (x)((∇ we obtain θ(y, ˙ ∇) = 0. Applying the Stokes formula on the domain D := D \ B(x, ) and using the asymptotics of Ψ− near the origin from Exercise 9.3.3, the rest of the proof follows as for Theorem 8.1.8. It is essential in Theorem 9.3.4 that the domain D = D+ is bounded. In the exterior domain D− = V \ D, we need appropriate decay of F at ∞. When k 6= 0, this takes the form of a radiation condition as follows. Definition 9.3.5 (Radiating fields). Let F be a multivector field that solves (D + ike0 )F = 0 in D− . We say that F radiates at ∞ if Z lim Ψ− k (y − x)ν(y)F (y) dy = 0, R→∞

|y|=R

for every x ∈ D− . Note that by applying Theorem 9.3.4 to the annulus R1 < |x| < R2 , the limit is trivial since the integrals are constant for R > |x|. We need an explicit description of this radiation condition. Note that x ( |x| + e0 )2 = 0.

Proposition 9.3.6 (Radiation conditions). Let F be a multivector field that solves (D + ike0 )F = 0 in D− and is continuous up to ∂D, and assume that Im k ≥ 0 and k 6= 0. If x ( |x| + e0 )F = o(|x|−(n−1)/2 e(Im k)|x| ) as x → ∞, then F radiates. Conversely, if F radiates, then Z F (x) = Ψ+ k (x − y)ν(y)F (y) dy, ∂D

(9.12)

Chapter 9. Dirac Wave Equations

306 for all x ∈ D− . In particular, F = O(|x|−(n−1)/2 e−(Im k)|x| )

and

x ( |x| + e0 )F = O(|x|−(n+1)/2 e−(Im k)|x| )

as x → ∞. Not only does this give an explicit description of the radiation condition, but it also bootstraps it in that the necessary condition is stronger than the sufficient condition. x Proof. Assuming the decay condition on ( |x| + e0 )F , it suffices to prove that Z |F |2V dx = O(e2Im kR ) (9.13) |x|=R

as R → ∞. Indeed, the Cauchy–Schwarz inequality and the asymptotics for Ψ− k from Exercise 9.3.3 then show that F radiates. To estimate |F |V , we note that x x x x |( |x| + e0 )F |2V = (F, ( |x| − e0 )( |x| + e0 )F iV = 2|F |2V + 2(F, |x| e0 F iV . − := {x ∈ D− ; |x| < R}, Applying Stokes’s theorem for bodies on the domain DR we obtain Z Z x (F, |x| (F, νe0 F iV dx e0 F iV dx = |x|=R ∂D Z + (−ike0 F, e0 F iV + (F, −e0 (−ike0 F )iV dx. − DR

In total, this shows that Z Z Z Z 1 x |F |2V dx = (F, νe0 F iV dx−2Im k |F |2V dx, |( |x| +e0 )F |2V dx− − 2 |x|=R |x|=R DR ∂D from which (9.13) follows from the hypothesis, since ∂D is independent of R and since Im k ≥ 0. − The converse is an immediate consequence of Theorem 9.3.4 applied in DR , − and the asymptotics for Ψk from Exercise 9.3.3. There are two important applications of the Dirac equation (D + ike0 )F = 0 to classical differential equations, namely time-harmonic acoustic and electromagnetic waves. Example 9.3.7 (Helmholtz’s equation). For a scalar function u, define F = ∇u + ike0 u. Then u solves the Helmholtz equation ∆u + k 2 u = 0 from Example 6.3.4 if and only if F solves the Dirac equation DF + ike0 F = 0. However, note that F is not a general solution to this equation: it is a vector field F (x) ∈ 41 Wc .

9.3. Time-Harmonic Waves

307

To investigate the reproducing formula (9.11) for this vector field F , we evaluate the time-like and space-like parts of the equation, and get Z u(x) = ∂ν Φk (y − x)u(y) − Φk (y − x)∂ν u(y) dy, (9.14) Z∂D ∇u(x) = ∇Φk (y − x) 4 ν(y) 4 ∇u(y) + k 2 Φk (y − x)u(y)ν(y) dy, (9.15) ∂D

for x ∈ D, where ∂ν denotes the derivative in the normal direction. Equation (9.14) we recognise as the Green second identity for solutions to the Helmholtz equation, whereas (9.15) is an analogue of this for the gradient. This latter equation can be further refined by expanding the triple Clifford vector product as ∇Φk 4 ν 4 ∇u = (∂ν Φk )∇u + ∇Φk (∂ν u) − (∇Φk , ∇uiν + ∇Φk ∧ ν ∧ ∇u. Evaluating the vector part of (9.14), we obtain Z ∇u(x) = ∂ν Φk (y − x)∇u(y) + ∇Φk (y − x)∂ν u(y) ∂D − (∇Φk (y − x), ∇u(y)iν(y) + k 2 Φk (y − x)u(y)ν(y) dy,

x ∈ D.

For solutions to the Helmholtz equation ∆u + k 2 u = 0, the classical decay condition at ∞ is the Sommerfield radiation condition ∂r u − iku = o(|x|−(n−1)/2 e(Im k)|x| ), with ∂r denoting the radial derivative. To see its relation to the radiation condition for D + ike0 , we compute x x x x ( |x| + e0 )F = ( |x| + e0 )(∇u + ike0 u) = (1 + e0 ∧ |x| )(∂r u − iku) + ( |x| + e0 ) ∧ ∇S u, x x where ∇S u := |x| y ( |x| ∧ ∇u) is the angular derivative. By considering the scalar part of this identity, we see that the Dirac radiation condition entails the Sommerfield radiation condition. In fact the two conditions are equivalent. To see this, we can argue similarly to the proof of Proposition 9.3.6 to show that Green’s second identity (9.14) holds for the exterior domain D− . This will yield an estimate on ∇S u, given the Sommerfield radiation condition.

Example 9.3.8 (Time-harmonic Maxwell’s equations). Consider a time-harmonic electromagnetic wave F in a spacetime with three space dimensions. As in Sec1/2 −1/2 tion 9.2 we have F = 0 e0 ∧ E + µ0 B ∈ 42 Wc solving (D + ike0 )F = 0, √ with wave number k := ω/c = ω 0 µ0 . To investigate the reproducing formula (9.11) for this bivector field F , we evaluate the time-like and space-like bivector

Chapter 9. Dirac Wave Equations

308

parts of the equation, and obtain two classical equations known as the Stratton– Chu formulas: Z Φk (x − y) ν(y) × E(y) dy E(x) = ∇ × Z ∂U −∇ Φk (x − y) ν(y) · E(y) dy Z∂U + ikc Φk (x − y) ν(y) × (∗B)(y) dy, ∂U Z Φk (x − y) ν(y) × (∗B)(y) dy ∗B(x) = ∇ × Z ∂U −∇ Φk (x − y) ν(y) · (∗B)(y) dy ∂U Z − ik/c Φk (x − y) ν(y) × E(y) dy. ∂U

For solutions to the time-harmonic Maxwell’s equations, the classical decay conuller radiation condition dition at ∞ is the Silver–M¨ ( x −(n−1)/2 (Im k)|x| e ), |x| × E(x) − c(∗B)(x) = o(|x| x c |x| × (∗B)(x) + E(x) = o(|x|−(n−1)/2 e(Im k)|x| ).

Since x |x|

+ e0 (e0 E + cB) = e0 cB −

x |x| E

x + − E + c |x| B ,

we see that the Dirac radiation condition for the electromagnetic field is equivalent to the Silver–M¨ uller radiation condition. Note that both radiation conditions also give decay of the radial parts of the vector fields E and ∗B. Given the Cauchy reproducing formulas for D + ike0 , we can extend the theory of Hardy subspaces from Section 8.3 to the case k 6= 0. Acting on functions h : ∂D → 4Wc we define traces of Cauchy integrals Z + Ek h(x) := lim Ψ− k (y − z)ν(y)h(y)dy, z→x,z∈D + ∂D Z Ek− h(x) := − lim Ψ− x ∈ ∂D, k (y − z)ν(y)h(y)dy, z→x,z∈D −

∂D

and the principal value Cauchy integral Z Ek h(x) := lim+ Ψ− k (y − x)ν(y)h(y)dy, →0

x ∈ ∂D.

∂D\B(x;)

As in the static case k = 0, we limit ourselves to proving splittings of H¨older spaces of multivector fields into Hardy subspaces.

9.4. Boundary Value Problems

309

Theorem 9.3.9 (Hardy wave subspace splitting). Let D = D+ ⊂ V be a bounded C 1 domain, with exterior domain D− = V \D, and let Im k ≥ 0. Consider the function space C α = C α (∂D; 4Wc ), for fixed regularity 0 < α < 1. Then the operators Ek+ , Ek− , and Ek are well defined and bounded on C α (∂D). The operators Ek± are complementary projections, with Ek± = 12 (I ± Ek ), and they split C α (∂D) into Hardy subspaces C α (∂D) = Ek+ C α ⊕ Ek− C α . There is a one-to-one correspondence, furnished by the Cauchy integral (9.11) and the trace map, between fields in the interior Hardy subspace Ek+ C α and fields in D+ solving DF + ike0 F = 0 and H¨ older continuous up to ∂D. Likewise, there is a one-to-one correspondence, furnished by the Cauchy integral (9.12) and the trace map, between fields in the exterior Hardy subspace Ek− C α and fields in D− solving DF + ike0 F = 0, H¨ older continuous up to ∂D and radiating at ∞. Proof. Define the operator Z Rk h(x) = (Ψ− k (y − x) − Ψ(y − x))ν(y)h(y)dy,

x ∈ V.

∂D

By the asymptotics at x = 0 from Exercise 9.3.3, Rk h(x) is a well-defined convergent integral for all x ∈ V . Furthermore, by differentiating under the integral sign, we have Z dy |∂j Rk h(x)| . khk∞ (9.16) . khk∞ ln(1/dist (x, ∂D)), |y − x|n−1 ∂D for 0 < dist (x, ∂D) < 1/2 and j = 1, . . . , n, where khk∞ := sup∂D |h|. Integrating (9.16) and (8.6) as in Proposition 8.3.3, it follows that Ek± are bounded on C α , since Ek± = E ± ± Rk . Note that integrating (9.16) in fact shows that Rk h is H¨older continuous across ∂D. Therefore it follows from Proposition 8.3.4 that Ek± h = E ± h ± Rk h = 21 h ± 12 (Eh + 2Rk h) = 12 (h ± Ek h) and that Ek = E + 2Rk is a bounded operator on C α . As in Theorem 8.3.6, we conclude that Ek+ + Ek− = I and that E ± are projections by Theorem 9.3.4 and Proposition 9.3.6 respectively. This proves the splitting into Hardy subspaces for D + ike0 .

9.4

Boundary Value Problems

For the remainder of this chapter, we study boundary value problems (BVPs) for Dirac operators, where our problem is to find a solution F to DF (x) + ike0 F (x) = 0

Chapter 9. Dirac Wave Equations

310

in a domain D that satisfies a suitable condition on the trace F |∂D . To make the problem precise, one needs to state assumptions on ∂D: How smooth is it? Is it bounded or unbounded? We also need to specify the space of functions on ∂D in which we consider F |∂D , and in what sense the boundary trace F |∂D is meant. To start with, we postpone these details, and assume only given a Banach space H of functions on ∂D. A concrete example is H = C α (∂D) from Theorem 9.3.9. We assume that the Cauchy integral operator Ek acts as a bounded operator in H, and we recall that Ek is a reflection operator, Ek2 = I, and it induces a splitting of H into Hardy wave subspaces. Solutions F to DF + ike0 F = 0 in D = D+ are in one-to-one correspondence with f = F |∂D in Ek+ H. A formulation of a Dirac BVP is ( DF + ike0 F = 0, in D, T f = g, on ∂D. Here T : H → Y is a given bounded and linear operator onto an auxiliary Banach function space Y, which contains the boundary datum g to the BVP. In such an operator formulation, well-posedness of the BVP means that the restricted map T : Ek+ H → Y

(9.17)

is an isomorphism. Indeed, if so, then for every data g ∈ Y we have a unique solution f ∈ Ek+ H, or equivalently a solution F to DF + ike0 F = 0 in D, which depends continuously on g. The main goal in studying BVPs is to prove such well-posedness. Almost as good is to prove well-posedness in the Fredholm sense, meaning that T is a Fredholm map. In this case, g needs to satisfy a finite number of linear constaints for f to exist, and this is unique only modulo a finite-dimensional subspace. Proposition 9.4.1. Let T : H → Y be a surjective bounded linear operator. Then the restriction T : Ek+ H → Y is an isomorphism if and only if we have a splitting Ek+ H ⊕ N(T ) = H. Proof. If T : Ek+ H → Y is an isomorphism, denote its inverse by T0 : Y → Ek+ H. Then P := T0 T : H → H is a projection with null space N(T ) and range Ek+ H, which proves the splitting. Conversely, if we have a splitting Ek+ H ⊕ N(T ) = H, then clearly T : Ek+ H → Y is injective and surjective. Without much loss of generality, we assume from now on that T is a bounded projection on H with range Y ⊂ H. We consider the following abstract formulation of the BVP, in terms of two bounded reflection operators A and B on H: A2 = I

and B 2 = I.

The operator A plays the role of the Cauchy integral Ek , so that A+ = 21 (I + A) projects onto traces of solutions to the differential equation in D+ and A− =

9.4. Boundary Value Problems

311

1 − 2 (I − A) projects onto traces of solutions to the differential equation in D , with appropriate decay at infinity. The operator B encodes two complementary boundary conditions: either T = B + = 12 (I + B) or T = B − = 12 (I − B) can be used to define boundary conditions. Note that we have null spaces N(B + ) = B − H and N(B − ) = B + H. We note that the algebra for each of the operators A and B is similar to that of Ek in Theorem 9.3.9. We have two different splittings of H:

H = A+ H ⊕ A− H

and H = B + H ⊕ B − H,

and A = A+ − A− and B = B + − B − . The core problem in the study of BVPs is to understand the geometry between on the one hand the subspaces A± H related to the differential equation, and on the other hand the subspaces B ± H related to the boundary conditions. Example 9.4.2 (BVP = operator 4R2 ). The algebra of two reflection operators A and B can be viewed as an operator version of the Clifford algebra 4R2 for the Euclidean plane R2 . Indeed, consider two unit vectors a, b ∈ V . Since a2 = b2 = 1 in 4R2 , we have here a very simple example of an abstract BVP. The geometry of a and b is described by the angle φ between the vectors. We recall that this angle can be calculated from the anticommutator 1 2 (ab

+ ba) = cos φ,

or from the exponential ab = eφj , where j is the unit bivector with orientation of a ∧ b. Definition 9.4.3 (Well-posedness). Let A, B : H → H be two reflection operators on a Banach space H. Define the cosine operator 1 2 (AB

+ BA)

and the rotation operators AB

and BA = (AB)−1 .

We say that the AB boundary value problems are well posed (in the Fredholm sense) if the four restricted projections B ± : A± H → B ± H are all isomorphisms (Fredholm operators). Exercise 9.4.4 (Simplest abstract BVP). Let H = C2 and consider the two orthogonal reflection operators 1 0 cos(2α) sin(2α) A= and B = , 0 −1 sin(2α) − cos(2α) for some 0 ≤ α ≤ π/2. Compute the cosine and rotation operators and show that the AB BVPs are well posed if and only if 0 < α < π/2. Show that we have spectra σ( 12 (AB + BA)) = {cos(2α)} and σ(AB) = {ei2α , e−i2α }, and that the AB BVPs fail to be well posed exactly when these spectra hit {+1, −1}.

312

Chapter 9. Dirac Wave Equations

Figure 9.2: The two splittings encoding an abstract BVP, with associated reflection operators. For two general reflection operators A and B, the associated cosine and rotation operators each contain the necessary information to conclude well-posedness of the AB BVPs. Useful identities include the following, which are straightforward to verify: 1 2 (I 1 2 (I

+ BA) = B + A+ + B − A− , +

−

−

+

− BA) = B A + B A ,

(9.18) (9.19)

2(I + C) = (I + BA)B(I + BA)B,

(9.20)

2(I − C) = (I − BA)B(I − BA)B.

(9.21)

Proposition 9.4.5 (Well-posedness and spectra). Let A, B : H → H be two reflection operators on a Banach space H. Then the following are equivalent: (i) The AB BVPs are well posed. (ii) The spectrum of the rotation operator BA does not contain +1 or −1.

9.4. Boundary Value Problems

313

(iii) The spectrum of the cosine operator C = 21 (AB + BA) does not contain +1 or −1. Similarly, the AB BVPs are well posed in the Fredholm sense if and only if I ±BA are Fredholm operators, if and only if I ± C are Fredholm operators. Proof. We note that B + A+ + B − A− is invertible if and only if the BVPs B + : A+ H → B + H and B − : A− H → B − H are well posed, and similarly for B + A− + B − A+ . Also ((I + BA)B)2 is invertible if and only if I + BA is invertible, and similarly for I − BA. The equivalences follow. With this general setup, and Proposition 9.4.5 as our main tool for proving well-posedness of Dirac BVPs, we now consider the two main examples that we have in mind. The boundary condition B, unlike A, is typically a pointwise defined multiplier, derived from the orientation of the tangent space to ∂D, described by the normal vector ν. For the remainder of this section we assume that D is a bounded C 2 domain. In this case, we note that ν is a C 1 smooth vector field on ∂D. We will see below that the cosine operators for such smooth BVPs tend to be compact, leading directly to BVPs that are Fredholm well posed by Proposition 9.4.5. Indeed, by the general Fredholm theory outlined in Section 6.4, the operators I ± C will then be Fredholm operators with index zero. The cosine operators typically are generalizations of the following classical integral operator from potential theory. Exercise 9.4.6 (Double layer potential). Consider the integral operator Z Kf (x) := hΨ(y − x), ν(y)if (y)dy, x ∈ ∂D, ∂D

with kernel k(x, y) = hΨ(y − x), ν(y)i. In three dimensions, a physical interpretation of k(x, y) is that of the electric potential from a dipole at y, in the direction ν(y), and for this reason K is called the double layer potential operator. The operator K is weakly singular on smooth domains. More precisely, show that on a C 2 boundary ∂D of dimension n − 1, we have kernel estimates |k(x, y)| . |x − y|2−n and |∇0x k(x, y)| . |x − y|1−n , x 6= y, x, y ∈ ∂D, where ∇0x denotes the tangential gradient in the x-variable. Lemma 9.4.7 (Weakly singular = compact). Let Z T f (x) = k(x, y)f (y)dy,

x ∈ ∂D,

∂D

be a weakly singular integral operator with kernel estimates |k(x, y)| . |x − y|2−n and |∇0x k(x, y)| . |x − y|1−n , x, y ∈ ∂D. Here ∇0x denotes the tangential gradient along ∂D in the variable x. Then T is a compact operator on C α (∂D) for all 0 < α < 1.

Chapter 9. Dirac Wave Equations

314

Proof. Assume that x, x0 ∈ ∂D with |x − x0 | = . Write T f (x0 ) − T f (x) Z Z (k(x0 , y) − k(x, y))f (y)dy + = |y−x|≤2

(k(x0 , y) − k(x, y))f (y)dy

|y−x|>2

=: I0 + I1 . For I1 , we obtain from the mean value theorem the estimate |k(x0 , y) − k(x, y)| . /|y − x|1−n when |y − x| > 2. This yields |I1 | . ln −1 kf kL∞ . For I0 , we estimate |k(x0 , y) − k(x, y)| . |y − x|2−n and obtain |I0 | . kf kL∞ . It follows that T : C α (∂D) → C β (∂D) is bounded for all β < 1, and that T : C α (∂D) → C α (∂D) is a compact operator. Example 9.4.8 (Normal/tangential BVP). Our main example of a Dirac BVP occurs when the differential equation is DF + ike0 F = 0, that is, A = Ek , for a fixed wave number k ∈ C, and the boundary conditions are encoded by the reflection operator B = N given by N f (x) := ν(x) 4 fd (x) 4 ν(x),

x ∈ ∂D.

We know from Section 4.1 that N reflects the multivector f (x) across the tangent plane to ∂D at x, and assuming that ν ∈ C 1 , we have that N is bounded on C α (∂D). The projection N + in this case will yield a boundary condition that specifies the part of f (x) tangential to ∂D in the sense of Definition 2.8.6. This can be verified using the Riesz formula (3.4), as N + f = 12 (f + ν fbν) = ν 12 (νf + fbν) = ν(ν ∧ f ) = ν y (ν ∧ f ). The corresponding calculation using (3.3) shows that N − f = ν ∧ (ν y f ) yields the boundary condition that specifies the normal part of f (x). The four Ek N BVPs consist of two BVPs for solutions to DF + ike0 F = 0 in the interior domain D+ , where the tangential or normal part of F |∂D is specified, and two BVPs for solutions to DF + ike0 F = 0 in the exterior domain D− , where the tangential or normal part of F |∂D is specified. By Proposition 9.4.5, the wellposedness of these four BVPs may be studied via the associated cosine operator Ek N +N Ek = (Ek +N Ek N )N . When k = 0, we calculate using Ψν = 2hΨ, νi−νΨ

9.4. Boundary Value Problems

315

that 1 2 (E

Z + N EN )f (x) = p.v.

Ψ(y − x)ν(y)f (y) + ν(x)Ψ(y − x)f (y)ν(y)ν(x) dy

Z∂D = 2p.v. hΨ(y − x), ν(y)if (y)dy ∂D Z + p.v. (ν(x) − ν(y))Ψ(y − x)f (y)dy ∂D Z + ν(x) p.v. Ψ(y − x)f (y)(ν(y) − ν(x))dy ν(x). ∂D

Assume now that D is a bounded C 2 domain. We can then apply Lemma 9.4.7 to each of these three terms, showing that EN + N E is a compact operator on C α . Moreover, the compactness of Ek − E on C α follows by yet another application of Lemma 9.4.7. We conclude that the Ek N BVPs are well posed in the sense of Fredholm in C α (∂D) for C 2 domains D. Example 9.4.9 (Spin BVP). The second example of a Dirac BVP that we shall consider is that in which the boundary conditions are induced by left Clifford multiplication by the normal vector ν. For technical reasons we study boundary conditions encoded by the reflection operator B = S given by Sf (x) := e0 4 ν(x) 4 f (x),

x ∈ ∂D.

Note that (e0 n)2 = −e20 n2 = 1, so indeed S is a reflection operator, and it is bounded on C α , since ν is C 1 regular. The factor e0 is motivated by Proposition 3.3.5 as in Proposition 9.1.5, and makes 4W ev invariant under S. As before, we study the differential equation DF + ike0 F = 0 encoded by the reflection operator A = Ek . It would be more natural to consider the operators Ek and S acting on spinor fields ∂D → 4W / , though, since both operators use only left multiplication by multivectors. So the true nature of the Ek S BVPs are BVPs for the 4-Dirac / operator. However, we here consider the 4-Dirac operator, since we aim to combine the Ek S and the Ek N BVPs in Section 9.5. The ranges of the projections S + f = 12 (1 + e0 ν)f

and S − f = 12 (1 − e0 ν)f

are seen to be the subspaces of multivector fields containing left Clifford factors that are respectively the light-like vectors ν ±e0 . The advantage of the S boundary conditions is that in some sense, the Ek S BVPs are the best local BVPs possible for the differential equation DF + ike0 F = 0. We will see several indications of this below.

Chapter 9. Dirac Wave Equations

316

For the cosine operator 21 (ES + SE), we calculate Z 1 (x) = p.v. Ψ(y − x)ν(y)f (y) + ν(x)Ψ(y − x)f (y) dy (E + SES)f 2 ∂D Z = 2p.v. hΨ(y − x), ν(y)if (y)dy Z∂D + p.v. (ν(x) − ν(y))Ψ(y − x)f (y)dy, ∂D

since e0 anticommutes with the space-like vectors ν and Ψ. As in Example 9.4.8, we conclude from this, using Lemma 9.4.7, that the Ek S BVPs are well posed in the sense of Fredholm in C α on C 2 domains. Having established well-posedness in the Fredholm sense for the Ek N and Ek S BVPs, we know that the BVP maps (9.17) are Fredholm operators, so that the null spaces are finite-dimensional and the ranges are closed subspaces of finite codimension. It remains to prove injectivity and surjectivity, whenever possible. Proposition 9.4.10 (Injectivity). Let 0 < α < 1 and Im k ≥ 0. • For the Ek N BVPs we have Ek+ C α ∩N + C α = Ek+ C α ∩N − C α = Ek− C α ∩N + C α = Ek− C α ∩N − C α = {0} if Im k > 0. Moreover, if D− is a connected domain and k ∈ R \ {0}, then Ek− C α ∩ N + C α = Ek− C α ∩ N − C α = 0. • For the Ek S BVPs we have Ek+ C α ∩ S + C α = Ek− C α ∩ S − C α = {0} whenever Im k ≥ 0. Proof. For the estimates we require the Hermitian inner product (w1 , w2 iV := (e0 w b1 e−1 0 , w2 i on 4Wc from Definition 9.3.1. Consider first the interior BVPs. Given f = F |∂D ∈ Ek+ C α , we define the linear 1-form D × V → C : (y, v) → (e0 vF (y), F (y)iV , which has nabla derivative (e0 ∇F (y), ˙ F (y)i ˙ V = (e0 (∇F ), F iV − (e0 F, ∇F iV = (e0 (−ike0 F ), F iV − (e0 F, −ike0 F iV = −2Im k|F |2V . From the Stokes formula (7.4), it follows that Z Z (Sf, f iV dy = −2Im k ∂D

D+

|F |2V dx.

9.4. Boundary Value Problems

317

If f ∈ N ± C α , then (Sf, f iV = 0, and we conclude that F = 0 if Im k > 0. So in this case, Ek+ C α ∩ N ± C α = {0}. If f ∈ S + C α , then (Sf, f iV = |f |2V , and we conclude that f = 0 whenever Im k ≥ 0, so Ek+ C α ∩ S + C α = {0}. Consider next the exterior BVPs. Let f = F |∂D ∈ Ek− C α , and fix a large − := D− ∩ {|x| < R}, radius R. From Stokes’s theorem applied to the domain DR we have Z Z Z x (e0 |x| F, F iV dx − (Sf, f iV dy = −2Im k |F |2V dx. |x|=R

− DR

∂D

Furthermore, on the sphere |x| = R, we note that x x |( |x| + e0 )F |2V = 2|F |2V − 2(e0 |x| F, F iV ,

and obtain the identity Z Z x (|F |2V − 12 |( |x| + e0 )F |2V )dx − |x|=R

Z (Sf, f iV dy = −2Im k

∂D

− DR

|F |2V dx.

R x Using Proposition 9.3.6, we have limR→∞ |x|=R |( |x| +e0 )F |2V dx = 0 for all Im k ≥ − α 2 0. If f ∈ S C , then (Sf, f iV = −|f |V , and we again conclude that f = 0, so Ek− C α ∩ S − C α = {0}. If f ∈ N ± C α , then (Sf, f iV = 0, and we have Z Z Z x |F |2V dx + 2Im k |F |2V dx = 12 |( |x| + e0 )F |2V dx → 0, R → ∞. |x|=R

− DR

|x|=R

When Im k > 0, this shows that F = 0. When k ∈ R \ {0}, we have Z lim |F |2V dy = 0. R→∞

|x|=R

Applying Rellich’s lemma (Lemma 6.3.5) to the component functions Fs of F , which satisfy Helmholtz’s equation ∆Fs + k 2 Fs = 0, we also in this case conclude that F = 0, so in either case, Ek− C α ∩ N ± C α = {0}. Summarizing our findings, we have obtained the following well-posedness results. Theorem 9.4.11 (C α well-posedness). For the Dirac BVPs with boundary function space C α (∂D), 0 < α < 1, on domains with C 2 regular boundary ∂D, we have the following well-posedness results. The four BVPs N ± : Ek± C α → N ± C α are well posed when Im k > 0. If the exterior domain D− is connected, then the exterior BVPs N ± : Ek− C α → N ± C α are well posed for all nonzero Im k ≥ 0. The two spin-Dirac BVPs S − : Ek+ C α → S − C α and S + : Ek− C α → S + C α are well posed for all Im k ≥ 0.

318

Chapter 9. Dirac Wave Equations

We remark that by applying analytic Fredholm theory, one can prove that in fact, also the interior Ek N BVPs are well posed for k ∈ R, except for a discrete set of resonances. Proof. We make use of the Fredholm theory outlined in Section 6.4. By Example 9.4.8 and Proposition 9.4.5, the Ek N BVPs are well posed in the Fredholm sense for all k. By Proposition 9.4.10 the four maps N ± : Ek± C α → N ± C α are injective when Im k > 0. We conclude that I ± 12 (Ek N + N Ek ) are injective Fredholm operators with index zero, and therefore invertible. So the Ek N BVPs are well posed when Im k > 0. For k ∈ R\{0}, we have injective semi-Fredholm maps N ± : Ek− C α → N ± C α by Proposition 9.4.10. By perturbing Ek− to Im k > 0, Lemma 9.4.12 below proves that they are invertible. The well-posedness of S − : Ek+ C α → S − C α and S + : Ek− C α → S + C α follows from Example 9.4.8 and Proposition 9.4.10, using Proposition 9.4.5. Note that I − 12 (Ek S + SEk ) = ((S − Ek+ + S + Ek− )S)2 is an injective Fredholm operator with index zero, and hence invertible. The following two techniques for proving existence of solutions to BVPs turn out to be useful. Lemma 9.4.12 (Perturbation of domains). Let At , t ∈ [0, 1], and B be reflection operators on a Banach space H, and consider the family of BVPs described by B + : + A+ t H → B H. If these are all semi-Fredholm maps and if t 7→ At is continuous, + + + + then the indices of B + : A+ 0 H → B H and B : A1 H → B H are equal. + Proof. We parametrize the domains A+ t H by the fixed space A0 H. Considering + + + + ˜ At := At : A0 H → At H as one of the four abstract A0 At BVPs, we note that

I + At A0 = 2I + (At − A0 )A0 . If kAt −A0 k ≤ 1/kA0 k, it follows that I +At A0 is invertible, and from (9.18) we see + + + ˜+ in particular for 0 ≤ t ≤ , that A˜+ t is invertible. Let Bt := B : At H → B H. + + + + ˜t A˜t , we conclude that Ind(B ˜ A˜ ) = Applying the method of continuity to B + + ˜ + A˜+ ). Since A˜+ ˜ ˜ Ind(B are invertible, we obtain Ind( B ) = Ind( B ). Repeating t 0 0 0 ˜ + ) = Ind(B ˜ + ). this argument a finite number of times, we conclude that Ind(B 1 0 Lemma 9.4.13 (Subspace duality). Let A and B be two reflection operators on a Banach space H, and consider the BVP described by B + : A+ H → B + H. This map is surjective if and only if the dual BVP described by (B ∗ )− : (A∗ )− H∗ → (B ∗ )− H∗ is an injective map. Proof. Note that A∗ and B ∗ are reflection operators in H∗ . By duality as in Section 6.4, we have (A+ H)⊥ = R(A+ )⊥ = N((A∗ )+ ) = R((A∗ )− ) = (A∗ )− H∗ and similarly (B − H)⊥ = (B ∗ )+ H∗ . Similarly to Proposition 9.4.1, since (A+ H + B − H)⊥ = (A+ H)⊥ ∩ (B − H)⊥ , this translates to the claim.

9.5. Integral Equations

319

We end this section with two applications of the techniques in this section to Dirac’s equation. Example 9.4.14 (The MIT bag model). Consider Dirac’s equation i~∂t ψ = H0 ψ from Proposition 9.2.8 on a bounded domain D ⊂ V . The MIT bag model is used in physics to describe the quarks in a nucleon, that is, a proton or neutron. The bag D represents the nucleon, and the boundary condition is ν.ψ = ψ, 2

or in the 4V representation e0 ν.ψ = ψ. This boundary condition implies in particular that the probability current hjp , νi = i(ψ, ν.ψi = i(ψ, ψi across ∂D vanishes, since jp is a real spacetime vector field. We see that with suitable modifications, such BVPs for time-harmonic solutions to Dirac’s equation can be studied with the methods described in this section. Example 9.4.15 (Chirality of (anti-)particles). What we refer to here as abstract BVPs, namely the algebra of two reflection operators describing the geometry between two splittings of a function space, appear in many places independent of any BVPs. One of many such examples we saw in connection to Proposition 9.2.8. Consider the Hilbert space H := L / c2 ), where we saw two different split2 (V ; 4V I 0 tings. The reflection operator B = encodes the Chiral subspaces of right0 −I and left-handed spinors, whereas / −i D m ˜ A = sgn(H0 ) = √ / . ˜ −D m ˜ 2 − ∆ −m / by Pauli matrices, the Fourier mulUsing, for example, the representation of 4V tiplier of the rotation operator AB at frequency ξ ∈ V is seen to have the four eigenvalues p λ = (±|ξ| ± im)/ ˜ ˜ 2. |ξ|2 + m Therefore the spectrum of AB is precisely the unit circle |λ| = 1. We conclude that although the spectral subspaces L± / 2 ) do not intersect the chiral subspaces, 2 (V ; 4V the angle between them is zero. The problem occurs at high frequencies: particles or antiparticles of high energy may be almost right- or left-handed.

9.5

Integral Equations

The aim of this section is to use the somewhat abstract theory from Section 9.4 to derive integral equations for solving Dirac BVPs, with good numerical properties, that have recently been discovered.

320

Chapter 9. Dirac Wave Equations

• It is desirable to extend the theory to nonsmooth domains, which have boundaries that may have corners and edges, as is often the case in applications. Ideally, one would like to be able to handle general Lipschitz domains. • To solve a given BVP, we want to have an equivalent integral formulation Z k(x, y)f (y)dy = g(x), x ∈ ∂D, ∂D

where the boundary datum gives g and the integral equation is uniquely solvable for f if and only if the BVP to solve is well posed. Ideally we want to have a function space without any constraints, meaning a space of functions ∂D → L with values in a fixed linear space L and coordinate functions in some classical function space. In this section we let D be a bounded strongly Lipschitz domain. At this generality, the normal vector field ν is only a measurable function without any further smoothness. To extend the theory from Section 9.4 and keep the basic operators Ek , N and S bounded, we shall use L2 = L2 (∂D; 4Wc ), which is the most fundamental space to use for singular integral operators like Ek . Indeed, the singular integral operator Ek is bounded on L2 (∂D) for every Lipschitz domain D, by Theorem 8.3.2 and Exercises 9.3.3 and 6.4.3. We first consider Fredholm well-posedness of the Ek N BVPs in L2 on bounded strongly Lipschitz domains. On such nonsmooth domains, it is not true in general that Ek N + N Ek , or even the classical double layer potential from Exercise 9.4.6, is compact. However, we recall from Proposition 9.4.5 that it suffices to show that the spectrum of 12 (Ek N + N Ek ) does not contain ±1. Theorem 9.5.1 (Rellich estimates). Let D be a bounded strongly Lipschitz domain, and let θ be a smooth compactly supported field that is transversal to ∂D as in Exercise 6.1.8. Define the local Lipschitz constant L := sup∂D (|θ ∧ ν|/hθ, νi) for ∂D. Then λI + Ek N is a Fredholm operator on L2 (∂D) of index zero whenever λ = λ1 + iλ2 , |λ2 | < |λ1 |/L, λ1 , λ2 ∈ R. Note that since Ek N and (Ek N )−1 = N Ek are bounded, we also know that the spectrum of Ek N is contained in an annulus around 0. Furthermore, since ((λI + Ek N )Ek )2 = λ(λ + λ−1 + Ek N + N Ek ), the resolvent set of the cosine operator contains the hyperbolic regions onto which λ 7→ 12 (λ + λ−1 ) maps the double cone |λ2 | < |λ1 |/L. And for λ = ±1 it follows in particular that the Ek N BVPs are well posed in the Fredholm sense in L2 (∂D).

9.5. Integral Equations

321

Proof. To motivate the calculations to come, we consider first the BVP described by N + : Ek+ L2 → N + L2 . To estimate kf kL2 in terms of kN + f kL2 , we insert the factor hθ, νi and express it with the Clifford product as Z Z Z 1 2 2 kf kL ≈ |f | hθ, νidy = (f, f (θν + νθ)i dy = Re (f ν, f θiV dy. V V 2 2 ∂D ∂D ∂D We next use the reversed twin of the Riesz formula (3.4) to write f ν = 2f ∧ ν − ν fb. We estimate the last term so obtained by applying Stokes’s theorem with the linear 1-form (y, v) 7→ (v fd (y), f (y)θ(y)iV , giving Z (ν fb, f θiV dy = ∂D

Z

(−ike0 fb, f θiV + (fb, (−ike0 f )θiV +

D

n X (fb, ej f (∂j θ)iV dy j=1

(9.22) Combining and estimating, we get kf k2L2 (∂D) . kf

∧

νkL2 (∂D) kf kL2 (∂D) + kF k2L2 (Dθ ) ,

where Dθ := D ∩ supp θ. The Cauchy integral L2 (∂D) → L2 (Dθ ) : f 7→ F can be shown to be a bounded operator by generalizing the Schur estimates from Exercise 6.4.3 to integral operators from ∂D to Dθ . Moreover, such estimates show by truncation of the kernel that this Cauchy integral is the norm limit of Hilbert–Schmidt operators, and hence compact. On the first term we can use the 1 absorption inequality kN + f kkf k ≤ 2 kN + f k2 + 2 kf k2 . Choosing small leads to + + a lower bound, showing that N : Ek L2 → N + L2 is a semi-Fredholm operator. Next consider the integral equation λh + Ek N h = g, where we need to estimate khkL2 in terms of kgkL2 . To this end, we note that Ek N h = g − λh, so that Ek± L2 3 f ± := 2Ek± N h = N h ± (g − λh). Applying (9.22) to f + and the corresponding application of Stokes’s theorem to f − , we obtain estimates Z ± ± b (ν f , f θiV dy . kF k2L2 (supp θ) . ∂D

We now expand the bilinear expressions on the left, writing f ± = N h ∓ λh ± g, and observe that the integrals Z Z (ν b h, hθiV dy and (νN b h, N hθiV dy ∂D

∂D

are bad in the sense that we have only an upper estimate by khk2L2 , whereas the terms Z Z (νN b h, λhθiV dy = λ (h, hθνiV dy ∂D

∂D

Chapter 9. Dirac Wave Equations

322

c R R and ∂D (νλb h, N hθiV dy = λ ∂D (h, hθνiV dy are good in the sense that they are comparable to khk2L2 . To avoid the bad terms, we subtract identities and obtain Z 2 b+ , f + θiV − (ν fb− , f − θiV dy f kF kL & (ν 2 (supp θ) ∂D Z & 2 Re λ (h, hθνiV dy − khkL2 kgkL2 − kgk2L2 . ∂D

Writing θν = hθ, νi + that |θ2 ∧ ν| ≤ Lhθ, νi for some L < ∞. It R θ ∧ ν, we know follows that 2 Re λ ∂D (h, hθνiV dy & khkL if |λ2 | < |λ1 |/L, and we conclude 2 that in this case, λI + Ek N is a semi-Fredholm operator. That it is a Fredholm operator with index zero follows from the method of continuity, by perturbing λ, for example, to 0, where Ek N is an invertible operator. Theorem 9.5.2 (L2 well-posedness for Ek N ). For the Ek N Dirac BVPs with boundary regularity L2 (∂D) on bounded strongly Lipschitz domains D, we have the following well-posedness results. The four BVPs N ± : Ek± L2 → N ± L2 are well posed for all Im k ≥ 0 except for a discrete set of real k ≥ 0. If the exterior domain D− is connected, then the exterior BVPs N ± : Ek− L2 → N ± L2 are well posed for all nonzero Im k ≥ 0. Proof. By Theorem 9.5.1 and Proposition 9.4.5, the Ek N BVPs are well posed in the Fredholm sense for all k. Proposition 9.4.10 can be verified when the C α topology is replaced by L2 . The proof can now be completed as in Theorem 9.4.11. For the remainder of this section we consider the second problem posed above, namely how to formulate a given Dirac BVP as an integral equation that is good for numerical applications. As a concrete example, we take the exterior BVP with prescribed tangential part, that is, N + : Ek− L2 → N + L2 .

(9.23)

This BVP has important applications, since a solution of it yields an algorithm for computing, for example, how acoustic and electromagnetic waves are scattered by an object D. Assuming that the exterior domain D− is connected and Im k ≥ 0, k 6= 0, we know that this BVP is well posed. Although it is an invertible linear equation by which we can solve the BVP, it is not useful for numerical applications. The reason is that the solution space Ek− L2 is defined by a nonlocal constraint on f ∈ L2 . What we need is an ansatz, meaning some operator U : Y → Ek− L2 , where Y is a function space that is good for numerical purposes and U has good invertibility properties. Using such a U , we can solve the BVP (9.23) by solving N +U h = g

9.5. Integral Equations

323

for h ∈ Y. This gives the solution f = U h ∈ Ek− L2 . As a first try, we swap the roles of Ek and N and consider U = Ek− : N + L2 → Ek− L2 . This leads to the operator N + Ek− |N + L2 , which can be shown to be closely related to the double layer potential operator from Exercise 9.4.6. The function space Y = N + L2 is good, but although this U is a Fredholm operator, it fails to be invertible for a discrete set of real k. Indeed, N(U ) = N + L2 ∩ Ek+ L2 will contain the eigenvalues of the self-adjoint operator −ie0 D with tangential boundary conditions on the bounded domain D+ . This explains a well-known problem in the numerical solution of BVPs by integral equations: the existence of spurious interior resonances k, where the integral equation fails to be invertible, even though the BVP it is used to solve is itself well posed. A better try, which should be more or less optimal, comes from the Ek S BVPs. Swapping the roles of Ek and S, we consider U = Ek− : S + L2 → Ek− L2 . Similarly, a good ansatz for an interior Dirac BVP is U = Ek+ : S − L2 → Ek+ L2 . It is important not to swap S + and S − in these ansatzes. Maybe the best way to see that the Ek S BVPs have well-posedness properties superior to those for the Ek N BVPs on L2 , even in the Fredholm sense and in particular on Lipschitz domains, is to consider the rotation operator Z Ek Sf (x) = 2p.v. x ∈ ∂D. Ψ− k (y − x)f (y)dy, ∂D

2

Note that we used that ν = 1. Since Ek −E is a weakly singular integral operator, it is compact on L2 (∂D), and when k = 0, we note that ES is a skew-symmetric operator, since Ψ is a space-like vector depending skew-symmetrically on x and y. In particular, this means that the spectrum of ES is on the imaginary axis and the operators I ± ES are invertible with k(I ± ES)−1 k ≤ 1. By the identities from Proposition 9.4.5, this means, for example, that kE − hk = 21 k(I − Ek S)hk ≥ 12 khk,

h ∈ S + L2 .

For general k, we note that there is still a major difference in well-posedness properties of the Ek S BVPs as compared to those for the Ek N BVPs. The operator λI + Ek S can fail to be Fredholm only when Re λ = 0, whereas λI + Ek N can fail to be Fredholm whenever | Re λ| ≤ L|Im λ|, not far away from λ = ±1 for large L. So, as compared to the Ek N BVPs, the well-posedness properties for the Ek S BVPs do not essentially depend on the Lipschitz geometry of ∂D. Theorem 9.5.3 (L2 well-posedness for Ek S). For the Ek S Dirac BVPs with boundary regularity L2 (∂D) on bounded strongly Lipschitz domains D, we have the following well-posedness results. The two spin-Dirac BVPs S − : Ek+ L2 → S − L2 and

Chapter 9. Dirac Wave Equations

324

S + : Ek− L2 → S + L2 are well posed for all Im k ≥ 0. Equivalently, the ansatzes Ek− : S + L2 → Ek− L2 and Ek+ : S − L2 → Ek+ L2 are invertible for all Im k ≥ 0. Proof. As before, we note the identity 21 (I − Ek S) = Ek+ S − + Ek− S + and its twin S + Ek− + S − Ek+ = 12 (I − SEk ) = 12 (Ek S − I)SEk . From the discussion above it follows that I − Ek S is a Fredholm operator of index 0, which directly shows that the two BVPs and the two ansatzes are Fredholm maps. By Proposition 9.4.10 adapted to L2 , the four maps are injective for all Im k ≥ 0. Therefore I − Ek N is injective, hence surjective. We conclude that the two BVPs and the two ansatzes are invertible. Example 9.5.4 (Asymptotic APS BVPs). Consider the Cauchy reflection operator A = Ek encoding the Dirac equation DF + ike0 F = 0, together with the abstract boundary conditions B = El , where k, l ∈ C. Clearly not all four Ek El BVPs are well posed, since El − Ek is a compact operator. However, since 1 2 (I

+ El Ek ) = El+ Ek+ + El− Ek−

clearly is a Fredholm operator with index zero, the two BVPs El+ : Ek+ L2 → El+ L2 and El− : Ek− L2 → El− L2 are Fredholm operators. Such BVPs with nonlocal boundary conditions defined by the differential equation itself are essentially the boundary conditions employed by Atiyah, Patodi, and Singer (APS) in their work on index theory for manifolds with boundary. We next let l → ∞ along the upper imaginary axis. The operators El are not norm convergent, but for a fixed function h, one can show that El h → −Sh. Note from the formula for Ψ− l how the singular integral operators El localize to the pointwise multiplier −S. This shows that indeed, the operator S is related to the differential equation as a local asymptotic Cauchy singular integral, and to some extent explains why the Ek S BVPs are so remarkably well posed. Example 9.5.5 (Spin integral equation). We now return to the exterior Dirac BVP (9.23) with prescribed tangential parts, which we know is well posed whenever Im k ≥ 0, k 6= 0, and D− is connected. Using the invertible ansatz Ek− : S + L2 → Ek− L2 from Theorem 9.5.3, we can solve the BVP (9.23), given datum g ∈ N + L2 , by solving N + Ek− h = g (9.24) for h ∈ S + L2 , giving the solution f = Ek− h ∈ Ek− L2 and Z F (x) = Ψ− x ∈ D− , k (y − x)(−ν(y))h(y)dy, ∂D

solving DF + ike0 F = 0 in D− with N + F |∂D = g. This is certainly numerically doable, since both spaces N + L2 and S + L2 are defined by a simple pointwise

9.5. Integral Equations

325

constraint determined by the normal ν. However, we can enhance the integral equation somewhat as follows. Consider the reflection operator T given by T f = −e0 fbe0 . We note that, similarly to N , replacing ν by the time-like vector e0 , indeed T 2 = I and T reflects time-like multivectors in the subspace of space-like multivectors. Computing relevant cosine operators, we have b (T S + ST )f = −e0 (e[ 0 νf )e0 + e0 ν(−e0 f e0 ) = 0, (T N + N T )f = −e0 νf νe0 − νe0 f e0 ν 6= 0, b (N S + SN )f = ν(e[ 0 νf )ν + e0 ν(ν f ν) = 0. By Proposition 9.4.5, this means that we have optimally well posed abstract BVPs T S and N S. In particular, this allows us to parametrize the domain space S + L2 of the integral equation (9.24) for example by T + L2 = L2 (∂D; 4Vc ), is an ideal space for applications. the space of space-like multivector fields, which √ In fact, we verify that S + : T + L2 → S + L2 is 1/ 2 times an isometry. Since T N + N T 6= 0, we cannot directly parametrize the range space N + L2 of (9.24) by T + L2 . However, we can go via the splitting L2 = S + L2 ⊕ S − L2 , since for example, T + S + : N + L2 → T + L2 √ is invertible. In fact, both S + : N + L2 → S + L2 and T + : S + L2 → T + L2 are 1/ 2 times isometries. To summarize, we propose that the exterior BVP (9.23) with prescribed tangential part is best solved using the integral equation T + S + N + Ek− S + h = T + S + g, for h ∈ T + L2 . Indeed, the derivation above shows that this integral equation is uniquely solvable, and the function space for the variable h and the datum T + S + g is simply T + L2 = L2 (∂D; 4Vc ). To write out this equation more explicitly, we compute that T + S + g = 12 (g0 + νg1 ), when g = g0 + e0 g1 and g1 , g2 ∈ N + L2 ∩ T + L2 , so the time-like part is mapped onto a normal part when the original multivector is tangential. We also compute that T + S + N + S + T + = 14 T + . Writing Ek− = 12 (I − Ek ), the integral equation for h ∈ L2 (∂D; 4Vc ) becomes Z 1 + M (x)p.v. Ψ− x ∈ ∂D. h(x) k (y − x)(ν(y) − e0 )h(y)dy = 2M (x)g(x), 2 ∂D

Chapter 9. Dirac Wave Equations

326

Here M denotes the multiplier that projects onto tangential multivectors and maps tangential time-like multivectors onto normal space-like multivectors by replacing a left factor e0 into ν. We refer to this integral equation as a spin integral equation for solving the BVP (9.23), since the key feature is that it uses an ansatz derived from the Ek S BVPs, which, as we have discussed in Example 9.4.9, really are / + ike0 ψ = 0. equation Dψ / BVPs for the 4-Dirac Example 9.5.6 (Transmission problems). Transmission problems generalize boundary value problems in that we look for a pair of fields F + : D+ → 4Wc and F − : D− → 4Wc such that DF + + ik2 e0 F + = 0, in D+ , (9.25) DF − + ik1 e0 F − = 0, in D− , M f + = f − + g, on ∂D. Here the wave numbers k1 , k2 ∈ C are different in the two domains, with Im k1 ≥ 0 and Im k2 ≥ 0. The relation between the traces f + = F + |∂D and f − = F − |∂D on ∂D is described by a multiplier M ∈ L(L2 ) and a given source g ∈ L2 . For solving the transmission problem (9.25), unlike in the case of BVPs, we have a good ansatz directly available, namely U : L2 → Ek+2 L2 ⊕ Ek−1 L2 : h 7→ (Ek+2 h, Ek−1 h). In the case k1 = k2 , it is clear from the L2 analogue of Theorem 9.3.9 that U is invertible. What is somewhat surprising is that U is invertible for all Im k1 ≥ 0 and Im k2 ≥ 0. To prove this, it suffices by the method of continuity to show that U is injective. To this end, note that U h = 0 means that h = F + |∂D = F − |∂D , − − + − where DF + + ik1 e0 F + = 0 in R D and DF + ik2 e0 F = 0 in D . Applying Stokes’s theorem twice to ∂D (e0 νh, hiV dy, computations as in the proof of Proposition 9.4.10 give Z Z Z 2Im k1 |F − |2V dx + |F + |2V dx + 2Im k2 |F − |V2 dx − DR

D+

=

1 2

Z |x|=R

|x|=R

x |( |x| + e0 )F − |2V dx.

Using radiation conditions and jumps, this shows that F + = F − = 0 and therefore h = 0. Using this invertible ansatz U , we can now solve the transmission problem (9.25) by solving the integral equation (M Ek+2 − Ek−1 )h = g

9.6. Boundary Hodge Decompositions

327

for h ∈ L2 . Note that this is an integral equation in L2 (∂D; 4Wc ) without any constraints. From the solution h, we finally compute the field Z Ψk−2 (y − x)ν(y)h(y)dy, F + (x) = ∂D Z − F (x) = − Ψk−1 (y − x)ν(y)h(y)dy, ∂D

solving the transmission problem. In Section 9.7, we apply this integral equation for Dirac transmission problems to solve scattering problems for electromagnetic waves.

9.6

Boundary Hodge Decompositions

We have considered Dirac BVPs in the previous sections and how to solve them by integral equations. Returning to Examples 9.3.7 and 9.3.8, one important issue remains. We saw there that both the Helmholtz equation and Maxwell’s equations can be viewed as special cases of the Dirac equation DF + ike0 F = 0. However, in these examples F is a vector field and a bivector field respectively, and not a general multivector field. If we intend, for example, to solve BVPs for Helmholtz’s or Maxwell’s equations by a spin integral equation as in Example 9.5.5 or a transmission problem with a Dirac integral equation as in Example 9.5.6, then we need a tool to ensure that the solution multivector field F is in fact a vector or bivector field. It turns out that there exists an exterior/interior derivative operator acting on multivector fields ∂D → 4Wc , which we shall denote by Γk , which is the tool needed. Applications to Maxwell scattering are found in Section 9.7. The point of departure for our explanations is Proposition 8.1.5, where we noted that for a monogenic field ∇ 4 F = 0, each of its homogeneous component functions Fj is monogenic if and only if ∇ ∧ F = 0 = ∇ y F . Generalizing this to time-harmonic waves with wave number k ∈ C, we have the following. Lemma 9.6.1 (Two-sided k-monogenic fields). Assume that F : D → 4Wc solves DF + ike0 F = 0 in some open set D ⊂ V . Write F = F0 + F1 + · · · + Fn+1 , where Fj : D → 4j Wc . Then DFj + ike0 Fj = 0 in D for all 0 ≤ j ≤ n + 1 if and only if ( dF + ike0 ∧ F = 0, δF + ike0 y F = 0. The way we use this result is that if we construct F solving DF + ike0 and some BVP, and if dF + ike0 ∧ F = 0, then we can conclude, for example, that F2 is a bivector field solving the Dirac equation, since the homogeneous parts of F decouple, and thus F2 is an electromagnetic field satisfying Maxwell’s equations. Proof. If (∇ + ike0 ) ∧ F = 0 = (∇ + ike0 ) y F , then (∇ + ike0 ) ∧ Fj = ((∇ + ike0 ) ∧ F )j+1 = 0

Chapter 9. Dirac Wave Equations

328

and (∇ + ike0 ) y Fj = ((∇ + ike0 ) y F )j−1 = 0, and so (∇ + ike0 ) 4 Fj = (∇ + ike0 ) y Fj + (∇ + ike0 ) ∧ Fj = 0 for all j. Conversely, if (∇ + ike0 ) 4 Fj = 0 for all j, then (∇ + ike0 ) ∧ Fj = ((∇ + ike0 ) 4 Fj )j+1 = 0 and (∇ + ike0 ) y Fj = ((∇ + ike0 ) 4 Fj )j−1 = 0. Summing over j, we obtain (∇ + ike0 ) ∧ F = 0 = (∇ + ike0 ) y F . To proceed with the analysis, we need to choose a function space. Since our theory for Hodge decompositions as well as for spin integral equations is set in Hilbert spaces, we choose L2 (∂D). Definition 9.6.2 (Boundary Γk operator). Consider the Hardy space splitting L2 (∂D) = Ek+ L2 ⊕ Ek− L2 on a strongly Lipschitz domain. Define the operator Γk by Γk f := g + + g − , where f = Ek+ f + Ek− f , F ± denote the Cauchy integrals of f in D± so that Ek± f = F ± |∂D , and g ± = G± |∂D ∈ Ek± L2 are such that their Cauchy integrals equal G± = (∇ + ike0 ) ∧ F ± in D± . The domain of Γk is the set of f for which such g ± exist. In a series of lemmas, we derive below a more concrete expression for this unbounded operator Γk as a tangential differential operator on L2 (∂D). It turns out that Γk acts by exterior differentiation along ∂D on tangential fields and by interior differentiation along ∂D on normal fields, modulo zero order terms determined by k. Definition 9.6.3 (Tangential derivatives). Consider the Lipschitz boundary M = ∂D, which is a Lipschitz manifold in the sense that the transition maps, as in Section 6.1, are Lipschitz regular. As in Definitions 11.2, 11.2.6, 12.1.1 and extending to Lipschitz regularity as in Section 10.2, we define tangential exterior and interior derivative operators d0 and δ 0 in L2 (M ; ∧M ), such that (d0 )∗ = −δ 0 . In the notation of this chapter, complexifying the bundle ∧M to ∧Mc , we have N + L2 = {f1 + e0 ∧ f2 ; f1 , f2 ∈ L2 (M ; ∧Mc )}, and extending d0 and δ 0 to operators in N + L2 acting as d0 f := d0 f1 − e0 ∧ (d0 f2 ), δ 0 f := δ 0 f1 − e0 ∧ (δ 0 f2 ), on f = f1 + e0 ∧ f2 , with f1 , f2 ∈ L2 (M ; ∧M ). The reader is kindly advised to consult the relevant sections of the following chapters, as indicated in Definition 9.6.3, for further details. Note that the minus sign in the actions on N + L2 occurs because the time-like e0 and the formally space-like tangential ∇0 anticommute.

9.6. Boundary Hodge Decompositions

329

Lemma 9.6.4 (Ek± L2 to N + L2 ). If f ∈ Ek+ L2 ∩ D(Γk ), then N + f ∈ D(d0 ) and d0 (N + f ) + ike0 ∧ (N + f ) = N + (Γk f ). The same holds for f ∈ Ek− L2 ∩ D(Γk ). Proof. Let f = F |∂D , where F is the Cauchy extension of f . Write f = f1 + e0 ∧ f2 and F = F1 + e0 ∧ F2 , where fj and Fj are space-like fields, j = 1, 2. Generalizing Exercise 11.2.3, with methods as in Lemma 10.2.4, to Lipschitz regular hypersurfaces, we have N + fj = ρ∗ Fj , where ρ : ∂D → V denotes the embedding of ∂D into V . The commutation theorem shows that d0 ρ∗ Fj = ρ∗ (dFj ), giving d0 (N + f ) = d0 ρ∗ F1 − e0 ∧ d0 ρ∗ F2 = N + (dF ). This proves the first statement, since e0 ∧ N + f = N + (e0 ∧ f ), and the proof for Ek− L2 is similar. Using Hodge star dualities, we next derive the corresponding result for the normal part. This uses left Clifford multiplication by ν, which is an isometry between N − L2 and N + L2 . Lemma 9.6.5 (Ek± L2 to N − L2 ). If f ∈ Ek+ L2 ∩ D(Γk ), then ν y f = νN − f ∈ D(δ 0 ) and δ 0 (ν y f ) + ike0 y (ν y f ) = ν y (Γk f ). The same holds for f ∈ Ek− L2 ∩ D(Γk ). Proof. Using nabla calculus with ∇k := ∇ + ike0 , given f = F |∂D ∈ Ek+ L2 ∩ D(Γk ), we write for example DF + ike0 F = 0 as ∇k F = ∇k 4 F = 0. Extending Proposition 8.1.13, such solutions form a right Clifford module, so G = F w = F ∗ = F y w, writing w = e012···n for the spacetime volume element, with dual volume element w∗ = −w ∈ 4n W ∗ , is also a solution to ∇k G = 0 in D+ . Moreover, ∗(∇k ∧ G) = −w x (∇k ∧ G) = (−w x G) x ∇k = F x ∇k = ∇k y F , making use of the algebra from Section 2.6. By Lemma 9.6.4 applied to G, we have N + (∇k ∧ G)|∂D = ∇0k ∧ (N + g) with g = G|∂D , writing d0 formally with nabla calculus using ∇0k = ∇0 + ike0 along ∂D. The spacetime Hodge dual of the left-hand side is ∗(N + (∇k ∧ G)|∂D ) = N − (∇k y F )|∂D . For the right hand side, we note for h := ∇0k ∧(N + g) ∈ N + L2 that ∗h = −(νw0 )h = −ν(w0 h) = −ν(w0 x h), where w0 := ν y w. We used here Corollary 3.1.10 and hwi2 = −1. We get ∗(∇0k ∧ (N + g)) = −ν((w0 x N + g) x ∇0k ) = ν((ν y f ) x ∇0k ) = ν∇0k y (f x ν).

Chapter 9. Dirac Wave Equations

330

Note that the first step uses a nonsmooth extension of Exercise 11.2.7. Reversing these two equations, multiplying them from the left by ν, and equating them yields ν y (∇k y F )|∂D = ν(∇k0 y (f x ν))ν = ∇k0 y (fb x ν) = −∇0k y (ν y f ). In the second step we used that νh = b hν whenever h ∈ N + L2 , and in the last step we applied the commutation relation from Proposition 2.6.3. This proves the lemma for Ek+ L2 , since Γk f = −(∇k0 y F )|∂D . The proof for Ek− L2 is similar. We next show the converses of Lemmas 9.6.4 and 9.6.5. Lemma 9.6.6 (N ± L2 to Ek± L2 ). If f ∈ N + L2 and f ∈ D(d0 ), then f ∈ D(Γk ) with Γk f = d0 f + ike0 ∧ f. Similarly, if f ∈ N − L2 and ν y f ∈ D(δ 0 ), then f ∈ D(Γk ) with Γk f = ν ike0 y)(ν y f ). Proof. Let f ∈ N + L2 and define Cauchy extensions Z 0 F ± (x) = ± Ψ− k (y − x)ν(y)f (y)dy,

∧

(δ 0 +

x ∈ D± .

Differentiating under the integral sign, with notation as in the proof of Lemma 9.6.5, we have Z ∇k ∧ F (x) = ∓ ˙ Ψk− (y − x)(∇ k ∧ ν(y) ∧ f (y))dy ∂D Z =± Ψ− k (y˙ − x)(∇−k ∧ ν(y) ∧ f (y))dy, ∂D

where we have used the algebraic anticommutation relation − − ∇k ∧ (Ψ− k h) = (∇k , Ψk ih − Ψk (∇k ∧ h) − and the first term vanishes, since Ψ+ k (·−y) = −Ψk (y −·) is a fundamental solution to D + ike0 . Aiming to apply a nonsmooth extension of Exercise 11.2.7, we form the inner product with a fixed multivector w ∈ 4Wc , and obtain Z hw, ∇k ∧ F (x)i = ± hν(y) y (∇k y (Ψk+ (y˙ − x)w)), f (y)idy. ∂D

We choose to use the complex bilinear pairing on 4Wc , but this is not important. By Lemma 9.6.5, we have ν(y)y(∇k y(Ψk+ (y˙ −x)w)) = −(δ 0 +ike0 y)(ν(y)y(Ψk+ (y − x)w)). Note that F in the proof of Lemma 9.6.5 need not solve a Dirac equation for such a trace result to be true. Duality yields Z 0 hw, Ψ− hw, ∇k ∧ F (x)i = ± k (y − x)ν(y)(d f + ike0 ∧ f )idy. ∂D

9.6. Boundary Hodge Decompositions

331

Since w is arbitrary, this proves the lemma for N + L2 . The proof for f ∈ N − L2 is similar. We calculate Z ∇k ∧ F (x) = −∇k y F (x) = ∓ Ψk− (y˙ − x)(∇−k y (ν(y) y f (y)))dy. ∂D

Pairing with w gives Z hw, ∇k ∧ F (x)i = ∓ h∇k ∧ (Ψ+ k (y˙ − x)w), ν(y) y f (y)idy ∂D Z h(d0 + ike0 ∧)N + (Ψk+ (y˙ − x)w), ν(y) y f (y)idy =∓ Z∂D hw, Ψk− (y − x)ν(y)(ν(y) ∧ ((δ 0 + ike0 y)(ν(y) y f (y))))idy. =± ∂D

The second equality uses that ν yf ∈ N + L2 and Lemma 9.6.4. Since w is arbitrary, this proves the lemma for N − L2 . Summarizing the above results, we obtain the following concrete expression for Γk . Given Lemmas 9.6.4, 9.6.5, and 9.6.6, the proof is straightforward. Proposition 9.6.7. The operator Γk is a nilpotent operator in L2 (∂D) in the sense of Definition 10.1.1. Its domain equals D(Γk ) = {f ∈ L2 (∂D) ; N + f ∈ D(d0 ) and ν y f ∈ D(δ 0 )} and Γk f = (d0 + ike0 ∧)N + f + ν ∧ (δ 0 + ike0 y)(ν y f ),

f ∈ D(Γk ).

The operator Γk commutes with Ek and with N . Having uncovered this nilpotent operator Γk , we now investigate the Hodge splitting of L2 (∂D) that it induces. We need a Hermitian inner product on L2 (∂D), and we choose Z (f (x), g(x)iV dx.

(f, gi = ∂D

Proposition 9.6.8 (Boundary Hodge decomposition). When k 6= 0, the nilpotent operator Γk induces an exact Hodge splitting L2 (∂D) = R(Γk ) ⊕ R(Γ∗k ), where the ranges R(Γk ) = N(Γk ) and R(Γk∗ ) = N(Γk∗ ) are closed. When k = 0, the ranges are still closed, but the finite-dimensional cohomology space H(Γk ) = N(Γk ) ∩ N(Γ∗k ) will be nontrivial.

Chapter 9. Dirac Wave Equations

332

Proof. Proposition 10.1.2 shows that Γ = Γk induces an orthogonal splitting L2 (∂D) = R(Γk ) ⊕ H(Γk ) ⊕ R(Γ∗k ). When D is smooth and k = 0, it follows from Propositions 12.1.3 and 10.1.6 that the ranges are closed and that the cohomology space is finite-dimensional for Γ = Γk . Adapting the methods from Theorem 10.3.1 to the manifold setting, this result can be extended to the case that D is merely a Lipschitz domain. However, on nonsmooth boundaries ∂D we do not have D(d0 ) ∩ D(δ 0 ) = H 1 (∂D), but still D(d0 ) ∩ D(δ 0 ) is compactly embedded in L2 (∂D). Assume next that k 6= 0 and define the nilpotent operator µf = ike0 ∧ N + f + ν ∧ (ike0 y (ν y f )) so that Γk = Γ + µ. We compute µ∗ f = ik c e0 y N + f + ν ∧ (ik c e0 ∧ (ν y f )). As in Example 10.1.7, we note that N(µ) ∩ N(µ∗ ) = {0}. Consider the abstract Dirac operators Γk + Γ∗k = (Γ + Γ∗ ) + (µ + µ∗ ) : D(Γ) ∩ D(Γ∗ ) → L2 . Since Γ + Γ∗ : D(Γ) ∩ D(Γ∗ ) → L2 is a Fredholm operator and µ + µ∗ : D(Γ) ∩ D(Γ∗ ) → L2 is a compact operator, it follows from Proposition 10.1.6 that Γk +Γ∗k : D(Γ) ∩ D(Γ∗ ) → L2 is a Fredholm operator. Thus the ranges are closed. To prove that in fact the cohomology space N(Γk ) ∩ N(Γ∗k ) in fact is trivial, we note that Γµ∗ + µ∗ Γ = 0. Thus, if Γf + µf = 0 = Γ∗ f + µ∗ f , then 0 = (f, (Γµ∗ + µ∗ Γ)f iV = (Γ∗ f, µ∗ f iV + (µf, Γf iV = −kµ∗ f k2 − kµf k2 . This shows that f = 0 and completes the proof.

Exercise 9.6.9. Roughly speaking, Γ∗k acts as interior derivative on N + L2 and as exterior derivative on N − L2 . Write down the details of this, and show that Γ∗k commutes with N , but not with Ek in general.

9.7

Maxwell Scattering

In this section, we demonstrate how classical Helmholtz and Maxwell boundary value and transmission problems can be solved using the operators Ek , N , and Γk . Recall that Ek is the reflection operator for the Hardy space splitting from Theorem 9.3.9, that N is the reflection operator for the splitting into normal and

333

9.7. Maxwell Scattering

tangential fields from Example 9.4.8, and that Γk is the nilpotent operator for the boundary Hodge decomposition from Proposition 9.6.8. The basic operator algebra is that Ek2 = N 2 = I, Γ2k

= 0,

6 N Ek , Ek N =

Γk Ek = Ek Γk ,

Γk N = N Γk .

The Ek N BVPs are essentially well posed, so by Proposition 9.4.5, roughly speaking Ek and N are closer to anticommuting than to commuting.

Figure 9.3: A rough sketch of the splittings involved in a Dirac BVP. The splitting into Ek± H encodes the Dirac equation. The splitting into N ± H encodes the boundary conditions. The circle indicates the boundary Hodge splitting, with the interior of the circle illustrating N(Γk ) where the Maxwell BVP takes place. We note that the operator S from Example 9.4.9 does not commute with Γk , but we will not need this, since we use only S as a computational tool for solving an Ek N BVP.

Chapter 9. Dirac Wave Equations

334

We consider as an example the exterior Dirac BVP (9.23) with prescribed tangential part. The other three Ek N BVPs can be analyzed similarly. Using the operator Γk , we have three relevant L2 (∂D; 4Wc )-based function spaces, namely H = L2 ,

H = D(Γ),

and H = N(Γk ).

Note that D(Γ) = D(Γk ) is a dense subspace of L2 , which does not depend on k, although the equivalent norms kf kD(Γk ) = (kf k2L2 + kΓk f k2L2 )1/2 do depend on k. Further note that N(Γk ) is a closed subspace of L2 , as well as of D(Γ), which is roughly speaking half of the latter spaces by Hodge decomposition. Since Ek and N commute with Γk , they act as bounded linear operators in each of the three function spaces H, and in each case we see that Ek2 = N 2 = I. Therefore we can consider the BVP (9.23), expressed as the restricted projection N + : Ek− H → N + H, in each of the three function spaces H. Our aim in this section is to solve BVPs in H = N(Γk ). This, however, is a function space defined by a differential constraint, which we may want to avoid numerically. For this reason, we prefer to enlarge the function space to either L2 or to the function space D(Γ) in which roughly speaking half of the functions have Sobolev regularity H 1 , since Γk is nilpotent, and to solve the integral equation in such a space. Proposition 9.7.1 (Constrained Dirac BVPs). Consider the exterior Dirac BVP N + : Ek− L2 → N + L2 with prescribed tangential part at ∂D, and assume that Im k ≥ 0 and k 6= 0, so we have L2 well-posedness of this BVP by Theorem 9.5.2. Then the restricted map N + : Ek− H → N + H is also invertible for each of the function spaces H = D(Γ) and H = N(Γk ). For the solution f ∈ Ek− L2 to the BVP with datum g = N + f ∈ L2 , the following holds. If g ∈ D(Γ), then f ∈ D(Γ). If Γk g = 0, then Γk f = 0. If Γk g = 0 and g is a j-vector field, then f is a j-vector field. Note that if g ∈ N + L2 is a j-vector field, then in general the solution f ∈ to the BVP will not be a homogeneous j-vector field. The constraint Γk g = 0 is crucial. Ek− L2

Proof. (i) Lower bounds for N + : Ek− D(Γ) → N + D(Γ)

(9.26)

hold, since kf kD(Γ) ≈ kf kL2 + kΓk f kL2 . kN + f kL2 + kN + Γk f kL2 = kN + f kL2 + kΓk (N + f )kL2 ≈ kN + f kD(Γ) .

9.7. Maxwell Scattering

335

To show surjectivity, we can proceed as follows. First apply Lemma 9.4.12 with A = Ek , B = N , H = D(Γ) and perturb k into Im k > 0. This shows that it suffices to show surjectivity for Im k > 0. Then we use that N and Ek commute with Γk , and similarly to the above, we derive lower bounds for λI + Ek N : D(Γ) → D(Γ), when |λ1 | > L|λ2 |, from Theorem 9.5.1. Therefore the method of continuity shows that I ± Ek N are Fredholm operators of index zero on D(Γ). The argument in Proposition 9.4.10 shows that all four Ek N BVPs are injective when Im k > 0, and so it follows from (9.19) that (9.26) is surjective. If f ∈ Ek− L2 solves the BVP with datum g ∈ D(Γ), then let f˜ ∈ Ek− D(Γ) be the solution to the well-posed BVP described by (9.26). By uniqueness of the solutions to the L2 BVP, we conclude that f = f˜ ∈ D(Γ). (ii) Next consider N + : Ek− N(Γk ) → N + N(Γk ). This map is clearly bounded and injective with a lower bound. To show surjectivity, let g ∈ N + N(Γ) ⊂ N + D(Γ). By (i) there exists f ∈ Ek− D(Γ) such that N + f = g. Since N + (Γk f ) = Γk (N + f ) = Γk g = 0, it follows from L2 well-posedness that f ∈ Ek− N(Γk ). If furthermore g ∈ N(Γk ) is a j-vector field, then the solution f satisfies Γk f = 0, and we conclude from Lemma 9.6.1 that each homogeneous component function fm belongs to Ek+ L2 . Since N + fm = gm = 0 if m 6= j, it follows in this case that fm = 0 by uniqueness of solutions to the BVP. Therefore f = fj is a j-vector field. Example 9.7.2 (Helmholtz BVPs). In Example 9.3.7 we saw how the Helmholtz equation for a scalar acoustic wave u is equivalent to the vector field F = ∇u + ike0 u solving the Dirac equation DF + ike0 F = 0. (i) The Neumann BVP for u amounts to specifying the normal part N − f = (∂ν u)ν of f = F |∂D . In this case, by Proposition 9.6.7 the condition Γk (N − f ) = 0 is automatic for a vector field f , since ∧−1 W = {0}. Therefore, solving the Dirac BVP for F with this prescribed datum on ∂D will produce a vector field F according to Proposition 9.7.1. From Proposition 9.6.8 it follows that F ∈ R(Γk ), which means that there exists a scalar function u such that F = ∇u + ike0 u. In particular, u solves ∆u + k 2 u = 0 with prescribed Neumann datum. (ii) The Dirichlet BVP for u amounts to specifying the tangential part N + f = ∇0 u + ike0 u of f = F |∂D . For a given tangential vector field g = g1 + e0 g0 , where g1 is a space-like vector field and g0 is a scalar function, we note that Γk g = ∇0 ∧ g1 + e0 ∧ (−∇0 g0 + ikg1 ), so g ∈ N(Γk ) amounts to ikg1 = ∇0 g0 .

336

Chapter 9. Dirac Wave Equations

Therefore, solving the Dirac BVP for F with such a tangential vector field g ∈ N(Γk ) on ∂D as datum will produce a vector field of the form F = ∇u + ike0 u by Proposition 9.7.1, where u solves the Helmholtz Dirichlet problem. Example 9.7.3 (Maxwell BVPs). In Example 9.3.8 we saw how Maxwell’s equations 1/2 for an electromagnetic wave F are equivalent to the bivector field F = 0 e0 ∧ −1/2 E + µ0 B solving the Dirac equation DF + ike0 F = 0. We now assume that the interior domain D+ ⊂ R3 is a perfect electric conductor, so that E = B = 0 in D+ . If Maxwell’s equations are to hold in the distributional sense in all R3 , by the vanishing right-hand sides in the Faraday and magnetic Gauss laws, we need N + f = 0 for f = F |∂D . If the electromagnetic wave in D− is the superposition of an incoming wave f0 and a reflected wave f1 ∈ Ek− L2 , then f1 needs to solve the BVP where N + f1 is specified to cancel the datum N + f0 . Note that for the classical vector fields E and ∗B, the tangential part N + f corresponds to the tangential part of E and the normal part of ∗B. For a given tangential bivector field g = e0 ∧ g1 + g2 , where g1 is a space-like tangential vector field and g2 is a space-like tangential bivector field, we note that Γk g = e0 ∧ (−∇0 ∧ g1 + ikg2 ) + ∇0 ∧ g2 , so g ∈ N(Γk ) amounts to ikg2 = ∇0 ∧ g1 . In terms of the electric and magnetic fields, the tangential part of B is given by the tangential curl of the tangential part of E. From Proposition 9.7.1 it follows that if we solve the Dirac BVP (9.23) with such a tangential bivector field g ∈ N(Γk ) on ∂D as datum, then the solution f will indeed be a bivector field representing an electromagnetic field. Example 9.7.4 (Maxwell transmission problems). When an electromagnetic wave propagates in a material and not in vacuum, we account for the material’s response to the field by replacing 0 and µ0 by permittivity and permeability constants and µ depending on the material properties. These may in general be variable as well as matrices, but we limit ourselves to homogeneous and isotropic materials for which and µ are constant complex numbers. Similar to (9.3), we define the electromagnetic field F := 1/2 e0 ∧ E + µ−1/2 B. √ Maxwell’s equations in such a material read DF + ike0 F = 0, with k = ω µ. Consider the following transmission problem. We assume that the exterior domain D− consists of a material with electromagnetic properties described by 1 , √ µ1 , giving a wave number k1 := ω 1 µ1 , and that the interior domain D+ consists of a material with electromagnetic properties described by 2 , µ2 , giving a wave √ number k2 := ω 2 µ2 . We obtain a transmission problem of the form (9.25) for a pair of electromagnetic fields F ± : D± → 42 Wc .

337

9.7. Maxwell Scattering 1.5

0.5 0.4

1

0.3 0.5

0.1

1

0.2

0.5

0.1

0

0

0

-0.1

0

-0.1 -0.5

-0.2 -0.3

-0.5

-0.2 -0.3

-1

-0.4 -0.5 0

0.2

0.4

0.6

0.8

1

-1.5

0.4

1

0.3

-1

-0.4 -0.5 0

(b) 1.5

0.5

0.2

0.4

0.6

0.8

1

-1.5

1.5

0.5 0.4

1

0.3

0.2

0.5

0.1

0.2

0.5

0.1

0

0

0

-0.1

0

-0.1 -0.5

-0.2 -0.3

-0.5

-0.2 -0.3

-1

-0.4 -0.5

(c)

0.4 0.3

0.2

(a)

1.5

0.5

0

0.2

0.4

0.6

0.8

1

-1.5

-1

-0.4 -0.5

(d)

0

0.2

0.4

0.6

0.8

1

-1.5

Figure 9.4: TM magnetic waves U = B12 . ∂Ω parametrized√by sin(πs) exp(i(s−1/2) π/2), 0 ≤ s ≤ 1. (a) Incoming wave U0 = exp(18i(x + y)/ 2) from south-west. (b) Wave reflected by a perfect electric conductor, computed with the spin integral equation in Example 9.5.5. (c) Waves reflected into Ω− and transmitted into a dielectric object Ω+ , computed with a tweaked version of the Dirac integral equation in Example 9.5.6. Wave numbers k1 = 18 and k2 = 27 as in Example 9.7.4. (d) As in (c), but Ω+ is now a conducting √ object described by the Drude model and an imaginary wave number k2 = i18 1.1838. Here the wave decays exponentially into Ω+ and surface plasmon waves, excited by the corner singularity, appear near ∂Ω. The jump condition M f + = f − + g is found by returning to the original formulation of Maxwell’s equations for E and B. For these to hold in the distributional sense across ∂D, Faraday’s law and the magnetic Gauss laws dictate that ν ∧ E and ν ∧ B do not jump across ∂D. Furthermore, assuming that we do not have any electric charges and current except for those induced in the material described by and µ, the Amp`ere and Gauss laws require that ν y(µ−1 B) and ν y(E) not jump across ∂D. If we translate this to spacetime multivector algebra, this specifies the multiplier q q q q M = µµ21 N + T + + 12 N + T − + µµ12 N − T + + 21 N − T − ,

Chapter 9. Dirac Wave Equations

338 0.5

3

0.5

0.4 2

0.3

-13

0.3

0.2

1

0.1

-13.5

0.2 0.1

0

0

-0.1 -1

-0.2

-14

0 -0.1

-14.5

-0.2

-0.3

-2

-0.4

-0.3

-15

-0.4 -3

-0.5 0

(d)

0.4

0.2

0.4

0.6

0.8

1

-15.5

-0.5 0

(b)

0.2

0.4

0.6

0.8

1

0.5

0.5 0.4

-13

-13

-13.5

-13.5

0.3 0.2 0.1 -14

0 -0.1

-14

0

-14.5

-14.5

-15

-15

-0.2 -0.3 -0.4 -15.5

-0.5

(c)

0

0.2

0.4

0.6

0.8

1

-15.5

-0.5

(d)

0

0.2

0.4

0.6

0.8

1

Figure 9.5: Upper left (d) is same as Figure 9.4(d), but scaled so the peaks of the plasmon wave are visible. (b), (c) and (d) show log10 of the estimated absolute error for the three scattering computations. (d) indicates the numerical challenge in computing surface plasmon waves. Here the parameters hit the essential spectrum, where the integral equation fails to be Fredholm. using the normal reflection operator N and the time reflection operator T . Note how the two commuting reflection operators N and T split the electromagnetic field into these four parts. With this formulation, and with the datum g being the boundary trace g = F0 |∂D of an incoming electromagnetic wave F0 in D− , we can use the Dirac integral equation proposed in Example 9.5.6 to compute the transmitted wave F + in D+ and the reflected wave F − in D− . We end this chapter with some examples of how the integral equations from Examples 9.5.5 and 9.5.6 perform numerically when applied to scattering problems for electromagnetic fields as in Examples 9.7.3 and 9.7.4. Results are shown in Figures 9.4 and 9.5. For simplicity, we consider a two-dimensional scattering problem in which the object represented by the domain D+ ⊂ R3 is a cylinder D+ = Ω+ × R

9.8. Comments and References

339

along the z-axis over the base Ω+ ⊂ R2 = [e12 ] in the xy-plane, and the field is transversal magnetic. This means that we assume that √ F = F (x, y) = e0 ∧ (E1 (x, y)e1 + E2 (x, y)e2 ) + √1µ B12 (x, y)e12 . In classical vector calculus notation this means that E is parallel to R2 and the 2 vector field ∗B is orthogonal √ to R , explaining the terminology. Maxwell’s equations, after dividing F by , read (∇ + ike0 )(e0 E + cB) = (c∇B − ikE) + e0 (−∇E + iωB) = 0, where ∇ = e1 ∂1 + e2 ∂2 is the nabla symbol for R2 . From the space- and time-like parts of this equation, we get ∆B = ∇(∇B) = (ik/c)∇E = (ik/c)iωB = −k 2 B, that is, U := B12 solves the Helmholtz equation, and E = (c/ik)∇B = (c/ik)(∇U )e12 . This means that Maxwell’s equations for transversal magnetic fields F are equivalent to the Helmholtz equation for U = B12 and that E is obtained from the gradient ∇U by rotation and scaling. In particular, it follows that for transversal magnetic fields F , the tangential boundary datum N + f corresponds to the Neumann data ∂ν U for U .

9.8

Comments and References

9.2 Building on the work of Michael Faraday and Andr´e-Marie Amp`ere, James Clerk Maxwell (1831–1879) collected and completed the system of equations governing electromagnetic theory in the early 1860s. His Treatise on Electricity and Magnetism was published in 1873. The equations that he obtained showed that electric and magnetic fields propagate at the speed of light, and they were relativistically correct decades before Einstein formulated relativity theory. The fundamental equation of quantum mechanics, the Schr¨odinger equation from Example 6.3.6, was first discovered in 1925 and describes physics at small scales. The famous Stern–Gerlach experiment from 1922 showed that the intrinsic angular momentum of particles is quantized. The Pauli equation from 1927 is a modification of the Schr¨odinger equation that takes this spin phenomenon into account, but neither of these is the correct equation at high speeds, that is, they are not relativistically correct. The Klein–Gordon equation from 1926 is a relativistically correct version of the Schr¨odinger equation, but it does not incorporate spin. Paul Dirac finally succeeded in

340

Chapter 9. Dirac Wave Equations 1928 in finding the equation that is correct from the point of view of both quantum mechanics and relativity theory, as well as correctly describing spin1/2 particles, which include all the elementary particles constituting ordinary matter. The classical derivation of the Dirac equation is to seek matrices γ0 , γ1 , γ2 , γ3 by which one can factorize the Klein–Gordon equation into a first-order wave equation. This amounts to using a matrix representation of the spacetime Clifford algebra, something that the pioneers of quantum mechanics were unaware of. Starting from the 1960s there has been a renewed interest in Clifford’s geometric algebra, where in particular, David Hestenes [55], Hestenes and Sobczyk [57], and Hestenes [56] have advocated geometric algebra as the preferred mathematical framework for physics. In particular, [55] is a reference for using Clifford algebra to study Maxwell’s and Dirac’s equations. The formulations (9.4) and (9.5) of Maxwell’s equations as wave 4-Dirac equations go back to M. Riesz. A further reference for the use of multivectors in electromagnetic theory is Jancewicz [60]. A standard mathematics reference for the analysis of Dirac’s equation is Thaller [93]. Further references on Dirac operators and spinors in physics include Benn and Tucker [19] and Hitchin [58].

9.3-9.6 The material covered in these sections, which aim to solve Maxwell BVPs using multivector calculus, builds on the author’s PhD thesis and publications [8, 9, 7, 14, 10]. The first basic idea for solving boundary value problems for Maxwell’s equations is to embed it into a Dirac equation as in Example 9.3.8. This was first used by McIntosh and M. Mitrea in [67] in connection with BVPs on Lipschitz domains. The second basic idea is to formulate Dirac boundary value problems in terms of Hardy projections Ek± and projections N ± encoding boundary conditions, and to show that these subspaces are transversal. This was first worked out by Axelsson, Grognard, Hogan, and McIntosh [11]. The third main idea is to extract a Maxwell solution from the Dirac solution as in Proposition 9.7.1, using the Hodge decomposition on the boundary defined by the operator Γk from Section 9.6. This was worked out in detail in [9]. We have chosen to use the spacetime formulation, but as in Propositions 9.1.5 and 9.1.6, we can equally well use a 4V formulation in which the Dirac equation reads DF = ikF for F : D → 4Vc . The main reason for our choice is that the operator Γk in Section 9.6 is difficult, although not impossible, to handle using the latter formalism. To minimize the algebra, the 4Vc formulation was used in [84, 80], where the spin integral equation from Example 9.5.5 was first introduced.

9.8. Comments and References

341

A main philosophy in [9] and associated publications is to handle the boundary value problems by first-order operators. It is clear what this means for the differential operators: in (9.10) the second-order Helmholtz operator is factored by the first-order Dirac operator. But we also have corresponding factorizations of the boundary integral operators. In the abstract formulation with Proposition 9.4.5, the second-order cosine operator is factored by the first-order rotation operator in (9.20)–(9.21). We think of the rotation operators being as of first order, since they essentially are direct sums of two restricted projections as in (9.18)–(9.19). Similarly, the cosine operator can be seen to be essentially the direct sum of compositions of two restricted projections, hence of second order. A reference for Bessel functions and Exercise 9.3.3 is Watson [95]. Standard references for the classical double and single layer potential integral equations are Colton and Kress [29, 30] and Kress [62]. The method to prove semi-Fredholm estimates of singular integral operators on Lipschitz domains as in Theorem 9.5.1 using Stokes’s theorem and a smooth transversal vector field as in Exercise 6.1.8 goes back to Verchota [94]. The spectral estimates in Theorem 9.5.1 are from [7]. 9.7 Figures 9.4 and 9.5 have been produced by Johan Helsing using the spin and tweaked Dirac integral equations. The state-of-the-art numerical algorithm RCIP, recursively compressed inverse preconditioning that he uses is described in [51], with applications to Helmholtz scattering in [52] and [53]. Since the Dirac equation is more general than the Helmholtz and Maxwell equations that it embeds, the spin and Dirac integral equations cannot quite compete with the most efficient Kleinman–Martin type integral equation [53, eq. 45] in terms of computational economy. In terms of achievable numerical accuracy in the solution, however, the two systems of integral equations perform almost on par with each other. Moreover, the spin and Dirac integral equations apply equally well to Maxwell scattering in three dimensions, where the present understanding of integral formulations for Maxwell’s equations is incomplete.

Chapter 10

Hodge Decompositions Prerequisites: The reader is assumed to have read Sections 7.5 and 7.6, which this chapter develops further. A good understanding of unbounded Hilbert space operators and the material in Section 6.4 is desirable. Some exposure to distribution theory and algebraic topology helps, but is not necessary. Road map: We saw in Section 7.6 that every multivector field F on a domain D can be decomposed into three canonical parts F = ∇ ∧ U + H + ∇ y V, where ∇ ∧ H = 0 = ∇ y H, and H and the potential V are tangential on ∂D. This is the Hodge decomposition of the multivector field F , which amounts to a splitting of the space of all multivector fields F into two subspaces R(d) and R(δ ) of exact and coexact fields respectively, and a small subspace Ck (D) of closed and coclosed fields, all with appropriate boundary conditions. Alternatively, we can instead demand that H and the potential U be normal on ∂D. At least four types of questions arise. (i) Are the subspaces R(d) and R(δ) transversal, that is do they intersect only at 0 and at a positive angle? This would mean that these subspaces give a splitting of the function space H that we consider, modulo Ck (D). In the case of H = L2 (D) which we only consider here, these subspaces are in fact orthogonal, but more generally this problem amounts to estimating singular integral operators realizing the Hodge projections onto these subspaces. We touch on this problem in Proposition 10.1.5 and Example 10.1.8. (ii) Are the ranges R(d) and R(δ) closed subspaces? This is a main problem that we address in this chapter, and we show that this is indeed the case for © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_10

343

344

Chapter 10. Hodge Decompositions

bounded domains D. See Section 10.3 and Example 10.1.8. We saw in Section 7.6 that such closedness yields well-posedness results for boundary value problems. (iii) What properties, in particular regularity, of the potentials U and V do we have? Note that the parts ∇ ∧ U and ∇ y V are uniquely determined by F , but not so for the potentials U and V . We show in Section 10.4 that the most obvious choice, the Hodge potentials of minimal L2 norm, are not always the best choices. Even more surprising is the fact that there exist Bogovski˘ı potentials V , for which we have full Dirichlet boundary conditions V |∂D = 0. (iv) Is the cohomology space Ck (D) finite-dimensional? More exactly, how do we go about calculating the dimension of this subspace for a given domain D? As compared to the first three questions, which belong to analysis, this fourth question belongs to algebraic topology and is addressed in Section 10.6. In the analysis of Hodge decompositions on domains, the regularity and curvature of the boundary play an important role through Weitzenb¨ock formulas. Hodge decompositions can also be considered on manifolds, and in this case also the curvature of the manifold in the interior of the domain enters the picture. This will be a central idea in Chapters 11 and 12. In the present chapter we avoid the technicalities of vector bundles and limit the discussion to domains in affine spaces. Highlights: • Compactness and Hodge decomposition: 10.1.6 • Natural boundary conditions for d and δ: 10.2.3 • Weitzenb¨ock boundary curvature: 10.3.6 • Bogovski˘ı and Poincar´e potentials: 10.4.3 ˇ • Cech computation of Betti numbers: 10.6.5

10.1 Nilpotent operators In terms of operators, a splitting of a function space corresponds to a projection P , along with its complementary projection I − P . Somewhat similarly, we show in this section how Hilbert space operators Γ with the property that Γ2 = 0 induce splittings of the function space in a natural way, generalizing Hodge decompositions. Usually the condition Γk = 0 for some k ∈ Z+ defines nilpotence, but we shall always assume index k = 2. Definition 10.1.1 (Nilpotent). A linear, possibly unbounded, operator Γ : H → H in a Hilbert space H is said to be nilpotent (with index 2) if it is densely defined, closed, and if R(Γ) ⊂ N(Γ). In particular, Γ2 f = 0 for all f ∈ D(Γ). We say that a nilpotent operator Γ is exact if R(Γ) = N(Γ).

345

10.1. Nilpotent operators

Recall that the null space N(Γ) is always closed if Γ is closed but that in general, the range R(Γ) is not a closed subspace. If Γ is nilpotent, then we have inclusions R(Γ) ⊂ R(Γ) ⊂ N(Γ) ⊂ D(Γ) ⊂ H. Let H0 denote any closed subspace complementary to N(Γ), for example H0 = N(Γ)⊥ , so that H = N(Γ) ⊕ H0 . Then the restricted map Γ : H0 → R(Γ) ⊂ N(Γ) is injective, which roughly speaking means that N(Γ) is at least half of H. For this reason it is natural to combine a nilpotent operator Γ1 with a “complementary” nilpotent operator Γ2 . Ideally one would like to have a splitting of the Hilbert space H = R(Γ1 ) ⊕ R(Γ2 ), where R(Γ1 ) = N(Γ1 ) and R(Γ2 ) = N(Γ2 ). Since N(Γ∗1 ) = R(Γ1 )⊥ , the natural choice in a Hilbert space is Γ2 = Γ∗1 . Proposition 10.1.2 (Abstract Hodge decomposition). Let Γ be a nilpotent operator in a Hilbert space H. Then so is Γ∗ , and there is an orthogonal splitting into closed subspaces H = R(Γ) ⊕ C(Γ) ⊕ R(Γ∗ ), (10.1) where C(Γ) := N(Γ) ∩ N(Γ∗ ), N(Γ) = R(Γ) ⊕ C(Γ),

and

∗

N(Γ ) = C(Γ) ⊕ R(Γ∗ ). Note that C(Γ) = {0} if and only if Γ is exact. Proof. If T is a densely defined and closed operator in H, then R(T )⊥ = N(T ∗ ) and therefore R(T ) = N(T ∗ )⊥ . This proves that R(Γ∗ ) = N(Γ)⊥ ⊂ R(Γ)⊥ = N(Γ∗ ), showing that Γ∗ is nilpotent and that we have orthogonal splittings H = N(Γ) ⊕ R(Γ∗ ) = R(Γ) ⊕ N(Γ∗ ). But R(Γ) ⊂ N(Γ), since Γ is nilpotent, so using the second splitting in the first, we get ⊥ N(Γ) = R(Γ) ⊕ N(Γ) ∩ R(Γ) = R(Γ) ⊕ N(Γ) ∩ N(Γ∗ ) , which proves the stated splitting.

The mapping properties of Γ and Γ∗ are as follows. In the Hodge decomposition (10.1), the operator Γ is zero on R(Γ) ⊕ N(Γ) ∩ N(Γ∗ ) = N(Γ), and Γ∗ is zero on N(Γ∗ ) ∩ N(Γ) ⊕ R(Γ∗ ) = N(Γ∗ ). On the other hand, we see that the

Chapter 10. Hodge Decompositions

346

restrictions Γ : R(Γ∗ ) → R(Γ) and Γ∗ : R(Γ) → R(Γ∗ ) are injective and have dense ranges: H

H

=

=

⊕

R(Γ)

R(Γ)

s

N(Γ) ∩ N(Γ∗ )

Γ∗

Γ

⊕

∗

N(Γ) ∩ N(Γ )

R(Γ∗ )

⊕

⊕

+

(10.2)

R(Γ∗ )

We have been using the formally skew-adjoint Dirac operator D = d + δ in Chapters 8 and 9. Using instead the anti-Euclidean Clifford product leads to a formally self-adjoint Dirac operator d − δ. For the following results we can use either the abstract Dirac operator Γ − Γ∗ or its self-adjoint analogue Γ + Γ∗ . To be able to use resolvents without complexifying the space, we choose to work with Γ − Γ∗ . Note from the mapping properties of Γ and Γ∗ that such operators swap the subspaces R(Γ) and R(Γ∗ ). Proposition 10.1.3 (Abstract Hodge–Dirac operators). Let Γ be a nilpotent operator in a Hilbert space H. Consider the operator Π := Γ − Γ∗ with domain D(Π) := D(Γ) ∩ D(Γ∗ ). Then Π is skew-adjoint, that is, Π∗ = −Π in the sense of unbounded operators, with N(Π) = C(Γ) and R(Π) = R(Γ) + R(Γ∗ ). We refer to operators Π = Γ − Γ∗ , derived from a nilpotent operator Γ, as an abstract Hodge–Dirac operator. Note that in Euclidean spaces, the 4-Dirac operator D from Definition 9.1.1 is an example of a Hodge–Dirac operator, whereas / from Definition 9.1.3 as a Hodge–Dirac operator to have the 4-Dirac / operator D requires a complex structure on our Euclidean space, as discussed at the end of Section 9.1. Proof. We use the Hodge decomposition from Proposition 10.1.2. If Γu + Γ∗ u = 0, then Γu = 0 = Γ∗ u by orthogonality, from which N(Π) = C(Γ) follows. If f = Γu1 + Γ∗ u2 , then f = Π(PΓ∗ u1 + PΓ u2 ), from which R(Π) = R(Γ) + R(Γ∗ ) follows. Note that u1 − PΓ∗ u1 ∈ N(Γ) ⊂ D(Γ) and similarly for u2 . It is clear that −Π is the formal adjoint of Π. It remains to prove that if hf, Πgi + hf 0 , gi = 0 for all g ∈ D(Π), then f ∈ D(Π) and f 0 = Πf . Writing f = f1 + f2 + f3 in the Hodge splitting, and similarly for f 0 , we have hf1 , Γgi + hf30 , gi = 0, 0 + hf20 , gi = 0, hf3 , −Γ∗ gi + hf10 , gi = 0, by choosing g ∈ R(Γ∗ ) ∩ D(Γ), g ∈ C(Γ) and g ∈ R(Γ) ∩ D(Γ∗ ) respectively. Since Γ and Γ∗ are adjoint in the sense of unbounded operators, we conclude that f1 ∈ D(Γ∗ ), f30 = −Γ∗ f1 , f20 = 0, f3 ∈ D(Γ) and f10 = Γf3 . This shows that f ∈ D(Π) and f 0 = Πf .

347

10.1. Nilpotent operators

Definition 10.1.4 (Hodge projections). Let Γ be a nilpotent operator in a Hilbert space H. The associated Hodge projections are the orthogonal projections PΓ and PΓ∗ onto the subspaces R(Γ) and R(Γ∗ ) respectively. The orthogonal projection PC(Γ) onto the Γ-cohomology space C(Γ) is PC(Γ) = I − PΓ − PΓ∗ . Proposition 10.1.5 (Formulas for Hodge projections). Let Γ be a nilpotent operator in a Hilbert space H. If Γ is exact, then PΓ f = ΓΠ−1 f = −Π−1 Γ∗ f = −ΓΠ−2 Γ∗ f, PΓ∗ f = −Γ∗ Π−1 f = Π−1 Γf = −Γ∗ Π−2 Γf, for f ∈ D(Π) ∩ R(Π). If Γ is not exact, let ∈ R\{0}. Then we have PC(Γ) f = lim→0 (I +Π)−1 f , and the Hodge projections are PΓ f = lim Γ(I + Π)−1 f, →0

PΓ∗ f = − lim Γ∗ (I + Π)−1 f, →0

with convergence in H, for f ∈ H. We also have PΓ f = − lim(I + Π)−1 Γ∗ f for f ∈ D(Γ∗ ) and PΓ∗ f = lim(I + Π)−1 Γf for f ∈ D(Γ). Proof. The formulas for exact operators Γ involving Π−1 are immediate from (10.11), and the final second-order formulas follow since PΓ = PΓ2 and PΓ∗ = PΓ2∗ . For nonexact Γ, consider first PC(Γ) f . If f ∈ C(Γ), then (I + Π)−1 f = f . If f = Πu ∈ R(Π), then (I + Π)−1 Πu = u − 2 (I + Π)−1 u → 0 as → 0. We have used the skew-adjointness of Π, which implies that k(I + Π)−1 k ≤ 1. These uniform bounds also allow us to conclude that (I +Π)−1 f → 0 also for all f ∈ R(Π). This proves the formula for PC(Γ) f , from which it immediately follows that Γ(I + Π)−1 f = PΓ Π(I + Π)−1 f → PΓ (f − PC(Γ) f ) = PΓ f, and similarly for PΓ∗ . Alternatively, for f ∈ D(Γ∗ ), we have −(I + Π)−1 Γ∗ f = (I + Π)−1 ΠPΓ f → (I − PC(Γ) )PΓ f = PΓ f, and similarly for PΓ∗ .

The following result describes an important property that a nilpotent operator may have, which we will establish for d and δ on bounded Lipschitz domains. Proposition 10.1.6 (Compact potential maps). For a nilpotent operator Γ in a Hilbert space H, the following are equivalent.

Chapter 10. Hodge Decompositions

348

(i) The subspaces R(Γ) and R(Γ∗ ) are closed and C(Γ) is finite-dimensional, and the inverses of Γ : R(Γ∗ ) → R(Γ) and Γ∗ : R(Γ) → R(Γ∗ ) are compact. (ii) There exist compact operators K0 , K1 : H → H, with R(K1 ) ⊂ D(Γ), such that the homotopy relation ΓK1 f + K1 Γf + K0 f = f holds for all f ∈ D(Γ). (iii) The Hilbert space D(Γ) ∩ D(Γ∗ ), equipped with the norm (kf k2 + kΓf k2 + kΓ∗ f k2 )1/2 , is compactly embedded in H. Carefully note that unlike (i) and (iii), property (ii) does not involve the adjoint Γ∗ . We exploit this in Theorem 10.3.1 below to reduce the problem of existence of potentials, from Lipschitz domains to smooth domains. Also note for (i) that when the ranges are closed, Γ : R(Γ∗ ) → R(Γ) has a compact inverse if and only if there exists a compact operator KΓ : R(Γ) → H such that ΓKΓ = IR(Γ) . Indeed, if we have such KΓ , then PΓ KΓ is a compact operator giving a potential u ∈ R(Γ∗ ). Proof. Assume (i). Define compact operators K0 := PC(Γ) , and ( Γ−1 f ∈ R(Γ∗ ), f ∈ R(Γ), K1 f := 0, f ∈ N(Γ∗ ). It is straightforward to verify that ΓK1 = PΓ and K1 Γ = PΓ∗ , from which (ii) follows. ∞ be a sequence such that fj , Γfj and Γ∗ fj all are Assume (ii). Let (fj )j=1 bounded sequences in H. We have (I −PΓ )fj = (I −PΓ )(ΓK1 fj +K1 (Γfj )+K0 fj ) = (I −PΓ )K1 (Γfj )+(I −PΓ )K0 fj . By duality, we also obtain from the homotopy relation that PΓ fj = PΓ (Γ∗ K1∗ fj + K1∗ (Γ∗ fj ) + K0∗ fj ) = PΓ K1∗ (Γ∗ fj ) + PΓ K0∗ fj . ∞ ∞ This shows that (PΓ∗ fj )∞ j=1 , (PC(Γ) fj )j=1 and (PΓ∗ fj )j=1 have subsequences that converge in H, and (iii) follows. Assume (iii). The operator I + Π is an isometry between the Hilbert spaces D(Γ)∩D(Γ∗ ) and H, since Π is a skew-adjoint operator. Since I is compact between these spaces, perturbation theory shows that

Π : D(Γ) ∩ D(Γ∗ ) → H is a Fredholm operator, and (i) follows.

349

10.1. Nilpotent operators

Nilpotent operators appear naturally from the exterior and interior products, since v ∧ v ∧ w = 0 and v y (v y w) = 0. Example 10.1.7 (Algebraic Hodge decomposition). Fix a unit vector v ∈ V in an n-dimensional Euclidean space and define nilpotent linear maps µ(w) := v ∧ w,

µ∗ (w) := v y w,

w ∈ ∧V.

We apply the abstract theory above to Γ = µ and H, the finite-dimensional Hilbert space ∧V . Lemma 2.2.7 shows that R(µ) = N(µ), so in this case µ is exact and the Hodge decomposition reads ∧V = R(µ) ⊕ R(µ∗ ), where R(µ) are the multivectors normal to and R(µ∗ ) are the multivectors tangential to the hyperplane [v]⊥ , in the sense of Definition 2.8.6. We have (µ−µ∗ )2 = −1, and the Hodge projections are µµ∗ onto normal multivectors, and µ∗ µ onto tangential multivectors. Note that R(µ) and R(µ∗ ), for the full algebra ∧V , both have dimension 2n−1 . However, this is not true in general for the restrictions to ∧k V . For example, the space R(µ) ∩ ∧1 V of vectors normal to [v]⊥ is one-dimensional, whereas the space R(µ∗ ) ∩ ∧1 V of vectors tangential to [v]⊥ has dimension n − 1. The smaller k is, the more tangential k-vectors exist as compared to normal k-vectors. At the ends, all scalars are tangential and all n-vectors are normal. Example 10.1.8 (Rn Hodge decomposition). Consider the exterior and interior derivative operators dF (x) = ∇ ∧ F (x) and δF (x) = ∇ y F (x) in the Hilbert space H = L2 (V ; ∧Vc ) on the whole Euclidean space X = V , where we complexify the exterior algebra in order to use the Fourier transform. These two nilpotent operators are the Fourier multipliers Z n X c ei ∧ dF (ξ) = ∂i F (x)e−ihξ,xi dx = iξ ∧ Fˆ (ξ), c (ξ) = δF

k=1 n X k=1

X

Z ei y

∂i F (x)e−ihξ,xi dx = iξ y Fˆ (ξ) = (−iξ) y F (ξ),

X

defining the interior product as the sesquilinear adjoint of the exterior product. Define the pointwise multiplication operators µξ (Fˆ (ξ)) := ξ ∧ Fˆ (ξ),

µ∗ξ (Fˆ (ξ)) := ξ y Fˆ (ξ).

We view µξ , µ∗ξ : L2 (X; ∧Vc ) → L2 (X; ∧Vc ) as multiplication operators by the Pn radial vector field X 3 ξ 7→ ξ = k=1 ξk ek ∈ V . Thus we have F(dF ) = iµξ (F(F )),

F(δF ) = iµ∗ξ (F(F )).

Chapter 10. Hodge Decompositions

350

In particular, F is closed if and only if Fˆ is a radial multivector field, that is, ξ ∧ Fˆ = 0, and F is coclosed if and only if Fˆ is an angular multivector field, that is, ξ y Fˆ = 0. From Plancherel’s theorem it is clear that Γ = d and Γ∗ = −δ, with domains D(Γ) := {F ∈ L2 ; ξ ∧ Fˆ ∈ L2 } and D(Γ∗ ) := {F ∈ L2 ; ξ y Fˆ ∈ L2 }, are nilpotent operators in H, and that d = −δ ∗ in the sense of unbounded operators. In this case, the Hodge decomposition reads L2 (V ; ∧Vc ) = R(d) ⊕ R(δ). That d is exact is a consequence of µξ being exact for each ξ ∈ V \ {0}. By considering Fˆ near ξ = 0, we see that the ranges are not closed, which is a consequence of the domain X not being bounded. Using the formulas from Proposition 10.1.5, we see that the Hodge projections are the singular integrals Z Pd F (x) = ∇ ∧ Ψ(x − y) y F (y)dy X Z k = F (x) + p.v. ∇ ∧ (Ψ(x˙ − y) y F (y))dy, n X Z Pδ F (x) = ∇ y Ψ(x − y) ∧ F (y)dy X Z n−k = F (x) + p.v. ∇ y (Ψ(x˙ − y) ∧ F (y))dy, n X for k-vector fields F ∈ L2 (X; ∧k Vc ). We have used the distributional derivative ∂i Ψ(x) = ei δ(x)/n + p.v.∂i Ψ(x).

10.2

Half-Elliptic Boundary Conditions

For the remainder of this chapter, we study the nilpotent operators d and δ on bounded domains D, at least Lipschitz regular, in Euclidean space X. The main idea in this section is to use the commutation theorem (Theorem 7.2.9) and reduce the problems to smooth domains. Realizing the operators that are implicit in Definition 7.6.1 as unbounded nilpotent operators, we have the following. Definition 10.2.1 (d and δ on domains). Let D be a bounded Lipschitz domain in a Euclidean space (X, V ). Define unbounded linear operators d, d, δ, δ in L2 (D) = L2 (D; ∧V ) as follows. Assume that F, F 0 ∈ L2 (D) and consider the equation Z hF 0 (x), φ(x)i + hF (x), ∇ y φ(x)i dx = 0. D

If this holds for all φ ∈ C0∞ (D), then we define F ∈ D(d) and dF := F 0 . If this holds for all φ ∈ C ∞ (D), then we define F ∈ D(d) and dF := F 0 .

351

10.2. Half-Elliptic Boundary Conditions Assume that F, F 0 ∈ L2 (D) and consider the equation Z hF 0 (x), φ(x)i + hF (x), ∇ ∧ φ(x)i dx = 0. D

If this holds for all φ ∈ C0∞ (D), then we define F ∈ D(δ) and δF := F 0 . If this holds for all φ ∈ C ∞ (D), then we define F ∈ D(δ) and δF := F 0 . We recall from Section 7.6 that by Stokes’s theorem we interpret F ∈ D(d) as being normal at ∂D in a weak sense, and F ∈ D(δ) as being tangential at ∂D in a weak sense. Basic properties of these operators are the following. Proposition 10.2.2 (Nilpotence). Let D be a bounded Lipschitz domain in a Euclidean space. Then the operators d, d, δ, δ are well-defined nilpotent operators on L2 (D). In particular, they are linear, closed, and densely defined. With the pointwise Hodge star and involution maps, we have δ(F ∗) = (dFb)∗, d(F ∗) = (δ Fb)∗,

F ∈ D(d), F ∈ D(δ).

Proof. Consider d. The other proofs are similar. That d is defined on C0∞ (D), linear, and closed is clear. It is well defined, since F = 0 implies F 0 = 0, since C ∞ (D) is dense in L2 (D). To show nilpotence, assume F ∈ D(d). Then Z Z 0 + hdF (x), ∇ y φ(x)i dx = − hF (x), ∇ y (∇ y φ(x))idx = 0 D

D

for all φ ∈ C ∞ (D), which shows that d(dF ) = 0. The relation between d, δ and δ, d follows from Proposition 7.1.7(i).

The goal of this section is to prove the following duality. Recall the definition (6.4) of adjointness in the sense of unbounded operators. Proposition 10.2.3 (Duality). Let D be a bounded Lipschitz domain in a Euclidean space. Then d and −δ are adjoint in the sense of unbounded operators. Similarly, d and −δ are adjoint in the sense of unbounded operators. From Propositions 10.1.2 and 10.2.3 we obtain a Hodge decomposition with tangential boundary conditions L2 (D) = R(d) ⊕ Ck (D) ⊕ R(δ), where Ck (D) := N(d) ∩ N(δ), and a Hodge decomposition with normal boundary conditions L2 (D) = R(d) ⊕ C⊥ (D) ⊕ R(δ), where C⊥ (D) := N(d) ∩ N(δ). We will prove in Section 10.3 that the ranges of the four operators are closed, so the closures here are redundant. For the proof of Proposition 10.2.3, we need the following results.

352

Chapter 10. Hodge Decompositions

Lemma 10.2.4 (Local nonsmooth commutation theorem). Let ρ : D1 → D2 be a Lipschitz diffeomorphism between domains D1 and D2 in Euclidean space. If F ∈ D(d) in D2 with supp F ⊂ D2 , then ρ∗ F ∈ D(d) with d(ρ∗ F ) = ρ∗ (dF ) in D1 . Similarly, if F ∈ D(δ) in D1 with supp F ⊂ D1 , then ρ˜∗ F ∈ D(δ) with δ(˜ ρ∗ F ) = ρ˜∗ (δF ) in D2 . We recall, for example, that supp F ⊂ D2 means that F = 0 in a neighborhood of ∂D2 . Note that for general Lipschitz changes of variables, ρ∗ F and ρ˜∗ F are defined almost everywhere by Rademacher’s theorem. Proof. By Proposition 7.2.7 it suffices to prove the first statement. Consider first F ∈ C0∞ (D2 ). We mollify and approximate ρ by ρt (x) := ηt ∗ ρ(x), where η ∈ C0∞ (X; R) with η = 1 and ηt (x) := t−n η(x/t). Note that ρt is well defined on every compact subset of D1 for small t. It follows that ρt ∈ C ∞ and R

d(ρ∗t F ) = ρ∗t (dF ) holds by Theorem 7.2.9. From the dominated convergence theorem we conclude that ρ∗t F → ρ∗ F in L2 (D1 ). Since for the same reason ρ∗t (dF ) → ρ∗ (dF ), and d is a closed operator, it follows that ρ∗ F ∈ D(d) and d(ρ∗ F ) = ρ∗ (dF ). Next consider general F ∈ D(d) with compact support in D2 . Similarly to above, we now mollify and approximate F by Fn ∈ C0∞ (D2 ), with Fn → F and dFn → dF in L2 (D2 ). We have shown above that d(ρ∗ Fn ) = ρ∗ (dFn ). Using that ρ∗ : L2 (D2 ) → L2 (D1 ) is bounded and that d is closed, it follows that ρ∗ F ∈ D(d) and d(ρ∗ F ) = ρ∗ (dF ). The following shows that the normal and tangential boundary conditions for d and δ are obtained by closure from C0∞ . Proposition 10.2.5 (Half Dirichlet conditions). Let D be a bounded Lipschitz domain in a Euclidean space. If F ∈ D(d), then there exists Ft ∈ C0∞ (D) such that Ft → F

and

dFt → dF

in L2 (D) as t → 0. Similarly, if F ∈ D(δ), then there exists Ft ∈ C0∞ (D) such that Ft → F and δFt → δF in L2 (D) as t → 0. Proof. By Hodge star duality it suffices to consider d. By the compactness of D, we can localize and assume that supp F ⊂ Dp ∩ D near p ∈ ∂D as in Definition 6.1.4. We note from Definition 10.2.1 that extending F by 0 outside D, we have F ∈ D(d) on X as in Example 10.1.8. Pulling back by the local parametrization ρ, Lemma 10.2.4 shows that ρ∗ F ∈ D(d) on Rn . We translate ρ∗ F up into Ωp and

10.2. Half-Elliptic Boundary Conditions

353

pull back by ρ−1 to define F˜t := (ρ∗ )−1 (ρ∗ F (x0 , xn − t)). This yields F˜t ∈ D(d) with supp F˜t ⊂ D. Finally, we mollify and approximate F˜t by Ft (x) := ηt (x) ∗ F˜t (x), x ∈ D, R where η ∈ C0∞ (X; R) with η = 1, supp η ⊂ B(0, r), and ηt (x) := t−n η(x/t). If 0 < r < t is chosen small enough depending on the Lipschitz geometry, we obtain Ft ∈ C0∞ (D) and can verify that Ft and dFt converge to F and dF respectively. Proof of Proposition 10.2.3. Consider the equation Z hF 0 (x), φ(x)i + hF (x), ∇ y φ(x)i dx = 0. D

This holds for all F ∈ D(d) with F 0 = dF , and all φ ∈ C0∞ (D) by Definition 10.2.1. By Proposition 10.2.5 and a limiting argument, this continues to hold for φ ∈ D(δ). This shows that d and −δ are formally adjoint. Furthermore, assume that the equation holds for some F and F 0 ∈ L2 (D) and all φ ∈ D(δ). In particular, it holds for all φ ∈ C0∞ (D), and it follows by definition that F ∈ D(d) and F 0 = dF . This shows that d and −δ are adjoint in the sense of unbounded operators. The proof that d and −δ are adjoint in the sense of unbounded operators is similar. We next remove the assumption of compact support in Lemma 10.2.4. Lemma 10.2.6 (Nonsmooth commutation theorem). Let ρ : D1 → D2 be a Lipschitz diffeomorphism between bounded Lipschitz domains D1 and D2 in Euclidean space. If F ∈ D(d) on D2 , then ρ∗ F ∈ D(d) on D(D1 ) with d(ρ∗ F ) = ρ∗ (dF ) in D1 . Similarly, if F ∈ D(δ) on D1 , then ρ˜∗ F ∈ D(δ) with δ(˜ ρ∗ F ) = ρ˜∗ (δF ) on D2 . Proof. By Proposition 10.2.2, it suffices to consider d. In this case, we must show that Z hρ∗ (dF ), φi + hρ∗ F, ∇ y φi dx = 0 D1

C0∞ (D1 ).

for φ ∈ By the Lipschitz change of variables formula (6.2), see Section 6.5, and Lemma 10.2.4, this is equivalent to Z hdF, ρ˜∗ φi + hF, ∇ y (˜ ρ∗ φ)i dx = 0, D2

which holds by Proposition 10.2.3.

It is clear from the definition that D(d) on D can be viewed as a subspace of D(d) on X, by extending F on D by zero to all X. The following existence of extension maps shows that D(d) on D can be identified with the quotient space D(dX )/D(dX\D ).

Chapter 10. Hodge Decompositions

354

Proposition 10.2.7 (Extensions for d and δ). Let D be a bounded Lipschitz domain in a Euclidean space X. Assume that F ∈ D(d) on D. Then there exists F˜ ∈ D(d) on X such that F˜ |D = F . Furthermore, there exists Ft ∈ C ∞ (D) such that Ft → F and dFt → dF in L2 (D) as t → 0. Similarly, assume that F ∈ D(δ) on D. Then there exists F˜ ∈ D(δ) on X such that F˜ |D = F . Furthermore, there exists Ft ∈ C ∞ (D) such that Ft → F and δFt → δF in L2 (D) as t → 0. Proof. As in the proof of Proposition 10.2.5, it suffices to consider d, and we may assume that supp F ⊂ Dp ∩ D, a small neighborhood of p ∈ ∂D. By Lemma 10.2.6 we have ρ∗ F ∈ D(d) on Ωp ∩ {xn > 0}. Define ( ρ∗ F (x), xn > 0, G(x) := ∗ ∗ R ρ F (x), xn < 0, where R(x0 , xn ) := (x0 , −xn ) denotes reflection in Rn−1 . We claim that G ∈ D(d) on all Ωp across Rn−1 . To see this, for φ ∈ C0∞ (Ωp ), we calculate Z Z ∗ ∗ hdρ F, φi + hρ F, ∇ y φi dx + hdR∗ ρ∗ F, φi + hR∗ ρ∗ F, ∇ y φi dx xn >0 xn 0

Since φ + R∗ φ is tangential on Rn−1 , we have φ + R∗ φ ∈ D(δ) on Ωp ∩ Rn+ , so by Proposition 10.2.3, the integral vanishes. By Lemma 10.2.4, the field F˜ := (ρ∗ )−1 G ∈ D(d) on X is an extension of F , and if we mollify and approximate F˜ by Ft (x) := ηt ∗ F˜ (x),

x ∈ D,

∞

as above, we obtain Ft ∈ C (D) and can verify that Ft and dFt converge to F and dF respectively.

10.3

Hodge Potentials

Our main result on Hodge decompositions is the following. Theorem 10.3.1 (Hodge decompositions on Lipschitz domains). Let D be a bounded Lipschitz domain in a Euclidean space X. Then the operators d, δ, d, δ in L2 (D; ∧V ) all have closed ranges, the cohomology spaces Ck (D) = N(d) ∩ N(δ) and C⊥ (D) = N(d) ∩ N(δ) are finite-dimensional, and we have Hodge decompositions L2 (D; ∧V ) = R(d) ⊕ Ck (D) ⊕ R(δ) = R(d) ⊕ C⊥ (D) ⊕ R(δ). Moreover, the inverses of d : R(δ) → R(d), δ : R(d) → R(δ), d : R(δ) → R(d), and δ : R(d) → R(δ) are all L2 compact.

355

10.3. Hodge Potentials

The proof follows from the following reduction and Theorem 10.3.3 below. Reduction of Theorem 10.3.1 to a ball. We prove that there are compact operators K0 and K1 on L2 (D) such that dK1 F + K1 dF + K0 F = F for all F ∈ D(d). By Propositions 10.1.6 and 10.2.2, this will prove TheoremS10.3.1. By Definition 6.1.4 we have a finite covering D = α Dα , with Lipschitz diffeomorphisms ρα : B → Dα from the unit ball B. Moreover, we have a partition of unity ηα ∈ C ∞ (D) subordinate to this covering. By Theorem 10.3.3 for the ball B, we have compact maps K1B and K0B on L2 (B) such that dK1B F + K1B dF + K0B F = F . Note that we need only part (i) in the proof of Theorem 10.3.3 for this. Define X K1 F := ηα (ρ∗α )−1 K1B (ρ∗α F |Dα ), α

which is seen to be compact on L2 (D). We calculate X X dK1 F = ηα (ρ∗α )−1 (I − K1B d − K0B )(ρ∗α F |Dα )+ ∇ηα ∧ (ρ∗α )−1 K1B (ρ∗α F |Dα ) α

α

= F − K1 dF − K0 F, where K0 F :=

X

ηα (ρ∗α )−1 K0B (ρ∗α F |Dα ) −

X

α

∇ηα ∧ (ρ∗α )−1 K1B (ρ∗α F |Dα )

α

is seen to be compact on L2 (D). Note the critical use of Theorem 7.2.9. This proves Theorem 10.3.1 for Lipschitz domains D. In the proof of Theorem 10.3.1 we used Proposition 10.1.6(ii). As for the characterization (iii), it is natural to ask whether D(d) ∩ D(δ) ⊂ H 1 (D), that is, whether the total derivative ∇⊗F belongs to L2 (D), whenever F, dF, δF ∈ L2 (D). This is not true for general Lipschitz domains, where the irregularities of ∂D may prevent F ∈ D(d) ∩ D(δ) from having full Sobolev H 1 regularity, but it does hold for smooth domains. Example 10.3.2 (Nonconvex corner). Let Dα ⊂ R2 be a bounded domain that is smooth except at 0, in a neighborhood of which Dα coincides with the sector {reiφ ; r > 0, 0 < φ < α}. Define a scalar function u : Dα → R such that u = rπ/α sin(πφ/α)η, where η ∈ C0∞ (R2 ), η = 1 in a neighborhood of 0, and η = 0 where Dα differs from the sector. Consider the gradient vector field F := ∇u ∈ R(d). Using the estimate |F | . rπ/α−1 , we verify that F ∈ D(d) ∩ D(δ). However, Z Z 1 |∇ ⊗ F |2 dxdy & (rπ/α−2 )2 rdr. D

0

Chapter 10. Hodge Decompositions

356

Therefore, when Dα is not convex, that is, when α > π, then F ∈ / H 1 (D). Figure 10.1 shows the case α = 3π/2.

Figure 10.1: The harmonic function r2/3 sin(2φ/3) in quadrants 1–3 in the unit circle, with Dirichlet boundary conditions but infinite gradient at the origin.

Theorem 10.3.3 (Full regularity of Hodge potentials). Let D be a bounded C 2 domain. Then D(d) ∩ D(δ) = Hk1 (D) := {F ∈ H 1 (D) ; ν y F |∂D = 0} 1

and

1

D(d) ∩ D(δ) = H⊥ (D) := {F ∈ H (D) ; ν ∧ F |∂D = 0}. For the proof of Theorem 10.3.3 we shall prove a Weitzenb¨ock identity for d and δ on D, involving a boundary curvature term. This requires the following definitions from differential geometry and uses that the boundary ∂D is C 2 regular. In this case, the unit normal vector field ν on ∂D is C 1 , and the curvature of the boundary is a continuous function. Proposition 10.3.4 (Derivative of normal). Let D be a bounded C 2 domain, with outward-pointing unit normal vector field ν on ∂D. At p ∈ ∂D, let Tp (∂D) denote the tangent hyperplane. Then the map p S∂D : Tp (∂D) → Tp (∂D) : v 7→ ∂v ν,

is linear and symmetric. Moreover, for any tangential C 1 vector fields u and v on ∂D, at each p ∈ ∂D we have p p hu, S∂D vi = −h∂u v, νi = hS∂D u, vi.

357

10.3. Hodge Potentials

Proof. We have 0 = ∂v |ν|2 = 2h∂v ν, νi, since |ν| = 1 on ∂D, which shows that p p at p ∈ ∂D, we note (v) is a tangential vector. To show the symmetry of S∂D S∂D that 0 = ∂u hv, νi = h∂u v, νi + hv, ∂u νi and 0 = ∂v hu, νi = h∂v u, νi + hu, ∂v νi, p since u and v are tangential on ∂D. The symmetry of S∂D now follows, since the Lie bracket ∂u v − ∂v u = [u, v] = Lu v is tangential.

Definition 10.3.5 (Second fundamental form). Let D be a bounded C 2 domain. The symmetric bilinear form p B∂D : Tp (∂D) × Tp (∂D) → R : (u, v) 7→ −h∂u v, νi

from Proposition 10.3.4 is called the second fundamental form for ∂D. The associated symmetric map p S∂D : Tp (∂D) → Tp (∂D) : v 7→ ∂v ν

from Proposition 10.3.4 is called the Weingarten map, or shape operator, for ∂D. p The eigenvalues {κ1 , . . . , κn−1 } of S∂D are called the principal curvatures of ∂D 0 at p, and a corresponding ON-basis {e10 , . . . , en−1 } for Tp (∂D) of eigenvectors to p is referred to as the principal directions of curvatures at p. S∂D Note that if D is a convex domain, then κj ≥ 0. Theorem 10.3.6 (Weitzenb¨ock identities). Let D be a bounded C 2 domain, and let e0j denote the principal directions of curvatures κj , j = 1, . . . , n − 1. Then Z

Z

2

2

(|dF | + |δF | )dx −

|∇ ⊗ F | dx = D

Z

2

D

|∇ ⊗ F |2 dx =

Z

j=1

(|dF |2 + |δF |2 )dx −

D

D

where |∇ ⊗ F |2 =

Pn

n−1 XZ

j=1

n−1 XZ j=1

κj |ej0

F |2 dy,

F ∈ H⊥1 (D),

κj |ej0 y F |2 dy,

F ∈ Hk1 (D),

∧

∂D

∂D

|∂j F |2 .

Example 10.3.7 (Kadlec’s formula). Consider a scalar function U : D → R satisfying Poisson’s equation ∆U = f in D, with Dirichlet boundary conditions U |∂D = 0. This means that its gradient vector field F = ∇U is normal at the boundary. Assuming that F ∈ H⊥1 (D), we have Kadlec’s formula n Z X i,j=1

D

2

Z

2

Z

|f | dx − (n − 1)

|∂i ∂j U | dx = D

∂D

H(y) |∇U |2 dy,

Chapter 10. Hodge Decompositions

358

y where H(y) := Tr(S∂D )/(n − 1) is the mean curvature of the boundary. Note that Lagrange’s identity Proposition 3.1.1 shows that |e0j ∧ F |2 = |e0j |2 |F |2 −|he0j , F i|2 = |F |2 . If instead U satisfies Neumann boundary conditions hν, ∇U i = 0, then we get a similar identity n Z X i,j=1

|∂i ∂j U |2 dx =

Z

|f |2 dx −

D

D

n Z X j=1

κj |hej0 , ∇U i|2 dy,

∂D

but where all the principal curvatures appear and not only the mean curvature. For convex domains, the Weitzenb¨ock identities imply that Z Z |∇ ⊗ F |2 dx ≤ (|dF |2 + |δF |2 )dx, for all F ∈ H⊥1 (D) ∪ Hk1 (D), D

D

since in this case all κj ≥ 0. In general, we have the following estimates. Corollary 10.3.8 (Gaffney’s inequality). Let D be a bounded C 2 domain. Then Z Z |∇ ⊗ F |2 dx . (|dF |2 + |δF |2 + |F |2 )dx, for all F ∈ H⊥1 (D) ∪ Hk1 (D). D

D

Proof. For a C 2 domain, we note that the principal curvatures κj are bounded 2 . To functions, which shows that the boundary integral terms are . kF kL 2 (∂D) 2 replace this by a term kF kL2 (D) , we apply Stokes’s theorem to obtain a standard trace estimate as follows. Let θ ∈ C0∞ (X; V ) be a vector field such that inf y∈∂D hθ(y), ν(y)i > 0, that is, θ is uniformly outward pointing on ∂D. Stokes’s theorem gives Z Z |F |2 hθ, νidy = 2h∂θ F, F i + |F |2 div θ dx. ∂D

D

Estimating, this shows that Z Z 2 2 kF kL2 (∂D) . |∇ ⊗ F ||F | + |F |2 dx. |F | hθ, νidy . ∂D

D

It follows from the Weitzenb¨ock identities that Z Z Z Z |∇ ⊗ F ||F |dx + C (|dF |2 + |δF |2 )dx + C |∇ ⊗ F |2 dx ≤ |F |2 dx, (10.3) D

D

D

D

for some constant C < ∞. We next use an estimate technique called the absorption inequality, which is 1 2 b . ab ≤ 2 a2 + 2

10.3. Hodge Potentials

359

√ √ This is, of course, nothing deeper than ( a − b/ )2 ≥ 0. To use this, we take a = |∇ ⊗ F (x)|, b = |F (x)|, and = C −1 . This shows that the second term on the right-hand side in (10.3) is Z |∇ ⊗ F ||F |dx ≤

C D

1 2

Z

2

|∇ ⊗ F | dx + D

C2 2

Z

|F |2 dx,

D

where the first term can be moved to the left-hand side in (10.3) and be absorbed there. Gaffney’s inequality follows. Proof of Theorem 10.3.6. (i) Let first F ∈ C 2 (D) and consider the 1-form θ(x, v) :=

n X hv, ej ihF (x), ∂j F (x)i − hv ∧ F (x), dF (x)i − hv y F (x), δF (x)i, j=1

for x ∈ D, v ∈ V . We calculate its exterior derivative θ(x, ˙ ∇) = (|∇ ⊗ F |2 + hF, ∆F i) − (|dF |2 + hF, δdF i) − (|δF |2 + hF, dδF i) = |∇ ⊗ F |2 − |dF |2 − |δF |2 , since ∆ = δd + dδ. The Stokes’ formula (7.4) gives Z

Z |∇ ⊗ F |2 − |dF |2 − |δF |2 dx =

D

hF, ∂ν F i − hν ∧ F, dF i − hν y F, δF i dy.

∂D

We continue and rewrite the right-hand side with nabla calculus as hF, hν, ∇iF i − hF, ν ∧ (∇ y F )i − hν ∧ F, ∇ ∧ F i = hF, ∇ y (n ∧ F˙ )i − hν ∧ F, ∇ ∧ F i = −hF, ∇ y (n˙ ∧ F )i + hF, ∇ y (n ∧ F )i − hν ∧ F, ∇ ∧ F i,

(10.4) (10.5)

where n ∈ C 1 (X; V ) denotes an extension of ν. The first step uses the algebraic anticommutation relation ν ∧ (∇ y F ) = hν, ∇iF − ∇ y (n ∧ F˙ ), and the second step uses the analytic product rule ∇ y (n ∧ F ) = ∇ y (n˙ ∧ F ) + ∇ y (n ∧ F˙ ). At p ∈ ∂D, we calculate the first term in the ON-basis {e01 , . . . , e0n−1 , ν} and get hF, ∇ y (n˙ ∧ F )i =

n−1 X

κj |e0j

∧

F |2 + hν ∧ F, (∂ν n) ∧ F i.

j=1

On the other hand, the normal derivatives in the last two terms in (10.4) are hF, ν y ∂ν (n ∧ F )i − hν ∧ F, ν ∧ ∂ν F i = hν ∧ F, (∂ν n) ∧ F i.

Chapter 10. Hodge Decompositions

360

Therefore these three terms involving the normal derivatives cancel, and we obtain the identity Z (|∇ ⊗ F |2 − |dF |2 − |δF |2 )dx D

=−

n−1 XZ j=1

κj |e0j ∧

2

Z

F | dy +

∂D

hF, ∇0 y (ν ∧ F )i − hν ∧ F, ∇0 ∧ F i dy, (10.6)

∂D

Pn−1 where ∇0 := ν y (ν ∧ ∇) = j=1 e0j ∂e0j . (ii) Next consider F ∈ H⊥1 (D). To obtain the first Weitzenb¨ock identity, we use the fact that C 2 (D) is dense in H 1 (D) and take Fj ∈ C 2 (D) such that Fj → F and ∇ ⊗ Fj → ∇ ⊗ F in L2 (D). On the C 2 manifold ∂D, we use the Sobolev spaces H 1/2 (∂D) and H −1/2 (∂D), as discussed in Example 6.4.1, where H 1/2 ⊂ L2 ⊂ H −1/2 . As usual, we allow the functions to be multivector fields, and require that each component function be in such Sobolev space. We need the following well-known facts. • The trace map H 1 (D) → H 1/2 (∂D) : F 7→ F |∂D is a bounded linear operator. • The tangential derivative ∇0 defines a bounded linear operator ∇0 : H 1/2 (∂D) → H −1/2 (∂D). • Multiplication by a C 1 function like ν is a bounded operation on H 1/2 (∂D). • The spaces H 1/2 (∂D) and H −1/2 (∂D) are dual; in particular, we have the estimate Z . kF kH 1/2 (∂D) kGkH −1/2 (∂D) . hF, Gidx ∂D

Given this, we apply (10.6) to Fj and take the limit as j → ∞. Since ν ∧ Fj → ν ∧ F = 0 in H 1/2 (∂D), we obtain the Weitzenb¨ock identity for F ∈ H⊥1 (D). (iii) To obtain the Weitzenb¨ock identity for F ∈ Hk1 (D), we instead rewrite θ as θ(x, ν) = hF, hν, ∇iF i − hF, ν y (∇ ∧ F )i − hν y F, ∇ y F i = hF, ∇ ∧ (n y F˙ )i − hν y F, ∇ y F i = −hF, ∇ ∧ (n˙ y F )i + hF, ∇ ∧ (n y F )i − hν y F, ∇ y F i, and proceed as in (i) and (ii).

We finally prove Theorem 10.3.3. From the Weitzenb¨ock identities, the Gaffney inequalities show that on C 2 domains we have Hk1 (D) ⊂ D(d) ∩ D(δ) and H⊥1 (D) ⊂ D(d) ∩ D(δ) and that k∇ ⊗ F k2 + kF k2 ≈ k∇ ∧ F k2 + k∇ y F k2 + kF k2

361

10.3. Hodge Potentials

in L2 (D) norm, for all F ∈ Hk1 (D) and all F ∈ H⊥1 (D). It is important to note that this equivalence of norms, without further work, does not imply that D(d)∩D(δ ) ⊂ Hk1 (D) and D(d) ∩ D(δ) ⊂ H⊥1 (D). It particular, the absorption technique in the proof of Corollary 10.3.8 fails to prove this. Proof of Theorem 10.3.3. It remains to prove D(d) ∩ D(δ) ⊂ H⊥1 (D). By Hodge duality as in Proposition 10.2.2, this will imply the corresponding result for normal boundary conditions. (i) Consider first the case that D is the unit ball B := {x ∈ V ; |x| < 1} and let F ∈ D(d) ∩ D(δ). Using a partition of unity, we write F = F0 + F1 , F0 , F1 ∈ D(d) ∩ D(δ), where F0 (x) = 0 when |x| > 1/2 and F1 (x) = 0 when |x| < 1/3. We use inversion R(x) = 1/x in the unit sphere, with derivative Rx h = −x−1 hx−1 to extend F1 to ( F1 (x), |x| < 1, ˜ F1 (x) := ∗ R F1 (x), |x| > 1. Arguing as in the proof of Proposition 10.2.7, replacing Rn−1 by the sphere |x| = 1, we conclude that F˜ ∈ D(d) on X. Moreover, R is a conformal map and R∗ = ˜ ∗−1 . From this it follows that R∗ F1 ∈ D(δ) on X, with |x|2(n−1) R ˜ ∗−1 F1 (x) + |x|2(n−1) R ˜ ∗−1 (∇ y F1 )(x) ∇ y R∗ F1 (x) = |x|2(n−2) x y R for |x| > 1. Recall that F1 , extended by 0 for |x| > 1 belongs to D(δ). We obtain an extension F˜ := F0 + F˜1 of F to X, with F˜ = 0 for |x| > 3 and ˜ ˜ F , dF , and δ F˜ all belonging to L2 (X). By Plancherel’s theorem and Langrange’s identity, we get Z Z Z (2π)n |∇ ⊗ F˜ |2 dx = |F(F˜ )|2 |ξ|2 dξ = |ξ ∧ F(F˜ )|2 + |ξ y F(F˜ )|2 dξ < ∞. X

X

X

Recall from Example 10.1.8 that d and δ on X are the Fourier multipliers iµξ and iµ∗ξ . This shows that F ∈ H 1 (D). (ii) Next consider a general bounded C 2 domain D. Localizing the problem with a partition of unity, we may assume that D is C 2 diffeomorphic to B. Moreover, we may assume that we have a C 2 map ρ : [0, 1] × B → X such that ρt = ρ(t, ·) defines a C 2 diffeomorphism B → ρt (B) =: Dt , with D0 = B and D1 = D. For fixed t ∈ [0, 1], we consider the inclusion Hk1 (Dt ) ⊂ D(d)∩D(δ) on the C 2 domain Dt . We note from Proposition 10.1.3 that I + d + δ : D(d) ∩ D(δ) → L2 (Dt ) is an invertible isometry, so the inclusion amounts to I + d + δ : Hk1 (Dt ) → L2 (Dt ) being an injective semi-Fredholm operator. See Definition 6.4.9. From (i), we know that it is surjective for the ball at t = 0. To apply the method of continuity and

Chapter 10. Hodge Decompositions

362

conclude that it is surjective for all t, and in particular for D at t = 1, we note that the normalized pushforward ρ˜t∗ defines invertible maps Hk1 (B) → Hk1 (Dt ) and L2 (B) → L2 (Dt ). The method of continuity therefore applies to the family of semi-Fredholm operators ρt∗ )−1 (I + d + δ)˜ (˜ ρt∗ : Hk1 (B) → L2 (B). We conclude that I + d + δ : Hk1 (Dt ) → L2 (Dt ) is invertible, which shows that D(d) ∩ D(δ ) = Hk1 (D) and completes the proof of Theorem 10.3.3.

10.4

Bogovski˘ı and Poincar´e Potentials

Recall that exterior and interior potentials in general are highly nonunique. In this section we prove the following surprising results about potentials on strongly Lipschitz domains D. • We have seen in Example 10.3.2 that in contrast to smooth domains, the potential U in the subspace R(δ) to F = dU ∈ R(d) may not belong to H 1 (D). We refer to this potential U as the Hodge potential for F , which is characterized by its minimal L2 norm. It follows from Theorem 10.4.3 below that every exact field F ∈ R(d) on any bounded and strongly Lipschitz domain D nevertheless has a potential ˜ , in general different from the Hodge potential, such that U ˜ ∈ H 1 (D) and U dU = F . We refer to such potentials as (regularized) Poincar´e potentials for F. • We have seen that the Hodge potential U ∈ R(d) to F = δU ∈ R(δ) is tangential on ∂D, meaning that half of the component functions of U vanish there. Theorem 10.4.3 below show that every field F ∈ R(δ) on any bounded ˜ , in general different and strongly Lipschitz domain D in fact has a potential U ˜ ∈ H 1 (D) and δU = F . This means from the Hodge potential, such that U 0 ˜ vanish on ∂D, and we note that this is that all component functions of U a nontrivial result also for smooth domains. We refer to such potentials as Bogovski˘ı potentials for F . ˜∈ Similarly, and related by the Hodge star, there exist Poincar´e potentials U ˜ ∈ H 1 for F ∈ R(d). We will H 1 (D) for F ∈ R(δ), and Bogovski˘ı potentials U 0 formulate the results only for d and δ, and leave it to the reader to translate the results in this section to d and δ. First consider a star-shaped domain D. In what follows, we extend the operators initially defined on k-vector fields, by linearity to act on general multivector fields. The method we use to construct a Poincar´e potential U to a given field F ∈ R(d) on D builds on Poincar´e’s Theorem 7.5.2. If D is shar-shaped with respect to p0 ∈ D, then this gives the potential Z 1 Tp0 (F )(x) := (x − p0 ) y F (p0 + t(x − p0 )) tk−1 dt, x ∈ D, (10.7) 0

10.4. Bogovski˘ı and Poincar´e Potentials

363

provided k ≥ 1 and F is a smooth k-vector field. For a scalar function F : D → ∧0 R, we let Tp0 F = 0. We would like to extend (10.7) to fields that are square integrable, without any assumption on regularity. To obtain a bounded operator, we need to average the formula (10.7) over base points p around p0 . In what follows, we assume that D is star-shaped not only with respect to a point, but to a whole ball. We fix θ ∈ C0∞ (B(p0 ; )) andR assume that D is star-shaped with respect to each p ∈ B(p0 ; ), where > 0 and θdx = 1. Then define the averaged operator Z TD F (x) :=

θ(p)Tp F (x)dp,

x ∈ D.

(10.8)

|p−p0 |≤

We rewrite this formula by changing the variables p and t to y := p + t(x − p) and s = 1/(1 − t) − 1. This gives Z TD F (x) = (x − y) y F (y) kθ (x, y)dy, (10.9) D

where Z

∞

kθ (x, y) :=

θ(y + s(y − x))sk−1 (1 + s)n−k ds.

0

This operator TD constructs the regularized Poincar´e potential for an exact kvector field on a bounded domain that is star-shaped with respect to B(p0 ; ). Exercise 10.4.1 (Kernel support). Show that kθ (x, y) 6= 0 is possible only when y lies on the straight line between x and a point p ∈ supp η, and that we have estimates 1 |kθ (x, y)| . , x, y ∈ D, |x − y|n so that TD is a weakly singular integral operator. Note how by averaging with θ we have replaced the line integral for Tp0 by a volume integral over a conical region for TD . The adjoint operator ∗ TD F (x) =

Z (y − x) ∧ F (y) kθ (y, x)dy

(10.10)

D

constructs the Bogovski˘ı potential for a (k − 1)-vector field F ∈ R(δ) on the star∗ ∗ shaped domain D. We see from Exercise 10.4.1 that TD F |∂D = 0, since for TD we integrate over a cone starting at x, away from B(p0 , ). For domains D that are C 2 diffeomorphic to a domain that is star-shaped with respect to a ball, we can pull back and push forward these operators TD and ∗ TD to obtain Bogovski˘ı and regularized Poincar´e potentials. Next we extend these constructions to general strongly Lipschitz domains, and provide the necessary analysis.

Chapter 10. Hodge Decompositions

364

Definition 10.4.2 (Bogovski˘ı and Poincar´e maps). Let D be a bounded and strongly S Lipschitz domain. Fix a finite cover D = α Dα by domains Dα that Rare starshaped with respect to balls B(pα ; ). Further fix θα ∈ C0∞ (B(pα ; )) with θα dx = 1 and a partition of unity ηα ∈ C ∞ (D) subordinate to the covering Dα . We assume that ηα = 1 on a neighborhood of supp θα . The regularized Poincar´e map with these choices Dα , θα , ηα , for d on D is X TD F (x) = ηα (x)TDα (F |Dα )(x), x ∈ D. α

The Bogovski˘ı map, with these choices Dα , θα , ηα , for δ on D is X ∗ ∗ TD (ηα F |Dα )(x), x ∈ D. TD F (x) = α α ∗ Here TDα and TD are the Poincar´e and Bogovski˘ı maps on the star-shaped doα mains Dα , constructed as above.

Unlike the star-shaped case, these Bogovski˘ı and regularized Poincar´e maps on general strongly Lipschitz domains do not straight away give potentials for (co-)exact fields. We proceed as follows. Theorem 10.4.3 (Bogovski˘ı and Poincar´e homotopies). Let D be a bounded and strongly Lipschitz domain. The regularized Poincar´e potential map from Definition 10.4.2, maps TD : C ∞ (D) → C ∞ (D) and extends by continuity to a bounded operator TD : L2 (D) → H 1 (D). ∗ The Bogovski˘ı potential map from Definition 10.4.2 maps TD : C0∞ (D) → C0∞ (D) and extends by continuity to a bounded operator ∗ TD : L2 (D) → H01 (D).

We have homotopy relations d(TD F ) + TD (dF ) + KD F = F, ∗ ∗ −δ(TD F = F, F ) − T ∗ (δF ) + KD

F ∈ D(d), F ∈ D(δ),

with perturbation terms KD F (x) =

X

Z ηα (x)

θα F0 dy +

α ∗ KD F (x) =

X α

X

∇ηα (x) ∧ TDα (F |Dα )(x),

α

Z θα (x)

ηα F0 dy +

X

∗ TD (∇ηα y F |Dα )(x), α

α

∗ which are bounded, KD : L2 (D) → H 1 (D) and KD : L2 (D) → H01 (D). Here F0 0 denotes the ∧ V part of F .

365

10.4. Bogovski˘ı and Poincar´e Potentials

To see how Theorem 10.4.3 implies the existence of Bogovski˘ı and Poincar´e potentials, we consider the following Hodge decomposition: L2 (D)

L2 (D)

⊕

R(d)

=

=

R(d)

t

Ck (D)

δ

d

⊕

⊕

Ck (D)

⊕

*

R(δ)

(10.11)

R(δ)

Given F ∈ R(d), we apply the homotopy relation to the Hodge potential U ∈ R(δ), with dU = F , to obtain U = dTD U + TD dU + KD U , and in particular, F = dU = d(TD F + KD U ). ˜ := TD F + KD U ∈ H 1 (D) is a regularized Poincar´e potential Therefore the field U for F . Similarly, for F ∈ R(δ) we apply the homotopy relation to the Hodge potential U ∈ R(d), with δU = F , to obtain ∗ ∗ F = δU = δ(−TD F + KD U ),

˜ := −T ∗ F + K ∗ U ∈ H 1 (D) is a Bogovski˘ı potential for F . where the field U 0 D D Proof of Theorem 10.4.3. (i) Let F ∈ C ∞ (D). Then F |Dα ∈ C ∞ (Dα ), and we see from (10.8) for the star-shaped domain Dα that TDα (F |Dα ) ∈ C ∞ (Dα ). Note that TDα acts on C ∞ (X), but the values TDα F (x), for x ∈ Dα , depend only on F |D . With the partition of unity ηα , we obtain TD F ∈ C ∞ (D). Let F ∈ C0∞ (D). Then ηα F |Dα ∈ C0∞ (Dα ), and we see from Exercise 10.4.1 ∗ that supp TD (ηα F |Dα ) is compactly contained in Dα . To verify smoothness, we α write Z ∞ Z ∗ k−1 n−k TDα G(x) = − z ∧ G(x − z) θα (x + sz)s (1 + s) ds dz. Dα

0

∗ Differentiation with respect to x shows that TD (ηα F |Dα ) ∈ C0∞ (Dα ), and thereα ∗ ∞ fore that TD F ∈ C0 (D). Averaging the homotopy relation in Exercise 7.5.6, we obtain

d(TDα F ) + TDα (dF ) + KDα F = F on Dα , with

Z KDα F :=

θα F0 dx.

As in the proof of Theorem 10.3.1, the product rule yields d(TD F ) + TD (dF ) + KD F = F on D, and duality yields the stated formulas for δ. ∗ . To this end, assume (ii) It remains to establish H 1 bounds for TDα and TD α that D = Dα is star-shaped with respect to a ball and consider the operators (10.9)

Chapter 10. Hodge Decompositions

366

and (10.10). By Exercise 10.4.1, these are weakly singular integral operators, and Schur estimates as in Exercise 6.4.3 show that TD is bounded on L2 (D). Expanding (1 + s)n−k with the binomial theorem, we may further replace kθ (x, y) by Z ∞ θ(y + s(y − x))sn−1 ds. 0

Indeed, in estimating k∇⊗TD F kL2 the difference will be a weakly singular operator ∗ that is bounded as above, and similarly for k∇ ⊗ TD F kL2 . Make the change of variables t = s|y − x|, fix a coordinate 1 ≤ i ≤ n, and define Z ∞ z n−1 zi η x+t k(x, z) := t dt . n |z| |z| 0 Estimating the multivector fields componentwise, we see that it is enough to consider the operators Z Z ∗ Sf (x) := k(y, y − x)f (y) dy and S f (x) := k(x, x − y)f (y) dy, D

D ∗

and prove bounds on k∇Sf kL2 and k∇S f kL2 . We note that k(x, z) is homogeneous of degree −n + 1 with respect to z. For fixed x, we expand k(x, z/|z|) in a series of spherical harmonics on the unit sphere S. We get k(x, z) =

1 |z|n−1

hj ∞ X X

kjm (x)Yjm (z/|z|) =

j=0 m=1

hj ∞ X X

kjm (x)

j=0 m=1

Yjm (z) , |z|n−1+j

h

j denotes an ON-basis for the space Pjsh (S) of scalar-valued spherwhere {Yjm }m=1 ical harmonics, for j ∈ N. See Section 8.2. In particular the coefficients are R kjm (x) := S k(x, z)Yjm (z) dz. Define weakly singular convolution integral operators Z Yjm (x − y) f (y) dy. Sjm (x) := |x − y|n−1+j D

With kjm as multipliers we have Sf (x) =

hj ∞ X X (−1)j Sjm (kjm f )(x), j=0

m=1

S ∗ f (x) =

hj ∞ X X

kjm (x)Sjm f (x).

j=0 m=1

The main estimate we need is kSjm kL2 (D)→H 1 (D) . (1 + j)n−2 .

(10.12)

To see this, we use zonal harmonics as in Section 8.2 to estimate Z |Yjm (z)| = Zj (z, y)Yjm (y)dy ≤ kZj (z, ·)kL2 (S) kYjm kL2 (S) . (1 + j)n−2 |z|j , S

ˇ Cohomology 10.5. Cech

367

which yields the L2 estimate. To bound ∇Sjm f on L2 (X), we use Proposition 6.2.1 to see that ∇Sjm is a Fourier multiplier with estimates ξ2c Γ((1 + j)/2) Yjm (ξ)/|ξ|1+j . (1 + j)n−2 , ξ ∈ X, Γ((n − 1 + j)/2) of the symbol. This proves (10.12). To estimate the multipliers Z kjm (x) =

k(x, z)Yjm (z)dz, S

we use that k(x, ·) is smooth on S, while Yjm becomes more oscillatory as j grows, to show that kjm decays with j as follows. By Proposition 8.2.15, the spherical Laplace operator ∆S is a self-adjoint operator in L2 (S) with ∆S Yjm = (2 − n − j)jYjm . Using self-adjointness N times shows that Z 1 kjm (x) = (∆N k(x, z))Yjm (z)dz. (2 − n − j)N j N S S Since ∆N S k(x, ·), for any fixed N , is bounded, we get the estimate |kjm (x)| . (1 + j)−2N . Similarly, we bound Z ∇kjm (x) =

∇x k(x, z)Yjm (z)dz S

uniformly by (1 + j)−2N . Collecting our estimates, we obtain kSf kH 1 (D) .

∞ X

hj (1 + j)n−2 (1 + j)−N kf kL2 (D) . kf kL2 (D) ,

j=0

kS ∗ f kH 1 (D) .

∞ X

hj (1 + j)−N (1 + j)n−2 kf kL2 (D) . kf kL2 (D) ,

j=0

provided we fix large enough N . This completes the proof.

ˇ 10.5 Cech Cohomology In this section we collect some tools from algebraic topology that we use in Section 10.6 to calculate the dimensions of the finite-dimensional cohomology space

Chapter 10. Hodge Decompositions

368

N(d) ∩ N(δ ), more precisely the Betti numbers bk (D), from Definition 7.6.3. We also use these tools in Chapters 11 and 12. Our starting point is the notion of a sheaf, where we only use the following simplified version of this concept. We consider some set D and some fixed finite covering of it by subsets D1 , . . . , DN , so that D = D1 ∪ · · · ∪ DN . By a sheaf F on D we mean a collection of linear spaces F(D0 ), one for each intersection D0 of the subsets Dj . In fact, it is only the additive structure of ˇ sheaves that is relevant, and in Chapter 11 we shall use Cech cohomology, where the spaces F(D0 ) are the smallest additive group Z2 = {0, 1}. The linear spaces that we use in this chapter are supposed to behave like spaces of functions defined on D0 in the sense that we require that there exist linear restriction maps F(D0 ) → F(D00 ) : f 7→ f |D00 whenever D00 ⊂ D0 ⊂ D. If an intersection D0 is empty, then we require that the linear space F(D0 ) be trivial, that is, F(D0 ) = {0}. The intersections Ds = Ds1 ∩ · · · ∩ Dsk , of distinct subsets Dsj that we consider, are indexed by the 2N subsets s = ˇ {s1 , . . . , sk } ⊂ {1, . . . , N }. Since the Cech algebra that we are about to construct is alternating, we choose below to index the intersections not by s, but by auxiliary basis multivectors es in ∧RN . This is only a formal notation, which turns out to be useful, since it allows us to recycle some, by now well known to us, exterior algebra. Definition 10.5.1 (k-cochains). Let F be a sheaf on D as above, with covering ˇ D = {D1 , . . . , DN }. A Cech k-cochain f associates to each (k + 1)-fold intersection Ds , |s| = k + 1, an element in the linear space F(D0 ), which we denote by hf, es i ∈ F(Ds ). This is not an inner product, but only a convenient notation for the value of f on Ds . We also extend the definition of f homogeneously by letting hf, αes i := αhf, es i, for α ∈ R. ˇ The space of all Cech k-cochains f on D with values in F is denoted by k k C (D; F). Viewing C (D; F) as ⊕s:|s|=k+1 F(Ds ), it is clear that this is a linear space. For k < 0 and k ≥ N we let C k (D; F) := {0}. ˇ The Cech coboundary operator ∂k : C k (D; F) → C k+1 (D; F) is the linear map defined by h∂k f, es i :=

N X hf, ej y es i|Ds ,

|s| = k + 2, f ∈ C k (D; F).

j=1

For k < 0 and k ≥ N − 1, we let ∂k = 0. ˇ We will see that Cech k-cochains and ∂k behave in many ways like k-covector fields and the exterior derivative d. We need some terminology.

ˇ 10.5. Cech Cohomology

369

Definition 10.5.2 (Complex of spaces). A complex of linear spaces is a sequence of linear maps between linear spaces ∂j−2

∂j−3

∂j−1

∂j+1

∂j

∂j+2

→ Vj−2 → Vj−1 → Vj → Vj+1 → Vj+2 →

such that R(∂j−1 ) ⊂ N(∂j ) in Vj . The complex is said to be exact at Vj if R(∂j−1 ) = N(∂j ). If it is exact at all Vj , we say that the complex is exact. More generally, the cohomology of the complex at Vj is the quotient space H j (V ) := N(∂j )/R(∂j−1 ). An important special case occurs when Vj = {0} for some j, so that ∂j = ∂j−1 = 0. In this case, exactness at Vj+1 means that ∂j+1 is injective, and exactness at Vj−1 means that ∂j−2 is surjective. Lemma 10.5.3. If (Vj , ∂j ) is a an exact complex of finite-dimensional linear spaces and Vj1 = Vj2 = 0, then X (−1)j dim Vj = 0. j1 <j<j2

Proof. The dimension theorem for linear maps shows that dim Vj = dim N(∂j ) + dim R(∂j ). Since R(∂j ) = N(∂j+1 ), we get a telescoping sum X X (−1)j (dim N(∂j ) + dim N(∂j+1 )) (−1)j dim Vj = j1 <j<j2

j1 <j<j2

= (−1)j1 +1 dim R(∂j1 ) + (−1)j2 −1 dim N(∂j2 ) = 0.

ˇ ˇ Lemma 10.5.4 (Cech sequence complex). The Cech ∂k−2

∂k−1

∂k+1

∂

∂k+2

k → C k−1 (D; F) → C k (D; F) → C k+1 (D; F) → C k+2 (D; F) →

is a complex of linear spaces. Proof. Let f ∈ C k (D; F) and |s| = k + 3. Then X XX h∂k+1 ∂k f, es i = h∂k f, ej y es i|Ds = hf, ei y (ej y es )i|Ds\{j} j

=

X

j

hf, (ej

∧ ei )

i

Ds

y es i|Ds = 0,

i,j

since ei ∧ ej = −ej

∧ ei

on performing the sum.

ˇ We denote the Cech cohomology spaces associated with this complex by k H (D; F). A key result that we now prove is, roughly speaking, that sheaves of functions defined without any constraints will have trivial cohomology spaces. A constraint here could mean that we consider functions that are constant or that satisfy some differential equation. More precisely, a sheaf F is defined to be a fine sheaf if every sufficiently smooth cutoff function η : D → R gives well-defined multiplication operators f 7→ ηf on each of the linear spaces F(D0 ). In particular, if supp η ⊂ D0 , then ηf is supposed to be extendable by zero to a function ηf ∈ F(D) on all D. When restricted to some D00 , this defines ηf ∈ F(D00 ).

Chapter 10. Hodge Decompositions

370

Proposition 10.5.5 (Cohomology of fine sheaves). If F is a fine sheaf on D, then H k (D; F) = {0} when k ≥ 1. For any sheaf F, the restriction map gives an invertible map F(D) → H 0 (D; F). Proof. First consider the second claim. If f ∈ C 0 (D; F) and ∂0 f = 0, then for all 1 ≤ i < j ≤ N , we have 0 = h∂0 f, e{i,j} i = hf, ej i|D{i,j} − hf, ei i|D{i,j} . Thus there is a unique function f ∈ F(D) such that f |Dk = hf, ek i. Since ∂−1 = 0, this proves the statement. Now let F be a fine sheaf, k ≥ 1, and f ∈ C k (D; F) with ∂k f =P 0. Pick a partition of unity {ηj }N subordinate to D, so that supp η ⊂ D and j j j=1 j ηj = 1 on D. Define a (k − 1)-cochain hg, et i :=

X

ηi hf, ei ∧ et i,

|t| = k.

i

Note that hf, ei ∧ et i is defined only on Dt ∩ Di , but that after multiplication by ηi , the product can be extended by zero across (∂Di ) ∩ Dt to all Dt . The anticommutation relation from Theorem 2.8.1 yields X XX h∂k−1 g, es i = hg, ej y es i|Ds = ηi hf, ei ∧ (ej y es )i j

=

j

XX j

Ds

i

ηi (δi,j hf, es i − hf, ej y (ei ∧ es )i)

i

Ds

X X X = ηi hf, es i − ηi hf, ej y (ei ∧ es )i|Di ∩Ds |Ds i

i

j

= hf, es i − 0 = hf, es i, where δi,j = 1 if i = j and otherwise 0. This shows that N(∂k ) = R(∂k−1 ), as desired. We finish this section with two algebraic techniques that are useful in studying complexes. By a diagram of maps being commutative, we mean that whenever we have two different compositions of maps A → D, A

f1

/B

f4

/D

f2

C then we have f3 ◦ f1 = f4 ◦ f2 .

f3

ˇ 10.5. Cech Cohomology

371

Lemma 10.5.6 (Snake lemma). Let (Uj , ∂ju ), (Vj , ∂jv ) and (Wj , ∂jw ) be complexes of linear spaces, and for each j, let fj

gj

0 → Uj → Vj → Wj → 0 be a short exact sequence such that ∂jv fj = fj+1 ∂ju and ∂jw gj = gj+1 ∂jv for all j. This hypothesis is summarized in the following commutative diagram with exact columns: 0 ···

···

0

u ∂j−3

/ Uj−2

v ∂j−3

u ∂j−2

/ Uj−1

v ∂j−2

fj−2

/ Vj−2

/ Vj−1

···

/ Wj−2

/ Uj

v ∂j−1

/ Vj

/ Wj−1

0

/ Uj+1

∂jv

/ Vj+1

0

/ Wj

∂jw

u ∂j+1

/ Uj+2

v ∂j+1

fj+1

gj w ∂j−1

0

∂ju

fj

gj−1 w ∂j−2

0

u ∂j−1

fj−1

gj−2 w ∂j−3

0

0

/

··· /

··· /

···

fj+2 v ∂j+2

/ Vj+2

gj+1

/ Wj+1

u ∂j+2

gj+2 w ∂j+1

w ∂j+2

/ Wj+2

0

0

˜ j : H j (W ) → H j+1 (U ) and induced linear Then there are connecting linear maps h j j ˜ maps fj : H (U ) → H (V ) and g˜j : H j (V ) → H j (W ), for all j, such that the cohomology sequence H j−2 C (U )

˜ j−3 h

f˜j−2

H

j−2

(V )

g ˜j−2

H

j−2

˜ j−2 h

H j−1 A (U ) f˜j−1

H

j−1

(V )

g ˜j−1

(W )

H

j−1

˜ j−1 h

H jB (U )

j

f˜j

H (V ) g ˜j

(W )

j

H (W )

˜j h

H j+1 B (U ) f˜j+1

H

j+1

(V )

g ˜j+1

H

j+1

˜ j+1 h

H j+2 A (U ) f˜j+2

H

j+2

g ˜j+2

(W )

C ˜ j+2 h

(V )

H j+2 (W )

is an exact complex. Exercise 10.5.7 (Diagram chasing). Prove the snake lemma through diagram chasing. To see an example of this, consider the definition of the connecting map ˜ j : H j (W ) = N(∂ w )/R(∂ w ) → H j+1 (U ) = N(∂ u )/R(∂ u ). Take w ∈ N(∂ w ). h j j−1 j+1 j j Surjectivity of gj gives v ∈ Vj , which maps to a v 0 ∈ Vj+1 . Commutativity ∂jw gj = gj+1 ∂jv shows that v 0 maps to 0 in Wj+1 , and so exactness at Vj+1 gives u ∈ Uj+1 such that fj+1 (u) = v 0 . Since v 0 maps to 0 in Vj+2 , commutativity

Chapter 10. Hodge Decompositions

372

v u fj+1 = fj+2 ∂j+1 ∂j+1 shows that u ∈ N(∂j+1 ), since fj+2 is injective.

u

/0

v

/ v0

/ 0

w

/ 0

˜ j ([w]) := [u]. Further diagram chasing shows The connecting map is defined as h u w ˜ j is a well-defined map between cohomolthat u ∈ R(∂j ) if w ∈ R(∂j−1 ), so that h ogy spaces. Through diagram chasing, one can similarly prove the following useful result. Lemma 10.5.8 (Five lemma). Consider the following commutative diagram of linear spaces and linear maps, where the two rows form complexes: U1

f1

h1

V1

/ U2

f2

h2

g1

/ V2

/ U3

f3

h3

g2

/ V3

/ U4

f4

h4

g3

/ V4

/ U5 h5

g4

/ V5

(i) If the row complexes are exact at U3 and V2 , and if h2 , h4 are injective and h1 is surjective, then the middle map h3 is injective. (ii) If the row complexes are exact at V3 and U4 , and if h5 is injective and h2 , h4 are surjective, then the middle map h3 is surjective. In particular, if the row complexes are exact, and if h1 , h2 , h4 , and h5 are invertible, then h3 is also invertible.

10.6

De Rham Cohomology

Let D be a bounded Lipschitz domain in an n-dimensional Euclidean space X, and consider the operators d and δ = −d∗ in L2 (D). In this section, we study the finite-dimensional subspace Ck (D) in the Hodge decomposition L2 (D) = R(d) ⊕ Ck (D) ⊕ R(δ). Definition 10.6.1 (De Rham cohomology spaces). Let D be a bounded Lipschitz domain. The De Rham cohomology spaces are the quotient spaces H k (D) := N(d; ∧k )/R(d; ∧k ). We identify their direct sum with H(D) := N(d)/R(d).

10.6. De Rham Cohomology

373

ˇ These cohomology spaces H k (D) should not be confused with the Cech cohomology spaces H k (D; F) from Section 10.5. We shall, however, show in this section that for the sheaf F = R they are indeed closely related. We note that the following spaces essentially are the same. The last three are indeed equal. • The De Rham space H(D). • The dual quotient space N(δ)/R(δ). • The intersection of the Hodge subspaces Ck (D) = N(d) ∩ N(δ). • The null space of the Hodge–Dirac operator d + δ from Proposition 10.1.3. • The null space of the Hodge–Laplace operator (d + δ)2 = dδ + δd. Note that the orthogonal complement Ck (D) of R(d) in N(d) is different from, but can be identified with the quotient space H(D). One can show that for smooth domains, Ck (D) is a subspace of C ∞ (D). For Lipschitz domains this is not true, but using the potential maps from Section 10.4, one can show, at least for strongly Lipschitz domains, that there is another complement of R(d) in N(d) that is contained in C ∞ (D). This means that the de Rham cohomology space H(D) can be represented by C ∞ (D) fields. The cohomology space Ck (D) splits into its homogeneous k-vector parts Ck (D; ∧k ), and it is our aim in this section to calculate the Betti numbers bk (D) = dim Ck (D; ∧k ) = dim H k (D) from Definition 7.6.3 for a given domain D. On a domain D with boundary, we can similarly consider the cohomology space C⊥ (D) with normal boundary conditions. But by Hodge star duality as in Proposition 10.2.2, we have dim C⊥ (D; ∧k ) = dim Ck (D; ∧n−k ) = bn−k (D). For the remainder of this section, we therefore consider only tangential boundary conditions. The following observation shows that the Betti numbers do not depend on the geometry, but only on the topology of the domain. Proposition 10.6.2 (Topological invariance). Let ρ : D1 → D2 be a Lipschitz diffeomorphism between bounded Lipschitz domains in Euclidean space. Then the pullback induces an invertible linear map ρ∗ : H(D2 ) → H(D1 ). In particular, bk (D1 ) = bk (D2 ) for all k = 0, 1, 2, . . . , n. Note that the pullback will not in general map between R(d)⊥ = N(δ) for the two domains, and that consequently the spaces Ck (D, ∧k ) depend on the geometry of D. It is only the dimensions that are topological invariants.

374

Chapter 10. Hodge Decompositions

Proof. The result is immediate from Lemma 10.2.6, which shows that ρ∗ yields invertible maps between the ranges R(d) as well as between the null spaces N(d) for the two domains. The Betti numbers give information about the topology of the domain, the simplest example being b0 (D) = number of connected components of D, which is clear, since dF = 0 for a scalar function means that F locally constant. Note also that by imposing Dirichlet boundary conditions dF = 0 for a scalar function forces F = 0. By Hodge star duality, this translates to bn (D) = 0 for every bounded Lipschitz domain in Euclidean space. The heuristic is that for general k, the Betti number bk (D) measures the number of k-dimensional obstructions in D. For the topologically trivial domains, those that are Lipschitz diffeomorphic to the ball, the following result is clear from Poincar´e’s theorem (Theorem 7.5.2) and its L2 extension in Section 10.4. Proposition 10.6.3 (Cohomology for balls). If D is Lipschitz diffeomorphic to a ball, then b0 (D) = 1 and bk (D) = 0 for k = 1, 2, . . . , n. To calculate the Betti numbers for a general domain, we cover D by subsets Dj , all diffeomorphic to balls, such that D = D1 ∪ · · · ∪ DN . We also require that all intersections be topologically trivial as follows. Definition 10.6.4 (Good cover). Let D be a bounded Lipschitz domain, and assume that D = D1 ∪ · · · ∪ DN is a finite cover of D by open subsets. We say that D = {Dj }N j=1 is a good cover of D if all nonempty intersections Ds = Ds1 ∩· · ·∩Dsk are Lipschitz diffeomorphic to balls. We use the algebra from Section 10.5 and three simple examples of sheaves F on D. • The sheaf D(d; ∧k ), where the Hilbert space associated with an open set D0 ⊂ D consists of k-vector fields F ∈ L2 (D0 ) such that dF ∈ L2 (D0 ). • The sheaf N(d; ∧k ), where the Hilbert space associated with an open set D0 ⊂ D consists of k-vector fields F ∈ L2 (D0 ) such that dF = 0. • The sheaf R, where a real line is associated with an open set D0 ⊂ D, and restriction is the identity map.

375

10.6. De Rham Cohomology

Note that D(d; ∧k ) is a fine sheaf, but not N(d; ∧k ) or R. Note also that we cannot use δ here to define sheaves, since restriction does not preserve the boundary conditions. The main result of this section is the following characterization of the Betti numbers. ˇ Theorem 10.6.5 (De Rham = Cech). Let D be a bounded Lipschitz domain in Euclidean space, and let D be a good cover as in Definition 10.6.4. Then bk (D) = dim H k (D; R),

k = 0, 1, . . . , n,

(10.13)

ˇ where H k (D; R) is the Cech cohomology space for the constant sheaf R. This shows in particular that the Betti numbers do not depend on the exterior and interior derivative operators, since the right-hand side in (10.13) does not. Conversely, dim H k (D; R) does not depend on the choice of good cover, since the left-hand side in (10.13) does not. Proof. Consider the following sequence of maps on Ds : i

d

0 → N(dDs ; ∧k ) → D(dDs ; ∧k ) → N(dDs ; ∧k+1 ) → 0, where i denotes inclusion. Assuming that the intersection Ds is Lipschitz diffeomorphic to a ball, we know that this is an exact complex for k ≥ 0. Acting componentwise, this induces an exact complex i

d

0 → C j (D; N(d; ∧k )) → C j (D; D(d; ∧k )) → C j (D; N(d; ∧k+1 )) → 0. Consider the following commutative diagram: 0

···

···

···

0

∂j−2

/ C j−1 (D; N(d; ∧k ))

∂j−2

∂j−2

/

∂j−1

/ C j (D; N(d; ∧k ))

∂j−1

i

/ C j−1 (D; D(d; ∧k ))

0

∂j−1

/

∂j

/ C j+1 (D; N(d; ∧k ))

∂j

i

/ C j (D; D(d; ∧k ))

d

C j−1 (D; N(d; ∧k+1 ))

0

0

∂j

/ C j+1 (D; D(d; ∧k )) /

/

···

∂j+1

/

···

∂j+1

/

···

i

d

C j (D; N(d; ∧k+1 ))

∂j+1

d

C j+1 (D; N(d; ∧k+1 ))

0

According to Proposition 10.5.5, the cohomology spaces H j (D; D(d; ∧k )) for the second row vanish when j ≥ 1. The exact cohomology complex provided by the snake lemma (Lemma 10.5.6), thus splits into exact sequences 0 → H j (D; N(d; ∧k+1 )) → H j+1 (D; N(d; ∧k )) → 0

376

Chapter 10. Hodge Decompositions

for k ≥ 0, j ≥ 1, and for j = 0 the exact complex 0 → N(d; ∧k ) → D(d; ∧k ) → N(d; ∧k+1 ) → H 1 (N(d; ∧k )) → 0. This shows that dim H j+1 (D; N(d; ∧k )) = dim H j (D; N(d; ∧k+1 )) for j ≥ 1, and dim H 1 (N(d; ∧k )) = dim(N(d; ∧k+1 )/dD(d; ∧k )) = bk+1 (D). Thus for k ≥ 1, we get dim H k (D; R) = dim H k (D; N(d; ∧0 )) = dim H k−1 (D; N(d; ∧1 )) = dim H k−2 (D; N(d; ∧2 )) = · · · = dim H 1 (D; N(d; ∧k−1 )) = bk (D), since the sheaves N(d; ∧0 ) and R coincide. Proposition 10.5.5 shows that H 0 (D; R) = R(D), which equals the space Ck (D; ∧0 ) of locally constant functions on D. Thus dim H 0 (D; R) = b0 (D), which completes the proof. Theorem 10.6.5 reduces the computation of the Betti numbers to a finite problem, although the construction of a good cover can be nontrivial. Note that we started by defining the Betti numbers as the dimension of the finite-dimensional space N(d; ∧k )/R(d; ∧k ). However, note that both the numerator and denominator are infinite-dimensional Hilbert spaces in general. On the other hand, we have now characterized the Betti numbers as the dimensions of the spaces N(∂k ; C k (D; R))/R(∂k−1 ; C k (D; R)), where all spaces involved are finite-dimensional. Example 10.6.6 (Annulus). The simplest domain with nontrivial topology is the two-dimensional annulus D = {x = (x1 , x2 ) ; r < |x| < R}. We see that a good cover of D requires three subsets D1 , D2 , D3 . For example, D1 := {x ∈ D ; x2 > 0}, D2 := {x ∈ D ; x1 > x2 }, and D3 := {x ∈ D ; x1 + x2 < 0} give a good cover. The nonempty intersections are D1 , D2 , D3 , D12 , D13 , and D23 . We see that C 0 (D; R) is a three-dimensional space, with a basis (ω1 , ω2 , ω3 ), where hωi , ei i = 1, and 0 on the other subsets. Similarly, C 1 (D; R) is a three-dimensional space, with a basis (ω12 , ω13 , ω23 ), where hωij , eij i = 1, and 0 on the other subsets. By Definition 10.5.1, the matrix for ∂0 is −1 1 0 −1 0 1 . 0 −1 1 This has a one-dimensional null space, so that b0 (D) = dim N(∂0 ) = 1. Since ∂0 has two-dimensional range and ∂1 = 0, since D123 = ∅, we get b1 (D) = dim N(∂1 ) − dim R(∂0 ) = 3 − 2 = 1. We have shown that the Betti numbers for D are bi (D) = (1, 1, 0).

377

10.6. De Rham Cohomology

Figure 10.2: (a) Vector field F to be decomposed. (b) The cohomology part H1 of F in the two-dimensional space Ck (D). By Exercise 10.6.8, the cohomology part of any F is a linear combination of this H1 and H2 appearing in Figure 10.3. (c) The curl-free part ∇U of F . (d) The divergence-free part ∇ y (V j) of F , with tangential boundary conditions.

Exercise 10.6.7 (3D spherical shell). Show by constructing a good cover of the three dimensional spherical shell D := {x = (x1 , x2 , x3 ) ; r < |x| < R} with four subsets that bi (D) = (1, 0, 1, 0). Exercise 10.6.8 (General plane domain). For a plane domain consisting of m disks with ni smaller interior disks removed in their respective disks, i = 1, . . . , m,

378

Chapter 10. Hodge Decompositions

Figure 10.3: (a) Vector field F to be decomposed. (b) The cohomology part H2 of F in the two-dimensional space Ck (D). By Exercise 10.6.8, the cohomology part of any F is a linear combination of this H2 and H1 appearing in Figure 10.2. (c) The curl-free part ∇U of F . (d) The divergence-free part ∇ y (V j) of F , with tangential boundary conditions. construct a good cover and show that the Betti numbers are m X bi (D) = m, ni , 0 . i=1

The case m = 1, n = n1 = 2 is illustrated in Figures 10.2 and 10.3. The disadvantage with Theorem 10.6.5 is that the construction of good covers soon gets complicated in higher dimensions. We therefore discuss two complemen-

379

10.6. De Rham Cohomology

tary techniques for computing Betti numbers: The Mayer–Vietoris sequence and unneth formula. These enable us to calculate Betti numbers for unions and the K¨ Cartesian products of domains. Theorem 10.6.9 (Mayer–Vietoris sequence). Let D1 , D2 be bounded Lipschitz domains such that D1 ∪ D2 and D1 ∩ D2 also are Lipschitz domains. Then we have the following exact complex, the Mayer–Vietoris sequences: H k−17 (D1 ∪ D2 )

H k (D : 1 ∪ D2 )

H k−1 (D1 ) ⊕ H k−1 (D2 )

H k+1 (D : 1 ∪ D2 )

H k (D1 ) ⊕ H k (D2 )

H k+1 (D1 ) ⊕ H k+1 (D2 )

H k−1 (D1 ∩ D2 )

7

H k (D1 ∩ D2 )

H k+1 (D1 ∩ D2 )

Proof. Consider the complex j

i

0 → D(dD1 ∪D2 ; ∧k ) → D(dD1 ; ∧k ) ⊕ D(dD2 ; ∧k ) → D(dD1 ∩D2 ; ∧k ) → 0, where i(f ) := (f |D1 , f |D2 ) is restriction, and j is the map j(g1 , g2 ) := g1 |D1 ∩D2 − g2 |D1 ∩D2 . We see that i is injective and that the sequence is exact at the middle space. To verify that j is surjective, take h ∈ D(dD1 ∩D2 ; ∧k ). Proposition 10.2.7 shows that we can extend h to g ∈ D(dD1 ; ∧k ) such that g|D1 ∩D2 = h. Then j(g, 0) = h. The stated Mayer–Vietoris sequence is obtained by applying the snake lemma (Lemma 10.5.6) with ∂ = d to this complex. Theorem 10.6.10 (K¨ unneth formula). Let D1 and D2 be bounded Lipschitz domains in Euclidean spaces X1 and X2 respectively, not necessarily of the same dimension. Then the Betti numbers of the Cartesian product D1 × D2 ⊂ X1 × X2 are given by the K¨ unneth formula bk (D1 × D2 ) =

k X

bj (D1 ) bk−j (D2 ).

j=0

Proof of Theorem 10.6.10. Let pi : D1 × D2 → Di , i = 1, 2, denote the coordinate projections and use pullbacks to form the bilinear map (f1 (x1 ), f2 (x2 )) 7→ (p∗1 (f1 ) ∧ p∗2 (f2 ))(x1 , x2 ) = f1 (x1 ) ∧ f2 (x2 ), where fi : Di → ∧ki Vi . Note that d(p∗1 f1 ∧ p∗2 f2 ) = p∗1 (df1 ) ∧ p∗2 f2 + (−1)k1 p∗1 f1 ∧ p∗2 (df2 ). This shows that p∗1 f1 ∧ p∗2 f2 is closed if f1 and f2 are closed, and that it is exact if one of the factors is exact and the other factor is closed. Using the universal property for tensor products, we obtain a well-defined linear map H k1 (D1 ) ⊗ H k2 (D2 ) → H k1 +k2 (D1 ×D2 ) : [f1 (x1 )] ⊗ [f2 (x2 )] 7→ [f1 (x1 ) ∧ f2 (x2 )].

380

Chapter 10. Hodge Decompositions

Assume that Di have a good cover by Ni subsets, i = 1, 2. We prove the K¨ unneth formula by induction on N1 + N2 , the induction hypothesis being that the map M (H i (D1 ) ⊗ H k−i (D2 )) → H k (D1 × D2 ), i

defined as the direct sum of the maps above, is invertible for all fixed k. Evaluating the dimensions will then give the K¨ unneth formula. If N1 + N2 = 2, then D1 , D2 , and D1 × D2 are all diffeomorphic to balls in the respective spaces, and the induction hypothesis is readily verified. For the induction step, write, for example, D1 = U ∪ V , where U is one of the sets in a good cover of D1 and V is the union of the rest. The induction hypothesis applies to U × D2 , V × D2 , and (U ∩ V ) × D2 , since both V and U ∩ V have good covers by at most N1 − 1 sets. Consider the diagram / H k−1 (U × D2 ) ⊕ H k−1 (V × D2 ) (H i−1 (U ) ⊗ H k−i (D2 )) ⊕ (H i−1 (V ) ⊗ H k−i (D2 ))

/ H k−1 ((U ∩ V ) × D2 )

/ H k ((U ∪ V ) × D2 )

H i−1 (U ∩ V ) ⊗ H k−i (D2 )

H i (U ∪ V ) ⊗ H k−i (D2 )

(H i (U ) ⊗ H k−i (D2 )) ⊕ (H i (V ) ⊗ H k−i (D2 ))

/ H k (U × D2 ) ⊕ H k (V

H i (U ∩ V ) ⊗ H k−i (D2 )

× D2 )

/ H k ((U ∩ V ) × D2 ).

The horizontal maps are defined as above in the natural way, and the vertical maps come from the Mayer–Vietoris sequence. On the left, we have taken the tensor product by H k−i (D2 ), and the maps act trivially in this factor. For the first and fourth rows we have a distributional rule for tensor products and direct sums. It is straightforward to verify that this diagram commutes. Taking the direct sum over i, we can apply the induction hypothesis and the five lemma (Lemma 10.5.8), which proves the theorem. Example 10.6.11 (Rn annulus). Consider an n-dimensional annulus An = {x = (x1 , . . . , xn ) ; 2 < |x| < 3}, and let An+ := An ∩ {xn > −1} and An− := An ∩ {xn < 1}. Then An± are both diffeomorphic to the n-dimensional unit ball B n , and An+ ∩ An− is diffeomorphic to An−1 × (0, 1). We prove by induction that bi (An ) = (1, 0, 0, . . . , 0, 0, 1, 0), so that bn−1 (An ) = 1. The cases n = 2, 3 are known from Example 10.6.6 and Exercise 10.6.7. For n > 3, the Mayer–Vietoris sequence and K¨ unneth formula give exact sequences 0 → H k−1 (An−1 ) → H k (An ) → 0

10.7. Comments and References

381

for k ≥ 2 and using that b1 (An ) = 1, we have for k = 1 that 0 → R → R2 → R → H1 (An ) → 0 is exact. This proves the stated formula for the Betti numbers. Note that x/|x|n is a divergence and curl-free vector field that is normal on the boundary of the annulus. Applying the Hodge star map, it follows that Ck (An ; ∧n−1 ) is spanned by the tangential (n − 1)-vector field ∗x/|x|n . Exercise 10.6.12 (3D cohomology). Let D be a three-dimensional ball with m smaller balls and n cylinders removed, all of these disjoint. Show by inductively applying a Mayer–Vietoris sequence that bi (D) = (1, n, m, 0).

10.7

Comments and References

10.1 The formulation of Hodge decompositions presented here is based on the survey paper by Axelsson and McIntosh [14]. The notation Γ for an abstract nilpotent operator, generalizing d, was introduced there, along with Π = Γ ± Γ∗ for an abstract Hodge–Dirac operator. Playing with symbols, Γ∗ can be viewed as a backward Γ, and together they combine to a Π. As in Sections 9.3–9.6, a main philosophy in this chapter is to handle Hodge decomposition by first-order operators. Concretely, this means that we study Hodge decompositions as far as possible using Γ, Γ∗ , and Π, which in applications are always first-order differential operators, rather than involving the abstract Laplace operator Π2 . The latter is standard in the literature, but may sometimes complicate the problems. A concrete example is the proof of Theorem 10.3.1. The reason that this works is that we work with the firstorder operator Γ = d for which the commutation theorem is available. A reference for the classical variational second-order approach to Hodge decompositions is Morrey [72]. 10.2 We have two natural choices of boundary conditions leading to skew-adjoint Hodge–Dirac operators d + δ and d + δ respectively. It is standard to consider the second-order Laplace operators. Here the generalized Dirichlet Laplacian (d + δ)2 = dδ + δd is said to have relative boundary conditions, and the generalized Neumann Laplacian (d + δ)2 = dδ + δd is said to have absolute boundary conditions. See, for example, [91]. For geometrical reasons, we use the terminology normal and tangential, rather than relative and absolute, in this book. 10.3 The beautiful and simple reduction from Lipschitz domains to smooth domains using the commutation theorem in Theorem 10.3.1 appears in the work by R. Picard [74].

382

Chapter 10. Hodge Decompositions For strongly Lipschitz domains there is a singular integral proof, based on Theorems 8.3.2 and 9.5.1, showing that D(d+δ ) and D(d+δ) are contained in H 1/2 (D). This exponent 1/2 is sharp for the class of strongly Lipschitz domains. See [14].

10.4 These regularity and support properties of potentials where proved by Mitrea, Mitrea, and Monniaux [70] and Costabel and McIntosh [31]. In fact, [31] ∗ shows the stronger result that the constructed operators TD and TD are pseudodifferential operators of order −1. By well-known estimates for such operators, bounds on a large number of scales of function spaces follow immediately. Rather than relying on the theory of pseudodifferential operators, however, we follow the proof from [70]. ˇ 10.5-10.6 Two references for Cech cohomology theory are Bott and Tu [21] and Fulton [40]. The author’s original inspiration for this section comes from lecture notes by Ben Andrews at the Australian National University. The snake and five lemmas are usually used in an algebraic context. However, Pryde [76] shows that such techniques also are useful in the context of Fredholm operators on Banach spaces in analysis. Two references for de Rham cohomology are [21] and Madsen and Tornehave [65].

Chapter 11

Multivector and Spinor Bundles Prerequisites: The reader should be familiar with the basic ideas of differential geometry. Section 11.1 gives a short survey of the required material and fixes notation. Section 11.2 builds on Chapter 7, and Section 11.6 builds on Chapter 5. Section 11.4 uses Section 4.5. The material from Section 10.5 is used in Section 11.6 and some in Section 11.2. Road map: In many situations the appropriate model of space is not that of an affine space, but rather that of a manifold as defined in Section 6.1. A manifold behaves only locally like affine space, whereas the global behavior can be quite different. So far we have considered only manifolds embedded as k-dimensional surfaces in some surrounding affine space. A general abstract C ∞ manifold M can always be embedded in some finite-dimensional affine space X. However, when the geometry of the manifold is prescribed by a Riemannian metric, most manifolds do not appear as k-surfaces embedded in some affine space. In the remaining chapters we consider general compact Riemannian manifolds M , from an intrinsic point of view. We also change focus: instead of being mainly concerned with local nonsmooth analysis as in the previous two chapters, we now focus on global smooth analysis in the remaining two chapters. Extending the concepts of multivectors and spinors to such manifolds presents new problems, although their relation is fairly straightforward after our having developed the affine theory thoroughly in the previous chapters. The first step is to construct the space Tp M of tangent vectors at a point p ∈ M . These form a family of vector spaces {Tp M }p∈M indexed by M , that is, a vector bundle. In contrast to the case in which M is © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_11

383

384

Chapter 11. Multivector and Spinor Bundles

embedded as a k-surface in an affine space (X, V ), in which case all tangent spaces Tp M are subspaces of V , in general two tangent spaces Tp M and Tq M , p 6= q, are incomparable. As a consequence, some care is needed to define directional derivatives, referred to as covariant derivatives in this context, of tangential vector fields along M . It is straightforward to construct a bundle of multivectors {∧(Tp M )}p∈M and to extend the multivector calculus from Chapter 7, which we do in Section 11.2. Particularly useful are the pullback and pushforward operations from Section 7.2, which allow us to pass between different charts on the manifold and give an invariant meaning to multivectors on the manifold. For spinors we also have induced maps from Proposition 5.3.5. However, these are defined only up to a sign, which makes the definition and even the existence of spinor bundles {4(T / p M )}p∈M a ˇ Z2 cohomology in Section 11.6 to investigate when delicate matter. We use Cech there are topological obstructions for spinor bundles to exist, and if they exist, how many different such spinor bundles there are globally over M . The most central concept for Riemannian manifolds is curvature, which in some sense measures how much the manifold locally differs from flat Euclidean space. The curvature operator in vector notation can be a confusing beast. Having access to multivectors, though, we show in Section 11.3 that this is a symmetric operator on bivectors ∧2 (Tp M ) → ∧2 (Tp M ) : b 7→ R(b). The input bivector b represents the oriented measure of an infinitesimal 2-surface Σ at p, and the output bivector R(b) represents, as in Proposition 4.2.3, an infinitesimal rotation determined by the total variation that a vector undergoes as we move it around ∂Σ, keeping it constant as much as the curved manifold allows. / acting on mulIn Sections 11.5 and 11.6, we define Dirac operators D and D tivector and spinor fields respectively. In flat Euclidean space we saw in Section 9.1 that both these Dirac operators are first-order partial differential operators that are square roots of the Laplace operator ∆. This is no longer true on a curved man2 / differ from ∆ by a zeroth-order term determined ifold, but the squares D2 and D by the curvature R of the manifold. We prove such Weitzenb¨ock identities D2 = ∆ + R in preparation for Chapter 12, although there they have many other applications to geometry. We have seen an analogue for bounded domains in Euclidean space in Theorem 10.3.6, where the curvature of the boundary appears rather than the curvature of space itself. Section 11.4 contains a proof of Liouville’s theorem on conformal maps in dimension n ≥ 3. This completes Theorem 4.5.12, which is otherwise not used elsewhere.

11.1. Tangent Vectors and Derivatives

385

Highlights: • Curvature as infinitesimal rotations around infinitesimal bivectors:11.3.2 • Liouville’s theorem on conformal maps: 11.4.2 • 4M and 4M / Weitzenb¨ock identities: 11.5.9 and 11.6.10 ˇ • Cech cohomology of spinor bundles: 11.6.3 and 11.6.5

11.1

Tangent Vectors and Derivatives

We define general vector bundles E following standard terminology, although from the discussion in Section 1.1 we keep in mind that we in general are considering bundles of linear spaces. Let M be a manifold, and let L be an N -dimensional linear space, over R or possibly C. When considering nonembedded manifolds and vector bundles over them, we shall assume C ∞ regularity. space L, is a collection of linear spaces • A vector bundle E over M , with fiber S {Ep }p∈M together with a cover M = α∈I Uα of M by open sets Uα ⊂ M , and linear invertible maps µα (p) ∈ L(L; Ep ), p ∈ Uα , α ∈ I, such that each bundle transition map µβα (p) := (µβ (p))−1 µα (p) ∈ L(L) is a C ∞ function of p ∈ Uβα := Uβ ∩ Uα for all α, β ∈ I. The linear space Ep is referred to as the fiber of E over p ∈ M . We refer to µα as bundle charts and to {µα }α∈I as a bundle atlas. More generally, a bundle chart is a family µ(p) ∈ L(L; Ep ) of linear invertible maps defined in open sets U such that the bundle transition maps (µ(p))−1 µα (p) ∈ L(L) are C ∞ functions of p ∈ U ∩ Uα for all α ∈ I. • A section w of E is a map that, to each p ∈ M associates a vector w(p) ∈ Ep . We let C ∞ (M ; E) be the space of all sections w such that µ(p)−1 (w(p)) ∈ L is a C ∞ -regular function of p ∈ U for every bundle chart µ. Spaces of sections like C k (M ; E) and L2 (M ; E) are defined similarly. • A (local) frame for E in an open set U ⊂ M is a set of sections ei ∈ C ∞ (U ; E), i = 1, . . . , N , such that {ei (p)} is a basis for Ep for each p ∈ U . PN Thus, if w is a section of E in U , then w(p) = i=1 wi (p)ei (p), where wi are the scalar coordinate functions for w in the frame {ei }. Note that in general there may not exist a frame defined on all M .

Chapter 11. Multivector and Spinor Bundles

386

• A metric g on E is a family of inner products g(·, ·) = h·, ·ip , one on each fiber Ep , p ∈ M . It is assumed that the symmetric nondegenerate bilinear forms L × L → R : (u, v) 7→ hµ(p)u, µ(p)vip are C ∞ functions of p ∈ U for all bundle charts µ. Let {ei }N i=1 be a frame in U ⊂ M . The metric coordinates in this frame are the functions gij (p) := hei (p), ej (p)ip ,

p ∈ U, i, j = 1, . . . , N.

A vector bundle equipped with a metric is called a metric vector bundle. We use only Euclidean metrics on real vector bundles and only Hermitian metrics on complex vector bundles. The metric coordinates for the constant standard inner product on Rn are the Kronecker delta ( 1, i = j, δij := 0, i 6= j. Example 11.1.1 (E ∗ = E identification). If E is a vector bundle over M , with bundle charts µ(p) : L → Ep , then the dual vector bundle E ∗ is defined to be the bundle {Ep∗ }p∈M of dual spaces, with fiber space L∗ and bundle charts (µ(p)−1 )∗ : L∗ → Ep∗ . When a vector bundle E is equipped with a metric, the dual bundle E ∗ and E can be identified by applying Proposition 1.2.3 to each fiber. However, some care has to be taken with regard to charts. We now have twins of bundle charts µ(p) : L → Ep ,

p ∈ U,

and (µ(p)−1 )∗ : L∗ → Ep∗ , Ep∗

p ∈ U,

∗

for = Ep . That L 6= L is a technicality; what is important is that these spaces do not depend on p. Fixing an auxiliary inner product on L, we have L∗ = L as the fiber space for E ∗ = E. This gives a transition map µ(p)∗ µ(p) : L → L∗ between the twin bundle charts. The most fundamental vector bundle is the tangent vector bundle. To define the tangent space Tp M to a nonembedded manifold M at a point p ∈ M , we proceed as follows. We use a C 1 curve γ : (−, ) → M with γ(0) = p and > 0 to define a tangent vector v. If µα is any chart for M around p, then t 7→ µ−1 α (γ(t)) yields a curve through x in Rn , where µα (x) = p. We define two curves γ1 and γ2 0 −1 0 n through p to be equivalent, γ1 ∼ γ2 , if (µ−1 α ◦ γ1 ) (0) = (µα ◦ γ2 ) (0) in R . From the chain rule, we see that this equivalence relation is independent of the choice of chart µα . A tangent vector v at p ∈ M is defined to be an equivalence class [γ] of curves γ through p on M , under the equivalence relation ∼.

11.1. Tangent Vectors and Derivatives

387

Definition 11.1.2 (Tangent bundle). Let (M, {µα }α∈I ) be an n-dimensional manifold, and let p ∈ M . We define the tangent space Tp M to be the set of equivalence classes v = [γ] of C 1 curves γ through p on M as above. Defining the bijective map −1 (11.1) Tp M → Rn : v = [γ] 7→ (µα ◦ γ)0 (0) to be linear gives Tp M the structure of an n-dimensional linear space. The tangent bundle T M is the vector bundle {Tp M }p∈M , with fiber space Rn . Each manifold chart µα : Dα → Mα induces a bundle chart µα (p) ∈ L(Rn ; Tp M ), p ∈ Mα , being the inverse of the map (11.1). We say that M is a Riemannian manifold if T M is equipped with a metric. We refer to sections of T M as tangent vector fields. From the standard basis {ei } for Rn , we define the coordinate frame {µα (ei )} in Mα . Note that for T M , the bundle transition functions µβα (p) : Rn → Rn equal the total derivatives µβα : Rn → Rn , p = µα (x), of the manifold transition x functions µβα . Definition 11.1.3 (Directional derivatives). Let f ∈ C 1 (M ; R) be a scalar-valued function on a manifold M , and let v ∈ Tp M be a tangent vector at p ∈ M . We define the tangential directional derivative (∂v f )(p) := (f ◦ γ)0 (0) if v = [γ]. In a chart µα around p = µα (x), let v = µα (p)u, u ∈ Rn . Then the chain rule shows that ∂v f (p) = ∂u fα (x), so ∂v f (p) is independent of the choice of curve γ representing v. Next consider a vector bundle E over M and a section w ∈ C ∞ (M ; E). We ask whether there is a well-defined tangential directional derivative ∂v w(p). Proceeding similarly to the above, we consider wα (p) := µα (p)−1 w(p) ∈ L and wβ (p) := µβ (p)−1 w(p) ∈ L. In this case, wβ (p) = µβα (p)(wα (p)). However, because of the p-dependence of the linear maps µβα (p), the derivatives ∂v wα (p) and ∂v wβ (p) will in general differ by a zeroth-order term according to the product rule. More precisely, (∂v wβ )(p) = µβα (p)(∂v wα )(p) + (∂v µβα (p))wα (p), where the directional derivative is taken componentwise. Directional derivatives of sections of vector bundles in differential geometry are traditionally called covariant derivatives, denoted by ∇v rather than ∂v , and amount to a choice of zeroth-order terms, consistent with the transition maps as above.

388

Chapter 11. Multivector and Spinor Bundles

Definition 11.1.4 (Covariant derivative). Let E be a vector bundle over a manifold M . A covariant derivative on E is a map C ∞ (U ; T M ) × C ∞ (U ; E) 7→ C ∞ (U ; E) : (v, w) 7→ ∇v w such that • ∇v1 +v2 w = ∇v1 w + ∇v2 w, for all v1 , v2 ∈ C ∞ (U ; T M ) and w ∈ C ∞ (U ; E), • ∇f v w = f (∇v w), for all f ∈ C ∞ (U ; R), v ∈ C ∞ (U ; T M ) and w ∈ C ∞ (U ; E), • ∇v (w1 + w2 ) = ∇v w1 + ∇v w2 , for all v ∈ C ∞ (U ; T M ) and w1 , w2 ∈ C ∞ (U ; E), • ∇v (f w) = f (∇v w) + (∂v f )w, for all f ∈ C ∞ (U ; R), v ∈ C ∞ (U ; T M ), and w ∈ C ∞ (U ; E). N Let E be a vector bundle over a Riemannian manifold M , and let {ei (p)}i=1 be a frame P for E in an open subset U ⊂ M . By the Leibniz rule, a section w(p) = i fi (p)ei (p) of E in U has covariant derivative

∇v w =

X X (∂v fi )ei + fi (∇v ei ). i

i

Since v 7→ ∇v w is linear, this uniquely defines vector fields wji ∈ C ∞ (U ; T M ) in U such that ∇v ei =

N X hωji , viej ,

v ∈ C ∞ (U ; T M ), i = 1, . . . , N.

j=1

This shows in particular that ∇v w is well defined at p ∈ M whenever v ∈ Tp M is given and w is a section differentiable at p. Definition 11.1.5 (Christoffel symbols). Let E be an N -dimensional vector bundle N over a manifold M , with a covariant derivative ∇v . Let e = {ei (p)}i=1 be a frame 2 for E in an open subset U ⊂ M . Then the N vector fields wij ∈ C ∞ (U ; T M ) specifying ∇v in the frame e are called the Christoffel symbols of ∇ in e. We write ωe ∈ C ∞ (U ; L(T M, L(E))) for the section such that the matrix for the map ωe (v) ∈ L(E) in the basis {ei } is {hωij , vi}ij . With this notation, we have X X X (∂v fi )ei + ωe (v) ∇v fi ei = fi ei i

i

i

in the frame e = {ei (p)}N i=1 . To single out a canonical covariant derivative, we demand that it be compatible with natural structures of the vector bundle. In

11.1. Tangent Vectors and Derivatives

389

particular, if E is equipped with a metric {h·, ·ip }p∈M , then we demand that the product rule ∂v hw1 , w2 i = h∇v w1 , w2 i + hw1 , ∇v w2 i hold for all w1 , w2 ∈ C ∞ (U ; E) and v ∈ C ∞ (U ; T M ) and say that ∇v is a metric covariant derivative. If ωij are the Christoffel symbols in an ON-basis, this happens if and only if ωji = −ωij holds for all i, j = 1, . . . , N . To obtain the unique existence of a covariant derivative on the tangent bundle T M , one imposes, besides it being metric, a condition in terms of Lie brackets of vector fields. Exercise 11.1.6 (Lie brackets on manifolds). Let µα : Dα → M and µα : Dβ → M be two charts on a manifold M . Push forward two vector fields u and v in Dα ⊂ Rn , by the transition map ρ := µβα , to u0 := ρ∗ u and v 0 := ρ∗ v in Dβ . Consider the Lie brackets [u, v] := ∂u v − ∂v u and [u0 , v 0 ] := ∂u0 v 0 − ∂v0 u0 . Show that [u0 , v 0 ] = ρ∗ [u, v], and deduce that the Lie bracket of tangent vector fields on M is well defined. Proposition 11.1.7 (Levi-Civita covariant derivative). Let M be a Riemannian manifold. Then there exists a unique covariant derivative on T M , the Levi-Civita covariant derivative, which is both metric and torsion-free, in the sense that ∂v hv1 , v2 i = h∇v v1 , v2 i + hv1 , ∇v v2 i, ∇v1 v2 − ∇v2 v1 = [v1 , v2 ], for all v, v1 , v2 ∈ C ∞ (M ; T M ). Proof. We use an ON-frame {ei (p)}N i=1 in some open set U ⊂ M . If a covariant derivative is metric and torsion-free, then its Christoffel symbols satisfy ωij = −ωji and hωkj , ei i − hωki , ej i = h[ei , ej ], ek i, for all i, j, k. Permuting (i, j, k) to (j, k, i), (k, i, j) and subtracting and adding these two obtained equations gives the formula 2hωkj , ei i = h[ei , ej ], ek i + h[ek , ei ], ej i − h[ej , ek ], ei i. Conversely, the covariant derivative defined through this formula is seen to be metric and torsion-free.

Chapter 11. Multivector and Spinor Bundles

390

Exercise 11.1.8. Generalize the above argument, and show that in a general frame {ei } for T M over a Riemannian manifold, the Levi-Civita covariant derivative is given by 2h∇ei ej , ek i = ∂ei hej , ek i + ∂ej hek , ei i − ∂ek hei , ej i + h[ei , ej ], ek i + h[ek , ei ], ej i − h[ej , ek ], ei i. Note that in a coordinate frame the last three terms vanish, whereas in an ONframe the first three terms vanish. The Christoffel symbols for a metric covariant derivative in an ON-frame give a skew-symmetric map on the fibers Ep . This is true in particular for the Levi-Civita covariant derivative on T M , and in this case we will identify these skew-symmetric maps and the corresponding bivector using Proposition 4.2.3. Definition 11.1.9 (T M Christoffel bivectors). Let M be a Riemannian manifold and let e = {ei (p)} be an ON-frame for the tangent bundle T M in an open set U ⊂ M . Denote the Christoffel symbols for the Levi-Civita covariant derivative by Γe ∈ C ∞ (U ; L(T M ; ∧2 M )), so that ∇v ei = Γe (v) x ei . For the definition of ∧2 M , see Section 11.2.

11.2

Multivector Calculus on Manifolds

In this section we show how the basic affine multivector calculus from Chapter 7 generalizes to manifolds. To simplify the presentation, we consider only compact Riemannian manifolds. Definition 11.2.1 (Multivector bundle). Let M be an n-dimensional Riemannian manifold with atlas {µα }α∈I . The multivector bundle ∧M over M is the vector bundle {∧(Tp M )}p∈M , with fiber space L = ∧Rn and a bundle atlas of bundle charts µα : ∧Rn → ∧(Tp M ) comprising the linear maps induced by the bundle charts of T M as in Definition 2.3.1. We refer to sections of ∧M as (tangent) multivector fields on M . Given a frame {ei }ni=1 for T M , we obtain an induced frame {es }s⊂n for ∧M . The bundle ∧M is a metric bundle equipped with the metric induced from T M , as in Definition 2.5.2, on each fiber. Using the standard inner product on the fiber space ∧Rn , in Mα = µα (Dα ) the manifold chart µα gives rise to the following three bundle charts for ∧M . • The pushforward chart µα (p) : ∧Rn → ∧Tp M . • The pullback chart (µ∗α )−1 (p) : ∧Rn → ∧Tp M .

11.2. Multivector Calculus on Manifolds

391

• The normalized pushforward chart |Jµα |−1 µα (p) : ∧Rn → ∧Tp M . Denote by ∧k M the subbundle of tangential k-vectors, k = 0, 1 . . . , n, so that ∧M = ∧0 M ⊕ ∧1 M ⊕ ∧2 M ⊕ · · · ⊕ ∧n M, where ∧1 M = T M and ∧0 M = M × R. Note that we shall not use the dual multicovector bundle (∧M )∗ , since we identify it with ∧M as in Example 11.1.1. Exercise 11.2.2 (Bundle transition maps). In a fixed manifold chart Mα = µα (Dα ), show that the transition map for ∧M from the pushforward chart to the pullback chart is G(p) : ∧Rn → ∧Rn , the ∧-extension of the linear map of Rn corresponding to the metric gij (x) = hµα (ei ), µα (ej )ip , p = µα (x). Also show that the transition map from the normalized pushforward chart to the √ pullback chart is G/ det g. If µβ : Dβ → Mβ is a second manifold chart and the manifold transition map from Dα to Dβ is µβα , show that the bundle transition maps between the pushforward, pullback, and normalized pushforward charts are µβα∗ , µ∗βα , and µ ˜βα∗ respectively. Note for ∧M that the pushforward charts µα are not only linear, but also ∧homomorphisms, as are the transition maps µβα . The same is true for the pullback charts, but not for the normalized pushforward charts. We have more precisely constructed not only a bundle of linear spaces but a bundle of associative algebras equipped with the exterior product. Applying the affine theory on each fiber, we also have interior products and Hodge star operations defined on ∧M . The latter require a choice of orientation of M , that is, a choice of frame en for the line bundle, that is, a vector bundle with one-dimensional fibers, ∧n M with |en | = 1. The local existence of such en in a chart Mα presents no problem. The global existence is discussed below. A C 1 map ρ : M → M 0 between two different manifolds M and M 0 induces, by working with appropriate types of charts, maps between the respective multivector bundles ∧M and ∧M 0 . Let µα : Dα → M be a chart around p ∈ M , let µ0β : Dβ0 → M 0 be a chart around q = ρ(p) ∈ M 0 , and define ρβα (x) := (µ0β )−1 (ρ(µα (x))). By Exercise 11.2.2, the following constructions are independent of choices of charts. • We define the pullback of a multivector field F 0 on M 0 to be the multivector ∗ field ρ∗ F 0 on M such that µ∗α (ρ∗ F 0 (p)) = ρ∗βα (µ0β F 0 (q)). • If ρ is a diffeomorphism, then we define the pushforward of a multivector field F on M to be the multivector field ρ∗ F on M 0 such that (µ0β∗ )−1 (ρ∗ F (q)) = ρβα∗ (µ−1 ˜∗ F on M 0 is defined similarly α∗ F (p)). The normalized pushforward ρ 0 −1 by demanding that (˜ µβ∗ ) (˜ ρ∗ F (q)) = ρ˜βα∗ (˜ µ−1 α∗ F (p)).

Chapter 11. Multivector and Spinor Bundles

392

Exercise 11.2.3 (Tangential restriction). Let M be a k-surface in a Riemannian manifold N with embedding ρ : M → N , and let p ∈ M ⊂ N . This means in particular that we can regard the tangent space Tp M as a subspace of Tp N . Generalize Exercise 7.2.2 and show that ρ∗ F (p), where F is a multivector field on N , equals the part of F (p) ∈ ∧Tp N tangential to Tp M . We next consider how directional, exterior, and interior derivatives generalize to manifolds. Proposition 11.2.4 (∧M covariant derivative). Let M be a Riemannian manifold. Then there exists a unique covariant derivative ∇v on ∧M that • equals the Levi-Civita covariant derivative from Proposition 11.1.7 on ∧1 M = TM, • equals the tangential directional derivative on ∧0 M = M × R, and • satisfies the product rule ∇v (w1 ∧ w2 ) = (∇v w1 ) ∧ w2 + w1 ∧ (∇v w2 ) for all C 1 multivector fields w1 and w2 and vectors v. Note that this induced covariant derivative, which we refer to as the LeviCivita covariant derivative on ∧M , preserves the subbundles ∧k M , k = 0, 1, . . . , n. Proof. Consider an induced ON-frame {es }s⊂n for ∧M . It is clear that a covariant derivative with the stated properties is unique, since ∇v (es1

∧

· · · ∧ esk ) =

k X

es1

∧

· · · ∧ ∇v esi

∧

· · · ∧ esk

(11.2)

i=1

and ∇v 1 = 0. To show existence, we define a covariant derivative by these identities. It suffices to verify the product rule for w1 = es1 ∧ · · · ∧ esk and w2 = et1 ∧ · · · ∧ etl . We note that (11.2) continues to hold for arbitrary indices 1 ≤ si ≤ n, not necessarily distinct or in increasing order, as a consequence of ∇v esi

∧

esi + esi

∧

∇v esi = 0.

This shows that ∇v (es ∧ et ) = (∇v es ) ∧ et + es ∧ (∇et ).

By its definition above, the Levi-Civita covariant derivative on ∧M satisfies the product rule with respect to exterior multiplication. The following shows that it also satisfies the natural product rules with respect to other products. This is also true for the Clifford product, but we postpone this discussion to Section 11.5. Proposition 11.2.5 (Covariant product rules). Let M be a Riemannian manifold, and let ∇v denote the Levi-Civita covariant derivative on ∧M . This is a metric covariant derivative, that is, ∂v hF, Gi = h∇v F, Gi + hF, ∇v Gi. Moreover, we have ∇v (F y G) = (∇v F ) y G + F y (∇v G), for all v ∈ C(M ; T M ) and F, G ∈ C 1 (M ; ∧M ). In particular, ∇v (F ∗) = (∇v F )∗. The analogous product rules for the right interior product also hold.

11.2. Multivector Calculus on Manifolds

393

Proof. It is clear from the definition that ∇v is metric on ∧0 M and ∧1 M . By bilinearity, it suffices to consider simple k-vector fields F = F1 ∧ · · · ∧ Fk and G = G1 ∧ · · · ∧ Gk . In this case, we have hF1 , G1 i · · · hF1 , Gk i X Y .. .. ... ∂ hFi , Gj i = ∂v hF, Gi = ∂v (±) v . . hFk , G1 i · · · hFk , Gk i X Y hFi , Gj i = · · · = h∇v F, Gi + hF, ∇v Gi, (±)∂v = Q since each vector field Fi and Gj appears exactly once in each product hFi , Gj i. The covariant derivative being metric, the identity for the left interior product follows by dualilty from the product rule for the exterior derivative, since h∇v (F y G), Hi = ∂v hF y G, Hi − hF y G, ∇v Hi = ∂v hG, F ∧ Hi − hG, F ∧ ∇v Hi = h∇v G, F ∧ Hi + hG, ∇v (F ∧ H) − F ∧ ∇v Hi = hF y ∇v G, Hi + h∇v F y G, Hi for all H ∈ C 1 (M ; ∧M ). From this, the identity for ∇v (F ∗) will follow if we prove that ∇v en = 0 for en ∈ C ∞ (M ; ∧n M ) such that |en | = 1. This follows from the fact that ∧n M has one-dimensional fibers and 0 = ∂v |en |2 = 2h∇v en , en i. The proof of the identities for the right interior product and Hodge star are similar. Using the pullback and normalized pushforward charts respectively, Exercise 11.2.2 and the commutation theorem (Theorem 7.2.9) show that the following gives well-defined exterior and interior derivatives of multivector fields on a manifold. Definition 11.2.6 (d and δ on manifolds). Consider the exterior bundle ∧M over a Riemannian manifold M . We define the exterior and interior derivatives dF and δF of a multivector field F ∈ C 1 (M ; ∧M ) as follows. In a chart µα : Dα → Mα , we define dF (p) := (µ∗α )−1 d(µ∗α F (x)), δF (p) := |Jµα |−1 µα δ (|Jµα |(µ−1 )F (x)), α at p = µα (x), x ∈ Dα , where d and δ on the right-hand sides denote the Rn derivatives from Definition 7.1.5. Exercise 11.2.7 (d and δ duality). Let M be a Riemannian manifold, with measure dp. Generalize Proposition 7.1.7 to Riemannian manifolds and prove the following duality relations.

Chapter 11. Multivector and Spinor Bundles

394

For all F ∈ C 1 (M ; ∧k M ), we have δ(F ∗) = (−1)k (dF )∗ at p ∈ M , where the Hodge star uses any fixed orientation of the manifold locally around p. For all F ∈ C 1 (M ; ∧M ) and G ∈ C 1 (M ; ∧M ) that vanish on ∂M , we have Z Z hF, δGidp. hdF, Gidp = − M

M

Hint: Use a partition of unity to localize to charts. The following main result in this section shows that the exterior and interior derivatives on a Riemannian manifold are nabla operators in the natural covariant sense. Proposition 11.2.8 (Covariant nabla). Let M be a Riemannian manifold and let ∇v be the Levi-Civita covariant derivative on ∧M . If {ei } is a frame for T M , with dual frame {ei∗ }, then X dF = e∗i ∧ ∇ei F, F ∈ C 1 (M ; ∧M ), i

δF =

X

ei∗ y ∇ei F,

F ∈ C 1 (M ; ∧M ).

i

Proof. The proof for the exterior derivative uses the expansion rule for d from Exercise 7.4.9. Applying pullbacks, pushforwards, and Exercise 11.1.6, this is seen to hold on M and shows that hdF, v0 ∧ · · · ∧ vk i =

k X (−1)i ∂vi hF, v0 ∧ · · · vˇi · · · ∧ vk i i=0

+

X

(−1)i+j hF, [vi , vj ] ∧ v0 ∧ · · · vˇi · · · vˇj · · · ∧ vk i,

0≤i<j≤k

for F ∈ C 1 (M ; ∧k M ) and vj ∈ C 1 (M ; ∧1 M ). Using that the Levi-Civita covariant derivative is metric, the first sum on the right-hand side is seen to equal k X (−1)i h∇vi F, v0 ∧ · · · vˇi · · · ∧ vk i i=0

−

X

(−1)i+j hF, (∇vi vj − ∇vj vi ) ∧ v0 ∧ · · · vˇi · · · vˇj · · · ∧ vk i

0≤i<j≤k

Pk Since ∇v is also torsion-free, it follows that hdF, v0 ∧ · · · ∧ vk i = i=0 h∇vi F, vi y (v0 ∧ · · · ∧ vk )i. This proves the stated formula, since both sides equal the gradient in the case k = 0 of scalar functions.

11.2. Multivector Calculus on Manifolds

395

To transfer this result to δ, we use Hodge duality. Write F = G∗, where we may assume that G is a k-vector section. Then X δF = (−1)k (dG)∗ = (−1)k e∗i ∧ ∇ei G ∗ i

=

X

∇ei G

∗ ∧ ei

∗=

i

X

e∗i y (∇ei G)∗ =

X

i

e∗i y ∇ei F,

i

where we have used Exercise 11.2.7 in the first step and Proposition 11.2.5 in the last step. In the case of an affine space (X, V ), the two one-dimensional linear spaces ∧0 V = R and ∧n V are isomorphic. For manifolds, the vector bundles ∧0 M and ∧n M need not be globally isomorphic: this happens precisely when M is orientable, that is, when there exists a globally defined section en ∈ C ∞ (M ; ∧n M ) with |en | = ˇ cohomology. 1. We can characterize orientability of the manifold M using Cech Example S 11.2.9 (The sheaf Z2 ). Let M be an n-dimensional manifold, with a cover M = α∈I Uα by open sets. For simplicity, we shall always assume that this is a good cover, although we will not always need all of this hypothesis. Analogous to Definition 10.6.4, this means that each nonempty intersection Us := Uα1 ∩ Uα2 ∩ · · · ∩ Uαk , s = {α1 , . . . , αk } is diffeomorphic to a ball in Rn . Consider the sheaf Z2 , which to each Us associates ˇ the additive group Z2 = {0, 1}. The Cech algebra from Section 10.5 generalizes to Z2 , since it really only requires the spaces to be abelian groups and not vector spaces. In this simplest group Z2 we have −1 = 1, which means that we do not need an alternating algebra. Therefore we write f (s) := hf, es i = −hf, es i ˇ for the value of the Cech cochain at the intersection Us . Also the coboundary operator simplifies to X ∂k f (s) = f (s \ {α}), |s| = k + 2, f ∈ C k (U ; Z2 ). α∈s

Proposition 11.2.10 (First Stiefel–Whitney class). Let M be a manifold with a bundle atlas of bundle charts µα (p) ∈ L(Rn ; Tp M ), p ∈ Uα , α ∈ I, for T M . Consider the sheaf Z2 on the good cover {Uα } from Example 11.2.9. Let f ∈ C 1 (U ; Z2 ) be the 1-cochain, for which the value on the intersection Uα ∩ Uβ is f ({α, β}) :=

( 0,

if det µβα > 0,

1,

if det µβα < 0.

Chapter 11. Multivector and Spinor Bundles

396

ˇ Then ∂1 f = 0, and the Cech cohomology class [f ] does not depend on the choice of local orientation in each Uα specified by the bundle charts µα . The manifold M is orientable if and only if there exists g ∈ C 0 (U ; Z2 ) such that ∂0 g = f , that is, if [f ] = [0]. ˇ The Cech cohomology class [f ] ∈ H 1 (U ; Z2 ) is called the first Stiefel–Whitney class w1 (M ) of M . The two orientations of M correspond to H 0 (U ; Z2 ) = Z2 . Proof. That ∂1 f ({α, β, γ}) = f ({α, β}) + f ({α, γ}) + f ({β, γ}) = 0

mod 2

follows from the transitivity µγβ µβα = µγα of the transition maps. Another choice of local orientation in each Uα corresponds to g ∈ C 0 (U ; Z2 ), where g({α}) = 1 if we change the orientation in Uα , and g({α}) = 0 otherwise. If f 0 ∈ C 1 (U ; Z2 ) is the 1-cochain for this new choice, then f 0 ({α, β}) = f ({α, β}) + g({α}) + g({β})

mod 2,

that is, f 0 = f + ∂0 g. Since M is orientable, we can choose f 0 = 0; this completes the proof. The integral calculus from Section 7.3 carries over with minor changes from affine spaces to general manifolds M . In the absence of an affine space X in which M is embedded, the main change is that we no longer can consider integrals like R oriented measure ∧k M or integrals M F (p)dp of tangent vector fields F . Such notions are not defined for general manifolds, for the simple reason that if E is a general vector bundle over a manifold M , then two fibers Ep and Eq , p 6= q, are incomparable. Two consequences are the following. • The vector sum vp + vq ,

vp ∈ Ep , vq ∈ Eq ,

is not defined, except as a direct sum vp ⊕ vq = (vp , vq ), which is not what we want. The continuous analogue of this, the integral, is therefore not possible either. • It is not well defined what is meant by a constant section of a general vector bundle. When the form we integrate takes values in a fixed linear space L for all p ∈ M , then we can proceed as before. We outline the main steps of the extension of Section 7.3 to an integral calculus of forms on k-surfaces M embedded in an n-dimensional manifold N . • Extending Definition 7.3.1, a general k-form on N is now a map that to each p ∈ N , associates a homogeneous function Θ : ∧k Tp N → L : w 7→ Θ(p, w). If these functions are linear, we refer to Θ as a linear k-form.

11.2. Multivector Calculus on Manifolds

397

• Using charts and a partition of unity, we define the integral Z Θ(p, dp) M

of a k-form Θ in N over an oriented C 1 -regular k-surface M . When R Θ is an even form, for example when we are integrating a scalar function M f (p)dp in a Riemannian manifold, where dp = |dp|, we do not need M to be oriented. • A linear k-form Θ on N can be uniquely written Θ(p, w) = hF1 (p), wiv1 + · · · + hFm (p), wivm , if {vj }m j=1 is a basis for L, where Fj and w are k-vector fields on N . By Proposition 11.2.8, its exterior derivative is Θ(p, ˙ ∇ y w) =

m m X n X X hdFj , wi = h∇ei Fj , e∗i y wivj j=1

j=1 i=1

in a frame {ei } for T M . Note that we cannot regard w as a constant section of ∧k M , and the product rule formula in Definition 7.3.7 needs to be adjusted by a term involving the covariant derivative of e∗i y w if we want to use this on a manifold N . • If the k-surface M in N is oriented and C 1 -regular, and if ∂M has the orientation induced by M , then for every C 1 -regular k-vector field F in a neighborhood of M in N , we have the Stokes formula Z Z hdF (p), dpi = hF (q), dqi. M

∂M

Here dF denotes the exterior derivative. Extending the nabla notation to manifolds, we may write ∇ ∧ F to avoid notational confusion between the exterior derivative dF and the oriented measure dp. For L-valued linear kforms, the Stokes formula reads Z Z Θ(p, ˙ ∇ y dp) = Θ(q, dq). M

∂M

Also the Hodge dual Stokes formula (7.4) extends to manifolds, by inspection of the affine proof, using Proposition 11.2.5. In particular, the following special case is useful. Exercise 11.2.11 (Divergence theorem). Let D be a domain in a Riemannian manifold N , with outward pointing unit normal vector field ν on ∂D. Show that Z Z div F (p) dp = hF (q), ν(q)i dq, D 1

∂D

for a C vector field F on D. Here div F = δF = tion 11.2.8.

P

i hei , ∇ei F i,

using Proposi-

Chapter 11. Multivector and Spinor Bundles

398

11.3 Curvature and Bivectors Consider a vector bundle E over a manifold M , with a given covariant derivative ∇v as in Definition 11.1.4 that allows us to differentiate sections of E. Given a n for E, we have Christoffel symbols ωe defined by frame {ei }i=1 X hωji , viej ∇v ei = j

as in Definition 11.1.5. Although in general we cannot give a definition of what it means for a section to be constant, using the covariant derivative, we can define what is meant by a section being constant along a given curve on M as follows. Definition 11.3.1 (Parallel sections). Let E be a vector bundle over a manifold M , with covariant derivative ∇v , and consider a curve γ, that is, a 1-surface, in M . A section F of E over γ is said to be parallel if ∇v F = 0 at each p ∈ γ and tangent vector v to γ at p. The following calculation shows that in general we cannot even extend this notion of parallel sections to 2-surfaces, which leads us to the fundamental notion of curvature. Example 11.3.2 (Parallel transport around triangles). Let E be a vector bundle over a manifold M , with covariant derivative ∇v . Consider a 2-surface Σ in M , at a given point p ∈ M , defined by a chart µ : D → Σ ⊂ M. Here D is the triangle D := {(x, y) ; x, y ≥ 0, x + y ≤ 1} ⊂ R2 and µ(0, 0) = p. For 0 < < 1 we also consider the subsurfaces Σ = µ(D ), where D := {(x, y) ; x, y ≥ 0, x + y ≤ }. Assume that F is a section of E over the curve ∂Σ that is parallel along the three sides and continuous at µ(, 0) and µ(0, ). We do not assume continuity at p. This defines a map Ep → Ep : lim F (µ(x, 0)) 7→ lim F (µ(0, y)), x→0+

y7→0+

which is referred to as the parallel transport around ∂Σ . We want to calculate this map to orderP 2 . To this end, fix a frame {ei }N i=1 for E in a neighborhood of Σ and write F = i Fi ei . The equation ∇v F = 0 becomes X X ∂v Fi + hωij , viFj ei = 0. i

j

11.3. Curvature and Bivectors

399

Figure 11.1: Parallel transport of a vector around a triangle Σ. Since the sphere has positive curvature, the resulting rotation will be in the direction corresponding to the orientation of Σ. 0 Write ωij := µ∗ (ωij ) for the pulled back vector field on D . Then the equation becomes N scalar differential equations X 0 ∂ u fi + hωij , uifj = 0, i = 1, . . . , N, j

along the triangle curve ∂D , where fi := µ∗ Fi = Fi ◦ µ and u is a tangent vector to ∂D . 0 Let zi := fi (0+ , 0) be given, and write aij (x, y) := hωij , e1 i and bij (x, y) := 0 hωij , e2 i. Also write aij := aij (0, 0), ∂x aij := ∂x aij (0, 0), ∂y aij := ∂y aij (0, 0), and similarly for bij . (i) We first solve the parallel transport equations ∂x fi (x, 0) +

X

aij (x, 0)fj (x, 0) = 0

j

along {(x, 0) ; 0 < x < }. With the ansatzes fi (x, 0) = φ0i + φ1i x + φ2i x2 + O(x3 )

400

Chapter 11. Multivector and Spinor Bundles

0 1 + αij x + O(x2 ), we obtain and aij (x, 0) = αij

fi (, 0) = zi −

X

aij zj +

j

2 X X aik akj − (∂x aij ) zj + O(3 ) 2 j k

under the inital condition φ0i = zi . (ii) We next solve the parallel transport equations X ∂x fi ( − x, x) + (−aij ( − x, x) + bij ( − x, x))fj ( − x, x) = 0 j

along {( − x, x) ; 0 < x < }. We approximate − aij ( − x, x) + bij ( − x, x) = (−aij + bij − ∂x aij + ∂x bij ) + (∂x aij − ∂y aij − ∂x bij + ∂y bij )x + O(2 ). 0 1 Letting φ0i = fi (, 0), αij = −aij + bij − ∂x aij + ∂x bij , and αij = ∂x aij − ∂y aij − ∂x bij + ∂y bij in (i), we get X fi (0, ) = zi − bij zj j 2

XX + (−aik bkj + bik akj + bik bkj ) − (∂y aij − ∂x bij − ∂y bij ) zj + O(3 ). 2 j k

(iii) We finally solve the parallel transport equations X ∂x fi (0, − x) − bij (0, − x)fj (0, − x) = 0 j

along {(0, − x) ; 0 < x < }. We approximate −bij (0, − x) = (−bij − ∂y bij ) + (∂y bij )x + O(2 ). 0 1 Letting φ0i = fi (0, ), αij = −bij − ∂y bij , and αij = ∂y bij in (i), we get

fi (0, 0+ ) = zi +

2 X X (−aik bkj + bik akj ) + (∂y aij − ∂x bij ) zj + O(3 ). 2 j k

To interpret this result, we push forward the oriented measure D to 2 ∧2 (Σ ) := µ(0,0) ( 2 e1 ∧ e2 ) ∈ ∧2 (Tp M ).

2 2 e1 ∧

e2 of

Although there is no notion of oriented measure of k-surfaces in manifolds in general, for ≈ 0 the bivector ∧2 (Σ ) approximately describes an oriented measure

401

11.3. Curvature and Bivectors

0 of the small 2-surface Σ . For the vector fields ωij = aij e1 + bij e2 in R2 , we note that 0 dωij = (∂x bij − ∂y aij )e12 , 0 0 ∧ ωkj = (aik bkj − bik akj )e12 . ωik

Therefore our calculation shows that X X 0 0 2 0 3 fi (0, 0+ ) = zi − ∧ ωkj , 2 e12 izj + O( ). ωik hdωij + j

k

Back on M we conclude that parallel transport around Σ equals X Ep → Ep : F 7→ F − hΩij , ∧2 (Σ )iFj ei + O(3 ), j

where Ω is the curvature operator defined below. The following encodes the curvature of a general manifold, as compared to Definition 10.3.5 which encodes the curvature of an n − 1 surface embedded in n-dimensional Euclidean space. Definition 11.3.3 (Curvature operator). Consider a vector bundle E with a covariant derivative ∇v , over a manifold M . Let {ei }N i=1 be a frame for E, and let ωij denote the Christoffel symbols of ∇v in this frame. Assuming that M is a Riemannian manifold, these are vector fields on M , defined in the domain of the frame. The curvature operator Ω for ∇v is the linear map that, at each p, to each bivector b ∈ ∧2 (Tp M ) associates the linear map Ω(b) ∈ L(Ep ), which in the given frame has matrix {hΩij , bi}N i,j=1 , where the bivector fields are Ωij := dωij +

N X

ωik ∧ ωkj ,

i, j = 1, . . . , N.

k=1

Note that we do not write Ωe as for ωe . As Example 11.3.2 indicates, unlike the Christoffel symbols, the operator Ω(b) is independent of the choice of frame, as the following shows. Proposition 11.3.4. Let E be a vector bundle with a covariant derivative ∇v , over a manifold M . Consider two frames {ei } and {˜ ei }, defined on the same open e set, with associated Christoffel symbols ω, ω ˜ and P curvature operators Ω and Ω ˜i = j ej αji , then respectively. If the relation between frames is e X X ω ˜ ij = αik ωkl αlj + αik (dαkj ), kl

e ij = Ω

X

k ik

α Ωkl αlj ,

kl

for all i, j = 1, . . . , N , where {αij }ij denotes the inverse matrix to A = {αij }ij .

402

Chapter 11. Multivector and Spinor Bundles

These transformation properties of Ωij mean that the curvature operator is a well-defined section Ω ∈ C ∞ (M ; L(∧2 M ; L(E))), globally on all M , that does not depend on the choice of frame. Note that the corresponding statement is not true for the Christoffel symbols, because of term A−1 dA. It is only when the change of frame matrix A is constant that the Christoffel symbols transform as an operator. ˜i = Proof. On the one hand, ∇v e ˜i = ∇v e

X

P

ej . ωji , vi˜ j h˜

∇v (ej αji ) =

j

On the other hand,

X X hωkj , viek αji + ej (∂v αji ), j

jk

and ∂v αji = hdαji , vi. Comparing these expressions, we obtain X X X hωkj , viek αji + h˜ ωji , viαkj ek = hdαki , viek , jk

jk

k

from which the relation for the Christoffel symbols follows. Furthermore, e ij = Ω

X

dαik ∧

k

+

X

X X αik ωkl αlj + dαkj + (dωkl )αlj − ωkl ∧ dαlj

l

X kmk0

α

ik

X

l

k

ωkl αlm + dαkm

∧

α

mk0

X

ωk0 l0 αl0 j + dαk0 j .

l0

l

to the stated relation, and the Here the terms containing dω and ω ∧ ω combine P remaining terms cancel pairwise, using 0 = k d(αik αkj ) = (dαik )αkj +αik (dαkj ). A fundamental fact about second-order partial derivatives in affine space is that ∂i ∂j = ∂j ∂i . The quantity that describes the failure of this for covariant derivatives on vector bundles turns out to be the curvature operator, as the following shows. Proposition 11.3.5 (Curvature as commutator). Let E be a vector bundle with a covariant derivative ∇v , over a manifold M . Then ∇u (∇v F ) − ∇v (∇u F ) = ∇[u,v] F + Ω(u ∧ v)F, for every section F ∈ C 2 (M ; E).

403

11.3. Curvature and Bivectors Proof. Calculating in a frame {ei }, we have X X ∇u (∇v F ) = ∇u hωki , viFi ek (∂v Fi )ei + i

ki

X X = (∂u ∂v Fi )ei + (∂u hωki , vi)Fi + hωki , vi(∂u Fi ) ek + (∂v Fi )hωki , uiek i

ik

X hωji , viFi hωkj , uiek . + ijk

Subtracting the corresponding formula for ∇v (∇u F ), it suffices to show that ∂u ∂v Fi − ∂v ∂u Fi = ∂[u,v] Fi , ∂u hωki , vi − ∂v hωki , ui = hdωki , u ∧ vi + hωki , [u, v]i, hωji , vihωkj , ui − hωji , uihωkj , vi = hωkj ∧ ωji , u ∧ vi. The first and last equations are straightforward to verify, and the second follows from Exercise 7.4.9, generalized to manifolds. This proves the proposition. Our interest is in metric vector bundles and covariant derivatives, in which case it is clear from Definition 11.3.3 and the skew-symmetry of the Christoffel symbols and the exterior product of vectors that Ωji = −Ωij in an ON-frame {ei }. The map Ω(b) ∈ L(E) being skew-symmetric, recalling Proposition 4.2.3 we can equivalently view it as the bivector X hΩij , biei ∧ ej ∈ ∧2 E, Ω(b) = i<j

so that Ω ∈ C ∞ (M ; L(∧2 M ; ∧2 E)). In this setup, the action of Ω(b) on a section F ∈ C ∞ (M ; E) is Ω(b)F = Ω(b) x F, using the right interior product in exterior bundle ∧E. Although possible for any metric vector bundle, we shall use only this bivector representation of the curvature operator for the tangent bundle E = T M . Definition 11.3.6 (Riemann curvature operator). Let M be a Riemannian manifold. By the Riemann curvature operator we mean the curvature operator R ∈ C ∞ (M ; L(∧2 M )) for the Levi-Civita covariant derivative on the tangent bundle T M . Using the relation ∧2 M = SO(T M ) ⊂ L(T M ) provided by Proposition 4.2.3, we have R ∈ C ∞ (M ; L(T M ) ⊗ L(T M )). Using the contractions L(T M ) ⊗ L(T M ) → L(T M ) = T M ⊗ T M → M × R,

Chapter 11. Multivector and Spinor Bundles

404

obtained by lifting the bilinear products A1 A∗2 and hv1 , v2 i of matrices and vectors respectively to the tensor products, we define from R the Ricci curvature Ric ∈ C ∞ (M ; L(T M )) and the scalar curvature S ∈ C ∞ (M ; R). If {ei } is an ON-frame for T M , then we define the Riemann curvature coefficients Rijkl := hei ∧ ej , R(ek ∧ el )i. Then the matrix for the Ricci curvature operator in this frame is Ricij := hei , Ric(ej )i =

X

Rikjk ,

k

and the scalar curvature is S =

P

i

Ricii =

P

ik

Rikik .

Proposition 11.3.7 (Symmetries of R). Let R ∈ C ∞ (M ; L(∧2 M )) be the Riemann curvature operator on a Riemannian manifold M . Then at each p ∈ M , the operator R : (∧2 M )p → (∧2 M )p is symmetric, and the Bianchi identities R(u ∧ v) x w + R(v ∧ w) x u + R(w ∧ u) x v = 0 hold for all vectors u, v, w ∈ Tp M . Proof. To prove the Bianchi identities, by trilinearity it suffices to consider a coordinate frame {ei }. In this case, [ei , ej ] = 0, so by Proposition 11.3.5 the identities follow from the computation ∇ei ∇ej ek − ∇ej ∇ei ek + ∇ej ∇ek ei − ∇ek ∇ej ei + ∇ek ∇ei ej − ∇ei ∇ek ej = ∇ei (∇ej ek − ∇ek ej ) + ∇ej (∇ek ei − ∇ei ek ) + ∇ek (∇ei ej − ∇ej ei ) = 0, since ∇u v − ∇v u = [u, v] for the Levi-civita covariant derivative. Using the Bianchi identities, we compute hu2 ∧ v2 , R(u1 ∧ v1 )i = hu2 , R(u1 ∧ v1 ) x v2 i = −hu2 , R(v1 ∧ v2 ) x u1 i − hu2 , R(v2 ∧ u1 ) x v1 i = hu1 , R(v1 ∧ v2 ) x u2 i + hv1 , R(v2 ∧ u1 ) x u2 i = −hu1 , R(v2 ∧ u2 ) x v1 i − hu1 , R(u2 ∧ v1 ) x v2 i − hv1 , R(u1 ∧ u2 ) x v2 i − hv1 , R(u2 ∧ v2 ) x u1 i = 2hu1 ∧ v1 , R(u2 ∧ v2 )i + hv2 , R(u2 ∧ v1 ) x u1 i + hv2 , R(u1 ∧ u2 ) x v1 i = 2hu1 ∧ v1 , R(u2 ∧ v2 )i − hv2 , R(v1 ∧ u1 ) x u2 i, from which the stated symmetry follows.

405

11.4. Conformal Maps and ON-Frames

11.4 Conformal Maps and ON-Frames The main result in this section, Liouville’s theorem on conformal maps, is difficult to place. It concerns the fractional linear maps from Section 4.5 and completes the proof of Theorem 4.5.12. It is also relevant for the hypercomplex analysis from Chapter 8. We have nevertheless placed it here, since the proof makes use of ON-frames and curvature. We start with the following method of Cartan for calculating the Christoffel symbols in an ON-frame. Proposition 11.4.1 (Cartan). Let {ei } be an ON-frame for T M on a Riemannian manifold M . Then the Christoffel symbols for the Levi-Civita covariant derivative in {ei } are the unique vector fields ωij that satisfy ωji = −ωij and dei +

n X

ωij

∧

ej = 0,

j = 1, . . . , n.

(11.3)

j=1

Note that the uniqueness part of the proposition means that any method of writing the exterior derivatives of the ON-frame vectors as in (11.3) must give the Christoffel symbols. Proof. Using Proposition 11.2.8, we have X XX X X X ωki ∧ ek dei = ej ∧ ∇ej ei = ej ∧ hωki , ej iek = hωki , ej iej ∧ ek = j

j

k

k

j

k

0 is any skew-symmetric for the Christoffel symbols. For uniqueness, assume that ωij ˜ ij := family of vector fields that satisfies (11.3), and consider the differences ω 0 . We have ωij − ωij X ω ˜ ij ∧ ej = 0, i = 1, . . . , n. j

Forming the inner product with frame bivectors gives X X 0= hek ∧ el , ω ˜ ij ∧ ej i = ˜ ij ihel , ej i − hek , ej ihel , ω ˜ ij i hek , ω j

j

˜ ik i. = hek , ω ˜ il i − hel , ω ωji for all i, j. This yields ˜ ij = −˜ So hek , ω ˜ il i = hel , ω ˜ ik i for all i, k, l, as well as ω ˜ ik i = hek , ω ˜ ki i = hej , ω ˜ jk i = −hei , ω hek , ω ˜ ij i = −hek , ω ˜ ji i ˜ ji i = hei , ω ˜ kj i = −hej , ω 0 for all k, i, j. This proves that the ωij coincide with the Christoffel symbols.

A two-dimensional Euclidean space abounds with conformal maps: all analytic functions are conformal where the derivative is nonzero. The goal of this

Chapter 11. Multivector and Spinor Bundles

406

section is to prove Liouville’s theorem on conformal maps, which shows that in n-dimensional Euclidean space, n ≥ 3, the situation is drastically different: the only conformal maps are the fractional linear maps / M¨obius maps from Theorem 4.5.12. These are very few indeed, since they depend only on as many parameters as the Lorentz isometries in dimension n + 2, that is, n+2 real parameters. 2 Theorem 11.4.2 (Liouville). Let D be a connected open subset of a Euclidean space (X, V ) of dimension dim X ≥ 3. Assume that f : D → V is a conformal map of class C 3 . Then there exists a fractional linear map g : V → V such that g|D = f . Here X = V with origin fixed, and extended V ⊂ V with notation as in Section 4.5. Even though the conclusion is global in the sense that we obtain a conformal diffeomorphism of the extended space V, the problem is local. Indeed, since fractional linear maps are real analytic, it suffices to prove that f coincides with a fractional linear map in D, when D is a ball. A concrete proof of this is as follows. Assume that we have proved that f = f1 on a ball B1 , and that f = f2 on a ball B2 , where B1 ∩ B2 6= ∅ and f1 and f2 are fractional linear maps. To prove that f1 = f2 on all V, write f0 (x) := f2−1 (f1 (x)) = (ax + b)(cx + d)−1 as in Theorem 4.5.16. Then x = (ax + b)(cx + d)−1 ,

for all x ∈ B1 ∩ B2 ,

or equivalently xcx + xd − ax − b = 0. If this second-order polynomial vanishes on an open set, then c = b = 0 and xd = ax for all vectors x. This proves that f0 (x) = axd−1 = x for all x ∈ V . Proof of Theorem 11.4.2. Without loss of generality, we assume that D is a ball with small enough radius that f : D → f (D) is a C 3 -diffeomorphism by the inverse function theorem. Since f is conformal, there exists a unique scalar function λ(x) > 0 in D such that λ(x)f x

is an isometry at each x ∈ D,

and by assumption λ is C 2 -regular. Fix an ON-basis {ei } for V and define the ON-frame ei (y) := λ(x)f x (ei ), y = f (x), in the open set f (D). By definition of λ, f∗ λ is an isometry, so f ∗ f∗ = λ−2 I when acting on vector fields in D. Therefore, pulling back {ei (y)} to D gives f ∗ (ei )(x) = 1/λ(x)ei , x ∈ D. Applying the exterior derivative and using Theorem 7.2.9, we get X f ∗ (∇ ∧ ei ) = −λ−2 (∇λ) ∧ ei = −λ−2 λj ej ∧ ei , j

since ei is a constant vector field, where λi := ∂i λ and ∂i is the partial derivative along ei . Pushing forward this equation to f (D), noting that f∗ f ∗ = µ−4 I

407

11.4. Conformal Maps and ON-Frames

when P acting on bivector fields in f (D), where µ := λ ◦ f −1 , gives µ−4 ∇ ∧ ei = −µ−2 j µi (µ−1 ej ) ∧ (µ−1 ei ), where µj := f∗ (λj ) = λj ◦ f −1 . We obtain X X ∇ ∧ ei = − µi ej − µj ei ∧ ej . µj ej ∧ ei = − j

j

Uniqueness of the Christoffel symbols in Proposition 11.4.1 shows that ωij = µi ej − µj ei . This gives X − µi ek ∧ ωkj + µj ek ∧ ωki ∇ ∧ ωij = (∇µi ) ∧ ej − (∇µj ) ∧ ei + k

= (∇µi ) ∧ ej − (∇µj ) ∧ ei +

X

− µi µk ek ∧ ej + µj µk ek ∧ ei ,

k

X

ωik ∧ ωkj

k

X = (µi ek − µk ei ) ∧ (µk ej − µj ek ) k

X = (µi µk ek ∧ ej − µ2k ei ∧ ej + µk µj ei ∧ ek ). k

Since Euclidean space has zero P curvature, it follows from Definition 11.3.3 that (∇µi ) ∧ ej − (∇µj ) ∧ ei = ( k µ2k )ei ∧ eP j . Pulling back this equation to D gives (∇λi ) ∧ (λ−1 ej ) − (∇λj ) ∧ (λ−1 ei ) = ( k λ2k )(λ−1 ei ) ∧ (λ−1 ej ). We obtain the following nonlinear system of second-order partial differential equations λ

n X

n X λ2k ei ∧ ej , λik ek ∧ ej − λjk ek ∧ ei =

k=1

i, j = 1, . . . , n,

(11.4)

k=1

where λij := ∂j ∂i λ. If dim X ≥ 3, then the system (11.4) is overdetermined, which we exploit as follows. Evaluating the em ∧ ej component of the equation (11.4), where i, j, m are distinct, we deduce that λim = 0 whenever i 6= m. On the other hand, evaluating the ei ∧ ej component of the same equation, it follows that λii + λjj = λ−1 |∇λ|2 . Since this equation holds for all i 6= j, we get λii = |∇λ|2 /(2λ), i = 1, . . . , n. Since λim = 0 whenever i 6= m, λi is a function of xi only, and so is λii . This holds for all i, so |∇λ|2 /λ =: c must be a constant independent of all xi . We have shown that λii (x) = |∇λ(x)|2 /(2λ(x)) = c, i = 1, . . . , n, x ∈ D. If c = 0, then ∇λ = 0 and λ(x) =: b is constant. Since ei = (bf∗ )(ei ) = (f −1 )∗ (b−1 ei ), it follows from Theorem 7.2.9 that ∇ ∧ ei = 0, since b−1 ei is constant. Proposition 11.4.1 shows that the Christoffel symbols vanish. Therefore ei must be constant in D since the covariant derivatives for the Euclidean space are the standard partial derivatives. Hence the total derivative f x is a constant matrix, and we conclude that the conformal map is of the form X X X f xi ei = b xi ei + d, xi ei ∈ D, i

i

i

Chapter 11. Multivector and Spinor Bundles

408

for some constant vector d ∈ V . Thus f is a restriction of a composition of an isometry, a dilation, and a translation, and thus a fractional linear map. On the other hand, if c 6= 0, then integration gives ∇λ = c(x − a) for some a ∈ V , since λi depends only on xi . This yields λ(x) =

1 c |c(x − a)|2 = |x − a|2 . 2c 2

This can be reduced to the case that λ(x) is constant, by composing y = f (x) with the fractional linear map x = g(z) = 1/z + a. Since λ > 0 on D, it follows that a ∈ / D, and therefore g maps a domain D0 ⊂ V bijectively onto D. From Exercise 4.5.18 we have that g z (h) = |z|−2 zhz −1 . Thus f ◦ g : D → f (D) has derivative f ◦ g z (h) = f x (|z|−2 zhz −1 ) = |x − a|2 f x (zhz −1 ) =

2 λf (zhz −1 ), c x

which is a constant times an isometry, so the scale factor λ associated to f ◦ g is constant. Thus the calculation above for constant λ applies to f ◦ g and shows that f ◦ g, and hence f , is a fractional linear map. This completes the proof. Exercise 11.4.3. In dimension n = 2, show that (11.4) reduces to the single equation λ∆λ = |∇λ|2 , that is, ∆(ln λ) = 0, saying that ln |f 0 (z)| is harmonic when f (z) is an analytic function.

11.5

Weitzenb¨ock Identities

Proceeding as in Section 3.1, we define on each fiber ∧(Tp M ) of the multivector bundle over a Riemannian manifold M a Clifford product 4. Definition 11.5.1 (Clifford bundle). Let M be a Riemannian manifold. The Clifford bundle 4M over M equals the multivector bundle ∧M as a bundle of linear spaces, but equipped with the Clifford product 4 on the fibers ∧(Tp M ), making it into a bundle of associative algebras over M . Just as in the case of a single affine space, we write 4M instead of ∧M when we use the Clifford product rather than the exterior and interior products, but we are not strict in this, since sometimes we use all these products and it may not be clear whether 4M or ∧M is the appropriate notation. Translating from the multivector bundle, the Clifford bundle is the direct sum of subbundles 4M = 40 M ⊕ 41 M ⊕ 42 M ⊕ · · · ⊕ 4n M. Note that since the charts µα : Rn → T M in general are not isometries, they do not induce isomorphisms between Clifford algebras. However, given an ON-frame {ei }ni=1 for T M , we obtain an induced ON-frame {es }s⊂n for 4M .

409

11.5. Weitzenb¨ock Identities

Exercise 11.5.2 (Clifford product rule). Show for the Levi-Civita covariant derivative on 4M = ∧M that the product rule ∇v (F

4

G) = (∇v F ) 4 G + F

4

(∇v G)

holds for all multivector fields F, G ∈ C 1 (M ; 4M ) on M . A first use of the Clifford product on 4M is to express the Christoffel symbols and curvature of ∧M = 4M in terms of the bivectors Γe and R for T M through Clifford commutators. Proposition 11.5.3 (Clifford bivector commutators). Let M be a Riemannian manifold, and let e = {ei (p)} be an ON-frame for the tangent bundle T M in an open set U ⊂ M . Denote by Γe ∈ C ∞ (U ; L(∧1 M ; ∧2 M )) and R ∈ C ∞ (M ; L(∧2 M ; ∧2 M )) the Christoffel symbols and curvature operator for T M . ˜ = {es (p)} be the induced ON-frame for 4M in U . Denote by ωe˜ ∈ Let e C ∞ (U ; L(∧1 M ; L(4M )) and Ω ∈ C ∞ (M ; L(∧2 M ; L(4M ))) the Christoffel symbols and curvature operator for 4M . Then ωe˜ (v)F = 21 [Γe (v), F ], Ω(b)F = 12 [R(b), F ], for v ∈ C(M ; T M ), b ∈ C(M ; ∧2 M ), and F ∈ C(M ; 4M ), using the Clifford commutator [w1 , w2 ] := w1 4 w2 − w2 4 w1 . Proof. Consider first the Christoffel symbols. For a basis vector field ei ∈ C(M ; T M ), we have ωe˜ (v)ei = ∇v ei = Γe (v) x ei = 12 (Γe (v) 4 ei − ei 4 Γe (v)) = 12 [Γe (v), ei ], using the analogue of (3.3) for the right interior product, and Definitions 11.1.5 and 11.1.9. For scalar functions F ∈ C(M ; 40 M ) both sides of the identity vanish, and for F ∈ C(M ; 4k M ), k ≥ 2, the result follows from the vector case k = 1 and the derivation property D(F1 4 F2 ) = (DF1 ) 4 F2 + F1 4 (DF2 ),

F1 , F2 ∈ C 1 (M ; 4M ),

which holds for D = Γe (v), as a consequence of Exercise 11.5.2, as well as for D : F 7→ 12 [Γe (v), F ], by general properties of commutators. For the curvature operator, we can argue similarly. Since Ω(b)ei = R(b) x ei = 12 (R(b) 4 ei − ei 4 R(b)) = 12 [R(b), ei ], it suffices to show that Ω(b)(F1 4 F2 ) = (Ω(b)F1 ) 4 F2 + F1 4 (Ω(b)F2 )

(11.5)

410

Chapter 11. Multivector and Spinor Bundles

for all F1 , F2 ∈ C ∞ (M ; 4M ). By Exercise 11.5.2, we have ∇u ∇v (F1 4 F2 ) = (∇u ∇v F1 ) 4 F2 + (∇v F1 ) 4 (∇u F2 ) + (∇u F1 ) 4 (∇v F2 ) + F1 4 (∇u ∇v F2 ), as well as ∇[u,v] (F1 4 F2 ) = (∇[u,v] F1 ) 4 F2 + F1 property (11.5) follows from Proposition 11.3.5.

4

(∇[u,v] F2 ), so the derivation

The fundamental first-order differential operator on M acting on sections of 4M is the following natural generalization of the Euclidean nabla operator from Definition 9.1.1. Definition 11.5.4 (The Dirac operator on 4M ). Let M be a Riemannian manifold. The 4-Dirac operator on 4M is the operator X DF := dF + δF = e∗i 4 ∇ei F, F ∈ C 1 (M ; 4M ), i

where {ei } is a frame for T M , with dual frame {e∗i }. Note that the Dirac operator D = d+δ is not locally similar to the Euclidean Dirac operator. This in contrast to d and δ, where by definition µ∗ d(µ∗ )−1 and (˜ µ∗ )−1 δ µ ˜∗ are the Euclidean exterior and interior derivatives, locally in any chart µ for M . As in Euclidean space, d2 = 0 and δ 2 = 0, and as a consequence, D2 = dδ + δd for every Riemannian manifold. We saw in Section 8.1 that D2 equals the Laplace operator ∆, acting componentwise on multivector fields, in Euclidean space. The situation on a manifold is more subtle, where D2 differs from the following Laplace operator, on the bundle E = 4M , by a curvature term. Proposition 11.5.5. Let E be a metric vector bundle with a metric covariant derivative, over a Riemannian manifold M . If {ei } is an ON-frame for T M in an open set U ⊂ M , then Z X Z X h∇ei F, ∇ei Gidp = − h∇ei ∇ei F − ∇∇ei ei F, Gidp (11.6) M

i

M

i

for all F, G ∈ C02 (U ; E). Here ∇ei ei denotes the Levi-Civita covariant derivative on T M . Proof. Using that the covariant derivative on E is metric, we get the pointwise identity ∂ei h∇ei F, Gi = h∇ei ∇ei F, Gi + h∇ei F, ∇ei Gi.

411

11.5. Weitzenb¨ock Identities Define the vector field v :=

X h∇ei F, Giei ∈ C01 (U ; T M ). i

From Proposition 11.2.8 we have X X hej , ∇ej (h∇ei F, Giei )i = hej , (∂ej h∇ei F, Gi)ei + h∇ei F, Gi(∇ej ei )i div v = ij

ij

=

X

∂ei h∇ei F, Gi + h∇ei F, Gi div ei .

i

Using the identity X X X X div(ei )ei = hej , ∇ej ei iei = − ∇ej ej , h∇ej ej , ei iei = − i

ij

ij

j

the stated identityR now follows from the divergence theorem in Exercise 11.2.11, which shows that M vdp = 0. Definition 11.5.6 (Laplace–Beltrami operator). Let E be a metric vector bundle with a metric covariant derivative, over a Riemannian manifold M . The Laplace– Beltrami operator is the second-order differential operator that in an ON-frame {ei } for T M is given by X ∆F := (∇ei ∇ei F − ∇∇ei ei F ). i

The left-hand side in (11.6), for G = F , can be viewed as the H 1 Sobolev (semi-)norm k∇ ⊗ F k2L2 (M ;E) , and −∆ is the L2 operator corresponding to this quadratic form. Exercise 11.5.7 (Second covariant derivative). Show that 2 F := ∇u ∇v F − ∇∇u v F ∇u,v

is bilinear in u and v. Deduce that the Laplace–Beltrami operator is well defined, that is, independent of choice of ON-basis. Exercise 11.5.8 (Scalar Laplace–Beltrami). Show that for the trivial vector bundle E = R, the Laplace–Beltrami operator acting on scalar functions f ∈ C 2 (M ; R) is ∆f = δdf. Verify also that this follows from Proposition 11.5.9 below. With a chart for M , and inverse metric (g ij ) = (gij )−1 and determinant g := det(gij ), generalize Example 7.2.12 and show that X √ √1 ∂i ( gg ij ∂j f ). ∆f = g ij

Chapter 11. Multivector and Spinor Bundles

412

The following main result in this section is the key identity used in the proof of the Chern–Gauss–Bonnet theorem in Section 12.3. Proposition 11.5.9 (Weitzenb¨ock identity for D). Let M be a Riemannian manifold M , and let {ei } be an ON-frame for T M . Then X 1 1X ei ej Ω(ei ∧ ej )F = −∆F + SF + −D2 F = −∆F − Rijkl ei ej F ek el , 4 8 i<j ijkl

∞

for F ∈ C (M ; 4M ), where ∆ and Ω denote the Laplace–Beltrami and curvature operators on 4M , and R denotes the Riemann curvature operator on T M with coefficients Rijkl and scalar curvature S. Proof. With notation as in Exercise 11.5.7 and suppressing 4, Proposition 11.3.5 yields X X D2 F = ei ∇ei (ej ∇ej F ) = ei (∇ei ej )∇ej F + ej ∇ei ∇ej F ij

=

X

=

X

ij

ei ej ∇2ei ,ej F

+ ej ∇∇ei ej F + (∇ei ej )∇ej F

ij

e2i ∇2ei ,ei F +

i

+

X

ei ej (∇2ei ,ej F − ∇2ej ,ei F )

i<j

X

ei ej ∇∇ei ej F + (∇ei ej )∇ej F

ij

= ∆F +

X

ei ej Ω(ei ∧ ej )F +

X

i<j

ei h∇ei ej , ek iej ∇ek F + h∇ei ej , ek iek ∇ej F .

ijk

Since h∇ei ej , ek i + hej , ∇ei ek i = ∂ei hej , ek i = 0, this proves the first identity. To express the identity in terms of the Riemann curvature operator, we use Proposition 11.5.3 to write X 1X ei ej Ω(ei ∧ ej )F = ei ej [R(ei ∧ ej ), F ] 2 i<j i<j 1 XX 1X Rijkl ei ej [ek el , F ] = Rijkl (ei ej ek el F − ei ej F ek el ). 4 i<j 8 kl ijkl P It remains to simplify ijkl Rijkl ei ej ek el . From the Bianchi identities in Proposition 11.3.7 and multivector algebra, we obtain X X X Rijkl ei ej ek = − (Rjkil + Rkijl )ei ej ek = − Rijkl (ek ei ej + ej ek ei ) =

ijk

=−

ijk

X

ijk

Rijkl (ei ej ek + 2ek y (ei ej ) + ei ej ek + 2(ej ek ) x ei )

ijk

= −2

X ijk

Rijkl ei ej ek − 4

X ij

Rijil ej + 2

X ij

Rijjl ei + 2

X ik

Riikl ek .

413

11.6. Spinor Bundles

P P Since Riikl = 0 and Rijil = −Rjiil , this yields − ijk Rijkl ei ej ek = 2 i Ricil ei , and in particular X X − Rijkl ei ej ek el = 2 Ricil ei el = 2S, (11.7) i

ijkl

since the Ricci curvature is symmetric. This proves the second form of the Weitzenb¨ock identity.

11.6 Spinor Bundles Let M be a Riemannian manifold. The multivector bundle ∧M = 4M over M is a globally well-defined vector bundle over M , built fiberwise from the tangent vector bundle T M . In this section we investigate when and how it is possible to / globally over M . What we want to achieve construct a bundle of spinor spaces 4M is the following. Definition 11.6.1 (Spinor bundle). Let M be an oriented Riemannian manifold of real dimension n. A (complex) spinor bundle over M is a complex vector bundle 4M / , of complex dimension 2bn/2c , together with linear maps ρ = ρp : Tp M → / )p ) depending smoothly on p ∈ M such that L((4M ρ(v)2 ψ = |v|2 ψ,

v ∈ C(M ; T M ), ψ ∈ C(M ; 4M / ).

/ is normed in the sense of Definition 5.3.3, that is, that each We assume that 4M / )p is equipped with a spinor inner product (·, ·i and a spinor conjugation fiber (4M ·† , which vary smoothly with p. Complexifying each fiber of the multivector bundle as in Section 1.5, we obtain a globally well-defined complex vector bundle 4Mc . Applying Lemma 5.1.4 fiberwise, we see that for a spinor bundle we have homomorphisms 4Mc → / ) of complex algebras at each p ∈ M . Assuming M to be oriented, we L(4M have globally defined main reflectors wn as in Definition 5.2.1. Since ρ(wn ) = ±I in odd dimension, we see that orientability is necessary for the existence of spinor bundles in this case. In even dimension, ρ(wn ) will yield globally defined subbun± dles 4 / M , which are fundamental in Chapter 12.4. For the construction of spinor bundles over an orientable Riemannian manifold M , we fix a bundle atlas for T M of bundle charts µα (p) ∈ L(Rn ; Tp M ), p ∈ Uα , α ∈ I. We assume that each µα (p) is an isometry and that {Uα } is a good cover of M . We also assume that all the charts µα (p) are orientation-preserving. In particular, we have orientation-preserving transition maps µβα (p) ∈ SO(Rn ). Locally over Uα , the existence of a spinor bundle is clear. Indeed, let ρ0 : Rn → L(4R / n)

414

Chapter 11. Multivector and Spinor Bundles

be the complex spinor space from Definition 5.2.4. Define (4M / )p := 4R / n and ρα (v) := ρ0 (µ−1 for v ∈ T M , p ∈ U . This clearly yields a spinor bundle (p)v), p α α locally over Uα . To construct a spinor bundle globally over M , we study the transition maps µβα (p) = µβ (p)−1 µα (p) ∈ SO(Rn ). These clearly satisfy the symmetry condition µαβ = µ−1 βα on Uβα , and the transitivity condition µγβ µβα = µγα on Uγβα := Uγ ∩ Uβ ∩ Uα . Definition 11.6.2 (Spin structure). Let M be an oriented Riemannian manifold, with a bundle atlas for T M with transition maps µβα (p) ∈ SO(Rn ) as above. A spin structure for M is a choice of smooth functions qβα (p) ∈ Spin(Rn ) ⊂ 4ev Rn such that µβα (p)v = qβα (p)vqβα (p)−1 ,

v ∈ Tp M, p ∈ Uβα , α, β ∈ I,

−1 on Uβα , and the transitivity and that satisfy the symmetry condition qαβ = qβα condition qγβ qβα = qγα , on Uγβα .

Since the map Spin(Rn ) → SO(Rn ) from Proposition 4.1.9 is a covering map, there are lifts qβα such that µβα v = −1 qβα vqβα , and there are two possible such lifts, differing only in sign. Choosing −1 appropriate signs, we can always satisfy the symmetry condition qαβ = qβα . However, Proposition 4.1.9 shows only that qγβ qβα = ±qγα . To investigate when it is ˇ possible to choose signs so that the transitivity condition holds, there is a Cech cohomology criterion for existence of spin structures analogous to that for orientability in Proposition 11.2.10.

Proposition 11.6.3 (Second Stiefel–Whitney class). Let M be an oriented Riemannian manifold, with a bundle atlas for T M with transition maps µβα (p) ∈ SO(Rn ), and let qβα (p) ∈ Spin(Rn ), p ∈ Uβα as above. Consider the sheaf Z2 on the good ˇ cover {Uα } as in Example 11.2.9. Define the Cech 2-cochain f ∈ C 2 (U ; Z2 ) by ( 0, qαγ qγβ qβα = 1, f ({α, β, γ}) := 1, qαγ qγβ qβα = −1. ˇ Then ∂2 f = 0, and the Cech cohomology class [f ] does not depend on the choice of lifts qβα . There exists a spin structure for M if and only if there exists g ∈ C 1 (U ; Z2 ) such that ∂1 g = f , that is, if [f ] = [0].

415

11.6. Spinor Bundles

ˇ cohomology class w2 (M ) := [f ] is called the second Stiefel–Whitney The Cech class of M . Proof. A straightforward verification shows that f is a well-defined element of C 2 (U ; Z2 ), that is, symmetric with respect to permutations of α, β, γ. To show that ∂2 f ({α, β, γ, δ}) = f ({α, β, γ}) + f ({α, β, δ}) + f ({α, γ, δ}) + f ({β, γ, δ}) =0

mod 2,

it suffices to see that (qαγ qγβ qβα )(qαβ qβδ qδα )(qαδ qδγ qγα )(qβγ qγδ qδβ ) = 1, by the symmetry of f . Since the left-hand side is ±1, it suffices to show that the scalar part w0 of this multivector w is 1. Using that (u 4 v)0 = (v 4 u)0 , this follows −1 from the assumed symmetry condition qβα = qαβ . 0 If qβα are other choices of lifts, define ( g({α, β}) :=

0, 1,

0 qαβ = qαβ , 0 qαβ = −qαβ .

This is a well-defined element of C 1 (U ; Z2 ) because of the assumed symmetry 0 condition on qβα and qβα , and we see that f 0 ({α, β, γ}) = f ({α, β, γ}) + g({α, β} + g({α, γ}) + g({β, γ})

mod 2,

where f 0 ∈ C 2 (U ; Z2 ) is the 2-cochain for this new choice. Therefore f 0 = f + ∂1 g, 0 so that [f 0 ] = [f ]. Since there exists a spin structure for M if and only if lifts qβα 0 can be chosen such that f = 0, this completes the proof. We now define a normed spinor bundle 4M / , using a given spin structure {qβα }. Locally in each Uα , by fixing an ON-frame, we have a spinor bundle over Uα with fiber 4R / n and representation ρα (v)ψ = ρ0 (µ−1 (p)v) α over each p ∈ Uα , α ∈ I. In terms of the spin structure, the bundle transition maps for T M are µβα (p)v = qβα (p)vqβα (p)−1 ,

αβ ∈ I.

To define a fiber of a global spinor bundle at p ∈ M , let I p := {α ∈ I ; p ∈ Uα }

416

Chapter 11. Multivector and Spinor Bundles

and consider the set I p × 4R / n . The symmetry and transitivity conditions on qβα show that the relation (α, ψ) ∼ (β, φ), defined to hold if φ = qβα (p).ψ is an equivalence relation on I p × 4R / n . We define the fiber of the global spinor bundle at p ∈ M to be the set of equivalence classes (4M / )p := {[(α, ψ)] ; α ∈ I p , ψ ∈ 4R / n }. With fiber space 4R / n and bundle charts ψ ∈ 4R / n , p ∈ Uα , α ∈ I,

qα (p)ψ := [(α, ψ)],

/ globally over M . Furthermore, defining we obtain a well-defined vector bundle 4M the Tp M representation on (4M / )p to be ρ(v)[(α, ψ)] := [(α, ρα (v)ψ)],

v ∈ Tp M, ψ ∈ 4R / n , p ∈ Uα , α ∈ I,

/ as a spinor bundle over M . we obtain 4M To norm 4M / , we fix spinor inner products and conjugations on 4R / n as in Definition 5.3.3. Define a spinor inner product and conjugation on each fiber (4M / )p by ([(α, ψ)], [(α, φ)]i := (ψ, φi, †

[(α, ψ)] := [(α, ψ † )], for ψ, φ ∈ 4R / n , p ∈ Uα , α ∈ I. / Exercise 11.6.4. Verify the details of the above construction and show that 4M indeed is a well-defined normed spinor bundle over M . Concretely, show • that ∼ is an equivalence relation and that (4M / )p is a linear space in a natural way, • that qα (p) defines bundle charts for a vector bundle 4M / , • that ρ is well defined independent of α as a consequence of qβα being the rotors representing the rotations µβα , • that the spinor inner product and conjugation are well defined independent of α as a consequence of Rn rotors qβα being real multivectors acting isometrically on 4R / n. ˇ Each spin structure on M , represented by a Cech 1-cochain g ∈ C 1 (U ; Z2 ),

∂1 g = [f ] = [0],

417

11.6. Spinor Bundles

yields a normed spinor bundle 4M over M as above. Consider two spinor bundles / / 1 M and 4 / 2 M corresponding to g1 , g2 ∈ C 1 (U ; Z2 ) with ∂1 g1 = ∂1 g2 = 0. We 4 say that 4 / 1 M and 4 / 2 M are isomorphic as normed spinor bundles over M if there exists a section F ∈ C ∞ (M ; L(4 / 1M ; 4 / 2 M )) such that at each p ∈ M , the linear map F = F (p) : (4 / 1 M )p → (4 / 2 M )p is an isometry that commutes with spinor conjugation and F (ρ1 (v)ψ) = ρ2 (v)(F ψ),

v ∈ Tp M, ψ ∈ (4 / 1 M )p .

Proposition 11.6.5 (Isomorphic spinor bundles). Let M be an oriented Riemannian manifold, and consider two spinor bundles 4 / 1 M and 4 / 2 M corresponding to g1 , g2 ∈ C 1 (U ; Z2 ) with ∂1 g1 = ∂1 g2 = 0 as above. Then 4 / 1 M and 4 / 2 M are isomorphic as normed spinor bundles over M if and only if there exists h ∈ C 0 (U ; Z2 ) such that g1 = g2 + ∂0 h. ˇ This means that there is a one-to-one correspondence between the Cech co1 homology classes in H (U ; Z2 ) and isomorphism classes of spinor bundles over M , provided that w2 (M ) = [0]. Analogously to the remark following Theorem 10.6.5, we note that this shows in particular that the number of elements in H 1 (U ; Z2 ) does not depend on the choice of good cover, and that the number of isomorphism classes of spinor bundles over M does not depend on the choice of Riemannian metric, but only the global topology of M . Proof. Assume that g1 = g2 + ∂0 h. This means that 1 2 qβα = (−1)hα +hβ qβα , j writing qβα for the two spin structures, j = 1, 2, and hα := h({α}). Consider the linear maps 4R / n → 4R / n : ψ 7→ (−1)hα ψ

in Uα . These are seen to be compatible, that is, give well-defined maps (4 / 1 M )p → (4 / 2 M )p by inspection of the equivalence relations ∼1 and ∼2 , and yield an isomorphism of normed spinor spaces. Conversely, assume that there exists an isomorphism F ∈ C ∞ (M ; L(4 / 1M ; 4 / 2 M )). In Uα , consider Fα : 4R / n → 4R / n defined by [(α, Fα (ψ))] = F ([(α, ψ)]),

ψ ∈ 4R / n.

We verify that Fα is a spinor map induced by the identity on Rn , as in Proposition 5.3.5. By uniqueness, there exists hα ∈ Z2 such that Fα = (−1)hα I. At

418

Chapter 11. Multivector and Spinor Bundles

p ∈ Uβα , we have 2 [(β, (−1)hα qβα ψ)] = [(α, (−1)hα ψ)] = [(α, Fα ψ)] = F ([(α, ψ)]) 1 1 1 = F ([(β, qβα ψ)]) = [(β, Fβ qβα ψ)] = [(β, (−1)hβ qβα ψ)],

where the equivalence classes on the first line are with respect to ∼2 , and those 1 2 on the second line are with respect to ∼1 . It follows that qβα = (−1)hα +hβ qβα , so that g1 = g2 + ∂0 h. We next proceed to analysis on a fixed normed spinor bundle 4M / over a spin manifold M . Definition 11.6.6 (Spinor fields and frames). Sections of a normed spinor bundle 4M / over an oriented Riemannian manifold M are referred to as spinor fields on M . We mainly use Ψ, Φ to denote spinor fields. We use induced ON-frames for 4M / , constructed as follows. Fix the standard representation of Rn from Example 5.1.5. This means that we index the standard basis {ei } for Rn by −m ≤ i ≤ m, including i = 0 if n = 2m + 1 and excluding i = 0 if n = 2m. The spinor space 4R / n equals ∧Cm as a complex linear space, with basis {es } indexed by s ⊂ {1, . . . , m}. When viewing ∧Cm as 4R / n , we write /es := es . Given an ON-frame {ei } for T M , indexed as the Rn standard basis above, consider the isometry Rn → Tp M that maps the standard basis onto the frame. By Proposition 5.3.5, there is an isometry 4R / n → (4M / )p compatible with the representation and spinor conjugation, unique up to sign. We refer to either of these two ON-bases {/ es } corresponding to {/es } under these isometries as a spinor frame for 4M / induced by {ei }. By construction, frame es multivector fields act on induced frame e /t spinor fields on M , in the same way as the es act on /et in Rn . Note that s and t are subsets of different index sets as above. To avoid technicalities, we shall not write out these index sets {−m, . . . , m}, modulo 0, and {1, . . . , m}. Note that if 4M / is constructed from local ON bundle charts µα (p) for T M in Uα as above, then the spinor frame induced by the vector ON-frame {µα (p)ei } equals {[(α, /es )]}. The first step in setting up the calculus on 4M / is to identify a canonical covariant derivative on 4M / . As in Chapter 5, we shall abbreviate the action of multivectors on spinors by writing w.ψ := ρ(w)ψ,

w ∈ 4p Mc , ψ ∈ 4 / p M, p ∈ M.

Recall from Lemma 5.1.4 that the representation ρ extends to a complex algebra representation of 4Mc on 4M / . Proposition 11.6.7 (4M / covariant derivative). Let 4M / be a normed spinor bundle over a Riemannian manifold M . Then there exists a unique covariant derivative ∇v on 4M / that is

419

11.6. Spinor Bundles • compatible with the representation in the sense that ∇v (F.Ψ) = (∇v F ).Ψ + F.(∇v Ψ), where ∇v F denotes the Levi-Civita covariant derivative on 4Mc , • metric in the sense that ∂v (Ψ, Φi = (∇v Ψ, Φi + (Ψ, ∇v Φi, and †

• compatible with spinor conjugation in the sense that (∇v Ψ) = ∇v (Ψ† ), for all vectors v ∈ C(M ; T M ), multivector fields F ∈ C 1 (M ; 4Mc ), and all spinor / ). fields Ψ, Φ ∈ C 1 (M ; 4M e v are two covariant derivatives Proof. To prove uniqueness, assume that ∇v and ∇ e v − ∇v . Since with the desired properties, and let Lv := ∇ e v Ψ) − ((∂v f )Ψ + f ∇v Ψ) = f Lv Ψ, Lv (f Ψ) = ((∂v f )Ψ + f ∇ for scalar functions f and spinor fields Ψ, it follows that Lv is a linear map on each fiber (4M / )p . A similar subtraction of the identities assumed to hold for ∇ e yields the following. From Lv (F.Ψ) = F.(Lv Ψ) and the uniqueness result and ∇ in Theorem 5.2.3, it follows that Lv Ψ = λv Ψ for some λv ∈ C ∞ (M ; C). From † (λv Ψ, Φi + (Ψ, λv Φi = 0 it follows that Re λv = 0, and from (λv Ψ) = λv (Ψ† ) it e v = ∇v . follows that Im λ = 0. This proves that ∇ To show the existence of a covariant derivative with these properties, it suffices to consider the problem locally in a frame. Fix a vector ON-frame e in U ⊂ M , with induced spinor frame e /. Let Γe ∈ C ∞ (U ; L(T M ; ∧2 M )) be the Christoffel symbols for T M and define a covariant derivative X es + 21 Γe (v).Ψ ∇v Ψ := (∂v Ψs )/ s

P

of spinor fields Ψ = s Ψs e /s . To prove ∇v (F.Ψ) = (∇v F ).Ψ + F.(∇v Ψ), it suffices to consider multivector and spinor frame fields F = es and Ψ = e et } is /t . Since {/ an induced spinor frame, it is clear that es ./ et is ±1 times a spinor frame element. Therefore, by Proposition 11.5.3 we need to prove 1 et ) 2 Γe (v).(es ./

= 12 [Γe (v), es ]./ et + es .( 12 Γe (v)./ et ),

which is clear. Further we note that ∇v is a metric covariant derivative, since Γe (v) = −Γe (v), because Γe (v) ∈ 42 M , and it is compatible with spinor conjugation since Γe (v) is a real bivector field. This completes the proof. Proposition 11.6.8 (Bivector derivations). Let 4M / be a normed spinor bundle over a Riemannian manifold M , with the covariant derivative from Proposition 11.6.7, and let e = {ei (p)} be an ON-frame for T M in an open set U ⊂ M . Denote by Γe ∈ C ∞ (U ; L(∧1 M ; ∧2 M )) and R ∈ C ∞ (M ; L(∧2 M ; ∧2 M )) the Christoffel symbols and curvature operator for T M .

Chapter 11. Multivector and Spinor Bundles

420

/ es (p)} be the induced ON-frame for 4M Let e in U . Denote by ωe/ ∈ / = {/ / )) and Ω ∈ C ∞ (M ; L(∧2 M ; L(4M / ))) the Christoffel symC ∞ (U ; L(∧1 M ; L(4M / . Then bols and curvature operator for 4M ωe/ (v)Ψ = 21 Γe (v).Ψ, Ω(b)Ψ = 12 R(b).Ψ, / ). for v ∈ C(M ; T M ), b ∈ C(M ; ∧2 M ), and Ψ ∈ C(M ; 4M Proof. The result for the Christoffel symbols is contained in the proof of Proposition 11.6.7. Consider therefore the curvature operator. Defining e Ω(b)Ψ := Ω(b)Ψ − 12 R(b).Ψ, we have e e Ω(b)(F.Ψ) = 12 [R(b), F ].Ψ + F.(Ω(b)Ψ) − 12 R(b).(F.Ψ) = F.(Ω(b)Ψ), by a computation similar to the proof of (11.5). The uniqueness result in Thee e orem 5.2.3 shows that Ω(b)Ψ = λb Ψ. Since Ω(b) is skew-symmetric, we have Re λb = 0, and since it commutes with spinor conjugation, we have Im λb = 0. The fundamental first-order differential operator on M acting on sections of 4M / is the following natural generalization of the Euclidean nabla operator from Definition 9.1.3. Definition 11.6.9 (The Atiyah–Singer Dirac operator). Let 4M / be a normed spinor bundle over a Riemannian manifold M . The 4-Dirac / operator on 4M / is the operator X / := DΨ e∗i .(∇ei Ψ), Ψ ∈ C 1 (M ; 4M / ), i

where {ei } is a frame for T M , with dual frame {e∗i }. 2

/ to the The analogue of Proposition 11.5.9 is the following result relating D Laplace–Beltrami operator ∆ on 4M / . This is the key identity used in the proof of the Atiyah–Singer index theorem in Section 12.4. / Let 4M Proposition 11.6.10 (Weitzenb¨ock identity for D). / be a normed spinor bundle over a Riemannian manifold M , and let {ei } be an ON-frame for T M . Then X 1 2 / Ψ = −∆Ψ − −D ei ej .Ω(ei ∧ ej )Ψ = −∆Ψ + SΨ, 4 i<j / as in Deffor Ψ ∈ C 2 (M ; 4M / ), where ∆ is the Laplace–Beltrami operator on 4M inition 11.5.6, Ω is the curvature operator on 4M / , and S is the scalar curvature S of T M as in Definition 11.3.6.

11.7. Comments and References

421

Proof. Calculations identical to those in Proposition 11.5.9 lead to the first identity. From Propositions 11.6.8, and (11.7) in the proof of Proposition 11.5.9, we then obtain X i<j

11.7

ei ej .Ω(ei ∧ ej )Ψ =

1X Rijkl ei ej ek el .Ψ = − 14 SΨ. 8

ijkl

Comments and References

11.1–11.3 A reference for the differential geometry needed in this book, and more, is Taubes [90]. It is standard in differential geometry to identify tangent vectors v and directional derivatives ∂v . In coordinates {x1 , . . . , xn }, it is standard to denote the coordinate basis vectors {e1 , . . . , en } by {∂/∂x1 , . . . , ∂/∂xn }, and to denote the dual basis {e∗1 , . . . , e∗n } by {dx1 , . . . , dxn }. There is of course a canonical one-to-one correspondence v ↔ ∂v , but to identify v and ∂v as objects is not natural, and leads to serious notational problems. We therefore refrain from doing so. A main example is the nabla operators, whereP we define exterior and interior derivatives, and Dirac operators using ∇ = j ei ∂i in an ON-basis. It leads to obvious notational problems if we write the basis vector ei as a derivative ∂i , and also writing the dual basis covector e∗i as dxi causes problems for the inexperienced reader. We use the terminology Christoffel symbols in a more general sense than the standard usage in the literature. Normally, Christoffel symbols Γijk refer to the tangent bundle and a coordinate frame. We more generally refer to the zero-order part of the covariant derivative, in a fixed frame for a general vector bundle, as Christoffel symbols. A reference for the use of differential forms, or multicovector fields in our terminology, in differential geometry is Darling [32]. For the curvature operator, one should note the two different formalisms. Following the Cartan tradition, using differential form, we have Definition 11.3.3. Without using differential forms, the standard definition of the curvature operator is by Proposition 11.3.5. 11.4 Using an ON-frame, equations (11.3), and the equations defining curvature in Definition 11.3.3 provide a way to compute curvature. This is called the orthonormal moving frame method, and the equations are referred to as the structure equations. Some references are [96], [32], and [57]. The proof of Liouville’s theorem presented in Section 11.4 follows Flanders [39]. The methods we use require the C 3 hypothesis. However, it has been shown by Reˇsetnjak [77] that this can be weakened to C 1 only, or even local integrability of |∂i f |n , i = 1, . . . , n.

422

Chapter 11. Multivector and Spinor Bundles The conformal maps in dimension n ≥ 3 are indeed very scarce. However, by relaxing the conformal requirement and only requiring the ratio between the largest and smallest singular value of the total derivative to be uniformly bounded, one obtains a much richer class of maps: the quasiconformal maps.

11.5–11.6 Identities like those in Propositions 11.5.9 and 11.6.10, with integral versions as formulated in Section 12.1, or versions for domains with a boundary as in Theorem 10.3.6, are named after Roland Weitzenb¨ock. In the literature they are sometimes referred to as Lichnerowicz identities, in particular that / which Andr´e Lichnerowicz derived. Also related are Bochner identities, for D which are second-order analogues of these identities. In the literature, the standard approach to defining spinor bundles and / operator is via principal bundles. These are similar to vector the 4-Dirac bundles, but instead of having a family of vector spaces indexed by M , we now have a family of copies {Px }x∈M of a given group G. But in contrast to vector bundles, we do not have a fixed identity element in the fibers Px , but the group G acts freely and transitively from the right on each Px . The typical way that this construction is used in connection to vector bundles E is that the fiber Px collects all bases for Ex of a certain type, described by G, and the coordinates for vectors in Ex are described by a representation of G. For example, to construct spinor bundles starting from T M , we collect all positively oriented bases into a principal bundle P with G = SO(Rn ), / similar to Proposition 4.1.9, then lift this to a principal Spin(Rn ) bundle P by combining the frames implicit in and finally obtain a spinor bundle 4M / / with coordinates coming from the representation of Spin(Rn ) on 4R / n. P In this book, we have avoided principal bundles to minimize technicalities. However, one can show that the spinor bundles obtained from a / as above are precisely the normed spinor bundles principal spin bundle P considered in Section 11.6. In particular, they come equipped with a spinor inner product and a spinor conjugation, and the induced spinor frames from /. Definition 11.6.6 correspond to P Our discussion of spin structures follows Gilkey [43].

Chapter 12

Local Index Theorems Prerequisites: Chapter 11 should contain the material from differential geometry needed to read the present chapter. Section 12.1 builds on part of Chapter 10. Road map: Let M be a two-dimensional closed Riemannian manifold. The famous Gauss– Bonnet theorem states that Z 1 χ(M ) = S(p)dp, 4π M where S = 2R1212 denotes the scalar curvature at p ∈ M and χ(M ) = b0 (M ) − b1 (M ) + b2 (M ) is the Euler characteristic for M . Here bj (M ) are the Betti numbers for M , with the obvious generalization of Definition 7.6.3 to the compact manifold M without boundary. Hiding behind this result is the 4-Dirac operator D and the splitting L2 (M ; 4M ) = L2 (M ; 4ev M ) ⊕ L2 (M ; 4od M ). Indeed, D is a skew-adjoint operator on L2 (M ; 4M ) that swaps these two subspaces, and χ(M ) equals the index of the restriction D : L2 (M ; 4ev M ) → L2 (M ; 4od M ). To appreciate the power of the Gauss–Bonnet theorem, one must note that it relates three fundamentally different quantities, S, χ(M ), and D, where S is a local geometric quantity, χ is a global topological quantity, and D is an analytic object. © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8_12

423

Chapter 12. Local Index Theorems

424

In this chapter, we study two generalizations of the Gauss–Bonnet theorem. The first is the Chern–Gauss–Bonnet theorem, which is the direct generalization to higher-dimensional manifolds. In odd dimension the Euler characteristic vanishes due to Poincar´e/Hodge duality. In the interesting case of even dimension, the scalar curvature in the integrand is replaced by the Pfaffian Pf(R), a quantity pointwise derived from the Riemann curvature operator R. The second generalization is the Atiyah–Singer index theorem for the 4/ / For this, we consider a spinor bundle 4M / over a closed oriented Dirac operator D. even-dimensional Riemannian manifold M , and consider the splitting +

+

/ M ) ⊕ L2 (M ; 4 L2 (M ; 4M / ) = L2 (M ; 4 / M) into the chiral subspaces of right- and left-handed spinor fields, which are swapped / The Atiyah–Singer index theorem, which ranks among the very top achieveby D. ments in twentieth-century mathematics, states in particular that the index of the restriction + − / : L2 (M ; 4 D / M ) → L2 (M ; 4 / M) ˆ obtained pointwise from the Riemann equals an integral over M of a quantity A(R) curvature operator. It turns out that this index is nonzero only for manifolds of ˆ does not depend on the dimension divisible by four, and the integral of A(R) choice of spinor bundle. It may happen that the integral over some manifolds M is not be integer-valued, which is for example the case for complex projective plane CP 2 , which shows that this four-dimensional real manifold does not possess a spin structure. It is not, however, the purpose of this chapter to pursue the large amount of interesting applications of these results, which can be found in the extensive existing literature. Rather our goal is to demonstrate how our systematic buildup of the multivector and spinor theory makes advanced results in modern mathematics rather easily accessible. For both proofs, we use the well known heat equation method. For a self/ adjoint elliptic differential operator D, with applications D = iD and D = iD − in mind, the strategy is as follows. Write L2 = L+ 2 ⊕ L2 for the splitting in this abstract formulation. Then by definition, the index we consider is dim N(D|L+ ) − dim N(D|L− ). 2

2

Aiming at curvature and having the Weitzenb¨ock identities in mind, we note that N(D|L± ) = N(D2 |L± ) and consider the eigenvalues of the nonnegative self-adjoint 2 2 ± 2 operator D2 . Write λ± j for the eigenvalues of D on the invariant subspaces L2 . − + Here the nonzero eigenvalues λj are the same as the nonzero eigenvalues λj , as a consequence of D+ : L2+ → L2− and D− : L2− → L2+ being adjoint operators. We now apply a suitable function f : R → R by Borel functional calculus to the self-adjoint operator D2 = D− D+ ⊕D+ D− . Choosing f (0) = 1 and f to decay fast enough toward ∞, we obtain a trace-class operator f (D2 ), and the index of

425

12.1. Fredholm Dirac Operators D+ equals the difference of traces X X f (λ− f (λj+ ) − j ) = Trf (D− D+ ) − Trf (D+ D− ). j

j

2

In the heat equation method, we choose f (λ) = e−tλ with a parameter t > 0, which yields trace-class operators by the Weyl asymptotics λj± ≈ j 2/n on an n2

dimensional manifold. The solution operators e−tD to the heat equation ∂t f + D2 f = 0 become local and converge to the identity in a suitable sense as t → 0+ . To prove the index theorems, we need to work with multivector and spinor calculus on M to identify the limit of trace differences lim+ Tr(e−tD− D+ ) − Tr(e−tD+ D− ) . t→0

Note that the existence of the limit is trivial, since the trace difference is independent of t. Sections 12.1 and 12.2 contain material on L2 Dirac operators and charts, preliminary to the proofs of the index theorems. Highlights: • The Chern–Gauss–Bonnet theorem: 12.3.1 • The Atiyah–Singer index theorem: 12.4.1

12.1 Fredholm Dirac Operators Throughout this chapter, we consider a closed compact Riemannian manifold M without boundary. In this section we study the L2 properties of the Dirac oper/ on M , analogously to our study of d and δ on affine domains in ators D and D Chapter 10. However, now our main work concerns the geometry of the manifold M rather than the boundary ∂M = ∅. In particular, no boundary conditions are needed for the operators. Consider first the multivector operators D, d, and δ. By transferring the Euclidean result to M using pullbacks and pushforwards as in Exercise 11.2.7, it follows that d and −δ defined on C ∞ (M ; ∧M ) are formally adjoint operators. From this it follows that D = d + δ, defined on C ∞ (M ; 4M ), is formally skew-adjoint. Using the covariant nabla expression for D from Definition 11.5.4, we can also verify this directly on M as follows. Localizing with a partition of unity, we may assume that F, G ∈ C0∞ (U ; 4M ) in an open set U ⊂ M , in which we have an ON-frame {ei } for T M . Define the vector field X v := hei 4 F, Giei . i

Chapter 12. Local Index Theorems

426 We compute div v =

X

h(∇ei ei ) 4 F, Gi + hei 4 ∇ei F, Gi + hei 4 F, ∇ei Gi

i

+

X hei 4 F, Gihej , ∇ej ei i.

(12.1)

ij

P P Since i ei hej , ∇ej ei i = − i ei hei , ∇ej ej i = −∇ej ej in the last term, this canRcels the first term. The skew-adjointness now follows from the divergence theorem div v dp = 0 from Exercise 11.2.11. M A similar calculation, replacing 4 by ∧, gives a second proof of the fact that −δ is a formal adjoint of d on M . The following extension of the domains of operators in the natural distributional sense yields closed and densely defined linear operators in L2 (M ; 4M ). Definition 12.1.1 (L2 operators). Let M be a closed Riemannian manifold. Consider the equation Z hF 0 , Gi + hF, DGi dp = 0. (12.2) M

The domain D(d) of the exterior derivative in L2 (M ; ∧M ) is the set of F ∈ L2 (M ; ∧M ) for which there exists F 0 ∈ L2 (M ; ∧M ) such that (12.2), with D = δ, holds for all G ∈ C ∞ (M ; ∧M ). For F ∈ D(d), we define dF := F 0 . The domain D(δ) of the interior derivative in L2 (M ; ∧M ) is the set of F ∈ L2 (M ; ∧M ) for which there exists F 0 ∈ L2 (M ; ∧M ) such that (12.2), with D = d, holds for all G ∈ C ∞ (M ; ∧M ). For F ∈ D(δ), we define δF := F 0 . The domain D(D) of the Dirac operator in L2 (M ; 4M ) is the set of F ∈ L2 (M ; 4M ) for which there exists F 0 ∈ L2 (M ; 4M ) such that (12.2), with D = D, holds for all G ∈ C ∞ (M ; 4M ). For F ∈ D(D), we define DF := F 0 . In the absence of boundary ∂M , we can prove full Sobolev H 1 regularity of fields F ∈ D(D) following the same route as in Section 10.3. We start with the integral form of the Weitzenb¨ock identity. Proposition 12.1.2 (Integral Weitzenb¨ock identity). Let M be a closed Riemannian manifold. Then Z Z 2 2 |DF |2 dp (|dF | + |δF | )dp = M M Z X Z Z X 1 1 hRijkl ei ej F ek el , F idp, (12.3) = |∇ei F |2 dp + S|F |2 dp + 4 M 8 M M i ijkl

for all F ∈ H 1 (M ; ∧M ). Here {ei } is an ON-frame for T M , and R and S denote the Riemann curvature operator and scalar curvature for T M .

427

12.1. Fredholm Dirac Operators

Proof. Assume first that F ∈ C 2 (M ; 4M ). Then the result follows from the nilpotence and formal adjointness of d and −δ, and duality from Propositions 11.5.5 and 11.5.9. Since all terms are continuous in H 1 norm, a limiting argument finishes the proof. Proposition 12.1.3 (Regularity). Let M be a closed Riemannian manifold. Then D = d + δ with D(D) = D(d) ∩ D(δ) = H 1 (M ; 4M ) and equivalences of norms. Proof. Clearly H 1 (M ; 4M ) ⊂ D(d) ∩ D(δ) ⊂ D(D), so it suffices to show that D(D) ⊂ H 1 (M ; 4M ). To this end, we argue as in the proof of Theorem 10.3.3, but replacing the ball B by the n-torus T n . (i) Consider first the case M = T n with the flat metric given by the constant standard inner product on Rn . This is the manifold T n = Rn /Zn , with charts obtained from quotient maps Rn → T n in the natural way. A multivector field on T n corresponds to a Zn -periodic field on Rn , and D(D) = H 1 (T n ; ∧T n ) follows from Plancherel’s theorem and the Fourier series analogue of Example 10.1.8. (ii) Next consider a general closed Riemannian manifold M , and F ∈ D(D). Using a partition of unity, we may assume that supp F is contained in the range Mα ⊂ M of a coordinate chart µα : Dα → Mα . Assuming that Dα ⊂ Rn is small, 1 (x) on T n such we identify Dα with a subset of T n , and define a C ∞ metric gij that µα : Dα → Mα is an isometry. In this way, we may regard Mα as an open subset of the n-torus 1 T n with geometry determined by the metric gij . 1 To show regularity of F , we perturb the metric gij on T n continuously to the flat Euclidean metric δij , by letting t 1 gij (x) := (1 − t)δij + tgij (x),

0 ≤ t ≤ 1.

t It is clear that gij defines a Riemannian metric on T n for each 0 ≤ t ≤ 1. Write Ttn for this Riemannian manifold, and note that all Ttn are the same as C ∞ manifolds, but their Riemannian geometries are distinct. Consider the family of bounded linear operators I + Dt : H 1 (Ttn ; 4Ttn ) → L2 (Ttn ; 4Ttn ), (12.4)

where Dt is the Dirac operator on 4Ttn . By Proposition 12.1.2, these are all injective semi-Fredholm maps with k∇ ⊗ F k2 + kF k2 ≈ kDF k2 + kF k2 = kF + DF k2 , using the formal skew-adjointness of Dt for the last equality. We want to show that these are all invertible maps. This is clear for I + D0 from (i). To apply the

428

Chapter 12. Local Index Theorems

method of continuity for semi-Fredholm operators, as explained in Section 6.4, we define auxiliary maps of multivector fields At : C(T0n ; 4T0n ) → C(Ttn ; 4Ttn ). Since the Ttn are all equal to T n as C ∞ manifolds, we can realize the multivector bundle 4Ttn as T n × ∧Rn , where ∧Rn is defined independently of any metric as in Section 2.1, before equipping each fiber with the inner product int . We therefore let At be the identity map on T n × ∧Rn , but with duced by gij different but equivalent metrics in the domain and range. In particular we obtain bounded and invertible linear maps At : L2 (T0n ; 4T0n ) → L2 (Ttn ; 4Ttn ) and At : H 1 (T0n ; 4T0n ) → H 1 (Ttn ; 4Ttn ), and the method of continuity applies to n n 1 n n A−1 t (I + Dt )At : H (T0 ; 4T0 ) → L2 (T0 ; 4T0 ).

It follows that I + D1 : H 1 (T1n ; 4T1n ) → L2 (T1n ; 4T1n ) is invertible. Since H 1 ⊂ D(D) on T1n and I +D is injective on D(D) by formal skew-adjointness, this shows that H 1 = D(D). This completes the proof. Recall the definition (6.4) of adjointness in the sense of unbounded operators. Proposition 12.1.4 (Duality). Let M be a closed Riemannian manifold. Then d and −δ, with domains as in Definition 12.1.1, are adjoint operators in L2 (M ; ∧M ) in the sense of unbounded operators. The Dirac operator D, with domain as in Definition 12.1.1, is a skew-adjoint operator in L2 (M ; 4M ) in the sense of unbounded operators. Proof. Consider first d∗ = −δ. As in Proposition 10.2.3, it is clear from Definition 12.1.1 that it suffices to show that for every F ∈ D(d), there exists a family of fields Ft ∈ C ∞ (M ; ∧M ) such that Ft → F and dFt → dF in L2 (M ; ∧M ) as t → 0+ . By localizing with a partition of unity, we may assume that F is supported in a chart Mα . Since d commutes with the pullback µ∗α , the result follows from the Euclidean case in Section 10.2. To prove D∗ = −D, we similarly note that it suffices to show that for every F ∈ D(D), there exists a family of fields Ft ∈ C ∞ (M ; 4M ) such that Ft → F and DFt → DF in L2 (M ; 4M ) as t → 0+ . In this case, we obtain from Proposition 12.1.3 that F ∈ H 1 (M ; 4M ). This completes the proof, since C ∞ is dense in H 1 . From Propositions 10.1.2 and 10.1.6, we now obtain the Hodge decomposition L2 (M ; ∧M ) = R(d) ⊕ C(D) ⊕ R(δ) of L2 multivector fields on M , with finite-dimensional cohomology space C(D) = N(d) ∩ N(δ), closed subspaces R(d) and R(δ) of exact and coexact fields, and compact potential maps.

429

12.1. Fredholm Dirac Operators In terms of the Dirac operator, this means that D : H 1 (M ; 4M ) → L2 (M ; 4M )

is a Fredholm operator. Since it is a skew-adjoint L2 operator, its index is zero. Concretely, N(D) = R(D)⊥ = C(D). Splitting the fields further into homogeneous k-vector fields, we write C(D; ∧k ) := C(D) ∩ L2 (M ; ∧k M ) and define the Betti numbers bk (D) := dim C(D; ∧k ),

k = 0, 1, 2, . . . , n.

As explained in the introduction, we are particularly interested in the following integer. Definition 12.1.5 (Euler characteristic). Let M be a closed Riemannian manifold. The Euler characteristic of M is the alternating sum X (−1)k bk (M ) χ(M ) := k

of Betti numbers, or equivalently the index of the restricted Dirac operator D : H 1 (M ; 4ev M ) 7→ L2 (M ; 4od M ). Exercise 12.1.6. Compute the three Betti numbers b0 (M ), b1 (M ), and b2 (M ) for the two-dimensional sphere S 2 as well as the two-dimensional torus T 2 = S 1 × S 1 , using Hodge star maps and the Gauss–Bonnet theorem stated in the introduction. Note that there exists a flat metric on T 2 , that is, a metric for which the curvature operator vanishes. Show also that there exists no flat metric on S 2 . / acting on sections of a given Next consider the spinor Dirac operator D normed spinor bundle 4M / over an oriented closed Riemannian manifold M . For a general real Riemannian manifold, without any further complex structure, we / in terms of some invariantly defined nilpotent operators Γ and Γ∗ , cannot write D analogously to D = d + δ for the 4-Dirac operator. But besides this, we have an / with similar properties to those of D above. Some details are as L2 operator D follows. / is formally • A calculation like (12.1), replacing ei 4 F by ei .Ψ, shows that D / ). skew-adjoint on C ∞ (M ; 4M / from C ∞ to D(D) / • Similar to Definition 12.1.1, we extend the domain of D / ) for which there exists Ψ0 ∈ consisting of spinor fields Ψ ∈ L2 (M ; 4M L2 (M ; ∧M ) such that Z / (Ψ0 , Φi + (Ψ, DΦi dp = 0 M

/ := Ψ0 . / we define DΨ / ). For Ψ ∈ D(D), holds for all Φ ∈ C (M ; 4M ∞

Chapter 12. Local Index Theorems

430

• From Propositions 11.5.5 and 11.6.10, we readily obtain an integral Weitzenb¨ock identity Z X Z Z 1 |∇ei Ψ|2 dp + |DΨ|2 dp = S|Ψ|2 dp, 4 M M i M / ). Here {ei } is an ON-frame and S is the scalar valid for Ψ ∈ H 1 (M ; 4M curvature for T M . / = H 1 (M ; 4M • Also for the 4-Dirac / operator, we have D(D) / ). However, some care concerning the map At is needed in adapting the perturbation argument in the proof of Proposition 12.1.3. On each n-torus Ttn we have an ON-frame {ei } for the tangent bundle globally defined on Ttn , for example by polar decomposition as in Definition 12.2.4. Upon mapping this onto the standard basis for Rn , we use the trivial spinor bundle 4T / tn = Ttn × 4R / n over Ttn . Note that there is no problem with topological obstructions here, since the problem is local. Defining At as the identity map on T n × 4R / n, but with different metrics on the domain and range, the proof proceeds as for D. / is a skew-adjoint operator in • From H 1 regularity it follows that D L2 (M ; 4M / ) in the sense of unbounded operators, and that / : H 1 (M ; 4M D / ) → L2 (M ; 4M / ) / = R(D) / ⊥, is a Fredholm operator. Since it is skew-adjoint, we have N(D) and in particular, its index is zero. • In contrast to the 4-Dirac operator, here we do not have access to a finer splitting into subspaces like L2 (M ; 4k M ) and a notion of Betti numbers, but assuming that the dimension n = 2m of M is even, we have the pointwise splitting into chiral subspaces. Let en be the unit n-vector field on M describing the orientation of M , and define the main reflector wn := i−m en as in Definition 5.2.1. Consider the pointwise orthogonal splitting +

−

L2 (M ; 4M / ) = L2 (M ; 4 / M ) ⊕ L2 (M ; 4 / M ), ±

where (4 / M )p are the ranges of the projections wn± = 12 (1 ± wn ), at each p ∈ M. ± / swaps the subspaces L2 (M ; 4 • It follows from Definition 11.6.9 that D / M ), since the ei swap them, while the ∇v preserve them, since the Γe (v) do so. This is so because the vectors ei anticommute with wn , while the bivectors Γe (v) commute with wn . Our goal in Section 12.4 is to calculate the index of the restricted 4-Dirac / operator + − / : H 1 (M ; 4 D / M ) → L2 (M ; 4 / M ).

12.2. Normal Coordinates

12.2

431

Normal Coordinates

Let M be a closed Riemannian manifold, with tangent bundle T M , Levi-Civita covariant derivative ∇v , and Riemann curvature operator R. Fix a base point q ∈ M . To do computations on M near q it is useful to choose a chart for M around that q is as good as possible. We use the exponential map expq : Tq M → M , which is the map taking a tangent vector v ∈ Tq M to the point p ∈ M at distance d(p, q) := |v| from q along the geodesic, that is, the length-minimizing curve γ, starting at γ(0) = q with tangent vector γ 0 (0) = v/|v| and parametrized by arc length. Definition 12.2.1. A normal chart for M around q is a chart µ : D → M, with D = B(0, r) ⊂ Rn , obtained by fixing an ON-basis for Tq M and identifying it with Rn , and applying the exponential map Tq M → M . The supremum over δ such that r ≥ δ can be chosen at each q ∈ M is called the injectivity radius of M . We start by formulating a condition for a chart to be normal, in terms of the metric gij := hei , ej i, where we write ei for the coordinate frame vector fields µ(ei ), by slight abuse of notation. In such a normal chart, we write ωij for the Christoffel symbols. Proposition 12.2.2 (Normal chart equations). Let µ : D → M be a chart such that µ(0) = q, with metric coordinates gij (x). Then µ is a normal chart if and only if X xi = gij (x)xj , x ∈ D, i = 1, . . . , n. j

Proof. That radial lines are geodesic is equivalent to ∇µ

x

(x) (µx (x/|x|))

x ∈ D,

= 0,

and that they are parametrized by arc length means that hµx (x/|x|), µx (x/|x|)iµ(x) = 1. In coordinates, the second equation reads X X gij (x)xi xj = x2i , (12.5) ij

i

P whereasP the first equation becomes ∂x (x/|x|) + ij ei hωij , xixj /|x| = 0, or equivalently j hωij , xixj = 0 for all i. Using Exercise 11.1.8, this reads X (2∂i gjk (x) − ∂k gij (x))xi xj = 0. (12.6) ij

Chapter 12. Local Index Theorems

432

P Under (12.5), equation (12.6) is seen to be equivalent to ij xi ∂i (gjk (x)xj −xk ) = 0. Interpreting this as the radial derivatives of gjk (x)xj − xk vanishing, the result follows. We next show that for normal coordinates, all second derivatives of the metric at q are given by the curvature coefficients. Proposition 12.2.3 (Metric Taylor expansion). Let µ : D → M be a normal chart for M at q, with metric gij (x). Then 1X gij (x) = δij − Rikjl (q)xk xl + O(|x|3 ), i, j = 1, . . . , n, 3 kl

where Rijkl (q) are the Riemann curvature coefficients in the coordinate basis {ei }, which is ON at q. In particular, gij = δij , ∂i gjk = 0, ∂k hωij , el i = −∂j ∂l gik , ∂i ∂j gkl + ∂i ∂k gjl + ∂j ∂k gil = 0, ∂i ∂j gkl = ∂k ∂l gij , Rijkl = ∂i ∂l gjk − ∂j ∂l gik ,

∂k ∂l gij = − 13 (Rikjl + Riljk ),

hold at q, for all i, j, k, l. Here the ωij denote the Christoffel symbols in the coordinate frame {ei }. Proof. We prove the identities at q, from which the stated Taylor expansion follows. Differentiating xi = gim (x)xm three times gives δij = (∂j gim )xm + gij , 0 = (∂k ∂j gim )xm + ∂j gik + ∂k gij , 0 = (∂l ∂k ∂j gim )xm + ∂k ∂j gil + ∂l ∂j gik + ∂l ∂k gij . At q, the first equation gives gij = δij , using the second equation three times gives ∂i gjk = −∂j gik = ∂k gij = −∂i gkj , thus ∂i gjk = 0, and the third equation gives ∂k ∂j gil + ∂l ∂j gik + ∂l ∂k gij = 0. Using this last equation three times gives ∂i ∂j gkl = −∂i ∂k gjl − ∂k ∂j gil = (∂k ∂l gji + ∂i ∂l gjk ) + (∂j ∂l gik + ∂k ∂l gij ) = 2∂k ∂l gij − ∂i ∂j gkl , thus ∂i ∂j gkl = ∂k ∂l gij . For the Christoffel symbols, we get from Exercise 11.1.8 that X 2h∇ei ej , ek i = 2 gαk hωαj , ei i = ∂i gjk + ∂j gik − ∂k gij . α 1 2 ∂k (∂l gji

At q, this gives ∂k hωij , el i = + ∂j gli − ∂i glj ) = 12 (−∂j ∂l gki − ∂k ∂i glj ) = −∂j ∂l gik . This gives curvature coefficients Rijkl = hdωij , ek ∧ el i = ∂k hωij , el i − ∂l hωij , ek i = −∂j ∂l gik + ∂i ∂l gjk ,

12.2. Normal Coordinates

433

using Exercise 7.4.9. Finally, we have Rikjl +Riljk = (∂i ∂l gkj −∂k ∂l gij )+(∂i ∂k glj − ∂l ∂k gij ) = −∂k ∂l gij − 2∂k ∂l gij = −3∂k ∂l gij . This completes the proof. Besides the coordinate frame {ei }, we require an ON-frame {ei } in which to do multivector calculus. We shall use the following construction. Definition 12.2.4 (Polar ON-frame). Let µ be a normal chart for M around q, with coordinate frame {ei }. Let G = (gij ) be the metric, with gij = hei , ej i, and consider its positive inverse square root G−1/2 = (αij )ij . The polar ON-frame for µ is the ON-frame {ei }, where X ei := αki ek = µ(G−1/2 ei ). k

This is a pointwise construction of {ei } from {ei } based on polar factorization as in Proposition 1.4.4. Note that {ei } is ON at q, whereas {ei } is an ON-frame in all of the chart. Indeed, hei , ej i = hµ(G−1/2 ei ), µ(G−1/2 ej )i = hG−1/2 ei , GG−1/2 ej )iRn = δij . Proposition 12.2.5. Let µ : D → M be a normal chart for M at q, with metric gij (x). Denote the Christoffel symbols and Riemann curvature coefficients in the associated polar ON-frame {ei } by Γij and Rijkl respectively. Then the identities Γij = 0,

∂k hΓij , el i = 12 Rijkl ,

hold at q for all i, j, k, l. Proof. It follows from Proposition 12.2.3 that G = (gij ) = I + O(|x|2 ). Thus the change-of-basis matrix from {ei } to {ei } is A := G−1/2 = I − 12 (G − I) + O(|x|4 ). By Proposition 11.3.4, Γij = (A−1 )ik (ωkl Alj + dAkj ) = ωij − 12 dgij + O(|x|2 ), where the {ωij } denote the Christoffel symbols in the coordinate frame {ei }. Using Proposition 12.2.3, we get at q that ∂k hΓij , el i = ∂k hΓij , el i = −∂j ∂l gik − 12 ∂k ∂l gij = −∂j ∂l gik − 12 (−∂j ∂l gik − ∂j ∂k gil ) = 12 (−∂j ∂l gik + ∂j ∂k gil ) = 12 Rijkl .

Note that at q, the curvature coefficients Rijkl are the same in the frames {ei } and {ei }, since these coincide there.

434

12.3

Chapter 12. Local Index Theorems

The Chern–Gauss–Bonnet Theorem

In this section, we prove the following local index theorem for the 4-Dirac operator. Recall from Definition 12.1.5 the relation between the Euler characteristic χ(M ), the Betti numbers bk (M ), and the index of D restricted to even multivector fields. Theorem 12.3.1 (Chern–Gauss–Bonnet). Let M be an n-dimensional compact and closed Riemannian manifold. If n = 2m is even, then 1 m Z χ(M ) = hPf(R), dpi, 2π M where Pf(R) denotes the Pfaffian of the Riemann curvature operator R. If n is odd, then χ(M ) = 0. We begin by explaining Pf(R). First we replace the antisymmetric matrix R = (Rij ) of bivectors by a scalar antisymmetric matrix A = (Aij ), where Aij ∈ R. Here Rij := R(ei ∧ ej ) ∈ ∧2 M in an ON-frame {ei }, or equivalently, by symmetry of R, the Rij are the bivectors from Definition 11.3.3. Definition 12.3.2 (Pfaffian). Let the dimension n = 2m be even, let A ∈ SO(Rn ) be an antisymmetric matrix, and let b ∈ ∧2 Rn be the bivector that represents A as in Proposition 4.2.3. Then the Pfaffian of A is Pf(A) := hb ∧ · · · ∧ b, e1 ∧ · · · ∧ en i/(n/2)!, where the first exterior product for b is m-fold. The Pfaffian behaves like a square root of the determinant for skew-symmetric matrices, as the following shows. Proposition 12.3.3 (Pfaffian algebra). Let the dimension n = 2m be even. We have the following properties of the Pfaffian functional of a skew-symmetric matrix A. (i) For A in standard form as in Proposition 4.3.6(ii), we have

0 −x1 Pf ... 0 0

x1 0 .. .

··· ··· .. .

0 0 .. .

0 0

··· ···

0 −xm

0 0 .. .

= x1 · · · xm . xm 0

435

12.3. The Chern–Gauss–Bonnet Theorem (ii) For general skew-symmetric A = (Aij ), we have the formula X 1 Pf(A) = m (i1 , j1 , . . . , im , jm )Ai1 j1 · · · Aim jm , 2 m! i ,j ,...,i ,j 1

1

m

m

where (i1 , j1 , . . . , im , jm ) = hei1 ∧ ej1 ∧ · · · ∧ eim ∧ ejm , e1 ∧ · · · ∧ en i is the sign of the permutation (i1 , j1 , . . . , im , jm ) → (1, . . . , n). (iii) If A ∈ SO(Rn ) and T ∈ L(Rn ), then Pf(T AT ∗ ) = det(T )Pf(A). Proof.P(i) and (ii) follow from the facts that b = x1 e12 + · · · + xm e2m−1,2m and b = 12 ij Aij eij in these cases respectively. To prove (iii), we use Proposition 2.7.1 to obtain T AT ∗ v = T (b x (T ∗ v)) = (T b) x v. Thus T b represents T AT ∗ , and (T b) ∧ · · · ∧ (T b) = T (b ∧ · · · ∧ b) = det(T )(n/2)!Pf(A)e1 ∧ · · · ∧ en , which proves (iii).

Definition 12.3.4 (Exterior Pfaffian). Let M be a Riemannian manifold of dimension n = 2m, with Riemann curvature operator R. Define the Pfaffian X 1 Pf(R) = m (i1 , j1 , . . . , im , jm )Ri1 j1 ∧ · · · ∧ Rim jm ∈ ∧n M. 2 m! i ,j ,...,i ,j 1

1

m

m

The generalization from A to R amounts to replacing the real field R by the algebra (∧ev Rn , ∧), which is commutative by Proposition 2.1.14. Proposition 12.3.3(iii) generalizes to show that Pf(R) is independent of the choice of positively oriented ON-frame. Lemma 12.3.5. The integrand hPf(R), dpi in Theorem 12.3.1 does not depend on the choice of ON-frame or orientation. P ˜i = Proof. Consider two ON-frames {ei } and {˜ ei }, related as e j ej αji . By ˜ ij = Proposition 11.3.4, the corresponding curvature coefficients are related as R P k,l αki Rkl αlj . This gives X ˜i j ∧ · · · ∧ R ˜i j (i1 , j1 , . . . , im , jm )R 1 1 m m i1 ,j1 ,...,im ,jm

=

X

X

hei1 αk1 i1

∧ ej1 αl1 j1 ∧

e1 ∧ · · · ∧ en i Rk1 l1 =

X

hA∗ (ek1

∧ el1 ∧

· · · ∧ ekm

∧

∧ ejk αlm jm ,

X k1 ,l1 ,...,km ,lm

Rk1 l1

∧

· · · ∧ Rkm lm

∧ elm ), e1 ∧

k1 ,l1 ,...,km ,lm

= det(A)

· · · ∧ eim αkm im

i1 ,j1 ,...,im ,jm

k1 ,l1 ,...,km ,lm

· · · ∧ Rkm lm ,

· · · ∧ en i Rk1 l1

∧

· · · ∧ Rkm lm

Chapter 12. Local Index Theorems

436

where A = (αij ). This, together with the observation that the oriented measure dp also changes sign if det(A) = −1, proves the proposition. 0 R12 , so that Example 12.3.6 (Gauss–Bonnet). If n = 2, then R = −R12 0 hPf(R), e12 i = R1212 = 12 S. Thus the Gauss–Bonnet theorem stated in the introduction to this chapter is the two-dimensional case of the Chern–Gauss–Bonnet theorem (Theorem 12.3.1). Exercise 12.3.7. Write down explicitly the Chern–Gauss–Bonnet integrand in dimension n = 4. We now embark on the proof of the Chern–Gauss–Bonnet theorem, which covers the remainder of this section. When n is odd, consider the Hodge star map ∗ : L2 (M ; 4k M ) → L2 (M ; 4n−k M ). This gives an isomorphism between Hk (M ) and Hn−k (M ) by Exercise 11.2.7. When n is odd, this Poincar´e duality implies that χ(M ) = 0. We next consider the nontrivial case of even dimension n = 2m. Following the heat equation method described in the introduction, we calculate χ(M ) = dim N(D|L2 (M ;4ev M ) − dim N(D|L2 (M ;4od M ) ) = dim N(D2 |L2 (M ;4ev M ) − dim N(D2 |L2 (M ;4od M ) ) 2

2

= Tr(etD |L2 (M ;4ev M ) ) − Tr(etD |L2 (M ;4od M ) ), for all t > 0. The second identity is valid because D is a normal operator, and the last identity follows from the general fact that the nonzero eigenvalues of operators A∗ A and AA∗ are the same, in particular for A = D : L2 (M ; 4ev M ) → L2 (M ; 4od M ). The idea of the proof is to compute this trace difference in the limit as t → 0+ . According to Proposition 11.5.9, the square D2 of the Dirac operator differs from the Laplace–Beltrami operator only by zero-order terms. Thus it is 2 reasonable to expect the operator etD to be an integral operator resembling the solution operator Z 2 t∆ e f (x) = (4πt)−n/2 e−|x−y| /(4t) f (y)dy Rn n

for the heat equation on R from Example 6.3.3. With this in mind, we make an ansatz of the form Z N X 2 (Ht f )(p) := (4πt)−m e−d(p,q) /(4t) tk H k (p, q)f (q)dq, (12.7) M

k=0

where N < ∞ is to be chosen. Here d(p, q) denotes the shortest distance between points p and q on M , and below, we shall choose linear maps H k (p, q) ∈

437

12.3. The Chern–Gauss–Bonnet Theorem

L(4(Tq M ); 4(Tp M )) depending smoothly on p, q ∈ M . We want to choose H k 2 such that Ht f approximates etD f well for small t, and in particular, H 0 (q, q) = I, 2

for all q ∈ M . This will ensure that limt→0+ Ht f = f = limt→0+ etD f for all 2 2 f . Secondly, since ∂t etD f = D2 etD f , we want (∂t − D2 )Ht f to be as small as possible. Lemma 12.3.8. Let µ : D → M be a normal chart for M around q = µ(0). Push forward the radial vector field x in D to the vector field rq (p) := µx (x), p = µ(x), and let dq (p) := |x| denote the shortest distance from p to q on M . Then ∇(d2q ) = 2rq

∆(d2q ) = 2n + ∂rq ln g,

and

where g = det(gij ) and ∆ is the scalar Laplace–Beltrami operator on M , and 2 2 (∂t − D2 )(tk−m e−dq /(4t) h) = tk−m e−dq /(4t) t−1 ∇rq + 14 ∂rq (ln g) + k h − D2 h , for all h ∈ C 2 (M ; 4M ). P Proof. Define the frame e∗i = j g ij ej dual to the coodinate frame {ei }, where (g ij ) denotes the inverse of the metric (gij ). Proposition 12.2.2 shows that X X X ∇(d2q ) = e∗i ∂i |x|2 = g ij ej 2xi = 2 xj ej = 2rq . i

ij

j

To compute ∆(d2q ), we use Exercise 11.5.8 to get ∆(d2q ) =

√1 g

X

√ ∂i ( gg ij 2xj ) =

√2 g

ij

X

√ ∂i ( gxi ) = 2n + ∂rq ln g.

i

For the last formula, clearly 2

2

∂t (tk−m e−dq /(4t) ) = (d2q /(4t2 ) + (k − m)/t)tk−m e−dq /(4t) . By Proposition 11.5.9, it suffices to prove the identity with D2 replaced by the Laplace–Beltrami operator on 4M . We compute X 2 2 2 2 ∆(e−dq /(4t) h) = (∆e−dq /(4t) )h + 2 (∂ei e−dq /(4t) )∇ei h + e−dq /(4t) ∆h, i

in an ON-frame {ei }. Here 2

∆e−dq /(4t) =

√1 g

X ij

d2 2n + ∂rq ln g −d2 /(4t) 2 √ q ∂i ( gg ij ∂j e−dq /(4t) ) = − e q 4t2 4t

Chapter 12. Local Index Theorems

438 and 2

X

2

ei (∂ei e−dq /(4t) ) = −

i

2 2 1 X ei (∂ei d2q )e−dq /(4t) = −t−1 e−dq /(4t) rq . 2t i

Combining these calculations proves the stated formula.

Applying ∂t − D2 to (12.7), we obtain (∂t − D2 )Ht f Z N X 1 −d2q /(4t) = ˙ q)f (q) e tk−m−1 ∇rq + 14 ∂rq (ln g) + k H k (p, m (4π) M k=0

−

N +1 X

tk−m−1 D2 H k−1 (p, ˙ q)f (q) dq.

k=1

This leads us to the following recursive definition of H k (p, q). For p, q ∈ M e k (p, q) such such that d(p, q) < δ, where δ is the injectivity radius of M , define H 0 k e e that H (q, q) = I and H (p, q) solves k e (p, e k−1 (p, ∇rq + 14 ∂rq (ln g) + k H ˙ q)f (q) = D2 H ˙ q)f (q) (12.8) e −1 (p, q) = 0. Note that for each k, for k = 0, . . . , N and f (q) ∈ (4M )q . Here H (12.8) is an ordinary differential equation along the geodesic from q to p, and that e k (q, q)f (q) = k −1 D2 H e k−1 (q, q)f (q) is specified, since for k ≥ 1, the initial value H rq (q) = 0. Existence theory for ordinary differential equations shows that this uniquely determines maps H k (p, q) ∈ L(4(Tq M ); 4(Tp M )) depending smoothly on p, q ∈ M , at distance < δ. To extend this construction to general p, q ∈ M , we make a smooth cutoff as follows. Let η ∈ C ∞ (R) be such that η(x) = 1 for x < δ/3 and η(x) = 0 for x > 2δ/3. Define e k (p, q), H k (p, q) := η(d(p, q))H where we understand that H k (p, q) = 0 if d(p, q) ≥ 2δ/3. 2 We have constructed Ht and next compare this ansatz to etD . Let Z Kt f (p) = Kt (p, q)f (q)dq := (∂t − D2 )Ht f (p). M

Here the kernel Kt (p, q) is a smooth function of p, q ∈ M and t > 0, and by construction we have 2

Kt (p, q) = −(4π)−m tN −m e−d(p,q)

/(4t)

D2 H N (p, ˙ q) 2

when d(p, q) < δ/3. Consider the difference Ht f − etD f , which satisfies 2

(∂t − D2 )(Ht f − etD f ) = Kt f

12.3. The Chern–Gauss–Bonnet Theorem

439

2

and initial conditions limt→0+ (Ht f − etD f ) = 0. Integration gives Ht f − e

tD2

t

Z

2

e(t−s)D Ks f ds,

f=

(12.9)

0

from which we deduce the following. Proposition 12.3.9 (Trace formula). We have the formula Z 1 m m ev M (H χ(M ) = (q, q)) − Tr (q, q)) dq Tr od M (H 4 4 (4π)m M for the index of D : H 1 (M ; 4ev M ) 7→ L2 (M ; 4od M ). Proof. We estimate the trace norm of (12.9). See Example 6.4.8. To estimate the trace norm kKs kL1 (L2 (M ) , we factorize into Hilbert–Schmidt operators Ks = (I + D)−j (I + D)j Ks . We use that the eigenvalues of D grow in size as |λk | = k 1/n , and as a consequence (I + D)−j is a Hilbert–Schmidt operator if j > m = n/2. For the second operator and d(p, q) < δ/3, we have |(I +D)j Ks (p, ˙ q)| . sN −m−j , whereas for d(p, q) ≥ δ/3, 2 we have |(I + D)j Ks (p, ˙ q)| . e−δ /(36s) s−m−j−2 . This shows that the Hilbert– Schmidt norm of (I + D)j Ks is bounded by sN −m−j for 0 < s < 1. Choosing N > n therefore shows that kKs kL1 (L2 (M ) . 1, and in particular, 2

kHt − etD kL1 (L2 (M )) ≤

Z

t

2

ke(t−s)D kL(L2 (M )) kKs kL1 (L2 (M )) ds . t → 0,

0

as t → 0+ , since D is a skew-adjoint operator, and trace-class operators form an ideal in L(L2 ). Since the trace functional is continuous in the trace norm, we get 2 2 χ(M ) = lim Tr(etD |L2 (M ;4ev M ) ) − Tr(etD |L2 (M ;4od M ) ) t→0+ = lim+ Tr(Ht |L2 (M ;4ev M ) ) − Tr(Ht |L2 (M ;4od M ) ) t→0

N

X 1 = lim tk−m (4π)m t→0+ k=0

Z

Tr4ev M (H k (q, q)) − Tr4od M (H k (q, q)) dq.

M

Since we know that the limit exists, all the terms 0 ≤ k < m must be zero, and we have proved the stated formula for χ(M ). It remains to compute Tr4ev M (H m (q, q)) − Tr4od M (H m (q, q)).

Chapter 12. Local Index Theorems

440

To this end, we fix a normal chart µ : x 7→ µ(x) around q and let {ei } be the associated polar ON-frame from Definition 12.2.4. To handle linear operators in L(∧M ), we use the frame − − {es+1 · · · e+ sk et1 · · · etl }s1 0

Equating the tm terms yields H m (0, 0) = im p0 (a1 , . . . , am ), where p0 is the polynomial from Definition 12.4.2. This coincides with the result for H0m from the k−1 k−1 recurrence (β + k)Hβk = D2 Hβ+2 + D0 Hβ−2 . (ii) By Definition 12.4.2, we have H m (0, 0) = im p(TrA2 , TrA4 , . . .).

448

Chapter 12. Local Index Theorems

By analytic continuation, this continues to hold for all aj ∈ C. Furthermore, it continues to hold for general A ∈ SO(Rn ), by changing basis to an orthogonal splitting for A as in Proposition 4.3.6 and using the invariance of the trace functional. Similarly, using instead the commutative algebra (∧ev Rn , ∧), it is clear that the recurrence (12.12) with D1 = 0 will result in H m (q, q) = im p(TrR2 , TrR4 , . . .). Using Exercise 12.4.5, we obtain that the index of + − / : L2 (M ; 4 D / M ) → L2 (M ; 4 / M)

is

1 (4π)m

Z

(2i)m im hp(TrR2 , TrR4 , . . .), dqi.

M

Since we have shown in Exercise 12.4.7 that the index vanishes when m is odd, this completes the proof of the Atiyah–Singer index theorem. To quote a famous mathematician: I think I’ll stop here.

12.5

Comments and References

12.3 The Gauss–Bonnet theorem goes back to Gauss, who, however, never published it. Bonnet first published a special case of the theorem in 1848. The full Chern–Gauss–Bonnet theorem for a general compact manifold was first proved by Chern in 1945. 12.4 The index theorem for general elliptic differential operators was proved by Atiyah and Singer in 1963, and is regarded as one of the great landmarks of twentieth-century mathematics. A main special case of this general index theorem is the index theorem for the 4-Dirac / operator. In fact, this operator was rediscovered by Atiyah and Singer in their work on the index theorem. The early proofs of the index theorem used different methods. The now dominant heat equation method, which we use, originates in the works of Atiyah, Bott, and Patodi [3] and Gilkey. Standard references for the index theory of Dirac operators include the books by Gilkey [43], Berline, Getzler, and Vergne [20] and Lawson and Michelsohn [63]. Further treatments of the index theory can be found in Taylor [92] and Gilbert and Murray [42]. The proofs of the index theorems given here follow the book [96] by Yanlin Yu, and they do not rely on the theory of pseudodifferential operators. Some minor variations on the setup from [96] include our use of the pointwise constructed polar ON-frame from Definition 12.2.4 and our use of skew-adjoint Dirac operators rather than self-adjoint Dirac operators. By a Wick-type rotation argument it is straightforward to see that this does not

12.5. Comments and References

449

affect the index of the operator. A reference for results on differential equations that we use, including Weyl’s law and existence results for ordinary ˆ differential equations, is Taylor [91, 92]. The local indices Pf(R) and A(R), the integrands appearing in the Chern–Gauss–Bonnet and Atiyah–Singer index theorems, are examples of what are called characteristic classes, which are certain polynomials in the curvature bivectors. A reference for the theory of symmetric polynomials, used in Definition 12.4.2, is Nicholson [73].

Bibliography [1] Ahlfors, L. V. M¨obius transformations in Rn expressed through 2 × 2 matrices of clifford numbers. Complex Variables Theory Appl. 5, 2-4 (1986), 215–224. [2] Arnold, D., Falk, R., and Winther, R. Finite element exterior calculus, homological techniques, and applications. Acta Numer. 255, 15 (2006), 1–155. [3] Atiyah, M., Bott, R., and Patodi, V. K. On the heat equation and the index theorem. Invent. Math. 19 (1973), 279–330. [4] Auscher, P., Axelsson, A., and Hofmann, S. Functional calculus of Dirac operators and complex perturbations of Neumann and Dirichlet problems. J. Funct. Anal. 255, 2 (2008), 374–448. [5] Auscher, P., Axelsson, A., and McIntosh, A. Solvability of elliptic systems with square integrable boundary data. Ark. Mat. 48 (2010), 253– 287. [6] Auscher, P., Hofmann, S., Lacey, M., McIntosh, A., and Tchamitchian, P. The solution of the Kato square root problem for second order elliptic operators on Rn . Ann. of Math. (2) 156, 2 (2002), 633–654. [7] Axelsson, A. Oblique and normal transmission problems for Dirac operators with strongly Lipschitz interfaces. Comm. Partial Differential Equations 28, 11-12 (2003), 1911–1941. [8] Axelsson, A. Transmission problems for Dirac’s and Maxwell’s equations with Lipschitz interfaces. PhD thesis, The Australian National University, 2003. Available at https://openresearch-repository.anu.edu.au/handle/1885/46056. [9] Axelsson, A. Transmission problems and boundary operator algebras. Integral Equations Operator Theory 50, 2 (2004), 147–164. [10] Axelsson, A. Transmission problems for Maxwell’s equations with weakly Lipschitz interfaces. Math. Methods Appl. Sci. 29, 6 (2006), 665–714. © Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8

451

452

Bibliography

[11] Axelsson, A., Grognard, R., Hogan, J., and McIntosh, A. Harmonic analysis of Dirac operators on Lipschitz domains. In Clifford analysis and its applications (Prague, 2000), vol. 25 of NATO Sci. Ser. II Math. Phys. Chem. Kluwer Acad. Publ., Dordrecht, 2001, pp. 231–246. [12] Axelsson, A., Keith, S., and McIntosh, A. Quadratic estimates and functional calculi of perturbed Dirac operators. Invent. Math. 163, 3 (2006), 455–497. [13] Axelsson, A., Kou, K., and Qian, T. Hilbert transforms and the Cauchy integral in euclidean spaces. Studia Math. 193, 2 (2009), 161–187. [14] Axelsson, A., and McIntosh, A. Hodge decompositions on weakly Lipschitz domains. In Advances in analysis and geometry, Trends Math. Birkh´’auser, Basel, 2004, pp. 3–29. [15] Axler, S. Down with determinants! Amer. Math. Monthly 102, 2 (1995), 139–154. [16] Axler, S., Bourdon, P., and Ramey, W. Harmonic function theory. No. 137 in Graduate Texts in Mathematics. Springer-verlag, 1992. ´n, A. Riesz continuity of the [17] Bandara, L., McIntosh, A., and Rose Atiyah-Singer Dirac operator under perturbations of the metric. Math. Ann. 370, 1-2 (2018), 863–915. ´n, A. Riesz continuity of the Atiyah-Singer [18] Bandara, L., and Rose Dirac operator under perturbations of local boundary conditions. To appear in Communications in Partial Differential Equations, 2019, DOI: 10.1080/03605302.2019.1611847. [19] Benn, I., and Tucker, R. An introduction to spinors and geometry with applications in physics. Adam Hilger, Ltd., 1987. [20] Berline, N., Getzler, E., and Vergne, M. Heat kernels and Dirac operators. No. 298 in Grundlehren der Mathematischen Wissenschaften. Springerverlag, 1992. [21] Bott, R., and Tu, L. Differential forms in algebraic topology. No. 82 in Graduate Texts in Mathematics. Springer-Verlag, 1982. [22] Bourguignon, J.-P. Spinors, Dirac operators, and changes of metrics. In Differential geometry: geometry in mathematical physics and related topics (Los Angeles, CA, 1990), vol. 54 of Proc. Sympos. Pure Math. Amer. Math. Soc., 1993, pp. 41–44. [23] Brackx, F., Delanghe, R., and Sommen, F. Clifford Analysis. No. 76 in Research Notes in Mathematics. Pitman, 1982. ¨ cker, T., and tom Dieck, T. Representations of compact Lie groups. [24] Bro Graduate Texts in Mathematics. Springer-Verlag, 1985.

Bibliography

453

[25] Cartan, E. Sur certaines expressions diff´erentielles et le probl`eme de Pfaff. Ann. Sci. Ecole Norm. Sup. 3, 16 (1899), 239–332. [26] Cartan, E. The theory of spinors. The M.I.T. Press, Cambridge, Mass., 1967. [27] Clifford, W. Applications of Grassmann’s Extensive Algebra. Amer. J. Math. 1, 4 (1878), 350–358. [28] Coifman, R. R., McIntosh, A., and Meyer, Y. L’int´egrale de Cauchy d´efinit un op´erateur born´e sur L2 pour les courbes lipschitziennes. Ann. of Math. (2) 116, 2 (1982), 361–387. [29] Colton, D., and Kress, R. Integral equation methods in scattering theory, first ed. John Wiley & Sons, New York, 1983. [30] Colton, D., and Kress, R. Inverse acoustic and electromagnetic scattering theory, second edition ed. Springer-Verlag, Berlin, 1998. [31] Costabel, M., and McIntosh, A. On Bogovskiˇi and regularized Poincar´e integral operators for the de Rham complexes on Lipschitz domains. Math. Z. 265, 2 (2010), 297–320. [32] Darling, R. Differential forms and connections. Cambridge University Press, 1994. [33] Delanghe, R., Sommen, F., and Souvcek, V. Clifford algebra and spinorvalued functions. A function theory for the Dirac operator. Mathematics and its Applications. Kluwer Academic Publishers Group, 1992. ´, J. The tragedy of Grassmann. Linear and multilinear algebra [34] Dieudonne 8, 1 (1979/80), 1–14. [35] Evans, L. Partial differential equations, vol. 19 of Graduate Studies in Mathematics. American Mathematical Society, 1998. [36] Evans, L., and Gariepy, R. Measure theory and fine properties of functions. Studies in Advanced Mathematics. CRC Press, 1992. [37] Fearnley-Sander, D. Hermann Grassmann and the creation of linear algebra. Amer. Math. Monthly 86, 10 (1979), 809–817. [38] Federer, H. Geometric measure theory. Die Grundlehren der mathematischen Wissenschaften, Band 153. Springer-Verlag, 1969. [39] Flanders, H. Liouville’s theorem on conformal mapping. J. Math. Mech. 15 (1966), 157–161. [40] Fulton, W. Algebraic topology. No. 153 in Graduate Texts in Mathematics. Springer-verlag, 1985. [41] Fulton, W., and Harris, J. Representation theory. A first course. No. 129 in Graduate Texts in Mathematics. Springer-verlag, 1991.

454

Bibliography

[42] Gilbert, J., and Murray, M. Clifford algebras and Dirac operators in harmonic analysis. Cambridge Studies in Advanced Mathematics. Cambridge University Press, 1991. [43] Gilkey, P. Invariance theory, the heat equation, and the Atiyah-Singer index theorem. No. 11 in Mathematics Lecture Series. Publish or Perish, Inc., 1984. [44] Grassmann, H. Die Lineale Ausdehnungslehre, ein neuer Zweig der Mathematik. 1844. andig und in strenger Form [45] Grassmann, H. Die Ausdehnungslehre: Vollst¨ bearbeitet. 1864. [46] Greub, W. Multilinear algebra. Second edition. Universitext. SpringerVerlag, 1864. [47] Griffiths, P., and Harris, J. Principles of algebraic geometry. Pure and Applied Mathematics. Wiley-Interscience, 1978. [48] Grisvard, P. Elliptic problems in nonsmooth domains. Monographs and Studies in Mathematics. Pitman, 1985. [49] Grove, L. Classical groups and geometric algebra. No. 39 in Graduate Studies in Mathematics. American Mathematical Society, 2002. [50] Harvey, F. R. Spinors and calibrations. No. 9 in Perspectives in Mathematics. Academic Press, Inc., 1990. [51] Helsing, J. Solving integral equations on piecewise smooth boundaries using the RCIP method: a tutorial. Available at https://arxiv.org/abs/1207.6737v9. [52] Helsing, J., and Karlsson, A. On a Helmholtz transmission problem in planar domains with corners. J. Comput. Phys. 371 (2018), 315–332. [53] Helsing, J., and Karlsson, A. Physical-density integral equation methods for scattering from multi-dielectric cylinders. J. Comput. Phys. 387 (2019), 14–29. obius differential geometry. [54] Hertrich-Jeromin, U. Introduction to M¨ No. 300 in London mathematical society lecture notes series. Cambridge university press, 2003. [55] Hestenes, D. Space-time algebra. Gordon and Breach, 1966. [56] Hestenes, D. New foundations for classical mechanics. No. 99 in Fundamental Theories of Physics. Kluwer Academic Publishers Group, 1999. [57] Hestenes, D., and Sobczyk, G. Clifford algebra to geometric calculus. A unified language for mathematics and physics. Fundamental Theories of Physics. D. Reidel Publishing Co., 1984.

Bibliography

455

[58] Hitchin, N. The Dirac operator. In Invitations to geometry and topology, Oxf. Grad. Texts Math. Oxford Univ. Press, Oxford, 2002, pp. 208–232. [59] Hladik, J. Spinors in physics. Graduate Texts in Contemporary Physics. Springer-Verlag, 1999. [60] Jancewicz, B. Multivectors and Clifford algebra in electrodynamic. No. 11 in Mathematics Lecture Series. World Scientific Publishing Co., Inc., 1988. [61] Kato, T. Perturbation theory for linear operators, second ed. SpringerVerlag, Berlin, 1976. Grundlehren der Mathematischen Wissenschaften, Band 132. [62] Kress, R. Linear integral equations. No. 82 in Applied Mathematical Sciences. Springer-Verlag, New York, 1999. [63] Lawson, H. B., J., and Michelsohn, M.-L. Spin geometry. No. 38 in Princeton Mathematical Series. Princeton University Press, 1989. [64] Lounesto, P. Clifford algebras and spinors. London Mathematical Society Lecture Note Series. Cambridge University Press, 2001. [65] Madsen, I., and Tornehave, J. From calculus to cohomology. de Rham cohomology and characteristic classes. No. 11 in Mathematics Lecture Series. Cambridge University Press, 1997. [66] McIntosh, A. Clifford algebras and the higher-dimensional Cauchy integral. In Approximation and function spaces (Warsaw, 1986), vol. 22 of Banach Center Publ. PWN, Warsaw, 1989, pp. 253–267. [67] McIntosh, A., and Mitrea, M. Clifford algebras and Maxwell’s equations in Lipschitz domains. Math. Methods Appl. Sci. 22, 18 (1999), 1599–1620. [68] McIntosh, A., and Monniaux, S. Hodge-Dirac, Hodge-Laplacian and Hodge-Stokes operators in Lp spaces on Lipschitz domains. Rev. Mat. Iberoam. 34, 4 (2018), 1711–1753. [69] Meyer, Y. Wavelets and operators. No. 37 in Cambridge Studies in Advanced Mathematics. Cambridge University Press, 1992. [70] Mitrea, D., Mitrea, M., and Monniaux, S. The Poisson problem for the exterior derivative operator with Dirichlet boundary condition in nonsmooth domains. Commun. Pure Appl. Anal. 7, 6 (2008), 1295–1333. [71] Mitrea, M. Clifford Wavelets, Singular Integrals and Hardy Spaces. No. 1575 in Lecture Notes in Mathematics. Springer, 1994. [72] Morrey, C.B., J. Multiple integrals in the calculus of variations. No. 130 in Die Grundlehren der mathematischen Wissenschaften. Springer-Verlag, 1966. [73] Nicholson, W. K. Introduction to abstract algebra. Wiley-Interscience. John Wiley & Sons, 2007.

456

Bibliography

[74] Picard, R. An elementary proof for a compact imbedding result in generalized electromagnetic theory. Math. Z. 187, 2 (1984), 151–164. [75] Porteous, I. Clifford algebras and the classical groups. No. 50 in Cambridge Studies in Advanced Mathematics. Cambridge University Press, 1995. [76] Pryde, A. J. The five lemma for Banach spaces. Proc. Amer. Math. Soc. 65, 1 (1977), 37–43. [77] Resetnjak, J. G. Liouville’s conformal mapping theorem under minimal regularity hypotheses. Sibirsk. Mat. 8 (1967), 835–840. [78] Riesz, M. Clifford numbers and spinors. With the author’s private lectures to E. Folke Bolinder. No. 54 in Fundamental Theories of Physics. Kluwer Academic Publishers Group, 1993. [79] Rindler, W. Relativity. Special, general, and cosmological. Oxford University Press, 2006. ´n, A. Boosting the Maxwell double layer potential using a right spin [80] Rose factor. To appear in Integral Equations and Operator Theory. ´n, A. Fredholm theory, singular integrals and T b theorems. Unpublished [81] Rose lecture notes from 2011, available at http://www.math.chalmers.se/∼rosenan/FST.html. ´n, A. Layer potentials beyond singular integral operators. Publ. Mat. [82] Rose 57, 2 (2013), 429–454. ´n, A. Square function and maximal function estimates for operators [83] Rose beyond divergence form equations. J. Evol. Equ. 13, 3 (2013), 651–674. ´n, A. A spin integral equation for electromagnetic and acoustic scat[84] Rose tering. Appl. Anal. 96, 13 (2017), 2250–2266. [85] Schwarz, G. Hodge decomposition - a method for solving boundary value problems. No. 1607 in Lecture Notes in Mathematics. Springer-Verlag, 1995. [86] Sommen, F. Spingroups and spherical means. In Clifford algebras and their applications in mathematical physics (Canterbury, 1985), vol. 183 of NATO Adv. Sci. Inst. Ser. C Math. Phys. Sci. Reidel, Dordrecht, 1986, pp. 149–158. [87] Sommen, F. Spingroups and spherical means. II. In Proceedings of the 14th winter school on abstract analysis (Srni, 1986), no. 14. Rend. Circ. Mat. Palermo (2) Suppl., 1987, pp. 157–177. [88] Sommen, F. Spingroups and spherical means. III. In Proceedings of the Winter School on Geometry and Physics (Srni, 1988), no. 21. Rend. Circ. Mat. Palermo (2) Suppl., 1989, pp. 295–323. [89] Stein, E., and Weiss, G. On the theory of harmonic functions of several variables. I. The theory of H p -spaces. Acta Math. 103 (1960), 25–62.

Bibliography

457

[90] Taubes, C. Differential geometry. Bundles, connections, metrics and curvature. No. 23 in Oxford Graduate Texts in Mathematics. Oxford University Press, 2011. [91] Taylor, M. Partial differential equations. I. Basic theory. No. 115 in Applied Mathematical Sciences. Springer-Verlag, 1996. [92] Taylor, M. Partial differential equations. II. Qualitative studies of linear equations. No. 116 in Applied Mathematical Sciences. Springer-Verlag, 1996. [93] Thaller, B. The Dirac equation. Springer-Verlag, 1992.

Texts and Monographs in Physics.

[94] Verchota, G. Layer potentials and regularity for the Dirichlet problem for Laplace’s equation in Lipschitz domains. J. Funct. Anal. 59, 3 (1984), 572–611. [95] Watson, G. N. A Treatise on the Theory of Bessel Functions. Cambridge University Press, New York, 1944. [96] Yu, Y. The index theorem and the heat equation method. No. 2 in Nankai Tracts in Mathematics. World Scientific Publishing Co., 2001.

Index ∗w, 54 :=, =:, 2 A∗ , 17 At , 17 E ± , 277 Ek N BVP, 314 Ek S BVP, 315 Ek± , 308 H k (p, q), 436 Jρ , 188 R, 403 Tp M , 387 V ∗, 6 V∞ , 135 Wl , 10 Ws , 10 Wt , 10 [v1 , v2 ], 236 , 197 ∆, 197, 411 Γ(z), 192 Γe , 390 Γk , 328 H, 84 Ω, 401 Psh , 266 Psm , 266 Ψ fundamental solution, 260 spinor field, 289 Ψ± k , 304 Θ∗, 54 ≈, 186 4M , 408 4V , 79

z c , 17 χ(M ), 429 b , 109 4V b 4V (2), 139 4, 76 ∂k , 368 curl F , 214 δΘ = ∇ y Θ, 212 DF = ∇ 4 F , 257 D, 410 div F , 214 D(T ), 202 x, ˙ 215 V, 135 ei , 385 (s, t), 32 grad f , 214 G(T ), 202 b k V , 34 ∧ ˆ 442 A, w, b 33 [w], 35 bwc, 36 y, 52 ↔, 2 ., 186 µ, 385 ρx , 187 C⊥ (D), 244 Ck (D), 244 Fvt , 235 L(V1 ; V2 ), 13, 205 Lv F , 236 | · |, 5, 201

© Springer Nature Switzerland AG 2019 A. Rosén, Geometric Multivector Analysis, Birkhäuser Advanced Texts Basler Lehrbücher, https://doi.org/10.1007/978-3-030-31411-8

459

460 ∇, 211 ∇ ⊗ F , 211 ∇v , 388 n, 27 N(T ), 202 ωe , 388 O(V ), 110 ⊕, 4 dwe, 36 ⊗, 14 k · k, 201 ∂i , 187 ∂v , 387 Pf, 434 π1 (M ), 118 Pin(V ), 110 R(T ), 202 x, 52 dx, 45 w, 33 ρ∗ Θ, 217 ρ∗ F , 217 /es , 418 4M / , 413 4V / , 163 h·, ·i, 5, 18 h·, ·iV , 304 / 420 D, / = ∇.Ψ, 289 DΨ e /s , 418 σn−1 , 194 SO(3), 86 SO(V ), 110 SO+ (W ), 128 ψ † , 168 Spin(V ), 110 Spin+ (W ), 128 hvi2 , 5 (·, ·i, 18 SU(2), 175 ρ˜∗ F , 217 ×, 58 Tr, 205 d, 351

Index δ, 351 SO(V ), 113 Spin(V ), 113 ∧, 27 ∧M , 390 ∧V , 30 ∧ev V , 30 ∧od V , 30 ∧k V , 27 ∧k (M ), 46 bk (D), 244 dF = ∇ ∧ F , 212 e∞ , 135 e∗i , 7 es , 27 f (x) ∗ g(x), 192 gij , 386 k-covector, 47 k-form, 225 linear, 225 k-vector, 27 v ± , 95 w.ψ := ρ(w)ψ, 163 adjoint formal, 203 operator, 16 unbounded operators, 203 affine space, 2 algebra associative, 4 completeness, 165 alternating, 24 alternating forms, 50 Amp`ere-Maxwell law, 293 anticommutation, 63, 76 antiparticles, 300 Atiyah–Singer–Patodi BVP, 324 Banach space, 200 Betti numbers, 244 Bianchi identities, 404 bilinear, 14 boundary orientation, 228

461

Index bundle Clifford, 408 multivector, 390 spinor, 413 vector, 385 BVP=boundary value problem, 309 Cartan formula, 238 lemma, 33 method, 405 Cartan–Dieudonn´e theorem, 107 Cauchy extension, 279 principal value integral, 277 singular integral, 277 Cauchy–Pompeiu formula, 260 Cauchy–Riemann equations, 258 celestial sphere, 143 character, 174 charge conjugation, 302 chart, 385 normal, 431 normalized pushforward, 391 pullback, 391 pushforward, 391 chiral, 300, 319 Christoffel bivectors, 390 symbols, 388 class function, 175 Clifford algebra, 90 algebra standard, 79 cone, 109 product, 75 trace, 87 closed, 240 closed range theorem, 206 coboundary operator, 368 cochain, 368 cofactor formula, 62 cohomology space, 373 Coifman–McIntosh–Meyer theorem, 279

commutation theorem, 221 commutative diagram, 370 complementary subspace, 203 complex, 369 complex eigenvector, 21 complex spinor space, 163 complex structure, 19 complexification, 20 composite, 34 conformal map, 134 cosine operator, 311 covariant derivative, 388 Cramer rule, 42 curvature operator, 401 coefficients, 404 commutator, 402 Ricci, 404 Riemann, 403 curvilinear coordinates, 222 d’Alembertian, 197 degenerate, 9 diagram chasing, 371 differential form, 253 Dirac equation, 298 Dirac matrix, 160 Dirac operator, 257 4, 287 4 on manifold, 410 4, / 289 Atiyah–Singer, 420 harmonic, 287 spherical, 274 wave, 287 Dirichlet problem, 249 domain C k , 188 Lipschitz, 188 starshaped, 190 strongly Lipschitz, 189 double covering, 112 double layer potential, 313 dual basis, 7

462 operator, 16 space, 6 duality, 5 elliptic rotation, 129 Euclidean, 5 Euler characteristic, 429 even Clifford subalgebra, 92 exact, 240, 344 expansion, 64 exponential function, 5 extension maps, 354 exterior algebra, 30 derivative, 212 product, 28 Faraday law, 293 fine sheaf, 369 five lemma, 372 flow, 235 Fourier transform, 192 fractional linear map, 135 frame, 385 polar ON, 433 fundamental group, 118 Gaffney inequality, 358 Gauss law, 293 magnetic, 293 Gegenbauer polynomial, 273 good cover, 374 Gram–Schmidt, 65 Grassmann cone, 34 Haar measure, 174 Hankel function, 196 Hardy projection, 278 subspace, 278 harmonic oscillator, 200 heat equation, 198 Helmholtz decomposition, 247 Helmholtz equation, 199, 306

Index Hermitian, 18 Hilbert space, 200 Hodge decomposition, 245 Rn , 349 abstract, 345 algebraic, 349 boundary, 331 Hodge projection, 347 Hodge star, 54 Hodge-Dirac operator, 346 homogeneous, 30 homotopy relations, 242, 364 hyperbolic rotation, 129 ideal, 164 induced basis, 27 induced map, 40 multivectors, 40 spinors, 171 inner product, 5 complex bilinear, 18 sesquilinear, 18 inner space, 36 interior derivative, 212 interior product left, 52 right, 52 inversion, 75, 134 involution, 33 isometry, 106 Jacobian, 188 K¨ unneth formula, 379 Kadlec formula, 357 Kelvin transform, 264 Lagrange identity, 74 Laplace operator, 197 spherical, 274 Laplace–Beltrami operator, 411 Legendre polynomial, 273 Levi-Civita covariant derivative, 389 Lie derivative, 236

463

Index bracket, 236, 389 light cone, 10 light-like, 10 linear space, 2 Liouville theorem conformal map, 406 monogenic field, 262 Lorentz boost, 125 M¨obius map, 137 main reflector, 161 manifold atlas, 190 chart, 190 closed, 191 oriented, 190 transition map, 190 with boundary, 191 Maxwell equations, 295 spacetime, 296 time-harmonic, 307 Mayer–Vietoris sequence, 379 Mehler formula, 200 method of continuity, 206 metric, 386 MIT bag model, 319 monogenic field, 257 two-sided, 259, 327 multicovector, 47 multilinear, 14 nabla symbol, 211 Neumann problem, 251 neutral Clifford algebra, 94 nilpotent, 213, 344 normal multivector, 66 normalized pushforward, 217 normed spinor space, 169 ON-basis, 8 operator compact, 204 Fredholm, 205

Hilbert–Schmidt, 204 semi-Fredholm, 205 trace class, 205 orientation, 44 oriented manifold, 395 measure, 46, 226 volume element, 45 orthochronous rotation, 128 rotor, 128 orthogonal complement, 7 outer space, 36 parabolic rotation, 126 parallel section, 398 transport, 398 partition of unity, 191 Pauli matrix, 101, 159 Peter–Weyl theorem, 174 pfaffian, 434 ucker relations, 67 Pl¨ Poincar´e theorem, 240 polar decomposition, 16 potential, 240 Bogovski˘ı, 362 Hodge, 246, 362 maps, 347 Poincar´e, 362 principal curvature, 357 principle of abstract algebra, 1 projection, 204 projective space, 143 pullback, 217 pushforward, 217 quaternion, 84 Rademacher theorem, 189 radiation condition Dirac, 305 Silver–M¨ uller, 308 Sommerfield, 307

464 real structure, 21 rectangular determinant, 28 reflection operator, 310 Rellich estimates, 320 Rellich lemma, 199 representation algebra, 156 group, 154, 173 irreducible, 173 minimal, 158 real Clifford algebras, 98 standard, 157 uniqueness, 162 vector, 156 reversion, 33 Riemann function, 193 Riemannian manifold, 387 Riesz formulas, 76 rotation operator, 311 rotor, 110 scalar curvature, 404 Schr¨ odinger equation, 200 Schur estimate, 202 Schur lemma, 173 second fundamental form, 357 section, 385 sheaf, 368 signature, 9 simple, 34 simply connected, 118 singular, 5 skew symmetric map, 114 snake lemma, 370 Sokhotski–Plemelj jump, 281 space-like, 10 spacetime, 10 spectral theorem, 203 sphere paradox, 146 spherical harmonics, 266 spherical monogenics, 266 spin, 86 group, 110 integral equation, 324

Index structure, 414 spinor abstract space, 154 conjugation, 168, 302 inner product, 168, 298 splitting, 185, 203 Hardy subspace, 282 Stein–Weiss vector field, 259 stereographic projection, 135 Stiefel–Whitney class first, 396 second, 415 Stokes theorem, 228 Cauchy formula, 232 classical, 230 Gauss formula, 231 Green formula, 231 Stratton–Chu formulas, 308 symmetric polynomial, 449 tangent bundle, 387 tangential multivector, 66 tensor abstract space, 154 product, 14 time-harmonic function, 18 time-like, 10 total derivative, 187 trace, 16, 205 transition map, 385 transmission problem, 326, 336 transversal magnetic, 339 ultrahyperbolic space, 117 universal property, 14, 26, 90 universality, 91 Vahlen cone, 139 Vahlen matrix, 139 vector product, 58 vector space, 2 volume-norm, 44 wave equation, 198

Index weakly singular integral, 202, 313 Weingarten map, 357 Weitzenb¨ock identity C 2 domain, 357 4M , 412 4M / , 420 integral 4M , 426 integral 4M / , 430 well-posedness, 311 Weyl representation, 161 zonal harmonic, 271 zonal monogenic, 271

465

Our partners will collect data and use cookies for ad personalization and measurement. Learn how we and our ad partner Google, collect and use data. Agree & close