file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=cover.html
Cover
file:///D|/Export1/...
174 downloads
1323 Views
8MB Size
Report
This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
Report copyright / DMCA form
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=cover.html
Cover
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=cover.html [4/21/2007 1:25:50 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_i.html
Page i
The Economics of Risk and Time
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_i.html [4/21/2007 1:25:51 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_ii.html
Page ii
This page intentionally left blank.
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_ii.html [4/21/2007 1:25:51 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_iii.html
Page iii
The Economics of Risk and Time Christian Gollier
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_iii.html [4/21/2007 1:25:52 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_iv.html
Page iv © 2001 Massachusetts Institute of Technology All rights reserved. No part of this book may be reproduced in any form by any electronic or mechanical means (including photocopying, recording, or information storage and retrieval) without permission in writing from the publisher. This book was set in Palatino in '3B2' by Asco Typesetters, Hong Kong. Printed and bound in the United States of America. Library of Congress Cataloging-in-Publication Data Gollier, Christian. The economics of risk and time / Christian Gollier. p. cm. Includes bibliographical references and index. ISBN 0-262-07215-7 (hc. : alk. paper) 1. Finance. 2. Financial engineering. 3. Risk. 4. Risk assessment. 5. Risk management. I. Title. HG101 .G65 2001 658.15'5—dc21 2001030213
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_iv.html [4/21/2007 1:25:52 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_v.html
Page v
Contents Preface Acknowledgments I General Theory 1 The Expected Utility Model 1.1 Simple and Compound Lotteries 1.2 Axioms on Preferences under Uncertainty 1.3 The Expected Utility Theorem 1.4 Critics of the Expected Utility Model 1.4.1 The Allais Paradox 1.4.2 The Allais Paradox and Time Consistency 1.5 Concluding Remark 1.6 Exercises and Extensions
2 Risk Aversion 2.1 Characterization of Risk Aversion 2.2 Comparative Risk Aversion 2.3 Certainty Equivalent and Risk Premium 2.4 The Arrow-Pratt Approximation 2.5 Decreasing Absolute Risk Aversion 2.6 Some Classical Utility Functions 2.7 Test for Your Own Degree of Risk Aversion 2.8 An Application: The Cost of Macroeconomic Risks 2.9 Conclusion 2.10 Exercises and Extensions
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_v.html [4/21/2007 1:25:53 AM]
xv xix 1 3 3 4 6 9 10 11 14 14 17 17 18 20 21 24 25 29 32 34 35
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_vi.html
Page vi
3 Change in Risk 3.1 The Extremal Approach 3.2 Second-Order Stochastic Dominance 3.3 Diversification 3.4 First-Order Stochastic Dominance 3.5 Concluding Remark 3.6 Exercises and Extensions
II The Standard Portfolio Problem 4 The Standard Portfolio Problem 4.1 The Model and Its Basic Properties 4.2 The Case of a Small Risk 4.3 The Case of HARA Functions 4.4 The Impact of Risk Aversion 4.5 The Impact of a Change in Risk 4.6 Concluding Remark 4.7 Exercises and Extensions
5 The Equilibrium Price of Risk 5.1 A Simple Equilibrium Model for Financial Markets 5.2 The Equity Premium Puzzle 5.3 The Equity Premium with Limited Participation 5.4 The Equity Premium and the Integration of International Financial Markets 5.5 Conclusion 5.6 Exercises
III Some Technical Tools and Their Applications 6 A Hyperplane Separation Theorem 6.1 The Diffidence Theorem 6.2 Link with the Jensen's Inequality 6.3 Applications of the Diffidence Theorem 6.3.1 Diffidence 6.3.2 Comparative Diffidence
39 40 42 45 46 47 48 51 53 53 55 57 58 59 61 62 65 65 68 71 73 75 76 79 81 81 88 89 89 90
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_vi.html (1 of 2) [4/21/2007 1:25:53 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_vi.html
6.3.3 Central Risk Aversion 6.3.4 Central Riskiness
91 92
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_vi.html (2 of 2) [4/21/2007 1:25:53 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_vii.html
Page vii 6.4 The Covariance Rule 6.5 Conclusion 6.6 Exercises and Extensions
7 Log-Supermodularity 7.1 Definition 7.2 Log-Supermodularity and Single Crossing 7.2.1 A Theoretical Result 7.2.2 Applications to the Standard Portfolio Problem 7.2.3 Jewitt's Preference Orders 7.3 Expectation of a Log-Supermodular Function 7.3.1 A Theoretical Result 7.3.2 Two Applications 7.4 Concluding Remark 7.5 Exercises and Extensions 7.6 Appendix
IV Multiple Risks 8 Risk Aversion with Background Risk 8.1 Preservation of DARA 8.2 The Comparative Risk Aversion Is Not Preserved 8.3 Extensions with Dependent Background Risk 8.3.1 Affiliated Background Risk 8.3.2 The Comparative Risk Aversion in the Sense of Ross 8.4 Conclusion 8.5 Exercises and Extensions
9 The Tempering Effect of Background Risk 9.1 Risk Vulnerability 9.2 Risk Vulnerability and Increase in Risk 9.2.1 Increase in Background Risk 9.2.2 Increase in the Endogenous Risk 9.3 Risk Vulnerability and the Equity Premium Puzzle
94 95 96 99 99 102 102 103 104 105 105 106 107 107 108 111 113 114 117 119 119 121 123 124 125 126 130 130 130 131
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_vii.html (1 of 2) [4/21/2007 1:25:54 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_vii.html
9.4 Generalized Risk Vulnerability 9.5 Standardness
132 135
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_vii.html (2 of 2) [4/21/2007 1:25:54 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_viii.html
Page viii 9.6 Conclusion 9.7 Exercises and Extensions
10 Taking Multiple Risks 10.1 The Interaction between Asset Demand and Small Gambles 10.2 Are Independent Assets Substitutes? 10.2.1 The i.i.d. Case 10.2.2 The General Case 10.3 Conclusion 10.4 Exercises and Extensions
11 The Dynamic Investment Problem 11.1 Static versus Dynamic Optimization 11.2 The Standard Portfolio Problem 11.2.1 The Model 11.2.2 The HARA Case 11.2.3 A Sufficient Condition for Younger People to Be More RiskAverse 11.3 Discussion of the Results 11.3.1 Nonlinear Risk Tolerance 11.3.2 Nondifferentiable Marginal Utility 11.4 Background Risk and Time Horizon 11.4.1 Investors Bear a Background Risk at Retirement 11.4.2 Stationary Income Process 11.5 Final Remark 11.6 Exercises and Extensions
12 Special Topics in Dynamic Finance 12.1 The Length of Periods between Trade 12.2 Dynamic Discrete Choice 12.3 Constraints on Feasible Strategies 12.4 The Effect of a Leverage Constraint 12.4.1 The Case of a Lower Bound on the Investment in the Risky Asset 12.4.2 The Case of an Upper Bound on the Investment in the Risky Asset
138 139 141 142 144 144 150 153 153 155 157 158 158 160 161 165 165 166 168 168 171 172 173 175 175 179 183 185 185 187
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_viii.html (1 of 2) [4/21/2007 1:25:54 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_viii.html
12.5 Concluding Remarks 12.6 Exercises and Extensions
190 190
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_viii.html (2 of 2) [4/21/2007 1:25:54 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_ix.html
Page ix
V The Arrow-Debreu Portfolio Problem 13 The Demand for Contingent Claims 13.1 The Model 13.2 Characterization of the Optimal Portfolio 13.3 The Impact of Risk Aversion 13.4 Conclusion 13.5 Exercises and Extensions
14 Risk on Wealth 14.1 The Marginal Propensity to Consume in State π 14.2 The Preservation of DARA and IARA 14.3 The Marginal Value of Wealth 14.4 Aversion to Risk on Wealth 14.5 Concluding Remark 14.6 Exercises and Extensions
VI Consumption and Saving 15 Consumption under Certainty 15.1 Time Separability 15.2 Exponential Discounting 15.3 Consumption Smoothing under Certainty 15.4 Analogy with the Portfolio Problem 15.5 The Social Cost of Volatility 15.6 The Marginal Propensity to Consume 15.7 Time Diversification and Self-Insurance 15.8 Concluding Remark 15.9 Exercises and Extensions
16 Precautionary Saving and Prudence 16.1 Prudence 16.2 The Demand for Saving 16.3 The Marginal Propensity to Consume under Uncertainty 16.3.1 Does Uncertainty Increase the MPC?
193 195 196 197 200 201 202 205 206 208 210 211 212 212 215 217 217 218 219 221 224 226 227 232 232 235 235 239 239 240
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_ix.html (1 of 2) [4/21/2007 1:25:55 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_ix.html
16.3.2 Does Uncertainty Make the MPC Decreasing in Wealth? 16.4 More Than Two Periods
241 242
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_ix.html (2 of 2) [4/21/2007 1:25:55 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_x.html
Page x 16.4.1 The Euler Equation 16.4.2 Multiperiod Precautionary Saving 16.5 Illiquid Saving under Uncertainty 16.6 Conclusion 16.7 Exercises and Extensions
17 The Equilibrium Price of Time 17.1 Description of the Economy 17.2 The Determinants of the Interest Rate 17.2.1 The Interest Rate in the Absence of Growth 17.2.2 The Effect of a Sure Growth 17.2.3 The Effect of Uncertainty 17.3 The Risk-Free Rate Puzzle 17.4 The Yield Curve 17.4.1 The Pricing Formula 17.4.2 The Yield Curve with HARA Utility Functions 17.4.3 A Result When There Is No Risk of Recession 17.4.4 Exploring the Slope of the Yield Curve When There Is a Risk of Recession 17.5 Concluding Remark 17.6 Exercises and Extensions
18 The Liquidity Constraint 18.1 Saving as a Buffer Stock 18.2 The Liquidity Constraint Raises Risk Aversion 18.3 The Liquidity Constraint and the Shape of Absolute Risk Tolerance 18.4 Numerical Simulations 18.5 Conclusion 18.6 Exercises and Extensions
19 The Saving-Portfolio Problem 19.1 Precautionary Saving with an Endogenous Risk 19.1.1 The Case of Complete Markets 19.1.2 The Case of the Standard Portfolio Problem
242 244 246 247 248 249 250 252 252 253 254 256 258 258 260 261 264 267 268 269 270 272 273 277 279 281 285 285 285 287
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_x.html (1 of 2) [4/21/2007 1:25:56 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_x.html
19.1.3 Discussion of the Results
288
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_x.html (2 of 2) [4/21/2007 1:25:56 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xi.html
Page xi 19.2 Optimal Portfolio Strategy with Consumption 19.3 The Merton-Samuelson Model 19.4 Concluding Remark 19.5 Exercises and Extensions
20 Disentangling Risk and Time 20.1 The Model of Kreps and Porteus 20.2 Preferences for an Early Resolution of Uncertainty 20.3 Prudence with Kreps-Porteus Preferences 20.4 Conclusion 20.5 Exercises and Extensions
VII Equilibrium Prices of Risk and Time 21 Efficient Risk Sharing 21.1 The Case of a Static Exchange Economy 21.2 The Mutuality Principle 21.3 The Sharing of the Social Risk 21.3.1 Decomposition of the Problem 21.3.2 The Veil of Ignorance 21.3.3 Efficient Sharing Rules of the Macro Risk 21.3.4 A Two-Fund Separation Theorem 21.3.5 The Case of Small Risk per Capita 21.4 Group's Attitude toward Risk 21.4.1 The Representative Agent 21.4.2 Arrow-Lind Theorem 21.4.3 Group Decision and Individual Choice 21.5 Introducing Time and Investment 21.6 A Final Remark: The Concavity of the Certainty Equivalent Functional 21.7 Conclusion 21.8 Exercises and Extensions 21.9 Appendix
22 The Equilibrium Price of Risk and Time
290 291 295 295 297 298 299 300 302 303 305 307 307 309 311 311 312 312 314 315 316 316 317 317 319 321 323 323 325 327
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xi.html (1 of 2) [4/21/2007 1:25:56 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xi.html
22.1 An Arrow-Debreu Economy 22.2 Application of the First Theorem of Welfare Economics 22.3 Pricing Arrow-Debreu Securities
327 328 329
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xi.html (2 of 2) [4/21/2007 1:25:56 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xii.html
Page xii 22.4 Pricing by Arbitrage 22.5 The Competitive Price of Risk 22.6 The Competitive Price of Time 22.7 Spot Markets and Markets for Futures 22.8 Corporate Finance in an Arrow-Debreu Economy 22.9 Conclusion 22.10 Exercises and Extensions
23 Searching for the Representative Agent 23.1 Analytical Solution to the Aggregation Problem 23.2 Wealth Inequality, Risk Aversion, and the Equity Premium 23.3 Wealth Inequality and the Risk-Free Rate 23.3.1 The Consumption Smoothing Effect 23.3.2 The Precautionary Effect 23.4 Conclusion 23.5 Exercises and Extensions
VIII Risk and Information 24 The Value of Information 24.1 The General Model of Risk and Information 24.1.1 Structure of Information 24.1.2 The Decision Problem 24.1.3 The Posterior Maximum Expected Utility Is Convex in the Vector of Posterior Probabilities 24.2 The Value of Information Is Positive 24.3 Refining the Information Structure 24.3.1 Definition and Basic Characterization 24.3.2 Garbling Messages and the Theorem of Blackwell 24.3.3 Location Experiments 24.4 The Value of Information and Risk Aversion 24.4.1 A Definition of the Value of Information 24.4.2 A Simple Illustration: The Gambler's Problem 24.4.3 The Standard Portfolio Problem
330 332 334 335 337 339 340 343 344 345 347 348 349 351 352 355 357 357 357 358 359 362 364 364 366 371 373 373 374 378
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xii.html (1 of 2) [4/21/2007 1:25:57 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xii.html
24.5 Conclusion
379
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xii.html (2 of 2) [4/21/2007 1:25:57 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xiii.html
Page xiii 24.6 Exercises and Extensions 24.7 Appendix
25 Decision Making and Information 25.1 A Technique for the Comparative Statics of More Informativeness 25.2 The Portfolio-Saving Problem 25.3 A Digression: Scientific Uncertainty, Global Warming, and the "Precautionary Principle" 25.4 The Saving Problem with Uncertain Returns 25.5 Precautionary Saving 25.6 The Value of Flexibility and Option Value 25.7 Predictability and Portfolio Management 25.7.1 Exogenous Predictability 25.7.2 Endogenous Predictability and Mean-Reversion 25.8 Conclusion 25.9 Exercises and Extensions
26 Information and Equilibrium 26.1 Hirshleifer Effect 26.2 Information and the Equity Premium 26.3 Conclusion 26.4 Exercises and Extensions 26.5 Appendix
27 Epilogue 27.1 The Important Open Questions 27.1.1 The Independence Axiom 27.1.2 Measures of Risk Aversion 27.1.3 Qualitative Properties of the Utility Function 27.1.4 Economics of Uncertainty and Psychology 27.2 Conclusion
Bibliography Index of Lemmas and Propositions Index of Subjects
380 382 383 383 386 389 390 392 393 397 399 400 405 405 407 407 413 418 418 420 423 423 423 424 425 426 427 429 441 443
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xiii.html [4/21/2007 1:25:57 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xiv.html
Page xiv
This page intentionally left blank.
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xiv.html [4/21/2007 1:25:58 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xv.html
Page xv
Preface Uncertainty is everywhere. There is no field in economics in which risk is not a dimension of the decisionmaking environment. The theory of finance provides the most obvious example of uncertainty in decision making. Similarly most recent developments in macroeconomics have been made possible by recognizing the importance of risk in explaining individual decisions. Consumption patterns, investments, and labor decisions can only be understood completely if uncertainty is taken into account into the decision-making process. Environmental economics provides another illustration. Public opinion is now very sensitive to the presence of potentially catastrophic risks related, for example, to the greenhouse effect and genetic manipulations. Environmental economists have introduced probabilistic scenarios in their models to exhibit socially efficient levels of prevention efforts. Finally, the extraordinary contributions of asymmetric information to game theory have reinforced interest in uncertainty among economists. We are lucky enough to have a well-accepted and unified framework to introduce uncertainty in economic modeling. John von Neumann and Oskar Morgenstern, in the mid-1940s, developed the expected utility theory, building on Daniel Bernoulli's idea that agents facing risk maximize the expected value of the utility of their wealth. Expected utility (EU) theory is now fifty years old. It is a ubiquitous instrument in economic modeling. Most economists recognize that the theory has been very useful for explaining the functioning of our economies. The aim of this book is to provide a detailed and unified analysis of the implications of the expected utility model to the economic theory. One of the innovation of this book is its refusal to use convenient utility functions to solve complex problems of decision and market
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xv.html [4/21/2007 1:25:58 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xvi.html
Page xvi equilibrium under uncertainty. The benefit of this method is twofold. First, it allows us to better understand the relevant basic concepts that determine optimal behavior under uncertainty. Second, it reestablishes the richness and diversity of the EU framework, something that tends to be forgotten in the strongly reductionist limitations imposed on the set of utility functions usually used for convenience. This book provides an overview of the most recent developments in expected utility theory. It is aimed at any audience that is interested in problems related to efficient, or optimal, public and private strategies for dealing with risk. The book heavily relies on concepts that are standard in the theory of finance. But most findings presented in this book are useful for our understanding of various economic problems ranging from macroeconomic fluctuations to global warming.
Organization of the Book Part I of this book is devoted to the presentation of the EU model and its basic related concepts. Chapter 1 presents the axioms that underlie the EU approach. It proves the von Neumann–Morgenstern EU theorem and discusses its main limitations. Chapter 2 shows how to introduce risk aversion, and "more risk aversion," in the EU model. The notion of risk premium is introduced here, together with the familiar utility functions that are used in macroeconomics and finance. Chapter 3 is devoted to the dual question of how to define risk, and "more risk," in this model. It also presents some recent developments in the literature on stochastic dominance. In part II we examine the standard problem of choice under uncertainty in which an agent must decide how much to invest in a portfolio of two assets, one of the two being risk free. Chapter 4 reviews the effect of a change in wealth, in the degree of risk aversion, or in the degree of risk on the optimal demand for the risky asset. This model is also helpful for understanding the behavior of a policyholder insuring a given risk at an unfair premium, or the behavior of a risk-averse entrepreneur who must determine her productions capacity under price uncertainty. Chapter 5 is devoted to an equilibrium version of this problem. It presents the simplest version of the equity premium puzzle. Part III introduces some technical tools that are useful to solve various classes of decision problems under uncertainty. It shows how the linearity of the EU objective function with respect to probabilities is very helpful. Simple hyperplanes separation theorems can
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xvi.html [4/21/2007 1:25:59 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xvii.html
Page xvii be used in such an environment. The existence of such simple technical tool in the case of EU is central to our understanding of the success of the EU model in economics. Chapter 6 is devoted to the presentation of the basic hyperplane separation theorem—a tool richer than Jensen's inequality—and its simplest applications in the economics of uncertainty. Chapter 7 examines more sophisticated applications that rely on the notion of logsupermodular functions. The use of these tools will allow us to unify the wide range of results that has appeared in the literature during the last thirty years. Part IV takes various decision problems to more than one risk. We take the standard portfolio model which was developed in the early 1980s and treat portfolio risk as a unique source of risk borne by the consumer. To this we introduce the real world, where people must manage and control several sources of risk simultaneously. Chapters 8 and 9 show how the presence of an exogenous risk in background wealth affects the demand for other independent risks. New considerations like properness, standardness, and risk vulnerability are introduced and discussed. Chapter 10 covers the problem of whether independent risks can be substitute. To illustrate, we characterize the conditions under which the opportunity to invest in one risky asset reduces the optimal demand for another independent risky asset. The remaining of this part deals with how to use dynamic programming methods to show how the time horizon length affects the optimal in behavior toward repeated risks. Chapter 11 does this for the simplest case where financial markets are frictionless; then chapter 12 explores the effects of various limitations in dynamic trading strategies on optimal dynamic portfolio management. The canonical decision problem in the theory of finance, namely the Arrow-Debreu portfolio problem, is developed in part V. Chapter 13 shows that the investor can exchange contingent claim contracts in competitive markets. We determine the impact of risk aversion on the investors optimal portfolio. Several properties of the value function for the investor's wealth are derived in chapter 14. Consumption and savings form the framework of part VI. We start with the consumption problem under certainty in chapter 15. There, we see how people evaluate their welfare when they consume their wealth over different periods. We learn that the structure of this problem is equivalent to the structure of the static Arrow-Debreu portfolio. We then explore the properties of the marginal propensity to consume and the concept of "time diversification." The assumption of a time-separable utility function makes the treatment of time
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xvii.html [4/21/2007 1:25:59 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xviii.html
Page xviii very similar to the treatment of uncertainty in classical economics. In chapter 16 the important notion of prudence is introduced. Saving is examined as a way to build a precautionary reserve to forearm oneself against future exogenous risks. The equilibrium version of this model is examined in chapter 17, where the equilibrium riskfree rate is derived. In these three chapters we assume that financial markets are perfect; that is to say, there is a single risk-free interest rate at which consumers can borrow and lend. This is obviously not a realistic assumption. So a market imperfection is introduced in chapter 18 by the way of a liquidity constraint. The existence of a liquidity constraint provides another incentive to save. Related to this is the Merton-Samuelson problem of the joint decisions on consumption and risk-taking. This problem is discussed in chapter 19. Finally, chapter 20 covers a model that allows us to disentangle the notions of risk aversion and aversion to consumption over time. In part VII several of the previous results are gathered in order to determine the equilibrium price of risk and time in an Arrow-Debreu economy. More generally, we look at an analysis of how risks are traded in our economies. Chapter 21 starts with the characterization of socially efficient risk-sharing arrangements. Chapter 22 covers how competition in financial markets can generate an efficient allocation of risks. It also discusses how to determine the equilibrium price of risk and time as a function of the characteristics of the particular economy. The discussion includes the properties of the capital asset pricing model. The chapter then provides an introduction to decision-making problems for corporate firms, and to the Modigliani-Miller theorem. Chapter 23 shows how individual risk attitudes can be aggregated to derive the society's attitude toward risk and time and its impact on asset prices. Last, part VIII focuses on dynamic models of decisions making under uncertainty when a flow of information on future risks is expected over time. The basic tools and concepts (value of information and Blackwell's theorem) are provided in chapter 24. Chapter 25 covers the optimal timing of an investment decision when there is a quasioptional value to waiting for future information about the investments profitability. The degree of irreversibility of the investment is taken into account. The concluding chapter 26 offers a smorgasbord of analyses related to the effect of the expectation of future information on current decisions and an equilibrium prices of financial assets.
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xviii.html [4/21/2007 1:26:00 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xix.html
Page xix
Acknowledgments The origin of this book goes back to my attendance of the enthusiastic and profound lectures on the economics of uncertainty by Jacques Drèze at the University of Louvain in the mid-1980s. Since those formative years, many people have contributed to the development of this book. Foremost I am indebted to Louis Eeckhoudt for all out years of joint work. Moreover he took a great responsability in reading the first versions of the manuscript in a very professional way. I am likewise grateful to my other co-authors on various papers related to this book: Bruno Jullien, Miles Kimball, John Pratt, Jean-Charles Rochet, Ed Schlee, Harris Schlesinger, Nicolas Treich, and Richard Zeckhauser. I further thank Rose-Anne Dana and Carole Haritchabalet for their helpful comments. My graduate students both at the University of Toulouse and at Studienzentrum Gerzensee have helped us with their questions, remarks, and corrections. Florence Chauvet, Joan Le Berre Desrochers, and Elisabeth Soulie provided expert secretarial support throughout the project. I also thank Elizabeth Murry at The MIT Press for her clever managerial advice concerning this book. This book would not have been possible without the exceptional environment in my life. In addition to the members of my own family, I want to thank my colleagues at GREMAQ and IDEI (Toulouse) for the quality of the scientific atmosphere that I enjoyed during the last seven years. Working on a long-term writing project is a risky activity that requires self-confidence. This would not have been possible without the friendly encouragement of Bruno Biais, Jean-Jacques Laffont, Jean-Charles Rochet, and Jean Tirole. I also benefited of invitations to work in other research centers that provided the environment ideal for big pushes in the progress of the manuscript: the
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xix.html [4/21/2007 1:26:00 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xx.html
Page xx CES in Munich, Yale University, Duke University, the University of Montreal, and the European University Institute at Florence. Finally I acknowledge the continuing financial support of various institutions at crucial times in the writing of the book: the Fédération Française des Sociétés d'Assurance, the Geneva Association, the Institut Universitaire de France, the Commissariat à l'Energie Atomique, and the COGEMA. Without the generous support of these institutions, I would not have been able to devote the time necessary to complete this work. C. G. Florence, Italy November 2000
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_xx.html [4/21/2007 1:26:01 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_1.html
Page 1
I General Theory
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_1.html [4/21/2007 1:26:01 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_2.html
Page 2
This page intentionally left blank.
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_2.html [4/21/2007 1:26:01 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_3.html
Page 3
1 The Expected Utility Model. Before addressing any decision problem under uncertainty, it is necessary to build a preference functional that evaluates the level of satisfaction of the decision maker who bears risk. If such a functional is obtained, decisions problems can be solved by looking for the decision that maximizes the decision maker's level of satisfaction. This first chapter provides a way—the now classical way—to evaluate the level of satisfaction under uncertainty. The rest of this book deals with applications of this model to specific decision problems. The modern theory of risk-bearing has been founded in 1944 by von Neumann and Morgenstern in their famous book entitled Theory of Games and Economic Behavior. In the 1930s and early 1940s, economists like John Hicks (1931), Jacob Marschak (1938), and Gerhard Tintner (1941) were debating whether the ordering of risks could be based on a function of their mean and variance alone. A first axiomatic approach to the ordering of probability distributions had been introduced by Frank Ramsey (1931). This approach was revived in successively clearer and simpler terms by von Neumann and Morgenstern (1947) and Marschak (1950). The aim of this chapter is to summarize this axiomatic approach.
1.1 Simple and Compound Lotteries The description of an uncertain environment contains two different types of information. First, one must enumerate all possible outcomes. An outcome is a list of variables that affect the well-being of the decision maker. The list might give a health status, some meteorological parameters, levels of pollution, and the quantities of different goods consumed. As long as we do not introduce saving (chapter
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_3.html [4/21/2007 1:26:02 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_4.html
Page 4 15), we will assume that the outcome can be measured in a one-dimensional unit, namely money (consumed at a specific date in time). But at this stage, there is no cost to allow for a multidimensional outcome. Let X be the set of possible outcomes. To avoid technicalities, we assume that the number of possible outcomes is finite, so X ≡ {xs}s=1,..., S. The second characteristic of an uncertain environment is the vector of probabilities of the possible outcomes. Let ps ≥ 0 be the probability of occurrence of xs, with
. We will assume that these probabilities are
objectively known. A lottery L is described by a vector (x1, p1; x2, p2; . . . ; xS, pS). Since the set of potential outcomes is invariant, we will define a lottery L by its vector of probabilities (p1, . . . , pS). The set of all lotteries on outcomes X is denoted . When S = 3, we can represent a lottery by a point (p1, p3) in R2, since p2 = 1 – p1 – p3. More precisely, in order to be a lottery, this point must be in the so-called Machina triangle . If the lottery is on a edge of the triangle, one of the probabilities is zero. If it is at a corner, the lottery is degenerated, by which we mean that it takes one of the values x1, x2, x3 with probability 1. This triangle is represented in figure 1.1. A compound lottery is a lottery whose outcomes are lotteries. Consider a compound lottery L which yields lottery with probability α and lottery the outcome of L be x1 is
with probability 1 – α. The probability that
. More generally, we obtain that
A compound lottery is a convex combination of simple lotteries. Such a compound lottery is represented in figure 1.1. From condition (1.1), it is natural to confound L with αLa + (1 – α)Lb. Whether a specific uncertainty comes from a simple lottery or from a complex compound lottery has no significance. Only the probabilities of potential outcomes matter (axiom of reduction).
1.2 Axioms on Preferences under Uncertainty We assume that the decision maker has a rationale preference relation over the set of lotteries that order is complete and transitive. For any pair (La, Lb) of lotteries in , either La is
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_4.html [4/21/2007 1:26:03 AM]
. This means
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_5.html
Page 5
Figure 1.1 Compound lottery, L = αLa(1 – α)Lb, in the Machina triangle
preferred to Lb (La Lb), or Lb is preferred to La (Lb La) (or both). Moreover, if La is preferred to Lb which is itself preferred to Lc, then La is preferred to Lc. To this preference order , we associate the indifference relation ~, with La ~ Lb if and only if La Lb and Lb La. A standard hypothesis that is made on preferences is that they are continuous. This means that small changes in probabilities do not change the nature of the ordering between two lotteries. Technically, this axiom is written as follows: 1 (CONTINUITY) The preference relation on the space of simple lotteries such that La Lb Lc, there exists a scalar α ∈ [0, 1] such that
AXIOM
∈
is such that for all La, Lb, Lc
As is well-known in the theory of consumer choice under certainty, the continuity axiom implies the existence of a functional U: → R such that
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_5.html (1 of 2) [4/21/2007 1:26:03 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_5.html
The preference functional U is an index that represents the degree of satisfaction of the decision maker. It assigns a numerical value to each lottery, ranking the lotteries in accordance to the individual's
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_5.html (2 of 2) [4/21/2007 1:26:03 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_6.html
Page 6 preference . Notice that U is not unique: take any increasing function g: R → R. Then the functional V that is defined as V(.) = g(U(.)) also represents the same preferences . The preference functional is ordinal in the sense that it is invariant to any increasing transformation. It is only the ranking of lotteries that matters. A preference functional can be represented in the Machina triangle by a family of continuous indifference curves. The assumptions above on preferences under uncertainty are minimal. If no other assumption is made on these preferences, the theory of choice under uncertainty would not differ from the standard theory of consumer choice under certainty. The only difference would be on how to define the consumption goods. Most developments in the economics of uncertainty and its applications have been made possible by imposing more structure on preferences under uncertainty. This additional structure originates from the independence axiom. AXIOM
Lc ∈
2 (INDEPENDENCE) The preference relation and for all α ∈ [0, 1],
on the space of simple lotteries
is such that for all La, Lb,
This means that if we mix each of two lotteries La and Lb with a third one Lc, then the preference ordering of the two resulting mixtures is independent of the particular third lottery Lc used. The independence axiom is at the heart of the classical theory of uncertainty. Contrary to the other assumptions made above, the independence axiom has no parallel in the consumer theory under certainty. This is because there is no reason to believe that if I prefer a bundle A containing 1 cake and 1 bottle of wine to a bundle B containing 3 cakes and no wine, I also prefer a bundle A′ containing 2 cakes and 2 bottles of wine to a bundle B′ containing 3 cakes and 1.5 bottles of wine, just because
1.3 The Expected Utility Theorem The independence axiom implies that the preference functional U must be linear in the probabilities of the possible outcomes. This is the essence of the expected utility theorem, which is due to von
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_6.html [4/21/2007 1:26:04 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_7.html
Page 7 Neumann and Morgenstern (1947). It has been extended later on by Savage (1954) to the case where there is no objective probabilities. 1 (EXPECTED UTILITY) Suppose that the rationale preference relation on the space of simple lotteries satisfies the continuity and independence axioms. Then can be represented by a preference functional that is linear in probabilities. That is, there exists a scalar us associated to each outcome xs, s = 1, . . . , S, in such a THEOREM
manner that for any two lotteries
and
, we have
Proof We would be done if we prove that for any compound lottery L = βLa + (1 – β)Lb, we have
Applying this property recursively would yield the result. To do this, let us consider the worst and best lotteries in , and . They are obtained by solving the problem of minimizing and maximizing U(L) on the compact set . By definition, for any L ∈ , we have scalars, αa and αb, in [0, 1] such that
. By the continuity axiom, we know that there exist two
and
Observe that La Then we have
Lb if and only if αa ≥ αb. Indeed, suppose that αa ≥ αb and take γ = (αa – αb)/(1 – αb) ∈ [0, 1].
The two equivalences are direct applications of the axiom of reduction. The second line of this sequence of preference orderings is due to the independence axiom together with the fact that file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_7.html (1 of 2) [4/21/2007 1:26:04 AM]
, by
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_7.html
definition of .
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_7.html (2 of 2) [4/21/2007 1:26:04 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_8.html
Page 8 , perfectly fits the definition (1.2) of a We conclude that U(L) = α, where α is such that preference functional associated to . Thus U(La) = αa and U(Lb) = αb. It remains to prove that
or equivalently that
This is true since, using the independence axiom twice, we get
This concludes the proof. The consequence of the independence axiom is that the family of indifference curves in the Machina triangle is a set of parallel straight lines. Their slope equals (u1 – u2)/(u3 – u2). This has the following consequences on the attitude toward risk. Consider the four lotteries, La, Lb, Ma, and Mb, depicted in the Machina triangle of figure 1.2. Suppose that the segment LaLb is parallel to segment
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_8.html (1 of 2) [4/21/2007 1:26:05 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_8.html
Figure 1.2 Allais paradox in the Machina triangle
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_8.html (2 of 2) [4/21/2007 1:26:05 AM]
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_9.html
Page 9 MaMb. This implies that La is preferred to Lb if and only if Ma is preferred to Mb. In this book, we will consider the outcome to be a monetary wealth. The ex-post environment of the decision maker is fully described by the amount of money that he can consume. A lottery on wealth can be expressed by a random variable whose realization is an amount of money w that can be consumed. This random variable can be expressed by a cumulative distribution function F, where F(w) is the probability that be less or equal than w. This covers the case of continuous, discrete, or mixed random variables. By the expected utility theorem, we know that to each wealth level w, there exists a scalar u(w) such that
where Fi is the cumulative distribution function of
. It is intuitive in this context to assume that the utility
function is constant. Notice that the utility function is cardinal: an increasing linear transformation of u, v(.) = au(.) + b, a > 0, will , we have not change the ranking of lotteries: if
To sum up, expected utility is ordinal, whereas the utility function is cardinal. Differences in utility have meaning, whereas differences in expected utility have no significance.
1.4 Critics of the Expected Utility Model The independence axiom is not applied without difficulties. The oldest and most famous challenge was proposed by Allais (1953). Table 1.1 Outcome as a function of the number of the ball Lottery
0
1–10
11–99
La Lb Ma Mb
50 0 50 0
50 250 50 250
50 50 0 0
file:///D|/Export1/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_9.html [4/21/2007 1:26:06 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_10.html
Page 10 The paradox that he raised generated thousands of papers. Whether the Allais's counterexample knocks out expected utility theory or comes from outwitting the players is still subject of lively discussions among the specialists. We first present the paradox. We then summarize the main arguments that have been proposed by the EU proponents to sustain that the paradox comes from a misconception by the players. 1.4.1 The Allais Paradox Allais proposed the following experiment: An urn contains 100 balls numbered from 0 to 99. There are four lotteries whose monetary outcomes depend in different ways on the number that is written on the ball that is taken out of the urn. An example of an outcome expressed in thousands of dollars is described in table 1.1. Decision makers are subjected to two choice tests. In the first test, they are asked to choose between La and Lb, and in the second test, they must choose between Ma and Mb. Many people report that they prefer La to Lb but subsequently prefer Mb to Ma. Notice that since La and Lb have the same outcome when the number of the ball is larger than 10, the independence axiom tells us that most people will prefer La′, which takes value 50 with certainty, to Lb′, which takes value 0 with probability 1/11 and value 250 with probability 10/11. The paradox is that the same argument can be used with the opposite result when considering the preference of Mb over Ma! Thus these choices among pairs of lotteries are not compatible with the independence axiom. This problem can be analyzed using the Machina triangle in figure 1.2. Let us define
The four lotteries in the Allais experiment are respectively La = (0, 1, 0), Lb = (0.01, 0.89, 0.10), Ma = (0.89, 0.11, 0) and Mb = (0.90, 0, 0.10). Segment LaLb is parallel to segment MaMb. Because indifference curves are parallel straight lines under expected utility, it must be the case that Ma Mb if La Lb.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_10.html [4/21/2007 1:27:51 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_11.html
Page 11 1.4.2 The Allais Paradox and Time Consistency Savage (1954), who was a strong supporter of the expected utility model, participated in the test organized by Allais. It happened that Savage was one of those people who preferred La to Lb and Mb to Ma. After realizing that this was not compatible with expected utility, he wanted to revise his choices. He claimed that he had been misled and that a more cautious reading of the problem would have been sufficient to avoid the mistake. A first argument in favor of the independence axiom is that individuals who violate it would find themselves subject to so-called Dutch book outcomes. To illustrate, suppose that a gambler is offered three lotteries La, Lb, and Lc. The gambler ranks them La Lb and La Lc, but contrary to the independence axiom, it turns out that Ld = 0.5Lb + 0.5Lc is preferred to La. Ld is a compound lottery in which a fair coin is used to determine which lottery, Lb or Lc, will be played. In making the choice from among La, Lb, and Lc, the gambler rationally selects La. Since Ld is preferred to La, we know that he is willing to pay a fee to replace La by the compound lottery Ld. This outcome is thus rationally accepted by the gambler. But, as soon as the coin is tossed, giving the gambler either Lb or Lc, one could get him to pay another fee to trade this lottery for La. At this point the gambler has paid two fees but is otherwise back to his original position. This outcome is dynamically inconsistent. The relation between the independence axiom and time consistency is central to the current debate between the EU and non-EU specialists. Wakker (1999) makes this relation clear by using the Allais's example. We summarize Wakker's analysis in figure 1.3. A simpler version of the Allais paradox is expressed by the comparison of figure 1.3a and e. As is usual, the square in the decision tree represents a decision (up or down), whereas the circle describes the occurrence of a random event. For example, problem a is to choose between a lottery that has a payoff of 250 with probability 10/11, and a sure payoff of 50. Standard choices are to select "down" in figure 1.3a, and to select "up" in figure 1.3e. Any combination of these two choices is not compatible with expected utility. To see how this is possible, let us examine the following sequence of comparisons: • Comparison between a and b. Upon being asked to express her preferences between lotteries "up" and "down" of problem a, an agent is
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_11.html [4/21/2007 1:27:51 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_12.html
Page 12
Figure 1.3 Decomposing Allais's paradox
told that this is only by chance that she has to participate. In fact, she has been selected at random with 11 other participants in a group of 100 persons. The decision problem in figure 1.3a is in fact a subtree of a larger problem that is represented by figure 1.3b. Should the agent modify her decision on the basis of this forgone event? Most people agree that forgone events should not affect the decision of the agent. Past events are irrelevant for the current decision. This rationale is called "consequentialism." It tells us that rational agents will make the same decision in both problem a and problem b. • Comparison between b and c. Let us now take the problem back one step. We have a group of 100 persons. Eleven will be selected at random to participate in game a. The agent under scrutiny does
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_12.html [4/21/2007 1:27:52 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_13.html
Page 13 not yet know whether she will be among the lucky eleven persons. However, she is asked to commit herself to the strategy that she will follow if she is selected. Figure 1.3c describes this problem. If the agent is time consistent, she should commit herself ex ante to do what will be optimal to do ex post. Why, for example, would she commit to the "up" strategy when she knows that "down" is optimal? Time consistency tells us that people will make the same decision in problems b and c. • Comparison between c and d. Problem d is essentially a rewriting of problem c. If the agent commits to strategy "up," she faces a selection risk and a gamble risk as is represented by the upper part of tree d. If she decides to commit to "down," she only faces the selection risk. Now, let us forget the story of selecting eleven people from a group of 100 persons. Problem d is a decision problem between two compound lotteries. This changes the context of the problem but not the underlying lotteries faced by the agent. Do we think that the context in the way the gamble is processed matters for the decision maker? Most people would say no. The context does not matter. Only the state probabilities and the state payoffs do. This assumption, called "context independence," implies that rational agents will make the same choice in problems c and d. • Comparison between d and e. Compounded lotteries can be reduced by using the standard law of probabilities. For example, if "up" is selected in problem d, there is an probability of obtaining 0. This is what we earlier called the reduction axiom. By repeating this calculation, the reader can easily check that problems d and e are the same. Assuming that the agent is able to reduce compounded lotteries in this way, we can conclude that she will make the same choice in both problems d and e. From this sequence of elementary assumptions—consequentialism, time consistency, context independence, and reduction—about preferences, it should be clear that if an agent prefers "down" to "up" in problem a, he will behave the same way in each of the other problems of figure 1.3, in particular, problem e. The violation of independence is possible only if at least one of the elementary assumptions above is also violated. While it may be possible that some people will violate one of these assumptions in experiments, we believe that there is a strong normative appeal to each assumption.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_13.html [4/21/2007 1:27:53 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_14.html
Page 14 Table 1.2 Outcome as a function of the number of the ball Lottery
Red
Black
White
La Lb Ma Mb
50 0 50 0
0 50 0 50
0 0 50 50
1.5 Concluding Remark We will learn more in the next chapters about the linearity of the preference functional with respect to probabilities and its central role in finding simple solutions to decision problems under uncertainty. If we relax the independence axiom, most problems presented in this book cannot be solved anymore. Our approach is thus pragmatic: we recognize that the independence axiom may fail to describe real behavior in certain risky environments, especially when there are low-probability events. But the combination of facts showing that the independence axiom, and its related axioms of time consistency and consequentialism, make common sense and make most problems solvable is enough to justify the exploration of its implications. The search for an alternative model of decision making under uncertainty that does not rely on the independence axiom has been a lively field of research in economics for more than fifteen years. There are several competing models, each with its advantages and faults.1 Most entail the expected utility model as a particular case. Nevertheless, the use of the expected utility model is pervasive in economics. Probably because of its strong normative basis.
1.6 Exercises and Extensions Exercise 1 Consider the Allais paradox summarized in table 1.1. Show that there exists no utility function such that the expected utility of La is larger than the expected utility of Lb, and the expected utility of Mb is larger than the expected utility of Ma.
1. For more about the history of the economic thought on this field, see Bernstein (1996) and Drèze (1987), for example.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_14.html [4/21/2007 1:27:53 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_15.html
Page 15 Exercise 2 (Ellsberg's paradox) Savage (1954) extended the expected utility theorem to situations where the probabilities are not objectively known. He introduced the sure thing principle, an axiom that is stronger than the independence axiom. He showed that this stronger axiom could be used to prove the "subjective EU" theorem by noting that • there exists a subjective probability measure p, • there exists a real-valued utility function u such that the decision maker ranks various distributions of consequences ω by using their subjective expected utility Σspsu(ωs). The following example is due to Ellsberg (1961). An urn contains 90 colored balls. Thirty balls are red, and the remaining 60 either black or white; the number of black (white) balls is not specified. There are 4 lotteries as described in table 1.2. Given that many people report ranking La Lb and Mb Ma, show that there exists no pair (p, u) that sustains such ranking of lotteries. The agents are averse to ambiguous probabilities, a possibility that is ruled out by the sure thing principle. (For a model with ambiguity aversion, see Gilboa and Schmeidler 1987.) Exercise 3 (Rank Dependent EU) Consider the lottery L = (x1, p1; x2, p2; . . . ; xs, ps), with x1 < x2 < . . . < xs. Quiggin (1982) proposed a generalization of the EU model by introducing a function g of transformation of the cumulative distribution:
with g(0) = 0 and g(1) = 1. • Draw the indifference curves in the Machina triangle for g(q) = q2 and
.
• Which restriction on g would you consider to solve the Allais paradox?
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_15.html [4/21/2007 1:27:54 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_16.html
Page 16
This page intentionally left blank.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_16.html [4/21/2007 1:27:54 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_17.html
Page 17
2 Risk Aversion In this chapter we introduce an additional restriction to the decision criterion under uncertainty. We now assume that decision makers are risk-averse. The concept of risk aversion is useful to measure the effect of risk on the level of satisfaction of the risk bearer. The main reference for this chapter is the dense paper by John Pratt (1964). The basic technical tool for this chapter is the well-known Jensen's inequality. Consider a real-valued function f is always smaller than if and any (vector of) random variable . Jensen's inequality just states that and only if f is concave. This property is actually a direct consequence of the definition of a concave function. That is to say, a function f is concave if and only if for all λ ∈ [0, 1] and all pairs (a, b) in the domain of f the following condition holds:
Graphically this means that any straight segment linking two points on the curve representing f must lie below the when is binary and distributed as (λ, a; 1 – λ, b). curve. Definition (2.1) is equivalent to Jensen's inequality extends this definition to any random variable. It should be noticed that the concavity comes from the absence of any restriction on the set of random variables . More precisely, this means that if f is not is larger than . Notice that we don't need concave, there must exist a random variable such that to know whether the function f is differentiable to apply the inequality. However, if f is twice differentiable, it is concave if and only if its second derivative is uniformly nonpositive, when the domain of f is R.
2.1 Characterization of Risk Aversion The effect of risk on well-being depends on three factors: the risk itself, the wealth level of the agent and his utility function. We focus
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_17.html [4/21/2007 1:27:54 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_18.html
Page 18 here on the effect of a pure risk, which is a zero-mean risk. It is widely believed that agents dislike pure risks, meaning that agents are averse to risk. DEFINITION
1 An agent is risk-averse if he dislikes all zero-mean risk at all wealth levels.
If the agent has a utility function u and an initial wealth level w0, this means that (w0) for all risk
with
. Equivalently this agent is risk-averse if and only if
is smaller than u is smaller than
for all random variable representing final wealth . An agent is risk-averse if replacing an uncertain final wealth by its expected value makes him better off. From Jensen's inequality this is true if and only if u is concave. PROPOSITION
1 An agent with utility function u is risk-averse if and only if u is concave.
Similarly an agent is risk-lover if his utility function is convex. He is risk-neutral if u is linear. The observation of human behavior strongly favors the assumption that human beings are risk-averse.1 For example, most households will insure their physical assets if an actuarially fair insurance premium is offered to cover them. Or, most investors will not purchase risky assets that cannot yield a larger expected return than the risk-free asset. These strategies are compatible with the assumption of risk aversion. However, horse betting, state-owned lotteries, and other unfair gambling operations contradict this assumption. There are various way to explain this. One is the element of optimism in the assessment of the probability of wining; another is the willingness of altruistic agents to participate in financing the public good often provided by the organizers of lotteries. Anyway, the size of markets for unfair lotteries is marginal with respect to the importance of insurance and financial markets, where risk aversion is necessary to explain observed strategies and equilibrium prices.
2.2 Comparative Risk Aversion. Human beings are heterogeneous in their genes and preferences. Some value safety at a high price, and others do not. The former will be reluctant to purchase stocks and will be willing to insure their endowed risk, even at a high insurance premium. In a word, they are
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_18.html [4/21/2007 1:27:55 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_19.html
Page 19 highly risk-averse. It is crucial for the discussion of this book to make this ordering of risk aversion more precise. Describing agents by their utility functions, we have the following definition. DEFINITION
2 Let agents u1 and u2 have the same initial wealth. Agent u1 is more risk-averse than agent u2 if u1
dislikes all lotteries that u2 dislikes, independent of their initial wealth. This is one way to say that u1 is more reluctant to take risk than u2. In order to characterize this distinction better, we can define function
in such a way that
for all u. Because u2 is increasing,
defined. By definition, we have that
for all z: function
is well
transforms u2 into u1. We check that
is positive. Thus function is increasing. Let us assume in addition that is concave. Then consider any risk that is disliked by agent u2, namely . Since is increasing and concave, Jensen's inequality implies that
This means that the concavity of u2 dislikes. It is also necessary. Indeed, if
is sufficient to guarantee that agent u1 dislikes all risks that agent is not concave, there must exist an interval in the image of u2 where
is convex. Consider an initial wealth w0 and a lottery interval, and which are such that locally convex , yields the result that u1 likes risk
such that
has its support in that
. Using Jensen's inequality as above, but with the , meaning that he is more risk-prone than agent u2, a
contradiction. We conclude that u1 is more risk-averse than u2 if and only if u1 is a concave transformation of function u2. If we assume that these two functions are twice differentiable, this requires that calculus, we obtain that
″(.) be nonpositive. Using standard
where file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_19.html (1 of 2) [4/21/2007 1:27:55 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_19.html
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_19.html (2 of 2) [4/21/2007 1:27:55 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_20.html
Page 20 is the so-called Arrow-Pratt degree of absolute risk aversion of agent ui measured at z. From this observation we see that
is concave if and only if A1 is uniformly larger than A2. We summarize these results in the next
proposition. PROPOSITION
2 Suppose that u1 and u2 are twice differentiable. The following conditions are equivalent:
1. The agent with utility function u1 is more risk-averse than the agent with utility function u2; namely the former rejects all lotteries that the latter rejects. 2. u1 is a concave transformation of u2. 3. A1 is uniformly larger than A2.
2.3 Certainty Equivalent and Risk Premium We continue to assume that agents are risk-averse: pure risks reduce the level of satisfaction of risk-bearers. Now we want to quantify this effect. This can be done by evaluating the maximum amount of money π that one is ready to pay to escape a pure risk. For a risk-averse agent, this amount is positive. π is called the risk premium. It can be calculated by solving the following equation:
One is indifferent between retaining the risk and paying the risk premium to eliminate the risk. The risk premium . is a function of the parameters appearing in the equation above: For the risk premium to be a valid measure of risk aversion, it must be that an increase in risk aversion increases the risk premium. Intuitively, a more risk-averse agent should be willing to pay more to evade a given risk. Suppose that u1 is more risk-averse than u2. We show that it implies that is larger than , or that
for all
, π2, and z0. The left-hand side equality means that π2 is the risk premium that agent u2 is ready to pay to
get rid of the risk, whereas the right-hand side inequality means that π1 is larger than π2. Defining w0 = z0 – π2 and
, the condition above is equivalent to requiring that agent u1 reject all lotteries
for which
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_20.html [4/21/2007 1:27:56 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_21.html
Page 21 agent u2 is indifferent, assuming they have the same wealth w0. We know that this is true if and only if u1 is more concave than u2. PROPOSITION
3 Agent u1 is always ready to pay more than agent u2 to evade a given risk so that for any w0 and
, if and only if u1 is more risk-averse than u2.
We now define the concept of certainty equivalent. Consider a lottery whose net gain is described by random variable , with . Suppose that you are offered a deal in which you can either play the lottery e or receive a sure gain C . What is the sure amount Ce that makes you just indifferent between the two options? The answer to this question is called the certainty equivalent of the (indivisible) lottery. It verifies condition (2.5):
The certainty equivalent is the value of the lottery. It is a function (2.3) and (2.5) yields
. Comparing conditions
The certainty equivalent of a lottery equals its expected gain minus the risk premium evaluated at the expected wealth of the agent. It implies in particular that an increase in risk aversion reduces certainty equivalents.
2.4 The Arrow-Pratt Approximation Let us examine the characteristics of the risk premium for small risks. To do this, let us consider a pure risk . Let g(k) denote its associated risk premium
, which is
We want to characterize the properties of g around k = 0. Observe that g(0) = 0. If we assume that u is twice differentiable, fully differentiating the equality above with respect to k allows us to write
and so g′(0) = 0 since
. Differentiating once again with respect to k yields
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_21.html (1 of 2) [4/21/2007 1:27:57 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_21.html
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_21.html (2 of 2) [4/21/2007 1:27:57 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_22.html
Page 22 This implies that
Finally, using a Taylor expansion of g around k = 0, we obtain that
or equivalently,
This result is the so-called Arrow-Pratt approximation. It allows us to disentangle the characteristics of risk and preferences to evaluate the impact of the former on welfare. In fact the Arrow-Pratt approximation should have been called the "de Finetti-Arrow-Pratt" approximation, for Bruno de Finetti (1952) discovered it first. By this approximation, we learn that the risk premium for a small pure risk is approximately proportional to its variance. As the size k of this risk approaches to zero, its risk premium approaches zero as k2. This property is called "second-order risk aversion" by Segal and Spivak (1990). By condition (2.6), it implies that
For small k the certainty equivalent of risk equals kμ, and the riskiness of the situation does not matter. Second-order risk aversion means that risk yields a second-order effect on welfare compared to the effect of the mean of the corresponding lottery. This is an important property of expected utility theory when the utility function is differentiable. We now turn to the analysis of the relationship between the risk premium and preferences. The Arrow-Pratt approximation (17.9) tells us that the risk premium is approximately proportional to the coefficient of absolute risk aversion measured at the initial wealth. Roughly speaking, A(w0) measures the maximum amount that an agent with wealth w0 and utility u is ready to pay to get rid of a small risk with a variance of 2. Because the variance has the dimensionality of the square of the monetary unit, A has the dimensionality of the inverse of the monetary unit. In figure 2.1 we draw the risk premium as a function of the size of the risk. We took the logarithmic utility function as an illustra-
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_22.html [4/21/2007 1:27:57 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_23.html
Page 23
Figure 2.1 Risk premium (thin solid line) and its approximation (thick solid line) as a function of the size k of the risk
tion. We assumed that w0 = 10 and that
takes value +1 and –1 with equal probabilities. The Arrow-Pratt
approximation gives us , which is represented by the tick curve in the figure, whereas the true risk premium is the thinner curve. We see that this approximation is excellent as long as k is not larger than, say, four. It underestimates the actual risk premium for larger sizes of the risk. Notice that the true risk premium tends to 10 ! when the k tends to 10, but the approximation gives only One can also examine the risk premium associated to a multiplicative risk, that is when final wealth product of initial wealth and a random factor
is the
. We define the relative risk premium
as the maximum share of one's wealth that one is ready to pay to escape a risk of losing a random share
If
of one's wealth:
and k is small, then the Arrow-Pratt approximation can be rewritten as
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_23.html [4/21/2007 1:27:58 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_24.html
Page 24 The relative risk premium is approximately proportional to
which is called the coefficient of relative risk aversion. It has the advantage over the coefficient of absolute risk aversion to be independent of the monetary unit for wealth.
2.5 Decreasing Absolute Risk Aversion We now turn to the relationship between the risk premium and initial wealth. It is widely believed that the more wealthy we are, the smaller is the maximum amount we are ready to pay to escape a given additive risk. This corresponds to the notion of decreasing absolute risk aversion (DARA), which is formally defined as follows. DEFINITION
3 Preferences exhibit decreasing absolute risk aversion if the risk premium associated to any risk is a
decreasing function of wealth:
for any
.
From the Arrow-Pratt approximation, we know that this is true for small risks if and only if A(w0) be decreasing in w0. We prove now that it is in fact the necessary and sufficient condition for π to be decreasing in w0, independent of the size of
. Fully differentiating condition (2.3) with respect to w0 yields
This is negative if wealth if the following property holds:
, where
. Thus the risk premium is decreasing in
This condition is formally equivalent to condition (2.4) with u2 ≡ u and u1 ≡ –u′. But we know that the necessary and sufficient condition for (2.4) is that u1 be more risk-averse than u2. In consequence the necessary and sufficient condition for the risk premium to be decreasing in wealth is that –u′ be more concave than u. Let
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_24.html [4/21/2007 1:27:58 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_25.html
Page 25 denote the degree of concavity of –u′. P is called the degree of absolute prudence (which is analyzed in more detail in chapter 15). Function –u′ is more concave than function u if the degree of concavity of –u′ is uniformly larger than the degree of concavity of u, meaning if
This is equivalent to the condition that A be decreasing in wealth, since it is easily verified that
We summarize these findings in the following proposition: PROPOSITION
4 Suppose that u is three times differentiable. The following conditions are equivalent:
1. The risk premium is a decreasing function of wealth. 2. The absolute risk aversion is decreasing with wealth, meaning that –u″(z)/u′(z) is decreasing. 3. –u′ is a concave transformation of u: –u′″(z)/u″(z) ≥ –u″(z)/u′(z) for all z. Notice that DARA requires that –u′ be sufficiently concave, or that u′ be (sufficiently) convex. A similar exercise could be performed on the notion of relative risk aversion. We would find that the relative risk premium for a given multiplicative risk is decreasing in wealth if and only if relative risk aversion is decreasing in wealth. However, there is no clear evidence in favor of such a hypothesis, contrary to what we have for DARA. Notice that decreasing relative risk aversion is stronger than DARA, since
The intuition is that as wealth increases, the (multiplicative) risk premium.
increases, which tends to raise the risk
2.6 Some Classical Utility Functions It is often the case that problems in the economics of uncertainty are intractable if no further assumption is made on the form of the utility
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_25.html [4/21/2007 1:27:59 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_26.html
Page 26 function. A specific subset of utility functions has emerged as being particularly useful to derive analytical results. This is the set of utility functions with harmonic absolute risk aversion (HARA). A function exhibits HARA if the inverse of its absolute risk aversion is linear in wealth. The inverse of absolute risk aversion is called absolute risk tolerance and is denoted T with
Utility functions exhibiting an absolute risk tolerance that is linear in wealth take the following form:
These functions are defined on the domain of z such that
. We then have that
To ensure that u′ > 0 and u″ < 0, we need to have ζ(1 – γ)γ–1 > 0. The different coefficients related to the attitude toward risk are thus equal to
Observe that the inverse of absolute risk aversion is indeed a linear function of z. By changing parameters η and γ, we can generate all utility functions exhibiting an harmonic absolute risk aversion. Three special cases of file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_26.html (1 of 2) [4/21/2007 1:27:59 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_26.html
HARA utility functions are well known:
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_26.html (2 of 2) [4/21/2007 1:27:59 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_27.html
Page 27 • Constant relative risk aversion (CRRA). If η = 0, then R(z) equals γ, which must be assumed to be nonnegative. Thus the case η = 0 corresponds to the situation where relative risk aversion is independent of wealth: the relative risk premium that one is ready to pay to escape a multiplicative risk does not depend on . This implies that this set wealth. Using condition (2.16), one can select ζ in such a way to normalize of utility functions is defined as
which means that
Some of these functions are represented in figure 2.2. It is noteworthy that the asymptotic properties of these functions are quite different depending upon whether relative risk aversion is larger or smaller than unity. When γ is less than unity, utility goes from 0 to infinity as wealth goes from zero to infinity. But when γ is larger than 1, utility comes from –∞ to zero. Notice that all these functions exhibit DARA, since A′(z) = –γ/z2. • Constant absolute risk aversion (CARA). We see from (2.18) that absolute risk aversion A(z) is independent of z if γ → + ∞. We obtain A(z) = A = 1/η in this case. It is easy to check that the set of functions u that satisfy the differential equation
corresponds to
All these functions exhibit increasing relative risk aversion. • Quadratic utility functions. We obtain quadratic utility functions by selecting γ = –1. There are two problems with this set of functions. First, they are defined only on the interval of wealth z < η, since they are decreasing above η. Second, they exhibit increasing absolute risk aversion, which is not compatible with the observation that risk premia for additive risks are decreasing with wealth. We want to stress here that these functions have been considered in the literature because they are easy to manipulate. There is no obvious reason to believe that they represent the attitude toward risk of agents in the real world.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_27.html [4/21/2007 1:28:00 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_28.html
Page 28
Figure 2.2a CRRA functions (γ < 1)
Figure 2.2b CRRA functions (γ > 1)
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_28.html [4/21/2007 1:28:00 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_29.html
Page 29
Figure 2.3 CARA utility functions
Another particularly useful class of utility function is represented by piecewise linear functions with a single kink. They take the following form:
This function is concave if a ≤ 1. Absolute risk aversion is zero everywhere except at kink z0 where it is not defined, since u is not differentiable at z0. The only utility functions that are more risk-averse than u are those similar to u, but with a constant a being replaced by b < a. One can verify that the latter function is a concave transformation of the former. Notice that a risk-neutral agent who has a gross income z that is taxed at a marginal rate 1 – a above z0 will have the same behavior toward risk as an agent with utility u who is not taxed. More generally, this analogy tells us that an increasing marginal tax rate increases the aversion toward risk of the taxpayer.
2.7 Test for Your Own Degree of Risk Aversion. Several authors tried to build the utility function of economic agents through questionnaires and tests in laboratories. The kind of questions was as follows:
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_29.html [4/21/2007 1:28:01 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_30.html
Page 30 Suppose that your wealth is currently equal to 100. This wealth is composed of a sure asset and a house whose value is 40. There is a probability p = 5% that a fire totally destroys your house. How much would you be ready to pay to fully insure this risk? If we normalize the utility function in such a way that u(100) = 1 and u(60) = 0, an answer k to this question would allow us to compute
Changing the value of p in the questionnaire would then allow to draw curve u point by point. However, the sensitivity of the answer to the environment and the difficulty to make the test realistic make the results relatively weak. This direction of research has long been abandoned. A less ambitious goal has been pursued, which is simply to estimate the local degree of risk aversion of economic agents, rather than to estimate their full utility function. This has been done via market data where the actual decisions made by agents have been observed. Insurance markets and financial markets are particularly useful, since risk and risk aversion play the driving role for the exchanges in those markets. Because we did not yet link decisions of agents in these markets to their degree of risk aversion, we postpone the presentation of these analyses to chapter 5. Still, you will not be in a situation to apply any theory presented here to your own life if you have no idea of your own degree of risk aversion. Our aim is thus to offer a very simple test—similar to the one in Barsky, Juster, Kimball, and Shapiro (1997)—that allows you to make a crude estimation of your degree of risk aversion. Also, for the discussion of this book, it is important that you have some idea of what is a reasonable degree of risk aversion. We stick to an estimation of relative risk aversion because this index has dimension. Since we will not attempt to estimate the functional form of u, let us make an assumption on it. More specifically, let us assume that your utility function takes the form (2.21). As said earlier, this is a strong assumption with no serious empirical justification. This assumption is done just for simplicity. Then, answer the following question: What is the share of your wealth that you are ready to pay to escape the risk of gaining or losing a share α of it with equal probability? Think about it for a few seconds before pursuing this reading.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_30.html [4/21/2007 1:28:01 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_31.html
Page 31 Table 2.1 Relative risk premium associated to the risk of gaining or losing (α% of wealth) RRA
α = 10%
α = 30%
γ = 0.5 γ=1 γ=4 γ = 10 γ = 40
0.3 0.5 2.0 4.4 8.4
2.3 4.6 16.0 24.4 28.7
Table 2.1 allows you to translate your answer, , to your degree of relative risk aversion (RRA) γ, if α = 10% or α = 30%. You can do the computation by yourself by solving the following equation:
If we focus on the risk of gaining or losing 10% of one's wealth, we would consider an answer as something reasonable. It implies that it is reasonable to believe that relative risk aversion is somewhere between 1 and 4. Saying it differently, a relative risk aversion superior to 10 seems totally foolish, as it implies very high relative risk premia. Notice in particular that a relative risk aversion of 40 implies that one would be ready to pay as much as 8.4% to escape the risk of gaining or losing 10% of one's wealth! Notice that a side product of this exercise is a test on the validity of the Arrow-Pratt approximation (2.10) which states that the relative risk premium must be approximately equal to half the product of the variance of the multiplicative risk and γ. For α = 0.1, the variance of the multiplicative risk equals 0.01. For γ = 1, it would predict a equaling to 0.005 which is the true value, at least for the first three decimals. For γ = 10, it would , which is 10% larger than the true value. The reader can verify that the Arrow-Pratt predict approximation entails a larger error for the larger risk α = 30%. Kandel and Stambaugh (1991) pointed out that this introspection is very sensitive to the size of the gamble under consideration. Consider an agent whose current wealth is $100,000 and a gamble to loose or win X with equal probabilities. On the one hand, if X = $30,000, then γ = 40 implies that the agent will pay about $28,700 to avoid the gamble. As explained above, this suggests a lower value of γ, such as γ = 2, which implies a more reasonable payment of $9,000. On the
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_31.html [4/21/2007 1:28:02 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_32.html
Page 32 other hand, if X = $500, then γ = 2 implies that the investor is willing to pay only $2.5 to avoid this gamble. This small risk premium is another illustration of second order risk aversion. Kandel and Stambaugh found this payment unrealistically small. They suggested that a large γ around 30 would be better to model the behavior towards small risks. Our viewpoint on this is that agents are almost risk-neutral toward risks that are small with respect to their aggregate (lifetime) wealth. After all, we don't insure a lot of small risk of life, and most people purchase deductible insurance. To end up, it should be noticed that this kind of tests exhibited important differences in the degree of risk aversion within the society. Observable characteristics of agents, as wealth, social status, gender, occupation and the like, have an important impact on the attitude toward risk. In particular, women are more risk-averse than men, on average (Schubert, Brown, Gysler, and Brachinger 1999).
2.8 An Application: The Cost of Macroeconomic Risks Let us consider a very simple society in which every agent would be promised the same share of the aggregate production in the future. This means that the risk borne by every agent is measured by the uncertainty affecting the growth of GDP per head. All other risks are washed away by diversification. How this can be made possible in more complex, decentralized economies is the subject of chapters 21 and 22. In order to assess the degree of risk associated to every one's income in the future, let us examine how volatile the growth of GDP per capita as been in the past. We reproduce in figure 2.4 the time series of US GDP per capita in 1985 dollars for the period from 1963 to 1992. By long-term historical standards, per capita GDP growth has been high: around 1.86% per year. However, this growth rate has been variable over time, as shown in figure 3.7. This series appear stationary and ergodic. The standard deviation of the growth rate has been around 2.41%. Suppose that we attach an equal probability that any of the annual growth rate observed during the period 1963 to 1992 occurs next year. That is, transform frequencies into probabilities. Let be the random variable with such probability distribution. If we assume that agents have constant relative risk aversion γ, one can estimate the certainty equivalent growth rate, gc, by using the following equality:
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_32.html [4/21/2007 1:28:03 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_33.html
Page 33
Figure 2.4 US Real GDP per capita in 1985 dollars (source: Penn World)
Figure 2.5 Growth rate of the US real GDP per capita, in %
The social cost of risk can be measured by the risk premium associated to . It is the reduction in the growth rate one would be ready to pay to get rid of the macroeconomic risk. We computed the certainty equivalent growth rate and the social cost of risk for different values of γ. The results are reported in table 2.2. We see that it does file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_33.html (1 of 2) [4/21/2007 1:28:03 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_33.html
not differ much from the expected growth rate for acceptable values of γ. If we take γ = 4 for a reasonable upper bound, the macroeconomic risk does not cost us more than one-tenth of a percent of growth. The
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_33.html (2 of 2) [4/21/2007 1:28:03 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_34.html
Page 34
Figure 2.6 Frequency table of the growth of real US GDP per capita, 1963 to 1992
Table 2.2 Social cost of the macroeconomic risk Certainty equivalent growth rate (%)
Social cost of macroeconomic risk (%)
1.86 1.85 1.83 1.74 1.56 0.50
— 0.01 0.03 0.12 0.30 1.36
RRA γ=0 γ = 0.5 γ=1 γ=4 γ = 10 γ = 40
implication of this is that not too much should be done to reduce this uncertainty. Policy makers should rather concentrate their effort in maximizing the expected growth. The existing macroeconomic risk is just not large enough to justify spending much of our energy to reduce it. This point has been made first by Lucas (1987).
2.9 Conclusion file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_34.html (1 of 2) [4/21/2007 1:28:03 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_34.html
Two important notions have been introduced in this chapter. The first one is risk aversion, which means that one rejects any zero-mean risk, whatever one's initial wealth. Risk aversion is equivalent to the
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_34.html (2 of 2) [4/21/2007 1:28:03 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_35.html
Page 35 concavity of the utility function. We also determined whether an agent is more risk-averse than another, that is, whether the first rejects any lottery for which the former is indifferent, whatever their identical initial wealth level. This is equivalent to requiring that the utility function of the first is a concave transformation of the utility function of the second. We showed how to measure the degree of risk aversion of an agent by using a simple thought experiment.
2.10 Exercises and Extensions Exercise 4 Risk aversion is a concept that is independent of the expected utility model. As an illustration of this, consider the rank-dependent EU model described by equation (1.4). Assume that u(z) = z. Characterize the set of functions g that makes the decision maker risk-averse (see Cohen 1995). Exercise 5 Show that
Exercise 6 Consider the utility function (2.23), with a positive constant a < 1. • Show that u is concave. • Show that u exibits first-order risk aversion at z = z0, namely that
tends to zero when k tends to zero
as k, rather than as k2. • Show that a reduction in the constant a increases the degree of risk aversion, in the sense of Arrow-Pratt. • Does it satisfy DARA? Exercise 7 Consider any pair of i.i.d. random variables is the payoff that one gets per dollar invested in the investment project i. You are allowed to allocate your initial wealth freely in these two investment projects. • Show that it is always optimal to split your initial wealth equally to invest in the two projects, under risk aversion. • Show that it is always optimal to invest 100% of your initial wealth in one of the two risky assets if you are a risk-lover.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_35.html [4/21/2007 1:28:04 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_36.html
Page 36 • Estimate the welfare gain (in certainty equivalent) with respect to an economy in which people cannot split their wealth into different investment projects. Exercise 8 Take the Machina triangle and draw the indifference curves of a risk-averse agent. Exercise 9 Consider a pair of utility functions (u, v) and an increasing and convex function f such that v(z) = u(f (z)) for all z. • Exhibit the condition on (u, v, f) that makes v more concave than u in the sense of de Finetti-Arrow-Pratt (see Ross 1999). • Consider case f(z) = (1 – t)z, with t ∈ [0, 1]. Exhibit the condition for this linear tax to reduce risk aversion. Examine, in particular, the HARA case (see also Arrow 1965). • Consider a tax on capital earnings which takes the following form: capital gains are taxed at t%, but capital losses are not compensated. How the introduction of such a tax affects the certainty equivalent of an uncertain capital income? • Consider function f(z) = exp z. Show that the absolute risk aversion of v and the absolute risk aversion of u are related in the following way:
Exercise 10 The Arrow-Pratt formula (17.9), if it would hold for all risks, would make the EU model a particular case of the mean-variance (MV) model. The theory of finance usually limits the analysis to the mean and variance by restricting the set of utility functions and distribution functions that make the Arrow-Pratt approximation exact. Show that the Arrow-Pratt formula is exact when the utility function is CARA and is normally distributed. Exercise 11 Using condition (2.7), show that the risk premium is an increasing function of the size k of risk . Exercise 12 The risk premium π is the maximum price that one is ready to pay to escape an endowed pure risk. The compensating risk premium p is the minimum compensation that induces one to accept a pure risk. Namely is defined by
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_36.html [4/21/2007 1:28:05 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_37.html
Page 37
• Show that Vallée 1968).
is smaller than
for all
such that
if and only if u is DARA (see La
• Show that the compensating risk premium is convex in the size of the risk, that is, convex in α. This is not necessarily the case for the risk premium (see Eeckhoudt and Gollier 1998).
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_37.html [4/21/2007 1:28:05 AM]
is
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_38.html
Page 38
This page intentionally left blank.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_38.html [4/21/2007 1:28:05 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_39.html
Page 39
3 Change in Risk In the previous chapter we examined the notion of risk premium, which measures the cost of the introduction of risk. The risk premium π basically depends upon the characteristics of the utility function and the characteristics of the risk. We also characterized the changes in the utility function that have an unambiguous effect on the risk premium, independently of the risk under consideration. This corresponded to the concept of "more risk aversion." In this chapter we perform the symmetric exercise, which is to look for changes in risk that have an unambiguous effect on the risk premium, independently of the utility function under consideration. Without loss of generality, let us normalize w0 to zero. Observe that
is larger than
if and
only if
The theory of stochastic dominance is devoted to determining under which condition on the distribution functions and the inequality above holds for any utility function u in a function set . Of course the answer to this of question depends on which set is considered. This will generate various stochastic dominance orders. At this stage it is important to observe that as soon as contains more than one utility function, the associated stochastic dominance ordering is incomplete in the sense that there exist pairs such that does not dominate and does not dominate . This is the case when two agents in diverge on their ordering of these two risks so that no unanimity emerges. To escape potential problems of convergence of the integrals, we limit the analysis to random variables whose supports are in interval [a, b].
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_39.html [4/21/2007 1:28:06 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_40.html
Page 40
3.1 The Extremal Approach. The general technique used in stochastic dominance is based on the fact that the expectation operator is additive. This means that
where uλ(z) = (1 – λ)u0(z) + λu1(z) for all z. More generally, this means that if condition (3.1) is satisfied for some utility functions (u0, u1, . . . , un), it is also satisfied for any utility function which is a convex combination of them. The extremal approach consists in searching for a small subset B of such that any element in can be written as a convex combination of elements in B. In our terminology, B is called a basis for . Requiring that condition (3.1) holds for elements in B will not only be necessary but also sufficient for condition (3.1) to hold for any element in . If the dimensionality of B is smaller than the dimensionality of , testing whether (3.1) holds for all u ∈ is made simpler. It will be enough to check the inequality for the subset of u is B. Let us illustrate the extremal approach with the following simple example. Suppose that is the set of increasing, concave piecewise linear functions with two kinks, one at z0 and the other at z1, and such that the level of utility is constant for wealth levels larger than
Figure 3.1 Function u as a convex combination of functions u0 and u1 file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_40.html [4/21/2007 1:28:06 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_41.html
Page 41 z1. That is,
is the set of functions u such that
This function is parametrized by a scalar a. contains an infinite number of functions, but its dimensionality is 1. Notice that any function in is a convex combination of u0(.) = min(., z0) and u1(.) = min(., z1). These two basic functions together with a function u of type (3.2) are depicted in figure 3.1. These two functions represent a basis for since function (3.2) equals (1 – a)u0 + au1. We conclude that it is enough to verify that
and
to guarantee that all functions of type (3.2) satisfy condition (3.1). In general, the function set is not as simple as in the example above. In the remaining of this chapter, b(., θ) will denote a function in the basis, where θ is an index in some index set Θ ⊂ R. It means that for any utility function in set , there exists a nondecreasing function H: Θ → R+ such that
for all z ∈ [a, b], where H satisfies the condition that
H is the transform of u with respect to basis B. dH(θ) can be interpreted as the weight associated to function b(., θ). It must be positive and it must sum up to 1. In fact, when u admits a bounded first derivative, H can be interpreted as a cumulative distribution function on a random variable , and u can be written as
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_41.html (1 of 2) [4/21/2007 1:28:07 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_41.html
To sum up, we can rewrite the initial problem as
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_41.html (2 of 2) [4/21/2007 1:28:07 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_42.html
Page 42 for all random variable
whose support is in Θ. The necessary and sufficient condition is written as
This is another way to say that by the subset of agents in B.
is unanimously preferred to
by population
if it is unanimously preferred
3.2 Second-Order Stochastic Dominance In this section we determine the conditions under which inequality (3.1) holds for any increasing and concave utility function. In that case we say that dominates in the sense of second-order stochastic dominance (SSD). This concept has first been examined by Hadar and Russell (1969) and Rothschild and Stiglitz (1970). Let 2 be the set of increasing and concave functions from [a, b] to R. It is a convex set. As observed for example by McFadden (1989), a simple basis for
2
is the set B2 of min-functions:
To show that any increasing and concave function can be expressed as a convex combination of min functions, let us take an arbitrary u ∈ 2. The weighting function H must satisfy
for any z ∈ [a, b]. Integrating by parts yields
where I(x) is the indicator function which takes value 1 if x is true, and 0 otherwise. Differentiating the above equality with respect to z yields u′(z) = H(b) – H(z). If u is twice differentiable, taking H′(θ) = –u″(θ) will give us the weight function. At wealth levels z where u′ is not differentiable, H discontinuously jumps by u′(z–) – u′(z+), which corresponds to an atom in the distribution of . Following the general argument presented in the previous section, this implies that the necessary and sufficient condition (3.3) is written here as
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_42.html [4/21/2007 1:28:08 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_43.html
Page 43
Let Fi denote the cumulative distribution function of
. Then the inequality above is equivalent to
or, integrating by parts, to
This is in turn equivalent to
This set of inequalities is called the Rothschild-Stiglitz's (1970) integral condition for SSD, although this condition already appeared in Hardy, Littlewood, and Polya (1929) and Hadar and Russell (1969). This condition means that all min-utility individuals prefer over . This is sufficient to guarantee that all risk-averse agents prefer over . Notice that a necessary condition for a SSD deterioration of a risk is that the mean is not increased. This is a consequence of definition (3.1) where we take u as the identity function, which is in 2. This can also be seen by recalling the following statistical relationship that can be obtained by integrating by parts:
Combining this with condition (3.4) for θ = b yields the result. A limit case is when the expectation of the risk is unchanged by the shift in distribution. A SSD deterioration in risk that preserves the mean is called an increase in such that is an increase in risk with respect to risk (IR). We present in figure 3.2 an example of a pair . Observe that we split the probability mass at 100 into two equal probability masses respectively at 50 and 150. This is an example of a mean-preserving spread (MPS) of probability mass. More generally, define an interval X ⊂ [a, b]. A MPS is obtained by taking some probability mass from X to be distributed outside X, in such a way to preserve the mean. In the case of a continuous
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_43.html [4/21/2007 1:28:08 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_44.html
Page 44
Figure 3.2 as an increase in risk of
random variable, a spread around X is defined by the condition that is positive (resp. negative) outside (resp. inside) X. The reader can easily verify that this is a sufficient condition for (3.4). Rothschild and Stiglitz (1970) showed that any increase in risk can be obtained by a sequence of such mean-preserving spreads. is obtained from by adding Another way to look at the change in risk presented in figure 3.2 is to say that , with being distributed as (–50, 1/2; 50, 1/2). More generally, if a white noise to the low realization of
a direct use of Jensen's inequality yields
Thus condition (3.5) is sufficient for SSD. Rothschild and Stiglitz (1970) showed that it is also necessary. To sum up, an increase in risk from
to
can be defined by any of four equivalent statements:
• The means are the same and risk-averse agents unanimously prefer .
over
: ∀u concave:
• The Rothschild-Stiglitz's condition (3.4) is satisfied; it is satisfied as an equality for θ = b. •
is obtained from
by a sequence of mean-preserving spreads.
•
is obtained from
by adding a white noise
to it, with
for all x.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_44.html (1 of 2) [4/21/2007 1:28:09 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_44.html
Several other preference orderings than expected utility with u″ < 0 are such that agents unanimously dislike a SSD change in distribution. Such preference orderings are said to satisfy the SSD property.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_44.html (2 of 2) [4/21/2007 1:28:09 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_45.html
Page 45
3.3 Diversification There are many ways to reduce risk. The most well-known one is diversification. Consider a set of n lotteries and are assumed to be independent and identically whose net gains are characterized by distributed. You are offered to participate to these gambles. If you are allowed to split your stake in different gambles, how should you do it? Common wisdom suggests that one should not "put all one's eggs in the same basket." This is because diversification is an efficient way to reduce risk, as we now show. A feasible strategy in this game is characterized by a vector A = (α1, . . . , αn) where αi is one's share in gamble i, with . It yields a net payoff equaling a generalization of Rothschild and Stiglitz (1971).
. We have the following proposition, which is
5 The distribution of final wealth generated by the perfect diversification strategy dominates the distribution of final wealth generated by any other feasible strategy. PROPOSITION
SSD-
denote final wealth under the perfect diversification strategy. Consider now any alternative feasible Proof Let strategy A = (α1, . . . , αn). We have that
We would be done if noise
By symmetry,
has a zero mean conditional to
. We have that
is independent of i. Let us denote it k. It implies that
This proves that the alternative strategy is second-order stochastically dominated by the perfect diversification strategy. file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_45.html [4/21/2007 1:28:09 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_46.html
Page 46 It implies that all risk-averse agents will select the perfect diversification strategy. At the limit, when n tends to infinity, this strategy allows to replace the unit risky gamble by its expectation. This is the Law of Large Numbers.
3.4 First-Order Stochastic Dominance A more restrictive stochastic order than SSD is obtained by enlarging the set
2
satisfy condition (3.1). This is the case when we allow risk-loving behaviors. Let
of utility functions that must 1
denote the set of all
increasing functions from [a, b] to R. If condition (3.1) holds for all utility functions in
1,
we say that
dominates in the sense of the first-order stochastic dominance (FSD) order. Using the same technique as above, it is straightforward to verify that the set B1 of step-functions
is a basis for
1.
For a differentiable utility function, this can be seen by observing that
For utility function b(., θ) ∈
1,
the expected utility of lottery
equals
The expected utility of a lottery can be read as the probability that this lottery exceeds threshold θ. It implies that is preferred to for all expected-utility maximizers (risk-averse or risk-lover) if and only if is preferred to for all expected-utility maximizers with a step utility function, that is, if
In words, the change in risk must increase the probability that the realized value of the risk be less than a threshold, whatever the threshold. Such a change in risk can be obtained by shifting probability masses to the left, or by adding a noise
to the original risk
such that
for all x.
A particular case of FSD shifts in distribution is when k(θ) = F1(θ)/F2(θ) is nonincreasing in θ. This property file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_46.html (1 of 2) [4/21/2007 1:28:10 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_46.html
defines the monotone
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_46.html (2 of 2) [4/21/2007 1:28:10 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_47.html
Page 47
Figure 3.3 Random variables dominated by A in the machina triangle
probability ratio (MPR) stochastic order. Since by definition k(b) equals 1, the fact that k is nonincreasing implies that k(θ) ≥ 1 for all θ less than b. This is equivalent to the FSD condition (3.7). A more restrictive condition is the so-called monotone likelihood ratio (MLR) property. Limiting our analysis to random variables having a density which is positive on [a, b],2 this stochastic dominance order is defined by be nonincreasing. We show in chapter 7 that MLR is a special case of the condition that MPR. When risks have support in {x1, x2, x3}, we can use the Machina triangle to represent all lotteries that are dominated by lottery A with respect to the dominance orders defined in this chapter. This is done in figure 3.3 by assuming without loss of generality that x1 < x2 < x3.
3.5 Concluding Remark In this chapter we showed that a utility function—when it has a bounded first derivative—is basically the same object as a random variable. Indeed, using a basis {b(., θ)}θ ∈ Θ of the set of functions under consideration, to each utility function in the function set generated by the basis, there exists an associated "random variable" such that
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_47.html (1 of 2) [4/21/2007 1:28:11 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_47.html
2. Green (1987) provides some more insights on this argument.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_47.html (2 of 2) [4/21/2007 1:28:11 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_48.html
Page 48 It implies that expected utility can be written as
Our objective here is to show that the notions of change in risk developed in this chapter are linked to the notions of change in preference developed in the previous chapter. For example, consider the set 1 of increasing utility functions, where b(x, θ) = I(x ≤ θ) and H(θ) = u(θ), with the normalization u(a) = 0, u(b) = 1. In this case the for all θ is dual to the notion monotone likelihood ratio order which states that . The symmetry is not perfect, of more risk aversion, which states that however. While it is economically meaningful to address the problem of the effect of a change of F on EU,
there is no sense to determine the effect of a change of H on U:
or . This is because expected utility is ordinal. It does not allow to make interpersonal comparisons of welfare. Otherwise, all results that have been presented in this chapter could have been used to in the case of 2! make such comparisons. Just replace Fi by ui in the case of 1, or by
3.6 Exercises and Extensions Exercise 13 Let H be the transform of u in the basis B1 of
1.
Let
be the random variable whose cumulative
function is H (we assume that u(a) = 0 and u(b) = 1). Show that the expected utility of a lottery the probability that this lottery outperforms lottery . (Castagnoli and Li Calzi 1996).
can be read as
Exercise 14 Find the basis for the convex set Ψ3 = {u | u′ ≥ 0, u″ ≤ 0, u′″ ≥ 0}. Characterize the stochastic dominance order associated to it (third-order stochastic dominance; see Whitmore 1970). Exercise 15 Consider now the set ΨD of utility functions exhibiting decreasing absolute risk aversion.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_48.html [4/21/2007 1:28:11 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_49.html
Page 49 • Show that ΨD is a convex set (Pratt 1964). • Show that any piecewise exponential function with a decreasing (stepwise) absolute risk aversion profile must be in the basis (Gollier and Kimball 1997). • Can we find a simple basis for
3
(Vickson 1975)?
Exercise 16 Pratt and Zeckhauser (1987) and Caballe and Pomansky (1996) consider the set ΨC of "completely monotone" utility functions whose basis is the set of CARA functions. In other words, ΨC is the set of Laplace transform of probability distributions. • Show that all functions in ΨC have all positive even derivatives and all negative odd derivatives. In fact, they are the only functions with such a property. • Show that power functions and the logarithm belong to ΨC. • Characterize the stochastic dominance order associated to ΨC. • Show that it is weaker than SSD. Exercise 17 Meyer (1977) introduces the concept of stochastic dominance with respect to a function. Consider any increasing function u1 and define the associated function set . • Show that
(u1) is the set of functions that are more risk-averse than u1.
(u1) is a convex set.
• Find a simple basis for
(u1), and derive the integral condition for the corresponding stochastic dominance
order. Exercise 18 Consider an economy where all agents face an idiosyncratic risk to lose 100 with probability p. N agents decide to create a mutual agreement where the aggregate loss in the pool will be equally split among its members. • Describe the residual loss of each member of the pool when N = 2 and N = 3.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_49.html (1 of 2) [4/21/2007 1:28:12 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_49.html
• Show that an increase from N = 2 to N = 3 generates a reduction in risk borne by each member of the pool, in the sense of SSD. Exercise 19 Consider a risk-neutral agent who owns a call option on an underlying asset whose value at the exercise date is . The call
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_49.html (2 of 2) [4/21/2007 1:28:12 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_50.html
Page 50 option gives to the agent the option to purchase the asset at a prespecified price—the exercise price—y. • Show that the agent likes an increase in risk in the value of the underlying asset. • How your answer would change if the agent owns n > 1 different call options, respectively with exercise price y1, . . . , yn?
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_50.html [4/21/2007 1:28:12 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_51.html
Page 51
II The Standard Portfolio Problem
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_51.html [4/21/2007 1:28:13 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_52.html
Page 52
This page intentionally left blank.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_52.html [4/21/2007 1:28:13 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_53.html
Page 53
4 The Standard Portfolio Problem. One of the classical problems in the economics of uncertainty is the standard portfolio problem. An investor has to determine the optimal composition of his portfolio containing a risk-free and a risky asset. This is a simplified version of the problem of determining whether to invest in bonds or equities. In fact the structure of this problem may be applied to several problems in which a risk-averse agent has to choose the optimal exposure to a risk. This is the case, for example, when an insured person has to determine the optimal coinsurance rate to cover a risk of loss, or when an entrepreneur has to fix his capacity of production without knowing the price at which he will be able to sell the output. The standard portfolio problem has multiple applications in the day-to-day life. In this chapter we examine the static version of this model. This model has first been analyzed by Arrow (1963) and Pratt (1964). Several extensions to this basic model will be considered later on.
4.1 The Model and Its Basic Properties Consider an agent with an increasing and concave utility function u ∈
2.
He has a sure wealth W that he can
invest in a risk-free asset and in a risky asset. The return of the risk-free asset is r. The return of the risky asset over the period is a random variable . The problem of the agent is to determine the optimal composition (W – α, α) of his portfolio, where W – α is invested in the risk-free asset and α is invested in the risky asset. The value of the portfolio at the end of the period is written as
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_53.html [4/21/2007 1:28:13 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_54.html
Page 54 where w0 = W(1 + r) is future wealth of the risk-free strategy and
is the excess return of the risky
asset. Let F be the cumulative distribution of random variable . We do not consider here the existence of shortsale constraints; that is, α is allowed to be larger than W or less than zero. The problem of the investor is to choose α in order to maximize the expected utility V(α) :
Let us assume that u is differentiable. The first-order condition for this problem is written as
where α* is the optimal demand for the risky asset. Notice that this is a well-behaved maximization problem, since the objective function V is a concave function of the decision variable α under risk aversion:
Since V is concave, the sign of V′(0) determines the sign of α*. Observe that
In consequence and α* have the same sign. In addition it is optimal to invest everything in the risk-free asset . This later result is obvious: buying some of the risky asset generates an increase in risk of the value of if the portfolio. When the expected excess return of the risky asset is positive, the optimal portfolio is a best compromise between the expected return and the risk. 6 In the standard portfolio problem with a differentiable utility function, risk-averse agents invest a positive (resp. negative) amount in the risky asset if and only if the expected excess return is positive (resp. negative). PROPOSITION
This is an important and somewhat surprising result. One could have conjectured that if the excess return be positive in expectation but very uncertain, it could be better not to invest in the risky asset. This is not possible: as soon as the expected excess return is positive, even if very small, it is optimal to be exposed to the risk. There is an intuition for this result. As we have seen in chapter 2, if the utility function is differentiable, the risk premium tends to zero as the square
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_54.html [4/21/2007 1:28:14 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_55.html
Page 55 of the size of the risk. This has been referred to as risk aversion being of the second order. In such a case, when considering purchasing a small amount of the risky asset, the expected benefit (return) is of the first order whereas the risk is of the second order. It is noteworthy that there is an important prerequisite in proposition 6: the utility function must be differentiable. If it is not, risk aversion can generate a first-order effect and it may be possible that α* = 0 even if . From now on, we assume that
is positive. If it is not, just replace
by
and α by –α.
Observe that the first-order condition (4.3) has no solution if does not alternate in sign. In particular, if potential realizations of are all larger than zero, V is increasing in α and the optimal solution is unbounded. This is because the risky asset has a return larger than the risk-free asset with probability 1. The risky asset is a money machine in this case. This cannot be an equilibrium. Therefore we hereafter assume that alternates in sign. Notice that this assumption alone does not guarantee that α* is bounded. Suppose that the domain of u is R. Then α* is bounded only if limα→∞ V′(α) is negative. This condition is rewritten as follows:
or equivalently,
If u′ tends to zero when wealth tends to plus infinity, or when u′ tends to +∞ when wealth tends to minus infinity, the above inequality will automatically be satisfied. A similar argument can be used when the domain of u is bounded upward or downward, as in the case of CRRA functions.
4.2 The Case of a Small Risk It can be useful to determine the solution to problem (4.2) when the portfolio risk is small. The problem with this approach is that the size of the risk is endogenous in this problem. The simplest method to escape the difficulty is to define
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_55.html [4/21/2007 1:28:15 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_56.html
Page 56
where and μ > 0. The optimal investment in the risky asset is α*(k), which is a function of k, with α*(0) = 0. When k is positive, we obtain α*(k) as the solution of the following equation:
Fully differentiating this equation yields
where
. Evaluating this expression at k = 0, we obtain
A first-order Taylor expansion of α*(k) around k = 0 yields
where α* is the solution to program (4.2) with asset is thus approximatly equal to
, or
. The relative share of wealth to invest in the risky
The optimal share of wealth to invest in the risky asset is approximately proportional to the ratio of the expectation and variance of the excess return. The coefficient of proportionality is the inverse of relative risk aversion. Let us consider a logarithmic investor (R = 1). If the excess return has an expectation of 5% and a standard deviation of 30%, the optimal portfolio contains around 5/9 in the risky asset. This seems to be a rather large proportion. We will come back to this problem when we will discuss the equilibrium price of assets. It is noteworthy that approximation (4.5) is exact if u is CARA and
is normally distributed. If
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_56.html (1 of 2) [4/21/2007 1:28:15 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_56.html
we obtain
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_56.html (2 of 2) [4/21/2007 1:28:15 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_57.html
Page 57
This shows that the Arrow-Pratt approximation is exact in this case. It implies that the optimal α is the one that maximizes αμ – (Aα2σ2)/2. This yields
4.3 The Case of HARA Functions We have just seen that the portfolio problem is much simplified if we assume that the utility function of the investor is CARA and if is normally distributed. Another type of simplification can be made if we assume that the utility function belongs to the class of HARA functions if
The value of parameters ζ, η and γ are selected in order to have u increasing and concave (see chapter 2). The first-order condition (4.3) is rewritten in this case as
Let us first solve the following problem:
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_57.html (1 of 2) [4/21/2007 1:28:16 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_57.html
Notice that a is the optimal investment if η + w0/γ = 1. It is a function of the distribution of
and of the
coefficient of concavity γ. Now Observe that
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_57.html (2 of 2) [4/21/2007 1:28:16 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_58.html
Page 58
is the general solution to equation (4.8). Indeed, we have
Thus, by solving the first-order condition for a single value of w0, we obtain the optimal solution for all values of w0. When the utility function is HARA, there is a linear relation between the optimal exposure to risk and the wealth level. Two special cases are usually considered. When the utility is CARA (γ → ∞), α* is independent of wealth. When the utility function is CRRA (η = 0), α* is proportional to wealth, that is, the share of wealth invested in the risky asset is a constant independent of wealth. Judging by empirical evidence, neither constant relative nor absolute risk aversion does seem to be a reasonable assumption. On the one hand, given CRRA, the optimal share of wealth invested in various assets would be independent of wealth. On the other hand, as Kessler and Wolff (1991) show, the portfolios of households with low wealth contain a disproportionately large share of risk free assets. Measuring by wealth, over 80% of the lowest quintile's portfolio was in liquid assets, whereas the highest quintile held less than 15% in such assets. This suggests that relative risk aversion is decreasing with wealth. It should be noticed, however, that the low share of wealth invested in stocks by the poor may alternatively be due to the existence of a fixed cost to participate to stock markets. Still CARA functions are disqualified by the fact that it would induce the relative share of wealth invested in risky assets to be decreasing in wealth! More generally, this is the case for all HARA utility functions with η > 0.
4.4 The Impact of Risk Aversion Under which condition does a change in preferences reduce the optimal exposure to risk, whatever its distribution? To answer this question, let us consider a change of the utility function from u2 to u1. By the concavity of the objective function V with respect to α, this change reduces the optimal investment in the risky asset if and only if
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_58.html [4/21/2007 1:28:16 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_59.html
Page 59 The equality to the left means that we normalized the optimal exposure to risk of u2 to unity. The inequality to the right means that
is less than unity.
It is easy to show that, according to the intuition, an increase in risk aversion in the sense of Arrow and Pratt generates such a reduction in the demand for the risky asset. Indeed, if is less than
is concave, then obviously
for all x. It implies that
is less than , which is equal to zero. This proves that an increase in risk aversion is sufficient for a reduction in the optimal exposure to risk. It can also be shown that it is necessary, in the sense that if u1 is not more risk-averse than u2, then there exists a w0 and a distribution of that violates condition (6.25). This can easily be done by contradiction. PROPOSITION
7 Consider a change in the attitude towards risk from u1 to u2. This reduces the optimal exposure to
any risk independent of initial wealth if and only if u1 is more risk-averse than u2 in the sense of Arrow and Pratt. A similar exercise can be performed on the effect of a change in w0 on the optimal structure of the portfolio. This exercise can be seen as an application of proposition 7 by defining u2(z) = u1(z – k). It directly yields the following proposition. 8 A reduction in wealth always reduces the investment in the risky asset if and only if absolute risk aversion is decreasing. PROPOSITION
Similarly a reduction in wealth always reduces the share of wealth invested in the risky asset if and only if relative risk aversion is decreasing.
4.5 The Impact of a Change in Risk The question dual to the one examined in the previous section is as follows: Under which condition does a change in the distribution of return increase the demand for the risky asset? To address this question, let us respectively with cumulative distribution function F1 and F2. In general consider two random returns and terms, the problem is to determine the conditions under which we have
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_59.html [4/21/2007 1:28:17 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_60.html
Page 60
The intuition would suggest that the demand for the risky asset would be reduced by all risk-averse agents if its return undergoes an increase in risk. Rothschild and Stiglitz (1971) and Hadar and Seo (1990) showed that this intuition is not sustained by the theory. Not even a first-order stochastically dominated shift in the distribution does necessarily generate a reduction in the demand for the asset! The origin of the problem of SSD not being sufficient is that (x) = xu′(w0 + x) is not in 2; in other words, it is not increasing and concave in x. Otherwise, being dominated by in the sense of SSD would have been sufficient to guarantee that , which is in turn sufficient for condition (4.11). Remember indeed that larger than dominated by in the sense of SSD means that
be being
Observe that
Thus, assuming that the domain of u is R+, is increasing if relative risk aversion R is less than unity. This condition will be sufficient to guarantee that any FSD-dominant change in the distribution of returns will increase the demand for the risky asset. We can also verify that
The first term in the right-hand side is negative if ′ is positive. The second term is negative if relative risk aversion is increasing and absolute risk aversion is decreasing. Notice that ″ may also be written as
where Pr(z) = zP(z) is relative prudence. This condition implies that is concave if relative prudence is less than 2, and prudence is positive. We thus conclude this section by the following proposition, which is in Hadar and Seo (1990). 9 Suppose that the domain of u is R+. Then, a shift of distribution of returns increases the demand for the risky asset if PROPOSITION
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_60.html [4/21/2007 1:28:18 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_61.html
Page 61 1. this shift is FSD-dominant, and if relative risk aversion is less than unity; 2. this shift is SSD-dominant, relative risk aversion is less than unity and increasing, and absolute risk aversion is decreasing; 3. this shift is SSD-dominant, relative prudence is positive and less than 2. The problem with this proposition is that it is rather unlikely that relative risk aversion be less than unity. Moreover, for HARA utility functions, relative prudence is less than 2 if and only if relative risk aversion is less than 1.
4.6 Concluding Remark The results in this chapter are important. The central one is that an increase in risk aversion reduces the demand for the risky asset. It is noteworthy that this condition is necessary: if the change in preference is not an increase in concavity, it is possible to find an initial wealth and a distribution of excess returns such that the investor reduces his demand. This result will frequently be used in the rest of this book. The effect of a change in distribution is much more problematic. These results are important also because the standard portfolio problem (4.2) has more than one economic interpretation. In addition to the portfolio investement problem, it can be interpreted as the problem of an entrepreneur with a linear production function who has to determine his production before knowing the competitive price at which he will be able to sell it. In that case, is the unit profit margin (output price minus marginal cost) and α is the production. This application has first been examined by Sandmo (1971). Alternatively, consider a risk-averse agent who owns an asset which is subject to a random loss . Suppose that the agent can select the share α of the risk that he will retain. The proportion (1 – α) is sold to an insurance company against the payment of an insurance premium that is proportional to the expected indemnity, which is also called the actuarial value of the policy. In consequence problem (4.2) can also be interpreted as a (proportional) insurance problem where α is the retention rate on the insurable risk. All results in this chapter can be reinterpreted for these two applications.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_61.html [4/21/2007 1:28:18 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_62.html
Page 62
4.7 Exercises and Extensions Exercise 20 Consider the standard portfolio problem with a zero return on the risk-free investment and a random return on the risky investment. • How do you react to the announcement that the return of the risky investment is not , but rather probability q and 0 with probability 1 – q?
with
• How do you react to the announcement that you cannot directly invest in the risky project, but rather in a risky asset which is a portfolio containing a proportion q of the risky investment project, and a proportion 1 – q of the risk free investment project? Exercise 21 (Mossin 1968a) You face a risk of a random loss on your capital that can be covered by an insurance contract. The contract guarantees an indemnity that is a proportion c of the incurred loss. The premium is equal to (1 + λ) times the expected indemnity, where λ is the loading factor. The coverage rate is selected by the policyholder. • Show that the decision problem of the policyholder is formally equivalent to the standard portfolio problem. • Show that the optimal rate of coverage is less than 1 if and only if λ is positive. • Show that an increase in risk aversion increases the optimal rate of insurance coverage. • Do you think that consumers facing a risk of loss should always purchase some insurance? Exercise 22 Consider a logarithmic investor (u(z) = ln z) who can invest in a risk-free asset with return r and in a risky asset whose distribution of return is (p, x_; 1 – p, x+), with x_ < r < x+. • Derive an analytical formula for the optimal demand for the risky asset. • Is it always finite? • Examine the effect of change in wealth on the demand for the risky asset.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_62.html [4/21/2007 1:28:19 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_63.html
Page 63 • Examine the effect of an increase of the risk free rate on the demand for the risky asset. • Examine the effect of a change in x_ or x+ on the optimal demand. • Show that an increase in x+ combined with a reduction of x_ that leaves the expected excess return unchanged reduces the demand. Exercise 23 Answer to the same questions as in exercise 22 when the utility function is u(z) = min(z, kz + (1 – k) z0). Exercise 24 Show that the demand for the risky asset is decreasing with the risk-free rate when absolute risk aversion is increasing (see also Fishburn and Porter 1976). Exercise 25 Examine the effect of the following tax scheme on the optimal portfolio: • Capital gains are taxed at rate τ, and capital losses are compensated at the same rate. • Capital gains are taxed at rate τ, but capital losses are not compensated (no loss offset). Exercise 26 (Eeckhoudt, Gollier, and Schlesinger 1995) Consider the so-called newsboy problem, which is generic for various inventory problems. A newsboy must order newspapers at a unit price c, before knowing the demand for news. He can resell the newspapers at a unit price p > c. All unsold newspapers are lost. • Write the maximization problem of the risk-averse newsboy. • Derive and discuss the optimal strategy of the risk-neutral newsboy. • Examine the effect of risk aversion on the optimal choice. • What is the effect of a mean-preserving spread in the distribution of the demand that is concentrated above (resp. below) the initially optimal command of newspapers? Exercise 27 (Roy and Wagenvoort 1996) Examine the condition under which the marginal propensity to invest in the risk-free asset 1 – (dα/dw0) is positive.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_63.html [4/21/2007 1:28:19 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_64.html
Page 64 Exercise 28 Consider two alternative random variables, and with , representing the distribution of returns of the risky asset. Suppose that is second-order stochastically dominated by . Show that all riskaverse expected-utility-maximizing investors prefer to live in the economy 2 than in economy 1. Exercise 29 We examine how the limited liability of the decision maker u affects his optimal risk taking. The , where k is the minimum level of subsistence that is left final consumption of the agent is that he to the agent in case of bankruptcy. Alternatively, k – w0 is the maximum loss on his exposure to risk retains. All losses in excess of that will be borne by a third party, at no cost for the decision maker. The decision maker's liability is limited to k. • Show that the agent u with limited liability k behaves in the face of risk exactly as an agent with utility v with full liability, with v(z) = u(max(k, z)). • (Gollier, Koehl, and Rochet 1997) Examine the properties of the optimal exposure α as a function of w0.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_64.html [4/21/2007 1:28:20 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_65.html
Page 65
5 The Equilibrium Price of Risk. In the previous chapter we saw how people might manage their portfolio given the current price of the assets. This chapter provides a general equilibrium framework to determine the price of the risky asset. The equilibrium price of the risky asset depends on both the degree of risk aversion of the investors and the degree of risk in the economy. In this chapter we examine this relationship in the most simplistic model similar to the one proposed by Lucas (1978). Following Mehra and Prescott (1985) and many others afterward, we also calibrate this model. Without much surprise, the model does not fit the data. The surprise, if there is any, comes from how big the discrepancy is between the theoretical equilibrium prices and the observed prices.
5.1 A Simple Equilibrium Model for Financial Markets We assume that the economy is composed of several identical agents who live for one period. Consumption takes place at the end of the period. Agents are identical not only on their preferences but also on their endowment. They have the same utility function u which is increasing and concave. There are n different assets in the economy, each of them representing a specific technology to produce an homogeneous consumption good. There is no production cost, and the consumption good is the numéraire. It implies that the profit is equal to the production. Each agent is endowed with one unit of each asset. Let be a random variable representing the production per capita in sector i of the economy, i = 1, . . . , n. By this we mean that is the dividend per share in that sector, since capital is the only input. Because the model is static, this is also the ex-post value of this
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_65.html [4/21/2007 1:28:20 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_66.html
Page 66 sector. Everyone agrees on the distribution of the . We see that
is the GDP per capita.
There is a simple financial market in this economy. Before the realization of the , agents can sell their property rights on the different sectors of the economy against the delivery of a certain quantity of the consumption good. Let Pi be the unit price of asset i, which is a sure quantity of the consumption good that must be delivered in exchange for one unit of asset i. There is a risk-free storage technology that yields Rf units of the good per unit stored at the beginning of the period. Thus rf = Rf – 1 is the risk-free rate in the economy. The decision problem of each agent in the economy is to determine the proportion αi of property rights he owns in sector i that he will retain, given price Pi of those rights. If αi is larger than 1, this means that the agent is purchasing additional shares of this sector. The maximization problem is written as
The standard portfolio problem that we examined in the previous chapter is a particular case of problem (7.10) with n = 1 and w0 = P1 and
. The solution to this generalized problem is uniquely determined by its
first-order conditions:
To determine the equilibrium price of firms, we need to impose a market-clearing condition on financial markets. The net demand for property rights must be zero. Notice that the optimal demand for them is the same for all agents in this model with identical preferences and identical endowments. It implies that the equilibrium condition is αi = 1 for all i: no one wants to sell his endowment of property rights at equilibrium. This condition together with condition (5.2) form a system of equations that must satisfied for all αi and Pi to characterize a competitive equilibrium. Combining them yields
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_66.html (1 of 2) [4/21/2007 1:28:21 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_66.html
The random variable from the fact that the joint
is called the pricing kernel of this economy. This terminology comes
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_66.html (2 of 2) [4/21/2007 1:28:21 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_67.html
Page 67 distribution of to
is a sufficient statistic to determine the price of asset i, since condition (5.3) is equivalent
It is a tradition to express this property by using stock returns rather than stock prices and cash flows. Let and obtain from (5.4) that
define respectively the gross and net rate of return of asset i. We directly
or equivalently,
At equilibrium the expected return of an asset equals the risk-free rate plus a risk premium . This risk premium is smaller, the larger its covariance with the pricing kernel. The intuition for this is simple: when the covariance is large, and when the marginal utility is high which is when consumption is low, the asset's payoff tends to be large. The asset then is more valuable in that it succeeds to deliver wealth when wealth is most valuable. The limit case is when asset i has a return that is independent of the market return, in which case φi = 0. At the margin, increasing the share of this asset in the portfolio does not affect the portfolio risk. Therefore, at equilibrium, the risk associated to asset i will not be compensated. This is due to second-order risk aversion. denotes the initial market value of the Let be the return of the stock market as a whole. If investors' portfolio, the return of the stock market is . The equilibrium price of the macroeconomic risk in this economy can be measured by the difference between the expected return of stocks compared to the return of the safe strategy. The expected excess return of holding the market portfolio of risky assets is called the equity premium and is denoted
. It yields
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_67.html [4/21/2007 1:28:22 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_68.html
Page 68
Using the standard technique presented in chapter 4, we can check that the equity premium is nonnegative and that it is an increasing function of risk aversion. An increase in risk aversion raises the supply of stocks, which must be compensated at equilibrium by a reduction in their price P. This increases their returns, thereby increasing the equity premium.
5.2 The Equity Premium Puzzle We see that computing the equity premium is a simple task in this framework. Remember that in equation (5.7) is nothing else than the real gross domestic product per capita. We can believe that agents form their expectation about the future growth of GDP per capita by looking at the past experience. We presented the data for the growth of real GDP per capita for the period from 1963 to 1992 in section 2.8. Let us assume that agents believe that each realized growth rate in the past will occur with equal probability. Let us also assume that agents have a constant relative risk aversion γ. Using equation (5.7) allows us to derive the equity premium in that economy. These computations are reported in table 5.1. Notice that a simpler method can be used to calculate the equity premium in this economy, since the risk associated to the annual change of real GDP is so small. Using Taylor approximations, we find that the equity premium is approximately equal to Table 5.1 Equity premium with CRRA and based on the actual growth of US GDP per capita, 1963 to 1992 RRA
Equity premium (%)
γ = 0.5 γ=1 γ=2 γ=4 γ = 10 γ = 40
0.03 0.06 0.11 0.23 0.61 2.86
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_68.html [4/21/2007 1:28:22 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_69.html
Page 69 The Equilibrium Price of Risk
where is the index of relative risk aversion evaluated at the mean. We see that the equity premium is approximately proportional to relative risk aversion and to the variance of the risk. In our data we have . We can compare this approximation yielding φ = 0.056γ to the true φ given in table 5.1. By this rule of thumb we obtain equity premia with only a very small margin of error with respect to the exact theoretical values. We see that in the range of acceptable values of relative risk aversion (γ ∈ [1, 4]), the equity premium does not exceed one-fourth of a percent. With γ = 2, the equity premium equals one-tenth of a percent. This means that those who purchase stocks rather than bonds will end up at the end of the year with a wealth that is only 0.1% larger on average. This is the compensation at equilibrium for accepting a variable income stream. Is this prediction about how risks should be priced in our economy compatible with the data? Historical data on asset returns are available from several sources. Shiller (1989) and Kocherlakota (1996) provides statistics on asset returns for the United States over the period 1889 to 1978. The average real return according to Standard and Poor 500 is 7% per year, whereas the average short-term real risk-free rate is 1%. The observed equity premium has thus been equal to 6% over the century. We reproduce the data for the period
Figure 5.1 Annual real return of Standard and Poor 500
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_69.html [4/21/2007 1:28:23 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_70.html
Page 70
Figure 5.2 Annual real return to nominally risk-free bonds
Figure 5.3 Excess returns of Standard and Poor 500
corresponding to the period examined above in figures 5.1, 5.2, and 5.3. Over the period 1963 to 1995, the real return of the S&P500 was 7.72% per year, which is close to the historical average. Second, by historical standards, the real interest rate has been high, averaging 3.14%. This is particularly true for the 1980s. The bottom line is an equity premium around 4.5%. Clearly, this simple version of the model proposed by Lucas (1978) does not fit the data. The observed equity premium is just 20 to 50 time larger than its theoretical value. This puzzle was first stated by Mehra and Prescott (1985). Said differently, the existing equity prefile:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_70.html [4/21/2007 1:28:23 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_71.html
Page 71 mium is compatible with the model only if relative risk aversion is assumed to be larger than 40. Recall that with this large risk aversion, agents are ready to pay as much as 29% of their wealth to get rid of the risk of losing or gaining 30% of it with equal probability. However, as argued by Rietz (1988), we cannot exclude the possibility that our sample is not large enough to describe the macroeconomic risk. This is the so-called peso problem, in which there is a small probability of a catastrophic event. This probability is taken into account by investors, but not by the model because this event is not in the sample.
5.3 The Equity Premium with Limited Participation It is a well-known fact that only a small proportion of households actually are stockholders. The average postwar rate of participation to stock markets has been around 20% in the United States. Could this fact explain the equity premium puzzle? Following Campbell and Mankiw (1990), Mankiw and Zeldes (1991), and Vissing-Jorgensen (1998), let us suppose that for some exogenous reason some agents are not allowed to hold stocks. In other words, contrary to the assumption made in the basic model, these agents have to sell the property rights that they initially owned. This could be due to their lack of knowledge about how to manage a portfolio of stocks, or to their inability to manage their business. The consequence is that they are fully insured against the macroeconomic risk. How would the exclusion of this category of agents from financial markets affect the equilibrium price of property rights on firms? The intuition is that this exclusion will increase the supply of stocks, thereby reducing their equilibrium price. It would yield an increase in their returns. The equity premium would be larger as a consequence. To simplify the notation, let us assume that there is only one risky asset whose payoff is . With this in mind, the decision problem for those who can invest in stock is not different from program (7.10). The only difference in the modeling comes from the market-clearing condition. The exogenously given proportion of agents allowed to participate is denoted by k. This is the participation rate on financial markets. Since all participants are alike, they will all hold the same number α of property rights. The equilibrium condition is that kα = 1,
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_71.html [4/21/2007 1:28:24 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_72.html
Page 72
Figure 5.4 Equity premium (in %) with limited participation (with γ = 2)
or α = 1/k. Since this α is larger than 1, all investors will go short on the risk-free asset to purchase stocks. Without loss of generality, we normalize the risk free rate to zero: Rf = 1. Using the first-order condition (5.2) with this α yields the following condition on the equilibrium price P(k):
Notice that the equilibrium price P(k) cannot in general be analytically extracted from this condition, contrary to the previous case where k equaled 1. Numerical methods must be used to get P(k) from which the equity premium can be obtained. We have done this exercise by using a power utility function with γ = 2, together with the data of the growth of US real GDP per capita for the period 1963 to 1992. The results are illustrated in figure 5.4. We see that the participation rate has no sizable effect on the equity premium as long as it exceeds, say, 30%. If it is less than that, the effect becomes big enough to explain the puzzle. The macroeconomic risk is borne by such a small fringe of the population that a large bonus must be given to them as a compensation. The driven force for the smaller participation rate to increase the equity premium is the necessity to induce stockholders to increase their bearing of the macroeconomic risk. At equilibrium it will make
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_72.html [4/21/2007 1:28:24 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_73.html
Page 73 the pricing kernel
more volatile. The excess volatility of the pricing kernel increases the absolute value of its covariance with stocks' returns, thereby, from (5.7), increasing the equity premium. It is, however, counterfactual that nonstockholders are fully insured against macroeconomic shocks. As shown by Campbell and Mankiw (1990) and Mankiw and Zeldes (1991) nonstockholders have a consumption pattern that covaries positively with stockholders' consumption. This will reduce the effect of the participation rate by reducing the volatility of the pricing kernel.
5.4 The Equity Premium and the Integration of International Financial Markets In our simple version of the Lucas's model, we assumed that US households could just purchase US stocks. This assumption relies on the observation that households usually do not purchase foreign assets. The existence of barrier to investment in foreign countries are still very strong, as can be seen by actual portfolios in various countries. For example, French and Poterba (1991) and Baxter and Jermann (1997) reported that US investors hold around 94% of their financial assets in the form of US securities. In Japan, the United Kingdom, and Germany the portfolio share of domestic assets exceeds 85% in each case. This is the so-called international diversification puzzle. Baxter and Jermann (1997) showed that this absence of international diversification is even made more puzzling due to the presence of nontradable human capitals that are strongly positively correlated with domestic capital returns. Suppose that, thanks to the growing integration of international financial markets, investors better diversify their portfolio by purchasing foreign assets. What would happen to the equity premium? Let us consider an economy with two countries of equal size where agents have the same attitude toward risk expressed by utility function u. Suppose that firms in country 1 generate revenues , whereas
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_73.html [4/21/2007 1:28:25 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_74.html
Page 74 firms in country 2 generate revenues . For the sake of simplicity, let us assume that liberalizing capital flows between the two countries raise the equity premium?
and
are i.i.d. Does
Observe first that because of the symmetry, asset prices after the integration will be the same at equilibrium in the two countries. Then, as we have seen from proposition 5, it will optimal for all agents to perfectly diversify their portfolios. The problem is thus to determine the demand for an asset with payoff
compared to the
is a SSD-dominant shift in the demand for the asset yielding payoff . But we know that distribution of risk with respect to . We can then use proposition 9, which provides a sufficient condition for any such shift to increase the demand for the risky asset. This yields the following result: 10 Suppose that countries face idiosyncratic shocks on their revenues. Suppose also that relative risk aversion is less than unity and nondecreasing, and absolute risk aversion is nonincreasing. Then liberalizing capital flows will raise the aggregate demand for stocks. At equilibrium it will reduce the equity premium. PROPOSITION
We see that under these conditions, the equity premium puzzle is even stronger than what we suggested in section 5.2. Allowing international diversification of portfolio would reduce the theoretical value of the equity premium. But the assumptions that are proposed in the proposition above are not satisfactory. In particular, the proposition requires that relative risk aversion be less than unity, a very doubtful assumption. Eeckhoudt, Gollier, and Levasseur (1994) provide other sufficient conditions. Also it should be noted that individual portfolios do not seem to be as internationally diversified as they should. This makes proposition 10 less relevant in this context, but it raises another issue of financial markets. We conducted some numerical computations on the effect of liberalizing capital flows in this simple two-country model. For liberalized financial markets the equity premium is
Once again, using historical data for the growth of real GDP in the United States, we can compare the equity premia in the two situa-
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_74.html [4/21/2007 1:28:25 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_75.html
Page 75 Table 5.2 Comparison of equity premia before and after the liberalization of capital flows between two countries
RRA
Equity premium in autharky (%)
Equity premium with international diversification (%)
γ = 0.5 γ=1 γ=2 γ=4 γ = 10 γ = 40
0.03 0.06 0.11 0.23 0.61 2.86
0.015 0.03 0.06 0.12 0.32 1.42
tions for different values of the constant relative risk aversion. Using approximation (5.8), we anticipate that the integration will approximately divide the equity premium by a factor 2, since the variance of half the variance of . The true equity premia are given in table 5.2.
is
Of course, this is a theoretical exercise, since real GDP of different countries are affected by common random factors. Still it provides an upper bound on the reduction of the equity premium that can be obtained by reducing barriers to capital mobility. In the case of i.i.d. shocks on national productions, allowing two equally sized countries to exchange capital ventures reduces the equity premium by two.
5.5 Conclusion In an extremely simplified model of asset pricing, the predicted equity premium is more than 20 times smaller than the equity premium observed over the century. There are two ways to interpret this puzzle. To put it simply, either markets do not work as they should or the model is wrong. If you believe that the first answer is correct, you should invest in stock like crazy—to take advantage of the incredibly large equity premium for the associated low risk. The other explanation seems more plausible, since a lot of simplifying assumptions have been made to make the model tractable. Let us list just a few of them: • Agents bear no other risk than their portfolio risk. There is no risk on human capital, which is normally a nontradable asset due to the prohibition of slavery.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_75.html [4/21/2007 1:28:26 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_76.html
Page 76 • Agents cannot internationally diversify their portfolio. There is no other risky investment opportunity than US stocks. • Agents consume all their wealth at the end of the period. There is no intent to invest for the long term. • There is no cost for trading stocks and bonds. • Only stocks and bonds can be traded, at the exclusion of more sophisticated assets as options. As a consequence agents cannot tailor their final consumption plan in a nonlinear way. • All agents have identical preferences and there is no wealth inequality. • Agents have constant relative risk aversion. • All agents have access to financial markets. The question is whether relaxing these assumptions could explain the puzzle. We examine this question in various places in the book.
5.6 Exercises Exercise 30 The aim of this exercise is to examine how does long term growth affect the equity premium of the , economy. Consider an economy whose GDP per capita z follows a stationary geometric process: where
, . . . are i.i.d. random variables that are distributed as .
• Derive the formula for the equity premium at each date t = 1,2, . . . . • Show that the equity premium is constant over time when the utility function is CRRA. The fact that the equity premium remained stable over time during the last century is an argument for using CRRA functions. • Derive the conditions under which the equity premium is a decreasing function of the GDP per capita in the economy. Answer to the same question in an economy where the GDP/cap follows an arithmetic process . Exercise 31 Consider an economy with no aggregate risk. Show that the price of any risky asset is equal to its expected payoff.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_76.html [4/21/2007 1:28:26 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_77.html
Page 77 Exercise 32 Prove that the equity premium is decreasing with the initial wealth of the representative agent under DARA. Exercise 33 (Capital asset pricing model) Suppose that all agents have a quadratic utility function u(z) = cz – 0.5z2. • Show that
at equilibrium, where
asset i and the market portfolio, and
is the covariance of the returns of
is the variance of the market portfolio. Notice that the ratio
is often called the "beta" of asset i. • Show that risk aversion measured at
is the equity premium in this economy, where R is the index of relative .
Exercise 34 Suppose that there is one risk-neutral agent and several risk-averse agents in the economy. Derive the equilibrium asset pricing formula in this economy. Exercise 35 Prove formally that the equity premium is an increasing function of the degree of risk aversion of the representative agent. Exercise 36 In this chapter we assumed that there is no possibility for firms to invest. Suppose alternatively that the good can be either consumed or used as an investment. Also agents are initially endowed with one unit of the good. Each firm i is contemplating the possibility to enlarge its size. In order to increase the output (per capita) by δ% in all states of nature, it needs δki units of the good as an investment. • Where would it be most efficient to invest the few units of the good that are available ex ante? • Would competitive financial markets allocate capital efficiently? Are equilibrium prices useful to help the social planner to organize this allocation? Exercise 37 (Baxter and Jermann 1997) Let us consider again the two-country model of section 5.4. Suppose that each firm's payoff are allocated as follows: a share 1 - α of it is distributed to shareholders, and the remaining α < 0.5 share is distributed to the
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_77.html [4/21/2007 1:28:27 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_78.html
Page 78 employees as salary. We maintain the assumption of independence between human capital cannot be traded on financial markets.
and
. Finally risks attached to
• Show that the equity premia are not affected by the fact that human capital cannot be traded. • Describe the optimal composition of portfolios in this integrated economy. In particular, show that one invests more in foreign stocks than in domestic ones. That fact reinforces the international diversification puzzle.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_78.html [4/21/2007 1:28:28 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_79.html
Page 79
III Some Technical Tools and Their Applications
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_79.html [4/21/2007 1:28:28 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_80.html
Page 80
This page intentionally left blank.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_80.html [4/21/2007 1:28:28 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_81.html
Page 81
6 A Hyperplane Separation Theorem Because of the linearity of the objective function with respect to probabilities, a limited number of analytical tools are necessary to solve most problems in expected utility theory. The Jensen's inequality is clearly one of these tools. But it requires concavity, which is often too demanding, as we will see. This chapter, together with the next one, has two objectives. The first one is to describe these tools that will be used in the next few chapters to derive new results. The second one is to apply these tools to some simple problems linked to what we did in the previous chapters. By doing so, we will see that several a priori unrelated results in the economics of uncertainty are in fact deeply interrelated, as they share the same technical structure, and as they use the same tool to be solved.
6.1 The Diffidence Theorem Since the well-being of an individual is measured by the expected utility of his final wealth, some problems in expected utility theory simplify to determining the impact of a change in a parameter in the model on the expectation of a function of a random variable with support in [a, b]. Let be this random variable, and let f1 and f2 be this real-valued function respectively before and after the change in the parameter of the model. We want to determine the conditions under which
In general, no restriction is imposed on the distribution of the random variable, except that its support is bounded in [a, b]. Obviously a necessary condition is that
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_81.html [4/21/2007 1:28:29 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_82.html
Page 82
Indeed, condition (6.2) is nothing else than condition (6.1) for the degenerated random variable which takes value x with probability 1. Condition (6.2) is also sufficient for condition (6.1), since it implies that
for any cumulative distribution function G. To summarize, in order to verify that condition (6.1) holds for any random variable, it is enough to verify that it holds for any degenerated random variable. This greatly reduces the dimensionality of the problem. Most of the time, the technical problem is a little more complex than the one just presented. More specifically, it happens that one has already some information about the random variable under consideration. For example, one knows that
and one inquires about whether condition (6.1) holds for this subset of random variables. Obviously, since one wants a property to hold for a smaller set of objects, the necessary and sufficient condition will be weaker than condition (6.2). We now determine this weaker condition. A simpler way to write the problem is as follows:
If one does not know the results presented below, the best way to approach this problem would probably be to try to find a counterexample. A counterexample would be found by characterizing a random variable such that , with positive. Our best chance is obtained by finding the cumulative distribution function G of that satisfies the constraint and which maximizes the expectation of f2. Thus we should first solve the following problem:
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_82.html [4/21/2007 1:28:29 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_83.html
Page 83
The second constraint simply states that G is indeed a cumulative distribution function. Let G* be the distribution function of the best candidate random variable to violate condition (6.4). The Kuhn-Tucker conditions for this maximization problem are written as
where m and k are the Lagrangian multipliers associated respectively to conditions ∫ f1(x)dG(x) = 0 and ∫ dG(x) = 1. Because of the linearity of both the objective function and the constraints with respect to the decision variables, condition (6.6) are necessary and sufficient. From condition (6.6) we obtain that
Because G* must satisfy the constraints of program (6.5), it yields
Thus a necessary and sufficient condition for condition (6.4) to hold is that k be nonpositive. From conditions (6.6) this means that we need to find a pair (m, k) such that
This condition is obviously equivalent to
for all x in [a, b]. This proves the following central result, which is due to Gollier and Kimball (1996). We hereafter refer to this proposition as the diffidence theorem. The origin of this term will be explained later on. PROPOSITION
11 The following two conditions are equivalent:
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_83.html (1 of 2) [4/21/2007 1:28:30 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_83.html
1.
with support in
.
2. ∃m: f2(x) ≤ mf1(x) ∀x ∈ [a, b].
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_83.html (2 of 2) [4/21/2007 1:28:30 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_84.html
Page 84 Because of the ubiquity of this result in this book, it is important to understand its nature. To help the reader, we hereafter provide an alternative proof, which is more intuitive. One can trace the origin of this approach to an important paper by Pratt and Zeckhauser (1987). Observe that the objective and the two (equality) constraints in program (6.5) are linear in dG, which is required to be nonnegative. Thus this is a standard linear programming problem. Linearity entails corner solutions. More specifically, the basic lesson of linear programming is that the number of choice variables that are positive at the optimal solution does not exceed the number of constraints. Thus we obtain the following lemma. 1 Condition (6.4) is satisfied if and only if it is satisfied for any binary random variable, i.e. for any random variable whose support contains only two points. LEMMA
For readers who are not trained in operations research, we hereafter provide an informal proof of this result. Suppose, by contradiction, that the solution of program (6.5) has three atoms, namely at x1, x2, and x3 with probabilities p1 > 0, p2 > 0, and p3 > 0 respectively. We can represent this random variable by a point A in the Machina triangle where we use the last constraint in (6.5) to replace p2 by 1 – p1 – p3. This is drawn in figure 6.1. Because this random variable satisfies condition segment CD that is defined by
, we have that point A is an interior point on
Figure 6.1 Nonoptimal interior points on segment CD
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_84.html [4/21/2007 1:28:30 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_85.html
Page 85 The problem is to find (p1, p3) that maximizes the following function:
As usual, this problem is graphically represented by iso-objective curves in figure 6.1. The important point is that these curves are straight lines. As we know, this is specific to expected utility, where the well-being of an agent is linear in probabilities. It clearly appears that an interior point on segment CD is never optimal. At the limit the iso-objective lines are parallel to segment CD, and it is enough to check the condition at C or D, as stated in the lemma. Thus, to check that condition (6.4) holds for any random variable, it is enough to check that it holds for any binary random variable. This problem is rewritten:
for all p ∈ [0, 1], and (x1, x2) ∈ [a, b] × [a, b]. For the left condition to be satisfied, we need f1(x1) and f1(x2) to alternate in sign. Without loss of generality, let f1(x1) < 0 < f1(x2). By eliminating p from problem (6.13), we can be rewrite it as
Since this must be true for all x1 such that f1(x1) < 0, the inequality above is equivalent to
This must hold for any x2 such that f1(x2) > 0. In a similar way the condition is rewritten as
This can be true only if there exists a scalar m, positive or negative, such that
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_85.html (1 of 2) [4/21/2007 1:28:31 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_85.html
The two inequalities above hold if and only if
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_85.html (2 of 2) [4/21/2007 1:28:31 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_86.html
Page 86 In short, condition (6.4) holds for any binary random variable if and only if condition (6.14) holds. This is the diffidence theorem. Observe that as expected, the necessary and sufficient condition for (6.4) is weaker that the necessary and sufficient condition for (6.1). Indeed, m is not forced to be equal to 1 in condition (6.14), contrary to what happens in condition (6.2). There still remains a difficulty to verify whether condition (6.4) is true: one has to look for a scalar m such that
be nonpositive for all x. This search may be difficult. However, the problem is much simplified if f1 and f2 satisfy the following conditions: 1. There is a scalar x0 such that f1(x0) = f2(x0) = 0. 2. f1 and f2 are twice differentiable at x0. 3.
.
From condition 1, we have that H(x0, m) = 0. Since, from condition 2, H is differentiable at x = x0, it must be true that
in order to have H nonpositive in the neighborhood of x0. The first necessary condition yields
whereas the second necessary condition is rewritten as
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_86.html (1 of 2) [4/21/2007 1:28:31 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_86.html
We obtain the following corollary, which is due to Gollier and Kimball (1996). COROLLARY
1 Suppose that conditions 1 to 3 are satisfied. Then a necessary and sufficient condition for (6.4) is
A necessary condition for (6.4) is condition (6.15).
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_86.html (2 of 2) [4/21/2007 1:28:31 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_87.html
Page 87 Again, the dimensionality of the problem is much reduced by this result. One has initially to check whether any random variable that verifies condition
also verifies condition
. We end up with
, is nonpositive.
checking whether a function of a single variable,
This problem is even further simplified by necessary condition (6.15) that is referred to as the local diffidence condition. This terminology comes from the fact that (6.15) is the necessary and sufficient condition for any small risk around x0 to verify property (6.4). Indeed, suppose that
. If k is small, condition
can be rewritten as
Using the assumption that f1(x0) = 0 implies that
In the same vein, condition
is equivalent to
Replacing by its expression from (6.17) makes the inequality above equivalent to condition (6.15). Thus this condition is necessary and sufficient for condition (6.4) to hold for any small around x0. Inequality (6.15) is only necessary for it to be true for any risk. In some applications the equality condition following case:
in problem (6.4) is replaced by an inequality, as in the
The necessary and sufficient condition for this problem is that there exists a nonnegative scalar m such that f2(x) ≤ mf1(x) for all x. The sufficiency of this condition is immediate: if this condition holds for any x, taking the expectation yields
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_87.html (1 of 2) [4/21/2007 1:28:32 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_87.html
The right-hand side of the inequality above is the product of a nonnegative scalar and of a nonpositive one. It implies that is nonpositive. Lemma 1 proves the necessity of the existence of a m. The fact that it must be nonnegative is most easily shown when
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_87.html (2 of 2) [4/21/2007 1:28:32 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_88.html
Page 88 . By contradiction, suppose that
conditions 1 to 3 are satisfied, where negative. It means that
and
have opposite signs. Then suppose that
is is positive (resp.
negative) and is negative (resp. positive). It implies that condition (15.13) is violated for some degenerated at a value slightly to the left (resp. right) of x0. COROLLARY
that is
2 Suppose that conditions 1 to 3 are satisfied. Then a necessary and sufficient condition for (15.13) is
A necessary condition for (15.13) is condition (6.15).
6.2 Link with the Jensen's Inequality. The previous chapters intensively used Jensen's inequality, which can be seen as an application of the hyperplane separation theorem presented above. The problem is to determine the effect of risk on the expectation of a function of a random variable. More specifically, we want to determine the condition under which the expectation of this function is smaller than the function evaluated at the expected value of the random variable:
In fact, this problem can be solved by using the diffidence theorem. Let us rewrite it as
for any scalar x0. Using proposition 11 with f1(x) = x – x0 and f2(x) = f(x) – f(x0) yields the following condition:
The left-hand side of the inequality above is the equation of a straight line with slope m that is tangent to curve f at point x0. The inequality means that such a tangent line must always lie above curve f, as is the case in figure 6.2. From (6.15) we know that a necessary condition is that f be locally concave at x0. Since this must be true for any x0, it is clear that condition (6.19) holds if and only if function f is concave. This is the Jensen's inequality, which is a direct application of the diffidence theorem.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_88.html [4/21/2007 1:28:33 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_89.html
Page 89
Figure 6.2 Concave function below tangent lines PROPOSITION
12 (Jensen's Inequality) The following two conditions are equivalent:
1.
for all
.
2. f is concave (resp. convex). If f is neither concave nor convex, the result is ambiguous in the sense that it is always possible to find a pair such that inequality (6.19) holds with and the opposite inequality holds with .
6.3 Applications of the Diffidence Theorem We now illustrate these techniques by presenting their most obvious applications to the economics of uncertainty. 6.3.1 Diffidence An agent with sure wealth w0 and utility function u dislikes any zero-mean risk if the following condition is satisfied:
When we require that this condition holds at all wealth level, we obtain the concept of risk aversion. However, if we let w0 fixed, the concavity of u is not necessary anymore to guarantee property (6.20). Before making this point, let us call agent u diffident at w0 if condition (6.20) holds for all
. Using the diffidence theorem with f1(x)
= x and f2(x) = u(w0 + x) – u(w0), we obtain the following necessary and sufficient condition for this:
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_89.html (1 of 2) [4/21/2007 1:28:33 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_89.html
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_89.html (2 of 2) [4/21/2007 1:28:33 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_90.html
Page 90 Graphically this means that there exists a straight line containing (w0, u(w0)) that is entirely above the curve representing u. If u is differentiable at w0, m = u′(w0) is the only candidate to satisfy condition (6.21). A necessary condition for the diffidence of u at w0 is that u be locally concave at that point. If u is twice differentiable, it means that u″(w0) is nonpositive. The difference between risk aversion and diffidence at some wealth level looks subtle at first sight. It will, however, play an important role in more complex risky environments, in particular in those explored in chapter 9. 6.3.2 Comparative Diffidence Similarly we say that agent u1 is more diffident than agent u2 at w0 if, for any ,
In words, agent u1 dislikes all lotteries that u2 dislikes, given the fact that both have the same fixed initial wealth level w0. Again, this is less demanding than "more risk aversion," since the agents' common initial wealth level is not arbitary. Using the diffidence theorem, we can write the necessary and sufficient condition for u1 to be more diffident than u2 at w0 as
If we assume that u1 and u2 are differentiable at w0, we know that
, so that this condition is
rewritten as
A couple of functions satisfying this condition is presented in figure 6.3, where we use the normalization u1(w0) = u2(w0) and
. Under this normalization, inequality (6.23) just states that curve u1 must lie below
curve u2. If the two functions are twice differentiable, the necessary condition (6.16) becomes
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_90.html (1 of 2) [4/21/2007 1:28:34 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_90.html
where
is the coefficient of absolute risk aversion of agent ui. Condition (6.24) is the
necessary and sufficient condition
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_90.html (2 of 2) [4/21/2007 1:28:34 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_91.html
Page 91
Figure 6.3 u1 more diffident than u2 at initial wealth level w0
for agent u1 to reject any small risk for which agent u2 is indifferent, assuming that they have the same wealth w0. If the risk is not "small" the necessary and sufficient condition is inequality (6.23), which is weaker than the condition that u1 be more concave than u2. In particular, it is apparent in figure 6.3 that A1 needs not to be larger than A2 everywhere. 6.3.3 Central Risk Aversion Let us now go back to the standard portfolio problem. We have shown in chapter 4 that agent u1 has a smaller demand for the risky asset than agent u2 at all wealth levels if and only if u1 is more concave than u2. If we fix the common initial wealth at some known level w0, this condition becomes again too restrictive. Before checking this, let us fix the terminology. We say that u1 is centrally more risk-averse than u2 at wealth level w0 if the former has a smaller demand for the risky asset than the latter at that wealth level, that is if
where we normalized the optimal exposure of u1 to risk
at unity.
In this case the diffidence theorem yields
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_91.html (1 of 2) [4/21/2007 1:28:34 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_91.html
as a necessary and sufficient condition. If we normalize
, this means that the slope of u1 is
smaller (resp. larger) than
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_91.html (2 of 2) [4/21/2007 1:28:34 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_92.html
Page 92 the slope of u2 to the right (resp. to the left) of w0. This condition is weaker than u1 being more concave than u2. However, if it holds for any w0, u1 is more risk-averse than u2. At this point it is important to notice that central risk aversion is a concept that is stronger than diffidence. If an agent u1 is centrally more risk-averse than agent u2 at w0, then agent u1 is more diffident than agent u2 at w0, but the opposite is not true. In other words, condition (6.26) implies condition (6.23). This is proved by integrating condition (6.26) (divided by x) with respect to x. Since the direction of integration cancels out the direction of the inequality when x is positive or negative, we obtain that condition (6.26) implies that
for all x. But this condition is precisely equivalent to condition (6.23). PROPOSITION
13 If agent u1 is centrally more risk-averse than agent u2 at w0, then he is also more diffident than u2
at w0. The condition that an agent asks always less of the risky asset than another is a more demanding condition than the one that the first rejects more lotteries than the second, when these conditions are for a given wealth level. When this must be true for any wealth level, these two conditions become equivalent. It is the notion of more risk aversion. 6.3.4 Central Riskiness We can also use the diffidence theorem to reexamine the effect of a change in risk of return on the demand for the risky asset, as we did in section 4.5. In this section we are looking for the stochastic dominance order that guarantees that the change in risk reduces the demand for the risky asset by all risk-averse agents. Following Gollier (1995), we call this order "central dominance," or "central riskiness." Let us consider two random returns
and
respectively with cumulative distribution function F1 and F2.
Distribution F1 is centrally dominated by distribution F2 if and only if
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_92.html [4/21/2007 1:28:35 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_93.html
Page 93 for all u ∈
2.
The necessary and sufficient condition of the change in risk that guarantees that all risk-averse
investors increase their demand for the risky asset is obtained as follows. We know that to any concave utility function u, there exists a random variable
such that
for all z. It implies that
. Let us define
It implies that one may rewrite condition (4.11) as
From the diffidence theorem, we know that this is true for any , namely for any u ∈
2
if and only if there
exists a scalar m such that f2(θ) ≥ mf1(θ) for all θ. This yields the following proposition. PROPOSITION 14 All risk-averse investors increase their demand for the risky asset due to a change in distribution of returns from F1 to F2 if and only there exists a scalar m such that for all θ,
If this condition is satisfied, one says that dominates in the sense of central dominance (CD). As seen in section 4.5, second-order stochastic dominance is not sufficient for central dominance: a change in risk can improve the expected utility of all risk-averse investors, although some of them reduce their demand for this risk! As pointed out by Gollier (1995), second-order stochastic dominance is not necessary either: all risk-averse investors can reduce their demand for the risky asset due to a change in distribution of returns although some of them like this change! SSD and CD cannot be compared. Meyer and Ormiston (1985) found a subset of changes in risk which is in the intersection of SSD and CD. They called it a strong increase in risk (SIR). An SIR is a mean-preserving spread in which all the probability mass that is moved is transfered outside the initial support of the distribution. The reader can verify that a SIR is a special case of CD, with condition (6.28) satisfied for m = 1. Eeckhoudt and Gollier (1995) and Athey (1997) have shown that the MLR and MPR orders are also special cases of CD.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_93.html [4/21/2007 1:28:36 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_94.html
Page 94
6.4 The Covariance Rule One of the most used formula in the economics of uncertainty is
Thus a necessary and sufficient condition for
is that
and
covary negatively. In the particular case where z is a deterministic transformation of y, this means
that z increases when y decreases. Let us define
and
. We have that
for all if f and g are monotonic and have opposite slopes: f′(x)g′(x) ≤ 0 for all x. It means that z and y go in opposite direction when x is increased. We now show that this condition is necessary and sufficient if no restriction is put on . This can be seen by rewriting condition (6.29) as follows:
for any random variable
and any scalar x0. A direct application of the diffidence theorem is that this property
holds if and only if there exists a scalar m(x0) such that
for all x and x0. Without loss of generality, let us assume that f is nondecreasing. Then the inequality above is equivalent to
The condition above is called a single-crossing condition: g can cross the horizontal axis at level m(x0) only once, from above.3 Since this must be true for any x0, this is possible only if g is decreasing.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_94.html (1 of 2) [4/21/2007 1:28:36 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_94.html
Suppose, alternatively, that condition (6.29) must not hold for any random variable, but only for those that satisfy property
for a specific scalar x0. Then single-crossing condition (6.31) for that x0 is the necessary
and sufficient condition for (6.29) to hold
3. If g is continuous, the only candidate for m(x0) is g(x0).
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_94.html (2 of 2) [4/21/2007 1:28:36 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_95.html
Page 95
Figure 6.4 (f, g) satisfying condition
for all
such that
for this set of random variables. We depict in figure 6.4 a pair (f, g) satisfying the property. We summarize these results in the following proposition. PROPOSITION
15 Suppose that f: R → R is nondecreasing. Condition
1. holds for any
if and only if g is nonincreasing;
2. holds for any
such that
if and only if there exists a scalar m(x0) such that the single-crossing
condition (6.31) holds. A direct application of the proposition is that the equity premium obtained in equation (5.7) is positive under risk aversion.
6.5 Conclusion In previous chapters we observed that an increase in risk aversion reduces either the set of acceptable risks and the demand for a risky asset. When the initial wealth level of the decision maker is arbitrary, there is no other change in the design of the utility function that generates such results. However, when the initial wealth level is fixed, this is not true. This yielded two new concepts: diffidence and central risk aversion. An increase in file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_95.html (1 of 2) [4/21/2007 1:28:37 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_95.html
diffidence at wealth w0 reduces the set of acceptable risks for the decision maker with that initial wealth level. An increase in central risk aversion at wealth w0 reduces
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_95.html (2 of 2) [4/21/2007 1:28:37 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_96.html
Page 96 the demand for the risky asset for the investor with that initial wealth level. The two orders are weaker than an increase in risk aversion. We also noticed that more diffidence is weaker than more central risk aversion. These concepts were developed by using a standard hyperplane separation theorem that we called the diffidence theorem in order to refer to its simplest application in the economics of uncertainty. Both the Jensen's inequality and the covariance rule are applications of this theorem. In the next five chapters we develop other applications of it.
6.6 Exercises and Extensions Exercise 38 Consider two agents with utility function u1 and u2 respectively. • Determine the conditions related to these functions that guarantee that the risk premium of agent 1 is always larger than twice the risk premium of agent 2. • Would you modify your answer to the question above if we restrict the set of acceptable lotteries to all lotteries with three atoms? • Find a pair of utility functions that satisfies this property. Exercise 39 Apply the hyperplane separation theorem to the following definition of decreasing absolute risk aversion: all undesirable risk are expected-marginal-utility increasing (see equation 2.11). Prove that this is true at all wealth levels if and only A is decreasing. Can we relax this condition if we limit the analysis to a specific wealth level w0? Exercise 40 Consider two agents, respectively with utility function u1 and u2. Use corollary 1 to prove the following results: • A necessary condition for u2 to like the optimal portfolio of u1 is that the absolute risk aversion of u2 be less than twice the absolute risk aversion of u1: A2 ≤ 2A1. • A sufficient condition for the property above is that the absolute prudence of u1 is larger than the absolute risk aversion of u2, but less than twice the absolute risk aversion than u1 : A2 ≤ P1 ≤ 2A1 (Gollier and Kimball 1996).
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_96.html [4/21/2007 1:28:37 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_97.html
Page 97 Exercise 41 Multi-antecedent diffidence (Gollier and Kimball 1996). • Consider a set of n + 1 real-valued functions {f0, f1, . . . , fn}. Find the necessary and sufficient condition for the following problem: for all ,
• Show the equivalence of the following two properties: 1. Agent u0 dislikes all lotteries that both u1 and u2 dislike. 2. There exist two positive scalars m1 and m2 such that u0 is more diffident than m1u1 + m2u2. Exercise 42 Compare functions u1(z) = min(z, 1) and
.
• Is there one more risk-averse than the other? • Is there one more diffident than the other at some wealth levels? • Is there one centrally more risk-averse than the other at some wealth levels? Exercise 43 Show that the monotone likelihood ratio order is a special case of the central dominance order. Exercise 44 Using the hyperplane separation theorem, show that the following two properties are equivalent: 1. If the agent is indifferent about , he dislikes
for all k > 1, at all wealth levels.
2. The agent is risk-averse. Exercise 45 (Cauchy-Schwarz inequality) The Cauchy-Schwarz inequality is used in various places in this book. It states that for any pair (f, g) of positive functions,
The objective of this exercise is to prove this well-known result by using the hyperplane separation theorem. • Show that you can rewrite the Cauchy-Schwarz condition as file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_97.html (1 of 2) [4/21/2007 1:28:38 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_97.html
• Use the hyperplane separation theorem to prove that this result always holds. [Hint: Take m(λ) = –1/λ.]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_97.html (2 of 2) [4/21/2007 1:28:38 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_98.html
Page 98
This page intentionally left blank.
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_98.html [4/21/2007 1:28:39 AM]
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_99.html
Page 99
7 Log-Supermodularity. Under scrutiny an analytical problem is often found to have more structure than the one represented by condition (6.4), which is a single-crossing property (with respect to the index of functions fi). The additional structure is often provided by the assumption of log-supermodularity. This chapter is devoted to the exploration of various properties of log-supermodular functions. We will see that these properties are really applications of the hyperplane separation theorem 11 which we encountered in the preceding chapter. We will also consider various applications of these properties for the expected utility theory. Most of the results presented here are adapted from Athey (1997).
7.1 Definition As an introduction to the useful concept of log-supermodularity,4 let us consider two vectors in Rn, y and z. The operations "meet" (∨) and "join" (∧) are defined as follows:
These two operators from Rn to Rn are represented in figure 7.1 for the case n = 2. Consider now a real-valued function h: Rn → R, n ≥ 2, which is nonnegative. 4. The terminology is not homogeneous. A log-supermodular function is sometimes called total positive of order 2, following Karlin (1968). Also, if this function is a density, its log-supermodularity is equivalent to the condition that the corresponding random variables are "affiliated."
file:///D|/Export2/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_99.html [4/21/2007 1:28:39 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_100.html
Page 100
Figure 7.1 Representation of meet and join operators DEFINITION
4 Function h is log-supermodular (LSPM) if for all z, y ∈ Rn,
Looking at figure 7.1 for the case n = 2, we see that whatever the rectangle under consideration in the plan, the product of h evaluated at the southwest–northeast corners of the rectangle is larger than the product of h at the northwest–southeast corners. Notice that it is a direct consequence of the definition above that the product of two log-supermodular functions is log-supermodular. An important example of LSPM functions in R2 is the indicator function h(x, θ) = I (x ≤ θ) which takes value 1 if x ≤ θ is true and 0 otherwise. Of course the same property holds also for . In the two-dimensional case (n = 2) we can rewrite the definition of LSPM as
If we now assume in addition that h is strictly positive, h is LSPM if and only if
or equivalently,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_100.html [4/21/2007 1:30:50 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_101.html
Page 101 If h is differentiable with respect to its first argument, this is equivalent to the condition that
Since this must be true for any x and x0 and for any θH > θL, this inequality holds if and only if
is nondecreasing in θ. This proves the following lemma (Topkis 1978). 2 Suppose that h: R2 → R+ is differentiable with respect to its first argument. Then h is LSPM if and only if one of the following two equivalent conditions holds:
LEMMA
We see that LSPM basically means that the cross-derivative of the logarithm of the function is nonnegative. As we know, the sign of cross-derivatives is essential for comparative static analyses. Not surprisingly, logsupermodularity comes into play when solving comparative statics problems involving decision under uncertainty. Log-supermodularity has various interpretations in the expected utility theory depending on which function we consider. For example, if we rewrite utility function ui(z) as u(z, i), agent u1 is more risk-averse than agent u2 if and only if ∂u(z, i)/∂z is LSPM. Alternatively, let F(., i) denote a family of cumulative distribution functions Fi(.) of random variable
. Then random variable
is dominated by random variable
in the sense of
the monotone probability ratio (MPR) order if and only if F(z, i) is LSPM. Similarly is dominated by in the sense of the monotone likelihood ratio (MLR) order if and only if F′(z, i) is LSPM, where F′(z, i) is the density function of random variable . A useful application of lemma 2 is for functions h that can be expressed as h(s, θ) = g(s + θ) for some positive function g. Then, from the preceding lemma, h is LSPM if and only if g is log convex,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_101.html [4/21/2007 1:30:51 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_102.html
Page 102 namely if ln g is convex. For example, agent u is DARA if and only if h(z, x) = u′(z + x) is LSPM.
7.2 Log-Supermodularity and Single Crossing 7.2.1 A Theoretical Result Consider a function g that crosses singly from below the horizontal axis at point x0, which is (x – x0)g(x) ≥ 0 for all x. Consider also a positive function h(x, θ). We want to determine the conditions under which
In words, we want to know whether function —a convolution—also crosses singly the horizontal axis from below. To solve this problem, we use the diffidence theorem with f1(x) = –g(x)h(x, θL) and f2 (x) = –g(x)h(x, θH). By corollary 1, we have
Since, by assumption, g(x) and (x – x0) have the same sign, the inequality above holds if and only if
If x0 is arbitrary, this condition is equivalent to the LSPM of h. LSPM is necessary in the sense that if h is not LSPM, it is possible to find a single-crossing g and a random variable such that condition (7.1) fails. The way the diffidence theorem has been proved provides the technique to build such a counterexample. In a similar fashion, if g does not cross singly, there must exist a LSPM function h that violates condition (7.1). This is the essence of a result obtained by Karlin (1968). It has been used in economics of uncertainty initially by Diamond and Stiglitz (1974), Jewitt (1987) and Athey (1997). We summarize this result as follows: PROPOSITION
16 Take any real-valued function g: R → R that satisfies the single-crossing condition: ∃x0 : ∀x : (x
– x0)g(x) ≥ 0. Then
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_102.html [4/21/2007 1:30:51 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_103.html
Page 103 condition (7.1) holds if and only if h is log-supermodular. Moreover, if g does not satisfy the single-crossing condition, there exists a LSPM function h that violates condition (7.1). We now present various applications of this result. 7.2.2 Applications to the Standard Portfolio Problem We have seen that an increase in risk aversion reduces the demand for the risky asset. Technically the problem is written as
This result is in fact a direct application of proposition 16 with g(x) = x and . We have seen before that h(x, i) is log-supermodular in this case if and only if u1 is more risk-averse than u2 in the sense of Arrow-Pratt. Thus condition (7.2) holds for all
and w0 if and only if u1 is more risk-averse than u2. This
provides an alternative proof of proposition 7. We can also reexamine the comparative statics of a change in the distribution of returns. A change in the distribution from F1 to F2 reduces the demand for the risky asset if condition (6.27) is satisfied. Then one can apply proposition 16 with g(x) = xu′(w0 + x) and
. Remember that as we saw in section 7.1, the
corresponds to the notion of a monotone likelihood ratio (MLR). The following log-supermodularity of result is presented in Milgrom (1981), Landsberger and Meilijson (1990), and Ormiston and Schlee (1993). PROPOSITION
17 A MLR-dominant change in the distribution of the risky asset always increases the demand for it.
This demonstrates that the MLR order is a special case of the central dominance order. Finally we can show that MLR is necessary for the comparative statics property if the risk free rate in the economy is arbitrary. 18 Consider the portfolio problem . A change in the distribution of increases the demand for the risky asset for all concave u and all r if and only if the change is MLR dominant.
PROPOSITION
Proof To prove the necessity of MLR, we have to return to the hyperplane separation theorem (proposition 11). Let Let λ(x) denote
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_103.html [4/21/2007 1:30:52 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_104.html
Page 104 the likelihood ratio
where
. Assume that the change from F1 to F2 increases the demand for the risky asset for all r:
has c.d.f. F1. By proposition 11, it must be true that there exists γ(r) such that
for all x and r. This is possible only if λ is increasing. Suppose, by contradiction, that there exists x′ > x such that λ (x′) < λ(x). Then condition (7.4) for r = x′ implies that λ(x) ≤ γ(x′). Thus we have λ(x′) < γ(x′). This implies that condition (7.4) is violated in the neighborhood of x = x′ for r = x′, a contradiction. 7.2.3 Jewitt's Preference Orders and . It is intuitive that is more risky than if all Consider an individual u that is indifferent between over . It is also intuitive that is more risky than if individuals who are more risk-averse than u prefer over . If this is true for any u ∈ , we say that all those who are more risk-loving than u prefer in the sense of Jewitt's stochastic dominance associated to . We address this question for the case dominates = 1. Let the utility function u(x, θ) be indexed by a risk-loving parameter θ; that is, an increase in θ represents a reduction in the degree of concavity à la Arrow-Pratt. This means that ∂u/∂x is LSPM. Our problem can then be written as ∀u(., θL) ∈ 1,
Let G ≡ F2 – F1. The property above can be rewritten as follows: ∀u(., θL) ∈
1,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_104.html [4/21/2007 1:30:52 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_105.html
Page 105 then integrating by parts obtains
By assumption, we know that ∂u/∂x is log-supermodular. Using proposition 16 directly yields the following result. PROPOSITION
19 Risk
dominates risk
in the sense of Jewitt's stochastic order (condition 7.5) if and only if F1
crosses F2 only once from above:
Examples of changes in risk that satisfy the Jewitt's stochastic dominance order associated to
1
are mean-
preserving spreads. This is, in general, not the case for other SSD shifts, as curves F1 and F2 may cross each other more than once. Jewitt (1989) and Athey (1997) derive a similar condition for the Jewitt's order associated to
2.
7.3 Expectation of a Log-Supermodular Function 7.3.1 A Theoretical Result Consider now a function h: Rn → R+ with n ≥ 3. Consider also a vector z in Rn–1. Suppose that h is logsupermodular. We are interested in determining whether
is also log-supermodular. In words, does the expectation operator preserve log-supermodularity? This question has been examined by Karlin and Rinott (1980), Jewitt (1987), and Athey (1997). They obtained the following positive result, whose proof relies directly on the diffidence theorem. PROPOSITION
20
is log-supermodular if h is log-supermodular.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_105.html (1 of 2) [4/21/2007 1:30:53 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_105.html
Proof See the appendix at the end of the chapter.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_105.html (2 of 2) [4/21/2007 1:30:53 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_106.html
Page 106 Notice that the log-supermodularity of h is sufficient, but not necessary, for the log-supermodularity of H. We now present two applications of this result. 7.3.2 Two Applications The first application deals with the relationship between decreasing absolute risk aversion and decreasing absolute prudence. Remember that absolute risk aversion and absolute prudence are defined respectively as A(z) = –u″(z)/u′(z) and P(z) = –u′″(z)/u″(z). Thus decreasing absolute risk aversion (DARA) means that u′(z + x) is LSPM in (z, x), whereas decreasing absolute prudence (DAP) means that –u″(z + x) is LSPM in (z, x). We have the following result, which is due to Kimball (1993): PROPOSITION
21 Decreasing absolute prudence is stronger than decreasing absolute risk aversion: DAP ⇒ DARA.
Proof We assume DAP. Because g(z, x, θ) = Ind(θ ≥ x) is LSPM in (z, x, θ), so is h(z, x, θ) = –Ind(θ ≥ x)u″(z + θ). Indeed, recall that the product of two LSPM functions is LSPM. By proposition 20, it implies that
is LSPM with respect to (z, x). This means that u is DARA. The dual result is that all MLR shifts in distribution satisfy the MPR criterion, meaning that MLR is a particular case of MPR. PROPOSITION
22 The monotone likelihood ratio order is stronger than the monotone probability ratio order: MLR
⇒ MPR. Proof This simple proof is taken from Athey (1997). To keep it simple, suppose that a density function exists. Then observe that
We know that the indicator function I(θ ≤ x) is LSPM in (x, i, θ) and that, by assumption, F′(θ, i) is LSPM in (x, i, θ). Since the product of two LSPM functions is LSPM, the integrand in equation (7.9) is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_106.html [4/21/2007 1:30:54 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_107.html
Page 107 LSPM. Using proposition 20 with z = (x, i) yields the result that F(x, i) is LSPM.
7.4 Concluding Remark. All results presented in the last two chapters rely on the diffidence theorem. We showed that this theorem itself relies on the linearity of expected utility with respect to probabilities. It implies that in many instances we can limit the analysis of risky situations by just looking at binary risks. This in turn generates a lot of simplifications whose the different propositions of these chapters are the expression. A side product of this presentation is that none of the tools presented here can be used for examining the implications of nonexpected utility models. This is probably why the NEU literature is currently poor in terms of comparative statics analysis of specific problems of decision under uncertainty. In the next part of this book, we will frequently use the results that we developed here.
7.5 Exercises and Extensions Exercise 46 Use proposition 16 to prove that an increase in wealth increases the demand for the risky asset if and only if the investor is DARA. Exercise 47 Prove that h(x, θ) = I (x ≤ θ) is LSPM. Exercise 48 Consider any utility function u whose domain is R+. Prove that h(x, y) = u′(xy) is LSPM if relative risk aversion is decreasing. Exercise 49 (Jewitt 1988) Consider the n-risky-asset portfolio model:
Suppose that the equilibrium condition is αi = 1 for all i = 1, . . . , n. Let u be decreasing absolute risk-averse, and let the joint density of
be LSPM. Show that an increase in risk aversion of
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_107.html [4/21/2007 1:30:54 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_108.html
Page 108 the representative investor will reduce the price of every risky asset. Exercise 50 (Jewitt 1988) Consider again the n-risky-asset portfolio model as above. As before, suppose that u is DARA. Let f, the joint density, be a function of a parameter θ, and suppose that f is LSPM with respect to (y1, . . . , yn, θ). Interpret the increase in θ as "good news" about the distribution of the returns. Show that it implies that an increase in θ will increase the price of every risky asset.
7.6 Appendix Proof of Proposition 20 For simplicity, we prove the particular case n = 3. Notice that all applications of this proposition in this book are for the case n = 3. Take an arbitrary vector (xL, xH, yL, yH). We have to prove that
The proposition is a direct application of the following lemma: LEMMA
for all
3 Suppose that h(., . , θ), h(xL, . , .), and h(., yH, .) are all LSPM in R2. Then the following inequality holds and all (xH, yL) such that xL < xH and yL < yH:
Proof Using the mean value theorem, we rewrite this condition as
Using the diffidence theorem, a necessary and sufficient condition for this to be true is that there exists a function m(θ0) such that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_108.html (1 of 2) [4/21/2007 1:30:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_108.html
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_108.html (2 of 2) [4/21/2007 1:30:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_109.html
Page 109 for all θ. This is what we have to prove. Without loss of generality, let us assume that θ > θ0. Since we know that h(., . , θ) is LSPM, we have that
But, since h(xL, . , .) is LSPM, we know that
Now, since h(., yH,.) is also LSPM,
Combining the last three inequalities yields
Thus condition (7.11) holds with
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_109.html (1 of 2) [4/21/2007 1:30:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_109.html
This concludes the proof.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_109.html (2 of 2) [4/21/2007 1:30:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_110.html
Page 110
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_110.html [4/21/2007 1:30:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_111.html
Page 111
IV Multiple Risks
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_111.html [4/21/2007 1:30:56 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_112.html
Page 112
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_112.html [4/21/2007 1:30:56 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_113.html
Page 113
8 Risk Aversion with Background Risk Up to now we have assumed that the decision maker faces a single source of risk. In the real world people face many risks simultaneously. This raises the question of whether the optimal risk-taking decision to one risk is independent of the existence of other risks in the agent's portfolio. This is generally not true, even if risks are independent. As an illustration, we may question the idea that all individual risks are traded on financial markets. Suppose, in contrast, that households bear an uninsurable risk on their wealth. It cannot be sold on markets. The best examples are risks associated to human capital. Those risks are reputed not to be insurable for obvious reasons related to asymmetric information and incentives. Under such circumstances the portfolio problem is more complex, since the decision makers must now determine their attitude toward one risk—the portfolio risk— in the presence of another risk, namely the risk on human capital. In this chapter we examine whether this other risk does affect the qualitative results obtained earlier in this book. In the next chapter we quantify the effect. The exogenous risk, , is called the "background risk." It cannot be insured and is therefore completely exogneous. The portfolio problem is now rewritten as
where is the return of the risky asset, which is assumed to be independent of background risk . Observe that the expectation operator is on and : the decision on α must be taken prior to the realization of the two sources of risk. How does the presence of one may question, for
affect the properties of the optimal demand for the risky asset? More specifically,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_113.html [4/21/2007 1:30:56 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_114.html
Page 114 example, whether decreasing absolute risk aversion is still sufficient to guarantee that an increase in w0 reduces the optimal demand for the risky asset. Or, does a more risk-averse agent always demand less of the risky asset? The standard method to solve these questions is to refer to the notion of the indirect utility function, as introduced by Kihlstrom, Romer, and Williams (1981) and Nachman (1982). The analysis of the effect of an independent background risk on an optimal decision under risk can actually be transformed into a problem related to a change in preference by defining the following indirect utility function: for all z,
The decision problem (8.1) is then equivalent to
The optimal decision of agent u with background risk is the same as for agent v who faces no background risk. Thus the question of the effect of on α is the same as the effect of a change in preference from u to v. The advantage of this method is that we know very well how a change in the shape of the utility function affects the optimal decision under uncertainty.
8.1 Preservation of DARA We start with a simple first remark on the properties that the indirect utility function defined by (8.2) inherits from the utility function. It comes from the observation that
where u(n) is the nth derivative of u. In consequence, if u(n) does not change sign, it transfers its sign to v(n). Thus risk aversion and prudence, for example, are preserved by background risk. A less obvious question is whether the indirect utility function inherits decreasing absolute risk aversion from u. Let Av and Pv denote respectively the coefficient of absolute risk aversion and absolute prudence of function v. We have that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_114.html [4/21/2007 1:30:57 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_115.html
Page 115 It is important to notice that Av and Pv are not just the expected value of
and
. In fact we have
that
Remember now that u DARA is equivalent to P being uniformly larger than A. Similarly v DARA is equivalent to Pv being uniformly larger than Av. From (8.4) it is not a priori obvious that Pv is uniformly larger than Av if P is uniformly larger than A. This is true, however, as stated in the following general proposition: 23 Assume risk aversion. The indirect utility function v defined by condition (8.2) inherits the property that P is uniformly larger than kA: PROPOSITION
Proof This is another application of the diffidence theorem. Condition follows:
may be rewritten as
By the diffidence theorem, we know that this condition holds for any z and if and only if there exists m = m(λ, k, z) such that
Now observe that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_115.html (1 of 2) [4/21/2007 1:30:57 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_115.html
Since we can assume that A is nonnegative, condition P ≥ kA implies that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_115.html (2 of 2) [4/21/2007 1:30:57 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_116.html
Page 116 Combining this condition to (8.5), we would be done if we could find a m such that
Taking m = –kλ is a good candidate, since condition (8.8) is then equivalent to
which is true for all k ≥ 0. It is noteworthy that the opposite property is not true: the value function does not in general inherit property P ≤ kA from u. A first illustration of the preceding result is obtained by taking k = 1, which implies that decreasing absolute risk aversion is preserved by the introduction of an independent background risk. COROLLARY
3 Decreasing absolute risk aversion is preserved by the introduction of an independent background
risk. As a consequence, if u exhibits DARA, the optimal solution of program (8.1) is increasing in w0. But it is possible to build an example of a utility function with increasing absolute risk aversion such that the optimal exposure to risk is locally increasing in w0. The preservation of DARA by the expectation operator is a special case of a property found by Pratt (1964) that a sum of DARA utility functions is DARA. Again, the opposite is not true: the sum of two functions that exhibit increasing absolute risk aversion may exhibit decreasing absolute risk aversion locally. Notice also that the sum of two CARA functions is DARA. A generalization of proposition 23 is as follows: suppose that u(n–1) does not alternate in sign and that –u(n)/u(n–1) is uniformly positive. Then, using the same method to prove the proposition for n = 2, we obtain that
An application of this result is that decreasing absolute prudence is inherited by the indirect utility function. This result has first been stated by Kimball (1993).
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_116.html [4/21/2007 1:30:58 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_117.html
Page 117
8.2 The Comparative Risk Aversion Is Not Preserved The positive results obtained in proposition 23 could induce us to believe that all the characteristics of u are transferred to v. This is not true. For example, we do not know whether the convexity of absolute risk tolerance is preserved. Another example is the following one, which is due to Kihlstrom, Romer, and Williams (1981): Consider two agents respectively with utilities u1 and u2. Suppose that u1 is more risk-averse than u2 in the sense of Arrow-Pratt; in other words, A1(z) is larger than A2(z) for all z, or
is log-supermodular in (z, i). Does this
imply that agent u1 will behave in a more risk-averse way than agent u2 if they both face the same background risk ? Technically the question is equivalent to whether v1 is more concave than v2 in the sense of Arrow-Pratt, where vi is defined as
Notice that v1 is more risk-averse than v2 if and only if
is log-supermodular in (z, i). Observe
also that
where . Finally recall that by proposition 20, log-supermodularity is preserved by the expectation operator. The problem is that the use of this proposition for our purpose here requires h to be LSPM. Function h is LSPM if indeed u1 is more risk-averse than u2, but it also requires that ui be DARA. An application of proposition 20 is thus that u1 will behave in a more risk-averse way than u2 in the presence of background risk if A1 is uniformly larger than A2 and A1, A2 are decreasing. Using lemma 3 rather than proposition 20 allows for a slight improvement of this result. 24 Comparative risk aversion is preserved by the existence of an independent background risk if at least one of the two utility functions exhibits nonincreasing absolute risk aversion. PROPOSITION
Technically we mean the following: suppose that
is larger than
for all z and that A1 or A2 is nonincreasing in z. Then agent u1 will behave in a more risk-averse way than u2 in the presence of background risk; mathematically,
is larger than
for all z. The more risk-averse agent will have a smaller demand for an independent
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_117.html [4/21/2007 1:30:59 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_118.html
Page 118 risky asset. He will also be ready to pay more to escape another independent risk. Kihlstrom, Romer, and Williams (1981) provide a counterexample when absolute risk aversion is increasing for both utility functions. Although decreasing absolute risk aversion is a simple and natural assumption, it is stronger than necessary. For example, consider the case where u1 ≡ λu2 + g, with λ a positive constant and g being decreasing and concave. In this case we say that u1 is more risk-averse than u2 in the sense of Ross (1981). This order is stronger than the usual one. Observe indeed that u1 is more risk-averse than u2 in the sense of Pratt, since we have that
which is equivalent to
This condition is automatically satisfied, since the left-hand side of this inequality is negative whereas the righthand side is positive. This shows that DARA is not necessary to obtain the result in proposition 24, since no such a restriction is made here on ui. PROPOSITION
25 Suppose that u1 is more risk-averse than u2 in the sense of Ross, meaning that there exist a
positive scalar λ and a decreasing and concave function g such that u1(z) = λu2(z) + g(z) for all z. Then agent u1 behaves in a more risk-averse way than agent u2 in the presence of background risk. Propositions 24 and 25 each have their comparative advantages. More specifically, proposition 25, contrary to proposition 24, does not require that any of the two utility functions be DARA. This is, however, at the cost of a quite restrictive concept of comparative risk aversion. The notion of more risk aversion in the sense of Ross is indeed very strong. It is in particular stronger than the notion of Arrow-Pratt. But this new concept will play an important role in the next section. It is useful to observe that an alternative characterization of u1 being more risk-averse than u2 in the sense of Ross is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_118.html [4/21/2007 1:30:59 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_119.html
Page 119 or equivalently,
8.3 Extensions with Dependent Background Risk. The advantage of the independence assumption is that it allows us to use the technique of the indirect value function. The derivation of the characteristics of the indirect value function generates in one single shot a full description of the effect of background risk for a wide set of decision problems under uncertainty. This is not possible if we allow the background risk to be correlated with the endogenous risk. The analysis of interaction among dependent risks is still in its infancy. In this section we cover two related questions. First, we examine the effect of an increase in risk aversion on the demand for stock in the presence of a correlated background risk. In the second part of this section, we provide more insight to the Ross's concept of more risk aversion. 8.3.1 Affiliated Background Risk Let us consider the standard portfolio problem (8.1). Suppose first that and are independent. We know from corollary 3 that an increase in risk aversion reduces the optimal α if the absolute risk aversion of u is decreasing. Of course this result does not necessarily hold if the return of the risky asset is statistically related to to the background risk. In particular, if they are negatively correlated, the background risk will serve as some form of "homemade" insurance to cover the portfolio risk. We therefore understand why an increase in risk aversion does not necessarily imply a reduction of the portfolio risk. We need some restrictions on the statistical relationship between and to obtain an unambiguous effect of an increase in risk under DARA. Suppose that u1 is more risk-averse than u2 and that they are both DARA. Let f(x, y) be the density function for the vector of random variables
. We want to find conditions on f to guarantee that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_119.html [4/21/2007 1:31:00 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_120.html
Page 120 The equality to the left means that agent u2 has a unit demand for the risky asset, whereas the inequality to the right means that agent u1 has a smaller demand. Let us define
We would be done if k be log-supermodular. Indeed, condition (8.11) which can be rewritten as
Proposition 16 tells us that the log-supermodularity of k is sufficient for condition (8.13). By proposition 20, we also know that k is LSPM in (y, i) if both and g(x, y, i) = f(x, y) are LSPM in (x, y, i). The LSPM of h is due to the assumption that be DARA, and that u1 is more risk-averse than u2. We conclude that a sufficient condition for the result presented in proposition 24 to be robust to the introduction of some correlation between the endogenous risk and the background risk is that these two risks be affiliated, i.e. that their density function be LSPM. The following result is due to Jewitt (1988). 26 Suppose that are affiliated, meaning that the joint density function of the return of the risky asset and of the background wealth be LSPM. Assuming that the two investors are DARA, the more riskaverse agent has a smaller demand for the risky asset.
PROPOSITION
In a similar way it can be proved that under DARA an increase in w0 increases the investment in the risky asset, even when the background risk is affiliated to its return (Athey 1997). The concept of affiliated random variables has been primarily used for other economic applications as auction theory (see Milgrom and Weber 1982). Let us take some time to analyze what is behind the condition that two random variables are affiliated. This in fact implies a form of positive dependence that is, in general, stronger than a positive correlation between the two random variables. For example, the LSPM of the density of tells us that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_120.html [4/21/2007 1:31:00 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_121.html
Page 121
is increasing in x. Indeed, this property can be rewritten as [g(x1) = k ⇒ g(x2) ≥ k] for all x2 ≥ x1. This is equivalent to
Proposition 16 applied to this condition yields the following result:5 27 Suppose that the pair of random variables nondecreasing in x.
PROPOSITION
are affiliated. Then
is
Karlin and Rinott (1980) provide examples of LSPM densities. For example, symmetric, positively correlated normal random variables are LSPM. This is also true for multivariate logistic and gamma distributions. 8.3.2 The Comparative Risk Aversion in the Sense of Ross In the analysis that we now provide we consider the simplest decision problem under uncertainty. Namely we examine a take-it-or-leave-it offer to gamble on a risk that is correlated to the background wealth of the decision maker, whose solution depends directly on the risk premium of the corresponding lottery. Let us define the premium lottery
as the price that agent ui is ready to pay to replace lottery
. It is obtained by
If
price πi is the risk premium that ui is ready to pay for the elimination of noise
to the initial risk
. If
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_121.html (1 of 2) [4/21/2007 1:31:01 AM]
by
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_121.html
and
is degenerated at zero, πi is the risk premium analyzed in section 2.3. In section 8.2 we obtained
conditions for π1 to be larger than π2 when the two sources of risk are independent.
5. A more general result of this can be found in Karlin (1968, ch. 1, prop. 3.1).
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_121.html (2 of 2) [4/21/2007 1:31:01 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_122.html
Page 122 Ross (1981) and Jewitt (1986) examined the case where to assume that is an increase in risk with respect to
and are stochastically dependent. The natural case is or equivalently that . In this case πi is the
premium that agent ui is ready to pay for the reduction of risk from
to
, or equivalently for the elimination
for the potentially correlated noise . It is intuitive that a more risk-averse agent will be ready to pay more for a given reduction of risk. However, we know from section 8.2 that it is not true in general. We want to guarantee that be larger than for all pairs such that is an increase in risk of . This question is actually dual to the concept of the Jewitt's stochastic where π2 is dominance order that we described in section 7.2.3. This would be true if defined by equation (8.16) for i = 2. Notice that a sufficient condition is that u1 be more risk-averse than u2 in the sense of Ross. Indeed, if u1 ≡ λu2 + g, with λ > 0 and g′, g″ ≤ 0, we have that
The first equality is by assumption. The second equality is due to the definition of π2. The first inequality is due to the concavity of g together with the fact that is an increase of risk with respect to inequality is obtained by observing that π2 is positive and g is decreasing.
. The second
More surprising is the fact that the Ross's notion of more risk aversion is necessary. This is claimed in the following proposition, which is due to Ross (1981): PROPOSITION
28 The following two conditions are equivalent:
1. Agent u1 is ready to pay more than agent u2 for any Rothschild-Stiglitz reduction in risk. 2. Agent u1 is more risk-averse than agent u2 in the sense of Ross. Proof It remains to prove that the Ross's notion of more risk aversion is necessary for property 1. To do this, let is distributed as and that the added noise to produce the us assume that file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_122.html [4/21/2007 1:31:01 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_123.html
Page 123 riskier
is degenerated at zero for any realization of , and is equal to
if
takes value z0. Define
as a function of the size k of the noise. Obviously πi(0) = 0, and by first-order risk aversion,
. Thus a necessary condition is that
be larger than
. This condition is written as
or
This condition holds for any distribution of only if
if and only if it holds for any degenerate distribution, that is, if and
This condition must hold for any pair (z, z0). Thus there must exist a scalar λ such that
It is then immediate to verify that this is equivalent to the condition that u1 ≡ λu2 + g, with g being decreasing and concave. This result is a good example of the difficulty encountered in extending existing results from the effect of an introduction/elimination of a pure risk to the analysis of a marginal increase/reduction in risk. In this case the fact that an agent is always willing to pay more than another for the elimination of a risk does not necessarily imply that he is willing to pay more for any marginal reduction of risk. In other words, even under DARA a more riskaverse agent (in the sense of Arrow-Pratt) may be more reluctant to pay for a risk-reducing investment.
8.4 Conclusion The presence of an uninsurable background risk in the wealth of a decision maker affects his behaviour towards other independent sources of risk. But under the standard assumptions that his utility function is concave and exhibits DARA, it is still true that he will file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_123.html [4/21/2007 1:31:02 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_124.html
Page 124 reject all unfair lotteries and that he will take more risk if his sure wealth is increased. Under DARA, an increase in risk aversion, namely concavifying the utility function, always increases the risk premium that the agent is ready to pay to escape an independent risk. However, this result does not hold if we consider the risk premium that the agent is ready to pay for a marginal reduction of risk, that is, for the elimination of a correlated noise. Ross's concept of an increase in risk aversion is the relevant concept to solve this last problem.
8.5 Exercises and Extensions Exercise 51 Pratt (1988) obtained the necessary and sufficient condition for the preservation of the comparative risk aversion when an independent background risk is added to the initially sure wealth. Show that this necessary and sufficient condition can be written as ∀x1, x2, ∀p ∈ [0, 1],
Exercise 52 Show that the Ross's notion of more risk aversion is stronger than that of Arrow-Pratt. Exercise 53 Show that an increase in wealth reduces the willingness to pay for a reduction in risk if and only if
This condition, called "Ross DARA," is equivalent to –u′ being more risk-averse than u in the sense of Ross. Exercise 54 Prove directly corollary 3 by using the properties of log-supermodular functions. Exercise 55 (Jewitt 1988) Consider the standard portfolio problem with background risk maxα , where the background risk is statistically related to the return of the risky asset. Suppose that is increasing in x and it is optimal for a given investor to hold α = 0, then everybody more risk-averse than the given investor will choose α < 0, and everybody less risk-averse will choose α > 0.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_124.html [4/21/2007 1:31:02 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_125.html
Page 125
9 The Tempering Effect of Background Risk In the previous chapter we were interested in determining whether the basic properties of the utility function are inherited by the indirect utility function in the presence of an exogenous background risk. More recent developments in this field focused on the impact of background risk on the degree of risk aversion. Does the presence of an exogenous background risk reduce the demand for other independent risks? In other words, are independent risks substitutes or complementary? This is an old question that was first raised by Samuelson (1963): I offered some lunch colleagues to bet each $200 to $100 that the side of a coin they specified would not appear at the first toss. One distinguished scholar . . . gave the following answer: "I won't bet because I would feel the $100 loss more than the $200 gain. But I'll take you on if you promise to let me make 100 such bets." This story suggests that independent risks are complementary. However, Samuelson went ahead and asked why it would be optimal to accept 100 separately undesirable bets. The scholar answered: "One toss is not enough to make it reasonably sure that the law of averages will turn out in my favor. But in a hundred tosses of a coin, the law of large numbers will make it a darn good bet." Obviously this scholar misinterpreted the law of large numbers. It is not by accepting a second independent lottery that one reduces the risk associated to a first one. If are independent and identically has a variance n time as large as the variance of each of these risks. What is distributed, stated by the law of large numbers is that —tends to almost surely as n tends to infinity. It is by subdividing—not adding
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_125.html [4/21/2007 1:31:03 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_126.html
Page 126 —risks that they are washed away by diversification, as shown in section section 3.3. An insurance company does not reduce its aggregate risk by accepting more independent risks in its portfolio. Once this fallacious interpretation of the law of large numbers is explained and understood, intuition suggests that independent risks should rather be substitute. The presence of one risk should have an adverse effect on the demand for another independent risk. This idea has been explored by various authors after Samuelson, such as Pratt and Zeckhauser (1987), Kimball (1993), and Gollier and Pratt (1996). In this chapter we examine the conditions under which the presence of an exogenous risk increases the aversion to other independent risks. This branch of the literature is referred to as the "background risk" problem. In the next chapter we will turn to the effect of the presence of an endogenous risk on the aversion to other independent risks.
9.1 Risk Vulnerability We now examine the impact of a background risk that has a nonpositive mean. The intuition is strong to suggest that such a background risk must increase the aversion to other independent risks. It should induce the agent to reject more lotteries, and to reduce his demand for any risky asset that has an independent return, whatever his initial wealth. The following definition has been proposed by Gollier and Pratt (1996): 5 Preferences exhibit "risk vulnerability" if the presence of an exogenous background risk with a nonpositive mean, namely an unfair risk, raises the aversion to any other independent risk. DEFINITION
Despite the intuitive content all expected utility preferences are not necessarily risk vulnerable, as shown in the following counterexample: Consider a risk-averse individual with piecewise linear utility u(w) that equals w if w is less than 100; otherwise, it equals 50 + 0.5w. If his initial wealth is 101, he dislikes a lottery offering –11 and +14 with equal probabilities, namely . But if one adds the mean-zero independent "background" risk to his initial wealth, the individual would find to be desirable. It is a paradox that an undesirable lottery can be made desirable by the presence of another (mean-zero) risk. Another illustration of this point is obtained by considering the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_126.html [4/21/2007 1:31:04 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_127.html
Page 127 standard portfolio problem. The risk-free rate is zero, whereas the return of the risky asset equals –100% or +190%, with equal probabilities. With a sure background wealth of 101, the agent would optimally purchase one unit of the risky asset. The same individual would purchase 10.5263 units of the risky asset in the presence of the undesirable risk background risk in addition to his 101 units of wealth. Introducing a zero-mean risk to initial wealth has multiplied the demand for the risky asset by a factor 10. From propositions 2 and 7, intuition is true if and only if the indirect utility function defined by condition (8.2) is more concave than the original utility function u, or if
or equivalently,
This condition is the technical condition for risk vulnerability: risk aversion is vulnerable to universally undesirable risks. We decompose the problem in two parts. We need first to show that any sure reduction in wealth raises risk aversion. Second, we also need to show that any zero-mean background risk raises risk aversion. The combination of these two conditions is necessary and sufficient for risk vulnerability, since any risk with a nonpositive mean can be decomposed into the sum of a degenerated random variable that takes a nonpositive value with probability 1, and a pure risk. The easy step is the first condition, which is trivially equivalent to decreasing absolute risk aversion. Thus DARA is necessary for risk vulnerability. We now present two sufficient conditions for risk vulnerability. The first one is that absolute risk aversion A(z) = –u″(z)/u′(z) be decreasing and convex. Decreasing absolute risk aversion, combined with proposition 15, implies that
If A is convex, Jensen's inequality implies that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_127.html [4/21/2007 1:31:04 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_128.html
Page 128 . The combination of the inequalities (9.3) and the second inequality being due to DARA combined with (9.4) implies condition (9.2), which defines risk vulnerability. The assumption of a convex absolute risk aversion is a natural assumption. It means in particular that the reduction of risk premium due to an increase in wealth is a decreasing function of wealth, at least for small risks (use the Arrow-Pratt approximation). Moreover absolute risk aversion cannot be positive, decreasing, and concave everywhere. Another sufficient condition for risk vulnerability is standardness. The concept of standardness has been introduced by Kimball (1993) to refer to the condition that absolute risk aversion and absolute prudence are decreasing with wealth. To prove the sufficiency of standardness for risk vulnerability, let us define the precautionary premium ψ associated to risk at wealth z as follows:
The analogy with the risk premium is obvious: ψ is the risk premium of for the agent with utility function –u′. In the terminology of chapter 3, ψ is equal to . Differentiating condition (9.5) with respect to z yields
and
By proposition 4, we know that ψ is decreasing in z because the decreasing absolute risk aversion of –u′ is equivalent to the decreasing absolute prudence of u. Combining this with the equality above implies that
Now recall that prudence is necessary for DARA. Thus the assumption that implies that is negative, since it is the sum of two negative terms. DARA concludes the proof that standardness is sufficient for risk vulnerability. These two sufficient conditions are not equivalent. Indeed, decreasing absolute prudence is itself equivalent to
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_128.html [4/21/2007 1:31:05 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_129.html
Page 129
Under DARA, this condition shows that, when P ≥ 2A, decreasing absolute prudence implies the convexity of absolute risk aversion. But nothing can be said when P is smaller than 2A. 29 The presence of an unfair background risk raises the aversion to other independent risks; that is to say, risk aversion is risk vulnerable, if at least one of the following two conditions is satisfied: PROPOSITION
1. Absolute risk aversion is decreasing and convex; 2. Absolute risk aversion and absolute prudence are decreasing. Notice that the subset of HARA utility functions defined by (23.5) that are DARA (γ ≥ 0) satisfies these two conditions. The necessary and sufficient condition is obtained by using the diffidence theorem on condition (9.2). It yields
This condition is not easy to manipulate. In addition, contrary to the other applications of the diffidence theorem presented earlier in this book, it does not simplify to the unidimensional condition (similar to A > 0 or P > 2A for example). Here we must verify that a function of two variables is positive everywhere. A necessary condition is derived from the local necessary condition (6.15). It is written here as
or
This means that absolute risk aversion may not be too concave. Careful differentiation shows that this condition can be rewritten as
To sum up, the necessary and sufficient condition for risk vulnerability is rather complex. We have two necessary conditions, and two sufficient conditions that are much easier to check. The two necessary conditions are DARA file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_129.html (1 of 2) [4/21/2007 1:31:05 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_129.html
and condition (9.8). The two sufficient conditions are stated in proposition 29.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_129.html (2 of 2) [4/21/2007 1:31:05 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_130.html
Page 130
9.2 Risk Vulnerability and Increase in Risk 9.2.1 Increase in Background Risk Risk vulnerability guarantees that the introduction of a pure risk increases the aversion to other independent risks. Does it also guarantee the same result for any marginal second order stochastically dominated change in background risk? In other terms, suppose that
dominates
in the sense of SSD. Does it imply that
is more concave than
for all z? Because
dominates
, or
in the sense of SSD, we know that
. Let us first suppose that noise
is independent of
is distributed as
, with
. In this case the problem simplifies to
whether is more concave than . Observe that risk vulnerability implies that v is more concave than u. Does this imply that is more concave than ? In other words, does an exogenous background risk preserve the comparative risk aversion order? This is precisely the question that we raised in section 8.2. We showed in proposition 24 that DARA of one of the two utility functions is sufficient for the preservation of the comparative risk aversion order. Since DARA is necessary for risk vulnerability anyway, this proves that adding an independent noise to background risk raises the aversion to other independent risks under risk vulnerability. The problem is less obvious in the case of a correlated noise. A full characterization of the solution to this problem is in Eeckhoudt, Gollier, and Schlesinger (1996). 9.2.2 Increase in the Endogenous Risk Risk vulnerability brings another view on the effect of an increase in risk of the return of a risky asset on its demand, something that we already examined in section 4.5. In the standard portfolio problem without to background risk, suppose that the return of the risky asset undergoes an increase in risk from with We have shown in that section that this does not necessarily imply that risk-averse investors will reduce their demand for the risky asset. We now provide another intuition for this phenomenon. Normaliz-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_130.html [4/21/2007 1:31:06 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_131.html
Page 131 ing the initial demand to unity with
, they reduce their demand if
We see that we can decompose the effect of an increase in risk of return into a wealth/background risk effect, on one hand, and a substitution effect, on the other hand. Let us start with the substitution effect, which is represented by the second term in the right-hand side of condition. This term is always negative, under risk aversion. Indeed, using proposition 15, we have that
The first term in the right-hand side of condition (9.9) is typically a background risk effect. How does the introduction of risk to wealth affect the demand for ? Because the effect of a background risk is ambiguous, the global effect of an increase in risk of return is also ambiguous. Moreover this background risk is potentially correlated to the initial risk . However, in the special case of a Rothschild-Stiglitz increase in risk that takes the form of adding an independent noise, this second term is unambiguously negative under risk vulnerability. We obtain the following proposition, which is due to Gollier and Schlesinger (1996). 30 Suppose that u is risk vulnerable. Then, in the standard portfolio problem, adding an independent zero-mean noise to the distribution of returns of the risky asset reduces the demand for it. PROPOSITION
9.3 Risk Vulnerability and the Equity Premium Puzzle. Does the existence of an uninsurable risk on human capital could explain the equity premium puzzle? This question was raised by Mehra and Prescott (1985), and it has been examined by Weil (1992). What the analysis above shows is that it can help solving the puzzle if one assumes that preferences are risk vulnerable. The question now is to determine whether it yields a sizable effect. The model is as follows: In addition to the one unit of physical capital, each agent in the economy is endowed with some human capital. As in the basic model presented in chapter 5, the revenue
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_131.html [4/21/2007 1:31:06 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_132.html
Page 132
Figure 9.1 Equity premium with background risk
generated by each unit of physical capital is denoted . It is perfectly correlated across firms. The revenue generated by the human capital is denoted . It is assumed that it is independently distributed across agents, and that it is also independent of . The equity premium in this economy will be equal to
To keep things simple, let us assume that background risk is distributed as (–k, 1/2; +k, 1/2). Parameter k is the standard deviation of the growth of labor incomes. Using the historical frequency of the growth of GDP per capita for as presented in section 2.8, and a CRRA utility function with γ = 2, we obtain the equity premium as a function of the size k of the uninsurable risk. We report these results in figure 9.1. We see that a very large background risk is required to explain the puzzle, with a standard deviation of the annual growth of individual labor income exceeding 80%. We conclude that the existence of idiosyncratic background risks, when considered in isolation, cannot explain the equity premium puzzle. Telmer (1993) and Lucas (1994) obtained a similar conclusion in a dynamic framework.
9.4 Generalized Risk Vulnerability A consequence of risk vulnerability is that any background risk with a nonpositive mean raises the risk premium that one is ready to pay
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_132.html [4/21/2007 1:31:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_133.html
Page 133 to escape another independent risk. It implies in particular that the presence of a nonpositive-mean background risk will never transform another undesirable independent risk into a desirable one. As in Pratt and Zeckhauser (1987) and Kimball (1993), one could raise the same question for other types of background risk. Let Σi(z, u) be a set of background risks that satisfy some properties that can depend on the preference u of the agent and on his initial wealth z. We say that the von Neumann–Morgenstern utility function u is vulnerable at z with respect to Σi (z, u) if an undesirable risk given background wealth z can never be made desirable by the introduction of any independent background risk in Σi(z, u):
If property (9.11) holds for every z, then u is said to be vulnerable to Σi (i.e., u ∈ Vi). Observe that risk vulnerability at z with respect to Σi(z, u) is equivalent to the condition that the indirect utility function is more diffident at z than the original utility function, namely that
for all y. If condition ∈ Σi(z, u) is linear in probabilities, it is a straightforward exercise to apply the diffidence theorem to condition (9.12) to obtain a necessary and sufficient condition for (9.11). Thus vulnerability at z with respect to Σi(z, u) just says that the indirect utility function v is more diffident at z than the original utility function u for any background risk in Σi(z, u). As we know, this does not necessarily imply that v is more concave than u. We can just infer of it that, locally at z, –v″(z)/v′(z) is larger than –u″(z)/u′(z). There are four sets Σi, i = 1, 2, 3, 4, that lead to known restrictions on the utility function. Decreasing Absolute Risk Aversion: V1 DARA is necessary and sufficient for a reduction in wealth to enlarge the set of undesirable risks. DARA is thus
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_133.html (1 of 2) [4/21/2007 1:31:08 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_133.html
equivalent to vulnerability
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_133.html (2 of 2) [4/21/2007 1:31:08 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_134.html
Page 134 to
with probability 1}. It is equivalent to A′ ≤ 0, or P ≥ A.
Risk Vulnerability: V2 By definition, risk vulnerability is equivalent to vulnerability to
. Obviously Σ1 belongs to Σ2.
It implies that risk vulnerability is more demanding than DARA, since the same property is required to hold for a larger set of background risks under risk vulnerability than under DARA. It is noteworthy that the concepts V1 and V2 rely on background risk sets Σ1 and Σ2 whose characterizations are independent of z and u, in contrast to the next two concepts. Properness: V3 The concept of proper risk aversion proposed by Pratt and Zeckhauser (1987) corresponds in our terminology to , the set of undesirable risks at w. It answers to the question vulnerability to raised by Samuelson (1963) that two independent risks that are separately undesirable are never jointly desirable. Observe that by risk aversion, Σ2 is a subset Σ3(z, u). Therefore Pratt and Zeckhauser's concept of properness is a special case of risk vulnerability V2. A necessary condition for properness is that
Using corollary 2, one can check that a necessary condition for this is that A″ be larger than A′A. This is the local properness condition found by Pratt and Zeckhauser. Standardness: V4 The concept of standard risk aversion introduced by Kimball (1993) corresponds to vulnerability to , i.e. the set of expected-marginal-utility-increasing risks. Given risk aversion, Σ4(z, u) includes Σ1 as a subset. Therefore decreasing absolute risk aversion is necessary for V4. Given this fact, Σ3(z, u) is a subset of Σ4(z, u), since decreasing absolute risk aversion means that –u′ is more concave than u. Therefore, as shown by Kimball,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_134.html [4/21/2007 1:31:08 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_135.html
Page 135 standardness implies properness (which itself implies risk vulnerability). A specific examination of standardness is presented in the next section. To sum up, we have that
Recall that risk vulnerability at z with respect to Σi(z, u) just says that the indirect utility function v is more diffident at z than the original utility function u for any background risk in Σi(z, u). In general, we cannot infer from this assumption anything about whether the agent with wealth z will invest less in the risky asset due to the presence of a background risk in Σi(z, u). That would require that the indirect utility function be centrally more risk-averse than the original utility function at z. As we know from proposition 13, this is a condition stronger than more diffidence at z. One may thus define a notion of "strong vulnerability" that is more demanding than the notion defined above. We say that the utility function u is strongly vulnerable at z with respect to Σi(., u) if any background risk in Σi(z, u) makes the indirect utility function centrally more risk-averse than u at z; that is, it reduces the demand for any independent risky asset:
for all y. We know that strong vulnerability and vulnerability do not differ for the case Σ1 and Σ2. We show in the next section that it is also true for Σ4. No general result has been established for case Σ3.
9.5 Standardness Let us take a closer look at the notion of standard risk aversion. We should find that decreasing absolute risk aversion and decreasing absolute prudence are necessary and sufficient for both conditions (9.12) and (9.14) for . These results are in Kimball (1993). Any risk in Σ4(z, u) can be decomposed into an expected-marginal-utility preserving risk and a sure reduction in wealth. DARA takes care of the wealth effect. Thus we now assume DARA,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_135.html [4/21/2007 1:31:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_136.html
Page 136 and we show that decreasing absolute prudence is necessary and sufficient for an expected-marginal-utility , or preserving risk to yield the result. Consider any background risk such that where ψ is the precautionary premium. The necessity of decreasing prudence is immediate. Recall that a necessary condition for both more diffidence at z and central more risk aversion at z is an increase in risk aversion locally at z. As we observed earlier, the absolute risk aversion of the indirect utility function satisfies the following property:
Because we need this to be larger than –u″(z)/u′(z) = A(z), a necessary condition is that ∂ψ/∂z be negative. By proposition 4, this condition is equivalent to decreasing absolute prudence. We now prove the sufficiency of decreasing absolute prudence for condition (9.14) to hold for any expectedmarginal-utility preserving risk. Notice that decreasing absolute prudence means that –u″(z + y + ξ) is logsupermodular with respect to (y, ξ). We write this as
for all ξ > 0, with the inequality reversed for ξ ≤ 0. Since the direction of integration is opposite the direction of the inequality, it can be integrated from ξ = 0 to ξ = x to yield
Since this condition holds for all x, taking the expectation yields
In consequence, decreasing absolute prudence is sufficient for implies that
. Dividing by
for all y, which means that the indirect utility function is centrally more risk-averse at z than the original utility function. We conclude that standardness is necessary and sufficient for any expected-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_136.html [4/21/2007 1:31:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_137.html
Page 137 marginal-utility-increasing background risk to reduce the demand for the risky asset. Because more diffidence is a weaker concept than central more risk aversion, this also proves that standardness is sufficient for the indirect utility function to be more diffident at z than the original utility function, for any . Since we already proved that standardness is necessary, we obtain the following proposition: PROPOSITION
31 Standard risk aversion is necessary and sufficient for any of the two following properties to hold:
1. Any expected-marginal-utility-increasing background risk reduces the demand for an independent risky asset. 2. Any expected-marginal-utility-increasing background risk makes the agent to reject more independent lotteries. Kimball (1993) provides a variant to this proposition. A problem with the result above is that it deals with a background risk that increases the expected marginal utility given sure wealth z, whereas we assume that the agent has the opportunity to invest this wealth in risky activities. Kimball shows that the same result applies for background risks, which raises Eu′ evaluated with the optimal exposure to the endogenous risk. More precisely, the problem is written as
The first condition states that the investor with sure initial wealth z invests one dollar in the risky asset whose return is . The second condition states that background risk increases the expected marginal utility in the presence of the optimal exposure to risk . The implication is that this background risk must reduce the demand for . PROPOSITION 32 Standard risk aversion is necessary and sufficient for property (9.15), which states that any background risk that raises Eu′ in the presence of the initial optimal investment in the risky asset reduces the demand for this asset.
Proof We focus on the proof of sufficiency, which is based again on the fact that decreasing absolute prudence is equivalent to the log-supermodularity of –u″(z + y + ξ) with respect to (y, ξ). It yields
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_137.html [4/21/2007 1:31:10 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_138.html
Page 138
which has the same sign as ξ(y2 – y1). Integrating with respect to ξ, we have that
which has the same sign as (y2 – y1). This expression implies that distribution of
, and . Taking
and
is nonnegative for any
i.i.d. and distributed as , this condition becomes
Recall now that DARA means that –u′ is more concave than u. It yields
In consequence the right-hand side of inequality (9.16) is nonnegative. A necessary condition for (9.16) is thus . To sum up, standardness generates the same comparative statics properties as risk vulnerability, at least for the standard portfolio problem and for the take-it-or-leave-it lottery problem. Because standardness does not imply that the indirect utility function be more concave, there is no guarantee that standardness can generate other comparative statics results. Still standardness has two important advantages. First, it is easy to characterize: contrary to risk vulnerability that is characterized by a two-dimension inequality (9.7), standardness just require that a one-variable function, P′, to be nonnegative. Second, the comparative statics property holds for a larger set of background risks (Σ2 ⊂ Σ4(z, u)). This is at the cost of a stricter condition on preferences (standardness implies risk vulnerability).
9.6 Conclusion It is widely believed that the presence of a risk in background wealth has an adverse effect on the demand for other independent risks. In
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_138.html [4/21/2007 1:31:10 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_139.html
Page 139 other words, the belief is that this background risk is substitute for other independent risks. This is not true in general, and the conditions that must be imposed on preferences to guarantee this result depend on the kind of background risk we have in mind. Not surprisingly, because they raise the question of the effect of risk on the degree of risk aversion –u″/u′, risk vulnerability and its various related concepts rely in one way or another on the fourth derivative of u. Despite this fact some of these conditions are intuitive. For example, a sufficient condition for a pure background risk to raise the aversion to other independent risks is that absolute risk aversion be decreasing and convex. This is a natural assumption, as it means that the risk premium is decreasing with wealth at a decreasing rate. However, Ross (1999) takes the opposite view that independent risks could well be complement: the two separately undesirable risks can be jointly desirable. There are two ways to resolve this dispute. The first would be to perform experiments on how people behave in the face of multiple risks. The other way is to examine the conditions on preferences that make independent risks substitutes or complements. If they can be related to other sensible properties of risk-averse agents, that would be an argument in favor of either substitutability or complementary. For example, the fact that a decreasing and convex absolute risk aversion is a sufficient condition for the Gollier-Pratt's notion of substitutability is clearly an argument in favor of the hypothesis that independent risks are substitutes.
9.7 Exercises and Extensions Exercise 56 Derive the necessary and sufficient condition for an undesirable background risk to raise local risk aversion (condition 9.13). Show that two necessary conditions are DARA and A″ ≥ A′A. Exercise 57 (Eeckhoudt, Gollier, and Schlesinger 1996) Consider a background risk – p; x2 – t, p). An increase in t deteriorates
that is distributed as (x1, 1
in the sense of first-order stochastic dominance. Define also the
indirect utility function . The intuition suggests that an increase in t should make the agent more averse to other independent risks, namely that is increasing in t. • Show that At(z) is increasing in t if and only if P(z + x2 – t) ≥ At(z).
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_139.html [4/21/2007 1:31:11 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_140.html
Page 140 • Let x1 and x2 be arbitrary. Show that the necessary and sufficient condition becomes P(z) ≥ A(z′) for all z, z′. In short, any FSD deterioration of the background risk raises the aversion to other risks only if the agent is RossDARA (see equation 8.18). • Consider now an increase in risk
, with
. Find a necessary condition for At to
be increasing in t. Exercise 58 (Athey 1999) Because the condition Ross-DARA that we obtained in the previous exercise is very restrictive, it may be interesting to relax it by constraining the set of FSD deterioration of the background risk. Let f(x, t) be the density function of background risk . The marginal indirect utility function can therefore be written as . Using the properties of log supermodular functions, prove that a MLR deterioration in the distribution of the background risk raises the degree of risk aversion of vt if u is DARA in the sense of Arrow-Pratt.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_140.html [4/21/2007 1:31:12 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_141.html
Page 141
10 Taking Multiple Risks In the previous chapter we examined the effect of an exogenous risk borne by an agent on his behavior toward other independent risks. We now turn to the analysis of the behavior towards independent risks that are endogenous. Typically we address here the question of how does investing in a portfolio influence the agent's behavior toward insuring her car. Or, how does the opportunity to invest in security A affect the demand for another independent security B? A last illustration is about the effect of the opportunity to invest in risky assets on the decision to invest in human capital. What is the effect of having the option to invest in one risky asset on the optimal demand for another independent risky asset? There is a priori no obvious reason for these two risks to be substitute or complement. This is due to the existence of two contradictory effects of the option to invest in a risk. To simplify, let us assume that the two risks are independent and identically distributed. The first effect of adding the second risk in the opportunity set of the agent is a diversification effect. If the budget devoted to the purchase of risks is fixed, it is optimal for the risk-averse agent to split this budget into two equal parts. As shown in proposition 5, any other composition of the portfolio would be second-order stochastically dominated by this perfectly diversified one. Thus, if the agent is constrained by a fixed budget for his investment in risky assets, the introduction of the second asset would reduce the investment in the first asset by 50%. This is a diversification effect. The second effect comes from the relaxation of the fixed budget constraint. The new diversified portfolio of risky assets is less risky than each of the individual assets. Since it becomes more attractive to risk-averse investors, it is likely that they will substitute the risk-free asset to this new portfolio of
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_141.html [4/21/2007 1:31:12 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_142.html
Page 142 risky assets.6 This portfolio substitution effect raises the demand for the initial risky asset. The diversification effect and the portfolio substitution effect go into opposite directions, and the global effect is a priori ambiguous. Another decomposition of the effect of one endogenous risk on the optimal exposure to another risk can be made. Introducing a new asset has a wealth effect and a pure risk effect. The wealth effect comes from the increase in the expected wealth generated by investing in this new risky asset. If investing in it would have a negative expected return, no risk-averse investor would participate. Under decreasing absolute risk aversion, this increase in expected wealth has a positive effect on the demand for other independent risks. If this would be the only effect, independent risky assets would be complements. But the opportunity to invest in a new risky asset has not the same comparative statics effect as just adding its certainty equivalent to the initial wealth of the investor. It also introduces a zero-mean, or pure, risk to background wealth. We know from chapter 9 that it has a tempering effect if the agent is risk vulnerable. Thus the pure risk effect is compatible with the substitutability of independent risks. It implies that risk vulnerability is a necessary condition for the substitutability of independent risks. Whether the effect of risk vulnerability dominates the wealth effect or not is the open question that we address in this chapter. Most results presented in this chapter are from Eeckhoudt and Gollier (1999).
10.1 The Interaction between Asset Demand and Small Gambles We start with a model in which the agent can invest in a risky asset and, at the same time, gamble a fixed amount in a risky activity. We examine the interaction between these two decisions. The problem is written as
Random variable is the return of the risky asset whereas is the net gain in gambling, which is assumed to be independent of . Decision
6. We should take note that it is not generally true that a reduction in the risk of an asset increases the demand for it. Gollier (1995) provides the necessary and sufficient condition on the change in the distribution of returns such that all risk-averse agents increase their demand for the risky asset.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_142.html [4/21/2007 1:31:13 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_143.html
Page 143 variable α is the amount invested in the risky asset, whereas δ is the amount invested in gambling, which is constrained to be 1 (take-it) or 0 (leave-it). The intuition suggests that the opportunity to invest in financial markets should reduce the incentive to take other risks. This means that the investor would reject more gambles than if he would not be allowed to invest in financial markets. In this section we limit the analysis to the case where the gamble is small with respect to portfolio risk. Thus the problem simplifies to determining whether the optimal portfolio risk raises local risk aversion:
Applying the diffidence theorem generates the following necessary and sufficient condition:
Rearranging, this is equivalent to
which is true if and only if absolute risk aversion is convex. PROPOSITION
33 Investing in a risky asset raises local risk aversion if and only if absolute risk aversion is convex.
Recall that the convexity of absolute risk aversion, combined with DARA, is sufficient for risk vulnerability, which itself is a natural assumption. There is an alternative way to prove proposition 33. Let F be the cumulative distribution function of . Let the risk-neutral cumulative distribution function G be defined by
Using Ê for the expectation operator associated to the probability distribution G, condition (10.1) may then be rewritten as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_143.html [4/21/2007 1:31:13 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_144.html
Page 144
The parallelism with condition (6.20) is perfect when one replaces u by –A. The convexity of A is clearly necessary and sufficient for the result. We can further define the substitution premium ι of , defined as
In words, the substitution premium associated to the opportunity to invest in a risky asset is the sure reduction in wealth that has the same effect on the demand for another small risk. It is positive if absolute risk aversion is decreasing and convex. Using the Arrow-Pratt approximation, it is approximately equal to
–A″/A′ is the local measure of risk substitutability. It is the sure increase in wealth that has the same effect on the attitude towards risk as the optimally selected risk whose risk-neutral variance equals 2.
10.2 Are Independent Assets Substitutes?. We now turn to the demand interaction between independent assets. It is intuitive to think that the demand for a risky asset should be reduced by the presence of other independent risky assets, meaning that independent risky assets are substitutes. We start with an analysis of the i.i.d. case, which is easier to treat. 10.2.1 The i.i.d. Case Suppose that there are two risky assets available in the economy, in addition to the risk free one. Their returns and are independent and identically distributed. Echoing Samuelson's story presented in the introduction of chapter 9, suppose that you are offered to invest in alone. Contrary to Samuelson (1963), we assume here that you can gamble how much you want on . Suppose now that you are also offered the opportunity to invest how much you want in another
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_144.html [4/21/2007 1:31:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_145.html
Page 145 identical and independent gamble. Would you invest less in as a consequence? Because we assume here that and are i.i.d., we can use proposition 5 to simplify the analysis. Indeed, this proposition implies that in the presence of two i.i.d. assets, it is optimal to invest the same amount in each of them. Normalizing the demand to unity when there is only 1 asset, the presence of a second i.i.d. asset reduces the demand for the first one if
where the condition to the right states that the derivative of is negative when evaluated at the initial optimal exposure α = 1. By symmetry, this right condition is equivalent to . In fact we will look for a weaker condition:
relaxing the condition that and
must be identically distributed.
This condition states that the presence of background risk that is optimal when taken in isolation reduces the is centrally demand for any other independent asset. This is true only if the indirect utility function : more risk-averse than u at z, for any background risk satisfying condition
Let us define a new set of random variables: notion of "strong vulnerability" with respect to diffidence theorem, this condition is equivalent to
. Condition (10.5) defines the , as defined in general terms by condition (9.14). Using the
being nonpositive for all (z, x, y) in the feasible domain.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_145.html (1 of 2) [4/21/2007 1:31:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_145.html
It is easy to prove that the convexity of absolute risk aversion is a necessary condition for (10.6). This is done by verifying that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_145.html (2 of 2) [4/21/2007 1:31:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_146.html
Page 146
and
The first two properties imply that a necessary condition is ∂2H/∂y2|y=0 be nonpositive, which in turn necessitates that A(z + x) ≥ A(z) + xA′(z). A similar exercise may be performed with respect to x, with the same result. Condition (10.6) is not easy to manipulate. It can thus be useful to try to provide a simpler sufficient condition. We use the same method as in the previous chapter where we showed for example that (strong) vulnerability with respect to Σ4(z, u) (standardness) is sufficient for (strong) vulnerability with respect to Σ2(z, u) (risk vulnerability) because Σ2(z, u) ⊂ Σ4(z, u). If we were able to show that
is in Σi(z, u), we would have in hand the
sufficiency of Vi. The problem is that any risk satisfying condition
must have a positive mean,
together with being desirable (otherwise, α = 0 would dominate strategy α = 1). Thus the only potential candidate is Σ4(z, u). We would have in Σ4(z, u) if
In words, we are looking for the condition under which the option to purchase stocks increases the marginal utility of wealth. As stated in the following proposition, this is the case if prudence is larger than twice the risk aversion (P ≥ 2A). 34 Condition (10.7) holds for any z and if and only if absolute prudence is larger than twice absolute risk aversion, that is, if and only if P(z) ≥ 2A(z) for all z.
PROPOSITION
Proof By the diffidence theorem (corollary 1), condition (10.7) holds if and only if
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_146.html [4/21/2007 1:31:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_147.html
Page 147 for all (z, x). If we divide the above condition by u′(z + x)u′(z), this inequality is rewritten as
where f(w) = 1/u′(w). Obviously, this condition holds for any (z, x) if and only if f is concave. But it is easily checked that f = 1/u′ is concave if and only if
or P(z) ≥ 2A(z).7 This concludes the proof of the proposition, which is due to Gollier and Kimball (1996).8 We conclude that this condition joint with standardness is sufficient for conditions (10.3), (10.4), (10.5), and (10.6). Because decreasing absolute prudence means that A″ ≥ –A′(P – 2A), we verify that this sufficient condition satisfies the necessary condition A″ ≥ 0. PROPOSITION 35 Two risky assets with i.i.d. returns are substitutes if absolute prudence is decreasing and larger than twice the absolute risk aversion. A necessary condition for the substitutability of independent risky assets is that absolute risk aversion be convex.
To examine how far away this sufficient condition is from the necessary and sufficient condition (10.6), we can look at a specific set of utility functions. As usual, let us consider the set of HARA functions, with
We know that these functions have a convex absolute risk aversion and are standard if γ is positive. They satisfy condition P ≥ 2A if γ is less than 1. Thus the sufficient condition described in proposition 35 is 0 < γ ≤ 1. Can we hope condition (10.6) to be satisfied for other values of γ? In the case η = 0, this sufficient condition is rewritten as 7. This condition is equivalent to T′(z) ≥ 1, where T = 1/A is absolute risk tolerance. 8. This result can also be proved by using proposition 16, with g(x) = x, h(x, 1) = u′(z + x) and
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_147.html [4/21/2007 1:31:15 AM]
.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_148.html
Page 148
Figure 10.1 Evaluation of H(1, 1, y) for the CRRA utility function, with γ = 5
being nonpositive for all (z, x, y) in the feasible domain
Fix x and y positive, and let z approach 0. Then the first and the third terms in the right-hand side of (10.10) tend to infinity. If γ is larger than 1, this is the first that dominates, and H tends to +∞. Function H is depicted in figure 10.1 for γ = 5, z = 1, x = 1, and η = 0. We see that H and ∂H/∂y are both zero at (z, x, 0). The convexity of A implies that ∂2H/∂y2 is negative at (z, x, 0), which implies that H is nonpositive in the neighborhood of y = 0. But H is not nonpositive everywhere. To be more explicit, take the CRRA utility function with γ = 5, z = 1, and and being i.i.d. and distributed as (– 0.1, 4/10; 10, 6/10). After some computations we get that the optimal investment when only one risky asset is offered is 0.1678. When the second risky asset is also offered, the investor increases his demand for the first asset up to 0.1707. Our conclusions from this example are twofold: first, the convexity of absolute risk aversion is not sufficient for the substitutability of large independent risks. Second, when constant relative risk aversion is larger than unity, there always exist pairs of inde-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_148.html [4/21/2007 1:31:16 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_149.html
Page 149
Figure 10.2 Relative risk aversion = 0.5
pendent risks that are complement, contrary to when constant relative risk aversion is less than unity. We may be interested in examining the optimal portfolio with more than two i.i.d. assets. Let us consider the following numerical examples. We consider a sequence of ten i.i.d. assets which generate return –0.1 with probability 0.4 and return 10 with probability 0.6, per euro invested. Initial wealth has been normalized to unity. We draw the optimal exposure to each individual risk as a function of the number of risk exposures that the agent faces. Figure 10.2 corresponds to the case where relative risk aversion equals 0.5. According to proposition 35, the optimal exposure to each risk is decreasing with the number n of risk exposures. In fact, given the large expected return of each risky project and the low risk aversion of the agent, the optimal exposure to each risk is almost equal to 10/n. This strategy maximizes the expected return under the constraint that the final wealth remains positive with probability 1. Figure 10.3 corresponds to a relative risk aversion equaling 2. The curve is hump-shaped: the initial two independent risks are complements. But as the number of risks increases, additional risks become substitutes to the others. Finally, in figure 10.4, the ten first i.i.d. risks are complements. We can also reinterpret the model presented above as the problem of an investor who lives for two periods. Random variables and are the return of the risky asset respectively in the first period and in
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_149.html [4/21/2007 1:31:16 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_150.html
Page 150
Figure 10.3 Relative risk aversion = 2
the second period. Suppose that the investor cannot rebalance his portfolio over time and that and are i.i.d. random variables. Then condition (10.3) means that the investor with two periods to live before consuming his wealth should take less risk than the investor with just one period to go. In the CRRA case we have shown that this is the case if relative risk aversion is less than unity. If it is larger than unity, we cannot conclude. Jagannathan and Kocherlakota (1996) performed some numerical simulations for CRRA utility functions by using the historical returns of stocks and bonds. They obtained that the relative share of wealth invested in stocks is decreasing in the time horizon even when relative risk aversion is as large as 5. Figure 10.4 shows that this is not true in general. To end up, observe that the analysis presented here is not realistic for examining the effect of time horizon and age on the optimal portfolio risk because, contrary to what we assume here, investors can rebalance their portfolio over time. The relevant model of the optimal dynamic portfolio management will be examined in chapter 11. 10.2.2 The General Case In this paragraph we remove the assumption that and are identically distributed. The absence of symmetry forces us to perform a comparative statics analysis with two decision variables.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_150.html [4/21/2007 1:31:17 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_151.html
Page 151
Figure 10.4 Relative risk aversion = 5
Figure 10.5 is less than 1 if
is negative
Consider a function g(αx, αy) that is concave with respect to (αx, αy). Let
denote the optimal solution of
the maximization of g(αx, αy). Under which condition is
less than 1? The standard method consists in first
obtaining the αx that maximizes g(αx, 1). Let us denote it
. Then
evaluated at
is less than 1 if ∂g/∂αy is negative when
. This idea is described in figure 10.5.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_151.html (1 of 2) [4/21/2007 1:31:17 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_151.html
We apply this technique for , which is concave under risk aversion. As before, normalize to 1 the demand for when it is the only asset on the market. Let us also normalize to unity. This means that the optimal demand for in the presence
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_151.html (2 of 2) [4/21/2007 1:31:17 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_152.html
Page 152 of one unit of equals 1. Thus, the demand for the following property holds:
is reduced by the opportunity to also invest in
if and only if
The condition (i) states that the optimal demand for asset equals 1, when it is the only asset offered on the market. Condition (ii) states that the optimal demand for asset when the investor already has one unit of is also 1. Condition (iii) requires that the investor then demands less than one unit of . Except for condition (ii), this problem is the same as problem (10.4). Applying the diffidence theorem twice— once by taking as given, and then again over —would yield a rather complex necessary and sufficient condition. We will limit our analysis to showing that the combination of standardness with condition P > 2A is also sufficient for (10.11). We would be done if we could show that background risk in condition (iii) raises the expected marginal utility of wealth. The trick here is to use proposition 32 rather than proposition 31, by showing that
when P is larger than 2A. Knowing that condition (9.15) holds under standardness would yield the result. To prove that condition (10.12) holds under P > 2A, define the indirect utility function problem can then be rewritten as
. The
Using proposition 34, this is true if and only if the indirect utility function satisfies condition Pv > 2Av. But we also know by proposition 23 that the indirect utility function inherits this property from u. We conclude that, under condition P > 2A, property (10.11) holds if property (9.15) holds. The use of proposition 32 concludes the proof of the following proposition, which generalizes proposition 35 to assets whose returns are not identically distributed.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_152.html [4/21/2007 1:31:18 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_153.html
Page 153 36 Suppose that absolute risk aversion is decreasing. Independent risky assets are substitutes if absolute prudence is decreasing and larger than twice absolute risk aversion. A necessary condition for the substitutability of independent risky assets is the convexity of absolute risk aversion. PROPOSITION
10.3 Conclusion In this chapter we approached the difficulty to determine the optimal composition of a portfolio when departing from mainstream finance which depends on the mean-variance paradigm. We looked here at the simplest case that we can imagine. We assumed the existence of one risk free and two risky assets with independent returns. Except when absolute risk aversion is constant, the presence of the second risky asset affects the demand for the first. The intuition suggests that it should reduce it. We showed that this is true if absolute risk aversion is standard and smaller than half the absolute prudence.
10.4 Exercises and Extensions Exercise 59 The objective of this exercise is to examine the substituability of independent assets in a meanvariance framework. Consider the utility function u(z) = cz – 0.5z2. • Show that the absolute risk aversion is convex. • Show that the necessary and sufficient condition H(z, x, y) ≤ 0—where H is defined by (10.6)—is written here as –x2y2 ≤ 0. • Show that
and
where αi is the optimal investment in each risky asset when i risky assets are available, and • Discuss and connect these results.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_153.html [4/21/2007 1:31:18 AM]
.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_154.html
Page 154 Exercise 60 You are allowed to take a stake k to a risky investment whose net return is . Let . Show that
What does this result imply for the effect of taking a marginal stake in a desirable risk on the demand for other independent risks? Exercise 61 Suppose that investors are allowed to invest in a risky project, but there is an upper bound k to their investment in it. Derive the necessary and sufficient condition for this new investment opportunity to raise local risk aversion.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_154.html [4/21/2007 1:31:19 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_155.html
Page 155
11 The Dynamic Investment Problem How should the length of one's investment horizon affect the riskiness of his portfolio? This question confronts start-up companies choosing which ventures to pursue before they go public, ordinary investors building a nest egg, investment managers concerned with contract renewals and executives seeking strong performance before their stock options come due, among others. Portfolio decisions are the focus of this chapter. However, the horizon-risk relationship extends beyond finance. Thus students may vary their strategy on grades—say how venturesome a paper to write—over the course of the grading period; presidents may adjust the risk of their political strategies from early through midterm and then as they approach an election. Popular treatments suggest that short horizons often lead to excessively conservative strategies. Thus, the decisions of corporate managers, judged on their quarterly earnings, are said to focus too much on safe, shortterm strategies, with underinvestment say in risky R&D projects. Privately held firms, it is widely believed, secure substantial benefit from their ability to focus on longer-term projects. Mutual fund managers, who get graded regularly, are also alleged to focus on strategies that will assure a satisfactory short-term return, with longterm expectations sacrified. Economists and decision theorists, speculators and bettors, have long been fascinated by the problem of repeated investment. In recent years this class of problems has been pursued in two different literatures, one in utility theory and the other in dynamic investment strategies. A recurring theme in both is that the opportunity to make further investments affects how one should invest today. These two literatures have not merged, in part because a key question has gone unresolved. How should the length of the investment horizon affect
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_155.html [4/21/2007 1:31:19 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_156.html
Page 156 the riskiness of one's investments? Some special cases have yielded results, as in the case of HARA utility functions. And with any particular utility function and set of investment opportunities actual calculations, perhaps using a simulation (e.g., see Viciera 1998 and Cocco, Gomes, and Maenhout 1998), could answer that question within a dynamic programming framework. But the central theoretical question of the link between the structure of the utility function and the horizon-riskiness relationship remained unresolved. Guiso, Jappelli, and Terlizzese (1996), among others, test the relation between age and risk-taking in a cross section of Italian households. Their empirical results show that young people actually hold the smallest proportion of risky assets in their portfolios. The share of risky assets increases by 20% to reach its maximum at age 61 (Guiso et al., p. 165). In the formal literature the horizon-riskiness issue has received the greatest attention addressing portfolios appropriate to age. Samuelson (1989) and several others have asked: "As you grow older and your investment horizon shortens, should you cut down your exposure to lucrative but risky equities?" Conventional wisdom answers affirmatively, stating that long-horizon investors can tolerate more risk because they have more time to recoup transient losses. This dictum has not received the imprimatur of science, however. As Samuelson (1963, 1989a) in particular points out, this "time-diversification" argument relies on a fallacious interpretation of the law of large numbers: repeating an investment pattern over many periods does not cause risk to wash out in the long run. Early models of dynamic risk-taking, such as those developed by Samuelson (1969) and Merton (1969) find no relationship between age and risk-taking. This is hardly surprising, since the models, like most of them in continuous time finance, employ HARA utility functions. This choice is hardly innocuous: If the risk-free rate is zero, myopic investment strategies are rational if and only if the utility function is HARA (Mossin 1968b). Given positive risk-free rates, empirically the normal case, only constant relative risk aversion (CRRA) utility functions allow for myopia. A myopic investor bases each period's decision on that period's initial wealth and investment opportunities and maximizes the expected utility of wealth at the end of the period. The value function on wealth exhibits the same risk aversion as the utility function on consumption; the future is disregarded. This chapter is devoted to the characterization of
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_156.html [4/21/2007 1:31:20 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_157.html
Page 157 optimal dynamic strategies when the utility function is not HARA. Most of the results presented in this chapter are from Gollier and Zeckhauser (1998).
11.1 Static versus Dynamic Optimization How does the possibility to take risk tomorrow affect the optimal exposure to today's risk? This problem is close to the one that we raised in chapter 10: How does the possibility to take risk today affect the optimal exposure to independent risk that is also available today? The difference lies on the timing of the decisions. The problem that we examined in chapter 10 was static: the decisions on and had to be taken simultaneously. We now consider the case where the decision on can be taken after having observed the realization of . Because we assume all along this chapter that and are independent, the benefit of this alternative assumption does not arise from a better knowledge of the distribution of . Rather, the benefit comes from enlarging the set of available strategies for the investor. Indeed, the investor can here adapt αx to losses or benefits that he incurred on . For example, under DARA the investor will reduce his exposure to risk if he incurred a loss on , and he will increase it otherwise. There is thus no doubt that this increased flexibility raises the lifetime expected utility of the investor. Indeed, the investor can at least duplicate his static strategy in this dynamic setting. What is more difficult to examine is the comparative statics effect. We consider a simple model with no intermediary consumption: the investor is willing to maximize the expected utility of his final wealth, which is the initial wealth plus the accumulated benefits and losses on the subsequent risks undertaken. Investors choose their portfolio to maximize the expected utility of the wealth that they accumulated at retirement. There are three dates t = 0, 1, 2. Young investors invest at dates t = 0 and 1, and they retire at date t = 2. Old investors invest at date t = 0, and they retire at t = 1. Each investor is endowed with wealth w0 at date 0, and has utility function u, which is assumed to be twice differentiable, increasing and concave. There is no serial correlation on the return of risky assets. The problem of young investors is solved by backward induction. Namely one first looks at the investment problem at date t = 1, which is a static one. One does it for any potential wealth level z that is attained at that date. One can thus define the value function v(z)
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_157.html [4/21/2007 1:31:21 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_158.html
Page 158 as the maximal expected utility that can be obtained on financial markets with this wealth level at date 1. The problem of the young at date t = 0 is then to find the investment strategy at the date that maximizes the expectation of the value v of wealth z generated over the first period. This is again a static investment problem, but with the utility function u that is replaced by the value function v. Thus the impact of time horizon on the optimal investment today can be measured by the change in the degree of concavity between u and v. If v is less concave than u, we will say that duration enhances risk.
11.2 The Standard Portfolio Problem. 11.2.1 The Model In this model the active investors at date t = 0 or 1 have the opportunity to invest in a risk free asset with a zero return and in a risky asset whose return is distributed as . Again, we assume that and are independent. The problem of the investors that are still active at date t = 1 with an accumulated wealth z is written as
The problem of young investors at date t = 0 is thus written as
whereas the problem of older investors who retire at date t = 1 is to maximize . Using proposition 7, we know that young investors will invest more (resp. less) in the risky asset than older ones, whatever w0 and , if and only if v is less (resp. more) concave than u. The existence of the second-period investment opportunity exerts two effects on first-period risk taking, which we label the flexibility and background risk effects. The background risk effect is similar to the one that we examined in the previous chapter. But an important difference appears in this dynamic framework. That is, the investor is flexible on how much of risk he will accept. More specifically, the investor will adjust optimal risk exposure in the second period to the outcome in the first. With decreasing absolute risk aversion, for example, the better the first-period outcome the more of the risky asset is purchased in the second period. The opportunity to adjust one's portfolio is an advantage; this flexibility
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_158.html [4/21/2007 1:31:21 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_159.html
Page 159 effect always reduces aversion to current risks. The relationship between the optimal exposure at date 1 and the accumulated wealth z can be evaluated by solving the first-order condition to problem (11.1) for every z:
Let us assume that u is twice differentiable. Fully differentiating this condition yields
As stated in proposition 8, α(z) is increasing, constant or decreasing depending upon whether absolute risk aversion is decreasing, constant or increasing. The envelope theorem yields
Fully differentiating again, this equality yields
where α = α(z). Combining conditions (11.3), (11.4), and (11.5) allows us to write
Our aim now is to compare –v″(z)/v′(z) to –u″(z)/u′(z). Our two effects emerge from condition (11.6). The flexibility effect is expressed by the second term in the right-hand side of (11.6), which is negative. The background risk effect corresponds to the first term in the right-hand side of (11.6). Future risk can be interpreted as a background risk with respect to the independent current risk. If α were fixed and independent of the realization of the first-period risk (i.e., dα/dz = 0), the degree of risk aversion of the young investor would . Notice that we know by proposition 33 that this is larger (resp. smaller) than –u equal ″(z)/u′(z) if and only if absolute risk aversion is convex (resp. concave). In consequence the two effects are in accordance with the hypothesis that duration
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_159.html [4/21/2007 1:31:22 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_160.html
Page 160 enhances risk if absolute risk aversion is concave. The concavity of A is sufficient, but not necessary, for this result. 37 Consider the two-period investment problem with a zero yield risk-free asset and another risky asset. Young investors are less risk-averse than old investors if absolute risk aversion is concave. PROPOSITION
11.2.2 The HARA Case The model above is not tractable to extract the solution to the problem of the young investor, expect in the HARA case with
We know that absolute risk aversion is convex for HARA utility functions. Proposition 37 is not helpful to solve the problem. But using results derived in section (4.3), we know that the optimal second period strategy α(z) is such that
with
It implies that
where
We conclude that in the HARA case the value function v represents the same attitude towards risk than the file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_160.html (1 of 2) [4/21/2007 1:31:22 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_160.html
original utility function u, since the first is a linear transformation of the second. Duration has no effect on risktaking in this case. An interpretation of this result is that the flexibility effect just compensate the background risk effect that goes the opposite direction. In this case we say that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_160.html (2 of 2) [4/21/2007 1:31:22 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_161.html
Page 161 myopia is optimal. An agent is myopic if he does as if his current investment period be the last one, at each period. The result that myopia is optimal in the HARA case can be found, for example, in Mossin (1968b) and Merton (1969). 11.2.3 A Sufficient Condition for Younger People to Be More Risk-Averse We now derive a sufficient condition for a longer horizon to reduce the demand for risky assets. Consider any . We have to determine the specific z. Without loss of generality, assume α(z) = 1, that is, conditions under which
for all z and satisfying . Let F denote the cumulative distribution function of also define as the random variable with cumulative distribution function G, with
. Let us
We verify that dG is positive under risk aversion, and that ∫ dG(x) = 1. Using this change of variable, the problem is rewritten as
where T(z) = –u′(z)/u″(z) and Tv(z) = –v′(z)/v″(z) are the indexes of absolute risk aversion of the old and the young respectively. The condition above is rewritten as
Observe that this condition cannot be analyzed using the diffidence theorem developed in chapter 6. This is because the right-hand condition in (11.8) cannot be expressed as a constraint on the sign of a linear function of the probabilities of . Thus more effort is required to characterize the solution. Our main result is the consequence of the following four lemmas. The first lemma has an interest in its own. file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_161.html [4/21/2007 1:31:23 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_162.html
Page 162 4 Condition (11.8) holds for any is concave. LEMMA
Proof Let us assume that
with a two-point support if and only if the absolute risk tolerance of u
is distributed as (x–, p; x+, 1 – p), with x– < 0 < x+. We use the following notation:
According to condition (11.8), we have to verify that
is nonnegative whenever
Eliminating p from (11.10) by using equation (11.11) and reorganizing the expression yields
Since the fraction in the right-hand side of this expression is negative, Γ is nonnegative if and only if
is nonpositive. This is equivalent to require that
where λ = –x–/(x+ – x–), y+ = z + x+, and y– = z + x–. Since this must be true for any λ, y+, and y–, the necessary and sufficient condition is that T be a concave function.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_162.html (1 of 2) [4/21/2007 1:31:23 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_162.html
When returns follow a binary process with just one "up" state and one "down" state, the concavity of absolute risk tolerance is enough to guarantee that younger people should invest less in stocks. Reciprocally, the convexity of T is enough to guarantee that younger people invest more in stocks, in the binary case. However, the assumption that is binary is not satisfactory, and we want to obtain a condition that does not rely on any limitation on the distribution of returns. In the next lemma we show that just looking at binary dis-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_162.html (2 of 2) [4/21/2007 1:31:23 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_163.html
Page 163 tributions is not enough to guarantee that the result presented above holds for any distribution. We must go one step further. LEMMA
5 Condition (11.8) holds for any
if it holds for any
with a three-point support.
Proof The structure of this lemma is the same as for lemma 1. We look for the characteristics of the which is the most likely to violate condition (11.8). To do so, we solve the following problem for any scalar λ:
If the solution of this problem for any λ is positive, condition (11.8) would be proved. Observe that the problem above is a standard linear programming problem. Therefore the solution contains no more than three x that are such that dG(x) > 0, since they are three equality constraints in the program. Thus condition (11.8) would hold for any if it holds for any with three atoms. In the next lemma, we show that any three-atom random variable expressed by a compound of two two-atom random variables LEMMA
6 To any three-atom random variable
random variables,
and
can be
, each satisfying condition
satisfying condition
, and a scalar λ ∈ [0, 1] such that
satisfying condition
.
, there exist two two-atom , i = 1, 2, and
is distributed as
. The proof of this lemma is left to the reader. LEMMA
7
is a convex function of the vector of probabilities of : ∀λ ∈ [0, 1]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_163.html (1 of 2) [4/21/2007 1:31:24 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_163.html
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_163.html (2 of 2) [4/21/2007 1:31:24 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_164.html
Page 164 Proof The proof of this claim is obtained by easy manipulations of (11.16). Denoting
and
, condition (11.16) is equivalent to
This is in turn equivalent to
which is always true. We can now gather these different pieces together. Suppose that T is concave. We want to prove that this is sufficient to guarantee that a longer horizon reduces risk taking, independent of the distribution of returns. From lemma 5 we know that it is enough to check condition (11.8) for all random variable with three atoms. Take any such random variable , with . From lemma 6 we know that there exist two binary random , such that is with some probability λ, and variables, and , with – λ. Because we assume that T is concave, lemma 4 implies that
with probability 1
for i = 1, 2. This implies that
The expectation operator is linear in probabilities. Combining conditions (11.16) and (11.18) yields
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_164.html (1 of 2) [4/21/2007 1:31:24 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_164.html
Since proof of the following proposition.
, this inequality is equivalent to condition (11.8). This concludes the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_164.html (2 of 2) [4/21/2007 1:31:24 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_165.html
Page 165 38 Consider the two-period investment problem with a zero yield risk-free asset and another risky asset. Young investors are more risk-averse than old investors if the absolute risk tolerance of final wealth is concave. PROPOSITION
The opposite is not true. This is confirmed by the following counterexample. Take T(w) = w–2, which is convex, and z = 1. Let be distributed as (–0.5, 0.10659; 1, 0.83266; 10, 0.06075) It is easily verified that , but
Condition (11.8) is satisfied for this three-point distribution, although absolute risk tolerance is convex. To sum up, the concavity of absolute risk tolerance is sufficient for younger people to purchase less of the risky asset. The age of the investor does not influence the optimal portfolio composition when absolute risk tolerance is linear. But the convexity of absolute risk tolerance does not imply that younger people purchase more of the risky asset. The propositions above provide qualitative results. We would like to know the quantitative magnitude of the duration or age effect on risk taking, and could determine that if we knew the first four derivatives of the utility function. For now, consider an illustration for the case of u(z) = z + lnz, with . The absolute risk tolerance of this utility function is convex. From lemma 4 we know that a longer time horizon increases the absolute risk tolerance towards immediate risks. After some tedious computations, we get T(5) = 30.0, whereas Tv (5) = 64.1: the young investor is more than twice as risk-tolerant than the old investor. This implies that if the expected excess return of the risky asset is small, the young will invest twice as much in it as will the old.
11.3 Discussion of the Results 11.3.1 Nonlinear Risk Tolerance We first link the convexity/concavity of absolute risk tolerance T to the convexity/concavity of absolute risk aversion A. Convex absolute risk tolerance is equivalent to
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_165.html [4/21/2007 1:31:25 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_166.html
Page 166
If absolute risk aversion is concave, absolute risk tolerance is automatically convex. But under the familiar condition of decreasing A, the concavity of A is not plausible, since a function cannot be positive, decreasing and concave everywhere. Condition (11.20) means in fact that absolute risk aversion may not be too convex to get a convex risk tolerance. On the contrary, a necessary condition for the concavity of absolute risk tolerance is that absolute risk aversion be convex, a condition that we already encountered. In fact there is a clear link between the concavity of absolute risk tolerance of u and standardness, as shown by the following result, which is originally due to Hennessy and Lapan (1998): PROPOSITION
39 If absolute risk tolerance is increasing and concave, then risk aversion is standard.
Proof T increasing means that A′ is negative and P is positive, whereas T concave means that A″ ≥ 2A′2/A. We also have that absolute prudence is decreasing if and only if
We conclude that absolute prudence is necessarily decreasing if T is increasing and concave. We learn from this proposition that the investor must be standard, proper, and risk vulnerable if the riskiness of his portfolio is increasing with age. Notice that standardness is not necessary for the concavity of T. This means that there is no obvious argument favoring the concavity or the convexity of absolute risk tolerance. 11.3.2 Nondifferentiable Marginal Utility Throughout this analysis we have assumed that utility was twice differentiable. Absent this property, absolute risk tolerance may not be defined correctly and the above theory is no longer relevant. We now show an example of a utility function that is not twice differentiable and that yields the property that young people are less riskaverse than older ones. Take
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_166.html [4/21/2007 1:31:26 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_167.html
Page 167
Figure 11.1 Value function with a piecewise utility function
with t ∈ [0, 1] and any scalar D. This function is continuous, piecewise linear and concave. It is drawn in figure 11.1. Consider the standard portfolio problem with no intermediate consumption, ρ = 1 together with and 0.5(x– + x+) > 0. It can be shown that a bounded solution exists for this problem with utility function (11.22) if and only if
Under this condition the optimal demand for the risky asset equals
The value function is then written as
whose graph is depicted in figure 11.1. file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_167.html (1 of 2) [4/21/2007 1:31:26 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_167.html
The value function is obviously less concave than the utility function in the usual sense: v is a convex transformation of u. Young investors purchase more of the risky asset than old investors. However, the absolute risk tolerance is not convex. Notice also that v is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_167.html (2 of 2) [4/21/2007 1:31:26 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_168.html
Page 168 piecewise linear as is u, so the argument can be reproduced recursively over more than two periods.
11.4 Background Risk and Time Horizon Up to now, investors confronted only portfolio risk. In the real world, investors also face risks on their human capital. Will they get or lose their job? Will they become disabled? How do such risks affect the influence of horizon length on the riskiness of savings portfolios? In the first section, we consider that the human capital risk occurs at retirement. It might be whether the individual is relatively healthy or not, with implications for ability to earn in retirement or medical expenses, which get deducted from wealth prior to consumption. In the second section, we look directly at lifetime labor market risk. It treats labor income as a random walk. In both sections we assume that all consumption is undertaken at the end and that the risk-free rate is zero. 11.4.1 Investors Bear a Background Risk at Retirement We return to our basic two-period model, but now confront the investor with a background risk when he is old. No such risk is borne when young. We know that this risk will have an adverse effect on the demand for stock from old investors under risk vulnerability.2 We want to determine how this risk influences the relationship between time horizon and portfolio risk. The answer to this question, as we know, depends basically on the concavity or convexity of the absolute risk tolerance of the following interim indirect utility function:
The agent with a utility function u and facing risk at retirement will have the same optimal dynamic portfolio strategy as agents having a utility function h and no such risk. The absolute risk tolerance of h equals
evaluated at z. Differentiating with respect to z yields
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_168.html [4/21/2007 1:31:27 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_169.html
Page 169 This implies that Th is concave if
is nonnegative. We are not aware of any conditions sufficient for these properties to hold other than the ones presented in the next proposition. PROPOSITION
40 Suppose that the utility function u is HARA and that the background risk
interim indirect utility function time horizons reduce the demand for stocks.
is small. Then, the
has a concave absolute risk tolerance. This implies that longer
Proof We have to prove that K is nonnegative. Without loss of generality we suppose that the expectation of is zero. Let be distributed as , with and . Then a second-order Taylor expansion around z yields
It implies that
or equivalently,
It is easy to verify that a0 = 0 (this is due to the linearity of T) and
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_169.html (1 of 2) [4/21/2007 1:31:27 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_169.html
where u(n) = u(n)(z). Following the assumption that preferences are HARA, suppose that u is defined by condition (23.5). Let x denote a + (z/γ). It implies that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_169.html (2 of 2) [4/21/2007 1:31:27 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_170.html
Page 170
Figure 11.2 ~ (–1, 1/2; 1, 1/2)
After some tedious manipulations, we obtain
We conclude that K(z) is positive if k is small enough. This result should be compared to the one obtained without background risk where myopia is optimal because of the HARA assumption. The presence of a small idiosyncratic risk at retirement tends to bias the effect of a longer time horizon in favor of more conservative portfolios. However, this result may not hold for background risks that are not small. This is shown by the following example: take a CRRA utility function with γ = 4. In figures 11.2 and 11.3 we have drawn when background risk is distributed as (–k, 1/2; k, 1/2). When k = 1, is uniformly decreasing, implying that the indirect utility function has a concave absolute risk tolerance. But for the larger background risk with k = 5, it happens that the absolute risk tolerance is first convex and then concave. This second graph thus provides a counterexample to the extension of the proposition above to large background risks. It should also be noticed that the effect of background risk on the relation between time horizon and portfolio risk is at best a second order effect under HARA. The following numerical simulation is an illustration of this fact. As before, let us consider a CRRA utility function with a relative risk aversion γ = 4. Suppose that the back-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_170.html [4/21/2007 1:31:28 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_171.html
Page 171
Figure 11.3 when
ground risk at retirement is and that the return on stocks at each period is . Take a wealth level z = 5. If there is only one period to go, the optimal investment in the stock is α = 0.251693. It goes down when more time remains before retirement. However, the effect is hardly observable, since the optimal demand is α = 0.251477 with six periods to go. 11.4.2 Stationary Income Process In the real world the risk on human capital takes the form of recurrent shocks on labor incomes. The simplest way to model this is to assume that at each period t, investors hold a lottery ticket that is actuarially fair. are i.i.d. and are independent of the returns of capital investments. We see that an Assume also that additional element comes into the picture now: younger people bear a larger risk on their human capital than older agents who own fewer lottery tickets. The general idea is that if utility is vulnerable to risk, the larger risk on human capital increases aversion to other independent risks, this will inhibit risk taking while young. It is difficult to extract the qualitative effect of time horizon on portfolio risk when a recurrent background risk is present. This remains to be done. We hereafter present a numerical estimation of this model. We took a CRRA utility function with γ = 2. At each period the investor faces a random shock on his labor income, which
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_171.html [4/21/2007 1:31:28 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_172.html
Page 172
Figure 11.4 Optimal demand for stocks as a function of wealth given shocks on labor income
takes the form . The per period rate of return on stocks in . In figure 11.4 we show the optimal demand αt(z) for stocks as a function of accumulated wealth for different time horizon. The main feature of this figure is that the demand for stocks goes down sharply as retirement is more distant, at least for low wealth levels. Risk vulnerability is the main reason for this effect.
11.5 Final Remark. This chapter has been devoted to the characterization of efficient dynamic portfolio strategies. We answered the following question: How does the presence of an option to reinvest in a portfolio in the future affect the optimal portfolio today? By solving this problem, we have been able to adapt the static portfolio problem to its intrinsically intertemporal nature. Our conclusions are mixed. The answer to the above-mentioned question basically depends on whether absolute risk tolerance is concave or convex. In the absence of background risk, we can summarize the solution as follows: 1. If absolute risk tolerance is linear (HARA case), the time horizon has no effect on the optimal portfolio. 2. If absolute risk tolerance is concave, a longer time horizon induces investors to reduce their demand for stocks. 3. If absolute risk tolerance is convex, we need more assumptions to conclude. In particular, if the portfolio risk is binary, a longer time
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_172.html [4/21/2007 1:31:29 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_173.html
Page 173 horizon increases the demand for stock. The same result is obtained by replacing the binary assumption by the assumption that absolute risk aversion is concave, a condition that is stronger than the convexity of absolute risk tolerance. There is a direct application of this result for the equity premium puzzle. In chapter 5 we calculated the equity premium based on the assumption that investors live just for one year. Under HARA this assumption has no effect on the resulting equity premium. But, if absolute risk tolerance is concave, taking into account the longevity of investors will make them more risk-averse than assumed in the static one-year model, thereby it will increase the equity premium. These results do not take into account the fact that younger people usually bear a larger risk on their human capital. Under risk vulnerability this is a strong motive for younger investors to purchase a more conservative portfolio.
11.6 Exercises and Extensions Exercise 62 Observe that is the upper envelope of a set of concave function. It is therefore not obvious a priori that v is concave. Show that v is concave, meaning that young investors who are risk-averse about their final consumption never behave in a risk-loving way. This can be shown for example by proving that . Exercise 63 (Roy and Wagenvoort 1996) The objective of this exercise is to examine whether the young investor increases his demand for the risky asset when he becomes wealtier if u is DARA. In other words, we examine . Prove that whether DARA is preserved in the portfolio problem. Define again •
for n = 1, 2, and 3, where f(n) is the nth derivative of f.
• v is DARA if u is DARA and α′ ∈ [0, 1]. To do this, use the previous results, together with the Cauchy-Schwarz inequality (6.33). Roy and Wagenvoort (1996) also provide a counterexample where u is DARA but v is not DARA. Their example is such that α′ > 1. Exercise 64 To extend the model to more than two periods, we need to determine whether the concavity/ convexity of the absolute
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_173.html [4/21/2007 1:31:29 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_174.html
Page 174 risk tolerance of u is transferred to the concavity/convexity of the absolute risk tolerance of v. Show that this property is not guaranteed in the general case. Exercise 65 In this chapter we assumed that the investor knew in advance the date at which she will have to consume her portfolio wealth. Suppose, alternatively, that the investor has a probability p to consume her accumulated portfolio wealth at date 1, and a probability 1 – p to consume her accumulated portfolio wealth at date 2. The utility level at the date without consumption is normalized to zero. Examine the effect of a change in the probability p on the optimal portfolio strategy at date 0.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_174.html [4/21/2007 1:31:30 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_175.html
Page 175
12 Special Topics in Dynamic Finance The standard portfolio model that we examined in the previous chapters does not take into account the presence of transaction costs and of constraints on financial strategies available to investors. The effects of these imperfections are in general trivial when considering a static model. They are generally not tractable in dynamic models. The large literature in finance on these questions is among the most complex fields in economic theory. Our aim in this chapter is to explore the effect of these imperfections on the optimal dynamic portfolio strategies. This chapter can be skipped for those readers that are not directly interested by these topics. We first examine the effect of an extreme case of transaction costs in section 12.1. We assume that there is just no way to rebalance one's portfolio through time. In the other sections we focus the analysis on various quantitative restrictions of the size of the future risk exposure. We first examine the case where the size of the risk is completely fixed, meaning when the risk decision is a take-it-or-leave-it offer. We then turn to the cases of an upper or lower bound on the risky investment.
12.1 The Length of Periods between Trade Most dynamic strategies entail a rebalacement of the portfolio composition at each date, almost surely. Even with a CRRA utility function where the investor follows the simple strategy to maintain constant the relative share of the portfolio value that is invested in the risky asset, the investor must trade whenever the excess return of the risky asset differs from zero. He must sell some of it if the excess return is positive, whereas he must buy some of it if the excess return has been negative. We suppose here that there is no dividend. The
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_175.html [4/21/2007 1:31:30 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_176.html
Page 176 return of an asset comes from an increase in its value. Thus we should observe that households are active on financial markets at every instant where relative prices of different securities diverge. That is, in fact, at every instant. This is obviously not the kind of investment behavior that we observe in the real world. One reason is that contrary to what we assumed, transaction costs prevail on financial markets. Investors must pay transaction costs every time that they want to rebalance their portfolio. The most obvious illustration of such costs is the bidask spread, which is the fact that the price at which one can sell an asset is smaller than the price at which one can purchase it. One should also take into account of the time or cost spent to get financial information, to learn prices, and to submit an order to the market. Sometimes distorting taxes prevail that limit the possibility to trade frequently. Most of these costs have a fixed component and a variable component. The presence of transaction costs will induce trading only if the benefit to trade exceeds its cost. The effect of transaction costs on the static optimal portfolio strategy is in general trivial. If there is a fixed cost to invest in risky assets, it may be optimal to be 100% in the risk free asset, although . The variable component of transaction costs may be analyzed as a leftward shift in the distribution of the return of the risky asset. In an intertemporal framework, the effect of a transaction cost is less obvious to tackle. In short, knowing that one will have to pay cost to rebalance one's portfolio in the future may have an impact on the optimal structure of the portfolio today. In this section we consider extreme case where there is an infinite cost to rebalance portfolios in the future. A good motivation for examining the infinite case is that it answer the following question: What is the impact of the length of periods between trade on the optimal portfolio risk? To be more precise, we consider the model with three dates and two assets that we examined in the previous chapter. In this analysis we look more generally at the relationship between flexibility—we label its opposite rigidity— and the risk of portfolios optimized to maximize expected utility. Common wisdom suggests that flexibility enhances risk. The strong form of rigidity arises when the investor is not able to modify his exposure to risk from period to period. We hereafter examine a somewhat weaker form of rigidity, whereby the investor can modify his risk exposure from one period to another, but he is forced to fix it a priori, which is before observing
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_176.html [4/21/2007 1:31:31 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_177.html
Page 177 realizations of the early returns on his investment. The individual invests at date 0, receives investment returns, and then must invest at date 1. In the rigid economy, date 1 investment decisions must be made before date 0 results are learned. The problem is written as follows:
where is the return of the risky asset between t and t + 1. As before, we assume that the returns of the risky asset follows a random walk, namely that and are independent. Observe that the decision β on the exposure on must be taken simultaneously with α. The optimal exposure to risk is denoted αr, where the upper index refers to the rigid economy. This problem was examined in section 10.2. We want to compare αr to the optimal investment in when the investors live in the flexible economy, which is one where he can determine β after the observation of . The investment problem in the flexible economy is written as
This problem is the one that we already examined in section 11.2. Intuition suggests that the opportunity to rebalance one's portfolio in the future contingent to the realization of the first-period portfolio should induce more risk-taking: αf ≥ αr. It may be useful to compare αr and αf to the solution of the myopic problem:
From proposition 36 we know that αr is less than αm if absolute prudence is decreasing and larger than twice the absolute risk aversion. Under these conditions, is substitute to . If the utility function is HARA, this is equivalent to the condition that γ be less than 1. Moreover we know that for the class of HARA functions, αm equals αf: myopia is optimal in that case. It yields the following result, which is due to Gollier, Lindsey, and Zeckhauser (1997). 41 Suppose that the utility function is HARA, with γ ≤ 1. Then αf is larger than αr; that is, flexibility enhances risk. PROPOSITION
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_177.html [4/21/2007 1:31:32 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_178.html
Page 178 We can extend this proposition by using sufficient conditions for αm being smaller than αf. These sufficient conditions are in proposition 37. 42 Suppose that absolute prudence is decreasing and larger than twice the absolute risk aversion. Suppose also that absolute risk aversion is concave. Then αf is larger than αr, that is, flexibility enhances risk. PROPOSITION
The sufficient conditions presented in this proposition are in fact sufficient for the stronger property that αr ≤ αm ≤ αf. We see that there are two effects of transforming a rigid economy into a more flexible one. These effects are obtained by decomposing the change via the single-period, myopic economy. Going from the rigid economy to the myopic one introduce a substitution effect. It implicitly eliminates the opportunity to invest in asset , which increases the demand for asset if they are substitutes. The second effect is obtained by moving now from the single-period economy to the flexible one, which raises the duration of the investment. This has a positive effect on risk-taking if duration enhances risk. We now go back to the initial motivation of this section, which was the analysis of the effect of the length of periods between trade on the optimal portfolio composition. If we assume that and are identically distributed, then by proposition 5, we get that
This is due to the fact that the optimal values of α and β coincide when the two risk are identically distributed. The problem to the right is an investment problem where the investor may not rebalance his portfolio at date 1. Its solution αr is thus interpreted now as the optimal investment in the risky asset when the period between trade can take place is between t = 0 and t = 2. Financial markets are closed at date t = 1. Comparing αr to αf is the relevant question to determine the effect of the length of periods between trade. The following result is a straightforward reinterpretation of the result above: 4 Suppose that the return of the risky asset follows a stationary random walk. Under the conditions of proposition 42, reducing the length of time between trade raises the demand for the risky asset. COROLLARY
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_178.html [4/21/2007 1:31:32 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_179.html
Page 179
12.2 Dynamic Discrete Choice In this section we examine the discrete choice version of the model presented in section 11.2. We analyze the effect of an option to gamble on a lottery with payoff in the future on the decision to gamble today on a lottery with payoff . The important restriction that is imposed here is that the investor can either accept or reject the lottery. This is a take-it-or leave-it problem. Contrary to the model presented in section 11.2 where the agent could determine the size of his exposure to risk, the size of the risk is exogenous in this section. An illustration of this problem is when an entrepreneur faces a sequence of real investment projects. It is often the case that the entrepreneur has no choice about the size of the investment; he can just do it or not, because of the presence of large fixed costs to implement the project. The attitude toward the current lottery is characterized by the degree of concavity of the value function v defined as
where K ≡ {0, 1} is the set of acceptable values for the decision variable. . Bell The value function is the upper envelope of two concave functions of z, respectively u(z) and (1988), in considering the same specific model, showed that v is not concave, except in the trivial case in which is never (or always) desirable. We reproduce here his figure (figure 12.1) in the specific case of decreasing absolute
Figure 12.1 Value function: Upper envelope of u(.) and
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_179.html [4/21/2007 1:31:33 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_180.html
Page 180 risk aversion. Use condition (2.11) with π = 0 to prove that under DARA, curve u(z) crosses above. Function v is locally convex at z0 where u(.) crosses
from
. The intuition of this result comes from the
optionlike payoff that is generated when the gambler gets the opportunity to bet on v as
. This is shown by rewriting
where CE is the certainty equivalent of , meaning that . The nonconcavity of v implies that the gambler could rationally accept unfair lotteries at date 0. As it is intuitively sensible, "an entrepreneur with an idea she believes will work, but without financial backers who agree, would be justified trying to raise the necessary capital in Atlantic City" (Bell 1988). We conclude from this discussion that v may never be more concave than u in all cases where the choice on switch from 0 to 1 depending on the wealth z of the agent. Is it possible that v be globally less risk-averse than u? The answer is no, at least for proper utility functions. The existence of an option to gamble on may induce gamblers to reject some lotteries that would have been accepted in the absence of an option. 43 Suppose that is desirable at some levels of wealth and undesirable at other levels. If risk aversion is proper, the indirect utility function v(.) defined by equation (12.1) is not a convex transformation of u. That is, there exist some lotteries that are accepted in the absence of an option to gamble on , but that are rejected if such an option is offered.
PROPOSITION
. Because DARA is necessary for properness, properness implies that there Proof Define exists a unique z0 such that m(z0) = u(z0). We first show that –m″(z0)/m′(z0) is larger than –u″(z0)/u′(z0). Because u is proper and
, we get that for any ,
or equivalently,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_180.html [4/21/2007 1:31:33 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_181.html
Page 181 At z0, m rejects all lotteries that u rejects: m is more diffident than u around z0. A necessary condition is that –m″ (z0)/m′(z0) be larger than –u″(z0)/u′(z0). By definition of z0, m and u must cross at z0. From definition (2.11) of DARA, m crosses u upward. Thus v is locally more concave than u to the right of z0. In order to illustrate proposition 43, suppose that the gambler with initial wealth w0 = 2.1 and utility function u (w) = In(w) has the opportunity to bet on . One can easily verify that it is optimal for him to bet. Suppose now that in addition to the opportunity to bet on today, the gambler knows that he . In that case it is not optimal anymore to bet on . The option to could bet tomorrow on bet in the future makes some current desirable risks undesirable. When u is proper, the absolute risk aversion index for v is decreasing everywhere except at z1 where an upward jump occurs. Thus the value function defined by (12.1) does not inherit the property of DARA from the utility function u. This jump in risk aversion at the wealth level where the option value becomes positive may yield counterintuitive comparative statics effects. To illustrate, let us consider again the logarithmic gambler facing the and the option to bet on . It is easy to current lottery verify that z0 = 2. It is noteworthy that there is more than one critical wealth level in the first period at which the gambler switches from rejecting to accepting lottery . One can verify that the gambler will not bet on if his wealth at the start of the game is less 1.529 or if it belongs to interval [2.094,2.282]. As wealth increases, the gambler switches thrice by first rejecting, then accepting, then rejecting again, and finally accepting the initial bet despite DARA. These paradoxes disappear if one assumes that and are identically distributed as . In that case the opportunity to gamble again on the same risk in the future always induce more risk-taking today. This is seen by proving that
Consider an agent with wealth w at t = 0. The condition to the left means that if he does not accept to gamble today, it is optimal to gamble tomorrow. The condition to the right means that it is optimal to gamble at t = 0. To prove this property, observe that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_181.html [4/21/2007 1:31:34 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_182.html
Page 182
where Et is the expectation with respect to
. Obviously this implies that
The left condition in (12.3) implies that . Thus the inequality . This result is easily extended for a larger sequence of i.i.d. risks. This above implies that proves the following proposition: 44 Consider a sequence of take-it-or-leave-it offers to gamble on i.i.d. lotteries. It is never optimal to postpone the acceptance of a desirable lottery. PROPOSITION
That corresponds to the common sense that one should not "put off till tomorrow what could be done today." The intuition of this result comes from the fact that postponing the acceptance of a desirable risk does nothing else than to reduce the opportunity set of the gambler. In other words, the optimal strategy with t periods remaining can always be duplicated when there are t + 1 periods remaining. This is done by acting as if the horizon is a period shorter, and by rejecting the last lottery. This strategy is clearly weakly dominated by the strategy of doing as if the horizon be a period shorter, and by doing what is optimal to do at the last period depending upon the state of the world at that time. The consequence of this result is that once the gambler decides to reject a lottery, he will afterward never gamble again. Two scenarios are then possible: 1. The agent never gambles. 2. The agent gambles from the first period till some period s, and then never gambles again. In short, the problem simplifies to a stopping time. Let
t
be the set of wealth levels for which it is optimal to
gamble when there are t lotteries remaining. Proposition 44 states that
1
⊂
2
⊂...⊂
T:
the acceptance set is
smaller for shorter horizon problems. As the horizon recedes, the gambler is more and more willing to gamble. This gives a potential rationale to the Samuelson's (1963) lunch colleagues who refused to gamble on a single risk but who would have accepted to gamble if many such i.i.d. risks were offered. Our
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_182.html [4/21/2007 1:31:35 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_183.html
Page 183 interpretation of the game is different from the one proposed by Samuelson in the sense that we do not assume that the player is forced to determine ex ante the number of i.i.d. gambles. Notice that the intuition of our result is in no way related to the law of large numbers because, as in Samuelson (1963), i.i.d. risks are added through time rather than subdivided. Gollier (1996a) shows that acceptance sets
t
are of the form
t
= {z ≥ zt} if u is DARA. As suggested by the
intuition, wealth must be large enough to gamble. From proposition 44 it must be that z0 ≥ z1 ≥ z2 ≥ . . . : the larger the number of option remaining, the smaller is the minimum wealth level at which one starts to accept to gamble. As in the standard static model, a DARA gambler accepts to gamble only if his wealth at the time of the decision is large enough. There is a critical wealth level at every period separating gambling states from nogambling states. But the existence of one or more options to bet in the future implies that the critical wealth level is reduced. The positive effect of the existence of options implies the counterintuitive property that a DARA gambler could well decide to stop gambling after the occurrence of a net gain on the previous lottery. The point is that the reduction in risk aversion due to the increase in wealth is not sufficient to compensate for the negative effect of the reduction of the opportunity set. To illustrate this result, let us consider the case of the logarithmic utility function with . The critical, minimal wealth level when there is no option to gamble again is z0 = 2.000. When there is a single option to gamble remaining, this critical wealth level goes down, z1 = 1.812. It is obtained by solving . We also obtain z2 = 1.702, z3 = 1.634, . . . , z17 = 1.316.
12.3 Constraints on Feasible Strategies The surprising result obtained in the previous section is that a risk-averse investor may behave in a risk-loving way when he faces a constraint on the size of his risk exposure in the future. This constraint was very specific, since the size of his exposure was completely exogenous when deciding to take risk: K = {0, 1}. Suppose, more generally, that the demand for the risky asset at date 1 must be in some subset K(z) of R, where z is the wealth level of the investor at that date. If K(z) is the set of integers, this constraint means that the investor may not purchase fractions of the risky asset. If K(z) is the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_183.html [4/21/2007 1:31:35 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_184.html
Page 184 set of real numbers less than z, this constraint is a no–short-sale constraint stating that the investor cannot borrow at the risk-free rate to invest in the risky asset. For the time being, we are just willing to determine whether the value function
is concave. We have shown in the previous chapter that it was the case when α is not constrained, that is, when K ≡ R. On the contrary, when K = {0, 1} as in the previous section, it was not the case. We provide now a sufficient condition for the value function defined by (12.4) to be concave. Suppose that K(z) is convex:
Following Pratt (1987), let us show that this condition together with the concavity of u imply that the value function is concave. Let us take any (z1, z2) with z1 < z2, and any λ ∈ [0, 1]. By definition, we have that
with
By the concavity of u, applying Jensen's inequality to the right-hand side of condition (12.6) for every realization of yields
Since α1 ∈ K(z1) and α2 ∈ K(z2), the convexity of K implies that λα1 + (1 – λ)α2 is in K(λz1 + (1 – λ)z2). This implies that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_184.html (1 of 2) [4/21/2007 1:31:36 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_184.html
The right-hand side of the inequality above is nothing but vK(λz1 + (1 – λ)z2). Combining conditions (12.6), (12.7), and (12.8) yields the result:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_184.html (2 of 2) [4/21/2007 1:31:36 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_185.html
Page 185
This means that vK is concave. 45 The value function (property 12.5).
PROPOSITION
is concave if u is concave and K is convex
This result is a generalization of a result by Pratt (1987). Observe that in the take-it-or-leave-it model with K(z) = {0, 1}, the condition that K be convex is not satisfied, and the proposition does not apply.
12.4 The Effect of a Leverage Constraint. In the preceding section we examined whether the value function is concave. We now turn to the comparison of the degrees of concavity of u and v. We consider two types of leverage constraints: α ≥ k(z) and α ≤ k(z). 12.4.1 The Case of a Lower Bound on the Investment in the Risky Asset Consider the case with
for some real-valued continuous function k(.). There is a minimum requirement on the size of the investment in the risky asset. This is the case, for example, in a French plan where, by law, pension plans will be required to be invested in stocks for at least 50% of their market value. This is also the case in the co-insurance model (see section 4.6) where the insured person must retain a positive deductible on his insurance contract. Let α(z) be the unconstrained solution of the static portfolio problem with wealth z:
Consider a wealth level where the constraint just becomes binding, yielding . We examine the is smaller than : the constrained exposure to risk is less flexible to changes in wealth than case where the optimal exposure. It implies that the constraint is binding for wealth levels slightly smaller than . We obtain that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_185.html [4/21/2007 1:31:37 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_186.html
Page 186 in a neighborhood of . We want to determine the impact of the minimum requirement constraint on the optimal portfolio at date t = 0. It is useful to start with the analysis of the degree of concavity of vK around . It is the same as the degree of concavity of v—the value function of the unconstrained problem—to the right of . For values of z slightly to the left of , we obtain that
whereas the degree of concavity of the unconstrained value function equals
Recall now that DARA means that directly implies that together with
is negative. The comparison of (12.10) and (12.11)
under DARA. We conclude that DARA is necessary and sufficient for the introduction of a minimum requirement on the investment in the risky asset to raise the aversion to risk of the value function in the neighborhood of a critical wealth level where the constraint becomes binding. The fact that such a minimum requirement raises aversion to date 0 risks is a very natural property. It is compatible with the idea that independent risks are substitutes. The fact that an agent is potentially forced to take more risk in the future induces the agent to undertake less risk today. The analysis above shows that DARA is necessary for the introduction of a minimum requirement to induce and –v″(z)/v′ more risk aversion. This condition is not sufficient for this result. The comparison of (z) is not trivial when they are not evaluated at a critical wealth level. The search for a sufficient condition will be limited now to the case where k(z) is a constant k that is independent of z. The easiest way to treat this problem is again to decompose the problem into, first, comparing vK to u and, second, comparing u to v. We know that vK is more concave than u if
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_186.html [4/21/2007 1:31:37 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_187.html
Page 187
The condition to the left means that we are in the binding region. This condition is close to condition (10.1), except that the left condition is an inequality rather than an equality. Using lemma 2 with f1(x) = xu′(z + kx) and f2 (x) = u″(z + kx)u′(z) – u′(z + kx)u″(z) yields the following set of necessary and sufficient conditions:
Thus vK is more concave than u if absolute risk aversion is decreasing and convex. We would now be done if u be (weakly) more concave than v. This is the case under HARA. 46 Suppose that u is HARA, with decreasing absolute risk aversion. Then a minimum requirement on the size of the investment in the risky asset (α ≥ k) in the future reduces the demand for the risky asset today. PROPOSITION
The restrictive condition HARA can be relaxed by using weaker sufficient conditions for u to be more concave than v. Using lemma 4, we obtain for example the following result: 47 Suppose that the risk on returns be binary. A minimum requirement on the size of the investment in the risky asset (α ≥ k) in the future reduces the demand for the risky asset today if absolute risk tolerance is convex and absolute risk aversion is decreasing and convex. PROPOSITION
From these three conditions only DARA is necessary. The two other conditions are probably too strong. The convexity of absolute risk aversion has already been found in proposition 33 as a condition for independent risks to be substitutes when taken simultaneously. 12.4.2 The Case of an Upper Bound on the Investment in the Risky Asset We now turn to the case where the investor faces an upper limit constraint α ≤ k(z) on his risky investment at date 1.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_187.html [4/21/2007 1:31:38 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_188.html
Page 188 As before, we first look at the effect of the leverage constraint on the degree of concavity of the value function around a critical wealth level . Also we examine the case where the constraint induces less sensitivity; that is, is smaller than . An exercise parallel to the one performed above would show that DARA is necessary and sufficient in that case to guarantee that vK be more concave than the unconstrained value function around . The intuition that a constraint limiting the size of the risky investment in the future should induce a reduction in the demand for the risky asset today is not obvious. Indeed, given the idea that independent risks are rather substitutes, the investor could prefer to increase his risky investment today to compensate for the anticipated limitation in the size of the risky investment tomorrow. This effect goes the opposite direction than the flexibility effect mentioned above, under DARA. We may thus expect an ambiguous global effect under assumptions compatible with the substitutability of independent risks. To show this, let us again consider the special case with k(z) = k. We decompose the global effect of the leverage constraint into a change from vK to u first and then a change from u to v. The first effect is an increase in risk aversion if
The condition to the right means that the investor with wealth z would like to invest more than k in the risky asset. One could use corollary 2 to get the necessary and sufficient condition that A be increasing and convex. It is more intuitive to consider the trick presented after proposition 33: let the cumulative distribution function of be defined by
Condition (12.12) may then be rewritten as
The ambiguity appears then clearly. The Jensen inequality combined with the assumption that A is convex is larger implies that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_188.html [4/21/2007 1:31:38 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_189.html
Page 189
Figure 12.2 Absolute risk aversion of the value function with a leverage constraint
than . This goes in the intuitive direction. But DARA plus positive goes in the opposite direction, with . Suppose, in particular, that A″(w)/A′(w) tends to zero when w tends to infinity. Then, using standard results on the symmetric concept of risk premium, the substitution equivalent ι of , defined as
which tends to zero with z. It implies that condition (12.13) is violated in that case for large values of z. Notice that conditions A′ < 0 and A″/A′ → 0 are satisfied for HARA utility functions that exhibit DARA (γ > 0). Thus, for HARA functions, the absolute risk aversion of vK is smaller than the absolute risk aversion of u when z is large. Since u and v are confounded for HARA functions, we obtain the following result: 48 Suppose that u is HARA. Then the introduction of a leverage constraint α ≤ k in the future has an ambiguous effect on the attitude toward current risks in the following sense: it raises local risk aversion around the critical wealth level where the leverage constraint becomes binding, and it reduces local risk aversion for large wealth levels. PROPOSITION
We represented in figure 12.2 the typical form of the absolute risk aversion of the value function with a leverage constraint. Grossman and Vila (1992) examined the same problem, but within a continuous-time infinite-horizon CRRA framework. They showed that a borrowing constraint make investors more (less) risk-averse if relative risk aversion is smaller (greater) than 1.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_189.html [4/21/2007 1:31:39 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_190.html
Page 190
12.5 Concluding Remarks The previous chapter characterized the efficient dynamic portfolio strategies when investors face no constraint on the way that they can rebalance their portfolio. This was a world of full flexibility. This chapter has been devoted to various limitations to flexibility. The general intuition is that a reduction of flexibility in the way the investor can rebalance his portfolio should make investors more averse to invest in risky assets. We first examined the effect on the optimal portfolio of the impossibility to rebalance one's portfolio through time. In a second part of the chapter, we looked at the effect of some limitations on the amount that can be invested in the risky asset in the future. We first showed that it may lead to a risk-loving behavior if the choice set is not convex. We then examined the effect of a leverage constraint. Our findings have been quite mixed. Restrictions on preferences are strong, and sometimes clearly unrealistic, to guarantee that less flexibility leads to less risk taking.
12.6 Exercises and Extensions Exercise 66 Consider the two-period standard portfolio problem. We want to examine the effect of the announcement that a participation fee to the stock market will be imposed in period 2 on the demand for stocks in . Of course, period 1. Let k > 0 be the fee in period 2, and define vk(z) as max {u(z), , the value function of the standard portfolio problem. We want to compare the degree of concavity of vk and v0. Show that • function vk has a convex kink at some z = z0 if k is not too large. • function vk is less concave than function v0 for all z < z0 if absolute risk tolerance is concave. • function vk is more concave than function v0 for all z > z0 if absolute risk aversion is decreasing and if dα/dz is less than unity. [Hint: Use the result derived from exercise 63 in the previous chapter.] Exercise 67 Consider the one-kink piecewise utility function u(z) = min {z, (1 – t)(z – D) + D}, and a gamble whose net payoff is distributed as (x– < 0, p; x+ > 0, 1 – p). We examine the properties of the indirect utility function
. Suppose that
. Show that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_190.html [4/21/2007 1:31:39 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_191.html
Page 191 • function right of D.
is piecewise linear with two concave kinks, one to the left of D, the other to the
• function v has five kinks, three being concave and two convex. From this, what can we say about the number of possible switches—as z increases—in the decision process toward another independent small gamble
whose outcome is announced before ?
Exercise 68 Examine the effect of a borrowing constraint α(z) ≤ k(z) imposed in period 2 on the optimal portfolio composition in period 1. Contrary to what has been done in section 12.4.2, consider the case where α′(z) is smaller than k′(z).
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_191.html [4/21/2007 1:31:40 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_192.html
Page 192
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_192.html [4/21/2007 1:31:40 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_193.html
Page 193
V The Arrow-Debreu Portfolio Problem
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_193.html [4/21/2007 1:31:41 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_194.html
Page 194
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_194.html [4/21/2007 1:31:41 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_195.html
Page 195
13 The Demand for Contingent Claims In the standard portfolio problem the final wealth of the investor is linear in the market return. This is because we assumed that there are only two assets, one being risk free and the other aggregating all risks in the economy. In the real world, people can make more sophisticated plans by purchasing specific assets whose returns are not perfectly correlated to the market. This is done, for example, by purchasing options or by negotiating contingent (insurance) contracts for specific risks with other agents on the market. It implies that people can select a risky position in a much wider class than the class of linear positions, namely a position implying a linear relationship between the investor's final wealth and the market return. In this chapter we assume that agents can select any risky position, not necessarily linear in the market return, that is compatible with their budget constraint. We will hereafter refer to x as the state of nature. Taking this view rather than the one in which x is specifically the return of the risky technology in the economy allows us to generalize the approach. The uncertainty in the economy can affect not only the return of various physical assets but also the labor income of the workers, the existence of damages to houses and cars, and so on. A state of nature is described by a specific realization of this vector of random variables. The assumption that the consumption function c can be made dependent on all these variables, means that the investor can sign a wide range of contingent contracts, as an insurance contract on her car, a bet on her neighbor's risk of accident, a call option on Microsoft, or an insurance on her future labor income. In a word, we assume that markets are complete. Market completeness is a rather extreme assumption. We do this in accordance with the theory of finance in which this assumption
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_195.html [4/21/2007 1:31:41 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_196.html
Page 196 played a central role in the progresses made in this field for the last thirty years. Introducing an idiosyncratic, uninsurable background risk as we did in chapter 9 is one way to relax the assumption that markets are complete. Another way is to impose portfolios to be linear, as in chapter 4.
13.1 The Model The uncertainty in the economy is represented by the set X of possible states of nature that could prevail at the end of the period. It is quantified by a random variable whose cumulative distribution function is objectively known. The definition of a state of the world must describe all the elements of the environment. If the only source of uncertainty is the aggregate wealth available in the economy, could be the GDP per capita, or any deterministic one-to-one function of it. Ex post, by definition, the Arrow-Debreu asset associated to state x provides one unit of the consumption good in state x and has no value otherwise. The assumption that markets are complete means that to each state of the world there is an associated Arrow-Debreu asset that is traded on financial markets. As is well known, this unrealistic assumption is equivalent to the condition that there are enough real assets on the market to build portfolios that duplicate the Arrow-Debreu assets. At the time of the portfolio decision, that is, before knowing the realization of x, Arrow-Debreu securities are traded, and equilibrium prices emerge. Let Π(x) denote the equilibrium price of the Arrow-Debreu asset associated to state x. It is also called the "state price." We take function Π(x) as given in this chapter. At the end of the period, a state x is revealed, and agents consume their income generated by the liquidation of their portfolio. The problem of a price-taking investor with wealth z is to select a portfolio of Arrow-Debreu securities that maximizes his expected utility under a budget constraint. A portfolio is represented by function c(.), where c (x) is the number of units purchased of the Arrow-Debreu security x. By definition, it is also the level of consumption in state x, since we assume that investors have no other contingent incomes than the ones generated by their portfolio. In this sense the portfolio c also describes a consumption plan. In the discrete version of the model where
has its support in a discrete set X, the problem is written as follows:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_196.html [4/21/2007 1:31:42 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_197.html
Page 197
where px is the probability of state x ∈ X, cx = c(x), πx = Π(x)/px and z is the initial wealth of the agent. Notice that we replaced the state price Πx by πx = Πx/px which is a state price density, or a price kernel. It is the price of asset x ∈ X per unit of probability. Observe that the expected gross return of asset x is the inverse of its price density:
We can now build a bridge between the model developed here and the standard portfolio problem. Let us suppose without loss of generality that Σx∈Xpxπx, which can be interpreted as the gross rate of interest, is equal to unity. If we impose that final consumption c be linear in the net return of the market π–1 – 1, we must have that . The budget constraint is satisfied under this restriction if and only if
This implies that, under the constraint that consumption must be linear in the market return, the decision problem (13.1) can be rewritten as
where . This is the standard portfolio problem. We conclude that the standard portfolio problem is equivalent to the Arrow-Debreu portfolio problem where investors are constrained to have linear portfolios. Of course, when there are only two states of nature, imposing linearity is not restrictive. This is only when there are three states or more that imposing a linear relationship in the state-dependent consumption may be constraining for investors.
13.2 Characterization of the Optimal Portfolio
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_197.html (1 of 2) [4/21/2007 1:31:43 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_197.html
For the general model where markets is written as
can be discrete, continuous or mixed, the portfolio problem with complete
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_197.html (2 of 2) [4/21/2007 1:31:43 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_198.html
Page 198
The Lagrangian takes the form
where ξ is the Lagrangian multiplier associated with the budget constraint. The first-order condition for this problem is written as follows:
In the finite case this condition is directly obtained by using the Lagrangian method with respect to the vector of decision variables (c1, c2, . . .). When there is a continuum of states, deriving the above condition requires the use of the simplest version of the variational calculus method. But there is nothing specific to be learned to examine this technical detail. Notice that since the problem is concave under risk aversion, condition (13.4) is the necessary and sufficient condition. It is a tautology in this model that the consumption in state x depends only on the price density π = π(x) in that state. Without any loss of generality, let us assume that there is a monotone relationship between π and x, so that function π admits an inverse. We can thus define a function C(π) such that
so that the optimal consumption plan is c(x) = C(π–1(x)). C(π) is the optimal consumption in all states whose price density is π. By the monotonicity and concavity of u and condition (13.5), C is decreasing in π: the larger the state price density, the smaller is the demand for the corresponding Arrow-Debreu security and consumption. Or, the smaller the expected return of an Arrow-Debreu security, the smaller is the demand for it. It is important to note that this is not a comparative statics result, since we are not changing any parameter of the economy. We just compare the demand for Arrow-Debreu securities associated to two different states in a given economy. A different exercise is to determine the effect of a change in the state price π(x) on the optimal demand c(x′) of the various Arrow-Debreu assets. See exercise 69 at the end of this chapter for more details.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_198.html [4/21/2007 1:31:43 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_199.html
Page 199 One can determine the optimal exposure to risk by differentiating condition (13.5). It yields
or eliminating ξ,
where T(C) = –u′(C)/u″(C) is the index of absolute risk tolerance. The sensitivity of consumption to the stateprice density equals the absolute risk tolerance of the agent evaluated at that level of consumption. Observe that – C′ is a local measure of the risk exposure that the agent is willing to take. A constant C means that the agent is not willing to take any risk. It yields the same level of consumption in all states of the world, although consuming in some states is more costly than in others. As shown by equation (13.6a), this is optimal only if the agent has a zero risk tolerance, namely only if he is infinitely risk-averse (see exercise 71). In all other cases, T is not zero, and so C′ is strictly negative: the agent bears risk. The finitely risk-averse agent is willing to take advantage of cheaper consumption in more favorable states despite it generates a risky consumption plan ex ante. These results parallel those obtained in section 4.1 for the standard portfolio problem. In both cases risk-averse agents always take some risk. We have the following results: 49 In an economy with complete markets and risk-averse investors, the demand for Arrow-Debreu securities satisfies the following properties: PROPOSITION
1. C′(π) ≤ 0: the demand for the Arrow-Debreu security associated to state x is smaller in states where π(x) is larger. 2. The local exposure to risk measured by –C′(π) equals the absolute risk tolerance measured in the corresponding state divided by the state price (equation 13.6a). 3. The local exposure to risk is decreasing (resp. increasing) in the state price if u′ is convex (resp. concave). Proof We just have to prove that –C″(π) is negative when u′ is convex. From equation (13.6a), this is true if
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_199.html [4/21/2007 1:31:44 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_200.html
Page 200
Figure 13.1 Optimal consumption plans for u and v, v being more risk-averse than u
Since from (2.14),
condition (13.7) becomes
Under risk aversion this is true if and only if P(C(π)) is nonnegative, that is, if u′ is convex. In other words, under risk aversion and prudence, the optimal portfolio strategy characterized by function C(π) is decreasing and convex. We draw two typical optimal portfolio strategies in figure 13.1. Following Mitchell (1994), for example, one may also inquire about how the total budget of the investor is allocated to the purchase of the different Arrow-Debreu securities. Since the budget associated to all contingent assets whose state price density is π equals B(π) = πC(π), an increase in π reduces B if C(π) + πC′(π) is negative. By condition (13.6a), this is equivalent to C – T(C) < 0, or C/T(C) < 1, that is, to relative risk aversion being smaller than unity. This result should be put in relation with property 1 in proposition 9. 50 The amount invested in a Arrow-Debreu security is smaller (larger) in states with a larger state price per unit of probability if relative risk aversion is smaller (larger) than unity. PROPOSITION
13.3 The Impact of Risk Aversion In the standard portfolio problem, an increase in risk aversion always reduces the optimal risk exposure, where file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_200.html (1 of 2) [4/21/2007 1:31:44 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_200.html
the risk exposure is mea-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_200.html (2 of 2) [4/21/2007 1:31:44 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_201.html
Page 201 sured by the amount invested in the risky asset. In the complete markets model, a global measurement of the exposure to risk is less obvious, since C′(π) just provides a local measurement of the consumption risk. What we can show is in the following proposition. It is a direct consequence of property (13.6a). PROPOSITION
51 Consider two agents with utility functions u1 and u2 that are twice differentiable. If u1 is more
risk-averse than u2 (possibly with a different wealth level), then the consumption plan of u1 can cross the consumption plan of u2 at most once, from below. Proof Suppose that there exists a π0 such that C1(π0) = C2(π0). Then from condition (13.6a) we have
where T1 and T2 are the absolute risk tolerance of respectively u1 and u2. The inequality is a direct consequence of the assumption that u1 is more risk-averse than u2. We conclude that whenever C1 crosses C2, it must cross it from below. By a standard continuity argument, it implies that the two curves can cross only once. We draw two representative consumption plans in figure 13.1. It is apparent that u1 has a safer position than u2. It is true in particular if we assume that the support of
is in a small neighborhood of π0.
A direct application of proposition 51 is that the presence of a non-tradable unfair background risk that is independent of will induce risk-vulnerable investors to select a safer consumption plan C. More generally, this proposition allows us to use all results that we already obtained about the conditions under which a well-defined indirect utility function is more concave than the original utility function.
13.4 Conclusion. We assumed in this chapter that there are as many independent securities as the number of possible states of the world. This implies that investors can decide how much to consume in each state without any other constraint than the ex ante budget constraint. Investors can select their exposure to risk in a much broader set than the set we considered in the alternative standard portfolio model. In this paradigm there is no more difference between the portfolio problem and the problem of determining how many bananas and apples to
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_201.html [4/21/2007 1:31:45 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_202.html
Page 202 consume under certainty. The decision problems are technically the same: maximizing the consumer's objective under a linear budget constraint. The only specificity comes from expected utility, which makes the objective function separable. This is different for the consumption problem under certainty with many commodities, where the marginal utility of bananas is a function of the number of apples already in the bundle. The property of separability due to the independence axiom allowed us to derive simple properties of the optimal structure of the portfolio by using standard algebra.
13.5 Exercises and Extensions Exercise 69 In this chapter we showed that consumption is larger in states where contingent prices per unit of probability are smaller. In this exercise we examine how the demand for Arrow-Debreu securities is affected by a change in the price of one of them. Suppose that there is a finite number of states s = 1, . . . , n. Consumption in state j is a function of the vector of state prices per unit of probability: cj = cj(π1, . . . , πj, . . . , πn). We examine the effect of a change in πi on the optimal portfolio (c1, . . . , cn). • Show that a small increase in the state price reduces the consumption in that state: dci/dπi ≤ 0. Arrow-Debreu securities are not Giffen goods in a complete market economy under expected utility. • Show that Arrow-Debreu securities are gross subtitutes (dcj/dπi > 0) if and only if relative risk aversion is less than unity. • (Clyman 1995) Show that the amount spent to consume in state i decreases with the corresponding state price (dπici/dπi < 0) if and only if relative risk aversion is less than unity. Notice that we could have considered an alternative model in which the investor is endowed with a state contingent income ω(x) ex post, rather than a sure wealth z ex ante. The budget constraint would then be written as . In that case an increase in π(x) makes the agent wealthier if c(x) is less than the endowment ω(x) in that state. That reverses the wealth effect, and the global effect of an increase in π(x) is ambiguous. Mitiushin and Polterovich (1978) and Mas-Colell (1991) showed that the law of demand holds·(i.e., the scalar product of the vector of the variation of demand with the vector of variation of prices is nonnegative) if relative risk aversion is uniformly less than 4.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_202.html [4/21/2007 1:31:46 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_203.html
Page 203 Exercise 70 Show that proposition 51 can be generalized in the following way: Define π0 as the contingent price at which the demand of agents 1 and 2 are equal. Then
for all π ≤ π0, assuming that agent 1 is
more risk-averse than agent 2 and that at least one of the two agents is DARA. Exercise 71 Solve program (13.3) when the utility function is piecewise linear, with parameters as in equation and there (2.23). Show that the optimal consumption is state-independent (full insurance) when exists a scalar λ such that a ≤ λπ(x) ≤ 1 for all x. Explore the characteristics of the optimal consumption plan, when it exists, for other values of the parameters. Exercise 72 Show that imposing that consumption be linear in the market return, namely that , is not retrictive when the investor has a logarithmic utility function.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_203.html [4/21/2007 1:31:46 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_204.html
Page 204
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_204.html [4/21/2007 1:31:47 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_205.html
Page 205
14 Risk on Wealth In the previous chapter we examined how a change in price and a change in risk aversion do affect the demand for contingent claims. We now turn to the analysis of the effect of a change in initial wealth z. This will allow us to determine the impact of z on the optimal expected utility. We will then extract some properties of the optimal portfolio strategy with complete markets similar to those that we obtained in the case of the standard portfolio model. Let v(z) denote the maximal expected utility that can be attained by an agent with wealth z who invests his wealth in the stock market. We called this the value function. In a slightly generalized version of the complete markets model that we examined in the previous Chapter, we allow for the utility function on consumption to depend on the state of the world. This means that the utility depends upon C and π. To each state of the world π, there is an associated function u(.,π) that is increasing and concave. With such an extension, the program of the investor is now written as
Notice that u′ = ∂u/∂C and –u′/u″ measure respectively the marginal utility of consumption and the absolute tolerance to risk associated to consumption. By defining v, we are now able to characterize the marginal value of wealth by v′ and the absolute tolerance to risk on wealth by –v′/v″. One of the main aims of this chapter is to examine the relationships that exist between the characteristics of the utility function of consumption and those of the value function of wealth.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_205.html [4/21/2007 1:31:47 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_206.html
Page 206 Whereas z was previously considered as a constant, we are now interested in a sensitivity analysis of v with respect to a change in z. Obviously both the Lagrangian multiplier and the optimal solution are a function of z. They are denoted respectively ξ(z) and C(π, z). The solution to this program satisfies the following condition:
where u′ denotes the marginal utility of consumption. Solving system of equations (14.2) and (14.3) allow us to obtain functions C(π, z) and ξ(z). Once the optimal function C(π, z) is obtained, one can calculate the value function by
14.1 The Marginal Propensity to Consume in State π In order to examine the impact of a change in wealth on the maximal expected utility, it is first useful to see how it affects the optimal consumption plan C(π, z). To do this, we fully differentiate the above system. It yields
and
Combining these two conditions implies that
In consequence C is increasing in z for all π. The intuition is that an increase in initial wealth raises the consumption level in all states of the world. From conditions (14.3), (14.4), and (14.6), we now have an explicit formula to calculate the marginal propensity to consume in state π:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_206.html [4/21/2007 1:31:48 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_207.html
Page 207 The marginal propensity to consume in a given state is equal to the absolute risk tolerance in that state divided by a weighted sum of the absolute risk tolerances in the various states. It can be useful to determine whether the marginal propensity to consume is increasing or decreasing, that is whether the consumption function C(π, z) is convex or concave in z. Notice first that by condition (14.7) the weighted average of ∂C/∂z is independent of z. Thus, if there exists some π where ∂C/∂z is increasing in z, there must exist other values of π where it is decreasing in z. Consider now a specific π. By differentiating condition (14.7), we obtain that the marginal propensity to consume C(π, z) is convex in z if the following inequality is satisfied:
Observe that the right-hand side of this inequality that we denoted is a weighted average of T′(z). In consequence is somewhere between the minimum and maximum realizations of . . The Thus C(π, z) is locally convex (resp. concave) in z wherever T′(C(π, z), π) is larger (resp. smaller) than limit case is when utility functions have linear absolute risk tolerance with the same slope, that is, when T′ is a constant independent of C and π. It implies that condition (14.8) is always an equality. It implies in turn that the marginal propensity to consume is constant.
52 The marginal propensity to consume is independent of z for utility functions exhibiting linear absolute risk tolerance with a state-independent slope. Otherwise, there exists a function defined by condition (14.8) such that PROPOSITION
Consider the case of a state-independent utility function that is not HARA, and suppose that T is convex. Then the proposition above states that the marginal propensity to consume out of wealth is decreasing if T′(C(π, z)) is smaller than . Let us define function such that . A restatement of the result for a state-independent convex T function is that the marginal propensity to
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_207.html [4/21/2007 1:31:48 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_208.html
Page 208
Figure 14.1 Optimal consumption plan as a function of wealth when absolute risk tolerance is convex, with π1 ≤ π2 ≤ π3 ≤ π4
consume is decreasing if C(π, z) is smaller than
. A representative illustration of this case is given in figure
14.1. Notice that it may be possible that a curve C(π, .) crosses curve concave and locally convex.
, implying that it is alternatively locally
14.2 The Preservation of DARA and IARA Let us now turn to the analysis of how the opportunity to take risk on Arrow-Debreu markets affect the marginal value of initial wealth. Conditions (14.3) and (14.5) imply that
and
Combining these equations, we obtain that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_208.html (1 of 2) [4/21/2007 1:31:49 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_208.html
for all π, where T(., π) and Tv(.) are the absolute risk tolerance of respectively u(., π) and v(.). Multiplying both side of the equality by
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_208.html (2 of 2) [4/21/2007 1:31:49 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_209.html
Page 209
, taking the expectation and using the property that
implies that
We can go one step further by differentiating condition (14.11a), which implies that
We conclude that
is a weighted average of the ex post T′, namely that the derivative of the absolute risk
tolerance is a martingale. Indeed, by (14.5), equals 1. We thus have that bound and the upper bound of T′. This directly yields the following result: PROPOSITION
is between the lower
53 Consider any pair (k1, k2) in R2. If the utility function on consumption satisfies condition k1P(z,
π) ≥ k2A(z, π) for all (z, π), so does the value function v defined by condition (14.1). Proof Observe that T′ = –1 + (P/A). It implies that
and
Combining these properties with the observation that result.
is weighted average of the state-dependent T′ yields the
This means that gambling on complete markets preserves the property that P is uniformly larger or uniformly smaller than kA. Several special cases of proposition 53 are noteworthy: • (k1, k2) = (0, –1): as we already know, risk aversion is preserved. • (k1, k2) = (1, 0): prudence is preserved. • (k1, k2) = (1, 1): decreasing absolute risk aversion is preserved. file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_209.html (1 of 2) [4/21/2007 1:31:49 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_209.html
• (k1, k2) = (–1, –1): increasing absolute risk aversion is preserved. • (k1, k2) = (1, 2): condition P ≥ 2A is preserved.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_209.html (2 of 2) [4/21/2007 1:31:49 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_210.html
Page 210 These results should be put in contrast with the fact that even DARA is not preserved in the standard portfolio problem (see Roy and Wagenvoort 1996 and exercise 63 in chapter 11). From now on, we assume that u is not state-dependent, meaning that u(., π) = u(.) for all π.
14.3 The Marginal Value of Wealth In this section we are interested in determining how the opportunity to gamble on complete markets affect the : if the agent marginal value of wealth. In order to eliminate a wealth effect, we assume hereafter that takes a risk free position, he would end up consuming his initial wealth z. This also means that the risk-free rate for the period covering the decision date to the consumption date is zero. Observe that we already addressed the question of how the opportunity to take risk affect the marginal value of wealth. More specifically, we addressed this question in the framework of the standard portfolio problem. In proposition 34 we showed that the option to invest in stocks raises the marginal value of wealth if and only if prudence is larger than twice the absolute risk aversion. We obtain the same result when the option is on investing in complete markets, as stated in the following proposition. 54 Suppose that . Then the option to invest in a complete set of Arrow-Debreu securities raises (resp. reduces) the marginal value of wealth if and only if absolute prudence is larger (resp. smaller) than twice the absolute risk aversion: –u′″(C)/u″(C) ≥ (≤)2[–u″(C)/u′(C)] for all C.
PROPOSITION
Proof Define function
If
(y) = u′–1(1/y). The combination of conditions (14.2) and (14.3) yields
is convex, Jensen's inequality implies that
Since, by condition (14.9), ξ = v′(z), we obtain u′(z) ≤ v′(z). Finally it is easy to verify that the convexity of equivalent to P(z) ≥ 2A(z)
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_210.html [4/21/2007 1:31:50 AM]
is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_211.html
Page 211
for all z. The proof of necessity is by contradiction, by selecting where is concave.
such that the support of
be in the zone
14.4 Aversion to Risk on Wealth We earlier obtained that the absolute risk tolerance on wealth is characterized by
If , meaning that the risk-free rate is zero, then the degree of tolerance to risk of the value function is a weighted average of the ex post absolute risk tolerance of the investor. This means that the absolute risk tolerance is a martingale under the risk-neutral probability distribution. Cox and Leland (1982) and He and Huang (1994) obtained the same property in a continuous-time framework. The use of Jensen's inequality together with the fact that
directly yields the following result:
55 Suppose that . Then, a complete set of Arrow-Debreu securities raises (resp. reduces) the absolute risk tolerance of the agent if and only if the absolute risk tolerance T(C) = –u′(C)/u″(C) is convex (resp. concave) in C. PROPOSITION
The reader certainly made a connection between this result and the one obtained in Proposition 38 in the case of the standard portfolio problem. To make this link more precise, consider the following dynamic investment problem: At the start of each period, the investor can bet on any gamble giving him a dollar if and only if a prespecified event x occurs during the period. For each possible exclusive event x, there is a price Π(x) for betting on it. These prices can be a deterministic function of time, but it is assumed that Σp(x)Π(x) = 1 in all periods. How does the option to gamble another time in the future affect the optimal gambling strategy today? Combining propositions 51 and 55 yields a complete answer to this question: if T is convex, the investor will take a riskier position, whereas he will take a safer position if T is concave. Two remarks must be made when comparing the effect of time horizon on the optimal dynamic investment strategy in the two frameworks. First, we obtain a much more clear-cut result in the case of the complete market model. Indeed, in the standard portfolio problem, the effect of time horizon is ambiguous when absolute risk tolerance
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_211.html [4/21/2007 1:31:51 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_212.html
Page 212 is convex. Second, it is noteworthy that the method of proof just use standard differential calculus in the complete markets framework. This is to be compared to the heavy artillery that we had to use to solve the dynamic version of the standard portfolio problem.
14.5 Concluding Remark In a sense the optimal portfolio decision problem is easier to solve in the case of complete security markets than in the one risk-free, one risky assets case. We just have to solve a system of n + 1 equations in the first case, where n is the number of states of the world. At the limit it is not necessary to know what is a probability or an expectation operator to do it. Thus the diffidence theorem and its corollaries are just useless in this world. The assumption that markets are complete allows for the most efficient use of the separability of the decision maker's objective function across states.9
14.6 Exercises and Extensions Exercise 73 (Dybvig 1988) Consider an economy with two assets. The risk-free asset has a zero return. The risky asset has two possible returns with equal probabilities: in the "up" state, the return is +100%. In the "down" state, the return is –50%. • What are the prices of the Arrow-Debreu securities associated to the "down" state and the "up" state? • Determine the optimal portfolio of Arrow-Debreu securities of an agent with a logarithmic utility function. Determine the corresponding portfolio containing the risk-free asset and the risky asset. • Suppose now that the agent lives for two periods, and that the return of the risky asset can be either –50% or +100% at each period with equal probability. A state of the world is now defined by a specific history. They are four states here: up-up, up-down, down-up, and down-down. What is the price of the Arrow-Debreu security at the beginning of the first period that is associated to the state up-up? What is the optimal dynamic strategy of the logarithmic agent?
9. Chateauneuf, Dana, and Tallon (1998) explore the consequences of nonadditive expected utility on optimal portfolio decisions and equilibrium prices.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_212.html [4/21/2007 1:31:51 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_213.html
Page 213 • Consider now an agent who lives for four periods. He consumes only at the end of the last period. The agent has an initial wealth of 16. He follows a stop-loss strategy with a maximum loss of 8. This means that he invests everything in the risky asset as long as the value of the portfolio remains above 8. If the value of the portfolio goes down to 8, he switches to the risk-free asset. Show that this cannot be an efficient strategy, namely that there is no differentiable utility function that makes this strategy optimal. [Hint: Use the necessary condition that the state-contingent consumption must be nonincreasing in the state price (per unit of probability).] Exercise 74 (Gollier 1997) There is another way to prove that stop-loss and lock-in strategies cannot be dynamically optimal when the expected excess return is positive. For each of these strategies, there are some states of the world where the agent is 100% in the risk-free position. This can be optimal only if the value function exhibits infinite risk aversion at the current wealth level. Show that this is not possible when the utility function is differentiable. Exercise 75 (Extension of proposition 54) In this exercise we examine the effect of a change in the distribution of
on the marginal utility of wealth. Define function
• Show that χ is decreasing in λ, and is concave (convex) in π if • Show that a mean-preserving spread of
, where
.
is concave (convex).
reduces (increases) the marginal value of wealth if P ≤ (≥)2A.
and that there Exercise 76 (Extension of proposition 54) Prove the following proposition: Suppose that exists y ≥ 0 such that limx→y u′(x) = +∞. Then the option to invest in a complete set of Arrow-Debreu securities reduces the marginal value of wealth if absolute prudence is smaller than twice the absolute risk aversion. [Hint: Use the property that if and concave, then for all k ≥ 1 and x ≥ 0 (subhomogeneity).] . Then a Exercise 77 (Extension of proposition 55) Prove the following proposition: Suppose that complete set of Arrow-Debreu securities raises (resp. reduces) the absolute risk tolerance if the absolute risk tolerance T is convex (resp. concave) and T(0) = 0. [Hint: Use the same hint as in the previous exercise.]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_213.html [4/21/2007 1:31:52 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_214.html
Page 214
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_214.html [4/21/2007 1:31:52 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_215.html
Page 215
VI Consumption and Saving.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_215.html [4/21/2007 1:31:53 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_216.html
Page 216
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_216.html [4/21/2007 1:31:53 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_217.html
Page 217
15 Consumption under Certainty 15.1 Time Separability Up to now we examined models where the agent consumes all his wealth at a single date in time. In this part of the book, we examine how an expected utility maximizer allocates his wealth to consumption over various dates. We have to examine the objective function of an agent who lives from date t = 0 to date t = n. Let ct denote consumption at date t. Suppose that the agent faces different feasible uncertain consumption streams where is the random variable characterizing the uncertain consumption at date t if stream i is selected. The problem of the agent is to rank them. If the agent satisfies the continuity axiom and the independence axiom for intertemporal lotteries, that is, for lotteries whose outcomes are in Rn+1, then we know from chapter 1 that there exists a function U: Rn+1 → R such that consumption stream is preferred to consumption stream if and only if
U(c0, c1, . . ., cn) is easily interpreted as the lifetime utility of the sure consumption flow (c0, c1, . . ., cn). In this chapter we examine this lifetime utility function under certainty. We will reintroduce uncertainty in the next chapters. This general approach raises several difficulties due to the fact that the marginal utility of consumption at date t is a function of past and future consumptions. Backward induction is not easy to use when such kind of interrelations are allowed. In consequence it is customary in intertemporal economics to assume that U is timeseparable, meaning that there exist n + 1 functions ut: R → R, t = 0, 1, . . ., n, such that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_217.html [4/21/2007 1:31:53 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_218.html
Page 218
Time separability implies that the ordering of consumption streams from t = 0 to t0 is independent of whatever consumption levels from t0 + 1 onward. This looks very much like the independence axiom. In fact this condition plays exactly the same role with respect to time than the role played by the independence axiom with respect to the states of the world. In particular, this independence condition implies time separability (see Blackorby, Primont, and Russell 1999). This condition is restrictive, since it does not allow for habit formations, a term for the idea that the marginal utility of future consumption is increasing with the level of past consumption (e.g., see Constandinides 1990; and Campbell and Cochrane 1995). Despite this deficiency we will maintain this assumption throughout. The reader more interested on separability in preferences should have a look at Debreu (1960), Gorman (1968), or more recently at Wakker (1989).
15.2 Exponential Discounting A further restriction on intertemporal preferences is often considered in the literature. Namely it is assumed that
where u = u0 is called the felicity/utility function on consumption, and β is the discount factor. This assumption is often referred to as "exponential discounting," since its equivalent formula with continuous time is ut(c) = e–rtu(c). Several arguments are in favor of exponential discounting. First, this assumption forces the time consistency of preferences. Time consistency in a world of certainty means that what is optimal to do tomorrow, seen from today, is still optimal the next day. Consider a problem in which you can save from t = 1 to t = 2 at a gross return ρ. Seen from today (t = 0), the optimal saving solves the following problem:
where wt is the income at t. One period later (t = 1), when the time to go to the bank arises, the optimal saving solves the somewhat different problem:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_218.html [4/21/2007 1:31:54 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_219.html
Page 219
Obviously what is optimal to implement at t = 1 is what it was planned one period earlier only if ut = βtu0. This is the only case where the marginal rate of substitution between consumptions at t = 1 and t = 2 is the same today and tomorrow. (See further Laibson 1997 and Laibson and Harris 1999 for an analysis of sophisticated consumption behaviors with hyperbolic discounting.) A second argument in favor of exponential discounting is that if we normalize the utility if death at zero and if we interpret β as the survival probability from one date to another, ut(c) = βtu0(c) has the interpretation of the expected utility of the consumption flow at date t seen from date 0. Finally discounting felicity is a simple way to guarantee that the objective function of the agent is well defined in the case of an infinite-horizon model (n → ∞). It is noteworthy that β is used to discount felicity, not money. It is commonly assumed that β is less than unity. This means that one unit of felicity tomorrow is valued less than one unit of felicity today. This is a notion of time impatience for happiness. Because βt tends to zero exponentially, it implies that the distant future does not matter for current decisions. In the next sections we reinterpret the concavity of the felicity function in relation with the substitution of sure consumptions over time.
15.3 Consumption Smoothing under Certainty There are two motives to saving in a risk-free environment: one could save to smooth consumption over time or one could save to take advantage of a high return on savings. We first examine the smoothing argument. Consider a two-date model with u0 ≡ u1 ≡ u. In this model there is no time impatience: one more unit of felicity tomorrow generates the same increase of lifetime utility as one more unit of felicity today. Furthermore we can assume that the return on savings is zero. This is to isolate the consumption smoothing effect of savings. Suppose additionally that the income flow is constant (smooth) over time: the agent earns w at each date and has no reserve of liquidity at date 0. The problem is to determine the optimal saving s at that time:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_219.html [4/21/2007 1:31:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_220.html
Page 220 Observe that V′(s) = u′(w + s) – u′(w – s) is zero for s = 0 and is decreasing in s if u is concave. Thus, under the concavity of the utility function, the optimal saving is zero. This means that the agent has a preference for a constant consumption over time. This would be the worst solution if the utility function were convex. Thus the concavity of the utility function represents the willingness to smooth consumption over time. One can measure the intensity of the willingness to smooth consumption over time by considering a situation where the income w0 at date 0 is strictly less than the income w1 at date 1. Since the marginal utility of consumption is larger at date 0 than at date 1 (u′(w0) ≥ u′(w1)) if u is concave, we know that the agent will not accept a deal to exchange one unit of consumption today for one unit of consumption tomorrow. He will require more than one unit of consumption tomorrow for the sacrifice of one unit of consumption today. Accepting the deal would increase the discrepancy between the consumption levels at the two dates. So the degree of resistance to intertemporal substitution can be measured by the additional reward k that must be given to the agent at date 1 to compensate for the loss in consumption at date 0. If the monetary unit is small with respect to the consumption level, k is defined by the following condition:
This condition states that the marginal loss in utility at date 0 must equal the marginal increase in utility at date 1. If w1 is close to w0, we can use a first-order Taylor expansion of u′(w0) around w1. It yields
Thus the resistance to intertemporal substitution is approximately proportional to the growth rate of consumption. The multiplicative factor, –w1u″(w1)/u′(w1), is hereafter called the relative fluctuation aversion, or the relative degree of resistance to intertemporal substitution of consumption. It is widely believed is in between 1 and 5. We can estimate the RISC by performing experiments or by observing saving behaviors under certainty. By regressing the rate of growth of consumption on the rate of interest, Hall (1988) found a RISC around 10, and Epstein and Zin (1991) found a value ranging from 1.25 to 5. The reader certainly did not miss the point that the relative degree of resistance to intertemporal substitution is equivalent to the index
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_220.html [4/21/2007 1:31:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_221.html
Page 221 of relative risk aversion. We flesh out the relationship between the consumption problem under certainty and the risk-taking problem in the next section.
15.4 Analogy with the Portfolio Problem The willingness to smooth consumption is only one aspect of the picture. If the interest rate is not zero, the desire to smooth consumption is in conflict with the possibility to increase total wealth by investing money over time. To simplify, let us assume that the interest rate ρ in the economy is the same for all time horizons (the yield curve is flat). There are zero-coupon bonds available for each date. A zero-coupon bond for date t guarantees to its holder 1 unit of the consumption good at date t and nothing for the other dates. The price of this zero-coupon is 1/ ρt. At date t = 1, the saving problem under certainty is written as
where wt is the sure income at date t and z0 is the current gross wealth. We defined z as the current net wealth, taking into account of the discounted value of future incomes. The budget constraint means that the discounted value of the consumption stream equals the current net wealth. Alternatively, writing the budget constraint as
means that the portfolio of zero-bonds that the agent purchases has a zero value. The problem above can be rewritten as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_221.html [4/21/2007 1:31:56 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_222.html
Page 222 Now observe that this problem is formally equivalent to the portfolio problem (13.1) where there is a complete set of Arrow-Debreu securities. The table of correspondence is
If we pursue the table of correspondence, we find that the zero-coupon bond is the exact twin of the ArrowDebreu security. The assumption of a state-independent utility function is technically equivalent to the assumption of a time-independent felicity function. This is the additive nature of the objective function with respect to time and uncertainty that generates these equivalences. All the results that we obtained for the static Arrow-Debreu economy can be used in this context. Here is a first illustration of the correspondence with the portfolio problem: we know that full insurance (constant consumption) is optimal with a differentiable utility function only if the state price density π is state independent. Since πt is proportional to (βρ)–t+1 in the saving version of the model, the state price density is constant only if βρ = 1. In the terminology of the saving problem, it is optimal to smooth consumption completely if and only if the discount rate equals the inverse of the gross return on saving. This is a generalization of the result presented in the previous section where we assumed that β = ρ = 1. More generally, by proposition 49, we know that the level of consumption in different states is a decreasing function of the state price density π in the corresponding state, under the differentiability and concavity of u. Again, using the table of correspondence with πt proportional to (βρ)–t+1, we obtain that consumption increases with time if βρ is larger than unity. This is when the willingness to increase lifetime wealth by saving is stronger than time impatience. We accept the notion that we consume less at the early stages of life in order to benefit from the positive return on savings. This is true even if βρ is just above unity. This means that consumption smoothing is only a second-order effect if u is differentiable, exactly as risk aversion is a second-order effect for the demand of a risky asset. By condition (13.6a), we see that the rate at which consump-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_222.html [4/21/2007 1:31:56 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_223.html
Page 223 Table 15.1 Correspondences between the portfolio problem and the consumption problem under certainty Arrow portfolio problem Separability across states Risk aversion Arrow-Debreu securities Full insurance optimal only if πs = π
Consumption problem Separability across time Fluctuation aversion Zero-coupon bonds Income smoothing optimal only if βρ = 1
Demand decreasing in π
Consumption increasing in time if βρ > 1
tion will increase over time is proportional to –u′(ct)/u″(ct). This means that saving under certainty is proportional to the absolute risk tolerance of the agent. Table 15.1 provides more details about these equivalences. One could also illustrate this equivalence by looking at the important problem in macroeconomics of the effect of an increase in the interest rate on the optimal saving. Let us suppose here that the agent has no future incomes (w1 = · · · = wn = 0). The increase in the interest rate provides an argument to increase saving in order to take advantage of the higher return of his investment (speculative motive). But because the agent is a net lender, the increase in the interest rate also increases his future wealth. It induces him to increase its current consumption to smooth it over time, thereby reducing saving (consumption smoothing effect). We know the condition under which the first effect dominates the second. The following proposition is similar to proposition 50: 56 Consider the consumption-saving problem of an agent with no future incomes. An increase in the interest rate increases (resp. reduces) savings if the relative aversion to consumption fluctuation –cu″(c)/u′(c) is smaller (resp. larger) than unity. PROPOSITION
The intuition of this result is simple. When the fluctuation aversion is small, the willingness to transfer the benefit of the increase in the future return of saving into an immediate increase in consumption is reduced. It implies that the speculative motive of saving dominates in that case. From the equivalence, we also obtain the following proposition, which is an application of proposition 51: 57 Consider the consumption-saving problem under certainty. An increase in the concavity of the felicity function in the PROPOSITION
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_223.html [4/21/2007 1:31:57 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_224.html
Page 224 sense of Arrow-Pratt increases (resp. reduces) early consumption if βρ is larger (resp. smaller) than unity. This is a confirmation that the degree of concavity of the felicity function is a measure of the resistance to intertemporal substitution. In the normal case where βρ is larger than 1, the agent accept to substitute current consumption by future consumption in order to take advantage of the relatively large return on investments. An increase in the concavity of u tends to reduce this substitution of consumption over time. It increases early consumption.
15.5 The Social Cost of Volatility The understanding of consumption behavior is probably one of the most important challenges in modern macroeconomics. There has been a lot of developments in this area of research since the seminal papers of Modigliani and Brumberg (1954) and Friedman (1957). Life cycle models of consumption rely on the assumption that households solve forward looking intertemporal consumption problems. Program (15.3) belongs to this family. Lucas (1987) raised an interesting puzzle of life cycle models of consumptions. This puzzle is the twin of the puzzle that we raised in section 2.8 about the low social cost of the macroeconomic risk. In order to present Lucas's puzzle, let us solve the consumption problem (15.3) under certainty in the case of a CRRA utility function. Using the first-order condition (13.5) and the definition (15.5) for π in the context of intertemporal consumption, it is easily verified that the optimal consumption path is
where a is equal to (βρ)1/γ. Constant c0 can be obtained by using the lifetime budget constraint of the agent. The important point here is that the growth rate g of consumption is constant over time. It equals
This formula generalizes approximation (15.1) to the case β ≠ 1. Under CRRA it is optimal to let consumption growth at a constant rate. If there is no impatience (β = 1), the optimal growth rate of consumption is approximately equal to the ratio of the interest rate
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_224.html [4/21/2007 1:31:58 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_225.html
Page 225
Figure 15.1 Actual and optimal consumption paths with γ = 0.5 (dashed line) and γ = 2 (solid line)
ρ – 1 and the relative fluctuation aversion γ. With a risk free rate around 1%, as observed in the period 1963 to 1992, the optimal growth rate should be somewhere one-fourth of a percent (γ = 4) and one percent (γ = 1). For the sake of illustration, consider an agent born on January 1, 1963, who is perfectly aware of the sequence of his disposable income for the next thirty years. Because he is a representative agent, the growth rate of his income is exactly equal to the growth rate of real US GDP per capita that was actually observed in the period 1963 to 1992, as reported in section 2.8. The geometric mean of the growth is g = 1.849% per year. Assuming the absence of impatience (β = 1) and an observed risk-free rate around ρ – 1 = 1%, this is compatible with the . We show in figure 15.1 the optimal consumption path of this agent for γ = 0.5 coefficient and γ = 2. We see that the actual consumption path is close to the optimal one for γ = 0.5. We also check that a more concave felicity function leads to more consumption smoothing. Of course, the actual consumption path is not optimal because the growth rate of consumption has not been constant over time. Business cycles occur that force households to adapt their consumption accordingly. These cycles make people worse off because of the volatility generated on their consumption. Given the concavity of the felicity function, people prefer to smooth their consumption around a constant trend. What is the social cost of business cycles, assuming that they are deterministic? Let us focus on the agent with γ = 0.5,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_225.html [4/21/2007 1:31:58 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_226.html
Page 226 since this agent might select a growth rate of consumption similar to that observed on average over the thirty years. One could measure the cost of business cycle by the reduction in the growth rate of consumption that this agent would accept in exchange for the complete elimination of business cycles. His optimal consumption path then would start at c0 = 10,708 dollars with a growth rate of 2.01% per year. The actual path provides the same discounted utility as a path starting at the same c0 but with a growth rate equaling 2.006%. In consequence business cycle "costs" amount to a reduction of 0.004% in the annual growth rate of the economy. This is obviously an insignificant effect. Lucas concluded that economists should deal with the determinants of the longterm growth rather than with the reduction of volatility.
15.6 The Marginal Propensity to Consume. An important macroeconomic question since Keynes is to determine how agents react to an increase in their wealth in terms of consumption and saving. The problem is thus to derive the vector of consumption functions (c1 (z), . . . , cn(z)) that solves program (15.3) for all z. The marginal propensity to consume is the share of the increase in wealth that is immediately consumed, . To examine the effect of a change in wealth on the optimal consumption plan, we use the analogy presented above and the results that we derived in section 14.1. We have that
where πt is defined by equation (15.5) and function C is the function that was studied in section 14.1. By analogy to condition (14.7), we obtain that the marginal propensity to consume equals
where T(.) is the absolute risk tolerance of u(.) and
is the random variable that takes value πt defined by (15.5)
with probability pt defined by (15.4), t = 1, . . . , n. By these definitions, we obtain that
An immediate consequence of this condition is that of nonnegative ele-
is less than unity. This is because the denominator is a sum
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_226.html [4/21/2007 1:31:59 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_227.html
Page 227 ments, the first of which is just the numerator itself. Moreover a longer longevity reduces the marginal propensity to consume. The intuition is that the same discounted wealth must be consumed with more saving to finance consumption in old age. The limit case is where β = ρ = 1, and so the marginal propensity to consume is equal to 1/n. Indeed, we know that perfect income smoothing is optimal in this case, with any increase in wealth equally distributed over time. When βρ = 1 but β and ρ are each different, the same rule applies. However, one must take into account of returns on the additional saving leading to the fair distribution of the additional wealth. It yields . We conclude that the length of the time horizon has a strong effect on the marginal propensity to consume. Another question is to determine whether the marginal propensity to consume is decreasing with wealth. The idea that the consumption function is concave in wealth is an old one, with Keynes (1936) writing that ". . . the marginal propensity to consume [is] weaker in a wealthy community. . . ." Lusardi (1992) and Souleles (1995) found a substantially smaller marginal propensity to consume for consumers with high wealth. This provides some support to the idea that absolute risk tolerance should be convex in wealth. Indeed, by proposition 52, the marginal propensity to consume is constant if the utility function exhibits the HARA property. If absolute risk tolerance is convex rather than linear, function
will be decreasing if c1(z) is less than a threshold
. Notice that this threshold is between the smallest element and the defined by condition largest element of vector (c0(z), . . . , cn(z)). In the normal case where βρ is larger than 1, we know that this vector is ordered in an increasing way. This implies that c1(z) is necessarily less than the threshold. The other cases follow in the same way. PROPOSITION 58 We consider the consumption problem under certainty in the normal case where βρ is larger than 1, meaning that consumption is increasing with age. The marginal propensity to consume is decreasing if and only if absolute risk tolerance is convex.
15.7 Time Diversification and Self-Insurance In chapters 11 and 14 we examined the argument that younger people should take more portfolio risk. We used a model in which the agent could repetitively take risks over time. Intermediary con-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_227.html [4/21/2007 1:32:00 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_228.html
Page 228 sumption was not allowed. We concluded that younger investors could take more or less portfolio risk depending on whether absolute risk tolerance is convex or concave. No time diversification was possible in that framework, since all risks are added up into risk on final consumption. The idea of time diversification is better described by a model in which an agent could take risk today and then allocate the induced shock on his wealth by adjusting his consumption accordingly over several periods. Young people are potentially more willing to take risks because current risks can be attenuated by spreading consumption over time. In a sense the ability to split an uncertain payoff at a given period in subsequent periods allows the agent to self-insure the risk. For example, in the case of the absence of future risk opportunity and two dates, if complete consumption smoothing is optimal, a $1 loss on current investment will be split into a fifty-cent reduction in current consumption and a fifty-cent reduction in future consumption. Given the concavity of the utility of consumption, this strategy has a smaller effect on lifetime utility than a straight $1 reduction in immediate consumption. This is risk-sharing with one's future self, or self-insurance. We now provide a theoretical foundation to these ideas. The timing of the problem is as follows: • At date 0, there is no consumption, but the agent takes a decision under uncertainty. It generates an uncertain payoff at date 1. • At date 1, the agent observes the realization x of . He allocates his discounted certain wealth over the n remaining dates by selecting a consumption stream (c1, . . . , cn) that is feasible. We use backward induction to solve this problem. We start with the solution to the consumption problem at date 1. The problem at date 1 is written as follows:
where y = w1 + x is the available wealth at date 1. Variable y is often called the cash-on-hand at date t = 1. It is important at this stage to stress the fact that y is not total wealth, which would take into
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_228.html [4/21/2007 1:32:00 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_229.html
Page 229 account the discounted value of future incomes. Total wealth is the right-hand side of budget constraint (15.10). In order to solve the first-period problem, it is useful to inquire about the properties of function h. Again, we can equate this problem with problem (14.1), using the table of equivalence in section 15.4, together with . Notice that whereas h is the value function of cash-on-hand, v is the value function of aggregate wealth. The equivalence property implies that we can directly use the results developed in section 14.3 for the value function of the complete market model. For example, we know that h inherits the properties of monotonicity and concavity from u: the possibility to smooth shocks on incomes do not transform a risk-averter into a risk-lover. Thus, if the decision problem at date 0 is a take-it-or-leave-it offer to gamble, the agent should reject all unfair gambles. The value function h also inherits DARA from u. This means that an increase in income will make the agent more willing to take risk at date 0. In short, the qualitative effects of risk on welfare and optimal decisions under uncertainty are unaffected by the possibility to smooth the shocks on wealth over time. We now come back to the important question of time diversification: does the possibility to distribute large gains and losses into small changes of consumption over long periods increase the optimal exposure to a single risk in time? Using equation (14.13), his tolerance to the risk offered at t = 1 is given by
where
is the discounted wealth of the consumer.
Let us start with the simplest case where we suppose that βρ = 1 and w1 = w2 = · · · = wn = w, which implies that
This means that the tolerance to a single small risk is proportional to the time horizon of the agent. This is a strong argument in favor of younger people to be less risk-averse. It is the mathematical translation of the intuition that we gave at the beginning of this section for the concept of time diversification: losses and gains will be split into n equal shares distributed over time. We may question whether the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_229.html [4/21/2007 1:32:01 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_230.html
Page 230 term "time diversification" is acceptable here, since there is only one risk involved in the analysis. One possible justification for this terminology comes from above: when consumption smoothing is optimal, the initial risk is at date 1, at date 2, and so on. Because of the assumption on the time separability of split in preferences, this is as if the consumer would take these as independent variables. Thus the optimal consumption strategy organizes the diversification of the risk by replacing lottery by n seemingly independent one-nth of it. Equation (15.12) holds only when perfect gains and losses smoothing must be optimal. Also it focus only on the case where the current cash-on-hand is close to the target consumption level per period (y = w = z/n). Let us now relax these two assumptions. We start by relaxing condition βρ = 1 When βρ is not equal to unity, perfectly smoothing gains and losses over time is not optimal. Another effect comes into the picture in that case. Suppose that T is convex. Then using this assumption together with Jensen's inequality yields
where
is the consumer's feasible lifetime consumption level. It is the constant consumption level that is feasible given the initial cash-on-hand y. 59 Under HARA the absolute risk tolerance to a single risk in time is proportional to the absolute risk tolerance of u evaluated at the feasible lifetime consumption level. The proportionality coefficient equals PROPOSITION
, where n is the consumer's remaining lifetime. This is a lower (resp. upper) bound if the absolute risk tolerance is convex (resp. concave). This bound is attained when β = ρ = 1. We are closest to the benchmark case by assuming that ρ = 1 (but β ≠ 1) and w2 = · · · = wn = w in which case formula (15.13) yields
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_230.html [4/21/2007 1:32:02 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_231.html
Page 231
under convex absolute risk tolerance. This means that the (optimal) consumption fluctuations over time reinforces the time-diversification argument in that case. Equation (15.13) also allows us to claim that a larger time horizon make the agent more willing to take a small risk, at least when the current cash-on-hand is around the feasible lifetime smooth consumption level, as is the case when the flow of income is constant. But if the cash-on-hand y is large with respect to the feasible smooth consumption c, a longer time horizon makes the consumer implicitly poorer. This is implicitly an income effect. Under DARA it has an adverse effect on his risk tolerance. To be more specific, let us consider the often considered case with w2 = · · · = wn = 0. It implies that y = z, and c is decreasing in the length of the time horizon. In the limit case with βρ = 1, we obtain that Th(y) = nT(y/n). It is not clear here whether a longer time horizon should induce more risk-taking. The problem here is to determine whether nT(y/n) is larger or smaller than T(y). There is a mathematical concept for that problem, which is named subhomogeneity. A function g is subhomogeneous if kg(z/k) > g(z) for all z and all k > 1. Thus, in the case of perfect consumption smoothing (βρ = 1), the subhomogeneity of absolute risk tolerance guarantees the property that, given the same aggregate wealth, the younger investor will take more risk. In the case of CRRA, absolute risk tolerance is homogeneous, which implies that Th(y) = nT(y/n) = T(y): the time-diversification effect is fully compensated by the income effect. This result remains true when βρ ≠ 1. 60 Suppose that the consumer has no future income and that he faces a single risk in time. He has n periods of consumption after having observed the realization of the risk. Then an increase in n PROPOSITION
• has no effect on the optimal exposure to the single risk under CRRA; • has a positive effect on the optimal exposure to the single risk if absolute risk tolerance is convex and subhomogeneous. Propositions 59 and 60 provide a precise content to the concept of time diversification of a single risk in time. In the next chapter we address the more difficult problem of time diversification when consumers face a recurrent risk over time.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_231.html [4/21/2007 1:32:02 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_232.html
Page 232
15.8 Concluding Remark A special case of the consumption problem under certainty is the so-called cake-eating problem, which can be stated in a somewhat unorthodox way as follows: An agent lives for n periods. There is only one state of the world. He is endowed with a cake in each period. The size of the cake can vary over time. He can transfer cakes across time at some rate of decay that can be negative (if the older cakes are better) or positive (if it is a perishable good). What is the consumption plan that maximizes the discounted value of the felicity generated by it? Now let us change some words to obtain the following question: An agent faces n possible states of the world. There is only one period. He is endowed with a cake in each possible state. The size of the cake can vary over different states. He can transfer cakes across states of the world by purchasing insurance contracts with an actuarially favorable or unfavorable premium rate. What is the consumption plan that maximizes the expected utility generated by it? We see that the change in the wording does not affect the mathematical model that is behind them. In fact the maximization problems are strictly equivalent. Since we already characterized the solution to the second problem, which is nothing else than the Arrow-Debreu portfolio problem, we had not much to do in this chapter to characterize the solution to the cake-eating consumption-saving problem under certainty.
15.9 Exercises and Extensions Exercise 78 Prove proposition 57 in a formal way. Exercise 79 Characterize the optimal consumption plan under certainty of an agent with a HARA felicity function. Examine more specifically the cases of power and exponential functions. Exercise 80 Examine the effect of a change in the discount factor on the absolute tolerance of risk on wealth, depending upon whether βρ be larger or smaller than 1. Begin with the HARA case as a benchmark.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_232.html [4/21/2007 1:32:03 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_233.html
Page 233 Exercise 81 Consider a time-inconsistent agent who consumes at three dates: 1, 2, 3. The risk-free rate is zero. He has a cash-on-hand y at t = 1 that he must allocate into consumptions at the three dates. His objective at date 1 is to maximize U1 = u(c1) + βu(c2) + βu(c3), whereas his objective at date 2 is to maximize U2 = u(c2) + βu(c3). In other words, he uses a discount rate of β–1 – 1 > 0 when contemplating consumption substitution between today and tomorrow, but he uses a zero discount rate when contemplating consumption substitution between tomorrow consumption and the day-after-tomorrow consumption. Finally suppose that the CRRA felicity function exhibits CRRA. • What would be the optimal consumption plan • Suppose that the agent consumes the actual consumption path
from the viewpoint of the agent at date 1?
at date 1. At date 2 it happens that
does not maximize U2. Derive
that will be selected by the agent at date 2.
• As above, the agent at date 1 behaves in a myopic way: he does not take into account the fact that he may not implement what he has planned to do in the future. From the viewpoint of the agent, the consumption path is dominated by another consumption plan where the date-1 agent plays a Stackelberg game with date2 agent. Find the equilibrium of that game when date-1 agent is the leader. in the logarithmic case. Show also that it is larger 1. Show that the equilibrium early consumption equals (smaller) than when relative risk aversion is less (larger) than unity. Find an intuition for these results. 2. Show that it dominates
from the viewpoint of date-1 agent.
• (Laibson 1996) What would have happened if the date-1 agent is able to commit itself to future saving (compulsory retirement scheme, purchase of illiquid assets, etc.)? Exercise 82 Consider the following nonseparable intertemporal preferences:
• Interprete k depending on whether it is positive or negative. • Show that the optimal consumptions c1 and c2 are respectively decreasing and increasing with k.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_233.html [4/21/2007 1:32:03 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_234.html
Page 234
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_234.html [4/21/2007 1:32:04 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_235.html
Page 235
16 Precautionary Saving and Prudence Under certainty the optimality of a consumption path depends on two behavioral factors: the degree of impatience of the consumer measured by his discount factor and his aversion to consumption fluctuations measured by the degree of concavity of his felicity function. When some uncertainty affects future incomes, the optimal consumption path will also be influenced by the degree of prudence of the agent. The principal recent innovation in life cycle models has been to introduce uncertainty, thereby allowing for an explicit treatment of the precautionary motive of saving. One of the main benefits is that we can accommodate a much wider range of behaviors in the precautionary model. This has been at the cost of a loss of tractability, as we already observed when we examined the dynamic portfolio problem. It is widely believed that the uncertainty affecting future incomes raises savings. Agents who behave in this way are said to be prudent. This an old idea originating from Keynes and Hicks, but it got the imprimatur of science only in the late 1960s with works by Leland (1968), Sandmo (1970), and Drèze and Modigliani (1972). We now introduce uncertainty in the basic model presented in the previous chapter.
16.1 Prudence We consider here a simple model with two dates t = 0, 1. Under certainty the agent has an income flow (w0, w1), and he selects a consumption plan (c0, c1). His saving in period 1 is thus s = w0 – c0. It allows the agent to consume ρs in addition to his income w1 at date 1. Under certainty the agent has an optimal saving s* that is the solution to problem (16.1):
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_235.html [4/21/2007 1:32:04 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_236.html
Page 236
Because we consider a two-date model, there is no problem of time inconsistency. Thus the analysis does not require that u1(z) = βu0(z). Under the concavity of u0 and u1, the necessary and sufficient condition for s* is written as
By differentiating this condition with respect to s* and w1, it is easily checked that s* is decreasing in w1 and increasing in w0. This expresses the preference for consumption smoothing. In order to determine the impact of a future income risk on optimal saving, we compare the optimal solution of the saving problem with or without this risk. Suppose that at the time of the decision on s, there is uncertainty about the income that will be earned in period 1. Namely the income at date 1 is , not w1. In order to isolate the effect of risk from the effect of smoothing expected consumption that has been analyzed in the . We also assume that this risk cannot be transferred to the market: it is previous sections, we assume that uninsurable. Kimball (1990) introduced the following terminology: DEFINITION
6 An agent is prudent if adding an uninsurable zero-mean risk to his future wealth raises his optimal
saving. We now characterize this concept in the state- and time-separable model. Under uncertainty, the consumptionsaving problem is to maximize
We inquire whether the optimal solution to this problem is larger than s*, the difference being the level of precautionary saving. It is noteworthy that H is concave in s if we assume that the agent has preferences for consumption smoothing, that is, if u0 and u1 are concave. In consequence there is precautionary saving if H′(s*) ≥ 0. By condition (16.2), this is true if
Intuitively, the willingness to save increases when the expected marginal utility of future wealth is increased. To summarize, prudence requires that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_236.html [4/21/2007 1:32:05 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_237.html
Page 237
where z = w1 + ρs*. Because w0, w1, and ρ are arbitrary, so is z. Thus we want this condition to hold for any acceptable z. This condition is formally equivalent to condition (6.20) which defines the notion of risk aversion, except that u1 is replaced by if and only if PROPOSITION
. Using the diffidence theorem yields that condition (16.3) holds for any z and
is concave, that is, if and only if
is convex. This result is due to Leland (1968).
61 An agent is prudent if and only if the marginal utility of future consumption is convex.
Similarly an agent is locally prudent at z if u′″(z). The introduction of a small risk to future income raises savings at date 0 if the agent is locally prudent at c1. Local prudence and local risk aversion are independent concepts in the sense that, at least locally, one could be risk-averse and prudent, risk-averse and imprudent, risklover and prudent, or, risk-lover and imprudent. Kimball (1990) proposes to measure prudence—or the intensity of the precautionary saving motive—by the precautionary equivalent premium ψ. This idea was already introduced in section 9.1 to prove that standardness implies risk vulnerability. The precautionary equivalent premium is the certain reduction in w1 that has the same effect on optimal saving as the addition of random noise to w1. It depends on z, the distribution of , and the degree of convexity of
. We obtain this premium by solving the following equation:
Observe again the equivalence with the symmetric notion in risk aversion, which is the risk premium π defined by condition (2.3). We have that
This equivalence has several important consequences. For example, one can approximate ψ by a formula that parallels the Arrow-Pratt approximation for the risk premium:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_237.html [4/21/2007 1:32:06 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_238.html
Page 238 where P(z) is the index of absolute prudence defined as
One can also define the notion of "more prudence." An agent with utility function on future consumption u1 is more prudent than the one with utility function u2, at all wealth levels, if and only if ψ(z, u1, ) is larger than ψ(z, u2, ) for all (z, ). By proposition 3, this is true if and only if –
is larger than –
for
all z. Finally, from proposition 4, we obtain that ψ is decreasing in z if and only if P is decreasing in z. This notion is called decreasing absolute prudence, which states that the sensitivity of saving to future risks is smaller for wealthier people. PROPOSITION
62 Define the precautionary equivalent premium ψ(z, u1, ) by condition (16.4) and absolute
prudence P by condition (16.5). They satisfy the following properties: 1. ψ(z, u1, ) is nonnegative for all (z, ) if and only if P(z) is nonnegative for all z. 2. ψ(z, u1, ) is larger than ψ(z, u2, ) for all (z, ) if and only if –
is larger than –
for
all z. 3. ψ(z, u1, ) is decreasing in z for all
if and only if P is decreasing in z.
Is prudence an assumption as realistic as risk aversion or decreasing absolute risk aversion? There are two main arguments in favor of prudence. First, many empirical studies have been conducted to verify whether people that are more exposed to future income risks save more. The results are overwhelmingly in favor of this assumption. See, for example, Guiso, Jappelli and Terlizesse (1996) and Browning and Lusardi (1996). The second argument is based on the property that
which implies that prudence is necessary for the widely accepted view that absolute risk aversion is decreasing. There is another importnat link between risk aversion and prudence that has been presented in proposition 21: if absolute prudence is uniformly decreasing, then absolute risk aversion must also be uniformly decreasing.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_238.html [4/21/2007 1:32:06 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_239.html
Page 239
16.2 The Demand for Saving. It is useful to examine the effect of a change in the risk free rate on the optimal saving under uncertainty. The first-order condition for the maximization of H is written as
Fully differentiating this condition yields
Since the bracketed term in the left-hand side of the inequality above is negative, we see that ds/dρ has the sign of . The term expresses the positive substitution effect, whereas the term is a wealth effect: if s is positive, an increase in the interest rate raises future wealth, thereby reducing the willingness to save. This effect disappears if the net saving s(ρ) is zero. More generally, the wealth effect reinforces the substitution effect if the optimal saving is nonpositive. If the optimal saving is positive, the global effect of an increase in the interest rate is ambiguous. Notice, however, that we can rewrite the right-hand side of equation (16.7) as
where R is the index of relative risk aversion, and . Thus, if R is uniformly less than 1, and if is nonnegative almost surely, then this expression is positive, and the optimal savings is increasing with the interest rate. The following proposition extends proposition 56: 63 Suppose that the future income is nonnegative almost surely. Then the optimal saving is increasing with the interest rate if relative risk aversion is less than unity. PROPOSITION
16.3 The Marginal Propensity to Consume under Uncertainty Up to now, we have examined how a future risk affects the level of current saving at a given level of wealth. Another question is to determine how a future risk with affects the sensitivity of saving with respect to changes in wealth, that is, the marginal
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_239.html [4/21/2007 1:32:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_240.html
Page 240 propensity to consume (MPC). Consider again the two-period saving problem under uncertainty:
where is the usual indirect utility function with an exogenous background risk. Observe that, for simplicity, we restrict our analysis here to the case where the felicity function is time-independent. We also normalized w1 to zero. Let c*(w0) = w0 – s*(w0) be the optimal consumption under uncertainty. Having elluded the background risk in using this technique, we have turned this problem into a particular case of the portfolio problem with complete markets, with a state-dependent utility function. We hereafter examine two questions. The first one is about the effect of uncertainty on the marginal propensity to consume. The second one is about the effect on its derivative with respect to wealth. 16.3.1 Does Uncertainty Increase the MPC? From conditions (14.10) and (14.11a) the marginal propensity to consume out of wealth under uncertainty equals
We compare it to the marginal propensity to consume under certainty. The consumption problem under certainty is written as
which yields an optimal consumption
. We know that
is smaller than c*(w0) under
prudence. The marginal propensity to consume under certainty equals
Observe that the MPC is increasing with the current risk tolerance and decreasing with the future risk tolerance. The comparison of equation (16.9) and (16.10) shows that there is little hope to obtain an unambiguous answer to the question of whether uncertainty increases the MPC. Indeed, there are two contradictory effects of uncertainty on the MPC. The first effect is due to the direct impact
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_240.html [4/21/2007 1:32:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_241.html
Page 241 of the risk on the future risk tolerance. Under risk vulnerability it reduces future risk tolerance; that is, Tv(z) is less than T(z) for all z. This raises the MPC. The second effect is indirect: under prudence the risk reduces the current consumption, and it increases the expected future consumption. Under DARA the current risk tolerance is reduced, but future risk tolerance increases. This means that the MPC is reduced. It is unclear which of the two effects dominates the other. A particular case is where the effect of uncertainty on the MPC is unambiguous. Let us assume that β = ρ = 1; here the MPC is obviously 0.5 under certainty. From condition (16.9) the MPC under uncertainty is larger than . We conclude that the uncertainty that if and only if T(c*) is larger than Tv(w – c*). Let denote increases the MPC if and only if the following property holds:
The equality to the left is the first-order condition for c*. The condition to the right is a rewriting of T(c*) ≥ Tv(w – c*). In words, does an expected-marginal-utility-preserving risk raises local risk aversion? This is equivalent to
We know that the answer is positive if and only if absolute prudence is decreasing, that is, whenever u″ is more concave than –u′. 64 Suppose that β = ρ = 1. The marginal propensity to consume is increased by the introduction of a zero-mean risk to future incomes if risk aversion is standard. PROPOSITION
Kimball (1998) provides other insights on the effect of risk on the MPC. 16.3.2 Does Uncertainty Make the MPC Decreasing in Wealth? We showed in proposition 58 that the marginal propensity to consume is decreasing with wealth under certainty if and only if
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_241.html [4/21/2007 1:32:08 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_242.html
Page 242 absolute risk tolerance is convex. On the contrary, if absolute risk tolerance is linear, then the marginal propensity to consume under certainty is not sensitive to a change in wealth. Carroll and Kimball (1996) showed that introducing uncertainty into the basic model will concavify the consumption function of HARA agents. The explanation of this phenomenon is obtained by combining results contained in propositions 23 and 52. By proposition 52, we know that the consumption in the first period will be concave in wealth if T′(w0 – s*) is smaller than
. The fact that this condition holds here is proved in two steps. First, we know that . Indeed, a HARA utility function is such that T′(z) = 1/γ, or
23 implies that
PROPOSITION
Proposition
or equivalently for all z. In particular, we have that . Second, because T′ is a constant in the HARA case, we obtain that . This concludes the proof of the Carroll-Kimball's result.
65 Under HARA the optimal consumption rule is concave in wealth, or s*″(w0) ≥ 0.
The reader can easily extend this result to the case where absolute risk tolerance is convex and expected consumption is increasing through time.
16.4 More Than Two Periods 16.4.1 The Euler Equation A dynamic version of the model is obtained by assuming that the labor income flow of the agent follows a random walk. Let be the income at date t = 0, 1, 2, . . . , n, with being independently distributed. Assuming a time-independent felicity function u, the optimal consumption strategy is obtained by backward induction. At the last date t = n, the agent consumes its remaining wealth. At date t = n – 1, after observing his income, his decision problem is written as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_242.html [4/21/2007 1:32:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_243.html
Page 243 where y is his "cash-on-hand" at t = n – 1, which is the sum of his accumulated financial wealth at that time and its current labor income. Solving this problem for each y yields the optimal state-dependent consumption cn–1(.) together with the value function of wealth vn–1(.). This is done by writing the first-order condition:
Observe also that the envelope theorem applied to program (16.13) yields
Now one can go one step backward by solving the decision problem at date t = n – 2. The decision problem at that date is written as
The first-order condition to this problem is written as
Using property (16.15), we see that this is equivalent to
where cn–2 and
are the optimal consumptions at the corresponding dates conditional to a given cash-on-hand
at t = n – 2. Condition (16.18) is often called the Euler equation, which states that the conditional (discounted) expectations of marginal felicity must be equalized over time. There is a simple intuition behind the Euler equation, as it just states that any small increase in saving at n – 2 used for increasing the consumption at n – 1 has no effect on the lifetime expected utility. More generally, at date t = 0, 1, . . . , n – 1, the decision problem is as follows:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_243.html (1 of 2) [4/21/2007 1:32:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_243.html
with vn ≡ u. The Euler equation is obtained using the same method as above. It yields
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_243.html (2 of 2) [4/21/2007 1:32:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_244.html
Page 244 which yields the optimal consumption ct(z) at t as a function of the cash-on-hand y at that date. 16.4.2 Multiperiod Precautionary Saving. The extension of the basic saving problem to more than two periods raises several questions. Some of them are still unexplored. This extension of the model implies that one must determine which of the properties of u are inherited by the value function vn–1. Because the vn–1 function combines an optimization operator with an expectation operator, we have propositions 23 and 53 to assist us in solving the various problems related to this extension. The first obvious problem is whether the uncertainty on the income at t + 1 increases saving at t. The question has been answered for t = n – 1. We just need u to be prudent for this. Let us go one date backward at t = n – 2. increases savings at Building on the equivalence between programs (16.1) and (16.16), the existence of risk t = n – 2 if and only if vn–1 is prudent. This is true if u itself is prudent. Indeed, this is a consequence of the fact that the maximization operator preserves prudence (proposition 53), together with the property that is positive if u′″ is positive. By recurrence, vt is prudent for all t. Our numerical example is based on the assumption that β = ρ = 1, which implies that perfect consumption smoothing is optimal under certainty. The optimal consumption policy is given by : consumption is set equal to the mean income plus a fraction of cash-onhand above the mean income. This strategy in the range y ∈ [5, 25] and t = n – 1, n – 2, n – 3, n – 4 is represented are random and i.i.d. by the dotted lines in figure 16.1, when w = 10. Suppose, alternatively, that More precisely, suppose that the labor income at each period is either 5 or 15 with equal probabilities. The optimal consumption strategy is represented in the same graphics by using plain curves when the felicity function is CRRA with a relative risk aversion of 2. We see that the uncertainty on the flow of incomes reduces the consumption at all levels of wealth and at all dates. Notice that all along this chapter we assumed that the agent is in a position to borrow and lend at rate ρ without any constraint. In particular, when the agent's cash-in-hand is small, he can go to the bank to borrow money in the expectation of reimbursing it with his future
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_244.html [4/21/2007 1:32:10 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_245.html
Page 245
Figure 16.1 Optimal consumption as function of cash-on-hand
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_245.html [4/21/2007 1:32:10 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_246.html
Page 246 incomes. In the real world this strategy is limited because of limited liability on the borrower side, which implies that banks are reluctant to borrow money to finance consumption. This implies that the consumers face a liquidity constraint. The effect of the liquidity constraint on the optimal consumption-saving behavior is examined in chapter 18.
16.5 Illiquid Saving under Uncertainty Up to now, we supposed that the decision on saving can be made after observing the income earned at each date. This is a natural assumption in a flexible economy where agents can cash on their savings whenever an adverse shock occurs on their incomes. However, a large part of households' savings is put in long-term funds in which withdrawal are costly. For various reasons, tax incentives are attached to long-term saving instruments, that make withdrawals very difficult or impossible. Good examples are mortgages and the money saved for retirement. Consider an extremely rigid world in which the long-term saving plan must be chosen at an early age. It implies that households consume at every period the difference between their actual income and the prespecified contribution to their retirement plan. In such a world, saving looses most of its capacity to forearm households against shocks to their incomes. The problem is to determine the optimal saving in such a rigid economy. More precisely, we want to determine whether the presence of risk on the flow of incomes affect the level of long-term saving. The benchmark is again the two-period saving problem (16.1) under certainty. In this section we need to assume exponential discounting: u1 = βu0. Let us now assume that the agent faces serially independent shocks on his income. Namely his income at date t is , where and are independent and identically distributed and . The optimal long-term saving in this uncertain rigid economy is the solution to the following program:
As usual, one can use the indirect utility function problem as follows:
in order to rewrite this
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_246.html [4/21/2007 1:32:11 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_247.html
Page 247 This problem is formally equivalent to the initial problem (16.1), except for the replacement of the original felicity function u0 = u1 by the indirect felicity function v. We know from chapter 9 that v is more concave than u0 if the original felicity function exhibits risk vulnerability. In this context more concavity means a larger fluctuation aversion. If we assume the normal case βρ > 1, a larger fluctuation aversion implies a smaller saving, according to proposition 57. The symmetric case is proved in the same way. 66 Suppose that the felicity function exhibits risk vulnerability. The presence of risk on the flow of income reduces (resp. increases) long-term illiquid saving in the rigid economy if βρ is larger (resp. smaller) than unity. PROPOSITION
In simpler words, the presence of risk on incomes makes the agent more resistant to intertemporal substitution. It is noteworthy that in the normal case we obtain exactly the opposite result than in the previous section: here, risk reduces saving. Whether the rigid economy that we examined here is a better representation of the real world than the flexible economy presented in the previous section is an empirical question that remains to be addressed. There is no doubt that a better model would be in between these two extreme models.
16.6 Conclusion In this chapter we addressed the question of how does a risk on the income flow affect the optimal level of saving. We gave some flesh to the assumption that human beings are prudent, namely that the third derivative of their utility function is positive. Otherwise, the presence of a risk on their future income would have the counterintuitive effect to reduce the optimal level of saving. We also discussed the idea that human beings have decreasing absolute prudence. Otherwise, wealthier agents would react more than poorer ones to the presence of a risk on future income. This also seems to be counterintuitive. The uncertainty on future income does not only affect the level of savings. It also affects the sensitivity of savings to changes in wealth, that is, it affects the marginal propensity to consume out of wealth. The effect of uncertainty on the MPC is unclear, but we know that it makes it an increasing function of wealth in the HARA case. Finally,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_247.html [4/21/2007 1:32:12 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_248.html
Page 248 the analysis of precautionary savings in the n-period case requires numerical simulations to solve the Euler equation.
16.7 Exercises and Extensions Exercise 83 We have shown that the introduction of a zero-mean risk to future incomes raises savings if and only if the agent is prudent. What is the condition to guarantee more generally that the agent increases his savings whenever his future income becomes riskier in the sense of Rothschild-Stiglitz? Exercise 84 Prove that absolute prudence may be uniformly decreasing only if it is uniformly positive. Exercise 85 Show that risk on future incomes has no effect on savings if the utility function is quadratic. Exercise 86 Consider the value function vn–1 defined by (16.13). Show that it is DARA if u is DARA. Exercise 87 Consider the (n + 1)-period consumption problem
under the budget constraint of consumption at all dates t < n.
. Show that the introduction of risk
with
reduces the level
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_248.html [4/21/2007 1:32:12 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_249.html
Page 249
17 The Equilibrium Price of Time As the saying goes, "Time is money." In this chapter we determine the equilibrium price of time, which is the equilibrium interest rate in the economy. There are two explanations for why time is money. On one hand, agents are impatient, and they often expect to have larger incomes in the future than in the present. On the other hand, firms have investment opportunities, some of them with a positive return. Therefore, if the interest rate were zero, consumers and entrepreneurs would borrow money like crazy. That may not be an equilibrium. Because of the excess demand for money, the market will react by increasing the interest rate. The level of the risk-free rate is one of the most important instruments used to control the level of activity of the economy. It affects the consumer's willingness to save and the entrepreneur's willingness to invest. It is the variable that organizes a balance between current and future pleasures. In this sense it determines the growth of the economy. Through its use as a discount factor to compute the net present value of potential investments, it guides efforts for improving the future by society. Let us take a specific example to illustrate this. At the current rate of emission of greenhouse gases on earth, there will be enough of such gases accumulated in the atmosphere in, say, 50 to 200 years from now, to generate a potentially devastating global warming on earth. Limiting industrial emissions could have a potentially large adverse effect on GDP. Suppose that we know perfectly the size of these damages. At the conferences of Rio and Kyoto, environmentalists discussed how much effort should be effected today to reduce greenhouse damages in the future. Notice that there are two ways to improve the welfare of future generations: either we attempt to limit the emissions in order to reduce future damages, or we increase
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_249.html [4/21/2007 1:32:13 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_250.html
Page 250 Table 17.1 Present value of one million in t years with discount rate r (in % per year) Percent
50 years
100 years
200 years
500 years
1 2 5 8
608,039 371,528 87,203 21,321
369,711 138,032 7,605 454
136,686 19,053 57 0.2
6,907 50 2.5 10–5 1.9 10–11
physical investments to increase the stock of capital available for these generations to compensate for damages generated by the global warming. This translates into our making the financial committment to reduce emissions only if this strategy dominates the strategy to invest more in physical capital. Obviously we should do this only if the internal rate of return of reducing our emissions of greenhouse gases exceeds the rate of return of risk-free physical investment. This is the argument one uses to justify net present value in cost-benefit analyses. The discount rate must be equalized to the equilibrium return of risk-free investments, that is, to the risk-free rate. The present value of a damage D in t years equals PV = D(1 + r)–t, where r is the risk-free rate. If a technology is available today to eliminate this future damage, it should be effected only if its present cost is less than PV. In table 17.1 we compute PV for different values of r and t, with a real damage D of one million. The exponential effect of discounting is clear. For example, one should spend no more than twenty cents this year to eliminate a one-million damage happening 200 years from now if one uses a discount rate of 8%. Notice that 8% is considered as an acceptable discount rate. Many countries, including the United States, France, and Germany, recommend the use of a discount rate between 5% and 8% for their public investment policy. Clearly, we have a problem if we use such a large discount rate for long horizons. Why is the value of a far distant benefit so low? What is the socially efficient interest rate? To answer these questions, we need to examine the determinants of the interest rate. To do this, we will examine a "tree economy" à la Lucas (1978).
17.1 Description of the Economy There is a fixed set of identical consumers who live from date 0 to T. At the origin each economic agent is endowed with one tree of size
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_250.html [4/21/2007 1:32:13 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_251.html
Page 251 z0. All trees have the same size zt at each date t. At each date a tree produces a quantity of fruits that is proportional to its current size. There is no possibility to plant additional trees in the economy, meaning that the stock of capital is exogenous. Fruits are perishable, namely they cannot be transferred from one period to the other: they must be consumed instantaneously. Thus z can also be interpreted as the GDP per capita in this economy. Agents extract utility by consuming fruits, which are the numéraire. The grow rate of the size of the tree from date t – 1 to t equals . We assume that there is a perfect correlation in the growth rate of the different trees in the economy during a specific period, but there is no correlation in the growth rate of a tree in different periods. Consumers are allowed to borrow and lend fruits over time. The equilibrium exchange rate for fruits tomorrow against fruits today is the equilibrium price of time. Let vt(z, b) be the value function at date t of an agent who has a tree of size z at that date and who has to pay back b fruits from his short-term debt. With a time-separable utility function, a consumer with a tree of size z and b fruits to reimburse at date t would choose his consumption of fruits ct as follows:
with vT(z, b) = u(z – b). At date t, the agent gets z fruits from his tree, he consumes ct fruits, he reimburses his short-term debt b, and he borrows the difference ct + b – z at the gross risk free rate ρt. Next period, the size of his tree will be
, and he will have to reimburse ρt(ct + b – z). The optimal consumption strategy is
characterized by function ct(z, b). The first-order condition to this problem is written as
where
is the derivative of vt+1 with respect to its second argument.
In this model with undiversifiable risks and with agents that are identical in their preferences and in their endowment, the autharky is a competitive equilibrium: [ct(z, b) = z – b for all z, b and t] is the market-clearing condition on the credit market at date t. The agent consumes the crop of his tree at each date, after having reimbursed of his debt, if any. That means that agents never borrow fruit at equilibrium under any circumstances. Because the agent
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_251.html [4/21/2007 1:32:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_252.html
Page 252 consumes his crop at each date, the value function takes the following form:
where
is the size of the tree at date τ. This implies that
The risk-free rate that sustains this equilibrium is obtained by using the first-order condition (17.2), together with the characterization (17.4) of the derivative of the value function. It yields
Notice that the risk-free rate is a function of the state of the economy at the corresponding date. In this simple economy, the state is given by the size z of the trees, namely by the current GDP per capita.
17.2 The Determinants of the Interest Rate. The short-term interest rate ρt(z) given by condition (17.5) is an equilibrium because it makes individual consumption plans compatible with each other in the economy. Three components are at play in the determination of the price of time that generates this outcome: impatience, the willingness to smooth consumption over time, and precautionary saving. These three components can be introduced separately by looking at specific distributions for , as we do now. 17.2.1 The Interest Rate in the Absence of Growth We can isolate the effect of impatience on the interest rate by assuming that there will be no growth. When almost surely, the equilibrium risk-free rate ρt is equal to 1/β. It is generally assumed that agents are impatient, namely that β is less than unity. This means that agents value future utility (utils) less than current ones. In the absence of growth, the discount rate on utils is the same than the discount rate on the numéraire.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_252.html [4/21/2007 1:32:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_253.html
Page 253 With a zero interest rate on credit markets, agents will want to borrow to increase current consumption at the cost of a smaller consumption later on, just because they are impatient to consume. Since all agents will do that, this cannot be an equilibrium. Impatience must be counterbalanced by a positive interest rate to reduce the willingness to borrow. Indeed, we have seen in section 16.2 saving that the optimal saving is increasing in the interest rate if the optimal saving is zero. Arrow (1994) suggests that this argument of pure preference for the present yields an interest rate of one to two percents. 17.2.2 The Effect of a Sure Growth When
takes a specific positive value gt+1 > 0 almost surely, the discount rate on the numéraire is not equal
anymore to the discount rate on utility: an argument of income smoothing enters into the picture. By expecting larger revenues in the future, agents want to borrow today in order to smooth consumption over time. An additional increase in the interest rate is necessary to induce agents not to borrow. The equilibrium interest rate equals
which is larger than 1/β when u is concave and gt+1 is positive. An increase in the concavity of the utility function raises the willingness to smooth consumption over time. When the growth is positive, this means that agents want to save less. The market will react by raising the interest rate. On the contrary, when the growth is negative, this means that agents want to save more. It implies a reduction of the equilibrium risk-free rate. 67 Suppose that there is no uncertainty on the future growth of the economy. Then an increase in the concavity of the representative agent's utility function PROPOSITION
• increases the equilibrium risk-free rate if the growth is positive (βρ > 1), • has no effect on the equilibrium risk-free rate if there no growth (βρ = 1), • reduces the equilibrium risk-free rate if the growth is negative (βρ < 1).
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_253.html [4/21/2007 1:32:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_254.html
Page 254 If gt+1 is small, a first-order Taylor approximation of the denominator around z yields
where R(z) = –zu″(z)/u′(z) is the relative fluctuation aversion. We see that the impact of a sure growth gt+1 on the equilibrium interest rate is approximately equal to the product of gt+1 by R(z). With an historical growth rate of real GDP per capita around 2% per year, and with R between 1 and 4, we obtain an effect increasing the interest rate by 2% to 8%. It is noteworthy that with a growth rate of 2% per year forever, the real GDP per capita 200 years from now will be 52 times larger than today. This puts a perspective to table 17.1: we should not make too much effort to improve the GDP of future generations because they will be so much wealthier than us anyway. This is a wealth effect. Because a constant growth rate of the economy means that the GDP per capita growth exponentially, we understand why the net present value (NPV) of a future cost or benefit should decrease exponentially with its maturity. However, this effect heavily relies on the assumption that the growth rate is known with certainty, even for distant futures. This assumption is hard to sustain. 17.2.3 The Effect of Uncertainty We now turn to the general case where is a nondegenerated random variable, which is a more realistic reduces the riskassumption. We see immediately from equation (17.5) that the presence of uncertainty on free rate ρt if and only if u′ is convex. The intuition is straight forward: when future incomes are uncertain, agents want to save for the precautionary motive. This must be compensated by a reduction of the risk-free rate. One way to quantify the effect of the uncertain growth on the interest rate is to define the "precautionary equivalent" growth rate, the certain growth rate that yields the same interest rate. This should not be confused with the certainty equivalent growth rate, the certain growth rate that generates the same expected utility of the representative agent in the future. The precautionary equivalent growth rate
is implicitly defined as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_254.html [4/21/2007 1:32:16 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_255.html
Page 255 Following Kimball (1990), the precautionary equivalent growth rate can be approximated by
Condition (17.9) indicates that the effect of the uncertainty on growth on the interest rate is the same as a sure reduction of the growth rate by the product of half its variance by relative prudence. The more prudent the people, the smaller is the precautionary equivalent growth rate, and the smaller is the equilibrium interest rate. If the precautionary effect is larger than the wealth effect combined with the effect of impatience, the equilibrium interest rate will be negative. To sum up, two different characteristics of the utility function affect the level of the equilibrium risk-free rate. A first concerns the effect of uncertainty. Which certain growth rate should we consider as equivalent to the uncertain growth rate we face? We have shown that the degree of relative prudence determines the impact of the riskiness of growth on the precautionary equivalent growth rate. The more prudent we are, the smaller will be the equivalent certain growth rate. Prudence is measured by the degree of convexity of u′. The second question is to determine by how much we should substitute consumption today by consumption tomorrow in the face of this precautionary equivalent growth rate. This depends on the degree of fluctuation aversion. It is measured by the degree of concavity of u. The more resistant to intertemporal substitution we are, the larger will be the interest rate in order to equilibrate the money market. Clearly, one can isolate the precautionary effect of a change in the utility function by assuming that the initial precautionary equivalent growth is zero so that βρ = 1. From proposition 67 we know that there is no consumption smoothing effect of a change in the utility function under such circumstances. We obtain the following result: 68 Suppose that the precautionary equivalent growth defined by equation (17.8) is zero, which is the case when βρ = 1. Then an increase in the degree of prudence will reduce the risk-free rate.
PROPOSITION
We can easily use proposition 67 to extend this result to the case where the precautionary equivalent growth is not zero. For example, if the precautionary equivalent growth is positive, a change in u
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_255.html [4/21/2007 1:32:17 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_256.html
Page 256 reduces the risk-free rate if u becomes less concave and u′ becomes more convex. This can be seen more explicitly by combining conditions (17.7) and (17.9). This yields
Hansen and Singleton (1983) obtained a similar formula using power utility functions and lognormal growth in continuous time. This is not a surprise. Indeed, we know that approximation (17.9) is exact in this case. In addition the size of growth is proportional to the size of the interval of time, implying that approximation (17.7) is exact at the continuous-time limit. An advantage of this formula is that it includes the three determinants of the equilibrium risk-free rate, with the three terms in the right-hand side of the equality above being respectively the pure preference for the present, the consumption smoothing effect, and the precautionary effect.
17.3 The Risk-Free Rate Puzzle We can calibrate this model by using CRRA utility functions for which –zu″(z)/u′(z) = γ and –zu′″(z)/u″(z) = γ + 1. Notice that CRRA implies that the risk-free rate is independent of the wealth level: it is scale invariant. We use data for the US growth of GDP per capita over the period 1963 to 1992, for which we have and
. For β = 1, condition (17.10) is rewritten as
The first term of right-hand side is the consumption smoothing effect, whereas the second term is the . Thus the precautionary effect. It makes the right-hand-side bell-shaped function with a maximum at equilibrium risk-free rate is increasing in the degree of risk aversion as long as γ is less than 31. Then the riskfree rate is decreasing in γ to eventually become negative when relative risk aversion is larger than 63. For γ = 1, the Hansen-Singleton formula yields an equilibrium risk-free rate approximately equal to 1.80%, whereas for γ = 4, it yields a risk-free rate around 6.86%. It should be noticed that the precautionary effect is obviously a secondorder effect in the realistic range of γ. For example, for γ = 1, the effect of uncertainty is to reduce the equilibrium rate from 1.83% to 1.80%.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_256.html [4/21/2007 1:32:17 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_257.html
Page 257 If we accept the idea that reasonable values of relative risk aversion are between 1 and 4, and if we take into account of a pure preference for the present around 1%, we obtain an equilibrium risk-free rate in an interval between 2.80% and 7.86%. This should be compared to the historical risk-free rate that fluctuated around an average value of 1% over the last century (see figure 5.2). Thus, following Weil (1989), we conclude that the theoretical model overpredicts the equilibrium risk-free rate. The growth of consumption has been so large and its volatility has been so low that the model is unable to explain why households saved so much during this period. A much larger risk-free rate than the actual one would have been necessary to explain the actual consumption growth. This is the risk-free rate puzzle. Another interesting point is to compare the risk free rate puzzle to the equity premium puzzle. We showed in chapter 5 that solving the equity premium requires a large γ around 40. But this choice of γ will yield an equilibrium annual risk-free rate around 25%. In effect, we need a low γ to solve the risk-free rate puzzle, whereas we need a large γ to solve the equity premium puzzle. Solving the two puzzles simultaneously would require a large γ together with a negative rate of pure preference (β » 1). This puzzle illustrates the lack of flexibility either of the expected utility model, or of the utility functions with constant relative risk aversion. In chapter 20, we explore a generalized expected utility model that is more flexible. The so-called recursive utility model explored there allows us to construct theoretical preferences with a small aversion to consumption fluctuations (to solve the risk free rate puzzle), and a large aversion to risk (to solve the equity premium puzzle). Alternatively, we can relax the assumption that relative risk aversion is constant with wealth. Considering other utility functions will make it possible to change the relation between risk aversion and prudence, which is linked by condition –zu′″/u″ = (–zu″/u′) + 1 in the CRRA case. In order to explore this possible solution, let us consider the more general set of HARA utility functions with
for γ > 0 and k is some fixed minimum level of subsistence. Note that relative risk aversion and relative prudence are increasing in k. Moreover the difference between risk aversion and prudence is decreasing with k. Normalizing z0 to unity, the Hansen-Singleton formula (17.10) becomes
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_257.html [4/21/2007 1:32:18 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_258.html
Page 258
Figure 17.1 Equilibrium risk-free rate when u′(z) = (z – k)–γ and z0 = 1
Recall that the second term of the right-hand side of the equality above represents the consumption smoothing effect, whereas the third term is the precautionary effect. When γ is small, only the consumption smoothing effect matters. Since the relative fluctuation aversion is increasing in k, we obtain that the equilibrium risk-free rate is increasing in the relative level of minimum subsistence. This is only when γ is large that the precautionary effect enters into the picture. Because relative prudence is increasing in k, a positive k reduces the certainty equivalent growth rate that can eventually become negative. We depict the relationship between the risk-free rate and γ in figure 17.1, for different values of the relative level of minimum subsistence. From this figure we see that allowing k to be different from 0 is not really helpful to solve the risk free rate puzzle. Moreover a negative minimum level of subsistence would be necessary.
17.4 The Yield Curve 17.4.1 The Pricing Formula Up to now we limited the analysis to the short-term interest rate at date t. This is the return at t of a zero-coupon bond that makes
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_258.html [4/21/2007 1:32:19 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_259.html
Page 259 a single payment at time t + 1. More generally, we can define the interest rate at time t of a zero-coupon bond maturing at time t + n. The gross return per period of such an asset is denoted ytn. At equilibrium, it must satisfy the following condition:
where z is the current consumption per capita. The right-hand side of this expression is the marginal utility cost of consuming one real dollar less at time t. Suppose that this additional dollar is invested in the zero-coupon bond with maturity at t + n. It will yield an additional income at that date that is equal to (ytn(z))n. The left-hand side is thus the discounted expected marginal utility benefit generated by consuming this additional income at t + n. At equilibrium the discounted marginal benefit must equal the marginal cost, meaning that condition (17.14) must be satisfied. This condition is rewritten as
When n = 1, we get back formula (17.5) with yt1(z) = ρt(z). In general, the interest rate of a zero-coupon bond will depend on its maturity, that is, ytn(z) ≠ ρt(z) for n > 1. The term structure of interest rates is the set of {ytn}n=1,2, . . . at a given time t. The yield curve is a plot of the term structure, that is, a plot of ytn against n. The observed yield curve is most commonly upward-sloping over a short time horizon, but it may happen that it is inverted, sloping down over some or all maturities. Brown and Schaefer (1997) report, for example, that over the period 1985 to 1994, the implied interest rate for a maturity at 25 years was lower than the interest rate for maturity at 15 years on 98.4% on the market for US treasury bonds. There is a wide literature on the equilibrium form of the yield curve. The most cited references on this topic are Vasicek (1977), Cox, Ingersoll, and Ross (1985a, b), and Longstaff and Schwartz (1990). The form of the yield curve is a complex function of the attitude toward risk and time, and of the statistical relationships that may exist in the temporal growth rates of the economy. The properties of the yield curve are not yet completely understood, even in the HARA case. In this section we assume that there is no serial correlation in the growth of the economy. This is an unrealistic assumption
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_259.html [4/21/2007 1:32:19 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_260.html
Page 260 because the instantaneous growth rate at t and t + n are in general influenced by the same underlying factors. It implies that and are correlated in the real world. We also assume that , . . . , are identically distributed. This implies that the interest rate ytn(z) = yn(z) is independent of t. The structure of interest rate is governed by rules that link future short-term rates to the current long-term rate. They can be derived either directly from equation (17.15) or by using an arbitrage argument. We have the following condition:
This shows that the long-term interest rate is a complex averaging of the current short-term interest rate and the future short-term interest rate. This equation will be used to determine whether y2(z) is smaller than y1(z), that is, whether the yield curve is decreasing. Also we can directly compare yield curve is decreasing if and only if
and
in condition (17.15) to see that the
Most results presented below are extracted from Gollier (1999). 17.4.2 The Yield Curve with HARA Utility Functions The simplest result is obtained with CRRA functions. When u′(z) = z–γ, we obtain that
The yield curve is flat when the representative agent has CRRA preferences. There is a simple intuition for that. For the CRRA function, the short-term interest rate is independent of GDP: y1(z) = y1. This implies that the future short-term interest rates are perfectly known even if there is some uncertainty about the growth rate of the economy. Moreover the future rates are equal to the current shortfile:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_260.html [4/21/2007 1:32:20 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_261.html
Page 261 term interest rate. By a simple arbitrage argument, or in using equation (17.16), we find that the long-term interest rate must be equal to the short-term one. This property cannot be extended to other utility functions in the HARA class. To illustrate, we take a family of HARA functions (17.12) where γ = 1. We consider a scenario where the growth rate of the GDP per capita is either –0.55% or 4.27% with equal probabily. This random variable has an expectation of 1.86% and a standard deviation of 2.41%, which is the mean and standard deviation of the yearly US growth rate of GDP per capita over the period 1963 to 1992. Current consumption is normalized to unity. We compute the interest rate yn for horizons up to n = 45 years, using various values of the parameter k. The results are drawn in figure 17.2. As a benchmark, we take k = 0.5 for which relative risk aversion equals 2 at current consumption and goes to 1 as consumption goes to infinity. The interest rate for a zero-coupon bond with maturity of 1 year is 3.50%. A zero-coupon bond with a maturity in 45 years has an equilibrium rate of return of only 1.80% per year. 17.4.3 A Result When There Is No Risk of Recession. For the sake of a simple notation, let xt denote the gross rate of growth 1 + gt. Let function h from
to R be
defined as h(x1, x2) = u′(zx1x2). Suppose that this function is log supermodular. This means that
for all (x1, x2) and
for all
for all
in
. In the case
the log supermodularity of h implies that
and x2 that are both larger than 1. This inequality is equivalent to
and g2 = x2 – 1 that are both positive. Suppose now that the growth rate
per period is
positive almost surely. The
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_261.html [4/21/2007 1:32:21 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_262.html
Page 262
Figure 17.2 Yield curve for u′(z) = (z – k)–1, β = 1 and
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_262.html [4/21/2007 1:32:21 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_263.html
Page 263 log supermodularity of h is sufficient to guarantee that condition (17.21) holds almost everywhere. Taking the expectation of this inequality directly yields inequality (17.17), which implies in turn that y2 is less than y1. If the utility function is three time differentiable, the log supermodularity of h also means that the cross derivative of log h is positive. It is easily seen that this is equivalent to the case where the relative risk aversion is decreasing (DRRA). Notice that all inequalities from (17.17) to (17.21) are reversed if relative risk aversion is increasing. 69 Suppose that the growth rate of consumption is nonnegative almost surely. The long-term discount rate is smaller (resp. larger) than the short-term rate if relative risk aversion is decreasing (resp. increasing). PROPOSITION
There is a simple reason why, in the absence of recession, the yield curve is decreasing when relative risk aversion is decreasing. The combination of these two conditions implies that the future short-term rate is decreasing in GDP per capita. Indeed, from (17.16), we have that
Clearly,
is negative if x2 is larger than unity almost surely and R′ ≤ 0. Now this implies that the future short-
term interest rate will be smaller almost surely, since we have assumed that the GDP per capita in the future will be at least as large as that today. By the standard arbitrage argument, it must be the case that the current longterm rate is smaller than the current short term one. Notice that we require that both and be nonnegative to get the result. The decreasing relative risk aversion is compatible with some well-documented observations that the relative share of wealth invested in risky assets is an increasing function of wealth. Kessler and Wolf (1991), for example, show that the portfolios of US households with low wealth contains a disproportionately large share of risk-free assets. Measuring by wealth, over 80% of the lowest quintile's portfolio was in liquid assets, whereas the highest quintile held less than
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_263.html [4/21/2007 1:32:22 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_264.html
Page 264 15% in such assets. Guiso, Jappelli, and Terlizzese (1996), using a cross section of Italian households, observed some portfolio compositions that are compatible with decreasing relative risk aversion. Ogaki and Zhang (1999) tested whether relative risk aversion is decreasing or increasing from various consumption data in developing countries. All these studies obtained strong evidence that relative risk aversion is decreasing. This suggests that the yield curve should be decreasing in an economy without serial correlation in the growth rate of GDP per capita, without any friction on the credit market, and without any risk of recession. The numerical examples provides some hope that the result above can be adapted to economies facing a risk of recession. Growth risks are mutually aggravating. 17.4.4 Exploring the Slope of the Yield Curve When There Is a Risk of Recession DRRA is not sufficient for a decreasing yield curve when there is a risk of recession. To illustrate, let us assume that βy1(z0) = 1: the precautionary effect just compensates the wealth effect, and this implies that the short-term interest rate is equal to the rate of pure preference for the present. This also means that This may be true only if there is a positive risk of recession. The yield curve would be decreasing if βy2(z0) is smaller than unity. This property is summarized as follows:
The two conditions to the left correspond to our assumption that βy1(z0) = 1. The condition to the right means that βy2(z0) is less than unity. The interpretation of condition (17.23) in terms of saving behavior is simple. Suppose that the agent does not want to save when the uncertain growth per period of her income is either or . Does ? Condition (17.23) states that she wants to save this imply that she will save in the presence of growth risk more. That would reduce the equilibrium interest rate. Define function v in such a way that v(z) = –u′(z) for all z. Next, we define function V as V(Z) = v(exp Z), and Z = ln z0 and Xt = ln xt. We can then rewrite property (17.23) as follows:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_264.html [4/21/2007 1:32:23 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_265.html
Page 265
In words, condition (17.24) shows that two lotteries on which the agent with utility function V is indifferent when taken in isolation are jointly undesirable. This means that V is (weakly) proper, as defined by Pratt and Zeckhauser (1987). From chapter 9 we know that a necessary condition for V to be proper is that
The violation of this condition would imply the existence of a pair (X1, X2), which would violate condition (17.24). This condition is necessary and sufficient for condition (17.24) to hold when With V(Z) = –u′(exp Z), we have that
and
are small risks.
This yields the following result: COROLLARY
5 A necessary condition for the yield curve to be nonincreasing is that
for all z, where P(z) = –u′″(z)/u″(z) is the index of absolute prudence. This condition is necessary and sufficient when the risk on growth is small and the short-term interest rate equals the rate of pure preference for the present. The necessary condition (17.29) requires the sophisticated computation of the fifth derivative of the utility function. This shows that introducing the risk of recession in the long term and in the short term makes it really a hard task to guarantee that long-term discount rates are smaller than short-term ones. So far we have not been able to fully characterize the set of utility functions generating an unambiguously sloped file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_265.html (1 of 2) [4/21/2007 1:32:23 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_265.html
yield curve, what-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_265.html (2 of 2) [4/21/2007 1:32:23 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_266.html
Page 266 ever the distribution of binary.
. In the next proposition, we provide the necessary and sufficient condition when
is
70 Define function H as H(x, y) = u′(z)u′(zxy) – u′(zx)u′(zy). Suppose that u exhibits DRRA. A necessary condition for the yield curve to be nonincreasing independent of the distribution of is written as PROPOSITION
for all (x, y) such that x < 1 < y. Limiting the analysis to binary distributions for necessary and sufficient. Proof We normalize z to unity. Let
makes condition (17.30)
be distributed as (x, p; y, 1 – p). Condition (17.17) is rewritten as
or equivalently,
Recall that DRRA infers that (x – 1)(y – 1)H(x, y) ≥ 0. In particular, it infers that H(x, x) and H(y, y) are nonnegative. When x and y are both larger than unity, we also have that H(x, y) is nonnegative, yielding K(x, y, p) Without CRRA, the difficulty ≥ 0. This is another proof of proposition 69 for binary distributions of occurs when x < 1 < y, implying that H(x, y) < 0. This is the case that we examine now. Observe that K is quadratic in p with ∂2K/∂p2 = H(x, x) – 2H(x, y) + H(y, y) ≥ 0. It is minimum at
which is between 0 and 1. The minimum value of K(x, y, p) for p ∈ [0, 1] is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_266.html (1 of 2) [4/21/2007 1:32:24 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_266.html
It will be nonnegative for all x < 1 < y if and only if condition (17.30) is satisfied.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_266.html (2 of 2) [4/21/2007 1:32:24 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_267.html
Page 267 The reader can check that symmetrically, condition [H(x, y)]2 ≥ H(x, x)H(y, y) for all x < 1 < y is necessary for the yield curve to be nondecreasing. Finally, observe that a sufficient condition for (17.30) is that H itself be log supermodular. This condition should not be confounded with DRRA, which means that h(x, y) = u′(zxy) is log supermodular.
17.5 Concluding Remark We showed that the equilibrium price of time is influenced by three distinct characteristics of the preferences of the representative agent: • The agent's pure preference for the present. This is characterized by β. • The agent's willingness to smooth consumption if the economy is growing. This is characterized by the degree of concavity of his utility function. • The agent's willingness to accumulate wealth if the rate of growth of the economy is uncertain. This is characterized by the degree of convexity of his marginal utility. A larger expected growth increases the equilibrium interest rate, whereas more uncertainty on the growth rate reduces it. When we calibrate the model to the actual mean and standard deviation of the historical economic growth, we obtain a short-term interest rate around 4% to 5% a year, which is much larger than the historical level (around 1%). The large historical growth rate of the economy and its low variability around the mean are such that economic agents should accumulate much less saving than they actually do at the historical rate of interest. This is the risk-free rate puzzle. Another intriguing phenomenon is that under the reasonable assumption that relative risk aversion is decreasing, the yield curve should be decreasing; in other words, the long-term interest rate should be less than the short-term one. This means that the so-called inverted yield curve should be the rule rather than the exception. A possible explanation of this puzzle is that ordinary households face a liquidity constraint, so an illiquidity premium is generated for long-term investments.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_267.html [4/21/2007 1:32:24 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_268.html
Page 268
17.6 Exercises and Extensions Exercise 88 One-switch utility functions are interesting because there is a simple proof for the property that the yield curve is not increasing. Following Bell (1988b), function u is one-switch if u′(z) = a + z–γ with a > 0 and γ > 0. It yields relative risk aversion R(z) = γ[az–γ + 1]–1, which is decreasing in z. • Show that condition (17.17) as a strict inequality holds for that utility function, except when Describe the yield curve in this latter case. • Show that conditions (17.29) and (17.30) hold for that utility function. Exercise 89 Find the necessary and sufficient condition for the yield curve to be decreasing when the growth rate is negative almost surely. Exercise 90 In this chapter we assumed that the growth is a multiplicative process. Suppose alternatively that Derive a necessary and sufficient condition for the yield growth is an additive process, with curve to be decreasing when there is no risk of recession. Give the intuition to it. Exercise 91 (Gollier 1999) In this exercise we relax the assumption that the growth risk is identically distributed Put differently if there is no in each period. Our assumption, ceteris paribus, is that change in the GDP per capita from date 1 to date 2, the short-term interest rate remains the same. Suppose that there is no risk of recession in the short run and that the future short-term interest rate is larger than the rate of pure preference for the present, almost surely. Show that the yield curve is then decreasing if and only if relative risk aversion –zu″(z)/u′(z) is decreasing and relative prudence –zu′″(z)/u″(z) is increasing.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_268.html [4/21/2007 1:32:25 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_269.html
Page 269
18 The Liquidity Constraint Up to now we assumed that agents can borrow and save as much as they want at an (equilibrium) interest rate ρ. This is not a realistic assumption. The rate at which we can borrow money is, in general, larger than the rate at which we can save (the opposite would generate a money machine). This is due to the cost of intermediation together with the asymmetric information that prevails between borrowers and lenders with respect to the solvency issue. In this chapter we take the extreme position that agent may not be a net borrower: net saving must be positive at any time. This is a rather strong version of the liquidity constraint. The existence of a liquidity constraint reduces the ability to smooth consumption over time. In particular, if agents expect a large increase in their earnings in the future, their inability to borrow forces them to have a rate of increase in consumption over time that is larger than the optimal one. Also the possibly binding liquidity constraint in the future increase the volatility of future consumption. This will induce consumers to save more. Deaton (1991), Hubbard, Skinner, and Zeldes (1994), and Carroll (1997) among others, used numerical simulations to show that the risk of facing a liquidity constraint in the future introduces another important motive to save. This will have a large impact on the optimal level of savings and on the impossibility to smooth consumption over time. In other words, saving is a "buffer stock" that reduces the probability that the liquidity constraint will be binding in the future. Agents accumulate assets to insulate themselves from temporary drop in income that cannot be compensated by short-term debts. Everyday life provides a lot of examples that a liquidity constraint can have a devastating effect on the lifetime welfare of a household. Importantly it prohibits consumption smoothing. We therefore need to examine some of the consequences
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_269.html [4/21/2007 1:32:26 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_270.html
Page 270 of the liquidity constraint on the saving behavior and on the attitude toward risk.
18.1 Saving as a Buffer Stock It is trivial to say that a binding liquidity constraint increases saving. But even when current earnings plus current cash are large enough to cover current expenses, the risk that the liquidity constraint will be binding in the future because of a sequence of adverse shocks on incomes can induce a larger saving rate. The existence of the liquidity constraint is costly for agents with a concave felicity function because of the misallocation of consumption that it generates. These agents are willing to pay for the reduction of the risk that this constraint will be binding in the future. The simplest way to reduce this risk is to accumulate more wealth. This is the intuition for why the liquidity constraint increases saving even when the constraint is currently not binding. Another way to explain the same idea is that the possibility of borrowing provides some insurance. Removing this opportunity is like introducing a downside risk which increases the willingness to save. This excess of saving with respect to its level when agents can freely save and borrow under the lifetime budget is called the "buffer stock" saving. In this section, we consider the simplest possible model to examine the effect of the liquidity constraint on the saving behavior. Our model is such that the consumer faces risk only once in his life. The specific timing of the problem is as follows: • At date 0, the consumer has a cash-on-hand w0. He decides how much to save (s), and he consumes the difference w0 – s. • At date 1, the agent observe his random income his cash-on-hand
He determines his consumption plan (c1, . . ., cn), given
and given his sure flow of income (w2, . . ., wn) for the remaining periods.
Essentially we have two consumption decisions here. One is done ex ante, before observing . The other is made ex post. As before, in order to characterize the first, we must first examine the second. For simplicity, let us assume that w2 = · · · = wn = w. Let us assume first that and β = ρ = 1. Without any liquidity constraint, an optimal level of saving s at date 0 solves the following program:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_270.html [4/21/2007 1:32:26 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_271.html
Page 271
where h is the lifetime utility from date 1 on as a function of cash-on-hand at that date. Function h is the solution of program (15.9). Let y denote the cash-on-hand at t = 1, namely y = s + w + ε. By our assumptions, the value function h takes the following form:
because perfect smoothing is optimal here. If y is larger than w, the consumer will consume w plus only 1/n of the difference between y and w at date 1. The remaining is equally allocated over the n – 1 other dates. On the contrary, if y is smaller than w, this negative shock is split equally over the n dates by borrowing (n – 1)/n of the | y – w| from an intermediary. Observe that in any case, the marginal propensity to save is only 1/n. The willingness to save at date t = 0 is measured by the expectation of function h′, with h′(y) = u′((y + (n – 1)w)/ n) for all y. Thus, if y is random, we see that the consumer should take into account only 1/n of the risk. This is the time diversification effect that we already illustrated in chapter 15. It reduces the precautionary saving accordingly. Suppose now that the agent is prohibited to be a net borrower. The problem can here be written as (18.1), but with a value function h that is replace by function hc that is defined as
If y is smaller than w, the liquidity constraint is binding and the agent is forced to absorb all his losses immediately, by reducing his consumption to c1 = y. Observe that an immediate effect of the liquidity constraint is to raise the MPC from 1/n to 1 in the range where it is binding. Another remark is that the ex post cost of the liquidity constraint can be measured by the difference between h(y) and hc(y), which is nonnegative under risk aversion. To make the problem interesting, suppose that the liquidity constraint is not binding at date 0 (s is positive). This can be done by assuming that w0 is large enough. However, the risk that it may be binding in the future if is unfavorable may have a positive impact
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_271.html [4/21/2007 1:32:27 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_272.html
Page 272 on the optimal saving at date 0. This is the notion of a buffer stock. The buffer stock is positive if and h′ coincide when s + ε is positive, we have to check whether than Eh′. Since negative s + ε, that is if
is larger
is larger than h′ for all
This is always true under the assumption of the concavity of the utility function. We conclude that the presence of a liquidity constraint always induces risk-averse households to increase their saving when β = ρ = 1. The inability to smooth negative shocks over time raises the expected marginal utility at the date of the shock. No other assumption than risk aversion is required in this case. PROPOSITION
71 Suppose that β = ρ = 1. Then the existence of a liquidity constraint increases savings.
Another important effect of the liquidity constraint on the optimal consumption is that it concavifies the consumption function. Indeed, the marginal propensity to consume out of wealth at date t = 1 equals
Because 1/n is (much) smaller than unity, the MPC is decreasing, and the consumption is concave. The volatility of consumption will thereby be much increased by the presence of a liquidity constraint.
18.2 The Liquidity Constraint Raises Risk Aversion We now turn to the effect of a liquidity constraint on the degree of risk tolerance. It is intuitive that an agent who is able to smooth a shock over several periods should be more willing to take risk than an agent who is forced to absorb the shock immediately through a reduction of its current consumption. This is because the liquidity constraint eliminates the time diversification effect for downside risks. We consider the same problem as above, except that the decision problem at date 0 is about an exposure to risk, not a saving problem. It can be for example a standard (or an Arrow-Debreu) portfolio that will be liquidated at date 1. We know that the effect of a change in the environment has a negative impact on the optimal risk exposure if it
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_272.html [4/21/2007 1:32:28 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_273.html
Page 273 increases the degree of concavity of the value function. The problem is thus to compare Th to coincide when y is larger than w, so do Th and
. Since h and hc
. Thus we just have to look at what happens only when y < w.
Differentiating condition (18.2) twice yields
For hc we directly obtain that
Thus a liquidity constraint reduces the willingness to take risk if Th is larger than
in the relevant domain,
namely if
When y is close to w, we have that the unconstrained agent has an absolute risk tolerance that is n times larger than the absolute risk tolerance of the agent facing a binding liquidity constraint. When y is smaller than w, the liquidity constraint induces a smaller consumption at date 1. This generates a "wealth effect" that can go both directions depending on whether absolute risk tolerance is increasing or decreasing. If it is increasing, we verify immediately that
We conclude that DARA is sufficient to guarantee that the a liquidity constraint induces more risk aversion, if β = ρ = 1. This effect is quite large, since a binding liquidity constraint divides risk tolerance by more than a factor n under DARA. 72 Suppose that ρ = β = 1. Where it is binding, the liquidity constraint divides the absolute risk tolerance by n, the remaining lifetime of the consumer, under CARA. The effect is larger under decreasing absolute risk aversion. PROPOSITION
18.3 The Liquidity Constraint and the Shape of Absolute Risk Tolerance file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_273.html (1 of 2) [4/21/2007 1:32:28 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_273.html
In the previous section we compared the risk-taking behavior with and without the liquidity constraint. We now examine the shape of the absolute risk tolerance of the value function hc under the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_273.html (2 of 2) [4/21/2007 1:32:28 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_274.html
Page 274 liquidity constraint. When β = ρ = 1, the analysis above directly yields that
There is an upward jump from T(w) to nT(w) to risk tolerance when the cash-on-hand attains y = w. Thus, when u is DARA, is increasing, which implies that hc is also DARA. But when u exhibits increasing absolute risk aversion (IARA), hc is not IARA because of the upward discontinuity of
at y = w. The liquidity constraint
introduces a bias in favor of decreasing absolute risk aversion. The intuition is simple: the liquidity constraint is more likely to be binding for poorer people, which makes them more risk-averse because of their inability to time diversify the risk. This phenomenon is more striking when we relax the assumption that β = 1. Let us now consider the case with β < ρ = 1. The consumption-saving problem at date 1 is now written as
Variable yt is the cash-on-hand available at the beginning of period t. It is the cash-on-hand available at the previous period, minus what has been consumed from it (ct), plus the current income w. The liquidity constraint states that it must always be larger than the current income. When βρ is less than 1 as it is the case here, we know that the unconstrained optimum consumption plan is downward sloping: because the consumer is impatient, she wants to borrow money to finance her early consumption. She would reimburse her debt later in her life, using her flow of income w. It is then easy to understand that, for a low initial cash-on-hand y = y1 < w, the optimal consumption path is to consume it in full (c1 = y), and to consume the income at each period (ct = w) afterward. On the contrary, if y = y1 > w, the presence of the liquidity constraint yt ≥ w
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_274.html [4/21/2007 1:32:29 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_275.html
Page 275 implies an optimal consumption path in which the cash-on-hand is positive but decreasing down to w at some date τ(y). From date τ(y) on, the agent consumes his income w. For any given y > w, the constrained optimal consumption is characterized by the following conditions: there is a integer τ(y) ≥ 2 such that
The consumer equalizes to λ(y) the discounted marginal felicity of consumption at all dates where the constraint is not binding. When it is binding, it must be that the discounted marginal utility is less than λ(y). In fact system (18.7) to (18.9) characterizes the unconstrained optimal consumption path where the time horizon would have been arbitrarily limited to n = τ(y). Thus, as long as a change in the initial cash-on-hand does not modify τ(y), the local properties of the solution are well known. In particular, we may use the condition, (15.11), that states that
By proposition 53, this implies that is decreasing (increasing). We now examine the discontinuities of
is decreasing (increasing) within each region where τ(y) is constant, if T
that occur at the set of initial cash-on-hand y where τ(y) changes.
First, it is easy to check that τ(y) is increasing in y. In a region of y where τ(y) is constant, λ(y) is equal to , which we know is decreasing with y. Indeed, in the unconstrained consumption problem, the value function is concave. Thus, as y increases, the right-hand side of inequality (18.8) decreases. Eventually constraint βt–1u′(w) ≤ λ(y) is violated for t = τ(y) + 1. Once this happens, date τ(y) + 1 must be included in the set of dates with an internally efficient consumption allocation. This means that τ(y) must be an increasing function of y. We conclude that there is a sequence w = k1 < . . . < kn < kn+1 = + ∞ such that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_275.html [4/21/2007 1:32:30 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_276.html
Page 276 τ(y) = i for all y ∈ [ki–1, ki[, i = 2, . . ., n + 1. At any given y = ki, there is an upward jump to the absolute risk tolerance, with
This condition is obtained from (18.10), together with ct(y–) = ct(y+)∀t, τ(y+) = τ(y–) + 1 and
.
The upward steps all have the same size T(w). To sum up, when u is DARA,
is piecewise increasing with upward jumps, which implies that the value
function hc is also DARA. But when u is IARA,
is piecewise decreasing with upward jumps, and hc is
neither IARA nor DARA. However, it happens that has a general upward shape even in this case, in the sense that is made precise in the following proposition: PROPOSITION
73 Suppose that β < ρ = 1. Consider the increasing sequence of cash-on-hands k1 < . . . < kn where
the boundedness of the liquidity constraint is delayed by one period. Then we have that and
.
Proof Parameter ki is defined by λ(ki) = βi–1u′(w). It implies that λ(ki+) = βλ(ki–1+). Using condition (18.7) respectively at y = ki and ki–1 yields
for all t = 1, . . ., i and all s = 1, . . ., i – 1. Taking t = s + 1, we have that
for all s = 1, . . ., i – 1. Using condition (18.10), we find that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_276.html (1 of 2) [4/21/2007 1:32:30 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_276.html
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_276.html (2 of 2) [4/21/2007 1:32:30 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_277.html
Page 277
Figure 18.1 Absolute risk tolerance of the value function when the felicity function is IARA
Thus we conclude that
is larger than
. The proof for the left derivatives is identical.
We describe a typical shape of the absolute risk tolerance of the value function in figure 18.1 in the case where u exhibits increasing absolute risk aversion. It is noteworthy that when we shorten the length of each periods, the horizontal distance between the upward discontinuities will decrease. At the limit of continuous time and infinite horizon, the proposition above basically states that all value functions will exhibit decreasing absolute risk aversion. This important result was obtained by Epstein (1983). This analysis provides a simple intuition for the result: a larger wealth better insulates the consumer from the liquidity constraint. That reduces his aversion to risk. In the continuous-time framework, this effect dominates even when the agent is IARA on his consumption.
18.4 Numerical Simulations. The problem is more complex to analyze if all future incomes are uncertain. By backward induction, the decision problem at t is written as
with vn ≡ u. The Euler equation is written as
with an equality if the liquidity constraint is not binding. file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_277.html [4/21/2007 1:32:31 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_278.html
Page 278
Figure 18.2 Optimal consumption as a function of cash-on-hand with a liquidity constraint
As in previous chapters, the effect of adding more risks must be treated by looking at the preservation of the relevant properties of the felicity function. This research remains to be done and will not be covered here. Rather, we provide numerical simulations based on the example presented in section 16.4.2. As a reminder, we assume that β = ρ = 1 and that the income at every period is either 5 or 15 with equal probability. Finally we consider an agent with a constant relative risk aversion equaling 2. The benchmark case is the optimal consumption strategy when there is no liquidity constraint, as described in section 16.4.2. In figure 18.2 we draw the optimal strategy with a liquidity constraint as a function of cash-on-hand for different time horizons. We see that the liquidity constraint becomes binding at a cash-on-hand around 6, for the various time horizons that are considered here. The concavity of the consumption function comes from the kink generated by the constraint, but also from proposition 65. If zero assets are carried forward, as is the case to the left of the graph, consumption changes are set equal to income changes over the period. Schechtman (1976) and Bewley (1977) showed that the optimal consumption converges to the mean income 10 at all dates as the time horizon tends to infinity. Such a result is made possible despite
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_278.html [4/21/2007 1:32:31 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_279.html
Page 279 the liquidity constraint because the optimal level of cash-on-hand tends to infinity. Zeldes (1989), Deaton (1991), and Carroll (1997) reexamine the optimal consumption behavior when the income process is not stationary. One can measure the size of the buffer stock behavior by the difference between the optimal consumption without the liquidity constraint and the optimal consumption with the liquidity constraint. These two functions of cash-on-hand are depicted in figure 18.3 for different time horizons. When the quantity of assets carried forward is large, the risk of the liquidity constraint being binding in the future is small, in particular, if the time horizon is short. Then the buffer stock does not need to be large. This is confirmed by this figure, since the optimal consumption quickly converge to the optimal consumption without the constraint. For example, with three periods to go and a cash-on-hand of 10, it is optimal to consume around 8 at that date as it appears in figure 18.3b. The agent will carry 2 units of consumption forward. If the bad state of the world occurs, it makes a cashon-hand of 7. We see on figure 18.3a that the liquidity constraint does not bind with that wealth level at that time. This explains why the buffer stock saving is zero with y = 10 with three periods to go. A striking feature of these figures is the small size of the buffer stock. We also measured the attitude toward risk as a function of the length of time horizon. We did it by computing the willingness to pay for the elimination of the current risk on income. We report these results on figure 18.4. We see a huge increase in risk aversion generated by the partial removal of the opportunity to time diversify due to the liquidity constraint.
18.5 Conclusion The existence of a liquidity constraint is extremely useful to explain several aspects of saving and risk-taking behaviors. First, it provides an additional motive to save. The "buffer stock" motive to save relies on the idea that a larger wealth accumulation is useful to get rid of the risk of the liquidity constraint. It generates a better smoothing of consumption over time. Second, it explains why the marginal propensity to consume is so large in the real world. A life cycle model with a perfect financial market would predict a very low MPC for young households, in particular, if there is no serial correlation in the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_279.html [4/21/2007 1:32:32 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_280.html
Page 280
Figure 18.3 Comparing the optimal consumption as a function of cash-on-hand with (solid) and without (dashed) a liquidity constraint
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_280.html [4/21/2007 1:32:32 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_281.html
Page 281
Figure 18.4 Risk premium for the current income risk as a function of cash-on-hand, with (solid) and without (dashed) a liquidity constraint
flow of labor incomes. The existence of the liquidity constraint can explain the fact that actual consumption flow is much parallel to the actual flow of incomes. The liquidity constraint also explain why young consumers may be more risk-averse than would have been expected with a perfect financial market. Remember that in a frictionless environment, the young consumer's tolerance to risk is basically proportional to her remaining lifetime. This is due to her better position to time diversify intertemporally independent risks. But in the real world we don't see many young agents insuring older ones. This suggests that time diversification cannot be used in a large extend by most consumers. The liquidity constraint is one explanation for that phenomenon to prevail.
18.6 Exercises and Extensions Exercise 92 Consider the two-date model with β = ρ = 1. Suppose that w0 is less than w1. Define the cost of the liquidity constraint to be the price that one is ready to pay today for the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_281.html [4/21/2007 1:32:33 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_282.html
Page 282 elimination of the constraint. Show that it is approximately equal to 0.125(w1 – w0)2 A(0.5(w1 + w0)). Exercise 93 (Epstein 1983) The aim of this exercise is to prove that the value function hc defined by (18.5) is DARA in the limit of continuous time and infinite horizon. The maximization problem in that framework is written as
(we normalized w to zero). Parameter δ is the discount rate. Using backward induction, we can rewrite this problem as
at the limit, when Δt → 0. • Use first-order Taylor expansions for functions e–δΔt and hc(y – cΔt) around Δt = 0 to rewrite the above problem as
• Change variable from y to . Define function ψ with . Rewrite the above equation using z and ψ rather than y and hc [Hint: k = maxx f(x) ⇔ –k = minx – f(x)]. Show that it implies that ψ must be concave. • Assume that hc is concave (this cannot be guaranteed in this framework). Show that ψ is concave is concave if and only if hc is DARA. • A simple example is obtained with the quadratic utility function with u′ (c) = α – c, with α > y. This illustrates the fact that the value function is DARA even when the felicity function is IARA. a. Show that the date τ(y) at which the liquidity constraint becomes binding satisfies the condition that e–δτ + δτ = 1 + δ(y/a).
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_282.html (1 of 2) [4/21/2007 1:32:34 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_282.html
b. Show that τ is increasing in y. c. Show that
and
.
Exercise 94 In this exercise we extend formula (18.3) to the case with β < ρ = 1. The reader should make use of the characterization of
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_282.html (2 of 2) [4/21/2007 1:32:34 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_283.html
Page 283 the optimal consumption obtained in section 18.3 to solve it. Prove the following statement: • The MPC is piecewise decreasing if absolute risk tolerance is concave. • The MPC has downwards discontinuities at the various level of y where the boundedness of the liquidity constraint is delayed by one period. • Use equation (18.14) to show that the MPC evaluated at the ki+ is decreasing in i under IARA, or under CRRA.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_283.html [4/21/2007 1:32:34 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_284.html
Page 284
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_284.html [4/21/2007 1:32:35 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_285.html
Page 285
19 The Saving-Portfolio Problem The concept of prudence has been introduced to treat the effect of an exogenous future risk on current savings. But many risks in life are endogenous. For example, people don't want to insure their house in full, and they purchase risky assets. The level of saving clearly affects the optimal behavior to these endogenous risks. Symmetrically the opportunity to take risk should influence the saving behavior. Thus it is important to examine the joint consumption and portfolio decisions. Samuelson (1969) and Merton (1969) were the first to provide a complete analysis of this problem in the case of HARA utility functions. Their motivation was not so much about the optimal saving behavior. Rather, they wanted to determine the optimal portfolio decision in a dynamic framework with intermediary consumption. Drèze and Modigliani (1972) focused instead on the effect of the portfolio strategy on the optimal saving strategy.
19.1 Precautionary Saving with an Endogenous Risk Following Drèze and Modigliani (1972), we start with the analysis of the impact of an endogenous risk on the optimal saving. We consider two cases. In the first case, the agent lives in an Arrow-Debreu world with the possibility to take any risky position on the aggregate risk in the economy. In the second case, there are only one risk-free asset and one risky asset in the economy. No other contingent contract can be exchanged. This refers to the standard portfolio problem. 19.1.1 The Case of Complete Markets Suppose that the state price density in the Arrow-Debreu economy be (.). The saving-portfolio problem is written as follows in this case:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_285.html [4/21/2007 1:32:35 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_286.html
Page 286
At date 0, the agent saves s from his current income w0 and sells his sure future income w1 on financial markets, yielding
. With this wealth he purchases a portfolio c(.) of Arrow-Debreu securities. His decision
problem is to select s and c(.) in order to maximize his lifetime expected utility. Define and . Parameter ρ is the risk-free interest rate. The budget constraint takes the following form:
with the property that . Notice that we can then rewrite this problem by using the value function v that has been defined by program (13.3) for price kernel π(.) and utility function u ≡ u1. It yields the following decision problem:
In parallel to what we have done for the exogenous future risk, we address the question of how does the presence of this endogenous risk affect the optimal saving. An application of this problem is for the effect of an economic policy related to pension funds. If one forces savings to be invested in bonds alone, will it induce more or less savings than if one allows pension funds to be invested in bonds and stocks? Intuitively two contradictory effects are at work here if investing in stocks is allowed. The global effect is split into the effect of a pure risk and the effect of an increase in the expected future income. First, because it is optimal to purchase some shares since , the presence of the portfolio risk has a positive precautionary saving effect under prudence. The intensity of this effect is increasing in P. Second, this option to take risk has a positive wealth effect: it increases the expected future income of the agent. We know that this increase in future income has a negative impact on the optimal saving if u1 is concave (income smoothing). The intensity of this opposite effect is increasing in the degree of concavity of u1; that is, it is increasing in
. Thus the global effect of the option to invest
in risky assets is ambiguous under prudence. But we may conjecture that it will have a positive impact on the aggregate saving if absolute prudence is large enough compared to absolute risk aversion. This is what we now show.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_286.html [4/21/2007 1:32:36 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_287.html
Page 287 We compare the solution of the above problem to the one in which the agent is forced to put all his wealth in a risk-free portfolio. This second problem is written as
where
is the return of the safe portfolio. The first-order condition is written as
The optimal saving when the agent is allowed to take a risky position is larger than s* if H′(s*) is positive, that is, if
The following proposition is then a direct application of proposition 54: 74 In an economy with a complete set of Arrow-Debreu securities, allowing investors to take a risky position increases (resp. reduces) their optimal saving if and only if absolute prudence is larger (resp. smaller) than twice absolute risk aversion. PROPOSITION
19.1.2 The Case of the Standard Portfolio Problem Let us suppose, alternatively, that the only possible risky positions that are allowed must be linear in the aggregate risk . Formally the problem is to choose s and α in order to solve the following problem:
This can be seen as a saving-portfolio problem in which s is the aggregate saving, α is the amount that is invested in a risky asset whose excess return is , whereas s – α is invested in a risk-free asset whose gross return is ρ. . Because the choice of α has no direct impact on the first Without loss of generality we assume that period utility, this problem can be seen as a dynamic problem in which s is selected in period 0 and α is selected in period 1. The value function can be written as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_287.html [4/21/2007 1:32:36 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_288.html
Page 288 and the first period problem can be rewritten as
We compare this solution to the one in which the agent cannot invest in the risky asset, in which case s maximizes u0(w0 – s) + u1(w1 + ρs). Let s* be the solution of this problem. For exactly the same argument as in the case of complete markets, the opportunity to take a risky position increases saving if , where α* is the optimal demand for the risky asset of the agent with wealth w0 + ρs*. Normalizing α* to unity and using condition (16.2), this condition is equivalent to
where z = w1 + ρs*. Because w0, w1 and ρ are arbitrary, we want this condition to hold for any z. We already know from proposition 34 that condition (19.4) holds if and only if P(z) ≥ 2A(z) for all z. This concludes the proof of the following proposition, which is due to Gollier and Kimball (1996): 75 Consider the standard portfolio problem. The opportunity to invest in a risky asset raises (resp. reduces) the aggregate saving if and only if absolute prudence is larger (resp. smaller) than twice the absolute risk aversion. PROPOSITION
19.1.3 Discussion of the Results For the two structures of financial markets, we obtain that P ≥ 2A is necessary and sufficient for the property that allowing risk-taking enhances the willingness to save. As expected, prudence is not sufficient to guarantee that allowing households to invest their savings in risky asset increases aggregate saving. Notice that the condition is also stronger than decreasing absolute risk aversion, which means that P ≥ A. Because credible estimations of absolute or relative prudence are still missing in the literature, let us look at the special case of HARA utility functions
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_288.html [4/21/2007 1:32:37 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_289.html
Page 289 for which we have
and
Thus, for HARA utility functions, absolute prudence equals (γ + 1)/γ times the absolute risk aversion. Clearly, absolute prudence is larger than twice absolute risk aversion if (γ + 1)/γ ≥ 2, or γ ≤ 1. In the specific case of CRRA functions where γ represents relative risk aversion, this would mean that relative risk aversion be less than 1. As explained in section 2.6, this condition is not true for most households in the real economy. Thus, if we believe in CRRA functions and in relative risk aversion being larger than unity, we should expect that the option to invest in risky assets would rather reduce aggregate saving. The (expected) consumption smoothing effect is here stronger than the precautionary saving effect. More generally, observe that condition P(z) ≥ kA(z) is equivalent to
where R(z) is the relative risk aversion measured at z. It implies that a set of sufficient conditions for P ≥ 2A is nonincreasing relative risk aversion with a coefficient of relative risk aversion bounded below unity. On the contrary, nondecreasing relative risk aversion combined with a relative risk aversion larger than 1 is enough for P ≤ 2A, namely for options to take risk to reduce saving. Our assumption that there are only two periods is without loss of generality, at least in the complete markets case. Indeed, we know that property that P is uniformly larger or uniformly smaller than 2A is inherited by the value function in this latter case. But only the property that P > 2A is preserved in the case of the standard portfolio problem. In consequence, removing the opportunity to invest in stocks reduces saving if –u′″/u″ is larger than –2u ″/u′. But it may be possible that it is also the case for some utility functions that violate this condition.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_289.html [4/21/2007 1:32:38 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_290.html
Page 290
19.2 Optimal Portfolio Strategy with Consumption. In the previous section we examined how does the possibility to purchase stocks influence consumption. In this section we examine how does the possibility to consume over time affects the optimal dynamic portfolio strategy. This is done basically by combining results from chapters 11 and 15. The optimal dynamic portfolio strategy when there is no intermediary consumption was the focus of chapter 11. In particular, we showed that a longer time horizon increases or reduces the demand for stocks depending on whether absolute risk tolerance is convex or concave, at least in an Arrow-Debreu economy. When investors consume their wealth over several periods, there is also a time diversification effect that is at play. We assume in this section that u0 ≡ u and u1 ≡ βu. We add a risk-taking decision at date 0 to the problem that we examined in the previous section. In short, the dynamic problem is written as follows:
where is defined as
The attitude toward risk at date t = 0 is given by the measure of concavity of the value function . The only difference of this dynamic problem with respect to problem (11.1) is that value function v is replaced by to express the time diversification effect. The effect of the existence of the second period on the optimal portfolio structure at date 0 is influenced by several factors: • A background risk effect generated by the option to invest in risk in the second period. • A wealth effect generated by this option. • A time diversification effect generated by the possibility to allocate gains and losses at date 0 over two periods. • An income effect generated by the flow of future incomes. Combining results on whether u ≡ u0 ≡ u1 is more concave than v and whether v is more concave than yields conditions for a positive effect of time horizon on risk-taking. An illustration is provided by the following proposition:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_290.html [4/21/2007 1:32:38 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_291.html
Page 291 PROPOSITION
76 Suppose that markets are complete and that investors have no income at old age (w1 = 0). The
young investor is always less risk-averse than the old investor if and only if the absolute risk tolerance of the felicity function u is convex and subhomogeneous. Proof A direct consequence of propositions 55 and 60. Another illustration is when w0 = w1, β = ρ = 1 and u is CRRA. In that case the agent with two periods to go should invest exactly twice as much as the agent who has only one period to go. Only the time diversification effect is at play. The case of CRRA is examined in more details in the next section.
19.3 The Merton-Samuelson Model The previous section is useful to understand the interactions between savings decisions and portfolio decisions. This analysis is not very helpful, however, to provide simple guidelines for taking decisions. Several authors (Leland 1968; Mossin 1968; Merton 1969; Samuelson 1969) simultaneously described the optimal decision rules in a particular case of the general model. The discrete version of the Samuelson-Merton saving-portfolio problem is written as follows:
with vn ≡ u. As in the previous section, we assume here that there is no other uncertainty than the portfolio risk. More specifically, we assume that the consumer receives sure income wt at date t = 1, . . ., n from his human capital. Variable y is interpreted as the cash-on-hand at date t. At that date with cash y, the agent consumes ct and carries the remaining by investing in assets to date t + 1. He invests αt+1 in stocks whose excess return from date t . The remaining is invested in the risk-free asset whose return is ρ. We assume here that there to date t + 1 is is no serial correlation in the returns of stocks. Finally we do not take into account of liquidity constraints. The first-order conditions for this problem are written as
and
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_291.html [4/21/2007 1:32:39 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_292.html
Page 292 The first condition is easily recognized as the condition on optimal savings: marginal discounted utilities are equalized. The second condition is also familiar, as it describes the condition for optimal risk taking. Using the envelope theorem together with condition (19.7) for date t + 1 rather than t, we obtain that . It implies that we can rewrite the condition above as follows:
Random variable is the optimal consumption at date t + 1 given the cash-on-hand y at t. We recognize the Euler equation, which requires that the discounted marginal utility of 1 dollar must be equalized over time. It happens that this model can be solved analytically when the utility function is HARA. To simplify the notation, let us limit the analysis to the case where u(c) = c1–γ/(1 – γ). When facing a dynamic problem like this, the only practical way to find a solution is to guess it, and then to check whether the trial solution verifies the Bellman equation (19.6). Recall that we learned in section 11.2.2 that the value function is HARA when u is HARA in the pure investment problem. We could try to check whether this is also the case when intermediary consumption is introduced. Our trial solution is thus k1, . . ., kn–1 to be specified. In particular, we will try a solution with
, for some constants K1, . . ., Kn–1 and . It is the net present
value of future labor incomes, that is, it is human capital. Let us define constants at, bt, and ηt as follows:
We then check that conditions (19.7) and (19.8) hold with
and
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_292.html (1 of 2) [4/21/2007 1:32:40 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_292.html
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_292.html (2 of 2) [4/21/2007 1:32:40 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_293.html
Page 293 or equivalently,
To check that this trial solution was a good guess, we need to check Bellman equation (19.6). After simplifying by
, it can be written as
Since this gives a well-defined recursive equation for H1, H2, . . ., Hn–1, Hn = 1, we obtain the confirmation that the trial solution at hand is indeed the optimal solution of our problem. Once the sequence (H1, . . ., Hn) is obtained, it is easy to describe the optimal saving-portfolio strategy, which is characterized by conditions (19.14) and (19.15).
is the aggregate wealth at date t, which is the sum of cash-on-hand at that Observe that time and the human capital. The optimal consumption strategy consists in fixing for each date a share of aggregate wealth that will be consumed at that date. The marginal consumption to consume at date t is equal to [Ht+1ηt+1 + 1]–1. Similarly
is the aggregate wealth at the beginning of period t +
1. Thus the optimal investment in stocks at any specific date t + 1 is a constant share at+1 of the aggregate wealth at that date. This is a consequence of the very specific properties of CRRA that yields an absolute risk tolerance that is linear and homogeneous. These rates can vary over time when the distribution of returns is not stable over time. A simple case is obtained when the distribution of returns is stationary. The solution to the recursive system (19.17) is characterized in this case as
In particular, it yields a marginal propensity to consume out of aggregate wealth amounting to
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_293.html (1 of 2) [4/21/2007 1:32:41 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_293.html
In the following numerical example we assume that the stable distribution of excess returns is normal, with a mean of 0.06 and a
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_293.html (2 of 2) [4/21/2007 1:32:41 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_294.html
Page 294
Figure 19.1 Optimal consumption expressed as a proportion of aggregate wealth with (solid) and without (dashed) the opportunity to invest in stocks
variance of 0.027. As reported by Kocherlakota (1996), this is close to the distribution of the excess return of Standard and Poor 500 over the last century. We also assume that ρ = 1.01, which is also its historical value over the same period. Finally we consider an agent with no impatience and a relative risk aversion equaling γ = 2. We obtain that a = 1.12. It is then easy to calculate the {Ht}t=1, . . ., n and the corresponding optimal marginal propensity to consume together with the optimal share of aggregate wealth invested in stocks. In figure 19.1 we reported the optimal MPC as a function of the time horizon. Not surprisingly, it decreases with n basically as 1/n. The important point here is to compare the optimal MPC with the one that would have been optimal without the opportunity to invest in stocks. This optimal MPC, when the risk-free asset is the only asset that can be carried forward, is represented in figure 19.1 by the dotted curve. Because P < 2A in this simulation, the opportunity to invest in stocks reduces the willingness to save. It increases the MPC at every date. It appears that the opportunity to invest in stocks has almost no effect the optimal MPC for short horizons. But the effect is quite important for the longer horizon. In our simulation, with twenty years to go, the MPC equals 0.0500 when there is no opportunity to invest in stock. But people who are allowed to invest in stock during their lifetime would rather select a MPC equaling 0.0595. This represents a 20% increase in the MPC.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_294.html [4/21/2007 1:32:41 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_295.html
Page 295
19.4 Concluding Remark The level of wealth accumulated is obviously an important element to determine the optimal portfolio structure. Reciprocally, the opportunity to take risk in the future is expected to affect the optimal strategy of wealth accumulation. In this chapter we examined the joint determination of the portfolio strategy and the consumption strategy. The opportunity to take risk has a mixed effect on saving. Under prudence it tends to increase the precautionary saving. But this option to take risk also generates a wealth effect. It makes the consumer implicitly wealthier in the future, and that tends to reduce the consumer's willingness to save today. We showed that the global effect of financial markets on saving is positive if prudence is larger than twice the absolute risk aversion. Researchers dealing with this problem use to think about it by using the Merton-Samuelson model with a CRRA utility function. This has the big advantage that the complete solution to the problem can be obtained analytically. We presented a simple simulation of the Merton-Samuelson model by using the historical data of stock returns.
19.5 Exercises and Extensions Exercise 95 Consider the Merton-Samuelson model where the risky asset would have a zero excess return at each period. Suppose also that β = ρ = 1. Show that • younger people invest more in risky assets. Explain why. • the marginal propensity to consume equals 1/τ where τ is the number of remaining consumption dates. Exercise 96 Consider again the Merton-Samuelson model where the risky asset has a zero excess return at each period. Suppose that ρ > β = 1. Find the condition on γ that the marginal propensity to consume is an increasing function of the risk free interest rate. Exercise 97 Consider the Merton-Samuelson with a stationary distribution of returns of the risky asset. • Find the condition on γ under which the marginal propensity to consume is reduced by the opportunity to invest in the risky asset.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_295.html [4/21/2007 1:32:42 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_296.html
Page 296 • Suppose that η is less than unity. Find the asymptotic value of the marginal propensity to consume out of wealth. Exercise 98 Using a two-period two-asset model, we showed that savings is increased by the option to invest in a risky asset if and only if absolute prudence is larger than twice absolute risk aversion. In the model under consideration we supposed that the agent is free to invest how much he wants in the risky asset. Suppose, alternatively, that there is an upper limit k to the amount that can be invested in the risky asset. Characterize the necessary and sufficient condition on the utility function that guarantees that the option to invest at most k dollars in the risky asset in the future increases savings today.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_296.html [4/21/2007 1:32:42 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_297.html
Page 297
20 Disentangling Risk and Time There is some logic for decreasing marginal utility of consumption to generate at the same time an aversion to risk in each period and an aversion to nonrandom fluctuations of consumption over time. If the marginal gain of one more unit of consumption is less than the marginal loss due to one unit reduction of consumption, agents will reject the opportunity to gamble on a fifty–fifty chance to gain or lose one unit of consumption. For the same reason, if their current consumption plan is smooth, patient consumers will reject the opportunity to exchange one unit of consumption today against one unit of consumption tomorrow. Thus the degree of concavity of the utility function on consumption measures the willingness to insure consumption across states and the willingness to smooth consumption across dates. The argument for having a unique function to characterize these two features of agents' preferences is based on the assumption that their objective over their lifetime is the sum of their expected utility at each periods. This makes the objective function additive in the two dimensional space statedate. If people live for two periods, the canonical model has the following functional form:
In this model the degree of aversion to consumption fluctuations over time and the degree of aversion to risk are identical and equal to –u″/u′. Kreps and Porteus (1978) and Selden (1978) proposed an alternative model to disentangle the attitudes toward consumption smoothing over time and across states. Their models are equivalent. Because the presentation by Kreps and Porteus (1978) is more intuitive, we hereafter follow their definitions.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_297.html [4/21/2007 1:32:43 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_298.html
Page 298
20.1 The Model of Kreps and Porteus The model is a direct extension of the additive model (20.1). It is written as
where u0, u1 and v are three increasing functions. It is easier to interpret this preference functional by using the certainty equivalent functional
which is defined as
. U can then be rewritten as
We see that the lifetime utility is computed by performing two different operations. First, one computes the by using utility function v. This is done in a certainty equivalent m of the future uncertain consumption atemporal context. Thus the concavity of v measures the degree of risk aversion alone. Second, one evaluates the lifetime utility by summing up the utility of the current consumption and the utility of the future certainty equivalent consumption, using functions u0 and u1. Because all uncertainty has been removed in this second operation, the concavity of these two functions are related to preferences for consumption smoothing over time. We see here that the lifetime utility function is not linear with repect to probability, except if u1 is linear. This is why the Kreps-Porteus model is often refered to as a nonexpected utility model. Some particular cases of Kreps-Porteus preferences are worthy to mention. For example, if v and u1 are identical, we are back to the additive model (20.1). Thus the standard additive model is a particular case of Kreps-Porteus preferences. But this model is much richer because of its ability to disentangle preferences with respect to risk and time. Suppose, for example, that v is the identity function but that u0 and u1 ≡ u0 are concave. In that case the agent is willing to smooth the expected consumption over time, though he is risk neutral. At the opposite side of the spectrum, we can imagine a risk-averse agent who is indifferent toward (certainty equivalent) consumption smoothing. This would be the case if v is concave but u0 and u1 are linear. This additional degree of freedom in the way we shape individual preferences does not simplify the analysis of optimal behavior and
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_298.html [4/21/2007 1:32:44 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_299.html
Page 299 the comparative statics. Not surprisingly, many researchers such as Epstein and Zin (1991) have considered a particular specification of the model by way of power functions. To illustrate, suppose that
with γ and α being two nonnegative scalars. Observe that γ is the degree of relative risk aversion whereas α is the relative resistance to intertemporal substitution. When γ = α, the model is additive. As we have seen before, there are some empirical evidence in favor of the hypothesis that agents are more risk-averse than they are resistant to intertemporal substitution.
20.2 Preferences for an Early Resolution of Uncertainty In the classical case with u1 ≡ v, the timing of the resolution of the uncertainty does not matter for the consumer. Let us be more precise about this statement. Suppose that the consumption plan under consideration be (c0,
),
where is random. This consumption plan is completely exogenous and cannot be modified by the agent. If the realization of is not expected to be known before t = 1, the lifetime expected utility would be measured by . Suppose, alternatively, that the realization of is expected to be known at t = 0, but that we , the lifetime utility is u0(c0) + u1 cannot revise the level of saving after this observation. Conditional to (c1). Ex ante, before knowing
, the expected lifetime utility is measured by
. Thus in the classical case agents are indifferent about the timing of the resolution of the uncertainty affecting their consumption plan. This is no longer the case with Kreps-Porteus preferences. Indeed, if is not expected to be known before t = 1, the lifetime utility of the agent is measured according to equation (20.3), with . On the is observed at t = 0. Then, since obviously m(c1) = c1 under contrary, suppose that the realization of certainty, the lifetime utility conditional to c1 is u0(c0) + u1(c1), as in the classical case. Ex ante, the lifetime utility of the agent equals . This is generally not equal to . We conclude that an agent with Kreps-Porteus preferences is in general not indifferent to the timing of the resolution of uncertainty.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_299.html [4/21/2007 1:32:44 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_300.html
Page 300 We say that an agent has preferences for an early resolution of uncertainty (PERU) if the agent prefers to observe at date t = 0 than at t = 1, whatever the distribution of . This is the case when
or equivalently, when
In words, PERU requires that the certainty equivalent be always larger when using function u1 than when using function v. By proposition 3, this is true if and only if u1 is less concave than v. PROPOSITION
77 Suppose that consumers have Kreps-Porteus preferences described by equation (20.3) with . Then they prefer an early resolution of uncertainty if and only if u1 is less concave than v
in the sense of Arrow-Pratt. In the last part of this book, we will examine in detail another reason why people prefer an early resolution of uncertainty. Early information can have a positive value because informed agents are able to take better decisions. For example, if the agent can revise his consumption plan after observing the realization of at date 0, the agent will in general be better off with the information, so, clearly, information is valuable. An analysis of the value of information is presented in chapter 24. This argument for preferences for an early resolution of uncertainty is completely orthogonal to the one presented in this section, since we have assumed here that agents do not make any decision.
20.3 Prudence with Kreps-Porteus Preferences In this section we reexamine the simplest decision problem in which time and risk are intricated, namely precautionary saving. We reconsider the model presented in section 16.1 in introducing Kreps-Porteus , with . The preferences. The income flow of the consumer who lives for two periods is gross return on savings between t = 0 and 1 is ρ. With Kreps-Porteus preferences the optimal level of savings is given by the solution of the following maximization problem:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_300.html [4/21/2007 1:32:45 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_301.html
Page 301 where M(s) is the certainty equivalent future consumption as a function of savings. It is implicitly defined by
Notice first that this problem is not necessarily well behaved, even if u0, u1, and v are concave. Indeed, the concavity of the objective function (20.7) requires that function M be concave in s. This is true in particular if we assume that the consumer is neutral to consumption fluctuations, namely that u0 = u1 are linear. It can be shown (see proposition 82) that M is concave if the absolute risk tolerance of v is concave. In particular, it is concave in the HARA case. If –v′/v″ is not concave, it may be possible that the solution to the first-order condition of problem (20.7) is a local minimum. We hereafter assume that this is not the case. As in section 16.1 we want to examine the impact of the presence of uncertainty on the optimal level of savings. Adding zero-mean risk to future consumption increases saving if it raises its marginal value. Without the marginal value of saving is measured by , whereas it equals to the future consumption. Consequently the agent is prudent if
when
is added
where
and
We must restrict the set of increasing and concave functions u1 and v to obtain prudence. One such condition is that u1 and v to be identical, with v′ being convex. We can generalize this result when u1 and v are not identical if we assume that u1 is more concave than v, which is prudent. Recall that this implies that u1 is centrally more riskaverse than v at w1 + ρs. Since M(s) is less than w0 + ρs (from 20.10 and risk aversion), we have that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_301.html [4/21/2007 1:32:45 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_302.html
Page 302 If we combine this result with the assumption that v′ is convex, which implies that , we obtain that
This is equivalent to condition (20.9). The problem with this sufficient condition is that we must assume that fluctuation aversion is larger than risk aversion. Another sufficient condition for prudence is that v be DARA, as shown by Kimball and Weil (1992). The proof of this statement is as follows: DARA means that
From (20.11) we know that M′(s) is larger than unity. Recall now that M(s) is smaller than w1 + ρs under the concavity of v. The concavity of u1 implies that
is larger than
. Combining these two
results yields inequality (20.9). We conclude that v DARA is sufficient for prudence in the framework of KrepsPorteus preferences. Notice that v DARA is also necessary in the limit case where u1 is linear, where condition (20.9) simplifies to M′(s) ≥ 1. To sum up, without any information on the felicity function u1 other than its concavity, the necessary and sufficient condition for prudence is that v be DARA. This is more restrictive than in the classical case in which we obtained the weaker condition v′″ ≥ 0. This is the cost to be paid to allow for any concave felicity function. These findings are summarized in the following proposition: PROPOSITION
78 Suppose that consumers have Kreps-Porteus preferences described by equation (20.3) with . Then the agent is prudent
1. if v′ is convex and u1 is more concave than v; 2. independent of the concave felicity function u1 if and only if v is DARA.
20.4 Conclusion We have seen that the classical additive model can be generalized to disentangle aversion to consumption file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_302.html (1 of 2) [4/21/2007 1:32:46 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_302.html
fluctuation from risk aversion.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_302.html (2 of 2) [4/21/2007 1:32:46 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_303.html
Page 303 The Kreps-Porteus criterion presented in this chapter has the merit to do that without eliminating the basic additive nature of the measurement of the static welfare under uncertainty, or of the lifetime welfare under certainty. This is only when risk and time are considered together that this criterion differs from what we have seen before. Our standard techniques can still be used in this generalized framework. Because the approach is more general, additional conditions will be required for the comparative statics analysis of risk taking and consumption. In particular, we have seen that consumers will accumulate precautionary saving in the face of uncertainty if the utility function is DARA, which is stronger than prudence in the classical framework. Several other extensions of classical results remain to be performed with Kreps-Porteus preferences.
20.5 Exercises and Extensions Exercise 99 (Kimball and Weil 1992) Show that the objective function in program (20.2) is concave if u1 is more concave than v. Exercise 100 (Kimball and Weil 1992) The objective of this exercise is to estimate the intensity of the , precautionary saving motive in the two-period Kreps-Porteus framework. Consider a future income . Let ψ(k) be the precautionary premium associated to risk , namely the sure reduction in future with income that has the same effect on savings than risk . It satisfies the following condition:
where m(k) is the certainty equivalent associated to
.
• Show that m(0) = m′(0) = 0, and m″(0) = σ2v″(w1)/v′(w1), with • Show that ψ(0) = ψ′(0) = 0, which implies that
. .
• Show that ψ(k) is approximately equal to
Consider also the special cases with u1 ≡ v, u1 linear or v linear.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_303.html [4/21/2007 1:32:47 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_304.html
Page 304 Exercise 101 The objective of this exercise is to reconsider the problem of section 19.1.2 with Kreps-Porteus preferences. The consumer has preferences as in (20.2), with u1(m) = βm. Initially the agent can invest just in a risk-free asset whose return is ρ. We examine the impact of allowing the agent to invest his savings in another asset which is risky with excess return . • Show that aggregate savings is increased if v is DARA, whereas it is reduced if v is IARA. Compare this result with the results obtained in section 19.1.2. • Derive the complete solution to the two-period saving-portfolio problem with v(z) = z1–γ/(1 – γ) and u1(z) = βu0 (z) = z1–α/(1 – α).
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_304.html [4/21/2007 1:32:47 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_305.html
Page 305
VII Equilibrium Prices of Risk and Time.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_305.html [4/21/2007 1:32:48 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_306.html
Page 306
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_306.html [4/21/2007 1:32:48 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_307.html
Page 307
21 Efficient Risk Sharing This chapter is devoted to the characterization of efficient allocations of risks. Up to now we were mostly interested about individual decisions related to risk and time. Even when we examined an equilibrium, as in chapters 5 and 17, the possibility to reallocate consumption among different agents across time and states was eluded by assuming that all agents were alike. In an economy with representative agents having the same attitude toward risk and time and the same state and time-dependent endowment, autharky is obviously Pareto-efficient, and it is a competitive allocation under the assumption of risk aversion. In the real world, agents are not alike. In particular, they are not "endowed" with the same perfectly correlated risks. They don't have the same wealth, and it is likely that there is some heterogeneity in their degree of risk aversion. In such a heterogenous economy it is essential that agents be allowed to trade their risk, to borrow and to lend. In this chapter we describe the set of allocations that are Pareto-efficient. The pioneering paper on this subject is Wilson (1968). We start with a simple static exchange economy.
21.1 The Case of a Static Exchange Economy There is a single consumption good in the economy. Agents have different characteristics that are represented by θ, where θ belongs to some characteristics set Θ. The distribution function of is denoted H. We normalize the population to unity. The characteristics θ of an agent influence his preferences that are represented by a utility function u(c, θ), where c denotes the quantity that is consumed by this agent. We assume that u is increasing and concave in c.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_307.html [4/21/2007 1:32:49 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_308.html
Page 308 Agents face risk. The simplest way to represent uncertainty in such an economy is to rely on the notion of a state of nature. The set of possible states of nature is denoted X. There is an agreed-upon cumulative distribution function F for the distribution of states x ∈ X in the economy. denotes the random variable whose cumulative distribution function is F. Each agent θ gets an endowment of the consumption good that is contingent to the state of the world that will prevail. Let ω(x, θ) denote the endowment of agent θ in state x. Let also z(x) denote the quantity of the consumption good that is available per head in state x, namely
z(x) can be interpreted as the GDP per capita in state x. We say that there is an aggregate uncertainty if the GDP per capita is uncertain, that is, if z is sensitive to variations of x. An allocation in this economy is characterized by a function c(x, θ) which is the quantity consumed in state x by agents having characteristics θ. This function is a risk-sharing rule. This allocation of risk is said to be feasible if the level of consumption per head in any state equals the quantity available per head in that state:
The expected utility that is enjoyed by agent θ with this allocation is
We say that a feasible allocation c(., .) is Pareto-efficient if there is no other feasible allocation that increases the expected utility of at least one category of agents without reducing the expected utility of any of the other categories. Consider now any positive function λ: Θ → R+, together with the corresponding maximization problem (21.4):
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_308.html [4/21/2007 1:32:49 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_309.html
Page 309 It is easy to check that any solution to this program is a Pareto-efficient allocation of risk. This is done by contradiction: suppose that the solution to program (21.4) is not Pareto-efficient. Then, by definition, there would exist an alternative feasible risk-sharing rule
such that
is weakly larger than
for all θ ∈ Θ, with a strict inequality for at least one θ of positive measure. It would imply that
which contradicts the assumption that c(., .) is a solution to program (21.4). We conclude that for any positive function λ(.), there is an associated Pareto-efficient risk-sharing rule c(., .) which is the solution to program (21.4). It can also be shown that, under risk-aversion, we get all Pareto-efficient risk-sharing rules by using this procedure.
21.2 The Mutuality Principle We now characterize the Pareto-efficient risk-sharing allocations. These allocations have two basic properties. The first one is governed by the so-called mutuality principle. 79 In a Pareto-efficient risk-sharing the consumption of each agent depends upon the state of nature only through the aggregate wealth available in that state. PROPOSITION
The intuition behind this result can be found in the way we prove it. Consider two different states of nature, y and y′, in which the wealth available per head is the same in the two states:
The mutuality principle states that for the allocation to be efficient, all agents must consume the same amount in the two states: c(y, θ) = c(y′, θ) for all θ ∈ Θ. By contradiction, suppose that there exists an agent θ0 such that c(x, θ0) is larger than c(x′, θ0). Then consider an alternative risk-sharing rule
that is exactly the same as c(., .),
except in states y and y′. More specifically, we assume that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_309.html [4/21/2007 1:32:50 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_310.html
Page 310 where p and p′ are the probability associated respectively to state y and y′. In words, agents replace their random consumption conditional to the realization of y or y′ by its expected conditional value. For simplicity we assumed here that there is a strictly positive probability that the state be y or y′. It is obvious that the alternative allocation is feasible. It is also obvious that all agents favor this alternative allocation of risk, since it yields a meanpreserving reduction in risk. This is seen by observing that for each θ ∈ Θ,
where
is a random variable that is zero if
is not y or y′, and that is otherwise distributed as
Observe that for all x. Thus is less risky than , in the sense of Rothschild-Stiglitz, for all θ. We conclude that c(., .) may not be Pareto-efficient, since is unanimously preferred to it. In short, the mutuality principle just states that all risks that can be diversified away must be done so. Said differently, agents should not gamble on events that are not statistically related to the macroeconomic risk. This is a necessary condition for Pareto-efficiency. Thus there is no loss of generality to characterize a Pareto-efficient allocation of risk by a function w(z, θ) where c(x, θ) = w(z(x), θ). In short, w(z, θ) is the consumption of agents θ in all states whose wealth per head is z. There is no doubt that this principle is violated in the real world. Most people bear risks that could be diversified in the economy. For example, real wages are not indexed on the GDP, and it is difficult to insure this risk, either on financial markets, or on insurance markets. There is a lot of reasons why our society has been unable to eliminate diversifiable risks: moral hazard, adverse selection, limited liabilities, and so on. Townsend (1994) tried to test whether the mutuality principle holds in three poor, high-risk villages in Southern India. He used panel data of individual consumptions. Without surprise, he rejected the model of efficient risk sharing in the economy as a whole, but he found that individual consumptions move together with village average consumption. More significantly, he found that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_310.html [4/21/2007 1:32:51 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_311.html
Page 311 individual consumptions are not much influenced by individual-specific events like sickness and unemployment. This indicates that some form of Pareto-improving risk sharing is organized inside each village.
21.3 The Sharing of the Social Risk Agents θ are fully insured against the social risk if w(z, θ) is independent of z. More generally, the sensitivity of w to changes of z measures the risk borne by category θ. We examine now examine how the macro risk should be allocated to the different categories of agents. 21.3.1 Decomposition of the Problem Let G denote the cumulative distribution function of the wealth per head follows:
. It is obtained from distribution F as
Using the mutuality principle, we may rewrite problem (21.4) as follows:
There is a simple intuitive way to solve this program. Let us define function v as
Obviously the z-dependent solution to problem (21.12) subject to (21.13) is a risk-sharing rule w that maximizes . It is enough to obtain the optimal solution to program (21.14) for all z to describe a Pareto-efficient allocation of risks.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_311.html [4/21/2007 1:32:51 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_312.html
Page 312 21.3.2 The Veil of Ignorance. . We can hence Here is the intuition for using this procedure. Without loss of generality, assume that interpret λ(θ)dH(θ) as a probability. When the group has to determine ex ante the rule that they will use to share the uncertain cake , obvious conflicts of interest take place in the debate. Each category θ wants to increase its share w(., θ). As suggested first by Vickrey (1960), suppose, alternatively, that agents have to determine the sharing rule w(., .) before observing their characteristics, that is, before knowing who they will be in the group. To be more precise, consider the following timing of the resolution of uncertainty: all agents are assigned the same probability λ(θ)dH(θ) to belong to type θ. Before learning their type, all members of the group vote about the sharing rule to be used. Agents have to select the sharing rule under the veil of ignorance about their characteristics. At the second stage, nature chooses in a random way the characteristics of the agents according to the distribution function H for . At the last stage, to the sharing rule that has been selected earlier.
is realized, and the allocation of the cake is made according
We see that we added an additional degree of uncertainty to the problem. The notion of the veil of ignorance transformed an interpersonal, distributional problem into a standard problem of decision of uncertainty. Indeed, at the first stage of the game, agents are all alike. They unanimously vote for the feasible allocation that maximizes their expected utility given the uncertainty about their characteristics. In fact the uncertainty about the size of the cake becomes secondary here: all agents will agree to maximize their expected utility conditional to state z. That is, the allocation of any size z of the cake that will unanimously be selected is the one that solves problem (21.14). 21.3.3 Efficient Sharing Rules of the Macro Risk Now, observe that problem (21.14) is technically equivalent to the selection of a portfolio of contingent claims, as described by program (14.1). Formally, this is seen by performing the following change of variable: define random variable as the one having cumulative distribution function K, with Using this change of variable in program (21.14) makes it equivalent to pro-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_312.html [4/21/2007 1:32:52 AM]
.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_313.html
Page 313 gram (14.1). Thus, in order to characterize the properties of an efficient risk-sharing rule, we can just recycle all properties that have been developed in chapter 14. The first-order condition of problem (21.14) is written as
where u′ is the derivative of u with respect to consumption. This condition is formally equivalent to condition (14.3), yielding w(z, θ) = C(1/λ(θ), z). This condition implies that
A Pareto-efficient allocation is such that the marginal rate of substitution between any two states of the world is the same for all agents. By properties (14.7) and (14.11a), the full differentiation of the first-order condition directly implies that
where T(w, θ) = –u′(w, θ)/u″(w, θ) and Tv(z) = –v′(z)/v″(z) are the indexes of absolute risk tolerance of respectively u(., θ) and v(.). This property of Pareto-efficient risk-sharing rules has first been obtained by Wilson (1968). It states that the share of the social risk that must be allocated to category θ is proportional to its degree of absolute risk tolerance. The larger the risk tolerance, the larger is the share of the social risk. The feasibility constraint implies that
Any reduction of wealth per head must be absorbed by an equivalent average reduction of consumption per head. Combining conditions (21.18) and (21.19) implies that
This result together with condition (21.18) yields the following proposition: 80 In a Pareto-efficient risk-sharing, the sensitivity of each agent's consumption to changes in the wealth per head in the PROPOSITION
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_313.html [4/21/2007 1:32:53 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_314.html
Page 314 economy equals the ratio of the corresponding agent's absolute risk tolerance to the average absolute risk tolerance in the economy. We now have a complete picture of how a group should share risk. The first rule is to put all individual risks in common, and to decide about a way to share the aggregate risk. This is the mutuality principle, which states that the final allocation of the consumption good should not depend on who brought what in the common pool. The second rule is as follows: if the aggregate wealth is uncertain, those who are less risk-averse should bear a larger share of the risk. If no member is risk neutral, Tv is finite, and all members should bear a share of the social risk. This is another application of second-order risk aversion that implies that nobody should be excluded from the sharing of the social risk, even if one's risk tolerance is low compared to the average degree of risk tolerance in the group. It is always optimal to allocate some of the risk to each member, because the effect of risk on welfare is a second-order effect. 21.3.4 A Two-Fund Separation Theorem It is interesting to determine the conditions under which the Pareto-efficient risk-sharing rules are linear in z. A linear risk-sharing rule is easy to implement, since it is done by organizing lump-sum, ex ante, transfers combined with assigning a fixed percentage of the size of the cake to each member. In a market economy this is done by offering to each agent a portfolio of a risk free asset and a share of all risky assets in the economy. It allows for a two-fund separation theorem. In such an economy all agents would have the same structure of their risky portfolio. Only the allocation of initial wealth into this fund of risky assets on one side, and the risk-free asset on the other side, would vary with the degree of risk aversion of the investors. When such a property of efficient allocations of risk would hold? Or said differently, when are efficient risksharing rules linear with the GDP per capita? Proposition 52 gives us the answer to this question. 81 (TWO-FUND SEPARATION THEOREM) The efficient compositions of the portfolio of risky assets are the same for all agents in the economy if and only if absolute risk tolerances be linear in w with the same slope: T (w, θ) = t(θ) + bw for all w and all θ ∈ Θ. PROPOSITION
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_314.html [4/21/2007 1:32:53 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_315.html
Page 315 This is the substance of the results obtained by Cass and Stiglitz (1970). If absolute risk tolerances are not linear with the same slope, agents will not invest in risky assets in the same way. In particular, those agents with a consumption that depends in a convex way on the GDP per capita should accept call options in their portfolio. A call option on GDP per capita with strike price z0 is a risky asset whose payoff is
which is
convex in z. 21.3.5 The Case of Small Risk per Capita It may be useful to provide an intuition for why the sharing rule (21.18) is Pareto-efficient. It will help us understanding how should the group behave in the face of uncertainty. Let us suppose that the wealth per head is initially certain and equal to z. An allocation w(z, θ) of this sure wealth is exogenously given. Suppose now that the group is considering taking a small risk per head, with mean μ > 0 and variance σ2. Should the group accept this risk? The answer will in general depend upon how this risk would be shared in the group. A possible procedure that is alternative to the one used above is to select the sharing rule that maximizes the mean of all certainty equivalents of the risk borne by the different members of the group. Let a(θ) represent the sensitivity of category's θ consumption to change in . This means that category's θ final wealth is . Feasibility imposes that . Using the Arrow-Pratt approximation, the certainty equivalent of the risk borne by category θ is measured by
The problem is thus to select the function a(.) that solves
where CE is the mean of the optimized members' certainty equivalent of the risk borne by the group. The firstorder condition to this problem is written as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_315.html [4/21/2007 1:32:54 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_316.html
Page 316 for all θ ∈ Θ. Using the feasibility constraint, we can eliminate the Lagrangian multiplier ξ to obtain
This is the same risk-sharing rule that we obtained above. Thus, at least for small risks, Pareto-efficient risksharing rules are those that maximize the sum of the certainty equivalents of the risk borne by the group's members.
21.4 Group's Attitude toward Risk Suppose that the group's members agree to share risk in order to maximize a weighted sum of members' expected utility. What should be the group's attitude toward risk in these circumstances? The answer to this question can be found in the characteristics of function v which were defined by equation (21.14). 21.4.1 The Representative Agent From the definition of value function v, it appears that Tv(z) is the relevant measure of absolute risk tolerance that must be taken into account to determine whether a risk should be undertaken by the group. Function Tv(.) is the group's absolute tolerance to risk per head. In fact the attitude toward risk of a group implementing the rule associated with weight function λ(.) is the same as the attitude toward risk of a group of identical agents v implementing the symmetric, fair allocation. They all have utility function v(.) as defined by (21.14). An agent with such a utility function is called a "representative agent." Of course, such an egalitarian group of identical agents has an attitude toward social risk per head that is identical to its member's attitude toward such a risk. This is because the group will transfer the risk unchanged to each of its members. Notice that there always exists such a representative agent when markets are complete. In general, the attitude toward risk of the representative agent depends upon the selection of the weight function λ (.). You are asked in exercise 102 to prove that an exception is when agents have linear risk tolerance with the same slope. When preferences of the representative agent are independent of the weight function, we say
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_316.html [4/21/2007 1:32:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_317.html
Page 317 that v is the representative agent in the strong sense (see Gorman 1953). 21.4.2 Arrow-Lind Theorem If the group is large enough, his behavior toward risk should be almost risk neutral. This point is made, for example, by Arrow and Lind (1970), but it is systematically used in the theory of finance. To explain this, consider an economy with n identical agents. Suppose also that the fair allocation of the consumption good is implemented: the weight function is constant. These assumptions are made for simplicity and they can easily be generalized (see exercise 103). Each agent is endowed with an initially sure amount w of the consumption good. The group is considering accepting an aggregate risk . Since this risk would be shared in a fair way, each agent would get 1/n of it. If the size n of the population is large enough, this is a small risk. Because risk aversion is a second-order effect in expected utility with a differentiable utility function, the members of the group will unanimously accept the risk as soon as is positive. That is, the group is neutral to an aggregate risk that is small in comparison to the group's size. This can also be seen by using condition (21.20) together with relaxing the assumption that the size of the population be one. Alternatively, take ∫Θ dH(θ) = n. Condition (21.20) is then written as
where y is the realization of the aggregate risk and T is the absolute risk tolerance of the identical agents in the economy. We see that as long as risk aversion is second-order (T(w) > 0), the group's degree of absolute risk tolerance tends to infinity when the size of the group tends to infinity. In short, the Arrow-Lind theorem basically means that the group's risk aversion is of the second order, exactly as individual risk aversion under expected utility. 21.4.3 Group Decision and Individual Choice We have just seen that groups unanimously accept or unanimously reject lotteries the size of the group.
that are small with respect to
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_317.html [4/21/2007 1:32:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_318.html
Page 318 This is not true in general when we relax the assumption that the risk is small, as we see now. Consider two distributions of the wealth per capita in the group, represented by random variable and . Under the assumption that the group acts against risk in order to maximize the weighted sum of expected utilities, the optimal choice is to prefer
over
if and only if
is larger than
. Suppose that we have
. Does it imply that all members are made better off with their efficient share of than with their efficient share of ? In other words, is there unanimity about the group's decision? This would be true if
That is, all agents rank lotteries in the same way. This would be true if and only if functions v(., θ) would all have the same degree of concavity, with v(z, θ) = u(w(z, θ), θ) for all (z, θ). Function v(., θ) describes the attitude towards the group's risk per head of agent θ, given the share w(., θ) of the risk that is allocated to her. We obtain that
and
Using condition (21.28), we obtain that
where function Tv is given by condition (21.20). Because of the second term in the right-hand side of this inequality, it appears that the degree of concavity of functions v(., θ) depends upon θ. In consequence, in general, there will be no unanimity in the group. The only agents that systematically agree with the group's decision are
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_318.html [4/21/2007 1:32:56 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_319.html
Page 319 . In fact all agents in the group satisfy those whose derived absolute risk tolerance, T′(w(z, θ), θ), is equal to this condition if absolute risk tolerance are linear with the same slope: T(w, θ) = t(θ) + bw. This is the only case where there can be unanimity about the ranking of lotteries in efficient groups. Another observation from equation (21.30) is that the least averse agents toward the group's risk are those whose absolute risk tolerances are increasing the most with consumption. These agents would like the group to take more risk than what it does by following the weighted–expected-utility-maximizing rule. This is because their stake on the group's risk is a convex function w(., θ) of the group's payoff z per head. It is quite paradoxical that it is the derivative of tolerance toward consumption risk, T′(., θ), and not T(., θ) itself, which is relevant for determining who is more risk tolerant toward the group's risk. This is because this risk is reallocate in an efficient way within the group. Pratt and Zeckhauser (1989) provide other insights about the complexity of efficient group decision making.
21.5 Introducing Time and Investment Taking risk is usually done by investing money today in the anticipation of an uncertain payoff in the future. Thus, taking risk has to do not only with the attitude toward risk but also with the attitude toward time. Introducing time in the model is easy if the members' utility function is time-additive. Consider a model with two dates indexed by t = 0 or t = 1. There is a single good that is perishable. At t = 0, there is uncertainty about the state of the world that will prevail at t = 1. The uncertainty is given by the state-contingent endowment at date 1. Similarly let w0(θ) denote the endowment of agents θ that is available to them at date t = 0. Let also z0 denote the wealth per head that is available at that time, i.e., . An allocation is characterized by a pair (c0(θ), c(x, θ)) where c0(θ) is the consumption of category θ at date t = 0. It is feasible if condition (21.2) is satisfied together with the date 0 feasibility constraint given by
. The lifetime utility attained by category θ is
where β is the discount factor. Using the same argument as in the static model, a feasible allocation is Pareto-efficient if it is a solution of the following class of problems:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_319.html [4/21/2007 1:32:56 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_320.html
Page 320
under the above-mentioned feasibility constraints. Let us define a new state of the world, compound random variable
takes value x0 with probability 1/β, and
. Moreover let
otherwise. Finally, let us define c(x0, θ)
= c0(θ) and z(x0) = z0. Then program (21.32) can be rewritten as
This is formally equivalent to program (21.4), with an additional state of the world that represents the date t = 0 allocation of the perishable good. Since program (21.4) is solved pointwise anyway, that does not affect at all the properties of the solution. This is another illustration of the symmetric role that are played by time and states of nature in an additive world. The mutuality principle applied in this intertemporal framework tells us that if there is a state of the world that yields the same wealth per head than at date 0, the distribution of consumption in that state should be the same as that of date 0. Turning now to the group's attitude toward time, we can use the same trick than for the analysis of the group's attitude toward risk. Let us define the representative agent's utility function v as we did in equation (21.14). Consequently the representative agent's objective function can be written as
It implies that the concavity of v does not only represent the degree of risk aversion of the group. It also represents the group's resistance to the intertemporal substitution of consumption. Therefore condition (21.20) is useful to measure the group's willingness to sacrifice the present for the future, i.e. to invest. To illustrate, suppose that z(x) = z1 > z0 for all x ∈ X. There is no uncertainty and the growth is positive. Consider the possibility for the group to make an investment with a positive sure return. Among the heterogeneous group, some agents have a high resistance to intertemporal substitution (small T). Those ones are intrinsically reluctant to invest. Others are intrinsically less resistant to this substitution (large T), and they would like to invest. The burden of the initial investment, together with its future benefits, will be mainly allocated to those with a large
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_320.html [4/21/2007 1:32:57 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_321.html
Page 321 T (see equation 21.18). As for the group's risk-taking decisions, the investment will be accepted if it increases v (z0) + βv(z1). As seen in chapter 15, the concavity of v given by condition (21.20) will be instrumental to solve this investment problem. However, the unanimity about this decision will be attained only in groups of HARA agents with the same slope.
21.6 A Final Remark: The Concavity of the Certainty Equivalent Functional In public economics it is a tradition to add up willingnesses to pay—and thus certainty equivalents—to evaluate the efficiency of public policies. We illustrated this method in section 21.4, where we showed that, when the aggregate risk is small, the Pareto-efficient risk-sharing arrangements maximize the sum of the certainty equivalent consumptions in the group. This is not true in general, as we show in this section. We show in particular that risk-sharing arrangements that maximize the sum of the certainty equivalent consumptions do not even need to satisfy the mutuality principle. To keep things simple, consider a pool composed of two types of agents, all with identical preferences represented by utility function u(.). There is a proportion λ of agents of type 1 in the pool. We consider the case of a finite number of states x = 1, . . . , n, and we let denote ω(x, θ). Under the veil of ignorance, the expectedutility-maximizing agents would unanimously select the fair risk-sharing arrangement with for all x, which is Pareto-efficient. Would they behave in the same way if their objective would be to maximize their certainty equivalent consumption rather than their expected utility? The certainty equivalent Ce of consumption plan C = (c1, . . . , cn) is measured by
or equivalently, by
We want to guarantee that the fair risk-sharing arrangement λw1 + (1 – λ)w2 dominates the autharky solution when the social
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_321.html [4/21/2007 1:32:58 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_322.html
Page 322 welfare function is the weighted sum of certainty equivalents, λCe(C1) + (1 – λ)Ce(C2). This is true if
or equivalently if the certainty equivalent functional is concave in the space of state-contingent plans. Condition (21.38) obviously holds in the case where is independent of x and Eω1 = Eω2 because of risk aversion. Thus, functional Ce cannot be globally convex under risk aversion. But it is not necessarily concave. The following Lemma is useful to link the concavity of Ce to some properties of the utility function. It is theorem 106 in Hardy, Littlewood, and Pólya (1934), and proposition 1 in Polak (1996). LEMMA
8 Consider a function g from R to R, twice differentiable, increasing and concave. Consider a vector with
and a function f from Rm to R, defined as
Define function h such that h(t) = –g′(t)/g″(t). Function f is concave in Rm if and only if h is weakly concave in R. Proof See the appendix at the end of this chapter. Using this lemma with g ≡ u, we directly obtain that the certainty equivalent is a concave function of the payoff vector if the absolute risk tolerance is linear or concave in wealth. A corollary of this is that the certainty equivalent Ce of a lottery , with Ce being defined by
is a concave function of the initial wealth w0 under this condition. We summarize these findings in proposition 82. 82 The certainty equivalent consumption Ce defined by condition (21.37) is a concave function of the vector of contingent consumptions if the absolute risk tolerance is concave or linear in wealth. It implies in particular that PROPOSITION
1. risk-sharing arrangements that maximize the sum of certainty equivalent consumptions in a pool need not to satisfy the mutuality principle, except if absolute risk tolerance is weakly concave.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_322.html [4/21/2007 1:32:58 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_323.html
Page 323 2. the risk premium associated to any lottery is convex in wealth under nonconvex absolute risk tolerance. We know that DARA implies that the risk premium associated to any lottery is a positive and decreasing function of wealth. There are some reasons to believe that it is decreasing at a decreasing rate, i.e., that the risk premium is a convex function of wealth. This may not be true when absolute risk tolerance is a convex function of wealth.
21.7 Conclusion. A feasible allocation is Pareto-efficient if it would be selected by agents under the veil of ignorance. In the ignorance about their future type, agents bear a sequence of two independent risks. First, they don't know who they will be. Second, they don't know which state of the world will occur. Therefore the risk-sharing problem can be decomposed into a sequence of state-specific problems where one must decide how to allocate the sure aggregate wealth in the corresponding state into individual consumptions. In each of these problems, the only source of uncertainty comes from the veil of ignorance. We showed that each of these problems is equivalent to an Arrow-Debreu portfolio problem. This analogy has been very helpful to describe the properties of the efficient risk-sharing rules and of the efficient consumption smoothing rules among the different agents. In particular, each agent will bear a share of the aggregate risk that is proportional to her absolute risk tolerance. There is a similar rule for the sharing of the fruits of growth in the economy. These rules imply that the group's degree of absolute tolerance to risk per capita is equal to the mean individual absolute risk tolerance in the corresponding state.
21.8 Exercises and Extensions Exercise 102 Suppose that all agents in the economy have linear absolute risk tolerance with the same slope: T (w, θ) = t(θ) + bw. • Show that all Pareto-efficient sharing rules are linear: w(z, θ) = k1(θ) + k2(θ)z. • Show that the absolute risk tolerance of the group is independent of the selected Pareto-efficient sharing rule.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_323.html [4/21/2007 1:32:59 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_324.html
Page 324 • Suppose that the group faces a standard portfolio problem: optimal α that should be selected by the group as a whole.
. Show that there is unanimity about the
Exercise 103 As in section 21.3.5 we consider a group who has to share a small risk whose variance is σ2. Show that • the certainty equivalent of the risk borne by category θ is
• the optimized certainty equivalent per head of risk up equals
• there is unanimity about the group decision towards risk . Exercise 104 (Leland 1980) We have shown above that Pareto-efficient rules are linear when all agents have linear absolute risk aversion with the same slope. Any linear sharing rule can be implemented by using just a riskfree asset and a second asset whose payoff is the GDP per capita. Alternatively, a consumption plan that is convex in the GDP per capita can be obtained by a bundle of call options (with various strike prices) whose underlying asset is the GDP per capita. Show that those who should purchase call options are those whose absolute risk tolerance is most sensitive to changes in consumption. Exercise 105 (Gollier 1996b) Consider an economy with two agents, θ = 1 or 2, having the same utility function u. There are two independent risks in the economy. The first one can be shared among the two agents, whereas the other is borne by only agent θ = 1. The sharing rule on the first risk cannot be made dependent to the realization of the nontradable risk. The group implements the risk-sharing rule that maximizes the sum of the expected utility of the two agents. Show that agent θ = 2 bears a larger share of the tradable risk (i.e., ∂w(z, θ)/∂z is larger for θ = 2 than for θ = 1, for all z) if and only if u exhibits decreasing absolute prudence.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_324.html [4/21/2007 1:33:00 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_325.html
Page 325
21.9 Appendix Proof of Lemma 8 Let us first observe that it is enough to look at the case m = 2. Indeed, defining we have
Since ∂fm/∂xj ≥ 0, applying the result by recurrence would yield the result. Observe also that
and
or equivalently
Let us define function k as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_325.html (1 of 2) [4/21/2007 1:33:00 AM]
,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_325.html
We can then rewrite the above equality as follows:
Similarly we have
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_325.html (2 of 2) [4/21/2007 1:33:00 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_326.html
Page 326 For m = 2, the determinant |∂2f/∂x2| of the Hessian is
Since g(f(x)) = q1g(x1) + q2g(x2), with q1 + q2 = 1, the right-hand side of the equality above is positive (resp. negative) if k is concave (resp. convex). Notice also that
if k is positive and concave. Thus, if k is concave, then
that is, ∂2f/∂x2 is semidefinite negative and f is concave. If k is convex, the determinant of the Hessian matrix is negative, implying that f is neither concave nor convex. Observe now that k(g(t)) = g′(t)h(t) with h(t) = –g′(t)/g″(t). It yields
and
We conclude that k is concave if and only if h is concave. This concludes the proof.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_326.html [4/21/2007 1:33:01 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_327.html
Page 327
22 The Equilibrium Price of Risk and Time In this chapter we examine the Pareto-efficient allocation of risk that is generated by competitive markets for contingent claims. It will allow us to examine the driving forces that govern the pricing of risk and time in such an economy. In fact the model that we present in this chapter is a two-period version of the continuous-time capital asset pricing model (CCAPM). The pioneering paper in this literature is the famous contribution of Arrow (1953).
22.1 An Arrow-Debreu Economy We start with a static model. Time will be added to this picture later in this chapter. The description of our economy is exactly the same as in section 21.1. We now assume that at the beginning of the period, viz before the . This is done by realization of , agents are in a position to trade their state-contingent endowment opening a complete set of markets for contingent claims. Agents are allowed to buy or sell any contingent claim that promises the delivery of one unit of the consumption good if and only if state x is realized. Following the notation that we used in chapter 13, let π(x) be the price of such a contract per unit of probability. Markets for contingent claims are assumed to be competitive. This means that consumers take prices as given, and that prices clear markets. The problem of agent θ is written as
where c(x, θ) is the consumption of agent θ in state x. In other terms, c(x, θ) – ω(x, θ) is his demand for the contingent claim associated
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_327.html [4/21/2007 1:33:01 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_328.html
Page 328 to state x. This problem is equivalent to problem (13.3), where at the market prices.
is the wealth of the agent evaluated
An allocation in this Arrow-Debreu economy is characterized by a couple (c(., .), π(.)). Such an allocation is a competitive equilibrium if 1. c(., θ) is the solution to program (22.1) given pricing kernel π(.), for all θ ∈ Θ; 2. all markets clear,
We hereafter assume that such an equilibrium exists and that it is unique. The rest of this chapter is devoted to the analysis of the properties of this equilibrium.
22.2 Application of the First Theorem of Welfare Economics One of the first achievements of mathematical economics was to prove that the competitive equilibrium is Paretoefficient. This framework involved writing the first-order condition associated with agent θ′ s portfolio problem (22.1):
It implies that
for all xi and xj in X. This means that the competitive equilibrium is such that the marginal rate of substitution between any two states of the world is the same for all agents, and is equal to the ratio of the corresponding state prices. But, as stated by equation (21.17), this property characterizes a Pareto-efficient allocation of risk. Thus the competitive equilibrium is Pareto-efficient. All properties of Pareto-efficient risk-sharing arrangements that we derived in the previous chapter hold for the competitive equilibrium. It means, for example, that if there are two states yielding the same wealth per head , all agents will consume the same quantity of the consumption good in these two states: c (x1, θ) = c(x2, θ). By condition (22.5), it implies that the equilibrium prices will be the same in these two states: π
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_328.html (1 of 2) [4/21/2007 1:33:02 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_328.html
(x1) = π(x2). Thus there is no loss of generality to characterize a state of the world by the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_328.html (2 of 2) [4/21/2007 1:33:02 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_329.html
Page 329 wealth per head that is available in that state (as we did in the previous chapter), or by the equilibrium price in that state (as we did in chapter 13).
22.3 Pricing Arrow-Debreu Securities Because the competitive equilibrium is Pareto-efficient, we know that there exists a representative agent that completely characterizes this economy. This is the competitive representative agent, whose existence is shown in Constandinides (1982). The competitive representative agent v with the mean GDP per capita in each state of the world would not want to trade in financial markets if he faces the equilibrium price kernel. By definition, we must have that
Because competitive prices are defined up to a multiplicative factor, we can take v′(z(x)) = π(x) for all x. The pricing kernel is just the marginal utility of the representative agent in autharky. Thus, obtaining the pricing kernel π(.) of the heterogenous economy directly yields the characterization of the corresponding representative agent of this economy. Now compare two economies. The first one is the one described earlier in this chapter, with potentially heterogeneous functions u(., θ) and ω(., θ). It yields competitive price kernel π(.). The second economy is an economy where people are all alike, with the same utility function v(.), with v′(z(x)) = π(x) for all x, and the same . Obviously the homogeneity of the population in this second state-dependent endowment economy eliminates any source of mutually advantageous transaction. It implies that the autharcic allocation in which agents just consume their endowment in every state is the competitive equilibrium of this economy. It implies that the price kernel in this second economy is given by the same π(.). By construction, these two economies have the same equilibrium price kernel. Thus the v function is sufficient to fully characterize the equilibrium pricing rule in the original economy. Consider an economy in which all agents are alike, with the . The competitive pricing kernel which same utility function u and the same state-contingent endowment sustains such an allocation is such that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_329.html [4/21/2007 1:33:03 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_330.html
Page 330 We presented in section 21.4 the technique to build the utility function of the representative agent from the utility functions of the different categories of agents in the heterogenous economy. We showed there how the risk aversion of actual agents is transferred to the representative agent's attitude toward risk. More specifically, the absolute risk tolerance of the representative agent is the mean absolute risk tolerance of actual agents in each state:
where c(x, θ) is the consumption of agent θ in state x, at the competitive equilibrium. If one agent is risk-neutral, so is the representative agent. Since π(x) = v′(z(x)), and given the risk aversion of the representative agent, the equilibrium price of a statecontingent asset is decreasing with the wealth per head available in that state. This induces the agent to reduce his consumption in states of relative scarcity of the consumption good. Let be the equilibrium state price a function of the wealth per head. The sensitivity of state-prices to differences in the state-wealth equals the absolute risk aversion of the representative agent:
The larger the absolute risk aversion, the larger the price sensitivity to differences in GDP. When the risk aversion is large, the agent is ready to pay much to smooth his consumption across states. Large price discrepancies in state-contingent prices are then necessary to induce him to bear the social risk.
22.4 Pricing by Arbitrage. In the previous section we provided a simple way to price Arrow-Debreu assets. But most assets in the real world are not Arrow-Debreu assets. In this section we explain how to price any asset that is not an Arrow-Debreu asset. This is done by using the technique of arbitrage. An asset in this static economy is entirely described by the state-contingent payoff that it will generate at the end of the period. Let qj(x) be the payoff of an asset j in state x. Knowing the competitive price π(.) of the ArrowDebreu assets, we can derive the com-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_330.html [4/21/2007 1:33:03 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_331.html
Page 331 petitive price Pj of asset j in the following manner: build a portfolio of Arrow-Debreu assets with, for each state x ∈ X, qj(x) units of the Arrow-Debreu asset associated to state x. The market value of this portfolio is
.
This portfolio has exactly the same risk profile as asset j itself. Thus we now have two strategies of taking risk . It implies that the cost of using these two strategies must be the same. Suppose, by contradiction, that the equilibrium price of asset j is larger than the market value of the equivalent portfolio of Arrow-Debreu assets: . A risk-free arbitrage that has a positive payoff is obtained by selling short asset j, and buying the portfolio of Arrow-Debreu assets that duplicates it. This strategy has no effect on the contingent consumption plan of the investor, since the promise to pay qj(x) due to the short-selling of asset j is just compensated by the payoff generated by the portfolio of contingent claims. Thus this strategy is risk free, with a zero net payoff with probability 1 at the end of the period. But it allows the investor to relax his budget constraint, since he can extract a sure profit from this strategy. This cannot be an equilibrium, as all agents will duplicate this strategy like crazy. It would raise the supply of asset j up to infinity. The symmetric argument by contradiction can be made if the price of the asset is less than the market value of the portfolio which duplicates must equal the market value of the portfolio of Arrow-Debreu it. We conclude that the price of any asset assets which has the same risk profile, namely
This equation is equivalent to equation (5.4). This should not be a surprise: since we assumed in chapter 5 that all agents are alike, no trade is an equilibrium, independently of the complete or incomplete structure of the markets. The market structure becomes important only when the final allocation is affected by it. In fact the model of chapter 5 is not tractable when agents have heterogenous preferences, or different wealth levels. There are two branches in the pricing theory in finance. The first branch is based on the optimal portfolio management of investors. Demand curves and supply curves for the different assets are built from solving the investors' expected utility maximization problem. This branch follows the line drawn in the previous section. An example of this is the capital asset pricing theory. The second branch is pricing by arbitrage, in which the price of a subset of assets is taken
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_331.html [4/21/2007 1:33:04 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_332.html
Page 332 as given. The equilibrium price of all other assets whose payoffs can be duplicated by this subset are derived by arbitrage. Option pricing formulas are typically obtained by following this approach. For example, the equilibrium price of a call option on asset j with strike price q0 equals max . No other information on preferences than that already contained in the prices of the subset is needed to derive such prices. Notice that we can now relax the assumption that there exists a complete set of Arrow-Debreu securities in the economy. We can suppose alternatively that there exists a set J of assets with payoff q(x, j) for asset j in state x. Suppose also that for each state x ∈ X, there exists a portfolio αx(.), with αx(j) characterizing the number of units of asset j in the portfolio, such that
Portfolio αx duplicates the Arrow-Debreu security associated to state x. If there exists such a portfolio for each x ∈ X, any state-contingent plan can be attained by a specific portfolio of real assets; that is, the market is complete. Under this assumption we say that the market "spans" the set of all possible consumption plans. Notice that the price of state x can be obtained from the price of real assets in J. By arbitrage, we have
22.5 The Competitive Price of Risk What are the determinants of the equilibrium price of risky assets? To answer this question, let us normalize to 1 the price of the asset promising an unconditional delivery of one unit of the good:
In this section the risk-free asset is the numéraire. It implies that the equilibrium price of the asset promising a payoff profile
equals
The equilibrium price of any asset equals its expected payoff plus a risk premium. This risk premium is measured by the covariance of
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_332.html [4/21/2007 1:33:05 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_333.html
Page 333 the payoff of the risky asset with the price kernel. This means that only the undiversifiable risk in has a positive risk premium. This classical result of the modern theory of finance has simple intuition. Consider an . This means that adding a marginal amount of this asset asset whose payoff is independent of the social risk in everyone's portfolio increases everyone's expected utility in proportion of the expected payoff of the asset:
if and are independent. This is another consequence of the fact that expected utility exhibits secondorder risk aversion. Because a marginal increase in the holding of this asset does not affect the portfolio's risk, no risk premium will be attached to this asset. On the contrary, if a risk premium would be included to the price, all investors would increase their demand for it. The risk premium equals the covariance of the state-contingent price with the payoff function of the asset. Since the sensitivity of the price kernel to changes in the wealth per head equals the absolute risk aversion of the representative agent, we understand that the risk premium will be an increasing function of the degree of risk aversion and of the covariance of the wealth per head and the payoff of the asset. This can be seen by approximating v′(z(x)) in equation (22.13) in a linear way around z0, where z0 is defined by . It yields
where Av = Av(z0) is the market's degree of risk aversion, and
is the market risk. This is the standard capital
asset pricing (CAPM) formula, which states that the equilibrium price of the risk associated to an asset equals the product of the market's degree of risk aversion by the covariance of the asset risk with the market risk. CAPM formula (22.15) is usually written in a slightly different way. Consider the price PM of an asset that exactly duplicates the risk profile of the market. Its price equals
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_333.html [4/21/2007 1:33:05 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_334.html
Page 334 Using the same linear approximation as above, we obtain
where and are respectively the mean and the variance of the market payoff. Eliminating Av from (22.15) and (22.17) yields
where is the so-called market price of risk, and is the so-called beta of asset j. The market price of risk, ΠM, is the expected return of equity. The equity premium is the market . The mean price of risk divided by the expected price of equity. It is approximated by value of the beta is one in the economy. It has the same sign as the covariance of the asset risk with the market risk. Both ΠM and the betas can quite easily be estimated by using market data.
22.6 The Competitive Price of Time Up to now in this chapter, we considered a static model. This did not allow us to examine the determinants of the risk free interest rate, that is, the price of time. To do this, we reexamine the two-date model that we used in section 21.5. The (perishable) wealth per head is z0 at date t = 0, and z(x) in state x at date t = 1. As explained in that section, adding time to the initial picture does not change the representative agent result: to any heterogeneous or unequal economy, there is an homogeneous and fair economy that duplicates its behavior toward risk and time. Let v be the utility of the representative agent. His objective is to maximize his lifetime utility
, where β represents here the discount factor. The equilibrium conditions are c0 = z0
and c(x) = z(x), for saying that the consumption is not transferable across time or states. The difference with what we did above is the possibility to consume at date t = 0. The Arrow-Debreu security associated to state x guarantees the delivery of one unit of the consumption good in state x, date 1, against the , because payment of price π(x) at date 0. Obviously we cannot normalize the price in such a way that the numéraire is now the consumption good itself: at date 0, promises of delivery at date 1 are exchanged against some units of the good today. It implies that π(.) contains information on the attitude
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_334.html [4/21/2007 1:33:06 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_335.html
Page 335 towards risk and time at the same time. The equilibrium price of time can be measured from the number of units of the consumption good that can be obtained at date t = 1 for certain when giving up one unit of the consumption good at date t = 0 on the credit market. Let 1 + r be this number. It is an equilibrium if, at the margin, the representative agent is indifferent upon signing up a credit contract, from the autharcic allocation. This is the case if
or equivalently, if
This condition is equivalent to equation (17.5). We now have a clear picture of the determinants of the price of time in an Arrow-Debreu economy, as they have been stated by Bohm-Bawerk. First, the larger the impatience of agents (β small), the larger is the equilibrium interest rate to induce them to select a feasible consumption pattern over time. This is a determinant generated by a pure preference for the present. There is a second effect due to the growth of GNP per head over time. When the expected growth rate is large, households are willing to transfer more of the future purchasing power into today consumption, for an income-smoothing reason. To induce them not to do that, the interest rate must be increased. By how much depends on the degree of concavity of the utility function of the representative agent. A third determinant of the equilibrium risk-free rate is uncertainty about the growth rate of incomes. The fact that future incomes are uncertain makes prudent agents more willing to accumulate precautionary savings to forearm themselves against potential losses of their incomes. In order to induce them not to do that, the risk-free rate must be reduced. By how much depends on the degree of convexity of the marginal utility of the representative agent.
22.7 Spot Markets and Markets for Futures As said above, the introduction of time makes the price kernel π(.) a little more complex to interpret, because time and risk are intertwined in it. Different prices can be inferred from it. In addition to the risk-free rate that we examined above, we obtain that the price of
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_335.html [4/21/2007 1:33:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_336.html
Page 336 equity in terms of today's consumption good is Pe that must satisfy the following equilibrium condition:
which implies that
Pe is the spot price of equity. The pure preference for the present and the expectations about the growth of the economy will affect the spot price of equity. Because equity on the spot market yields an uncertain payoff tomorrow against a payment today, the price of equity, Pe, is a measure that combines the price of risk and the price of time. The equilibrium price of risk can be obtained by looking at the price of a financial asset called a "future." A "future" contract on equity is exactly the same as the equity itself, except that the payment of the price will be called at date t = 1, not at t = 0. Thus a future contract organizes the exchange of a sure amount of the good tomorrow against an uncertain amount of the same good tomorrow. Its price is thus the price of risk in the economy. The price of this future is denoted Pf. It must satisfy the following equilibrium condition:
which implies that
Not surprisingly, we obtain that Pf = Pe(1 + r). Because Pf is the equilibrium price of risk in this economy, we may inquire about whether an increase in risk aversion of the representative agent reduces this price. Let v1 be more risk-averse than v2. More risk aversion reduces the equilibrium price of equity if
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_336.html (1 of 2) [4/21/2007 1:33:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_336.html
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_336.html (2 of 2) [4/21/2007 1:33:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_337.html
Page 337 Let
denote random variable
. The property above may then be rewritten as
This condition is formally equivalent to condition (6.25). Using proposition 7, we conclude that an increase in risk aversion reduces the equilibrium price of equity in the economy.
22.8 Corporate Finance in an Arrow-Debreu Economy It is easy to determine the rules that should determine investment strategies of investors and entrepreneurs in an Arrow-Debreu economy. For the sake of simplicity, let us go back to the static version of the model, with consumption taking place only at the end of the period. Consider an agent θ who has an investment project that , where α is a decision variable that affects the risk profile of the would generate a state contingent payoff investment. Should she invest in this project, and if yes, how much should she invest? Without any financial markets or any other risk-sharing arrangement, she would select the α that solves
Thus the degree of risk aversion of agent θ and the riskiness of the project will be the determinants of her investment in the project. Obviously the existence of financial markets will affect how entrepreneurs will select their investment strategy. The problem of corporate finance is not only to determine where and how much to invest but also how to finance the selected projects. It is important to observe that the possibility to share the risk of the project with other investors greatly affects the optimal investment strategy. In the previous chapter we have seen that it is socially efficient to aggregate all risks in the economy in order to optimize diversification. We have also seen in this chapter that it will be in the private interest of all agents to do so if financial markets are competitive, frictionless, and complete. This can be easily seen by observing that the true problem of our investor θ is to select at the same time her financial investment strategy on contingent claim markets and the characteristics of her physical investment in the project. This problem is written as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_337.html [4/21/2007 1:33:08 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_338.html
Page 338
Observe that we assume that the investor is a price-taker: she does not take into account of the possibility that her investment may affect equilibrium state-prices. It is clear that the above maximization program can be decomposed into two steps: 1. Select the α that maximizes the market value of the project
.
2. Rebalance the investor's risk profile by an active strategy on financial markets. This is because it is optimal to select the investment strategy that maximizes its market value in order to compensate the risk taken by investing in the project by trading on financial markets. The riskiness of the project does not directly affect the risk on consumption of the investor. We see that the only thing that is important in determining whether a project should be implemented is the statistical relationship that exists between the payoff of this project and the market risk. Observe that the notion of a firm does not exist in an Arrow-Debreu economy, since the value of a bundle of different investment projects that a firm could have is just the sum of the value of these projects. Mergers and acquisitions are just irrelevant for improving the market value of a firm. There are several ways for an entrepreneur to rebalance the risk profile of her income. She can do it by purchasing insurance contracts for specific contingencies, which are nothing else that contingent claim contracts. Or she can purchase a portfolio of risky assets whose returns are negatively correlated with the return of her investment project. She can also sell shares of her project on financial markets, creating a stockholders company. Finally, she can issue bonds or borrow from banks. The point is that these strategies have no effect on the market value of the project. Again, the value of the firm is just the sum of the market value of its component. Because the market did already diversify risks that can be washed out in the economy, purchasing a contingent claim contract on the market to reduce the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_338.html [4/21/2007 1:33:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_339.html
Page 339 riskiness of the firm will not affect its value. The increase in the market value of the firm due to this risk reduction is just compensated by the market price that the firm has to pay for this contingent claim contract. The shareholders of the firm are just indifferent about the financial engineering used by the firm to finance its investment. Any financial arrangement that could be done by the firm can be undone by shareholders purchasing the symmetric arrangement on financial markets. This is the essence of the famous Modigliani-Miller theorem. The only thing that matters for shareholders is to make sure that managers select the investment project—the α— that maximizes the market value of the firm. That may be a problem if managers have private informations about the risk profiles of the investment projects that are available to the firm. Most of the developments in corporate finance for the last twenty years are contained in this sentence. Notice that price-taker investors without asymmetric information are unanimous about which risks firms should undertake.
22.9 Conclusion There is no doubt that the functioning of modern economies relies heavily on the existence of efficient financial markets. Their role is to allocate physical capital optimally among a very large set of potential risky investment projects. They do not only allow for the elimination of diversifiable risks through their dissemination in individual portfolios; they also provide a lot of information about how society values risk and time. This information contained in the equilibrium prices is useful for entrepreneurs to determine whether a new project could increase the market value of the firm. We have seen in this chapter the determinants of these prices when markets are complete. They depend on the degree of risk aversion and prudence of the representative agent. The problem is thus to characterize the utility function of the representative agent. When the economy is homogeneous, with identical agents and with an equal distribution of endowments, this is nothing else than the utility function of the identical agents. In that circumstance, the results presented in this chapter are identical to those obtained in chapters 5 and 22. An aggregation problem arises when agents are heterogeneous, either in their preferences or in the value of their initial endowment. Except
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_339.html [4/21/2007 1:33:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_340.html
Page 340 in very specific instances, the search for the representative agent is a complex exercise. We will explore this subject in the next chapter.
22.10 Exercises and Extensions Exercise 106 Consider an economy whose consumers have linear risk tolerance with the same slope: T(w, θ) = t (θ) + bw. Characterize the pricing kernel of such an economy. Show that it is independent of the initial distribution of wealth in the economy. Exercise 107 The theorem of Modigliani-Miller states that the value of firms is independent of their financial structure. We illustrate this by the following example: Consider an entrepreneur who has an investment project yielding state-dependent payoffs qj(.). This investment costs I ex ante. The entrepreneur has no wealth ex ante; that is, he must raise the money required for the investment on financial markets. The price kernel is π(.). • Suppose that the entrepreneur raises the money on financial markets by emitting shares on qj. What is the share of the project that the entrepreneur must sell to finance the project? What is the market value of the residual property rights of the entrepreneur on this project? Is it optimal to implement the project? • Suppose alternatively that the entrepreneur raises the money with a standard debt contract that forces the entrepreneur to use the cash flow qj to repay the debt up to its face value D. What is the equilibrium value of D necessary to finance I? What is the market value of the residual property rights of the entrepreneur on this project? Is it optimal to implement the project? • Show that the market values of the residual property rights of the entrepreneur are the same. This means that the entrepreneur is indifferent to finance the project by emitting stocks or bonds. Moreover the financial structure of the firm does not affect its investment behavior. Exercise 108 (Arrow 1965) Consider an economy with n + 1 consumers. They are all endowed with sure wealth . The utility function of consumer 1 is u1, whereas the utility function w. But consumer 1 faces a risk of loss of all other consumers is u2. Consumer 1 can make a risk-sharing arrangement with a single consumer u2. The
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_340.html [4/21/2007 1:33:10 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_341.html
Page 341 risk-sharing contract, or insurance contract, takes the form of a premium P to be paid by consumer u1 ex ante, and an indemnity schedule I(.), where I(L) stipulates the indemnity to be paid by agent u2 when agent u1 incurs loss L. The indemnity schedule is constrained to be nonnegative: I(.) ≥ 0. Moreover there are transaction costs ex post: indemnity I costs C(I) ≥ I dollars to the insurer. • Show that the competitive premium associated indemnity schedule I(.) must satisfy the following constraint:
• Suppose that agents u2 are risk-neutral and that the cost function is linear: C(I) = (1 + k)I. Show that the equilibrium insurance contract contains a straight deductible d, meaning that I*(L) = max(0, L – d). • Suppose that agents u2 are risk-averse and that the cost function is linear. Show that the equilibrium contract contains a coinsurance clause above the deductible. The marginal rate of coverage above d is
• (Raviv 1979) Suppose that agents u2 are risk-neutral. Suppose also that the cost function satisfies the following condition: C′(I) > 1 and C″(I) > 0. Show that the equilibrium contract contains a straight deductible plus coinsurance above the deductible. What is the rate of insurance coverage I′? Exercise 109 (Franke, Stapleton, and Subrahmanyam 1998) Consider an economy with two agents having the same CRRA utility function u. Agent θ = 1 bears a zero-mean nontradable risk . All other risks are tradable; they yields a payoff per capita. There exists a complete set of competitive markets for claims contingent to . • Show that the agent-1 equilibrium portfolio of claims contingent to is such that c1(z) = k1z + k2ψ(c1(z)), where k1 and k2 are positive constants, and ψ is the precautionary premium associated to risk . • Show that c1(.) is convex if ψ(.) is convex. • Using lemma 8, prove that ψ(.) is convex. • We conclude that the agent with the background risk buys call options at equilibrium.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_341.html [4/21/2007 1:33:11 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_342.html
Page 342
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_342.html [4/21/2007 1:33:11 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_343.html
Page 343
23 Searching for the Representative Agent As shown in the previous chapter, the computation of the equilibrium price of financial assets is simple when the preferences of the representative agent are known. The difficulty lies in fact in characterizing these preferences. Often these characteristics are not the mean of the corresponding characteristics in the population. For example, the degree of absolute risk aversion of the representative agent is not the mean value of the actual absolute risk aversion in the population. The same remark holds for marginal utility, prudence, and so on. In this chapter we reexamine this aggregation problem. We consider exactly the same model as in the previous chapter: each agent θ is characterized by her utility function u(., θ) and by a endowment ω(z, θ) that is a function of the GDP per capita z that will prevail. Observe that we simplified the notation by assimilating the state of world x to the GDP per capita z that prevails in that state. By the mutuality principle, we know that this is without loss of generality. We suppose that there is a complete set of competitive, frictionless, markets for contingent claims. We suppose that a competitive , there exists an equilibrium distribution of consumption equilibrium exists. Then, for any distribution of that is characterized by function c(z, θ) and a representative agent characterized by the utility function v. They are obtained by solving the following system of equations that we derived in the previous chapter:
As a reminder, the first equation is the first-order condition of the decision problem of agent θ associated to the Arrow-Debreu security z, whereas the second equation is her budget constraint. The last
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_343.html [4/21/2007 1:33:12 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_344.html
Page 344 condition is the market-clearing condition for the Arrow-Debreu security associated to state z. The unknowns are functions c, v′, and ξ. Once this system is solved, we can easily derive the equilibrium prices of the ArrowDebreu securities, since we have that π(z) = v′(z).
23.1 Analytical Solution to the Aggregation Problem. In general, there is no simple way to characterize the attitude toward risk of the representative agent. There are four exceptions, however. The first exception has already been encountered in this book: when there is no evidence of heterogeneity in the economy, that is, where u and ω are independent of θ. Then the system above is solved trivially by setting
In this trivial case, v coincides with u. The second exception is where there exists a risk-neutral agent in the economy. System (23.1) is solved by setting v′(z) = 1 for all z in such a case. The third exception is where the utility functions are CARA: u′(c, θ) = exp(–θc). Then the system of equations (23.1) is solved by setting
where c0(θ) represents the budget constraint of agent θ, and θ* is defined as follows:
When all agents have CARA preferences, the representative agent will also be CARA. Notice that we could have proved this result directly by using property (22.7) of the representative agent's absolute risk tolerance. Her constant degree of absolute risk aversion is the harmonic mean of the different coefficients of absolute risk aversion in the population. Since the harmonic mean is smaller than the arithmetic mean, we conclude that in showing CARA preferences, the representative agent discloses a degree of absolute risk aversion that is smaller than the mean absolute risk aversion in the popula-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_344.html [4/21/2007 1:33:12 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_345.html
Page 345 tion. The intuition is simple: those who are less risk-averse will seek to purchase a larger share of the aggregate risk. So their degree of risk aversion will be overrepresented in the aggregation. This result will tend to reduce the risk aversion of the representative agent. The third exception may be extended to the case where all agents have linear risk tolerance with the same slope: T (w, θ) = θ–1 + w/γ. The CARA case is obtained when γ tends to infinity. This more general assumption is equivalent to
The first-order condition yields
Taking the expectation of both sides of this equality over , using the market-clearing condition, yields
where θ* is defined by condition (23.4). Finally we obtain that
Thus, when all agents have linear absolute risk tolerances with the same slope, the representative agent also has linear absolute risk tolerance. In particular, if all agents have the same HARA utility function ( degenerated), then the representative agent has the same attitude toward risk—and the same elasticity to intertemporal substitution—as the original agent in the economy, even if the distribution of wealth is unequal. Wealth inequality will have no effect on asset pricing in this economy.
23.2 Wealth Inequality, Risk Aversion, and the Equity Premium From now on, we assume that all agents have the same utility function u(.). Our objective will be to determine the effect of wealth inequality on the degree of risk aversion of the representative agent. We just proved that the distribution of wealth has no effect if u is HARA. More generally, we have that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_345.html [4/21/2007 1:33:13 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_346.html
Page 346 Along with the market-clearing condition
, this implies that Tv(z) is larger than T(z) if T is convex,
whereas Tv(z) is smaller than T(z) if T is concave. We conclude that wealth inequality increases (resp. reduces) the degree of risk aversion of the representative agent if the actual agents' absolute risk tolerance is concave (resp. convex). An application of this result is found in the analysis of the effect of wealth inequality on the equity premium. The equity premium with wealth inequality equals
where v is the solution of system (23.1). This should be compared to the equilibrium price in the absence of . We know that the equity premium is an increasing function of wealth inequality which equals the degree of risk aversion. Thus in order to sign the effect of wealth inequality on the equity premium, we just need to know whether v is more or less concave than u. So we end up with the following proposition: 83 Assume that all agents have the same utility function u. If absolute risk tolerance of u is concave (resp. convex), then the equilibrium price of equity in an unequal economy is smaller (resp. larger) than the equilibrium price in the egalitarian one, with the same aggregate uncertainty. PROPOSITION
This proposition generalizes what we already know when absolute risk tolerance is linear, which corresponds to HARA utility functions for which wealth inequality has no effect on asset pricing. To obtain this result, we do not need to make any assumption about whether risk aversion is increasing or decreasing. But under increasing absolute risk tolerance, the intuition is as follows: introducing wealth inequality in the economy alters the aggregate demand for equity in two ways. First, poorer agents are more risk averse. This reduces their demand for equity. Second, wealthier people have a larger demand for equity, since they are less risk averse. Under linear risk tolerance we know that the demand for equity is linear in wealth. This implies that the two effects exactly compensate each other, so the aggregate demand at the original equilibrium price is not affected. On the contrary, when absolute risk tolerance is concave, the demand for equity is concave in wealth, at least for a small aggregate
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_346.html [4/21/2007 1:33:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_347.html
Page 347 risk (see equation 4.5). This implies that a mean-preserving increase in the distribution of initial wealth reduces the aggregate demand for risk in this case. It reduces the equilibrium price of stocks. Although this technique cannot be extended to larger risks, proposition 83 indicates that the same result holds in general. A consequence of this analysis is that the equity premium is increased due to wealth inequality if T is concave, whereas it is decreased if T is convex. Following Weil (1992), suppose that an outside observer is willing to confront the equity premium obtained from the model above with the observed equity premium data. Suppose that the observer believes that an egalitarian allocation is implemented in the economy. This will lead to an underprediction (resp. overprediction) of the equity premium if T is concave (resp. convex). In order to help solving the equity premium puzzle, we thus need T to be concave.
23.3 Wealth Inequality and the Risk-Free Rate We now turn to the analysis of the impact of wealth inequality on the risk-free rate. Here, the equilibrium rate of return of a zero-coupon bond (i.e., the risk-free interest rate) equals
where z0 is the current GDP per capita and
is the GDP per capita at the maturity of the zero-coupon bond. The
level of the risk-free rate depends on three different aspects of the representative agent's preference. First, the riskfree rate is affected by the pure preference for the present, characterized by 1/β. Assets yielding payoffs in the distant future are less valued in order to compensate investors for their patience. Second, the risk-free rate depends on the growth of the representative agent's consumption. If v is concave, smoothing consumption over time increases lifetime utility. In the absence of uncertainty about the positive growth of consumption, increasing the risk-free rate above 1/β is necessary in order to compensate investors for not smoothing their consumption over time. By how much the risk-free rate must be increased depends on the elasticity of intertemporal substitution of the representative agent. Third, if the representative agent is prudent and if future consumption is uncertain, a reduction in the risk-free rate is required in order to convince
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_347.html [4/21/2007 1:33:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_348.html
Page 348 agents not to accumulate wealth for a precautionary saving motive. The size of this effect depends on the degree of absolute prudence of the representative agent. In order to address the question of the effect of wealth inequality on the risk-free rate, we must determine its impact on each of these three motives to save. We do this by starting with a simple case, which we extend using three subsequent steps. The benchmark case is where there is no uncertainty on future GDP per head (to exclude the precautionary saving motive) and no growth in aggregate consumption (to exclude the income-smoothing effect). In the benchmark case, credit markets work at date 0 to smooth all consumption plans at equilibrium in the unequal economy. The equilibrium is sustained by a gross risk-free rate equaling 1/β. Wealth inequality has no effect on it. We now introduce growth. 23.3.1 The Consumption Smoothing Effect In this section, we suppose that
takes value z1 with probability 1. Moreover we assume that z1 > z0. There is no
uncertainty about the positive growth of the GDP per head. In this economy, insurance markets will be at work to eliminate all individual risks. Because growth is certain and positive, the concavity of v in equation (23.9) implies that ρ = 1 + r is larger than 1/β. This is the pure income smoothing effect. Because the level of wealth of agents affect their willingness to smooth consumption over time, we may expect that wealth inequality affects the equilibrium risk-free rate in this economy. Proposition 67 tells us that we just have to determine the impact of the wealth inequality on the degree of concavity of v. Because result which parallels proposition 83:
, we directly obtain the following
84 Consider an economy with homogeneous preferences. Suppose that there is no uncertainty about the growth of GNP per head. Suppose also that the growth is positive, meaning that z1 > z0 or β(1 + r) > 1. If the PROPOSITION
absolute risk tolerance is concave (resp. convex), then the equilibrium interest rate in an unequal economy is larger (resp. smaller) than the equilibrium interest rate in the egalitarian one with the same growth. The opposite results hold if growth is negative. Another way of understanding this result is to observe that the marginal propensity to consume out of wealth is decreasing in wealth
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_348.html [4/21/2007 1:33:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_349.html
Page 349 under certainty if and only if absolute risk tolerance is concave (proposition 58). In other words, the concavity of T is equivalent to the concavity of saving with respect to wealth. Jensen's inequality yields that wealth inequality reduces the average/aggregate saving under this condition. This must be compensated by an increase in the interest rate in order to sustain the equilibrium. 23.3.2 The Precautionary Effect We now turn to the general case with an aggregate uncertainty at date 1. It is useful to decompose the global effect of wealth inequality on the risk-free rate into a consumption smoothing effect and a precautionary effect. Observe that there is no consumption smoothing effect under certainty if ρ = 1 + r equals 1/β. If we maintain the assumption ρ = 1/β in the uncertain but egalitarian world, we obtain a pure precautionary effect of wealth inequality. We know from proposition 68 that wealth inequality reduces the equilibrium risk free rate if v′ is more convex than u′. Thus, to determine whether wealth inequality reduces the risk-free rate, we have to examine how it affects the degree of prudence of the representative agent. This is in parallel with what we have done in the previous section where we had to look at the degree of risk aversion of the representative agent. But contrary to this first result, this new problem is much more complex. Indeed, by condition v′(z) = u′(c(z, θ))/ξ(θ), and after some tedious manipulations, we can verify that
and
Combined, they yield
On the left-hand side of this equation, we have the degree of absolute prudence of the representative agent that we want to be larger than P(z) to get the result. We learn from this equality that the degree of
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_349.html [4/21/2007 1:33:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_350.html
Page 350 prudence of the representative agent is a rather complex function of the degree of prudence of the original agent. In particular, contrary to what we got for the degree of risk tolerance, the degree of prudence (or its inverse) of the representative agent is not a weighted sum of the degree of prudence (or its inverse) of the original agent measured at different levels of consumption. To solve this problem, let us define the precautionary equivalent consumption C of consumption plan 1 as
at date
We would be done if this functional were concave. Concavity is best understood by assuming that there is a finite number of states, with being distributed as (z1, p1; . . . ; zm, pm). Assuming that βρ = 1 in the egalitarian economy means that z0 = C(z1, . . . , zm). At the given interest rate, wealth inequality has the effect to transform the consumption plan for agent θ from (z1, . . . , zm) to (c(z1, θ), . . . , c(zm, θ)), with
, for all i. The
willingness to consume at date 0 of agent θ is measured by C(c(z1, θ), . . . , c(zm, θ)). If C is a concave function of the vector of contingent consumptions, then the mean-preserving spread of the distribution of these vectors in the population will reduce the mean value of C. This reduction in the precautionary equivalent future consumption will induce an increase in the saving rate. The market will react to this shock by reducing the equilibrium riskfree rate. Lemma 8 in chapter 21 provides a sufficient condition for C to be concave. Indeed, using this lemma with g(t) = – u′(t) yields that the precautionary equivalent consumption is concave in the vector describing the consumption plan if the inverse of absolute prudence is concave. Proposition 85 is a consequence of this lemma. PROPOSITION 85 Suppose that agents are prudent (u′″ ≥ 0) and that ρ = 1/β in the egalitarian economy. Then wealth inequality reduces the equilibrium risk-free rate, if the inverse of absolute prudence (P–1(z) = –u″(z)/u′″ (z)) is concave.
Proposition 85 states that, if the inverse of absolute prudence is positive and concave, then wealth inequality has a negative precautionary effect on the risk-free rate. Thus the concavity of the inverse of absolute prudence may explain why the representative agent à la Lucas overpredicts the rate of return on safe bonds if wealth inequality is not taken into account.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_350.html [4/21/2007 1:33:16 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_351.html
Page 351 We are now in a situation to combine the consumption smoothing effect with the precautionary effect to obtain a general picture of the effect of wealth inequality on the equilibrium risk-free rate. Our findings are summarized in the following proposition. PROPOSITION 86 (Gollier 2000) Suppose that the risk-free rate is less (resp. larger) than the rate of pure preference for the present, βρ < (>)1, in the egalitarian economy. Then wealth inequality reduces the risk-free rate if the following two conditions are satisfied:
1. The inverse of absolute prudence is concave. 2. The absolute tolerance to risk is concave (resp. convex). Observe that the inverse of absolute risk aversion and the inverse of absolute prudence are both linear in wealth when the utility function is HARA. We know that the distribution of wealth does not affect asset prices in that case. Again, HARA is a limit case.
23.4 Conclusion The attractiveness of many asset pricing models available in the finance literature comes from the assumption that there is a representative agent in the economy. There are two potential deficiencies to this approach however. First, when the economy is heterogenous, the existence of a representative agent who maximizes the expected value of his utility should not be taken as granted. But as seen in the previous chapter, this assumption is automatically satisfied if markets are complete. The second potential deficiency comes from the difficulty to characterize the preferences of the representative agent when it exists. A helpful result to solve this problem is that the absolute risk tolerance of the representative agent equals the average absolute risk tolerance in the population, in each state of the world. But, for example, no similar formula can be obtained for the measure of prudence. The most part of the chapter has been devoted to the search of the representative agent, and of the corresponding equilibrium prices, when the only source of heterogeneity in the population is wealth inequality. If we consider an egalitarian economy whose risk-free rate is small, we concluded that under the concavity of both the inverse of the absolute risk aversion and the inverse of absolute prudence, the equity premium will be underpredicted and the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_351.html [4/21/2007 1:33:17 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_352.html
Page 352 risk-free rate will be overpredicted by a calibrator who does not take into account of wealth inequality.
23.5 Exercises and Extensions Exercise 110 Consider an economy with agents having the same CRRA utility function. Markets are incomplete in the sense that agents can only invest in two assets, one being risk free with a zero return, and the other being risky with return . The only source of heterogeneity in the economy is ex ante wealth inequality. • Show that there exists a representative agent in this economy. This agent endowed with the ex ante average wealth in the economy has an optimal demand for the risky asset which is equal to the mean demand in the economy, whatever the distribution of the risky asset returns. Notice that it is in general not true that such a representative agent exists when agents don't have linear absolute risk tolerance with the same slope (Can you prove this?). • Show that wealth inequality does not affect the aggregate demand for the risky asset in the economy. Thus, the result obtained in the case of complete markets with CRRA remains true in the two-asset model. This should not be a surprise, since restricting risk-sharing arrangements to linear ones does not affect the equilibrium allocation of risk. Explain why. Exercise 111 Consider an economy with two logarithmic agents living for two periods. They both have an initial wealth w that they must allocate to consumption at date 0 and 1. Agent θ = 1 is not impatient (β1 = 1), whereas agent θ = 2 has a discount factor β2 < 1 for felicity at date 1. • Show that a representative agent exists in this economy. This agent has a felicity function v and discount factor β* such that his competitive level of saving out of initial wealth w is always equal to the mean level of saving of the two agents at equilibrium, whatever the level of w. • Show that this representative agent is logarithmic with discount factor β* = (1 + 3β2)/(3 + β2). Exercise 112 Consider an economy with two agents, with u1 (c) = In (c) and
. The two agents are
endowed with the same random GDP per capita .
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_352.html [4/21/2007 1:33:17 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_353.html
Page 353 • Describe the set of Pareto-efficient risk-sharing rules in this economy. To whom should call options on equity be allocated? • Suppose that there are two states of nature. In the bad (good) state, both agents are endowed with one (two) unit of the consumption good. Characterize the competitive equilibrium of this economy. Describe the utility function of the competitive representative agent.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_353.html [4/21/2007 1:33:18 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_354.html
Page 354
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_354.html [4/21/2007 1:33:18 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_355.html
Page 355
VIII Risk and Information
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_355.html [4/21/2007 1:33:19 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_356.html
Page 356
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_356.html [4/21/2007 1:33:19 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_357.html
Page 357
24 The Value of Information. An important aspect of dynamic risk management is the existence of a flow of information linked to the distribution of future risks. Young workers can observe signals about their long-term productivity in the early stages of their career. Entrepreneurs obtain information linked to the profitability of their investment project. Policy-makers may want to wait for better scientific knowledge about the risk of global warming before deciding how much effort should be spent to reduce emissions of greenhouse gases. Technological innovations provide information about the prospect of future growth of the economy. The common feature of these examples is that some signals are expected to be observed before taking a final decision on the exposure to the risk. This part of the book is devoted to the analysis of the interactions among risk, time, and information. We begin with the examining how people value future risks in the presence of a flow of information related to its distribution. The notion of the value of information is introduced. We also examine here which information structure should be preferred by agents. We finish with the examination of the link between the value of information and risk aversion.
24.1 The General Model of Risk and Information 24.1.1 Structure of Information We consider a risky situation that is characterized by X possible states of the world indexed by x = 1, . . ., X. measures the probability of occurrence of these states. Prior to the realization of the Vector state, the decision maker is in a position to observe a signal. Let M be the number of potential signals indexed by
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_357.html [4/21/2007 1:33:19 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_358.html
Page 358 m = 1, . . ., M. Vector Q = (q1, . . ., qM) denotes the vector of unconditional probabilities of the different signals. Signals are potentially statistically related to the states. Thus, observing the signal allows the decision maker to revise the probability distribution of the states, using Bayes's rule. This statistical relationship is characterized by the matrix
of posterior probabilities. Namely pmx is the probability that state x be realized if signal m is observed. Matrix P has M rows and X columns. Vector pm. = (pm1, . . ., pmX) denotes the posterior probabilities conditional to signal m. It corresponds to row m of matrix P. Basic probability calculus yields that the prior probability of state equals Σmqmpmx. Using the matrix notation, this is equivalent to
Since we will not use the matrix of joint probabilities, its notation is not introduced here. In the remaining, we use the terms "prior" and "posterior" to refer to situations respectively before and after the observation of the signal. 24.1.2 The Decision Problem Posterior to the observation of the signal, but before the observation of the state of the world, the agent must take a decision in the face of the remaining uncertainty. In its most general formulation, the agent's final utility v(α, x) depends on his decision α and the state of the world x. The agent is limited in his choice of α, which must belong to some nonempty set B ⊂ Rn, where n is the dimensionality of the decision variable α. Except when explicitly stipulated, we will hereafter assume that n = 1. Suppose that the agent's preferences under uncertainty satisfy the axioms of von Neumann and Morgenstern. His decision problem is then written as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_358.html [4/21/2007 1:33:20 AM]
,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_359.html
Page 359
The bracketed term is the maximal expected utility that the decision maker can obtain when he observes signal m, which generates a vector of posterior probabilities pm. = (pm1, . . ., pmX). Prior to the observation of the signal, the maximum expected utility U(P, Q) is the sum of these maximum posterior expected utilities, weighted by the probability of the different possible signals that he can receive. Let us define function g from the simplex in RX to R as follows:
This represents the maximal expected utility given a vector of posterior probabilities pm. of the states. Associated to it is the optimal solution α(pm.) given this vector of posterior probabilities. We can rewrite equation (24.3) as
It is important to examine the main properties of the g function. 24.1.3 The Posterior Maximum Expected Utility Is Convex in the Vector of Posterior Probabilities Observe that function g is linear in the vector of posterior probabilities pm. when the optimal decision α is independent of this vector. This is by the way the main feature of expected utility. But in general, the optimal decision will be affected by a change in the probabilities. By definition (24.4), function g is the upper envelope of a set of functions f(α, pm.) = Σx pmxv(α, x), α ∈ B, that are linear in pm.. Therefore it is a convex function of vector pm.. Remember that a convex function g from 2
⊂ RX to R is defined by the property that for any couple (a, b) ∈
and any scalar q ∈ [0, 1], we have
Because this result is of primary importance for the remaining of this book, we illustrate this property by an file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_359.html (1 of 2) [4/21/2007 1:33:21 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_359.html
example. Let us consider an uncertain environment with three states of the world. In this case
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_359.html (2 of 2) [4/21/2007 1:33:21 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_360.html
Page 360
Figure 24.1 Expected utility as a function of the posterior probabilities with three possible states and two choices
a specific posterior probability distribution is characterized by a point (pm1, pm3) in the Machina triangle, with pm2 = 1 – pm1 – pm3. We consider the standard portfolio decision problem in which v(α, x) = v(w0 + αyx) where α is the number of euros invested in a risky project, and yx is the state-dependent net payoff of the project per euro invested. Suppose first that B = {0, 1}, that is, that the agent can invest either zero or one euro in the project. Then we have that g(pm.) equals
This is indeed the maximum of two linear functions of (pm1, pm3). In figure 24.1 we depict these two linear functions by a representation in the Machina triangle where y1 = –0.5, y2 = 0.5, y3 = 1, w0 = 1, and v(z) = –z–1. The two plans describe the expected utility respectively for α = 0 and α = 1 as a function of the posterior probabilities. The upper envelope of the two plans represents the maximum expected utility. It describes a convex function. When pm1 is large, it is better not to invest in the project and g(pm.) = v(w0). On the contrary, when it is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_360.html (1 of 2) [4/21/2007 1:33:21 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_360.html
small enough, it is optimal to invest in the project and
.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_360.html (2 of 2) [4/21/2007 1:33:21 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_361.html
Page 361
Figure 24.2 Maximal posterior expected utility in the portfolio problem with three possible states
In figure 24.2 we describe function g when B ≡ R, which is when there is no restriction on the amount that can be invested in the risky project. To each scalar α, there is a plan above the Machina triangle that represents the expected utility if this decision α is taken. For each potential vector of posterior probabilities, the decision maker selects the α that maximizes the conditional expected utility. Because the optimal α is a smooth function of this vector, so is the optimal expected utility. In Figure 24.2 we just represented the optimal expected utility as a function of the posterior probabilities, which is the upper envelope of an infinite number of plans that have not been represented. Thus the g function is convex. This result holds independent of the decision problem under consideration. It just requires the independence axiom of expected utility. Reciprocally one can prove that any convex function from ⊂ RX to R can be obtained by (24.4) from a specific decision problem (v, B). This is due to a well-known separating hyperplane theorem which implies that to any convex surface, there exists a set of hyperplanes for which this surface is the upper envelope. We summarize this in the following lemma:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_361.html [4/21/2007 1:33:22 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_362.html
Page 362 LEMMA
9 Consider the definition of function g:
⊂ RX → R as given by equation (24.4). We have that
• function g is convex; • to any convex function g, there exists a decision problem (v, B) such that equation (24.4) holds for all pm. ∈
.
24.2 The Value of Information Is Positive In this section we examine how the attitude toward the uncertainty about the state of the world x is affected by the flow of information. To do this, we compare the optimal prior expected utility with the structure of information (P, Q) to the optimal expected utility without any information. In the absence of any source of information before having to take a decision, the problem of the decision maker is written as:
We want to compare this to the maximal expected utility that we can attain when the decision can be taken after observing the signal coming from information structure (P, Q), which we denoted U(P, Q) in equation (24.5). The information structure has a value if the optimal expected utility U(P, Q) obtained with it is larger than the optimal expected utility g(Φ) obtained without it. In short, the information structure has a value if
since Φ = QP. Inequality (24.9) always holds. It is a direct consequence of our earlier observation that g is convex in the vector of probabilities. We conclude that any information structure has a nonnegative value. This is true independent of matrixes (P, Q) and of the characteristics of the decision problem given by B and v. At worst, the information structure has no value. This is the case when g is linear. We know that this happens only when observing the signal never affects the decision of the agent. As soon as there is at least one signal that generates a different action, g becomes convex and the inequality in (24.9) is strict: the information structure has a positive value. This
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_362.html [4/21/2007 1:33:23 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_363.html
Page 363
Figure 24.3 Simple information structure with two possible messages
remark has an important consequence: the value of information is extracted from the fact that the observation of a signal allows the agent to better adapt his decision to the risky environment that he faces. Flexibility is valuable. There is a simple intuition for why the value of information may not be negative. The action that is chosen without information is also feasible when information is available. Therefore the informed agent can at least duplicate his rigid strategy with the information structure by choosing uniformly the same action, no matter what information is received. As explained by Savage (1954), "the person is free to ignore the observation. That obvious fact is the theory's expression of the commonplace that knowledge is not disadvantageous." It is up to you to decide whether or not to take your umbrella when you heard that rain is forecasted for that day. In figure 24.3 we again consider the simple example that we gave in the previous section: there are three possible states of the world and v(α, x) = –(1 + αyx)–1 and α ∈ B ≡ {0, 1}. The story goes as follows: the decision maker has a CRRA utility function with γ = 2. He has the opportunity to purchase a lottery ticket whose possible net payoffs are (y1, y2, y3) = (–0.5, 0.5, 1). In the absence of any additional information, the distribution of probabilities is Φ = (0.35, 0.55, 0.1). This probability distribution is represented by point a in the Machina triangle of figure 24.3. The reader can easily check that it is not optimal for him to take this risk, in spite of the positive unconditional expectation of the net payoff. His expected utility equals –1. It is represented by point A in figure 24.4. Suppose now that the agent has access to a source of information about his chance to win the lottery. There are two possible messages of equal probabilities (q1 = q2 = 0.5). The vector of posterior
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_363.html [4/21/2007 1:33:23 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_364.html
Page 364
Figure 24.4 Value of information in the Machina triangle
probabilities is represented respectively by points b and c in the Machina triangle of figure 24.3. If message b is received, the posterior distribution is (0, 0.8, 0.2). The probability of a net loss being zero, the agent purchases the lottery in this case, and his expected utility equals –0.633. If message c is received, the posterior distribution is (0.7, 0.3, 0). Because the expectation of the net payoff is negative, the agent does not purchase the lottery ticket, and his expected utility equals –1. Prior to receiving the message, the expected utility equals 0.5((–0.633) + (–1)) = –0.8166. This is larger than –1, the expected utility of the uninformed agent. In figure 24.4 the ex ante expected utility of the agent is represented by the point at the center of the interval BC. It is strictly above point A, which means that the information structure has a positive value.
24.3 Refining the Information Structure 24.3.1 Definition and Basic Characterization In this section we compare pairs of information structures like (P1, Q1) and (P2, Q2). We assume that they describe the same uncertainty ex ante, namely that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_364.html [4/21/2007 1:33:24 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_365.html
Page 365
More specifically, we try to determine whether it is possible to rank (P1, Q1) and (P2, Q2) in the following sense. We say that an information structure is finer than another if the first is always preferred to the second. 7 (P1, Q1) is a better information structure than (P2, Q2) if and only if all agents enjoy a larger expected utility with (P1, Q1) than with (P2, Q2), no matter which decision problem (v, B) is considered. This is rewritten as follows: ∀B ⊂ R, ∀v: R2 → R: DEFINITION
or equivalently
This is a relevant question when the decision maker has the opportunity to choose between various information structures. Blackwell (1951) uses the terminology of an "experiment" to characterize the process of information acquisition. The agent faces various possible experiments. He must select one from whose he will extract a signal before taking decision α. For each possible experiment i, he determines the probability signals m = 1, . . . , Mi, and their corresponding vector (
of occurrence of
) of posterior probabilities. Does there exist
an experiment that is preferred by all decision makers? One faces this question when one has to evaluate medical diagnosis processes, experts in finance, but also to select a sample of the population, to improve weather forecasting, to select among various oil drilling strategies, and the list goes on. We will hereafter indifferently use the terms of a "better information structure," a "more informative structure," a "finer information structure," or an "earlier resolution of uncertainty" to describe structure 1 with respect to structure 2 if condition (24.12) is satisfied. A direct consequence of lemma 9 is given in the following proposition, which is used in Marschak and Miyasawa (1968), Epstein (1980), and Jones and Ostroy (1984), for example. PROPOSITION
87 (P1, Q1) is a better information structure than (P2, Q2) if and only if
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_365.html [4/21/2007 1:33:24 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_366.html
Page 366
for all convex function g:
⊂ RX → R.
24.3.2 Garbling Messages and the Theorem of Blackwell. Proposition 87 is not very helpful to provide an intuition. We hereafter assume that the two structures have the same set of signals m = 1, . . . , M. This is without loss of generality, since we can always set some of the to zero if signal m never occurs in structure i. The most extreme forms of information structure are respectively the completely uninformative one and the completely conclusive one. (P, Q) is completely uninformative if the vector of posterior probabilities is the same for all signals. It is completely conclusive if there is only one possible signal for each state of the world. In that latter case, the vector of posterior probabilities is of the form (0, . . . , 0, 1, 0, . . . , 0): the observation of the signal completely specifies the state. In such a situation the uncertainty is completely resolved by observing the signal. If the two structures describe the same prior uncertainty, it is obvious that the completely conclusive one is better than the completely uninformative one. This is a direct consequence of our observation that the value of information is always nonnegative. More generally, any information structure is better than the corresponding completely uninformative structure. This is the meaning of property (24.9). Similarly the completely informative experiment is better than any incomplete one with the same prior. As an example, consider three possible messages m = 1, 2, 3 that are equally likely. Information structure i = 1 gives a posterior distribution a in figure 24.5 for all three messages. This information structure is uninformative. Information structure i = 2 generates posterior distribution c if signal m = 1 is received, and posterior distribution b otherwise. This information structure is better than information structure 1. Observe that information structure 2 is a mean-preserving spread of information structure 1 in the space of probability distributions (here in the Machina triangle). Equivalently a is the center of gravity of an object having one-third of its mass in c and twothirds of it in b. Because the maximal expected utility is a convex function of the various posterior probability distributions, it is clear that structure 2 is preferred to structure 1: all decision makers like
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_366.html [4/21/2007 1:33:25 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_367.html
Page 367
Figure 24.5 Information structures with three messages
mean-preserving spreads in the space of posterior distributions function. And this has nothing to do with a riskloving behavior, which corresponds to mean-preserving spread in the space of wealth. Now consider information structure i = 3 that generates posterior distributions c, d, and e respectively for signals m = 1, 2, and 3. Observe again that b is the center of gravity of (d, e) where the mass is equally split in the two points. In other terms, (d, e) is a mean-preserving spread of b. Thus information structure i = 3 is a meanpreserving spread of information structure i = 2. Again, because the maximal expected utility is convex in the posterior distribution, information structure i = 3 is preferred to information structure i = 2, independent of the decision problem under consideration. This can be seen as follows:
because pb = 0.5(pd + pe) and g is convex in p. We conclude that any mean-preserving spread in the distribution of the vectors of posterior probabilities improves the degree of informativeness of the structure. An alternative way to explain this is by using the technique of garbling messages. One can duplicate information structure 2 from information structure 3 by using the following procedure: consider a garbling machine that receives the true signal from information structure 3 and send another message by using these rules: • If message 1 is received, the machine sends message 1. • If message 2 is received, the machine sends message 2 or 3 with equal probabilities.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_367.html [4/21/2007 1:33:26 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_368.html
Page 368 • If message 3 is received, the machine sends message 2 or 3 with equal probabilities. Suppose that the decision maker only observes the message coming out of this garbling machine. Then, if message 1 is observed, he immediately infers that the message received by the machine was message m = 1 from the original information structure 3: this message is not garbled by the machine. He deduces from observing message 1 that the posterior probability is pc. But if message 2 is received, the agent infers that the original message received by the machine was either m = 2 or m = 3 with equal probabilities. Thus the vector of probabilities on the states is either d or e with equal probabilities. In other words, the vector of posterior probabilities in this case is pb = 0.5(pd + pe). He concludes in the same way if the machine gives him message 3. To sum up, using information structure 3 coupled with the garbling machine generates an information structure equivalent to information structure 2. It gives him posterior pc with probability 1/3 and posterior pb with probability 2/3. More generally, any sequence of mean-preserving spread in the space of probability distribution can be obtained by this kind of garbling machine. A garbling machine is characterized by a function μ: R2 → R and a random variable that
independent of
such
for all x, where and are the messages coming respectively in and out of the machine. It is important to observe that the behavior of the garbling machine is independent of the state of nature x. This technique is easier to represent by using the matrix notation. Consider a machine which sends message j if it receives message i with probability hij:
The knowledge of matrix H = [hij] allows us to completely describe the distribution of the message coming out of the machine if we would know the message that has been received by it, independent of the state. Because the random noise introduced by the machine is not correlated to the state, observing the message coming out of it is a does not bring more information than observing the message coming in. In statistics, we say that does not give more information. This sufficient statistic for : if we already observed , observing garbling machine is represented in figure 24.6.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_368.html [4/21/2007 1:33:26 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_369.html
Page 369
Figure 24.6 Garbling machine
Figure 24.7 Experiment 1 is better than experiment 2 in case a, but not in case b
From matrix H, we can determine kij, the probability that message j has been received by the garbling machine if it sent message i:
Using Bayes's rule, we have that
Of course, we have kij ≥ 0 and ∑jkij = 1 for all i. From the kij, we can obtain the posterior distribution of the garbled information structure:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_369.html (1 of 2) [4/21/2007 1:33:27 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_369.html
Thus all posterior probabilities in structure 2 are a convex combination of the posterior probabilities of structure 1. In other words, all points representing posterior distributions of experiment 2 are within the smallest convex envelope of points representing posterior distributions of experiment 1. Two examples are given in figure 24.7. This is another way to say that structure 2 is a mean-preserving contraction of structure 1. In matrix notation, this means that P2 = KP1
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_369.html (2 of 2) [4/21/2007 1:33:27 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_370.html
Page 370 where K = [kij]. In order to preserve the distribution of prior probabilities Φ = Q1P1 = Q2P2 = Q2KP1, it must be true that
To sum up, one can deteriorate an information structure either by garbling messages, or by a sequence of meanpreserving contractions in the space of posterior probability distributions. These are two equivalent representation of the same phenomenon. We show in the following proposition, due to Blackwell (1951), that they characterize the set of finer information structures. Ponssard (1975), Cremer (1982), and Kihlstrom (1984) provided alternative proofs of this theorem. 88 Consider two information structures, (P1, Q1) and (P2, Q2), which have the same number M of possible messages, and the same vector of unconditional probabilities Φ = Q1P1 = Q2P2. Information structure 1 is finer than information structure 2 if and only if there exists a M × M matrix K = [kij] such that all kij ≥ 0 and PROPOSITION
for all i, and
Proof We start with the proof of sufficiency. We have that
We used the convexity of g to obtain the inequality on line 3 of this sequence of operations. This concludes the proof of sufficiency. The proof of necessity is relegated to the appendix of this chapter. We can summarize these results in the following way. We introduced four different notions of refinement of file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_370.html (1 of 2) [4/21/2007 1:33:27 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_370.html
information. Information structure 1 is finer than information structure 2 if
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_370.html (2 of 2) [4/21/2007 1:33:27 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_371.html
Page 371 1. all decision makers prefer structure 1 to structure 2, independent of the underlying decision problem; 2. the distribution of posterior probabilities of structure 2 is a mean-preserving spread of the distribution of posterior probabilities of structure 1; 3. the messages of structure 2 are obtained by garbling the message coming from structure 1; 4. condition (24.13) is satisfied for all convex function g. All these definitions are equivalent. 24.3.3 Location Experiments An important class of information structures is such that for all x,
where
is independent of x. An information structure belonging to such a set is called a "location experiment."
If , signal is an unbiased estimator of the true state of the world. Random variable linked to experiment i can be interpreted as the error of observation of the unknown variable x. The intuition suggests that an experiment with a smaller dispersion of the error around its mean should provide a better information. To illustrate, let us compare two location experiments. Experiment i = 1 has an error over interval [–1/2, 1/2], whereas experiment i = 2 has an error distributed over interval [–1, 1]:
that is uniformly distributed
that is more dispersed, since it is uniformly
It is easily shown that all agents prefer experiment 1 over experiment 2; that is to say, experiment 1 is better than experiment 2 in the sense of Blackwell. Indeed, remember that one way to prove this is to show that there exists a is distributed as function μ(., .) and a random variable such that signal immediate in our case by taking μ(m, ξ) = m + ξ and ξ ~ (–1/2, 0.5; 1/2, 0.5). Another example is when
and
for all x. This is
are both normally distributed:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_371.html (1 of 2) [4/21/2007 1:33:28 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_371.html
Suppose that is less dispersed than structure 2? The answer
, meaning that V1 is smaller than V2. Is structure 1 more informative than
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_371.html (2 of 2) [4/21/2007 1:33:28 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_372.html
Page 372 is distributed as plus a normally distributed variable is positive again, since V2 – V1. This is because the sum of two normal random variables is normal.
with mean zero and variance
These two examples suggests the possibility that the comparison of two location experiments can be reduced to the comparison of the dispersion of the errors in the observation of x. We hereafter show that this suggestion is misleading. As proved by Boll (1955), cited by Lehmann (1988), condition (24.15) can be true for two location experiments (24.23) only when function μ takes the form μ(m, ξ) = m + ξ. This greatly simplifies the analysis, but it also shows how restrictive is Blackwell's notion of better information. is normally distributed. It is well-known that if a random variable For example, suppose again that normal and if it is the sum of two independent random variables and , then and also be normal. In consequence cannot be more informative than a normally distributed unless ) is also normal. If is not normal, it can never be more informative than dispersed the distributions are. Another example is in Lehmann (1988) and Persico (1998). Suppose that distributed, with
and
is must (and
, no matter how concentrated or
are both uniformly
with 0 < ρ < 1, which implies that the error is less dispersed in experiment 1 than in experiment 2. We showed above that experiment 1 is better than experiment 2 when ρ = 1/2. The same proof can be used to show that this result holds for all ρ such that ρ = 1/k for some positive integer k. But Lehmann (1988) showed that this condition is not only sufficient but necessary. Thus taking ρ, which is not the inverse of an integer, does not lead to a finer information structure, whatever small ρ! These two examples confirm two claims: First, the Blackwell's order is partial. Second, it is very restrictive. This is due to the fact that Blackwell requires that all agents must prefer structure 1 to structure 2 independent of the decision problem under scrutiny. Some recent progress have been made to enlarge the degree of comparability by restricting the set of decision problems. In particular, Lehmann (1988) and Persico (1998) focus on problems satisfying the single crossing property, meaning problems where the derivative of v(α, x) with respect to α crosses zero at most once from below as a function of x. Athey and Levin (1998) examines decision
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_372.html [4/21/2007 1:33:29 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_373.html
Page 373 problems where the optimal decision α is an increasing function of the signal.
24.4 The Value of Information and Risk Aversion 24.4.1 A Definition of the Value of Information In section 24.2 we saw that the value of information is nonnegative. The ability to adapt the decision to the information makes the information structure always more valuable than no information at all. Notice that this result is obtained without making any assumption about the utility function of the decision maker, who can be risk-averse or risk-lover. We now examine the link between the value of information and risk aversion. Because we will evaluate information structures in monetary terms, we need to assume that the agent's utility v is a function of his consumption, which is itself a function c(α, x) of his decision α and of the state x:
Under this additional assumption, one can define the value of information structure (P, Q) as the sure amount V of consumption that is necessary to compensate the agent for the loss of the access to the information. V is thus formally defined as
where
x
is the unconditional probability of state x. The right-hand side of this equality is the expected utility of
the agent who has access to the information. The left-hand side is the expected utility of the uninformed agent who got compensation V. We can also interpret the value of information as the difference between two certainty equivalents. Let C(α, p., z) denote the certainty equivalent associated to lottery (c(α, 1), p1; . . .; c(α, X), pX) when the agent has an exogenous initial wealth z. This means that u(z + C(α, p.)) = Σpxu(z + c(α, x)). Let also C+(p.) be the supremum of the C(α, p., 0) over all α in B. C+(p.) is the maximal posterior certainty equivalent that the agent can obtain if the posterior probability distribution is p.. Before observing the message, the agent who will have access to the signal faces lottery
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_373.html [4/21/2007 1:33:30 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_374.html
Page 374 whose certainty equivalent is denoted IC. This certainty equivalent of the informed agent is such that u(IC) = Σqmu(C+(pm.)). It is the certainty equivalent of the maximum posterior certainty equivalents. IC represents the value of the game against nature for the informed agent. Let UC be the maximum certainty equivalent of the game for the uninformed agent with wealth V. This means that UC is the supremum of the C(α, Φ, V) over all α in B. This implies that condition (24.28) can be rewritten as
or equivalently
The value of the information is the difference between the certainty equivalents of the game respectively for the informed agent and the uninformed agent. The value of information is a function of the structure of the information (P, Q), of the decision problem (c(., .), B), and of the decision maker's attitude towards risk characterized by u. Section 24.3 has been devoted to the analysis of the link between the information structure and the value of information. Indeed, we could have defined a finer information structure as any structure that yields a larger value of information, independent of the decision problem and of the attitude toward risk of the decision maker. As in Willinger (1989) and Treich (1997), for example, we hereafter examine the link between V and u. We immediately see from (24.30) that an increase in risk aversion has an ambiguous effect on the value of information, since it reduces IC and UC at the same time. 24.4.2 A Simple Illustration: The Gambler's Problem The initial reaction to most people would be that more risk-averse agents would value information more. Because the flow of information will allow agents to better manage the risk, an increase in risk aversion would make the access to the information more valuable. This intuition is not true in general. Let us illustrate this problem with an example. A CRRA gambler is endowed with a wealth of 10. He concedes to gamble this amount in a coin toss where the payoff is +10 if the outcome is a head, and –5 if it is a tail. Without further information the gambler thinks that there is an equal chance of a
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_374.html [4/21/2007 1:33:30 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_375.html
Page 375
Figure 24.8 Value of information is not a monotone function of risk aversion
head or tail outcome. We must consider the possibility that there is access to some information. There are two possible messages for the outcome q1 = q2 = 0.5. If signal m = 1 is received, the posterior probability to get head becomes 0.75, whereas if signal m = 2 is received, this posterior probability is only 0.25. In the latter case the ex post expected payoff of the lottery is negative, so the risk-averse decision maker does not gamble. The value of the information V satisfies the following condition:
In figure 24.8 we show the value of information as a function of the degree of relative risk aversion. We see that the value of information is at first increasing, and then decreasing, in the degree of risk aversion. In fact the kink in this curve corresponds to the critical level of risk aversion above which the uninformed agent does not gamble. We are now able to explain why this curve is hump-shaped, that is, why the value of information is not necessarily increasing with risk aversion. Suppose that the gambler's risk aversion is large
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_375.html [4/21/2007 1:33:31 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_376.html
Page 376 enough to make him to decline the lottery when he is uninformed. This implies that UC = 0. In this case, the value of information comes from the decision to gamble when the signal is good (m = 1). The informed gambler gains the conditional certainty equivalent associated to this signal. The value of this posterior gamble is decreasing with risk aversion. In consequence a more risk-averse gambler will value information less. This comes from equality V = IC, together with the fact that the certainty equivalent of gambling is decreasing with risk aversion. The misleading idea is that more information yields less risk-taking, which implies that an increase of risk aversion reduces IC less than UC, thereby raising V. But more information can just as well increase the optimal exposure to risk, as we saw in the example above. To sum up, an increase in risk aversion has two competing effects on the value of information. First, it reduces the value of all risks taken ex post. Therefore it has an adverse effect on the value of information. Second, it reduces the value of the risk taken by the uninformed agent. This makes the information structure more valuable. But this second effect does not exist when the uninformed agent does not gamble at all. We formalize this idea in the following proposition. The gambler's problem is defined as any decision problem such that B = {0, 1} and c(α = 0, x) = c0 for all x ∈ X. In words, the gambler has only two options, one of which is risk free. 89 In the gambler's problem, the value of information is decreasing in the degree of risk aversion if the uninformed agent does not gamble. PROPOSITION
Proof We consider two agents j = 1, 2 with utility functions uj. They don't gamble if they are not informed. We assume that agent j = 1 is more risk-averse than agent j = 2. This implies that is less than all m. By assumption, the value of information for agent j is obtained by solving the following equation:
We know that
is dominated by
in the sense of first-order stochastic dominance, which implies that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_376.html [4/21/2007 1:33:31 AM]
for
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_377.html
Page 377 Because j = 1 is more risk-averse than j = 2, proposition 2 together with the above inequality implies that
This implies in turn that V1 is smaller than V2. This corresponds to relatively large degrees of risk aversion. But consider now the case where risk aversion is low enough for the decision maker to gamble even though he has no access to information. In that case the value of the information comes from the decision to switch to the sure position if the bad signal m = 2 is received. The gambler saves the conditional expected loss and the risk premium associated with this message. What makes this case more difficult is that the value of information is now the difference between two functions, IC and UC, that are decreasing with risk aversion. The risk premium ex post is increasing with risk aversion, and this has a positive effect on the value of the information. However, intuition suggests that the informed agent faces a reduced risk with respect to the risk taken by the uninformed one. In other words, an increase of risk aversion will reduce UC more than it will reduce IC. This has a positive effect on the value of information. This intuition is confirmed in figure 23.8, but we need a few more restrictions before we can formulate a theoretical foundation for this effect. Indeed, we know from section 7.2.3 that an increase in risk aversion does not necessarily induce people to value more a given reduction in risk. This is why the following proposition is restricted to the case of complete information. PROPOSITION 90 Consider the case of full information in the gambler's dilemma. The value of information increases in risk aversion if the uninformed agent gambles.
Proof Suppose that u1 is more concave than u2. The two agents gamble despite their lack of information. By assumption, the value of information is defined as follows:
where
is the random variable that is distributed has (c(1, 1),
1,
. . ., c(1, X),
). Now consider a case of
X
complete information where X = M and where the observation of signal m implies that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_377.html [4/21/2007 1:33:32 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_378.html
Page 378
state m is true. We have that . This implies that and are identically distributed. Moreover, their cumulative distribution single crosses the cumulative distribution of at 0, from below. Indeed, we have that
for all t > 0, and
for all t < 0. This means that implies that
Because
and
is riskier than
in the sense of Jewitt (1987). By proposition 19, this
are the same, this implies in turn that V2 is smaller than V1. Notice that when the
information is not complete,
and
do not coincide and we may not conclude.
24.4.3 The Standard Portfolio Problem Let us now examine the effect of more risk aversion on the value of information when c(α, x) = w0 + αyx and B = R. As for the gambler's problem, this effect would unambiguously be negative if UC were zero. This is the case here only if the unconditional expectation of is zero. Because the uninformed agent does not take any risk, the information will induce the decision maker to increase his risk exposure ex post. Therefore an increase of risk aversion reduces the value of information, which is the certainty equivalent of the posterior certainty equivalents of the optimal signal-dependent portfolios. Persico (1998) shows that this result holds with CARA functions and normal returns, even when the unconditional expectation of is not zero. Treich (1997) obtains the same result without assuming CARA, but only in the case of small portfolio risks. The general intuition is again that the informed agent will take more risk than the uninformed one. Therefore an increase in risk aversion has a larger impact on the certainty equivalent IC of the informed agent than on the certainty equivalent UC of the uninformed agent. It yields a reduction of the value of the information. This can be shown formally as follows: Let yx be equal to x for all x. Suppose that the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_378.html (1 of 2) [4/21/2007 1:33:33 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_378.html
unconditional distribution of
is normal with mean μx ≥ 0 and variance
. From conditions (4.6) and (4.7) we
know that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_378.html (2 of 2) [4/21/2007 1:33:33 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_379.html
Page 379
where A is the constant absolute risk aversion of u. Suppose also that is normal with mean 0 and variance
is distributed as
and is independent of . It implies that
for all x, with
is also normal with
Using again conditions (4.6) and (4.7), we obtain that
Notice at this stage that our basic intuition that the information increases the optimal exposure to risk is true in this case, since
Remember that V is defined as
. Combining (24.39) and (24.41), we have
This immediately implies that V is inversely proportional to the degree of absolute risk aversion.
24.5 Conclusion A Bayesian decision maker revises his belief about the distribution of a risk by observing signals that are correlated with it. If he cannot adapt his exposure to the risk to the posterior distribution, he is indifferent to any change in the information structure that preserves the prior distribution of the risk. This is because expected utility is linear in the probabilities of the various states of the world. The information has no value in that case. An experiment generating a signal that is correlated to the unknown state of the world has a value only if the agent can modify his decision from the information that he gets. Flexibility is valuable. The intuition is that the informed agent can at worst duplicate the rigid strategy of the uninformed file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_379.html [4/21/2007 1:33:33 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_380.html
Page 380 agent to secure the same level of expected utility. But, in general, he can do better by using the information for a risk management purpose. We examined how the value of information is affected by a change in the structure of the information and by a change of the degree of risk aversion of the decision maker. The first exercise lead us to the notion of a finer information structure. Again, the independence axiom greatly simplified the analysis by making the expected utility linear with respect to probabilities. In non-expected utility models, only functional forms that are convex with respect to probabilities will generate nonnegative values of information. Otherwise, agents may dislike information. Suppose, for example, that the lifetime utility function is not time-additive, meaning that the felicity function of the agent at any time is affected not only by his current level of consumption but also by his future level of expected felicity. One can be happy this month even if one's current consumption is low, but just because one is offered a free dream vacation the next month. Suppose that the current felicity of an anxious person is affected by his future certainty equivalent consumption in a concave way, as is the case for some Kreps-Porteus preferences. Such anxious persons can have negative values of information. To illustrate, consider the risk of a fatal illness and a medical test to determine who will be in good health and who will be ill. An early resolution of uncertainty may be undesirable because the positive effect of knowing oneself in good health is more than compensated by the risk of learning oneself close to death. If this is not compensated by the possibility to better adapt the medical treatment to the result of the test, its value is negative. This is an explanation for why some anxious people may prefer not to test themselves for AIDS and other predictable illnesses even when the test is free. This story has been discussed in section 20.2 for Kreps-Porteus preferences where agents may prefer or dislike an early resolution of uncertainty even without any flexibility in the decision process. In chapter 26 we will consider an alternative explanation that is based on the effects of information on equilibrium prices and on the opportunity to share risks efficiently.
24.6 Exercises and Extensions Exercise 113 (Lehmann 1988) Consider location experiments with both uniformly distributed,
. Suppose that
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_380.html [4/21/2007 1:33:34 AM]
and
are
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_381.html
Page 381 with positive integer.
and
. Show that structure 1 is finer than structure 2 if ρ is the inverse of a
Exercise 114 The decision problem here is an Arrow-Debreu portfolio problem:
Suppose that u is logarithmic. • Show that
.
• Show that g is convex. Notice that is the entropy. Function –h is often used in information theory as a specific g function to measure the degree of informativeness of an information structure. This is quite arbitrary however, since other convex g functions could be used. • Suppose that X = 2, Π1 = Π2 = 1 and
1
=
2
= 0.5. Calculate
a. the expected utility of the uninformed agent; b. the expected utility of the fully informed agent; c. the value of perfect information. Exercise 115 Consider a social planner who has to share a cake of fixed size z among X identical agents with utility function u. The social planner selects the allocation rule (c1, . . . , cX) that maximizes a weighted sum of agents' utility:
The social planner faces some uncertainty about the weight vector λm. that he will use to evaluate the social welfare ex post. The uncertainty takes the following form: with probability qm ≥ 0, the weight vector will be λm., for m = 1, . . . , M, with ∑qm = 1.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_381.html (1 of 2) [4/21/2007 1:33:34 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_381.html
• Show that the social planner would prefer to select the allocation after observing which λm. will preveal, rather than before observing it. • Compare two structures of uncertainty on the weight vector. Give a sufficient condition for the first to be preferred to the other by the social planner.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_381.html (2 of 2) [4/21/2007 1:33:34 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_382.html
Page 382 . Suppose that u(c) =
Exercise 116 Consider the standard portfolio problem kc – 0.5c2. Show that
. Prove that it is convex.
24.7 Appendix Proof of Necessity of Proposition 88 The proof of necessity follows from lemma 9: let C be the subset of the simplex in RX which corresponds to the . Suppose, by contradiction, that information structure 1 is smallest convex envelope of not obtained by garbling messages from information structure 2; in other words, condition (24.19) is not satisfied with kij ≥ 0 for all i. Or, suppose that some posterior probability distributions of structure 2 are outside C. See an example of this in figure 24.7b. By lemma 9, we know that there exists a decision problem (v, B) such that the corresponding g function is linear in C and is strictly convex outside C. This implies that
and
because g is convex outside C. We conclude that
This result implies that experiment 1 may not be better than experiment 2.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_382.html [4/21/2007 1:33:35 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_383.html
Page 383
25 Decision Making and Information. In the preceding chapter we examined how information affects ex ante welfare. In this chapter we focus on the effect of information on ex ante actions. The optimal dynamic management of risk is enriched when we introduce the process of information acquisition over time. For example, an important element in the theory of investment is about the optimal timing of the investment when the entrepreneur anticipates getting more information over time about the market value of his product. It would be optimal for him to delay investment in a project that has a positive risk-adjusted net present value. So waiting is an option. The entrepreneur could save the sunk cost of the investment if it happens that a bad information emerges that eliminates the chance that the investment will be profitable in the long run. This example shows that the structure of the information about a future risk can affect the optimal timing of decisions. We now explore the comparative statics property of an earlier resolution of uncertainty for various dynamic decision problems.
25.1 A Technique for the Comparative Statics of More Informativeness We consider here a class of problems written as
There are four dates in this game against nature: • At date 0, the agent must take a decision α0 that affects his date-0 felicity v0(α0).
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_383.html [4/21/2007 1:33:36 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_384.html
Page 384 • At date 1, the agent receives a message m from an experiment. He revises his beliefs according to the Bayes's rule. • At date 2, he must take another decision α within an opportunity set B(α0). • At date 3, the true state of the world x is revealed, and the agent enjoys a (discounted) felicity v which is a function of α0, α, and x. The structure of information is as in the previous chapter. The important dynamic aspect of this general problem is that two decisions must be taken in the face of uncertainty: one before the partial resolution of the uncertainty given by the signal, and another one after it. Decision α0 can be seen as a decision having long-lasting effects on the environment of the decision maker, whereas decision α is an ex post decision which is taken in relation to the observed signal. Relying on the notation that has been introduced in the previous chapter, we can rewrite problem (25.1) as follows:
where function g is defined as
Observe that g(α0, pm.) is the maximum posterior expected utility that the decision maker can attain if the vector of posterior probabilities is pm., when he previously selected early action α0. We assume hereafter that the objective functions in programs (25.2) and (25.3) are concave in the corresponding decision variable. The firstorder condition for program (25.2) is therefore necessary and sufficient:
Let us now compare the optimal decisions and that are respectively associated to information structures (P1, Q1) and (P2, Q2), where (P1, Q1) is a more informative than (P2, Q2). Since we assumed that the objective function in (25.2) is concave, we will have that is larger than if and only if file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_384.html [4/21/2007 1:33:36 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_385.html
Page 385
We can eliminate
from the inequality to obtain the following equivalent property:
Since we want that this be true for any felicity function v0, this condition is equivalent to require the following condition:
which holds for all α0. Our aim is to characterize the conditions that guarantee that the decision maker increases the level of his early action α0 when the experiment becomes more informative. By proposition 87, this means that we want that condition (25.7) holds for any (P1, Q1) and (P2, Q2) that satisfy the condition that
Thus we conclude that condition (25.7) holds for all (P1, Q1) and (P2, Q2) that satisfy condition (25.8) if and only if the derivative of g with respect to α0 is convex in the vector of posterior probabilities. This result is summarized in the following proposition, which is due to Epstein (1980). PROPOSITION
91 Consider the dynamic decision problem (25.1) and its corresponding optimal posterior expected
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_385.html (1 of 2) [4/21/2007 1:33:37 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_385.html
utility function g defined by (25.3). An increase in the degree of informativeness of (P, Q) leads to an increase in the optimal early action α0 for all
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_385.html (2 of 2) [4/21/2007 1:33:37 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_386.html
Page 386 function v0(.) if and only if ∂g/∂α0 is convex in the vector of posterior probabilities for all α0. Of course, if ∂g/∂α0 is concave, an increase in the degree of informativeness of the experiment would lead to a reduction in the optimal α0. This proposition provides a useful tool to perform the comparative statics of an earlier resolution of uncertainty. We now illustrate this in various applications.
25.2 The Portfolio-Saving Problem An illustration of proposition 91 is provided by the following example. Suppose that the decision maker extracts utility u from his final consumption c that is a function of α0, α and x: v(α0, α, x) = u(c(α0, α, x)). Suppose also that the utility function is HARA and that the payoff function is additive. More specifically, suppose that
Finally, suppose that B(α0) = R for all α0, and that z′(α0) is positive. Relying on results in section 4.3, we know that the optimal posterior risk exposure α* satisfies the property that
for some function a(p., γ). This implies in turn that there is a function h(p.; γ) related to function a such that
Differentiating condition (25.11) with respect to α0 yields
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_386.html (1 of 2) [4/21/2007 1:33:38 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_386.html
Now recall that c(α0, 0, x) = z(α0) must be larger than –γη; otherwise, it would not be in the domain of definition of u. This implies that η + [z(α0)/γ] is positive. Recall also that g(α0, p.; γ) is convex with
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_386.html (2 of 2) [4/21/2007 1:33:38 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_387.html
Page 387 respect to p., by lemma 9. We can conclude that
is convex with respect to p. if γ is in [0, 1]; otherwise, it is
concave. Observe that γ ∈ [0, 1] is equivalent to condition P ≥ 2A, where A and P are respectively the degree of absolute risk aversion and of absolute prudence of u. In consequence proposition 91 directly yields the following result, which generalizes those obtained by Gollier, Jullien, and Treich (1999). PROPOSITION
92 Consider the dynamic decision problem (25.1) with v(α0, α, x) = u(c(α0, α, x)) satisfying
conditions (25.9) together with B(α0) = R and z′ > 0. Then a more informative experiment increases the optimal early action α0 if γ is in interval [0, 1], that is, if absolute prudence is larger than twice the absolute risk aversion. It reduces it otherwise. To illustrate this result, let us consider the two-period saving-portfolio problem that we examined in section 19.1.2. An agent saves α0 from initial wealth w0 at date 0, and consumes the remaining wealth. It yields utility v0 (α0) = u0(w0 – α0). At date 1, he receives a signal that is correlated to the distribution of returns of the risky asset. At date 2, he determines the optimal investment α in that risky asset, and the remaining α0 – α is invested in a risk-free asset with gross return R. He faces no borrowing constraint. At date 4, he liquidates his portfolio and consumes c(α0, α, x) = w + Rα0 + αy(x), where y(x) is the excess return of the risky asset in state x, and w is an exogenous income. Assuming that the utility function on final consumption is HARA, we are here with the ideal illustration of the proposition above. We conclude that a better information structure on the assets future returns tends to increase aggregate savings in the HARA case if P ≥ 2A; otherwise, it tends to reduce it. The intuition of this result is as follows: crudely said, a better information structure induces the investor to take more risk on average, as seen in section 24.4.3. In consequence the prudent investor will accumulate more saving to forearm himself against this increased future risk. This precautionary effect is proportional to the absolute prudence P. But the better information structure also induces the investor to be more efficient ex post, yielding larger expected future capital earnings. This has an adverse effect on current savings. This wealth effect is proportional to the degree A of resistance to consumption fluctuations. The better information structure increases aggregate savings if the precautionary effect is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_387.html [4/21/2007 1:33:38 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_388.html
Page 388 sufficiently larger than the wealth effect. It so happens that the technical condition is P ≥ 2A. This result should not come as a complete surprise for the reader. The risk-averse investor of section 19.1.2 encountered a similar effect. Indeed, let us take an information structure with two equally likely signals such that be distributed as . This implies that α(α0, p1.) = –α(α0, p2.) = α*(α0), so the . In contrast, the expected utility of the informed agent is independent of the signal and is equal to uninformed agent has an expected utility that is equal to u(α0). This reflects the fact that by construction, the expected return of a risky asset is zero. Clearly, the uninformed agent should not purchase a risky asset. Now observe that having access to the information generates exactly the same effect as opening an asset market whose excess return is distributed as . We know from proposition 75 that having information increases saving if P is larger than 2A. There are at least two interesting points related to this special structure of information. First, we see that the effect of information is here equivalent to the effect of introducing an asset market. Second, this tells us that the condition P ≥ 2A that we obtained to be necessary and sufficient for the HARA case is at least necessary when u is not HARA. For small risks we obtain the same result by using Taylor expansions. Suppose that the expectation of small with regard to its variance, for all m. From the results in section 4.2, we have that
In turn
with
Differentiating g with respect to α0 yields
Since we know that g is convex in pm., we can conclude that ∂g/∂α0 is convex for P larger than 2A. By proposition 91, aggregate saving is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_388.html [4/21/2007 1:33:39 AM]
is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_389.html
Page 389 increased with the improvement in the information structure. So we have another proof for the necessity of condition P ≥ 2A. To sum up, we have seen three different ways in which the condition P > 2A provides a better information structure for increasing savings. The first, based on proposition 92, shows this condition to be necessary and sufficient when preferences exhibit HARA. The second introduces the information structure , which brings this condition to cases outside of HARA. The third admits Taylor approximations that confirm that P > 2A is a necessary and sufficient condition for portfolio risks that are small.
25.3 A Digression: Scientific Uncertainty, Global Warming, and the "Precautionary Principle" For readers who are interested in environmental economics, there is an alternative interpretation to model (25.9). In an environment of scientific uncertainty, it is often the case that one must take action for the prevention of a risk before the resolution of scientific uncertainty. Program (25.1) is perfectly fitted to analyze the optimal timing of the preventive actions in such a situation. Recent events provide several illustrations of the problem: global warming, genetic manipulations, mad cow disease, radiological exposures, cellular phones, and the lest goes on. Here, we focus on global warming. A big aspect of global warming is the scientific uncertainty about the extent of risk of damage to the economy. This question is related to the prospect of scientific progress and its connection to the timing in the effort to reduce emissions of greenhouse gases. Some scientists particularly those in the United States, suggest that we should wait for better information about the effects of global warming before proceeding with drastic fiscal or market policies that seek to drastically to limit emissions. Most leaders of the Green parties in Europe, in particular, are opposed to this idea. The Rio Conference favored the European position by promoting a "Precautionary Principle," so-called because it states that the absence of certainty, given our current scientific knowledge, should not delay the use of measures, at an acceptable cost, that prevent the risk of large and irreversible damages to the environment. What the Precautionary Principle puts at stake is the expectation of future scientific progress and its effect on the timing of any effort to prevent further environmental damage from occurring. Mathematically we could consider this using a two-period model
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_389.html [4/21/2007 1:33:40 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_390.html
Page 390 where et represents the emission of carbon dioxide in period t = 0, 1. Carbon dioxide emission is a well-known side effect in the production of energy. Say, there is a one-to-one relationship between the quantity of energy produced in period t in the quantity of gas emitted in that period. We let et represent both energy and emissions in period t. We will assume that the sources of energy are free. In period 0, the population extracts a surplus v0(e0) without consuming energy e0. In period 1, the population consumes energy e1, and contributes to the risk of global warming a quantity proportional to the total quantity of gases emitted during the two periods, e0 + e1. The damage to the economy expressed in energy-equivalent is equal to (e0 + e1)d(x), where x is a random variable. The uncertainty on x reflects the uncertainty in scientific knowledge. Between periods 0 and 1, scientific findings are made based on some experimental studies. The problem of the policy maker may be written as follows:
Let us define function y as y(x) = 1 – d(x). This makes the problem above equivalent to
where α is the aggregate concentration of greenhouse gases in period 1. Following proposition 92, and setting α0 = –e0, we can conclude that in HARA, the prospect of scientific discovery increasing our understanding of the greenhouse effect is enough to induce society to reduce current emissions of greenhouse gases provided that P > 2A. This strategy has the same structure as the increase in saving in the portfolio-saving problem. Both strategies transfer welfare to the future. Gollier, Jullien, and Treich (2000) provide more insights on this environmental result. Ulph and Ulph (1997) approach the problem using a different framework.
25.4 The Saving Problem with Uncertain Returns Let us consider now a model where the decision maker consumes at three different dates. At date 0, he is endowed with wealth w0, and he decides how much to consume from it, and how much to save. Let
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_390.html [4/21/2007 1:33:40 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_391.html
Page 391 α0 denote savings at date 0. There is a single asset whose gross return ρ0 over the period between dates 0 and 2, and it is certain. At date 2, the agent determines how much to consume out of α0ρ0, and how much to save for the last date, date 3. Let α denote the remaining savings at date 2. As of date 0, there is some uncertainty about the of the single asset between dates 2 and 3. But at date 1, the agent observes a signal that is gross return correlated with the state of the world x that will prevail at the last date. What is the effect of this information on saving at the initial date? Mathematically the problem can be stated as
This is an application of problem (25.1) with v(α0, α, x) = u(ρ0α0 – α) + βu(ρ(x)α). Epstein (1980) solved the problem for constant relative risk aversion using u′(c) = c–γ. Constant relative risk aversion implies that
with
By the envelope theorem, we have
Combined with condition (25.20), it is rewritten as
Now it is easy to check that function f defined as f(t) = (1 + t1/γ)γ is convex if γ is in interval [0, 1], and is concave otherwise. We can use proposition 91. Since is linear in the vector of probabilities of , we conclude with Epstein (1980) that the level of saving at date 0 increases with the degree of informativeness if relative risk aversion is less than 1, and that the level of saving at date 0 is reduced by it if relative risk aversion is larger than unity. Contrary to what we obtained in the saving-portfolio problem, it is not clear that this result can be extended to other utility functions besides those of CRRA.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_391.html [4/21/2007 1:33:41 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_392.html
Page 392
25.5 Precautionary Saving Again, let us look at a model where the decision maker consumes at three different dates. There is a single asset whose return is sure and normalized to zero. At the last date of consumption, the agent earns an uncertain income . If he receives some information about it between the two consumption dates 0 and 2, the decision problem is written as
We want to determine the effect of a better information on optimal saving before the information is received. We are not aware of any theoretical result on this question, and we have not been able to determine whether the function ∂g/∂α0 associated with this problem is concave. You are asked to show that it is concave in the CARA case at the end of this chapter. So in this section, we prove that perfect information always reduces the level of savings under prudence. The intuition is straightforward for why an earlier resolution of uncertainty can induce a smaller level of savings. Better information allows for the future risk to be diversified over time. The future reduction in risk is an incentive for prudent agents to reduce their precautionary savings. To show this, let us simplify the problem by assuming β = 1 and u0 ≡ u. Under perfect information it is always optimal for the agent to smooth his consumption perfectly. The problem at date 0 then becomes
Its first-order condition is written as
where is the optimal saving of the informed agent ex ante. In contrast, in the absence of information at date 1, the agent will want to smooth his consumption over the two first consumption dates. He must swallow all the risk associated with at the last date. His decision problem in this case is written as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_392.html [4/21/2007 1:33:42 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_393.html
Page 393
The optimal level of saving of the uninformed agent is larger than condition is rewritten as
Now, we can make the following changes in variables, with rewrite conditions (25.25) and (25.27) as
if and only if
is positive. This
and
. We
By the hyperplane separation theorem 11, the reader can immediately verify that this is true if and only if u′ is convex. So we see that a complete early resolution of the income uncertainty reduces the level of savings by all prudent agents.
25.6 The Value of Flexibility and Option Value. Up to now we have been considering decision problems (25.1) that are unconstrained: B(α0) = R for all α0. In this section we examine the case where the opportunity set B for the future decision α is a function of the early decision α0. To focus this aspect of problem (25.1), we assume that the final payoff v does not directly depend on α0, meaning that v(α0, α, x) = u(α, x) for all α0. We have seen that the value of information comes from the ability of the decision maker to adapt his decision to information. This flexibility makes more information valuable. If there is no cost to getting more flexibility, the agent will always select the environment that have the maximum flexibility. But flexibility is costly: holding open options for the future is often costly today. The decision maker does not really maintain maximum flexibility if this is costly to do so. He must trade off the future benefit of some open options with the immediate cost of doing so. The idea is that an increase in informativeness will always induce the decision maker to bias his ex ante decision in favor of more flexibility. Intuitively flexibility is at a premium when information is more forthcoming. We let α0 indicate the degree of flexibility in a decision. Now B(α0) is a subset of
whenever
than α0. Increasing α0
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_393.html [4/21/2007 1:33:42 AM]
is larger
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_394.html
Page 394 enlarges the opportunity set in the future. To keep things simple, let us assume that B(α0) is the set of all α that are less than α0: B(α0) = {α|α ≤ α0}. Observe that v0(α0) in problem (25.1) can then be interpreted as the immediate cost associated with the degree of flexibility α0. The Kuhn-Tucker conditions for problem
are written as
We also know that ∂g/∂α0 is zero if the constraint is not binding; it is equal to
when it is binding,
that is, when α* = α0. The Kuhn-Tucker conditions imply that
We see that ∂g/∂α0 is the maximum of two linear functions of the vector of probabilities, so the right-hand side of (25.31) is a convex function of it. From proposition 91 we conclude that a more information induces the decision maker to adopt a more flexible strategy. Arrow and Fischer (1974) and Henry (1974) were the first to obtain this result. It was later extended by Freixas and Laffont (1984). PROPOSITION
93 Consider problem (25.1). Suppose that the payoff function v is independent of α0 and that an
increase in α0 means more flexibility:
. Then all decision makers increase their
degree of flexibility α0 as the experiment becomes more informative. This result can be applied for the typical capacity choice problem. As in the standard model (25.1), there are three dates. At date 0, an entrepreneur must determine his future capacity of production α0. Each unit of capacity implemented generates a sunk cost k. The utility of the entrepreneur at date 0 is v0(α0) = u0(w0 – kα0). At date 2, the entrepreneur faces some uncertainty about its profit margin per unit of production. The profit margin in state x is denoted y(x). He selects the level of production α that maximizes his expected utility
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_394.html [4/21/2007 1:33:43 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_395.html
Page 395 under the constraint that α does not exceed the preselected capacity α0. Before doing that, but after the capacity choice α0, at date 1, he observes a signal
which is correlated to . By proposition 93, a
better information structure à la Blackwell necessarily increases the optimal capacity α0. The general result that more information induces the selection of a more conservative early action has been widely recognized in the modern theory of real investment. McDonald and Siegel (1986), Pindyck (1991), and Dixit and Pindyck (1994) show that it may be optimal to delay an investment with a positive marginal net present value (NPV) if more information is expected to come about the distribution of future cash flows. Said differently, the NPV rule must be adapted to take into account of the value of information that can be extracted if the investment is delayed. An option value to wait must be included in the NPV. We can illustrate this using a simple example. Suppose that contrary to the capacity problem above, the investment problem is now a discrete choice problem. The question is not how much to invest but when to invest. At date 0, the agent may decide to invest in a project whose implementation cost is K. It is a sunk cost. If he does invest at that date, this will give him the option to produce a widget at date 2 and at date 3. The margin profit of the widget at date 2 is sure and is equal to y2 > 0. The margin profit of the widget produced at date 3 depends on the state and is denoted y3(x). If he decides not to implement the project at date 0, he can still do it at date 2, in which case he will pay K to produce the at date 1 that is correlated to . Let us further assume widget at date 3. He can do so after observing a signal that markets are complete and that is not correlated with the market risk. This implies that the agent will fully insure the profit margin risk at a fair price. Finally, the risk-free rate is constant over time and is equal to r. Let α0 be the discrete choice variable at date 0. It takes value 1 if the agent implements the project, and it equals zero otherwise. Decision variable α takes value 1 if the agent has the opportunity to produce the widget at date 3, and is zero otherwise. We see that B(α0 = 0) = {0,1}, but B(α0 = 1) = {1}: in other words, the agent looses flexibility by deciding to invest immediately at date 0. In doing so, he becomes unable to obtain any value from the information coming at date 1. The dynamic decision problem of the agent is written as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_395.html [4/21/2007 1:33:44 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_396.html
Page 396
This problem has α0 both in the constraint α ∈ B(α0) and in the value function v(α0, α, x). But because we assumed risk neutrality, only the flexibility effect of a change in α0 will matter. To keep things simple, let us compare the optimal decision rule of this problem with the optimal decision rule with uninformative signals. Without information it is optimal to undertake the investment at date 0 if and only if
The left-hand side of this inequality is the NPV of the project at date 0, whereas the right-hand side is the maximum NPV that we can obtain from delaying the investment. If we suppose that it is optimal for the uninformed agent to invest at date 2 if he did not do so at date 0, condition (25.33) is simply rewritten as y2 ≥ rK: the marginal NPV of not delaying the project must be positive. Given an informative signal, it is optimal to invest in the project at date 0 if and only if
Because the max operator is convex, IC is larger than UC. The existence of information to come about future cash flows provides an incentive to delay investments. In fact, in the presence of signals correlated to cash flows, the NPV rule (25.33) must be adapted to include an option value IC – UC to the strategy of postponing the project. This option value is the value of information in the decision problem
As is obvious, this value of information is increasing with the degree of informativeness of the experiment. We conclude that more information à la Blackwell always tends to favor the option to delay the investment. Delaying the investment provides more flexibility for the future. In particular, it allows the investor to save the sunk cost if it
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_396.html [4/21/2007 1:33:44 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_397.html
Page 397 happens that the prospect of profitability of the project, given the signal, is poor. This is another illustration of proposition 93. McDonald and Siegel (1986) provide a rule-of-tumb to use in computing the option value of irreversible investments when the profit margin of the output follows Brownian motion. They take the current profit margin as the indicator of future profit margins. The concept of option value has also been applied in environmental economics. The literature on it originated from a question raised by Arrow and Fischer (1974). They considered the problem in terms of a cost–benefit analyses for environmental projects that have some kind of irreversibility, like the construction of a dam in a remote valley. The unexploited valley could be used by future generations for a different purpose whose added value is unknown as of today. But building a dam in the valley today is an irreversible phenomenon: if it is built, there will simply be no other way for the valley to be used in the future. Not building the dam is a strategy that leaves the options open for the future. The option value associated to this conservative strategy must be included in the cost–benefit analysis of the dam project. The same argument has been used for endangered animal species (if a species is lost, there is no way to recreate it), for prohibiting genetic manipulations (once done, there is no return), for determining the optimal reduction of emissions of greenhouse gases, for determining the degree of reversibility in the disposal of nuclear wastes, and so on.
25.7 Predictability and Portfolio Management The rest of this chapter is devoted to a dynamic decision problem that cannot be written as a program (25.1). We look at the problem of a two-period portfolio with complete markets in each period, as in section 14.4. In that section, as in chapter 11, it was assumed that time does not bring new information about future returns. Said in other words, future returns are not predictable. The absence of predictability is not compatible with the data, according to the results of Campbell (1987), Campbell and Shiller (1988a, b), Fama and French (1988, 1989), and Campbell (1991). They showed that there is some predictability in asset returns. In the real world, investors and econometricians can observe signals that are correlated with future returns. In this regard the investor's horizon may affect the optimal portfolio of the agent even in the HARA case, as demonstrated by Merton (1973) and Barberis (2000), for example. Our objective here is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_397.html [4/21/2007 1:33:45 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_398.html
Page 398 to examine more generally how the predictability of future returns affects the optimal exposure to risk prior to the revelation of the prediction/information. Let us consider the simplest model that deals with this question. This is a two-period model indexed by t = 0, 1. Consumption takes place only at the end of the second period (the investment problem). In each period there is a complete set of markets contingent on the state of the world that will be revealed at the end of the period. is is revealed and also a signal is received. The the price kernel in period t. At the end of the first period, existence of a first-period risk makes the model richer than the models examined earlier in this chapter, as it allows for information to be conveyed about future returns through a risk occurring in the first period. The , making this the predictability hypothesis. distribution of is conditional to the pair We solve the dynamic problem by backward induction. In period 2, given the realization (π0, m) observed at the end of the first period, the investor selects his portfolio that maximizes the expected utility of his final consumption:
where z is the accumulated wealth at the beginning of the second period, and v(z; π0, m) is the maximum expected utility that one can obtain with wealth z, given information (π0, m). The optimal consumption plan is given by c(π1, z; π0, m). The maximum utility that one can extract in turn from z prior to signal
is thus equal to
V(z; π0), which is given by
Finally, at the beginning of the first period, the investor selects the portfolio c0(.) which maximizes V:
where w0 is the initial wealth of the investor. is We examine two extremes. In one case, which is most in line with what we did before, the realization of is uninformative for : these random variables are assumed to be independent. Only the observation of is informative, and is not correlated to . As we will informative. In the other case, we assume that only see,
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_398.html [4/21/2007 1:33:45 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_399.html
Page 399 the effects of predictability on the optimal dynamic portfolio are quite different depending on whether or not informative signals are correlated with current returns. 25.7.1 Exogenous Predictability Let us begin by assuming that the source of predictability is exogenous to financial markets, which means, that there is no serial correlation in assets returns. So functions v, c, and V do not directly depend on π0. This is a model where the investor does not know the true distribution of returns. But he obtains exogenous information about it over time. This information is not conveyed by the observation of past returns. For example, the exogenous information about future returns could be the outcome of an election, or the occurrence of a technical innovation. Signal provides this information. How does the existence of such uncertainty about the distribution of future returns affect the optimal portfolio prior to the observation of the information? To answer this question, we just have to examine how the concavity of the single-variable function V is affected by this uncertainty. Notice that V(.) is a weighted average of the v(.; m) functions, whereas the degree of absolute risk tolerance of v(.; m) is given by
In this section we assume that information.
for all m, namely that the risk-free rate is unaffected by the
Suppose first that u is HARA, meaning that T is linear. Suppose also that signals affect the distribution of the is independent of m. Now, from excess return, but not the risk-free rate. In short, we can assume that conditions (25.37) and (25.40), we have that Tv(z; m) = T(z). In particular, Tv is independent of m. By property (25.38), we see that V is proportional to u, and the information structure has no effect on the optimal portfolio ex ante. In consequence, under HARA, investors are myopic twice: they are myopic with respect time horizon, and they are myopic toward exogenous predictability. Hedging against unfavorable changes in investment opportunities is not possible when the opportunity shifts are statistically independent of earlier returns. We can also claim that convex (concave) absolute risk tolerance is necessary for exogenous predictability to raise (reduce) the optimal exposure to risk ex ante. Treich (1997) shows this by examining
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_399.html [4/21/2007 1:33:46 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_400.html
Page 400 small standard portfolio risks. Another way to prove this claim is to examine the case where the unconditional distribution of future returns has a zero expectation. In that case the uninformed agent takes no risk in period 2. The uninformed agent has a function V identical to function u in such a situation. In contrast, in the economy with exogenous predictability, some signals will yield positive and other negative expected conditional returns, and the investor will purchase a risky portfolio. In that situation, V is a weighted sum of various v functions that are all less risk-averse than u if T is convex. Thus, under this condition, exogenous predictability makes the agent more willing to take a risk ex ante. This is because when the unconditional expectation of future returns is zero, the information structure has the same effect as giving investors the valuable option to become risk-takers. 94 Consider the case of exogenous predictability: there is no serial correlation of asset returns, but some exogenous independent signals affect future asset returns. Assume that the risk-free rate is unaffected by the information. Under linear absolute risk tolerance (HARA), this predictability has no effect on the optimal portfolio risk ex ante. A necessary condition to increase (reduce) it is that absolute risk tolerance be convex (concave). PROPOSITION
25.7.2 Endogenous Predictability and Mean-Reversion Suppose, alternatively, that is correlated with but is not informative. Here, functions v, c, and V ≡ v do not depend on m, which we eliminate from our notation. A good example of this problem is given by McCardle and Winkler (1992): in a casino, a certain coin is tossed twice, and players can gamble on each toss. The problem is that players do not know the probability of a head or tail outcome. Because the same coin is tossed twice, the first outcome conveys information about the distribution of future payoffs. What is the optimal dynamic strategy in this game against nature? In particular, would one have a different risk factor in stage 1 of the game if two different coins are used? Detemple (1986), Gennotte (1986), King and Leape (1987) and Brennan (1998) examined this question for the portfolio problem. Another example of predictability is provided by bond portfolios. Suppose that there is a short-term bond and a long-term bond. Suppose also that there is some uncertainty about the future short-term interest rate. In such a situation, any good news in the second period
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_400.html [4/21/2007 1:33:47 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_401.html
Page 401 (high interest rate) is compatible only with any bad news in the first period (low return on the long-term bond) because of the negative relationship between the interest rate and the price of long bonds. A larger first-period return yields a reduction in the second-period return. Brennan, Schwartz, and Lagnado (1997) solve numerically a simple version of this dynamic portfolio problem. Let us assume that
where is independent of . We have mean-reversion in state prices if function g is decreasing in its first argument as is the case in the bond portfolio problem. Intuition suggests that the presence of mean-reversion is an incentive for taking more portfolio risk. Indeed, the negative serial correlation in returns allows for a better diversification of the lifetime risks. In the example with a biased coin, there is rather a positive serial correlation of the payoffs. Intuition suggests that a long-horizon agent is less of a risk-taker in the initial stage of the game. Merton (1973) was the first to examine an economy where extra demand for stocks emerges when the available investment opportunities change as returns change. This additional holding by long-horizon investors is referred to as "hedging demand." Early investment assets are used to hedge against future changes in the investment opportunity set. The surprise is that hedging demand may be negative. We will see that the reason relates to the optimal portfolio rebalancement strategy of the investor. The following proposition allows us to begin to examine this question. PROPOSITION
95 Consider first the case where
follow a mean-reversion process: ∂g/∂π0 < 0. Then the marginal
value of wealth is a decreasing (increasing) function of π0 if relative risk aversion R(c) = –cu″(c)/u′(c) is uniformly larger (smaller) than unity:
The inequalities are reversed in case of positive serial correlation [∂g/∂π0 > 0]. Proof We take the following problem:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_401.html [4/21/2007 1:33:47 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_402.html
Page 402 We know that v′(z; π0) is equal to the Lagrangian multiplier ξ(z; π0), which must satisfy the budget constraint
where to π0 yields
where
, and is positive and increasing. Fully differentiating the budget constraint with respect
is distributed as
. It is easy to check that
is equivalent to
. Condition (25.45), together with ∂g/∂π0 < 0, implies that ξ and v′ are decreasing in π0. If R ≤ 1, we have that
for all y. This implies that ξ and v′ are increasing in π0. The opposite results
hold when g is increasing in π0. An FSD-dominated shift in the state–price density has substitution and wealth effects. The substitution effect comes from the increased efficiency of savings due to the increase in asset return in every state. It raises the marginal value of wealth. But the FSD-dominated shift in the state–price density increases consumption in every state. Because the marginal utility on consumption is decreasing, this tends to reduce the marginal value of wealth. This wealth effect is increasing with the degree of relative risk aversion. Thus, if we want the wealth effect to dominate the substitution effect, we need to put a lower bound to relative risk aversion. The proposition above states that then bound is set to unity. An application of proposition 95 is in the two-period portfolio-saving problem examined in section 19.1.1. In the model an FSD-improvement in market returns reduces savings when relative risk aversion is larger than unity. We can now turn to the period-0 portfolio decision, whose first-order condition to problem (25.39) is written as
The superscript p refers to the solution for the predictable world. We compare the optimal portfolio in the in the unpredictable world. Without loss of generality, suppose predictable world to the optimal portfolio that there exists a scalar such that the uncon-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_402.html [4/21/2007 1:33:48 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_403.html
Page 403 ditional be distributed as written as
. It implies that the period-0 portfolio problem in the unpredictable world is
It yields the first-order condition
where is the optimal portfolio in the unpredictable world and λu is the associated Lagrangian multiplier of the budget constraint. Combining the two first-order conditions (25.46) and (25.48) yields
Intuition suggests that mean-reversion should induce risk-averse investors with R > 1 to take more risk in period 0. Indeed, by proposition 95, the marginal utility of wealth is increased in low-price states, which are states in which wealth is already large. This property certainly holds if λu = λp. Now compare the two state prices and . Without mean-reversion, we would have : consumption in the high-price state is smaller. and in the unpredictable world measures the risk The difference between optimal consumptions taken ex ante in that world. R > 1 makes v′ decreasing in π in the predictable world. If λu = λp in equation (25.49), this directly implies that and cp(π) < cu(π). This means that cp single-crosses cu from above at . From our discussion in section 13.3, we learn here that investors increase initial risk-taking because of the meanreversion in returns. There is then a positive (hedging) demand for risky assets. In general, however, λp is not equal to λu: the ex ante marginal utility of wealth is affected by the statistical relationship in returns per period. This means that in addition to the effect presented above, mean-reversion also yields a wealth effect. It can potentially go in a direction opposite to that of hedging demand. To fully characterize the situation, following Merton (1971), let us assume that u is the CRRA function u′(z) = z–γ. As is well-known, the marginal value function can here be written as
By proposition 95, we know that h is increasing when γ < 1, whereas h is decreasing when γ > 1. Condition (25.49) is then
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_403.html [4/21/2007 1:33:49 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_404.html
Page 404 rewritten as
Now we seek to define π* such that
. The equality above is thus equivalent to
Obviously, if h is decreasing (increasing), cp single-crosses cu from above (below) at π*. Combining this with proposition 95 yields the following result. 96 Consider the two-period investment problem with mean-reversion and CRRA preferences. Demand hedging for risky assets is positive (negative) if relative risk aversion is larger (smaller) than unity. The opposite result holds in case of positive serial correlation. PROPOSITION
In particular, the logarithmic investor does not hedge demand at all. The logarithmic agent is thus not only myopic with respect to the time horizon and exogenous predictability; he will also be myopic to serial correlation in returns. This proposition further tells us that the CRRA utility function is no longer myopic when there is serial correlation in returns. That is to say, when relative risk aversion is constant and larger than unity, the presence of mean-reversion induces the agent with a longer time horizon to take more risk. This observation is intuitive, for taking risk in period 1 is a way to reduce the risk taken in period 0. On the other hand, when relative risk aversion is constant and less than one, hedging demand for risky assets will be negative, and the investor with a longer time horizon will want to reduce his risk exposure starting at date 0. The explanation has to do with the greater intensity of the substitution effect that induces a large marginal utility of wealth when the period-0 return is small (∂v′/∂π0 > 0). In this case where wealth is small, the marginal utility of wealth at the end of period 0 is larger due to mean-reversion. As a result the agent will take less risk in period 0.
25.8 Conclusion The comparative statics of incremental information effects is a field of research that is still in its infancy. The results available mostly
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_404.html [4/21/2007 1:33:49 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_405.html
Page 405 hold for very specific utility functions, or under risk neutrality. The parallel that can be drawn to the effect of an increase in risk makes this already complex matter considerably more interesting, as we have seen in section 4.5, and in the more complicated model we examined in this chapter. The complexity increases as more informative experiments are made and the mean-preserving effect spreads in a spatial dimension X – 1 ≥ 1. The set of meanpreserving spreads is made much richer than in the one-dimensional wealth space. We still must learn whether the comparative static property holds for a wider set of possible distribution changes.
25.9 Exercises and Extensions Exercise 117 Consider the two-period saving-portfolio problem as in section 25.2, but where markets are complete:
• Define function
as follows:
(y) = u′–1(1/y). Show that
is convex if and only if P ≥ 2A.
• Show that gs(s, p.) = ξ(s, p.), where ξ is the associated Lagrangian multiplier that satisfies the following condition:
for all λ ∈ [0, 1] and all pair of p1 and p2
• Show that belonging to the simplex of RX if and only if function
is convex.
• Show that a more informative experiment leads to an increase in saving if and only if P ≥ 2A. Exercise 118 Consider the two-period saving-portfolio problem of section 25.2 again, but with a no-short-selling constraint α ≤ α0:
Show that in the CRRA case a more informative experiment increases the optimal saving if relative risk aversion is less than unity. file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_405.html [4/21/2007 1:33:50 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_406.html
Page 406 Exercise 119 Consider problem (25.23) with a utility function u being CARA. Show that a better information à la Blackwell always reduces the optimal level of precautionary saving α0. Exercise 120 Consider a risk-neutral entrepreneur who has a new investment opportunity. He lives for n periods, and the interest rate is a constant 10% per period. At any time he can decide to pay a sunk cost K that will allow him to produce for free one widget per period from the next period until his retirement. The entrepreneur is pricetaker. The price of the widget in period 1 is P1. At each period, the price can be increased or reduced by 50% with equal probability, with respect to the price one period earlier. • Take n = 3. Describe the optimal investment strategy in period 1 as a function of K and P0. Explain why this optimal strategy departs from the standard NPV rule, which states that it is optimal to invest in period 1 if and only if K is less than (1.1)–1 P0 + (1.1)–2 P0. • Calculate the optimal price-contingent timing of investment when n = 4. Exercise 121 (Detemple 1986; Gennotte 1986; Brennan 1998; Brennan and Xia 1999) Reconsider the two-period portfolio problem without serial correlation that we examined in chapter 11. Suppose that the return of the risky asset is distributed as (a, p; b, 1 – p). The investor does not know parameter p, but he has some prior beliefs about it. The observation of the first-period return allows him to update their beliefs by using Bayes's rule. • Show that a large return in the first period yields a mean-preserving spread in the distribution of state prices in the second period. • Show that a mean-preserving spread in the distribution of state prices reduces the marginal value of wealth if and only if the ratio of absolute prudence to absolute risk aversion is smaller than 2. • Show that this implies that CRRA investor with a relative risk aversion larger than unity reduce their demand for stocks in the first period due to this parameter uncertainty.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_406.html [4/21/2007 1:33:51 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_407.html
Page 407
26 Information and Equilibrium. In the previous two chapters, we examined the effect of information on optimal decisions and on the decision maker's optimal expected utility, taking prices as given. But information, when it is available for all participants to a market, affects prices. This must be taken into account in the value of information. We examine the effect of information on welfare in the first part of this chapter. In the second part we determine how information affects average prices.
26.1 Hirshleifer Effect Realized risks cannot be insured. This simple maxim is at the heart of the paradox that information can be bad for the economy as a whole. Hirshleifer (1971) was the first to challenge the fact that the value of information is always nonnegative, by taking into account the effect of information on equilibrium prices. Stochastic changes in equilibrium prices introduce some additional uncertainty on risk-averse agents' wealth ex ante. Obviously this would have an adverse effect on their welfare. To illustrate this more fully, we consider an Arrow-Debreu economy with risk-averse agents indexed by θ ∈ Θ. Agent θ has utility function u(c, θ), as in chapter 22. The initial endowment of agent θ is a function of the state of . There is a complete set of markets for claims contingent to states. Prior to the the world x and is denoted opening of these markets, all agents observe the same signal, which is correlated to the realization of the state of the world.10 As
10. When the signal is observed only by a subgroup of the population, or when the signal is not the same for everyone, a problem of asymmetric information emerges in the economy. A huge part of modern economic theory is devoted to this problem, mostly taking a partial equilibrium approach. The book by Salanié (1997) provides an excellent introduction to this literature.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_407.html [4/21/2007 1:33:51 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_408.html
Page 408 before, qm denotes the probability of occurrence of signal m, with m = 1, . . . , M. Let pmx be the posterior probability of state x when signal m is observed. Contingent markets subsequently operate, and an equilibrium associated to that signal emerges. Let πm(x) and cm(x, θ) denote respectively the state price per unit of probability and agent θ's consumption is state x at the equilibrium associated with signal m. The definition and the properties of these posterior equilibria are as described in chapters 22 and 23. We are here performing a welfare analysis ex ante, that is, before observing the signal. An allocation in this economy is denoted by A = (cm(., .))m=1, . . . , M. Note that we assume that agents cannot invest in preventive activities. The only possible actions are portfolio rebalancements. In this way we minimize the value of information in this economy. Various forms of the following proposition are due to Marshall (1974), Green (1981), Hankansson, Kunkel and Ohlson (1982), and Schlee (1999). It states that if some agents are better off with information, some other agents must be hurt by it. Any information structure must be bad for at least one agent in the economy, when compared to the economy where the information is "prohibited." 97 Suppose that agents cannot invest in preventive activities. Then the competitive allocation from partial information cannot Pareto-dominate the competitive allocation Au = (cu (., .)) with no information. PROPOSITION
Proof Let Define allocation
characterize the signal-contingent structure of the competitive allocation. in such a way that
Figure 26.1 Pareto frontier and competitive equilibria
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_408.html [4/21/2007 1:33:52 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_409.html
Page 409
for all m, x, and θ, where qkx is the probability of signal k in state x. This is in fact a signal-independent allocation in which each agent's consumption is averaged over the various possible signals in each state of the world. It is trivial to verify that this allocation is feasible. By risk aversion, all agents prefer allocation A° over allocation Ai, since
where
x
is the probability of state x. The ex ante expected utility obtained at equilibrium with information is
less than the expected utility with allocation A°. Thus there is a feasible allocation in which no information is used and in which Pareto-dominates the competitive allocation with information. We conclude that this allocation is below the Pareto frontier of the economy with no information. By the first theorem of welfare economics, the equilibrium with information may not be Pareto-superior to the competitive equilibrium without information. Figure 26.1 illustrates the proof of proposition 97. Suppose that there are two agents in the economy. Without any information, efficient risk-sharing arrangements are represented by the concave Pareto frontier, and the competitive equilibrium, Au, is one of them. There are two possible signals. The figure does not show the Pareto frontiers associated with these two signals, but just the corresponding competitive equilibria. Agent 2 is particularly favored if signal 1 is received (he has a large endowment in a more likely state), and agent 1 is more favored if signal 2 is received. Their ex ante expected utility is along the segment linking the two competitive equilibria, since expected utility is linear in probabilities. Proposition 97 is proved by setting up a feasible allocation A° that does not use information and that is preferred to Ai by the two agents. In fact Ai is obtained from A° by adding zero-sum payoffs that are conditional to the signal in each state of the world x ∈ X. That is to say, going from
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_409.html [4/21/2007 1:33:53 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_410.html
Page 410 Ai to A° provides full insurance against signals at fair price. This makes all risk-averse consumers better off. The simplest case that illustrate this result is where signals are completely conclusive, meaning that they always reveal the true state of the world. In a situation like this, people have nothing to exchange after the signal is observed, and they will consume their endowment ω(x, θ). In other words, after information that state x will occur is revealed, the price of all Arrow-Debreu securities associated to x′ ≠ x goes down to zero. The autharcy is an equilibrium, ex post. From an ex ante point of view, everyone is hurt by the information, since all mutually advantageous risk-sharing arrangements are eliminated. Realized risks cannot be insured. The most striking case is where there is no aggregate risk. That is to say, without information all consumers are fully insured at fair price. In this extreme case the value of information is negative for everyone. It is equal to minus the risk premium associated to the initial endowment. The discussion above seems to suggest that agents are always hurt by the introduction of information in the economy. This is not true in general, however. Because competitive prices are related to probabilities in a complex manner, the average market value of the initial endowment of some players can be made larger by information. That can compensate the (partial) loss of insurability which we stressed above. An example of this is shown in figure 26.1, where agent 1 is made better off by the information. The following lemma provides a technical condition guaranteeing that information is always detrimental to welfare. 10 Consider the ex post competitive equilibrium allocation. The introduction of information is Paretoinferior, meaning that no agent is better off at equilibrium, if, for each possible signal m, there exists km such that LEMMA
where πm(.) and π(.) are the equilibrium price kernels respectively conditional to signal m and in the uninformative economy. Proof This proof is based on the proof of proposition 97. Consider allocation A° characterized by equation (26.1). We show that consumption plan c°(., θ) is in the uninformed budget set of agent θ. Indeed, we have the following sequence of relations:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_410.html [4/21/2007 1:33:53 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_411.html
Page 411
The first equality is from the definition (26.1) of allocation A°. The second equality is the usual property of conditional probabilities. The third is just a rewriting of the previous property, using our assumption that πm(x) = kmπ(x). The fourth equality is a direct use of the budget constraint of agent θ in the economy receiving signal m. Thus agent θ can at least consume c°(., θ) in the competitive economy with no information. This implies that agent θ prefers his optimal consumption plan cu(., θ) to c°(., θ). But we also know from the proof of proposition . We 97 that consumption plan c°(., θ) is preferred to the signal-contingent consumption plan u i conclude that A dominates A°, which in turn dominates A , in the sense of Pareto. What do we know about the relationships between the competitive prices of Arrow-Debreu securities and the probability distributions of states? Because equilibrium prices are also equal to the marginal utility of the representative agent, this problem can be answered by determining the relationship between the distribution of state probabilities and the attitude toward risk of the representative agent. For this, we need to go back to chapter 23. The system of equations (23.1) describes this relationship, where π(x) = v′(z(x)) is the marginal utility of the representative agent consuming z(x). Because the distribution of (and therefore of ) explicitly appears in the budget constraints (second equation of the system), its solution in general depends on this distribution. There are exemptions to this, however.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_411.html [4/21/2007 1:33:54 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_412.html
Page 412 They are listed in section 23.1. Suppose, for example, that there exists a risk neutral agent in the economy. In such a situation we know that all risk-averse agents are fully insured at fair price by the risk-neutral agent at the competitive equilibrium. Fair price means that contingent prices are proportional to state probabilities, or that state prices per unit of probability are constant. Notice that state prices per unit of probability are also constant when there is no macroeconomic risk. Another exception is when all agents have linear absolute risk tolerance with the same slope: T(c, θ) = t(θ) + bc. As shown by equation (26.3), the degree of concavity of the utility function of the representative agent is independent of the probability distribution of the states. In other words, a change in the market value of initial endowments, due to the information, does not affect the attitude toward risk of the representative agent. Since the only thing that a change in state probabilities does to the economy is to affect the market value of initial endowments, we conclude that equilibrium contingent prices (per unit of probability) and the attitude toward risk of the representative agent are not affected by a change in probabilities. These results are summarized in the following proposition, which is due to Schlee (1999). 98 Consider the competitive equilibrium allocation. The introduction of information is Paretoinferior, meaning that no agent is better off at equilibrium if any one of the following conditions is satisfied: PROPOSITION
1. There is a risk neutral agent in the economy. 2. There is no aggregate risk. 3. All agents have a linear absolute risk tolerance with the same slope: T(c, θ) = t(θ) + bc. An important application of the potentially Pareto-deteriorating effect of information concerns genetic testing. Recent advances in genetic testing allow us to predict more accurately whether an individual has a specific health risk. Genetic tests are valuable in the effort to focus health prevention measures for individuals who face a larger risk to develop the disease. But with competitive insurance markets, these individuals will also pay much higher premiums for their medical and life insurance, whereas those individuals with better genes will enjoy lower insurance costs. Gene testing is thus a risky prospect for some individuals. Testing may be Pareto-
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_412.html [4/21/2007 1:33:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_413.html
Page 413 deteriorating if the negative uninsurability effect is larger than the positive prevention effect. As a result individuals may decline undertaking genetic tests, even when they are offered for free. Lerman et al. (1996), cited by Schlee (1999), report that 57% of a group of 279 adult subjects with a family history of breast/ovarian cancer declined to receive for free the results of tests for mutation of a gene responsible for hereditary breast and ovarian cancer. Among the reasons listed against being tested, the most frequent, and important, reason was fear of insurance consequences should their results become public. This so-called "Hirshleifer effect" could be eliminated by organizing insurance markets ex ante. Why not insure against bad signals? In many instances there are practical reasons for that. The most important one is that some agents may get the signals earlier than others, thereby developing an adverse selection problem on these markets. In the genetic testing illustration, insurers will never know who has already learned about the quality of their genes. Obviously, controlling the flow of information in our complex societies is largely impractical. This discussion is related to a famous work by Grossman and Stiglitz (1980) who raised the following puzzle: consider an economy with identical preferences, without wealth inequality and without any idiosyncratic risk. In such an economy the autharcy is always an equilibrium, and information has no value. This is problematic if the information is costly to acquire, since no one would want to pay the cost to get valueless information! This situation is, in general, not an equilibrium: if no one on the market holds the information, it is, in general, highly valuable to acquire it. Grossman and Stiglitz (1981) present an equilibrium with asymmetric information where a subgroup of investors decides to pay for the information but is compensated by a positive value of information. In such equilibrium the positivity of the value of information is a corollary of the fact that financial markets are not informationally efficient; that is, prices do not completely reveal the information available in the economy.
26.2 Information and the Equity Premium In the previous section we reexamined the welfare effect of information in an equilibrium framework. A related problem is to examine the effect of information on prices at equilibrium. At every instant new information is revealed about the future value of firms. This knowledge is immediately translated to their prices. Information
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_413.html [4/21/2007 1:33:55 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_414.html
Page 414 about assets' future value creates price volatility. Information also improves the predictability of future prices. But information reduces the ex post uncertainty in the sense made clear in chapter 24. This should reduce the average equity premium, which is the premium to be paid to risk-averse investors to bear the social risk. This section is devoted to this question. We again consider the static model that used in the previous section. To simplify, we assume that all agents have the same utility function and the same state-dependent endowment: u(c, θ) = u(c) and ω(x, θ) = z(x) for all θ ∈ Θ. This implies that autharcy is always an equilibrium, and that information has no value. The price per unit of probability of the Arrow-Debreu security associated to state x if signal m is received is given by
We normalize these prices in such a way that it costs $1 to raise consumption by one unit in all states. The riskfree rate is zero. The ex post equity premium is thus equal to
or as in equation (5.7),
Under risk aversion we know that the equity premium is nonnegative. , is reduced by better As said above, the intuition suggests that the expected equity premium, information. Let us, for example, compare the average equity premium with complete information to the one with incomplete information. When the information is perfect, there is no aggregate risk ex post. At equilibrium the return of all assets must be equal to the return of the risk free asset. The equity premium φm is always zero. Thus the average equity premium is also zero. In contrast, in the less-than-perfect information situation, the equity premium is nonnegative for all signals, and is positive for the signals in which some uncertainty remains ex post. Thus the expected equity premium is positive. We conclude that the intuition that the average equity premium must be reduced by
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_414.html [4/21/2007 1:33:56 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_415.html
Page 415 information is confirmed when comparing two information structures, one of which is completely informative. However, this result cannot be extended to the comparison of two imperfect information structures, as we show now. We start with a simple counterexample. Let M = 2 and X = 3. Suppose that z(x) = x, for x = 1, 2, and 3. The utility function of the representative agent is such that u′(z) = 10 – z2. The reader can verify that this utility function is increasing and concave in the relevant domain of consumption. We compare two experiments. The first experiment is noninformative since pm1 = pm3 = 1/4 and pm2 = 1/2 for m = 1 and m = 2. The equity premium in this economy is equal to
The other experiment generates some information. When signal m = 2 is received from this experiment, the ex post distribution over X is degenerated at x = 2. It yields an equity premium φ2 = 0, since there is no uncertainty remaining. When signal m = 1 is received, the ex post distribution over X is such that p11 = p13 = 1/2 and p12 = 0. The reader can verify that the equity premium with this signal is equal to 2/3. The priors are the same for the two experiments if there is an equal probability to observe m = 1 or m = 2. This implies that the mean equity premium in the economy with the informative experiment is equal to 1/3 for the second experiment, which is larger than the equity premium in the uninformative experiment. By proposition 87, a better information structure would reduce the average equity premium if and only if function φ: RX → R is concave. The following lemma is useful to determine the condition under which φ is concave. LEMMA
11 Consider three real functions h1, h2, and h3, with h3 ≥ 0. Let functions ψ, V1 and V2 from
⊂ RX R be
defined as follows:
Consider any pair (p1., p2.) of vectors of probabilities in
⊂ RX. The following two conditions are equivalent:
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_415.html [4/21/2007 1:33:57 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_416.html
Page 416 1. ∀λ ∈ [0, 1] : λψ(p1) + (1 – λ)ψ(p2) ≤ ψ(λp1. + (1 – λ)p2.). 2. [V1(p1.) – V1(p2.)][V2(p1.) – V2(p2.)] ≥ 0. Proof See the appendix. The first condition in this lemma, if required for all pairs of vector of probabilities, is the definition of ψ concave. The second condition means that functionals V1 and V2 must rank probability distributions exactly in the same way. This is a very restrictive condition, as shown in our next lemma. 12 Consider function ψ as defined in equation (26.10). It is concave if and only if there exists a nonpositive scalar m such that h2(x) = mh1(x) for all x. LEMMA
Proof See the appendix. We can also easily verify that ψ is convex if and only if h1 = mh2, but with a positive m. We conclude that functionals ψ, like those defined in (26.10), are usually neither concave nor convex. The only exception is when functions h1 and h2 are proportional to each other. Notice that we already encountered this particular configuration in section 25.2. If we apply these results to the equity premium φ defined by equation (26.8), we need to use a specification with h1(x) = z(x), h2(x) = u′(z(x)) and h3(x) = z(x)u′(z(x)). By lemma 12, the equity premium is a concave function of the vector of state probabilities only if u′(z) = mz, with m ≤ 0. This is not compatible with the assumption of an increasing utility function on the relevant domain of consumption levels, except if we take m = 0. 99 Consider any strictly concave utility function. The effect of better information on the average equity premium is ambiguous in the following sense: it is always possible to find two improved information structures one of which reducing the average equity premium and the other increasing it. PROPOSITION
The ambiguity can be reduced by making additional assumptions on the information structure and on preferences. Let us consider the following possible approach: we depart from no information with an experiment with two possible signals, one of which generates perfect knowledge of the state. So let p1. be a vector of 0, except for the component corresponding to state x0, with z0 = z(x0). Let
be
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_416.html [4/21/2007 1:33:57 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_417.html
Page 417 the random variable with vector of probability p2.. Then lemma 11 states that the average equity premium is reduced by this experiment if
This is true if
for all z0 and . We can use proposition 1 to solve this problem. The necessary and sufficient condition is written as
for all z and z0, where R(z0) = –z0u″(z0)/u′(z0) is the index of relative risk aversion. The necessary condition corresponding to (6.15) is written as
Thus, when one of the two signals removes all uncertainty, a necessary condition for the information to reduce the expected equity premium is that relative prudence be larger than 2 plus twice the relative risk aversion. This condition is necessary and sufficient when the ex post uncertainty with the other signal is small. Notice that it is not satisfied for the utility function with u′(z) = 10 – z2 that we used in the counterexample, since relative prudence is negative for this u. Moreover it is not satisfied for CRRA function. In fact a necessary condition for (26.14) is that relative risk aversion be decreasing. We can use the same method to determine the effect of information on the expected risk-free rate. Using the standard two-period model of saving under uncertainty, the equilibrium risk-free rate as a function of the vector of state probabilities is written as
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_417.html (1 of 2) [4/21/2007 1:33:58 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_417.html
where z0 is the consumption at the first date. There is no doubt here that ρ is convex in the vector of probabilities. This can be seen by
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_417.html (2 of 2) [4/21/2007 1:33:58 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_418.html
Page 418 using lemma 12 with h1(x) = u′(z0), h2(x) = 1, h3(x) = u′(z(x)), and m = u′(z0). This implies that a better information structure always increases the expected risk-free rate in the economy.
26.3 Conclusion Public information affects beliefs and, as a consequence, equilibrium prices of assets. The price volatility that information generates is detrimental to risk-averse agents. On the other side, information is beneficial because it allows agents to rebalance their risk exposure according to it. Hirshleifer (1971) showed that the first negative effect can dominate the second positive effect, yielding a smaller ex ante expected utility for everyone. This is not true in general, however, since some agents may end up with a larger ex ante expected utility. Following Schlee (1999), we exhibited some conditions on preferences that makes all agents worse off by the information. We also considered the effect of information on average asset prices, and more specifically on the equity premium. A conclusive information structure yields a zero equity premium, since it eliminates the aggregate uncertainty ex post. It is not true, however, that a better information always reduces the average equity premium. We showed that there is no concave utility function that would guarantee this result. Thus we examined some restrictions on the information structure. In particular, we examined the effect of an information structure in which one of the signal is completely conclusive and the other is not. In doing so, we are in a situation to use the tools that we developed in chapter 6. We obtained that the information reduces the average equity premium for that particular information structure only if relative prudence is larger than 2 plus twice the relative risk aversion. The analysis of the effect of symmetric information on asset prices and risk sharing is a topic on which more research is needed. This chapter is merely an exploratory attempt to provide some insights on this topic.
26.4 Exercises and Extensions Exercise 122 In proposition 98 we compared the economy with information with the economy without information. Do you think
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_418.html [4/21/2007 1:33:59 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_419.html
Page 419 that we can extend this comparison to two informative experiments, one being more informative than the other in the sense of Blackwell? Exercise 123 Define the rate of return on equity R and the forward price of equity Pf as
Examine the conditions under which these economic variables are, on average, increasing with better information. Exercise 124 Consider a static fruit economy à la Lucas. All agents are identical with utility function u and an initial endowment of a fruit tree. The crop per tree is the same for all trees and is a function z(x) of the state of is subject to the information contained to a signal m. There are m = 1, . . . , M nature x. The distribution of possible signals, with probability distribution (q1, . . . , qm). Agents may trade shares of trees after observing the signal, but before the realization of the state of nature. • Determine the signal-dependent competitive price of trees in this economy. • Do people like to have information in this economy? • Examine the effect of information on the average equilibrium price of trees. Exercise 125 We consider a large economy of identical agents with utility function u. They are subject to a health risk because of the presence of a bad gene. Those with this bad gene have a probability p of manifesting the disease. Those without this bad gene will never have the disease. One could completely cure prevent the disease in the affected agents by spending k, which is less than the willingness to pay by the affected persons. The proportion of agents with the bad gene in the population is λ, which is small. We assume that there is a (public or private) market for health insurance, with a fair insurance pricing. • Suppose that there is no way to determine who has the bad gene before one manifests symptoms of the disease. What is the
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_419.html [4/21/2007 1:33:59 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_420.html
Page 420 equilibrium price of health insurance? Who buys that insurance, and how much of it? What is the level of welfare in the economy at equilibrium? • Suppose that geneticists find a cost-free way to test for the presence of the bad gene. Answer to the following questions: a. How will the insurance market react to this innovation? Will insurers discriminate agents depending on the results of the test? Who will purchase insurance, and at which price? b. Show why, ex ante, everyone prefers to vote against geneticists doing this research. c. How is this result affected by the existence of a preventive treatment that costs q < k per person and eliminates the health risk for any one who undergoes the treatment?
26.5 Appendix Proof of Lemma 11 Let aij denote
. Condition 1 can then be rewritten as
Because h3 is positive, so are the
, j = 1, 2. After some manipulations, this inequality can then be rewritten as
The inequality above holds if and only if
This is equivalent to condition 2. Proof of Lemma 12 file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_420.html (1 of 2) [4/21/2007 1:34:00 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_420.html
From lemma 11, we must show that functionals V1 and V2 rank all pairs of probability distributions in the same way. By the standard continuity argument, a necessary condition is
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_420.html (2 of 2) [4/21/2007 1:34:00 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_421.html
Page 421
This is equivalent to
Equation (26.21) is true if
or equivalently if
Since (26.23) must be true for any pair (p1., p2.), it can be true only if
is
independent of vector p.. Let m denote this constant. Thus we have that
for all vectors p.. This is equivalent to require that h1 = mh2. The nonpositivity of m comes directly from inequality (26.19). This proves the necessity part of the lemma. The proof of sufficiency is immediate from condition 2 in lemma 11.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_421.html [4/21/2007 1:34:00 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_422.html
Page 422
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_422.html [4/21/2007 1:34:01 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_423.html
Page 423
27 Epilogue. In this book we have attempted to understand how uncertainty affects welfare. More important we have explored how to determine the optimal actions in the face of uncertainty. Now that we have developed the tools and the analyses of the various underlying questions, we can assess whether the classical economics of uncertainty has achieved these goals.
27.1 The Important Open Questions Despite the great progress observed over the past half-century, we must admit that a lot of work remains to be done before any consensus can be achieved on a number of issues. No doubt, most of the concepts necessary to solve most problems of risk tolerance, prudence, stochastic dominance, informativeness of information structures, and the like, have been discovered. We now need to digest these concepts. Digestion means structuring the theory, testing it, and manipulating and interrelating the tools and results. Organizing this knowledge was a central goal of this book. Let us examine a few questions of theory for which there is a continuing debate. 27.1.1 The Independence Axiom All throughout this book we took the position that decision makers are expected utility maximizers; we showed that they satisfy the independence axiom. We justified this approach by following two different lines of defense. First, those agents that would not satisfy this assumption would face the time inconsistency of their decisions. As a consequence they would be subject to the acceptance of
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_423.html [4/21/2007 1:34:01 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_424.html
Page 424 "Dutch books," something that could threaten their life. Moreover we showed that they could not depend on Bayes's rule to update their beliefs over time. In short, we showed that it is not clear how nonexpected-utility agents would behave and survive in the face of a dynamic uncertainty. This challenging topic is currently under much discussion among my very competent colleagues specialized in this field. A second argument in favor of the expected utility model is that the linearity of expected utility with respect to the vector of probabilities allows for the use of standard hyperplane separation theorems that make life much easier for research. This leads often to clear-cut testable results, not something that we can expect in nonexpectedutility models. This does not mean that for some decision makers, and for some specific circumstances, the predictive power of the classical model is low, as illustrated by the famous Allais's paradox. This calls for pursuing our efforts to develop alternative models that will work better in these circumstances, and for these agents. I fear that this alternative model will be even more complex than what we obtained in this book. The problem is thus to find a good compromise between realism and tractability, something not easy to do. In any case, the expected utility model will serve as a simple, clear-cut benchmark for a long time. One of the critical aspect of the expected utility model is the analogy of the static portfolio problem and of the consumption-saving problem under certainty. This is because that these two problems are equivalent that it is so easy to mix risk and time in the classical model with complete markets. But this can also be seen as a disadvantage, since the concavity of the utility function plays two roles a priori different: it measures simultaneously the aversion to consumption fluctuations across states and over time. The recursive utility model of Kreps-Porteus-Selten is a clever generalization that overcomes this problem without giving up the substance of the expected utility model. 27.1.2 Measures of Risk Aversion It is quite surprising and disappointing for me that almost 40 years after the establishment of the concept of risk aversion by Pratt and Arrow, our profession has not been able to attain a consensus about
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_424.html [4/21/2007 1:34:02 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_425.html
Page 425 the measurement of risk aversion. Without such a consensus, there is no hope to quantify optimal portfolios, efficient public risk prevention policies, optimal insurance deductibles, and so on. It is vital that we put more effort on research aimed at refining our knowledge about risk aversion. For unclear reasons, this line of research is not in fashion these days, and it is a shame. Traditionally we can estimate risk aversion using experimental studies or market data. On one hand, existing experimental studies are not convincing, mainly because of the low level of payoffs and the unrealistic environment of the experiments. On the other hand, financial market data lead to the famous equity premium puzzle. Much of this book is devoted to describing the problem and the ways to solve it (background risk, aggregation of preferences, liquidity constraints, asymmetric information, etc.). But the difficulty is that the real world is too complex to be summarized in a testable model. The challenge is more to improve the theoretical ground rather than the econometrics. A similar problem can be raised for the measurement of the degree of prudence. But we have here the excuse that this concept is much more recent. 27.1.3 Qualitative Properties of the Utility Function Most comparative static properties of decision under uncertainty and equilibrium prices rely on fine conditions on the utility function of the decision makers. The most common problem, decreasing absolute risk aversion, benefits from a clear consensus. For the remaining problems, everything remains to be clarified and tested. The list of useful conditions is long, ranging from decreasing relative risk aversion to decreasing prudence, to convex absolute risk aversion, and to concave absolute risk tolerance. Without these assumptions we would not be able to determine the effect of time horizon, wealth inequality, and background risk on the equity premium and the risk-free rate, to cite just a few comparative static exercises. In the face of such a complex nexus of theoretical results, there is a clear tendency to make simplifying assumptions. Exactly as the macroeconomic theory is much simplified by the assumption that firms have CobbDouglas production functions, researchers in finance used to consider a very specific set of utility functions, namely functions exhibiting constant relative risk aversion (CRRA). This
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_425.html [4/21/2007 1:34:02 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_426.html
Page 426 well-established tradition is useful, as we know that it makes the dynamic portfolio problem with i.i.d portfolio risks over time equivalent to a static decision problem. This is because absolute risk tolerance is linear in wealth for CRRA utility function. CRRA also implies that relative prudence is equal to relative risk aversion plus one. The likelihood of such restrictions to be satisfied on the entire range of potential final wealth is actually zero. If risk aversion is genetic, there is just no theoretical (evolutionary) argument for genetic risk aversion to be constant. The absence of any empirical test of this restriction is another surprise for those external evaluators of the theory of finance observing this tradition.11 It is still possible that CRRA be a good approximation for some applications. 27.1.4 Economics of Uncertainty and Psychology A recent trend in economic theory is to recognize that agents may extract utility not only from their own current consumption but also from other elements of their environment. Below is a list of interesting generalizations of the standard egocentric model that we developed in this book: • Agents may be positively or negatively affected by the consumption level of their neighborhood. If the marginal utility of one's own consumption is increasing with the consumption level of others, the assumption is referred to as "keeping up with the Joneses," "envy," or "conformism." Abel (1990) is the first to examine an economy where investors are conformists. The standard approach consists in determining a Nash equilibrium on financial markets where each investor determine his/her optimal portfolio, assuming the portfolio selected by others. • The current felicity of an agent may be a negative function of his past consumption. This class of models corresponds to the notion of "habit formation." It is natural to think that an individual who consumes a lot today will get used to a large consumption level and will meet a higher consumption level tomorrow. Constantinides (1990) examines the optimal saving and consumption of agents having this psychological characteristics.
11. A recent paper by Guiso and Paiella (2000) provides such an empirical test. It strongly rejects the hypothesis that relative risk aversion is constant with respect to wealth.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_426.html [4/21/2007 1:34:03 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_427.html
Page 427 • It is often observed that consumers use a smaller discount rate for discounting felicity from period t + 1 to period t than for discounting felicity next period as of today. This leads to a form of time inconsistency in the selection of consumption plans under certainty. It is worthwhile to examine how such an agent would manage a portfolio of risky assets. The tools that we presented and developed in this book can be used to examine these generalized versions of the classical decision problem.
27.2 Conclusion The economics of uncertainty has been slow to progress since the 1960s and 1970s after the initial impetus by Arrow and Pratt. The models under scrutiny during that era were most of the time static with a single risk. Considerable progress was made in the 1980s by examining models with more than one risk. Multiple risks, static or dynamic, enriched the model considerably. Several new concepts and hypotheses were developed to make more progress possible. Now is a good time to try to summarize and to unify all the new knowledge in one book. But this does not mean that we have attained the ultimate theory. Far from that I believe that this book calls for another round of theoretical and empirical research.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_427.html [4/21/2007 1:34:03 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_428.html
Page 428
This page intentionally left blank.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_428.html [4/21/2007 1:34:04 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_429.html
Page 429
Bibliography. Abel, A. B. 1990. Asset prices under habit formation and catching up with the Joneses. American Economic Review Papers and Proceedings 80: 38–42. Allais, M. 1953. Le comportement de l'homme rationnel devant le risque, Critique des postulats et axiomes de l'école américaine. Econometrica 21: 503–46. Arrow, K. J. 1953. Le rôle des valeurs boursières pour la répartition la meilleure des risques. Econométrie, Paris: CNRS. (Translated as: Arrow, K. J. 1964. The role of securities in the optimal allocation of risk-bearing. Review of Economic Studies 31: 91–96.) Arrow, K. J. 1963. Liquidity preference. Lecture VI in Lecture Notes for Economics 285: The Economics of Uncertainty. Stanford University, pp. 33–53. Arrow, K. J. 1965. Yrjo Jahnsson Lecture Notes. Helsinki. Reprinted in Arrow (1971). Arrow, K. J. 1971. Essays in the Theory of Risk Bearing. Chicago: Markham Publishing. Arrow, K. J. 1994. Discounting climate change: Planning for an uncertain future. IDEI lecture. University of Toulouse. Arrow, K. J., and A. C. Fischer. 1974. Environmental preservation, uncertainty and irreversibility. Quarterly Journal of Economics 88: 312–19. Arrow, K. J., and R. C. Lind. 1970. Uncertainty and the evaluation of public investment decision. American Economic Review 60: 364–78. Athey, S. 1997. Comparative statics under uncertainty: Single crossing properties and log-supermodularity. Discussion paper. MIT. Athey, S., and J. Levin. 1998. The value of information in monotone decision problems. Mimeo. MIT. Battalio, R. C., J. H. Kagel, and D. N. MacDonald. 1985. Animals' choices over uncertain outcomes. American Economic Review 75: 597–613. Barberis, N. 2000. Investing for the long run when returns are predictable. Journal of Finance 55: 224–64.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_429.html (1 of 2) [4/21/2007 1:34:04 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_429.html
Barsky, R. B., F. T. Juster, M. S. Kimball, and M. Shapiro. 1997. Preference parameters and behavioral heterogeneity: An experimental approach in the health and retirement study. Quarterly Journal of Economics 112: 537–79.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_429.html (2 of 2) [4/21/2007 1:34:04 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_430.html
Page 430 Baxter, M., and U. J. Jermann. 1997. The international diversification puzzle is worse than you think. American economic Review 87: 170–80. Bell, D. E. 1988a. The value of pre-decision side bets for utility maximizers. Management Science 34: 797–800. Bell, D. E. 1988b. One-switch utility functions and a measure of risk. Management Science 34: 1416–24. Bernstein, P. L. 1996. Against the Gods: A Remarkable Story of Risk. Wiley, New York. Bewley, T. 1977. The permanent income hypothesis: A theoretical formulation. Journal of Economic Theory 16: 252–92. Blackorby, C., D. Primont, and R. R. Russell. 1999. Separability: A survey. In S. Barbera, P. J. Hammond, and C. Seidl, eds., Handbook of Utility Theory: Principles. vol. 1, Boston: Kluwer Academic. Blackwell, D. 1951. Comparison of Experiments. In J. Neyman, ed., Proceedings of the Second Berkeley Symposium on Mathematical Statistics and Probability. Berkeley: University of California Press, pp. 93–102. Bodie, Z., R. C. Merton, and W. F. Samuelson. 1992. Labor supply flexibility and portfolio choice in a life cycle model. Journal of Economic Dynamics and Control 16: 427–49. Bohnenblust, H. F., L. S. Shapley, and S. Sherman. 1949. Reconnaissance in game theory. The Rand Corporation, Santa Monica, CA. Boll, C. H. 1955. Comparison of experiments in the infinite case and the use of invariance in establishing sufficiency. Ph. D. thesis. Stanford University. Brennon, M. J. 1998. The role of learning in dynamic portfolio decisions. European Finance Review 1: 295–306. Brennan, M. J., E. S. Schwartz, and R. Lagnado. 1997. Strategic asset allocation Journal of Economic Dynamic and Control 21: 1377–1403. Brennan, M. J., and Y. Xia. 1999. Assessing asset pricing anomalies. Discussion paper 19–99, Anderson School at UCLA. Brown, R. H., and S. M. Schaefer. 1997. Why do long term forward rates (almost) always slope downwards? Discussion Paper. London Business School. Caballe, J., and A. Pomansky. 1996. Mixed risk aversion. Journal of Economic Theory 71: 485–514.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_430.html (1 of 2) [4/21/2007 1:34:05 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_430.html
Campbell, J. 1987. Stock returns and the term structure. Journal of Financial Economics 18: 373–99. Campbell, J. 1991. A variance decomposition of stock returns. Economic Journal 101: 157–79. Campbell, J., and J. Cochrane. 1995. By force of habit: A consumption-based explanation of aggregate stock market behavior. NBER Working Paper 4995. Cambridge, MA. Campbell, J., and N. G. Mankiw. 1990. Permanent income, current income, and consumption. Journal of Business and Economic Statistics 8: 265–78.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_430.html (2 of 2) [4/21/2007 1:34:05 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_431.html
Page 431 Campbell, J., and R. Shiller. 1988a. The dividend-price ratio and expectations of future dividends and discount factors. Review of Financial Studies 1: 195–227. Campbell, J., and R. Shiller. 1988b. Stock prices, earnings, and expected dividends. Journal of Finance 43: 661– 76. Carroll, C. D. 1997. Buffer-stock saving and the life cycle/permanent income hypothesis. Quarterly Journal of Economics 112: 1–55. Caroll, C. D., and M. Kimball. 1996. On the concavity of the consumption function. Econometrica 64: 981–92. Cass, D., and J. Stiglitz. 1970. The structure of investor preferences and asset returns, and separability in portfolio allocation. Journal of Economic Theory 2: 122–60. Castagnoli, E., and M. Li Calzi. 1996. Expected utility without utility. Theory and Decision 41: 281–301. Clyman, D. R. 1995. Unreasonable rationality? Management Science 41: 1538–48. Cocco, J. F., F. J. Gomes, and P. J. Maenhout. 1998. Consumption and portfolio choice over the life-cycle. Discussion paper. Harvard University. Constantinides, G. M. 1982. Intertemporal asset pricing with heterogeneous consumers and without demand aggregation. Journal of Business 55: 253–67. Constantinides, G. M. 1990. Habit formation: A resolution of the equity premium puzzle. Journal of Political Economy 98: 519–43. Cohen, M. 1995. Risk-aversion concepts in expected- and non-expected-utility models. Geneva Papers on Risk and Insurance Theory 20: 73–91. Cox, J. and H. Leland. 1982. Notes on dynamic investment strategies. Proceedings of Seminar on the Analysis of Security Prices, Center for Research in Security Prices (CRSP). Graduate School of Business, University of Chicago. Cox, J., J. Ingersoll, and S. Ross. 1985a. A theory of the term structure of interest rates. Econometrica 53: 385– 403. Cox, J., J. Ingersoll, and S. Ross. 1985b. An intertemporal general equilibrium model of asset prices. Econometrica 53: 363–84.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_431.html (1 of 2) [4/21/2007 1:34:06 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_431.html
Cremer, J. 1982. A simple proof of Blackwell's "comparison of experiments" theorem. Journal of Economic Theory 27: 439–43. Deaton, A. 1991. Saving and liquidity constraints. Econometrica 59: 1221–48. Debreu, G. 1959. Topological methods in cardinal utility theory. In K. J. Arrow, S. Karlin, and P. Suppes, eds. 1959. Mathematical Methods in the Social Sciences. Stanford: Stanford University Press, pp. 16–26. De Finetti, B. 1952. Sulla preferibilita. Giornale degli Economisti e Annali di Economia 11: 685–709. Detemple, J. B. 1986. Asset pricing in an economy with incomplete information. Journal of Finance 61: 383–92. Diamond, P. A., and J. E. Stiglitz. 1974. Increases in risk and in risk aversion. Journal of Economic Theory 8: 337–60.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_431.html (2 of 2) [4/21/2007 1:34:06 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_432.html
Page 432 Dixit, A. K., and R. S. Pindyck. 1994. Investment under Uncertainty. Princeton: Princeton University Press. Drèze, J. H. 1987. Essays on Economic Decisions under Uncertainty. Cambridge: Cambridge University Press. Drèze, J. H., and F. Modigliani. 1972. Consumption decisions under uncertainty. Journal of Economic Theory 5: 308–35. Dybvig, P. H. 1988. Inefficient dynamic portfolio strategies or How to throw away a million dollars in the stock market? Review of Financial Studies 1: 67–88. Eeckhoudt, L., and C. Gollier. 1994. Risk: Evaluation, Management and Sharing. Oxford: Hearvester Wheatsheaf. Eeckhoudt, L., and C. Gollier. 1995. Demand for risky assets and the monotone probability ratio order. Journal of Risk and Uncertainty 11: 113–22. Eeckhoudt, L., and C. Gollier. 1998. Which shape for the cost curve of risk? Mimeo. University of Toulouse. Eeckhoudt, L., and C. Gollier. 1999. Are independent risks substitutes? Mimeo. University of Toulouse. Eeckhoudt, L., C. Gollier, and H. Schlesinger. 1995. The risk-averse (and prudent) newsboy. Management Science 41: 786–94. Eeckhoudt, L., C. Gollier, and H. Schlesinger. 1996. Changes in background risk, and risk taking behaviour. Econometrica 64: 683–90. Ellsberg, D. 1961. Risk, ambiguity, and the Savage axioms. Quarterly Journal of Economics 75: 643–69. Epstein, L. S. 1980. Decision-making and the temporal resolution of uncertainty. International Economic Review 21: 269–84. Epstein, L. G. 1983. Decreasing absolute risk aversion and utility indices derived from cake-eating problems. Journal of Economic Theory 29: 245–64. Epstein, L. G., and S. Zin. 1991. Substitution, risk aversion and the temporal behavior of consumption and asset returns: An empirical framework. Journal of Political Economy 99: 263–86. Fama, E., and K. French. 1988. Dividend yields and expected stock returns. Journal of Financial Economics 22: 3–27.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_432.html (1 of 2) [4/21/2007 1:34:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_432.html
Fama, E., and K. French. 1989. Business conditions and expected returns on stocks and bonds. Journal of Financial Economics 25: 23–49. Fishburn, P., and B. Porter. 1976. Optimal portfolios with one safe and one risky asset: Effects of changes in rate of return and risk. Management Science 22: 1064–73. Franke, G., R. C. Stapleton, and M. G. Subrahmanyam. 1998. Who buys and who sells options: The role of options in a n economy with background risk. Journal of Economic Theory 82: 89–109. Freidman, M. 1957. A theory of the consumption function. Princeton: Princeton University Press.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_432.html (2 of 2) [4/21/2007 1:34:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_433.html
Page 433 Freixas, X., and R. Kihlstrom. 1984. Risk aversion and information demand. In M. Boyer and R. Kihlstrom, eds., Bayesian Models in Economic Theory. Dordrecht: Elsevier. Freixas, X. and J.-J. Laffont. 1984. On the irreversibility effect. In M. Boyer and R. Kihlstrom, eds., Bayesian Models in Economic Theory. Dordrecht: Elsevier, ch. 7. French, K., and J. Poterba. 1991. International diversification and international equity markets. American Economic Review 81: 222–26. Gennotte, G. 1986. Optimal portfolio choice under incomplete information. Journal of Finance 61: 733–49. Gollier, C. 1995. The comparative statics of changes in risk revisited. Journal of Economic Theory 66: 522–36. Gollier, C. 1996a. Repeated optional gambles and risk aversion. Management Science 42: 1524–30. Gollier, C. 1996b. Decreasing absolute prudence: Characterization and applications to second-best risk sharing. European Economic Review 40: 1799–1815. Gollier, C. 1997. On the inefficiency of bang-bang and stop-loss portfolio strategies. Journal of Risk and Uncertainty 14: 143–54. Gollier, C. 1999. Time horizon and the discount rate. Mimeo. University of Toulouse. Gollier, C. 2000. Wealth inequality and asset prices. Review of Economic Studies, forthcoming. Gollier, C. 2001. Optimal portfolio risk with first-order and second-order predictability. Discussion paper, University of Toulouse. Gollier, C., B. Jullien, and N. Treich. 2000. Scientific progress and irreversibility: An economic interpretation of the "Precautionary Principle." Journal of Public Economics 75: 229–53. Gollier, C., and M. S. Kimball. 1996. Toward a systematic approach to the economic effects of uncertainty: Characterizing utility functions. Discussion paper. University of Michigan. Gollier, C., and M. S. Kimball. 1997. New methods in the classical economics of uncertainty: Comparing risks. Discussion paper. University of Michigan. Gollier, C., P.-F. Koehl, and J.-C. Rochet. 1997. Risk-taking behaviour with expected utility and limited liability: Applications to the regulation of financial intermediaries. Journal of Risk and Insurance 64: 347–70.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_433.html (1 of 2) [4/21/2007 1:34:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_433.html
Gollier, C., J. Lindsey, and R. J. Zeckhauser. 1997. Investment flexibility and the acceptance of risk. Journal of Economic Theory 76: 219–42. Gollier, C. and J. W. Pratt. 1996. Risk vulnerability and the tempering effect of background risk. Econometrica 64: 1109–24. Gollier, C., and H. Schlesinger. 1996. Portfolio choice under noisy asset returns. Economics Letters 53: 47–51. Gollier, C., and R. J. Zeckhauser. 1998. Horizon length and portfolio risk. Discussion paper. University of Toulouse.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_433.html (2 of 2) [4/21/2007 1:34:07 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_434.html
Page 434 Gorman, W. M. 1953. Community preference fields. Econometrica 21: 63–80. Gorman, W. M. 1968. The structure of the utility function. Review of Economic Studies 35: 367–90. Green, J. 1981. The value of information with sequential future markets. Econometrica 49: 335–58. Green, J. 1987. Making book against oneself, the independence axiom, and nonlinear utility theory. Quarterly Journal of Economics 98: 785–96. Grossman, S., and J. Stiglitz. 1980. On the impossibility of informationally efficient markets. American Economic Review 70: 393–408. Grossman, S., and J. L. Vila. 1992. Optimal investment strategies with leverage constraints. Journal of Financial and Quantitative Analysis 27: 151–68. Guiso, L., T. Jappelli, and D. Terlizzese. 1996. Income risk, borrowing constraints, and portfolio choice. American Economic Review 86: 158–72. Guiso, L., and M. Paiella. 2000. Risk aversion, wealth and financial markets. Discussion paper. University of Sassari. Hadar, J., and W. Russell. 1969. Rules for ordering uncertain prospects. American Economic Review 59: 25–34. Hadar, J., and T. K. Seo. 1990. The effects of shifts in a return distribution on optimal portfolios. International Economic Review 31: 721–36. Hall, R. E. 1988. Intertemporal substitution of consumption. Journal of Political Economy 96: 221–73. Hankansson, N., J. Kunkel, and J. Ohlson. 1982. Sufficient and necessary conditions for informative to have social value in pure exchange. Journal of Finance 38: 1169–81. Hansen, L., and K. Singleton. 1983. Stochastic consumption, risk aversion and the temporal behavior of assets returns. Journal of Political Economy 91: 249–68. Hardy, G. H., J. E. Littlewood, and G. Pólya. 1929. Some simple inequalities satisfied by convex functions. Messenger of Mathematics 58: 145–52. Hardy, G. H., J. E. Littlewood, and G. Pólya. 1934. Inequalities. Reprinted in 1997 by Cambridge University Press.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_434.html (1 of 2) [4/21/2007 1:34:08 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_434.html
Harris, C., and D. Laibson. 1999. Dynamic choices and hyperbolic consumers. Mimeo. Harvard University. He, H., and C.-F. Huang. 1994. Consumption-portfolio policies: An inverse optimal problem. Journal of Economic Theory 62: 257–93. Hennessy, D. A., and H. E. Lapan. 1998. On the nature of certainty equivalent functionals. Discussion paper. Iowa State University. Henry, C. 1974. Investment decisions under uncertainty: The irreversibility effect. American Economic Review 64: 1006–12. Hicks, J. R. 1931. The theory of uncertainty and profit. Economica 11: 170–89. Hirshleifer, J. 1971. The private and social value of information and the reward to inventive activity. American Economic Review 61: 561–74.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_434.html (2 of 2) [4/21/2007 1:34:08 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_435.html
Page 435 Hirshleifer, J., and J. G. Riley. 1992. The Analytics of Uncertainty and Information. Cambridge: Cambridge University Press. Hubbard, R. G., J. Skinner, and S. P. Zeldes. 1994. The importance of precautionary motives for explaining individual and aggregate saving. In A. H. Meltzer and C. I. Plosser, eds, The Carnegie-Rochester Conference Series on Public Policy 40. Amsterdam: North-Holland. Huberman, G., and S. Ross. 1983. Portfolio turnpike theorems, risk aversion, and regularly varying utility functions. Econometrica 51: 1345–61. Jagannathan, R., and N. R. Kocherlakota. 1996. Why should older people invest less in stocks than younger people? Federal Reserve Bank of Minneapolis Quarterly Review 20: 11–23. Jewitt, I. 1986. A note on comparative statics and stochastic dominance. Journal of Mathematical Economics 15: 249–54. Jewitt, I. 1987. Risk aversion and the choice between risky prospects: The preservation of comparative statics results. Review of Economic Studies 54: 73–85. Jewitt, I. 1988. Risk and risk aversion in the two risky asset portfolio problem. Mimeo. University of Bristol. Jewitt, I. 1989. Choosing between risky prospects: the characterization of comparative statics results, and location independent risk. Management Science 35: 60–70. Jones, J. M., and R. A. Ostroy. 1984. Flexibility and uncertainty. Review of Economic Studies 51: 13–32. Kandel, S., and R. F. Stambaugh. 1991. Assets returns and intertemporal preferences. Journal of Monetary Economics 27: 39–71. Karlin, S. 1968. Total Positivity. Stanford: Stanford University Press. Karlin, S., and Y. Rinott. 1980. Classes of orderings of measures and related correlation inequalities. I. Multivariate totally positive distributions, Journal of Multivariate Analysis 10: 467–98. Kessler, D., and E. N. Wolff. 1991. A comparative analysis of household wealth patterns in France and in the United States. Review of Income and Wealth 37: 249–66. Keynes, J. M. 1936. The General Theory of Employment, Interest and Money. London: Macmillan. Kihlstrom, R. E., A Bayesian exposition of Blackwell's theorem on the comparison of experiments. In M. Boyer file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_435.html (1 of 2) [4/21/2007 1:34:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_435.html
and R. E. Kihlstrom, eds., Bayesian Models of Economic Theory. Amsterdam: Elsevier. Kihlstrom, R., D. Romer, and S. Williams. 1981. Risk aversion with random initial wealth. Econometrica 49: 911–20. Kimball, M. S. 1990. Precautionary savings in the small and in the large. Econometrica 58: 53–73. Kimball, M. S. 1993. Standard risk aversion. Econometrica 61: 589–611. Kimball, M. S., and P. Weil. 1992. Precautionary saving and consumption smooting across time and possibilities. NBER Working Paper 3976. Cambridge, MA.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_435.html (2 of 2) [4/21/2007 1:34:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_436.html
Page 436 King, M., and J. Leape. 1987. Asset accumulation, information, and the life-cycle. NBER Working Paper 2392. Cambridge, MA. Kocherlakota, N. R. 1996. The equity premium: It's still a puzzle. Journal of Economic Literature 34: 42–71. Kreps, D. M., and E. L. Porteus. 1978. Temporal resolution of uncertainty and dynamic choice theory. Econometrica 46: 185–200. Laibson, D. I. 1997. Golden eggs and hyperbolic discounting. Quarterly Journal of Economics 62: 443–79. Landsberger, M., and I. Meilijson. 1990. Demand for risky financial assets: A portfolio analysis. Journal of Economic Theory 50: 204–13. La Vallée, I. 1968. On cash equivalence and information evaluation in decisions under uncertainty, Part I: Basic theory. Journal of the American Statistical Association 63: 252–75. Lehman, E. L. 1988. Comparing location experiments. Annals of Statistics 16: 521–33. Lerman, C., et al. 1996. BRCA1 testing in families with hereditary breast-ovarian cancer. Journal of the American Medical Association 275: 1885–92. Leland, H. E. 1968. Dissertation. Stanford University. Leland, H. E. 1980. Who should buy portfolio insurance? Journal of Finance 35: 581–96. Longstaff, F. A., and E. S. Schwartz 1990. Interest rate volatility and the term structure: A two-factor general equilibrium model. Journal of Finance 47: 1259–82. Lucas, R. 1978. Asset prices in an exchange economy. Econometrica 46: 1429–46. Lucas, R. 1987. Models of Business Cycle. Oxford: Basil Blackwell. Lucas, D. J. 1994. Asset pricing with undiversifiable income risk and short sales constraints: Deepening the equity premium puzzle. Journal of Monetary Economics 34: 325–41. Lusardi, A. 1992. Permanent income, current income, and consumption: Evidence from panel data. Manuscript. Dartmouth College. Machina, M. 1982. Expected utility analysis without the independence axiom. Econometrica 50: 277–323.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_436.html (1 of 2) [4/21/2007 1:34:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_436.html
Machina, M. 1987. Choice under uncertainty: Problems solved and unsolved. Journal of Economic Perspectives 1: 281–96. Machina, M. J. 1995. Non-expected utility and the robustness of the classical insurance paradigm. Geneva Papers on Risk and Insurance Theory 20: 9–50. Marschak, J. 1938. Money and the theory of assets. Econometrica 6: 311–25. Marschak, J. 1950. Rational behaviour, uncertain prospects, and measurable utility. Econometrica 18: 111–41. Marschak, J., and K. Miyasawa. 1968. Economic comparability of information systems. International Economic Review 9: 137–74.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_436.html (2 of 2) [4/21/2007 1:34:09 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_437.html
Page 437 Marshall, J. 1974. Private incentives and public information. American Economic Review 64: 373–90. Mankiw, N. G., and S. Zeldes. 1991. The consumption of stockholders and nonstockholders. Journal of Financial Economics 29: 97–112. Mas-Colell, A. 1991. On the uniqueness of equilibrium once again. In W. Barnett, B. Cornet, C. d'Aspremont, J. Gabszewicz, and A. Mas-Colell, eds., Equilibrium Theory and Applications. Cambridge: Cambridge University Press, ch. 12. McCardle, K. F., and R. L. Winkler. 1992. Repeated gambles, learning, and risk aversion. Management Science 38: 807–18. McFadden, D. 1989. Testing for stochastic dominance. In T. B. Fomby and T. K. Seo, eds., Studies in the Economics of Uncertainty. pp. 113–34. McDonald, R., and D. Siegel. 1986. The value of waiting to invest. Quarterly Journal of Economics 101: 707–28. Mehra, R. and E. Prescott. 1985. The equity premium: A puzzle. Journal of Monetary Economics 10: 335–39. Merton, R. C. 1969. Lifetime portfolio selection under uncertainty: The continuous-time case. Review of Economics and Statistics 51: 247–57. Merton, R. C. 1971. Optimum consumption and portfolio rules in a continuous-time model. Journal of Economic Theory 3: 373–413. Merton, R. C. 1973. An intertemporal capital asset pricing model. Econometrica 41: 867–87. Meyer, J. 1977. Choice among distributions. Journal of Economic Theory 14: 326–36. Meyer, J., and M. Ormiston. 1985. Strong increases in risk and their comparative statics. International Economic Review 26: 425–37. Milgrom, P. 1981. Good news and bad news: Representation theorems and applications. Bell Journal of Economics 12: 380–91. Milgrom, P., and R. Weber. 1982. A theory of auctions and competitive bidding. Econometrica 50: 1089–22. Mitchell, D. W. 1994. Relative risk aversion with Arrow-Debreu securities. International Economic Review 35: 257–58.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_437.html (1 of 2) [4/21/2007 1:34:10 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_437.html
Mitiushin, L. G., and W. M. Polterovich. 1978. Criteria for monotonicity of demand functions (in Russian). Ekonomika i Mathematichskie Metody 14: 122–28. Modigliani, A., and R. Brumberg. 1954. Utility analysis and the consumption function: An interpretation of the cross-section data. In K. Kurihara, ed., Post-Keynesian Economics. New Brunswick, NJ: Rutgers University Press. Mossin, J. 1968a. Aspects of rational insurance purchasing. Journal of Political Economy 76: 533–68. Mossin, J. 1968b. Optimal multiperiod portfolio policies. Journal of Business 42: 215–29. Nachman, D. C. 1982. Preservation of "more risk averse" under expectations. Journal of Economic Theory 28: 361–68.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_437.html (2 of 2) [4/21/2007 1:34:10 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_438.html
Page 438 Ogaki, M., and Q. Zhang. 1999. Decreasing relative risk aversion and tests of risk sharing. Mimeo. Ohio State University. Ormiston, M. B., and E. E. Schlee. 1992. First and second order transformations and comparative statics under uncertainty. International Economic Review 33: 33–44. Ormiston, M., and E. Schlee. 1993. Comparative statics under uncertainty for a class of economic agents. Journal of Economic Theory 61: 412–22. Persico, N. 1998. Information acquisition in affiliated decision problems. Mimeo. UCLA. Pindyck, R. 1991. Irreversibility, uncertainty and investment. Journal of Economic Literature 29: 1110–48. Ponssard, J. P. 1975. A note on information value theory for experiments defined in extensive form. Management Sciences 21: 22. Polak, B. Notes on Anand and Sen's Notes on equally distributed equivalent incomes. Mimeo, Yale University. Pratt, J. 1964. Risk aversion in the small and in the large. Econometrica 32: 122–36. Pratt, J. W. 1987. Multiattribute utility and derived utility. In Y. Sawaragi, K. Inoue, and H. Nakayama, eds., Towards Interactive and Intelligent Decision Support Systems. Proceedings of the Seventh International Conference on Multiple Criteria Decision Making, Kyoto, pp. 149–54. Pratt, J. W., and R. Zeckhauser. 1987. Proper risk aversion. Econometrica 55: 143–54. Pratt, J. W., and R. Zeckhauser. 1989. The impact of risk sharing on efficient decision. Journal of Risk and Uncertainty. 2: 219–34. Quiggin, J. 1982. A theory of anticipated utility. Journal of Economic Behavior and Organization 3: 323–43. Ramsey, F. P. 1931. Truth and probability, In The Foundations of Mathematics and Other Essays. London: Routledge and Kegan Paul, pp. 156–98. Raviv, A. 1979. The design of an optimal insurance policy. American Economic Review 69: 84–96. Rietz, T. 1988. The equity premium puzzle: A solution? Monetary Economics 21: 117–32. Ross, S. A. 1981. Some stronger measures of risk aversion in the small and in the large with applications. Econometrica 3: 621–38. file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_438.html (1 of 2) [4/21/2007 1:34:11 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_438.html
Ross, S. A. 1999. Adding risks: Samueslon's fallacy of large numbers revisited. Journal of Financial of Quantitative Analysis 34: 323–40. Ross, S. A. 1999. Some notes on compensation, agency theory, and the duality of risk aversion and riskiness. Discussion paper. MIT. Rothschild, M., and J. Stiglitz. 1970. Increasing risk: I. A definition. Journal of Economic Theory 2: 225–43. Rothschild, M., and J. Stiglitz. 1971. Increasing risk: II Its economic consequences. Journal of Economic Theory 3: 66–84.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_438.html (2 of 2) [4/21/2007 1:34:11 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_439.html
Page 439 Roy, S., and R. Wagenvoort. 1996. Risk preference and indirect utility function in portfolio-choice problems. Journal of Economics 63: 139–50. Salanié, B. 1997. The Economics of Contracts: A Primer. Cambridge: MIT Press. Samuelson, P. A. 1963. Risk and uncertainty: the fallacy of the Law of Large Numbers. Scientia 98: 108–13. Samuelson, P. A. 1969. Lifetime portfolio selection by dynamic stochastic programming. Review of Economics and Statistics 51: 239–46. Samuelson, P. A. 1989a. The judgement of economic science on rationale portfolio management: Indexing, timing, and long-horizon effects. Journal of Portfolio Management (Fall). 3–12. Samuelson, P. A. 1989b. A case at last for age-phased reduction in equity. Proceedings of the National Academy of Sciences 86: 9048–51. Sandmo, A. 1971. On the theory of the competitive firm under price uncertainty. American Economic Review 61: 65–73. Savage, L. J. 1954. The Foundations of Statistics. New York: Wiley. Revised and enlarged ed., New York: Dover. 1972. Schechtman, J. 1976. An income fluctuation problem. Journal of Economic Theory 12: 218–41. Schlee, E. E. 1999. The value of information in efficient risk sharing arrangements. Mimeo. Arizona State University. Schubert, R., M. Brown, M. Gysler, and H. W. Brachinger. 1999. Financial decision-making: Are women really more risk-averse? American Economic Review 89: 381–85. Segal, U., and A. Spivak. 1990. First order versus second order risk aversion. Journal of Economic Theory 51: 111–25. Selden, L. 1978. A new representation of preferences over "certain x uncertain" consumption pairs: The "ordinal certainty equivalent" hypothesis. Econometrica 46: 1045–60. Selden, L. 1979. An OCE analysis of the effect of uncertainty on saving under risk independence. Review of Economic Studies 46: 73–82. Shiller, R. 1989. Market Volatility. Cambridge: MIT Press. file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_439.html (1 of 2) [4/21/2007 1:34:11 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_439.html
Siegel, J. 1994. Stocks for the Long Run. Burr Ridge, IL: Irwin. Souleles, N. S. 1995. The response of household consumption to income tax refunds. Manuscript. MIT. Tintner, G. 1941. The theory of choice under subjective risk and uncertainty. Econometrica 9: 298–304. Telmer, C. I. 1993. Asset-pricing puzzles and incomplete markets. Journal of Finance 48: 1803–32. Topkis, D. 1978. Minilizing a submodular function on a lattice. Operations Research 26: 305–21.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_439.html (2 of 2) [4/21/2007 1:34:11 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_440.html
Page 440 Townsend, R. M. 1994. Risk and insurance in village India. Econometrica 62: 539–92. Treich, N. 1997. Risk tolerance and value of information in the standard portfolio model. Economics Letters 55: 361–63. Ulph, A., and D. Ulph. 1997. Global warming, irreversibility and learning. Economic Journal 107: 636–50. Vasicek, O. 1977. An equilibrium characterization of the term structure. Journal of Financial Economics 5: 177– 88. Viciera, L. M. 1998. Optimal portfolio choice for long-horizon investors with nontradable labor income. Discussion paper. Harvard Business School. Vickrey, W. 1960. Utility, strategy, and social decision rules. Quarterly Journal of Economics 74: 507–35. Vickson, R. G. 1975. Stochastic dominance for decreasing absolute risk aversion. Journal of Financial and Quantitative Analysis 10: 799–811. Vissing-Jorgensen, A. 1998. Limited stock market participation. Discussion paper. MIT. Vissing-Jorgensen, A. 1999. Towards an explanation of household portfolio choice heterogeneity: Nonfinancial income and participation cost structure. Discussion paper. University of Chicago. von Neumann, J., and O. Morgenstern. 1944. Theory of Games and Economic Behavior. Princeton: Princeton University Press (2nd ed., 1947). Wakker, P. P. 1989. Additive Representations of Preferences: A New Foundation in Decision Analysis, Dordrecht: Kluwer Academic. Wakker, P. P. 1999. Justifying Bayesianism by dynamic decision principles. Mimeo. Tilburg University. Weil, P. 1989. The equity premium puzzle and the risk free rate puzzle. Journal of Monetary Economics 24: 401– 21. Weil, P. 1990. Nonexpected utility in macroeconomics. Quarterly Journal of Economics 105: 29–42. Weil, P. 1992. Equilibrium asset prices with undiversifiable labor income risk. Journal of Economic Dynamics and Control 16: 769–90. Whitmore, G. 1970. Third-degree stochastic dominance. American Economic Review 60: 457–59. file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_440.html (1 of 2) [4/21/2007 1:34:12 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_440.html
Willinger, M. 1989. Risk aversion and the value of information. Journal of Risk and Insurance 56: 104–12. Wilson, R. 1968. The theory of syndicates. Econometrica 36: 113–32. Yaari, M. E. 1987. The dual theory of choice under risk. Econometrica 55: 95–115. Zeldes, S. P. 1989. Consumption and liquidity constraints: An empirical investigation. Journal of Political Economy 97: 305–46.
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_440.html (2 of 2) [4/21/2007 1:34:12 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_441.html
Page 441
Index of Lemmas and Propositions Lemma 1 Lemma 2 Lemma 3 Lemma 4 Lemma 5 Lemma 6 Lemma 7 Lemma 8 Lemma 9 Lemma 10
84 101 108 162 163 163 163 322 362 410
Proposition 1 Proposition 2 Proposition 3 Proposition 4 Proposition 5 Proposition 6 Proposition 7 Proposition 8 Proposition 9 Proposition 10 Proposition 11 Proposition 12 Proposition 13 Proposition 14 Proposition 15 Proposition 16 Proposition 17 Proposition 18 Proposition 19 Proposition 20 Proposition 21 Proposition 22 Proposition 23
18 20 21 25 45 54 59 59 60 74 83 89 92 93 95 102 103 103 105 105 106 106 115
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_441.html (1 of 2) [4/21/2007 1:34:13 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_441.html
Proposition 24 Proposition 25 Proposition 26 Proposition 27 Proposition 28 Proposition 29 Proposition 30 Proposition 31 Proposition 32 Proposition 33 Proposition 34 Proposition 35 Proposition 36 Proposition 37 Proposition 38 Proposition 39 Proposition 40 Proposition 41 Proposition 42 Proposition 43 Proposition 44 Proposition 45 Proposition 46 Proposition 47 Proposition 48 Proposition 49 Proposition 50 Proposition 51 Proposition 52 Proposition 53 Proposition 54 Proposition 55 Proposition 56 Proposition 57 Proposition 58 Proposition 59 Proposition 60 Proposition 61 Proposition 62 Proposition 63
117 118 120 121 122 129 131 137 137 143 146 147 153 160 165 166 169 177 178 180 182 185 187 187 189 199 200 201 207 209 210 211 223 223 227 230 231 237 238 239
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_441.html (2 of 2) [4/21/2007 1:34:13 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_442.html
Page 442 Proposition 64 Proposition 65 Proposition 66 Proposition 67 Proposition 68 Proposition 69 Proposition 70 Proposition 71 Proposition 72 Proposition 73 Proposition 74 Proposition 75 Proposition 76 Proposition 77 Proposition 78 Proposition 79 Proposition 80 Proposition 81 Proposition 82 Proposition 83 Proposition 84 Proposition 85 Proposition 86 Proposition 87 Proposition 88 Proposition 89 Proposition 90 Proposition 91 Proposition 92 Proposition 93 Proposition 94 Proposition 95 Proposition 96 Proposition 97 Proposition 98 Proposition 99
241 242 247 253 255 263 266 272 273 276 287 288 291 300 302 309 313 314 322 346 348 350 351 365 370 376 377 385 387 394 400 401 404 408 412 416
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_442.html [4/21/2007 1:34:13 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_443.html
Page 443
Index of Subjects Allais paradox, 10 Ambiguity aversion, 15 Absolute risk aversion. See also Decreasing absolute risk aversion definition, 19 harmonic, 26 estimation, 31 convexity of, 129, 143, 160, 187 Arrow-Pratt approximation, 22, 57 Absolute prudence definition, 24, 238 smaller than, 21, 115, 146, 209, 287, 288, 294, 387 decreasing, 106, 129, 136, 238, 324 Absolute risk tolerance definition, 26 convexity of, 165, 166, 187, 207, 211, 227, 230, 231, 291, 301, 346, 348, 350, 400 Affiliated random variables, 119, 121 Arrow-Debreu securities, 196, 286, 328, 331 Arrow-Lind theorem, 317 file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_443.html (1 of 4) [4/21/2007 1:34:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_443.html
Arrow-Pratt approximation, 22 Arbitrage, 330 Background risk, 114 increase in, 130 time horizon and, 168 Buffer stock, 270 Blackwell theorem, 370 Bond portfolio, 401 Compound lottery, 4 Continuity (axiom of), 5 Cardinal utility, 9 Consequentialism, 12 Concave function, 17, 359, 416 Concave transformation, 19 Certainty equivalent, 180, 315, 373 definition, 21 concavity, 301, 321 Compensating risk premuim, 36 Completely monotone function, 49 Call option, 49 Capital asset pricing model, 77
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_443.html (2 of 4) [4/21/2007 1:34:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_443.html
Central risk aversion, 92, 135 Central riskiness, 92 Central dominance. See Central riskiness Cauchy-Schwarz inequality, 97, 173 CAPM, 327, 333 Dutch book, 11 Decreasing absolute risk aversion, 25, 59, 61, 106, 115, 208, 273, 302 DARA. See Decreasing absolute risk aversion Diversification, 45, 74 Diffidence theorem, 83 Diffidence, 89 DAP. See Absolute prudence Discounting, 218 Expected utility theorem, 6 External approach, 40 Equity premium, 67, 334, 345 Equity premium puzzle, 68, 173, 257, 413 Exponential discounting, 218 Euler equation, 242 Early resolution of uncertainty, 299, 365 Entropy, 381
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_443.html (3 of 4) [4/21/2007 1:34:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_443.html
First-order stockastic dominance, 46, 402 Flexibility, 393
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_443.html (4 of 4) [4/21/2007 1:34:14 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_444.html
Page 444 Garbling messages, 366 Global warming, 389 Genetic test, 412, 420 Harmonic absolute risk aversion, 26, 58, 147, 166, 187, 227, 230, 242, 261, 288, 292, 345, 386, 399 HARA. See Harmonic absolute risk aversion Habit formation, 218 Hedging, 401 Hirshleifer effect, 407 Independence axiom, 6, 361 Increase in risk, 43, 60, 130, 131 Increasing relative risk aversion, 61 Insurance, 61 International portfolio diversification, 73 Indirect utility function, 114 Illiquid saving, 246 Interest rate, 252 Irreversible investment, 397 Informational efficiency, 413 Jensen's inequality, 17, 88, 211 Jewitt's preference order, 104, 122 file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_444.html (1 of 4) [4/21/2007 1:34:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_444.html
Kreps-Porteus preferences, 298 Lottery, 4 Local diffidence condition, 87 Log-supermodularity, 99, 261, 267 Leverage constraint, 185, 189 Liquidity constraint, 270 Location experiment, 371 Machina triangle, 4 Macroeconomic risk measure, 33 social cost, 34 Mean-preserving spread, 43, 366 Monotone likelihood ratio, 47, 103, 140 Monotone probability ratio, 46, 103 Myopia, 160, 177, 399, 404 Marginal propensity to consume, 206, 226, 239, 272 Martingale, 209, 211 Marginal value of wealth, 210 Mutuality principle, 309 Market price of risk, 334 Modigliani-Miller theorem, 339
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_444.html (2 of 4) [4/21/2007 1:34:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_444.html
Mean-reversion, 400 Newsboy problem, 63 No-short-sale constraint, 184 Ordinal preference, 6 One switch function, 268 Option pricing, 332 Option value 393 Pricing kernel, 66, 333 Peso problem, 71 Participation to financial markets, 71 Proper risk aversion, 134, 180, 265 Precautionary premium, 128, 237, 350 Price kernel, 197 Prudence, 236, 255, 300, 350, 393. See also Absolute prudence Precautionay saving, 236, 244, 285, 392 Precautionary equivalent growth rate, 254 Pareto efficient risk sharing, 308, 328, 407 Precautionary principle, 389 Predictability, 397 Reduction (axiom of), 4 Rank dependent EU, 15
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_444.html (3 of 4) [4/21/2007 1:34:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_444.html
Risk aversion definition, 18 comparative, 19 central, 92 Risk premium. See also Certainty equivalent definition, 20 convex in wealth, 323 Relative risk aversion, 223, 299, 417 definition, 24 and the portfolio problem, 56 decreasing, 58, 263, 266 larger than unity, 31, 61, 147, 223, 239, 289, 387, 401, 404 Relative prudence, 60, 417 Ross comparative risk aversion, 118, 122 Risk vulnerability, 126, 133, 247 Risk substituability, 144 Resistance to intertemporal substitution, 220, 255, 299, 320 Risk-free rate puzzle, 256 Representative agent, 316, 329, 343 Sure thing principle, 15 Second-order risk aversion, 22, 35, 55, 314, 317
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_444.html (4 of 4) [4/21/2007 1:34:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_445.html
Page 445 Second-order stochastic dominance, 42, 60 Single-crossing condition, 94, 102 Standard risk aversion, 128, 134, 135, 137, 241 Standard portfolio problem, 54, 158, 166, 197 Stopping time, 182 Self-insurance. See Time diversification Subhomogeneity, 231, 291 Structure of interest rate, 260 Structure of information definition, 357 finer, 365 Time consistency and Allais paradox, 11, 13 and discounting, 218 Taxation and risk aversion, 29, 36, 63 Time separability, 217 Time diversification, 227 Two-fund separation, 314 Unanimity in group decision, 318 Veil of ignorance, 312 file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_445.html (1 of 2) [4/21/2007 1:34:15 AM]
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_445.html
Value of information, 362, 372, 410 Yield curve, 258 Wealth inequality, 345
file:///D|/Export3/www.netlibrary.com/nlreader/nlreader.dll@bookid=62736&filename=page_445.html (2 of 2) [4/21/2007 1:34:15 AM]