Wiley Finance : Financial Risk Management A Practitioner's Guide To Managing Market And Credit (2nd Edition) (Wiley Finance) Steve L. Allen An
User Manual:
Open the PDF directly: View PDF .
Page Count: 609
Download | |
Open PDF In Browser | View PDF |
Financial Risk Management Founded in 1807, John Wiley & Sons is the oldest independent publishing company in the United States. With offices in North America, Europe, Australia, and Asia, Wiley is globally committed to developing and marketing print and electronic products and services for our customers’ professional and personal knowledge and understanding. The Wiley Finance series contains books written specifically for finance and investment professionals as well as sophisticated individual investors and their financial advisors. Book topics range from portfolio management to e-commerce, risk management, financial engineering, valuation, and financial instrument analysis, as well as much more. For a list of available titles, visit our Web site at www.WileyFinance.com. Financial Risk Management A Practitioner’s Guide to Managing Market and Credit Risk Second Edition STEVEN ALLEN John Wiley & Sons, Inc. Cover image: John Wiley & Sons, Inc. Cover design: © Tom Fewster / iStockphoto, © samxmeg / iStockphoto Copyright © 2013 by Steven Allen. All rights reserved. Published by John Wiley & Sons, Inc., Hoboken, New Jersey. Published simultaneously in Canada. The First Edition of this book was published in 2003 by John Wiley & Sons, Inc. No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning, or otherwise, except as permitted under Section 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per-copy fee to the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, (978) 750-8400, fax (978) 646-8600, or on the Web at www.copyright.com. Requests to the Publisher for permission should be addressed to the Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, (201) 748-6011, fax (201) 748-6008, or online at http://www.wiley.com/go/permissions. Limit of Liability/Disclaimer of Warranty: While the publisher and author have used their best efforts in preparing this book, they make no representations or warranties with respect to the accuracy or completeness of the contents of this book and specifically disclaim any implied warranties of merchantability or fitness for a particular purpose. No warranty may be created or extended by sales representatives or written sales materials. The advice and strategies contained herein may not be suitable for your situation. You should consult with a professional where appropriate. Neither the publisher nor author shall be liable for any loss of profit or any other commercial damages, including but not limited to special, incidental, consequential, or other damages. For general information on our other products and services or for technical support, please contact our Customer Care Department within the United States at (800) 762-2974, outside the United States at (317) 572-3993 or fax (317) 572-4002. Wiley publishes in a variety of print and electronic formats and by print-on-demand. Some material included with standard print versions of this book may not be included in e-books or in print-ondemand. If this book refers to media such as a CD or DVD that is not included in the version you purchased, you may download this material at http://booksupport.wiley.com. For more information about Wiley products, visit www.wiley.com. Library of Congress Cataloging-in-Publication Data: Allen, Steven, 1945– Financial risk management [electronic resource]: a practitioner’s guide to managing market and credit risk / Steven Allen. — 2nd ed. 1 online resource. Includes bibliographical references and index. Description based on print version record and CIP data provided by publisher; resource not viewed. ISBN 978-1-118-17545-3 (cloth); 978-1-118-22652-0 (ebk.); ISBN 978-1-118-23164-7 (ebk.); ISBN 978-1-118-26473-7 (ebk.) 1. Financial risk management. 2. Finance. I. Title. HD61 658.15'5—dc23 2012029614 Printed in the United States of America. 10 9 8 7 6 5 4 3 2 1 To Caroline For all the ways she has helped bring this project to fruition And for much, much more Financial Risk Management Contents Foreword xvii Preface xix Acknowledgments xxiii About the Author xxvii CHAPTER 1 Introduction 1.1 Lessons from a Crisis 1.2 Financial Risk and Actuarial Risk 1.3 Simulation and Subjective Judgment CHAPTER 2 Institutional Background 2.1 2.2 2.3 2.4 2.5 Moral Hazard—Insiders and Outsiders Ponzi Schemes Adverse Selection The Winner’s Curse Market Making versus Position Taking CHAPTER 3 Operational Risk 3.1 Operations Risk 3.1.1 The Risk of Fraud 3.1.2 The Risk of Nondeliberate Incorrect Information 3.1.3 Disaster Risk 3.1.4 Personnel Risk 3.2 Legal Risk 3.2.1 The Risk of Unenforceable Contracts 3.2.2 The Risk of Illegal Actions 3.3 Reputational Risk 3.4 Accounting Risk 1 1 2 4 7 7 17 19 21 24 29 31 31 35 36 36 37 37 40 41 42 ix x CONTENTS 3.5 3.6 3.7 3.8 Funding Liquidity Risk Enterprise Risk Identification of Risks Operational Risk Capital CHAPTER 4 Financial Disasters 4.1 Disasters Due to Misleading Reporting 4.1.1 Chase Manhattan Bank/Drysdale Securities 4.1.2 Kidder Peabody 4.1.3 Barings Bank 4.1.4 Allied Irish Bank (AIB) 4.1.5 Union Bank of Switzerland (UBS) 4.1.6 Société Générale 4.1.7 Other Cases 4.2 Disasters Due to Large Market Moves 4.2.1 Long‐Term Capital Management (LTCM) 4.2.2 Metallgesellschaft (MG) 4.3 Disasters Due to the Conduct of Customer Business 4.3.1 Bankers Trust (BT) 4.3.2 JPMorgan, Citigroup, and Enron 4.3.3 Other Cases CHAPTER 5 The Systemic Disaster of 2007–2008 5.1 Overview 5.2 The Crisis in CDOs of Subprime Mortgages 5.2.1 Subprime Mortgage Originators 5.2.2 CDO Creators 5.2.3 Rating Agencies 5.2.4 Investors 5.2.5 Investment Banks 5.2.6 Insurers 5.3 The Spread of the Crisis 5.3.1 Credit Contagion 5.3.2 Market Contagion 5.4 Lessons from the Crisis for Risk Managers 5.4.1 Subprime Mortgage Originators 5.4.2 CDO Creators 5.4.3 Rating Agencies 5.4.4 Investors 5.4.5 Investment Banks 42 44 44 45 49 49 52 53 55 57 59 61 66 68 68 75 77 77 79 80 83 83 85 86 88 89 92 93 106 108 108 109 111 111 111 111 111 112 Contents 5.4.6 Insurers 5.4.7 Credit Contagion 5.4.8 Market Contagion 5.5 Lessons from the Crisis for Regulators 5.5.1 Mortgage Originators 5.5.2 CDO Creators 5.5.3 Rating Agencies 5.5.4 Investors 5.5.5 Investment Banks 5.5.6 Insurers 5.5.7 Credit Contagion 5.5.8 Market Contagion 5.6 Broader Lessons from the Crisis CHAPTER 6 Managing Financial Risk 6.1 Risk Measurement 6.1.1 General Principles 6.1.2 Risk Management of Instruments That Lack Liquidity 6.1.3 Market Valuation 6.1.4 Valuation Reserves 6.1.5 Analysis of Revenue 6.1.6 Exposure to Changes in Market Prices 6.1.7 Risk Measurement for Position Taking 6.2 Risk Control CHAPTER 7 VaR and Stress Testing 7.1 VaR Methodology 7.1.1 Simulation of the P&L Distribution 7.1.2 Measures of the P&L Distribution 7.2 Stress Testing 7.2.1 Overview 7.2.2 Economic Scenario Stress Tests 7.2.3 Stress Tests Relying on Historical Data 7.3 Uses of Overall Measures of Firm Position Risk CHAPTER 8 Model Risk 8.1 How Important Is Model Risk? 8.2 Model Risk Evaluation and Control xi 114 115 115 115 116 116 117 118 118 126 126 129 132 133 133 133 144 147 152 156 157 159 161 169 170 173 187 192 192 193 197 201 209 210 212 xii CONTENTS 8.2.1 8.2.2 Scope of Model Review and Control Roles and Responsibilities for Model Review and Control 8.2.3 Model Verification 8.2.4 Model Verification of Deal Representation 8.2.5 Model Verification of Approximations 8.2.6 Model Validation 8.2.7 Continuous Review 8.2.8 Periodic Review 8.3 Liquid Instruments 8.4 Illiquid Instruments 8.4.1 Choice of Model Validation Approach 8.4.2 Choice of Liquid Proxy 8.4.3 Design of Monte Carlo Simulation 8.4.4 Implications for Marking to Market 8.4.5 Implications for Risk Reporting 8.5 Trading Models CHAPTER 9 Managing Spot Risk 9.1 9.2 9.3 9.4 Overview Foreign Exchange Spot Risk Equity Spot Risk Physical Commodities Spot Risk CHAPTER 10 Managing Forward Risk 10.1 Instruments 10.1.1 Direct Borrowing and Lending 10.1.2 Repurchase Agreements 10.1.3 Forwards 10.1.4 Futures Contracts 10.1.5 Forward Rate Agreements 10.1.6 Interest Rate Swaps 10.1.7 Total Return Swaps 10.1.8 Asset‐Backed Securities 10.2 Mathematical Models of Forward Risks 10.2.1 Pricing Illiquid Flows by Interpolation 10.2.2 Pricing Long‐Dated Illiquid Flows by Stack and Roll 10.2.3 Flows Representing Promised Deliveries 10.2.4 Indexed Flows 213 214 219 222 223 226 232 234 237 241 241 243 245 247 249 250 253 253 257 258 259 263 270 270 271 272 272 274 275 276 278 282 284 291 293 295 Contents 10.3 Factors Impacting Borrowing Costs 10.3.1 The Nature of Borrowing Demand 10.3.2 The Possibility of Cash‐and‐Carry Arbitrage 10.3.3 The Variability of Storage Costs 10.3.4 The Seasonality of Borrowing Costs 10.3.5 Borrowing Costs and Forward Prices 10.4 Risk Management Reporting and Limits for Forward Risk CHAPTER 11 Managing Vanilla Options Risk 11.1 11.2 11.3 11.4 11.5 11.6 Overview of Options Risk Management The Path Dependence of Dynamic Hedging A Simulation of Dynamic Hedging Risk Reporting and Limits Delta Hedging Building a Volatility Surface 11.6.1 Interpolating between Time Periods 11.6.2 Interpolating between Strikes—Smile and Skew 11.6.3 Extrapolating Based on Time Period 11.7 Summary CHAPTER 12 Managing Exotic Options Risk 12.1 Single‐Payout Options 12.1.1 Log Contracts and Variance Swaps 12.1.2 Single‐Asset Quanto Options 12.1.3 Convexity 12.1.4 Binary Options 12.1.5 Contingent Premium Options 12.1.6 Accrual Swaps 12.2 Time‐Dependent Options 12.2.1 Forward‐Starting and Cliquet Options 12.2.2 Compound Options 12.3 Path‐Dependent Options 12.3.1 Standard Analytic Models for Barriers 12.3.2 Dynamic Hedging Models for Barriers 12.3.3 Static Hedging Models for Barriers 12.3.4 Barrier Options with Rebates, Lookback, and Ladder Options 12.3.5 Broader Classes of Path‐Dependent Exotics 12.4 Correlation‐Dependent Options xiii 299 299 300 301 302 303 304 311 313 318 321 329 344 346 346 347 352 355 359 364 367 369 370 371 377 378 378 378 379 381 383 385 387 402 403 404 xiv CONTENTS 12.4.1 Linear Combinations of Asset Prices 12.4.2 Risk Management of Options on Linear Combinations 12.4.3 Index Options 12.4.4 Options to Exchange One Asset for Another 12.4.5 Nonlinear Combinations of Asset Prices 12.4.6 Correlation between Price and Exercise 12.5 Correlation‐Dependent Interest Rate Options 12.5.1 Models in Which the Relationship between Forwards Is Treated as Constant 12.5.2 Term Structure Models 12.5.3 Relationship between Swaption and Cap Prices CHAPTER 13 Credit Risk 13.1 Short‐Term Exposure to Changes in Market Prices 13.1.1 Credit Instruments 13.1.2 Models of Short‐Term Credit Exposure 13.1.3 Risk Reporting for Market Credit Exposures 13.2 Modeling Single‐Name Credit Risk 13.2.1 Estimating Probability of Default 13.2.2 Estimating Loss Given Default 13.2.3 Estimating the Amount Owed at Default 13.2.4 The Option‐Theoretic Approach 13.3 Portfolio Credit Risk 13.3.1 Estimating Default Correlations 13.3.2 Monte Carlo Simulation of Portfolio Credit Risk 13.3.3 Computational Alternatives to Full Simulation 13.3.4 Risk Management and Reporting for Portfolio Credit Exposures 13.4 Risk Management of Multiname Credit Derivatives 13.4.1 Multiname Credit Derivatives 13.4.2 Modeling of Multiname Credit Derivatives 13.4.3 Risk Management and Reporting for Multiname Credit Derivatives 13.4.4 CDO Tranches and Systematic Risk CHAPTER 14 Counterparty Credit Risk 14.1 Overview 14.2 Exchange‐Traded Derivatives 405 409 413 415 417 422 425 426 430 437 445 446 447 451 456 457 458 465 468 471 479 479 482 486 490 493 493 495 498 500 505 505 506 Contents 14.3 Over‐the‐Counter Derivatives 14.3.1 Overview 14.3.2 The Loan‐Equivalent Approach 14.3.3 The Collateralization Approach 14.3.4 The Collateralization Approach— Wrong‐Way Risk 14.3.5 The Active Management Approach xv 512 512 513 515 521 526 References 533 About the Companion Website 547 Index 553 Foreword isk was a lot easier to think about when I was a doctoral student in finance 25 years ago. Back then, risk was measured by the variance of your wealth. Lowering risk meant lowering this variance, which usually had the unfortunate consequence of lowering the average return on your wealth as well. In those halcyon days, we had only two types of risk, systemic and unsystematic. The latter one could be lowered for free via diversification, while the former one could only be lowered by taking a hit to average return. In that idyllic world, financial risk management meant choosing the variance that maximized expected utility. One merely had to solve an optimization problem. What could be easier? I started to appreciate that financial risk management might not be so easy when I moved from the West Coast to the East Coast. The New York–based banks started creating whole departments to manage financial risk. Why do you need dozens of people to solve a simple optimization problem? As I talked with the denizens of those departments, I noticed they kept introducing types of risk that were not in my financial lexicon. First there was credit risk, a term that was to be differentiated from market risk, because you can lose money lending whether a market exists or not. Fine, I got that, but then came liquidity risk on top of market and credit risk. Just as I was struggling to integrate these three types of risk, people started worrying about operational risk, basis risk, mortality risk, weather risk, estimation risk, counterparty credit risk, and even the risk that your models for all these risks were wrong. If model risk existed, then you had to concede that even your model for model risk was risky. Since the proposed solution for all these new risks were new models and since the proposed solution for the model risk of the new models was yet more models, it was no wonder all of those banks had all of those people running around managing all of those risks. Well, apparently, not quite enough people. As I write these words, the media are having a field day denouncing JPMorgan’s roughly $6 billion loss related to the London whale’s ill-fated foray into credit default swaps (CDSs). As the flag bearer for the TV generation, I can’t help but think of reviving a 1970s TV show to star Bruno Iksil as the Six Billion Dollar Man. As eyepopping as these numbers are, they are merely the fourth largest trading loss since the first edition of this book was released. If we ignore Bernie Madoff’s R xvii xviii FOREWORD $50 billion Ponzi scheme, the distinction for the worst trade ever belongs to Howie Hubler, who lost $9 billion trading CDSs in 2008 for another bank whose name I’d rather not write. However, if you really need to know, then here’s a hint. The present occupant of Mr. Hubler’s old office presently thinks that risk management is a complicated subject, very complicated indeed, and has to admit that a simple optimization is not the answer. So what is the answer? Well, when the answer to a complicated question is nowhere to be found in the depths of one’s soul, then one can always fall back on asking the experts instead. The Danish scientist Niels Bohr, once deemed an expert, said an expert is, “A person that has made every possible mistake within his or her field.” As an expert in the field of derivative securities valuation, I believe I know a fellow expert when I see one. Steve Allen has been teaching courses in risk management at New York University’s Courant Institute since 1998. Steve retired from JPMorgan Chase as a managing director in 2004, capping a 35-year career in the finance industry. Given the wide praise for the first edition of this book, the author could have rested on his laurels, comforted by the knowledge that the wisdom of the ages is eternal. Instead, he has taken it upon himself to write a second edition of this timeless book. Most authors in Steve’s enviable situation would have contented themselves with exploiting the crisis to elaborate on some extended version of “I told you so.” Instead, Steve has added much in the way of theoretical advances that have arisen out of the necessity of ensuring that history does not repeat itself. These advances in turn raise the increasing degree of specialization we see inside the risk management departments of modern financial institutions and increasingly in the public sector as well. Along with continued progress in the historically vital problem of marking to market of illiquid positions, there is an increasing degree of rigor in the determination of reserves that arise due to model risk, in the limits used to control risk taking, and in the methods used to review models. The necessity of testing every assumption has been made plain by the stress that the crisis has imposed on our fragile financial system. As the aftershocks reverberate around us, we will not know for many years whether the present safeguards will serve their intended purpose. However, the timing for an update to Steve’s book could not be better. I truly hope that the current generation of risk managers, whether they be grizzled or green, will take the lessons on the ensuing pages to heart. Our shared financial future depends on it. Peter Carr, PhD Managing Director at Morgan Stanley, Global Head of Market Modeling, and Executive Director of New York University Courant’s Masters in Mathematical Finance Preface his book offers a detailed introduction to the field of risk management as performed at large investment and commercial banks, with an emphasis on the practices of specialist market risk and credit risk departments as well as trading desks. A large portion of these practices is also applicable to smaller institutions that engage in trading or asset management. The aftermath of the financial crisis of 2007–2008 leaves a good deal of uncertainty as to exactly what the structure of the financial industry will look like going forward. Some of the business currently performed in investment and commercial banks, such as proprietary trading, may move to other institutions, at least in some countries, based on new legislation and new regulations. But in whatever institutional setting this business is conducted, the risk management issues will be similar to those encountered in the past. This book focuses on general lessons as to how the risk of financial institutions can be managed rather than on the specifics of particular regulations. My aim in this book is to be comprehensive in looking at the activities of risk management specialists as well as trading desks, at the realm of mathematical finance as well as that of the statistical techniques, and, most important, at how these different approaches interact in an integrated risk management process. This second edition reflects lessons that have been learned from the recent financial crisis of 2007–2008 (for more detail, see Chapters 1 and 5), as well as many new books, articles, and ideas that have appeared since the publication of the first edition in 2003. Chapter 6 on managing market risk, Chapter 7 on value at risk (VaR) and stress testing, Chapter 8 on model risk, and Chapter 13 on credit risk are almost completely rewritten and expanded from the first edition, and a new Chapter 14 on counterparty credit risk is an extensive expansion of a section of the credit risk chapter in the first edition. The website for this book (www.wiley.com/go/frm2e) will be used to provide both supplementary materials to the text and continuous updates. Supplementary materials will include spreadsheets and computer code that illustrate computations discussed in the text. In addition, there will be classroom aids available only to professors on the Wiley Higher Education website. Updates will include an updated electronic version of the References T xix xx PREFACE section, to allow easy cut‐and‐paste linking to referenced material on the web. Updates will also include discussion of new developments. For example, at the time this book went to press, there is not yet enough public information about the causes of the large trading losses at JPMorgan’s London investment office to allow a discussion of risk management lessons; as more information becomes available, I will place an analysis of risk management lessons from these losses on the website. This book is divided into three parts: general background to financial risk management, the principles of financial risk management, and the details of financial risk management. ■ ■ ■ The general background part (Chapters 1 through 5) gives an institutional framework for understanding how risk arises in financial firms and how it is managed. Without understanding the different roles and motivations of traders, marketers, senior firm managers, corporate risk managers, bondholders, stockholders, and regulators, it is impossible to obtain a full grasp of the reasoning behind much of the machinery of risk management or even why it is necessary to manage risk. In this part, you will encounter key concepts risk managers have borrowed from the theory of insurance (such as moral hazard and adverse selection), decision analysis (such as the winner’s curse), finance theory (such as the arbitrage principle), and in one instance even the criminal courts (the Ponzi scheme). Chapter 4 provides discussion of some of the most prominent financial disasters of the past 30 years, and Chapter 5 focuses on the crisis of 2007–2008. These serve as case studies of failures in risk management and will be referenced throughout the book. This part also contains a chapter on operational risk, which is necessary background for many issues that arise in preventing financial disasters and which will be referred to throughout the rest of the book. The part on principles of financial risk management (Chapters 6 through 8) first lays out an integrated framework in Chapter 6, and then looks at VaR and stress testing in Chapter 7 and the control of model risk in Chapter 8. The part on details of financial risk management (Chapters 9 through 14) applies the principles of the second part to each specific type of financial risk: spot risk in Chapter 9, forward risk in Chapter 10, vanilla options risk in Chapter 11, exotic options risk in Chapter 12, credit risk in Chapter 13, and counterparty credit risk in Chapter 14. As each risk type is discussed, specific references are made to the principles elucidated in Chapters 6 through 8, and a detailed analysis of the models used to price these risks and how these models can be used to measure and control risk is presented. Preface xxi Since the 1990s, an increased focus on the new technology being developed to measure and control financial risk has resulted in the growth of corporate staff areas manned by risk management professionals. However, this does not imply that financial firms did not manage risks prior to 1990 or that currently all risk management is performed in staff areas. Senior line managers such as trading desk and portfolio managers have always performed a substantial risk management function and continue to do so. In fact, confusion can be caused by the tradition of using the term risk manager as a synonym for a senior trader or portfolio manager and as a designation for members of corporate staff areas dealing with risk. Although this book covers risk management techniques that are useful to both line trading managers and corporate staff acting on behalf of the firm’s senior management, the needs of these individuals do not completely overlap. I will try to always make a clear distinction between information that is useful to a trading desk and information that is needed by corporate risk managers, and explain how they might intersect. Books and articles on financial risk management have tended to focus on statistical techniques embodied in measures such as value at risk (VaR). As a result, risk management has been accused of representing a very narrow specialty with limited value, a view that has been colorfully expressed by Nassim Taleb (1997), “There has been growth in the number of ‘risk management advisors,’ an industry sometimes populated by people with an amateurish knowledge of risk. Using some form of shallow technical skills, these advisors emit pronouncements on such matters as ‘risk management’ without a true understanding of the distribution. Such inexperience and weakness become more apparent with the value‐at‐risk fad or the outpouring of books on risk management by authors who never traded a contract” (p. 4). This book gives a more balanced account of risk management. Less than 20 percent of the material looks at statistical techniques such as VaR. The bulk of the book examines issues such as the proper mark‐to‐market valuation of trading positions, the determination of necessary reserves against valuation uncertainty, the structuring of limits to control risk taking, and the review of mathematical models and determination of how they can contribute to risk control. This allocation of material mirrors the allocation of effort in the corporate risk management staff areas with which I am familiar. This is reflected in the staffing of these departments. More personnel is drawn from those with experience and expertise in trading and building models to support trading decisions than is drawn from a statistical or academic finance background. Although many readers may already have a background in the instruments—bonds, stocks, futures, and options—used in the financial markets, I have supplied definitions every time I introduce a term. Terms are italicized in the text at the point they are defined. Any reader feeling the need for a xxii PREFACE more thorough introduction to market terminology should find the first nine chapters of Hull (2012) adequate preparation for understanding the material in this book. My presentation of the material is based both on theory and on how concepts are utilized in industry practice. I have tried to provide many concrete instances of either personal experience or reports I have heard from industry colleagues to illustrate these practices. Where incidents have received sufficient previous public scrutiny or occurred long enough ago that issues of confidentiality are not a concern, I have provided concrete details. In other cases, I have had to preserve the anonymity of my sources by remaining vague about particulars. My preservation of anonymity extends to a liberal degree of randomness in references to gender. A thorough discussion of how mathematical models are used to measure and control risks must make heavy reference to the mathematics used in creating these models. Since excellent expositions of the mathematics exist, I do not propose to enter into extensive derivations of results that can readily be found elsewhere. Instead, I will concentrate on how these results are used in risk management and how the approximations to reality inevitable in any mathematical abstraction are dealt with in practice. I will provide references to the derivation of results. Wherever possible, I have used Hull (2012) as a reference, since it is the one work that can be found on the shelf of nearly every practitioner in the field of quantitative finance. Although the material for this book was originally developed for a course taught within a mathematics department, I believe that virtually all of its material will be understandable to students in finance programs and business schools, and to practitioners with a comparable educational background. A key reason for this is that whereas derivatives mathematics often emphasizes the use of more mathematically sophisticated continuous time models, discrete time models are usually more relevant to risk management, since risk management is often concerned with the limits that real market conditions place on mathematical theory. This book is designed to be used either as a text for a course in risk management or as a resource for self‐study or reference for people working in the financial industry. To make the material accessible to as broad an audience as possible, I have tried everywhere to supplement mathematical theory with concrete examples and have supplied spreadsheets on the accompanying website (www.wiley.com/go/frm2e) to illustrate these calculations. Spreadsheets on the website are referenced throughout the text and a summary of all spreadsheets supplied is provided in the “About the Companion Website” section at the back of the book. At the same time, I have tried to make sure that all the mathematical theory that gets used in risk management practice is addressed. For readers who want to pursue the theoretical developments at greater length, a full set of references has been provided. Acknowledgments he views expressed in this book are my own, but have been shaped by my experiences in the financial industry. Many of my conclusions about what constitutes best practice in risk management have been based on my observation of and participation in the development of the risk management structure at JPMorgan Chase and its Chemical Bank and Chase Manhattan Bank predecessors. The greatest influence on my overall view of how financial risk management should be conducted and on many of the specific approaches I advocate has been Lesley Daniels Webster. My close collaboration with Lesley took place over a period of 20 years, during the last 10 of which I reported to her in her position as director of market risk management. I wish to express my appreciation of Lesley’s leadership, along with that of Marc Shapiro, Suzanne Hammett, Blythe Masters, and Andy Threadgold, for having established the standards of integrity, openness, thoroughness, and intellectual rigor that have been the hallmarks of this risk management structure. Throughout most of the period in which I have been involved in these pursuits, Don Layton was the head of trading activities with which we interacted. His recognition of the importance of the risk management function and strong support for a close partnership between risk management and trading and the freedom of communication and information sharing were vital to the development of these best practices. Through the years, my ideas have benefited from my colleagues at Chemical, Chase, JPMorgan Chase, and in consulting assignments since my retirement from JPMorgan Chase. At JPMorgan Chase and its predecessors, I would particularly like to note the strong contributions that dialogues with Andrew Abrahams, Michel Araten, Bob Benjamin, Paul Bowmar, George Brash, Julia Chislenko, Enrico Della Vecchia, Mike Dinias, Fawaz Habel, Bob Henderson, Jeff Katz, Bobby Magee, Blythe Masters, Mike Rabin, Barry Schachter, Vivian Shelton, Paul Shotton, Andy Threadgold, Mick Waring, and Richard Wise have played in the development of the concepts utilized here. In my consulting assignments, I have gained much from my exchanges of ideas with Rick Grove, Chia‐Ling Hsu, Neil Pearson, Bob Selvaggio, Charles Smithson, and other colleagues at Rutter Associates, and Chris Marty and Alexey Panchekha at Bloomberg. In interactions with risk T xxiii xxiv ACKNOWLEDGMENTS managers at other firms, I have benefited from my conversations with Ken Abbott, John Breit, Noel Donohoe, and Evan Picoult. Many of the traders I have interacted with through the years have also had a major influence on my views of how risk management should impact decision making on the trading desk and the proper conduct of relationships between traders and risk management specialists. I particularly want to thank Andy Hollings, Simon Lack, Jeff Larsen, Dinsa Mehta, Fraser Partridge, and Don Wilson for providing me with prototypes for how the risk management of trading should be properly conducted and their generosity in sharing their knowledge and insight. I also wish to thank those traders, who shall remain anonymous here, who have provided me equally valuable lessons in risk management practices to avoid. This book grew out of the risk management course I created as part of the Mathematics in Finance MS program at New York University’s Courant Institute of Mathematical Sciences in 1998. For giving me the opportunity to teach and for providing an outstanding institutional setting in which to do it, I want to thank the administration and faculty of Courant, particularly Peter Carr, Neil Chriss, Jonathan Goodman, Bob Kohn, and Petter Kolm, with whom I have participated in the management of the program, and Caroline Thompson, Gabrielle Tobin, and Melissa Vacca, the program administrators. I have gained many insights that have found their way into this book by attending other courses in the program taught by Marco Avellaneda, Jim Gatheral, Bob Kohn, and Nassim Taleb. Ken Abbott began participating in the risk management course as a guest lecturer, later became my co‐teacher of the course, and now has full responsibility for the course with my participation as a guest lecturer. Many of the insights in this book have been learned from Ken or generated as part of the debates and discussions we have held both in and out of the classroom. The students in my risk management course have helped clarify many of the concepts in this book through their probing questions. I particularly want to thank Karim Beguir, who began as my student and has since graduated to become a Fellow of the program and a frequent and valued contributor to the risk management course. Several of his insights are reflected in the second edition of the book. I also wish to thank Otello Padovani and Andrea Raphael, students who became collaborators on research that appears on the website for the book (www.wiley.com/go/frm2e). Mike Fisher has provided greatly appreciated support as my graduate assistant in helping to clarify class assignments that have evolved into exercises in this book. The detailed comments and suggestions I have received from Neil Chriss on large portions of this manuscript far exceed the norms of either friendship or collegiality. In numerous instances, his efforts have sharpened both the ideas being presented and the clarity of their expression. I also wish Acknowledgments xxv to thank Mich Araten, Peter Carr, Bobby Magee, Barry Schachter, Nassim Taleb, and Bruce Tuckman for reading the text and offering helpful comments. For the second edition, I would like to thank Ken Abbott and Rick Grove for reading new chapters and offering helpful suggestions. I also wish to extend my thanks to Chuck Epstein for his help in finding a publisher for this book. Bill Falloon, Meg Freeborn, and Michael Kay, my editors at John Wiley & Sons, have offered very useful suggestions at every stage of the editing. At MacAllister Publishing Services, Andy Stone was very helpful as production manager and Jeanne Henning was a thorough and incisive copy editor for the first edition of this book. The individual to whom both I and this book owe the greatest debt is my wife, Caroline Thompson. The number of ways in which her beneficial influence has been felt surpass my ability to enumerate, but I at least need to attempt a brief sample. It was Caroline who introduced me to Neil Chriss and first planted the idea of my teaching at Courant. She has been a colleague of Neil’s, Jonathan Goodman’s, and mine in the continued development of the Courant Mathematics in Finance MS program. From the start, she was the strongest voice in favor of basing a book on my risk management course. At frequent bottlenecks, on both the first and second editions, when I have been daunted by an obstacle to my progress that seemed insurmountable, it was Caroline who suggested the approach, organized the material, or suggested the joint effort that overcame the difficulty. She has managed all aspects of the production format, and style of the book, including efforts from such distant ports as Laos, Vietnam, India, and Holland. About the Author teve Allen is a risk management consultant, specializing in risk measurement and valuation with a particular emphasis on illiquid and hard‐ to‐value assets. Until his retirement in 2004, he was Managing Director in charge of risk methodology at JPMorgan Chase, where he was responsible for model validation, risk capital allocation, and the development of new measures of valuation, reserves, and risk for both market and credit risk. Previously, he was in charge of market risk for derivative products at Chase. He has been a key architect of Chase’s value‐at‐risk and stress testing systems. Prior to his work in risk management, Allen was the head of analysis and model building for all Chase trading activities for over ten years. Since 1998, Allen has been associated with the Mathematics in Finance Masters’ program at New York University’s Courant Institute of Mathematical Sciences. In this program, he has served as Clinical Associate Professor and Deputy Director and has created and taught courses in risk management, derivatives mathematics, and interest rate and credit models. He was a member of the Board of Directors of the International Association of Financial Engineers and continues to serve as co‐chair of their Education Committee. S xxvii CHAPTER 1 Introduction 1.1 LESSONS FROM A CRISIS I began the first edition of this book with a reference to an episode of the television series Seinfeld in which the character George Costanza gets an assignment from his boss to read a book titled Risk Management and then give a report on this topic to other business executives. Costanza finds the book and topic so boring that his only solution is to convince someone else to read it for him and prepare notes. Clearly, my concern at the time was to write about financial risk management in a way that would keep readers from finding the subject dull. I could hardly have imagined then that eight years later Demi Moore would be playing the part of the head of an investment bank’s risk management department in a widely released movie, Margin Call. Even less could I have imagined the terrible events that placed financial risk management in such a harsh spotlight. My concern now is that the global financial crisis of 2007–2008 may have led to the conclusion that risk management is an exciting subject whose practitioners and practices cannot be trusted. I have thoroughly reviewed the material I presented in the first edition, and it still seems to me that if the principles I presented, principles that represented industry best practices, had been followed consistently, a disaster of the magnitude we experienced would not have been possible. In particular, the points I made in the first edition about using stress tests in addition to value at risk (VaR) in determining capital adequacy (see the last paragraphs of Section 7.3 in this edition) and the need for substantial reserves and deferred compensation for illiquid positions (see Sections 6.1.4 and 8.4 in this edition) still seem sound. It is tempting to just restate the same principles and urge more diligence in their application, but that appears too close to the sardonic definition of insanity: “doing the same thing and expecting different results.” So I have looked for places where these principles need strengthening (you’ll find a summary in Section 5.4). But I have also reworked the organization of 1 2 FINANCIAL RISK MANAGEMENT the book to emphasize two core doctrines that I believe are the keys to the understanding and proper practice of financial risk management. The first core principle is that financial risk management is not just risk management as practiced in financial institutions; it is risk management that makes active use of trading in liquid markets to control risk. Risk management is a discipline that is important to a wide variety of companies, government agencies, and institutions—one need only think of accident prevention at nuclear power plants and public health measures to avoid influenza pandemics to see how critical it can be. While the risk management practiced at investment banks shares some techniques with risk management practiced at a nuclear facility, there remains one vital difference: much of the risk management at investment banks can utilize liquid markets as a key element in risk control; liquid markets are of virtually no use to the nuclear safety engineer. My expertise is in the techniques of financial risk management, and that is the primary subject of this book. Some risks that financial firms take on cannot be managed using trading in liquid markets. It is vitally important to identify such risks and to be aware of the different risk management approaches that need to be taken for them. Throughout the book I will be highlighting this distinction and also focusing on the differences that degree of available liquidity makes. As shorthand, I will refer to risk that cannot be managed by trading in liquid markets as actuarial risk, since it is the type of risk that actuaries at insurance companies have been dealing with for centuries. Even in cases that must be analyzed using the actuarial risk approach, financial risk management techniques can still be useful in isolating the actuarial risk and in identifying market data that can be used as input to actuarial risk calculations. I will address this in greater detail in Section 1.2. The second core principle is that the quantification of risk management requires simulation guided by both historical data and subjective judgment. This is a common feature of both financial risk and actuarial risk. The time period simulated may vary greatly, from value at risk (VaR) simulations of daily market moves for very liquid positions to simulations spanning decades for actuarial risk. But I will be emphasizing shared characteristics for all of these simulations: the desirability of taking advantage of as much historical data as is relevant, the need to account for nonnormality of statistical distributions, and the necessity of including subjective judgment. More details on these requirements are in Section 1.3. 1.2 FINANCIAL RISK AND ACTUARIAL RISK The management of financial risk and the management of actuarial risk do share many methodologies, a point that will be emphasized in the next Introduction 3 section. Both rely on probability and statistics to arrive at estimates of the distribution of possible losses. The critical distinction between them is the matter of time. Actuarial risks may not be fully resolved for years, sometimes even decades. By the time the true extent of losses is known, the accumulation of risk may have gone on for years. Financial risks can be eliminated in a relatively short time period by the use of liquid markets. Continuous monitoring of the price at which risk can be liquidated should substantially lower the possibility of excessive accumulation of risk. Two caveats need to be offered to this relatively benign picture of financial risk. The first is that taking advantage of the shorter time frame of financial risk requires constant vigilance; if you aren’t doing a good job of monitoring how large your risks are relative to liquidation costs, you may still acquire more exposure than desired. This will be described in detail in Chapter 6. The second is the need to be certain that what is truly actuarial risk has not been misclassified as financial risk. If this occurs, it is especially dangerous—not only will you have the potential accumulation of risk over years before the extent of losses is known, but in not recognizing the actuarial nature, you would not exercise the caution that the actuarial nature of the risk demands. This will be examined more closely in Sections 6.1.1 and 6.1.2, with techniques for management of actuarial risk in financial firms outlined in Section 8.4. I believe that this dangerous muddling of financial and actuarial risk was a key contributor to the 2007–2008 crisis, as I argue in Section 5.2.5. Of course, it is only an approximation to view instruments as being liquid or illiquid. The volume of instruments available for trading differs widely by size and readiness of availability. This constitutes the depth of liquidity of a given market. Often a firm will be faced with a choice between the risks of replicating positions more exactly with less liquid instruments or less exactly with more liquid instruments. One theme of this book will be the trade‐off between liquidity risk and basis risk. Liquidity risk is the risk that the price at which you buy (or sell) something may be significantly less advantageous than the price you could have achieved under more ideal conditions. Basis risk is the risk that occurs when you buy one product and sell another closely related one, and the two prices behave differently. Let’s look at an example. Suppose you are holding a large portfolio of stocks that do not trade that frequently and your outlook for stock prices leads to a desire to quickly terminate the position. If you try selling the whole basket quickly, you face significant liquidity risk since your selling may depress the prices at which the stocks trade. An alternative would be to take an offsetting position in a heavily traded stock futures contract, such as the futures contract tied to the Standard & Poor’s™ S&P 500 stock index. This lowers the liquidity risk, but it increases the 4 FINANCIAL RISK MANAGEMENT basis risk since changes in the price of your particular stock basket will probably differ from the price changes in the stock index. Often the only way in which liquidity risk can be reduced is to increase basis risk, and the only way in which basis risk can be reduced is to increase liquidity risk. The classification of risk as financial risk or actuarial risk is clearly a function of the particular type of risk and not of the institution. Insurance against hurricane damage could be written as a traditional insurance contract by Metropolitan Life or could be the payoff of an innovative new swap contract designed by Morgan Stanley; in either case, it will be the same risk. What is required in either case is analysis of how trading in liquid markets can be used to manage the risk. Certainly commercial banks have historically managed substantial amounts of actuarial risk in their loan portfolios. And insurance companies have managed to create some ability to liquidate insurance risk through the reinsurance market. Even industrial firms have started exploring the possible transformation of some actuarial risk into financial risk through the theory of real options. An introduction to real options can be found in Hull (2012, Section 34) and Dixit and Pindyck (1994). A useful categorization to make in risk management techniques that I will sometimes make use of, following Gumerlock (1999), is to distinguish between risk management through risk aggregation and risk management through risk decomposition. Risk aggregation attempts to reduce risk by creating portfolios of less than completely correlated risk, thereby achieving risk reduction through diversification. Risk decomposition attempts to reduce a risk that cannot directly be priced in the market by analyzing it into subcomponents, all or some of which can be priced in the market. Actuarial risk can generally be managed only through risk aggregation, whereas financial risk utilizes both techniques. Chapter 7 concentrates on risk aggregation, while Chapter 8 primarily focuses on risk decomposition; Chapter 6 addresses the integration of the two. 1.3 SIMULATION AND SUBJECTIVE JUDGMENT Nobody can guarantee that all possible future contingencies have been provided for—this is simply beyond human capabilities in a world filled with uncertainty. But it is unacceptable to use that platitude as an excuse for complacency and lack of meaningful effort. It has become an embarrassment to the financial industry to see the number of events that are declared “once in a millennium” occurrences, based on an analysis of historical data, when they seem in fact to take place every few years. At one point I suggested, only half‐jokingly, that anyone involved in risk management who used the words perfect and storm in the same sentence should be permanently Introduction 5 banned from the financial industry. More seriously, everyone involved in risk management needs to be aware that historical data has a limited utility, and that subjective judgment based on experience and careful reasoning must supplement data analysis. The failure of risk managers to apply critical subjective judgment as a check on historical data in the period leading to the crisis of 2007–2008 is addressed in Section 5.2.5. This by no means implies that historical data should not be utilized. Historical data, at a minimum, supplies a check against intuition and can be used to help form reasoned subjective opinions. But risk managers concerned with protecting a firm against infrequent but plausible outcomes must be ready to employ subjective judgment. Let us illustrate with a simple example. Suppose you are trying to describe the distribution of a variable for which you have a lot of historical data that strongly supports a normal distribution with a mean of 5 percent and standard deviation of 2 percent. Suppose you suspect that there is a small but nonnegligible possibility that there will be a regime change that will create a very different distribution. Let’s say you guess there is a 5 percent chance of this distribution, which you estimate as a normal distribution with a mean of 0 percent and standard deviation of 10 percent. If all you cared about was the mean of the distribution, this wouldn’t have much impact—lowering the mean from 5 percent to 4.72 percent. Even if you were concerned with both mean and standard deviation, it wouldn’t have a huge impact: the standard deviation goes up from 2 percent to 3.18 percent, so the Sharpe ratio (the ratio of mean to standard deviation often used in financial analysis) would drop from 2.50 to 1.48. But if you were concerned with how large a loss you could have 1 percent of the time, it would be a change from a gain of 0.33 percent to a loss of 8.70 percent. Exercise 1.1 will allow you to make these and related calculations for yourself using the Excel spreadsheet MixtureOfNormals supplied on the book’s website. This illustrates the point that when you are concerned with the tail of the distribution you need to be very concerned with subjective probabilities and not just with objective frequencies. When your primary concern is just the mean—or even the mean and standard deviation, as might be typical for a mutual fund—then your primary focus should be on choosing the most representative historical period and on objective frequencies. While this example was drawn from financial markets, the conclusions would look very similar if we were discussing an actuarial risk problem like nuclear safety and we were dealing with possible deaths rather than financial losses. The fact that risk managers need to be concerned with managing against extreme outcomes would again dictate that historical frequencies need to be supplemented by informed subjective judgments. This reasoning 6 FINANCIAL RISK MANAGEMENT is very much in line with the prevailing (but not universal) beliefs among academics in the fields of statistics and decision theory. A good summary of the current state of thinking in this area is to be found in Hammond, Keeney, and Raiffa (1999, Chapter 7). Rebonato (2007) is a thoughtful book‐length treatment of these issues from an experienced and respected financial risk manager that reaches conclusions consistent with those presented here (see particularly Chapter 8 of Rebonato). The importance of extreme events to risk management has two other important consequences. One is that in using historical data it is necessary to pay particular attention to the shape of the tail of the distribution; all calculations must be based on statistics that take into account any nonnormality displayed in the data, including nonnormality of correlations. The second consequence is that all calculations must be carried out using simulation. The interaction of input variables in determining prices and outcomes is complex, and shortcut computations for estimating results work well only for averages; as soon as you are focused on the tails of the distribution, simulation is a necessity for accuracy. The use of simulation based on both historical data and subjective judgment and taking nonnormality of data into account is a repeated theme throughout this book—in the statement of general principles in Section 6.1.1, applied to more liquid positions throughout Chapter 7, applied to positions involving actuarial risk in Section 8.4, and applied to specific risk management issues throughout Chapters 9 through 14. EXERCISE 1.1 The Impact of Nonnormal Distributions on Risk Use the MixtureOfNormals spreadsheet to reproduce the risk statistics shown in Section 1.3 (you will not be able to reproduce these results precisely, due to the random element of Monte Carlo simulation, but you should be able to come close). Experiment with raising the probability of the regime change from 5 percent to 10 percent or higher to see the sensitivity of these risk statistics to the probability you assign to an unusual outcome. Experiment with changes in the mean and standard deviation of the normal distribution used for this lower‐probability event to see the impact of these changes on the risk statistics. CHAPTER 2 Institutional Background financial firm is, among other things, an institution that employs the talents of a variety of different people, each with her own individual set of talents and motivations. As the size of an institution grows, it becomes more difficult to organize these talents and motivations to permit the achievement of common goals. Even small financial firms, which minimize the complexity of interaction of individuals within the firm, must arrange relationships with lenders, regulators, stockholders, and other stakeholders in the firm’s results. Since financial risk occurs in the context of this interaction between individuals with conflicting agendas, it should not be surprising that corporate risk managers spend a good deal of time thinking about organizational behavior or that their discussions about mathematical models used to control risk often focus on the organizational implications of these models. Indeed, if you take a random sample of the conversations of senior risk managers within a financial firm, you will find as many references to moral hazard, adverse selection, and Ponzi scheme (terms dealing primarily with issues of organizational conflict) as you will find references to delta, standard deviation, and stochastic volatility. For an understanding of the institutional realities that constitute the framework in which risk is managed, it is best to start with the concept of moral hazard, which lies at the heart of these conflicts. A 2.1 MORAL HAZARD—INSIDERS AND OUTSIDERS The following is a definition of moral hazard taken from Kotowitz (1989): Moral hazard may be defined as actions of economic agents in maximizing their own utility to the detriment of others, in situations where they do not bear the full consequences or, equivalently, 7 8 FINANCIAL RISK MANAGEMENT do not enjoy the full benefits of their actions due to uncertainty and incomplete or restricted contracts which prevent the assignment of full damages (benefits) to the agent responsible. . . . Agents may possess informational advantages of hidden actions or hidden information or there may be excessive costs in writing detailed contingent contracts. . . . Commonly analyzed examples of hidden actions are workers’ efforts, which cannot be costlessly monitored by employers, and precautions taken by the insured to reduce the probability of accidents and damages due to them, which cannot be costlessly monitored by insurers. . . . Examples of hidden information are expert services—such as physicians, lawyers, repairmen, managers, and politicians. In the context of financial firm risk, moral hazard most often refers to the conflict between insiders and outsiders based on a double‐edged asymmetry. Information is asymmetrical—the insiders possess superior knowledge and experience. The incentives are also asymmetrical—the insiders have a narrower set of incentives than the outsiders have. This theme repeats itself at many levels of the firm. Let’s begin at the most basic level. For any particular group of financial instruments that a firm wants to deal in, whether it consists of stocks, bonds, loans, forwards, or options, the firm needs to employ a group of experts who specialize in this group of instruments. These experts will need to have a thorough knowledge of the instrument that can rival the expertise of the firm’s competitors in this segment of the market. Inevitably, their knowledge of the sector will exceed that of other employees of the firm. Even if it didn’t start that way, the experience gained by day‐to‐day dealings in this group of instruments will result in information asymmetry relative to the rest of the firm. This information asymmetry becomes even more pronounced when you consider information relative to the particular positions in those instruments into which the firm has entered. The firm’s experts have contracted for these positions and will certainly possess a far more intimate knowledge of them than anyone else inside or outside the firm. A generic name used within financial firms for this group of experts is the front office. A large front office may be divided among groups of specialists: those who negotiate transactions with clients of the firm, who are known as salespeople, marketers, or structurers; those who manage the positions resulting from these negotiated transactions, who are known as traders, position managers, or risk managers; and those who produce research, models, or systems supporting the process of decision making, who are known as researchers or technologists. However, this group of experts still requires the backing of the rest of the firm in order to be able to generate revenue. Some of this dependence Institutional Background 9 may be a need to use the firm’s offices and equipment; specialists in areas like tax, accounting, law, and transactions processing; and access to the firm’s client base. However, these are services that can always be contracted for. The vital need for backing is the firm’s ability to absorb potential losses that would result if the transactions do not perform as expected. A forceful illustration of this dependence is the case of Enron, which in 2001 was a dominant force in trading natural gas and electricity, being a party to about 25 percent of all trades executed in these markets. Enron’s experts in trading these products and the web‐enabled computer system they had built to allow clients to trade online were widely admired throughout the industry. However, when Enron was forced to declare bankruptcy by a series of financing and accounting improprieties that were largely unrelated to natural gas and electricity trading, their dominance in these markets was lost overnight. Why? The traders and systems that were so widely admired were still in place. Their reputation may have been damaged somewhat based on speculation that the company’s reporting was not honest and its trading operation was perhaps not as successful as had been reported. However, this would hardly have been enough to produce such a large effect. What happened was an unwillingness of trading clients to deal with a counterparty that might not be able to meet its future contractual obligations. Without the backing of the parent firm’s balance sheet, its stockholder equity, and its ability to borrow, the trading operation could not continue. So now we have the incentive asymmetry to set off the information asymmetry. The wider firm, which is less knowledgeable in this set of instruments than the group of front‐office experts, must bear the full financial loss if the front office’s positions perform badly. The moral hazard consists of the possibility that the front office may be more willing to risk the possibility of large losses in which it will not have to fully share in order to create the possibility of large gains in which it will have a full share. And the rest of the firm may not have sufficient knowledge of the front office’s positions, due to the information asymmetry, to be sure that this has not occurred. What are some possible solutions? Could a firm just purchase an insurance contract against trading losses? This is highly unlikely. An insurance firm would have even greater concerns about moral hazard because it would not have as much access to information as those who are at least within the same firm, even if they are less expert. Could the firm decide to structure the pay of the front office so that it will be the same no matter what profits are made on its transactions, removing the temptation to take excessive risk to generate potential large gains? The firm could, but experience in financial firms strongly suggests the need for upside participation as an incentive to call forth the efforts needed to succeed in a highly competitive environment. 10 FINANCIAL RISK MANAGEMENT Inevitably, the solution seems to be an ongoing struggle to balance the proper incentive with the proper controls. This is the very heart of the design of a risk management regime. If the firm exercises too little control, the opportunities for moral hazard may prove too great. If it exercises too much control, it may pass up good profit opportunities if those who do not have as much knowledge as the front office make the decisions. To try to achieve the best balance, the firm will employ experts in risk management disciplines such as market risk, credit risk, legal risk, and operations risk. It will set up independent support staff to process the trades and maintain the records of positions and payments (the back office); report positions against limits, calculate the daily profit and loss (P&L), and analyze the sources of P&L and risk (the middle office); and take responsibility for the accuracy of the firm’s books and records (the finance function). However, the two‐sided asymmetry of information and incentive will always exist, as the personnel in these control and support functions will lack the specialized knowledge that the front office possesses in their set of instruments. The two‐sided asymmetry that exists at this basic level can be replicated at other levels of the organization, depending on the size and complexity of the firm. The informational disadvantage of the manager of fixed‐income products relative to the front office for European bonds will be mirrored by the informational disadvantage of the manager of all trading products relative to the manager of fixed‐income products and the firm’s CEO relative to the manager of all trading products. Certainly, the two‐sided asymmetry will be replicated in the relationship between the management of the firm and those who monitor the firm from the outside. Outside monitors primarily represent three groups—the firm’s creditors (lenders and bondholders), the firm’s shareholders, and governments. All three of these groups have incentives that differ from the firm’s management, as they are exposed to losses based on the firm’s performance in which the management will not fully share. The existence of incentive asymmetry for creditors is reasonably obvious. If the firm does well, the creditors get their money back, but they have no further participation in how well the firm performs; if the firm does very badly and goes bankrupt, the creditors have substantial, possibly even total, loss of the amount lent. By contrast, the firm’s shareholders and management have full participation when the firm performs well, but liability in bankruptcy is limited to the amount originally invested. When we examine credit risk in Section 13.2.4, this will be formally modeled as the creditors selling a put option on the value of the firm to the shareholders. Since all options create nonlinear (hence asymmetric) payoffs, we have a clear source of incentive asymmetry for creditors. Institutional Background 11 It is less clear whether incentive asymmetry exists for shareholders. In principle, their interests are supposed to be exactly aligned with those of the firm’s management, and incentives for management based on stock value are used to strengthen this alignment. In practice, it is always possible that management will take more risk than shareholders would be completely comfortable with in the hope of collecting incentive‐based compensation in good performance years that does not have to be returned in bad performance years. Kotowitz (1989) quotes Adam Smith from Wealth of Nations: “The directors of such companies, however, being managers rather of other people’s money than of their own, it cannot well be expected, that they should watch over it with the same anxious vigilance with which the partners in a private company frequently watch over their own.” Government involvement arises from the asymmetric dangers posed to the health of the overall economy by the failure of a financial firm. If an implicit government guarantee is given to rescue large financial firms from bankruptcy (the notion of “too big to fail”), then moral hazard is created through management’s knowledge that it can try to create profit opportunities, in which the government has only limited participation through taxes, by taking risks of losses that will need to be fully absorbed by the government. If the government is not willing to prevent the failure of large financial firms, then it will want to place restrictions on the externalities that those firms can create by not having to bear their share of the cost to the overall economy of a firm’s potential bankruptcy. In all three cases of moral hazard involving outside monitors, the information asymmetry is even more severe than when the information asymmetry takes place wholly inside the firm. Senior management and its risk monitors are at least on the premises, are involved in day‐to‐day business with more junior managers, and can utilize informal measures, such as the rotation of managers through different segments of the firm, to attempt to diffuse both incentives and knowledge. Outside monitors will have only occasional contact with the firm and must rely mostly on formal requirements to obtain cooperation. Let us look at some of the outside monitors that creditors, shareholders, and governments rely on: ■ ■ In addition to their own credit officers, creditors rely on rating agencies such as Moody’s Investors Service and Standard & Poor’s (S&P) to obtain information about and make judgments on the creditworthiness of borrowers. Shareholders and creditors rely on investment analysts working for investment bankers and brokerage firms to obtain information about and make judgments on the future earnings prospects and share values of 12 ■ ■ FINANCIAL RISK MANAGEMENT firms. Although neither rating agencies nor investment analysts have any official standing with which to force cooperation from the firms they analyze, their influence with lenders and investors in bonds and stocks gives them the leverage to obtain cooperation and access to information. Governments can use their regulatory powers to require access to information from financial firms and employ large staffs to conduct examinations of the firms. For example, for the U.S. government, the Federal Reserve System and the Comptroller of the Currency conduct examinations of commercial banks. A similar function is performed by the Securities and Exchange Commission (SEC) for investment banks. Creditors, shareholders, and governments all rely on independent accounting firms to conduct audits of the reliability of the financial information disclosures that are required of all publicly held firms. Over the years, many critical questions have been raised about how truly independent the judgment of these outside monitors really is: ■ ■ ■ ■ Credit rating agencies have been accused of being too slow to downgrade ratings in response to adverse changes in a firm’s financial condition because their source of revenue comes from the firms whose debt they rate. Similarly, independent auditors have been suspected of being too deferential to the firms they monitor since these firms are the ones who pay their audit fees and hire them for consulting services. The fear is that the desire for more revenue will blunt objections to companies choosing accounting methods that cast their results in a favorable light. Investment banks have a built‐in conflict of interest from competing for the business of the firms whose performance their investment analysts are monitoring. It has long been noted that analysts’ buy recommendations far outnumber sell recommendations. Accusations have been leveled that government regulatory agencies are more concerned with protecting the interests of the firms being monitored than with protecting the public interest. These charges have particular force when personnel flow freely between employment in the regulatory agencies and in the firms they regulate. All of these criticisms seemed to be coming to a head in 2002 amid the scandals involving the now‐defunct auditing firm of Arthur Andersen, Enron’s declaration of bankruptcy only a week after being rated investment grade, and the massive declines in the stock values of technology firms highly touted by investment analysts. Some useful reforms have been undertaken, such as forbidding auditing firms to sell consulting services to firms Institutional Background 13 they audit and not allowing the bonuses of investment analysts to be tied to investment banking fees collected from clients whose stocks they cover. However, the basic sources of conflict of interest remain, and investors and lenders will continue to need to employ a skeptical filter when utilizing input from outside monitors. Although the conflicts between insiders and outsiders due to the two‐ sided asymmetry of moral hazard cannot be eliminated, a frank understanding by both sides can lead to a cooperative relationship. In a cooperative relationship, insiders will acknowledge the need to have outsiders exercise controls and will voluntarily share information and knowledge with outsiders. In a cooperative relationship, outsiders will acknowledge their need to learn from the insiders and will ease controls in response to a track record of openness, although both must recognize the need to always have some level of controls (the ancient folk wisdom states that “I trust my grandmother, but I still cut the cards when she deals”). A lack of understanding of moral hazard can lead to an uncooperative relationship fueled by mutual resentments between an insider, such as a trader or structurer, with an outsider, such as a corporate risk manager or regulator. An insider who does not understand the purely situational need to have someone less knowledgeable “look over my shoulder” will attribute it to an insulting lack of personal trust, an arrogant assumption of more knowledge than the other possesses, or a simple desire by the outsider to create a job or grab power (which is not to say that some of these motivations do not exist in reality, mixed in with the need to control moral hazard). The insider’s response will then probably be to withhold information, obfuscate, and mislead, which will drive the outsider to even closer scrutiny and more rigid controls, which is clearly a prescription for a vicious circle of escalation. An outsider who lacks an understanding of the situation may defensively try to pretend to have more knowledge than he actually has or may denigrate the knowledge of the insider, which will only exacerbate any suspicions of the process the insider has. Moral hazard has long been a key concept in the analysis of insurance risks. A typical example would be an insurance company’s concern that an individual who has purchased insurance against auto theft will not exercise as much care in guarding against theft (for example, parking in a garage rather than on the street) as one who has not purchased insurance. If the insurance company could distinguish between individuals who exercise extra care and those who don’t, it could sell separate contracts to the two types of individuals and price the extra losses into just the type sold to those exercising less care. However, the information advantage of an individual monitoring his own degree of care relative to the insurance company’s ability to monitor it makes this prohibitively expensive. So the insurance company 14 FINANCIAL RISK MANAGEMENT needs to settle for cruder measures, such as establishing a deductible loss that the insured person must pay in the event of theft, thereby aligning the interests of the insured more closely with the insurer. It has become increasingly common for moral hazard to be cited in analyses of the economics of firms in general, particularly in connection with the impact of the limited liability of shareholders willing to take larger gambles. The shareholders know that if the gamble succeeds, they will avoid bankruptcy and share in the profits, but will suffer no greater loss in a large bankruptcy than in a smaller one. To quote W. S. Gilbert: You can’t embark on trading too tremendous, It’s strictly fair and based on common sense, If you succeed, your profits are stupendous, And if you fail, pop goes your eighteen pence. (from Gilbert and Sullivan’s Utopia, Limited) A firm’s creditors can exercise some control over their actions and might be able to forbid such gambles, assuming they have sufficient knowledge of the nature of the firm’s investments. This is where the informational advantage of the managers over the creditors with respect to the firm’s investments comes in. What sort of actions can we expect from a trader based on the concept of moral hazard? We can certainly expect that the trader may have a different degree of risk aversion than the firm’s management, since traders’ participation in favorable results exceeds their participation in downside results. Taleb (1997, 66) refers to this as the trader “owning an option on his profits” and states that in such circumstances “it is always optimal to take as much risk as possible. An option is worth the most when volatility is highest.” This will probably become even more noticeable if the trader has been having a poor year. Knowing that she is headed toward a minimal bonus and possible dismissal may incline the trader to swing for the fences and take a large risk. The trader knows that if the risk turns out favorably, it might be enough to reverse previous losses and earn a bonus. If it turns out poorly, then “you can’t get less than a zero bonus” and “you can’t get fired twice.” (You can damage your reputation in the industry, but sharing information about a trader’s track record between competitor firms cannot be done that efficiently—more information asymmetry.) For this reason, firms may severely cut the trading limits of a trader having a poor year. Beyond the differences in risk aversion, moral hazard can even result in the perverse behavior (for the firm) of having a trader willing to increase risk Institutional Background 15 exposure when faced with a lower expected return. Consider the following advice to traders from Taleb (1997, 65): How aggressive a trader needs to be depends highly on his edge, or expected return from the game: ■ ■ When the edge is positive (the trader has a positive expected return from the game, as is the case with most market makers), it is always best to take the minimum amount of risk and let central limit slowly push the position into profitability. This is the recommended method for market makers to progressively increase the stakes, in proportion to the accumulated profits. In probability terms, it is better to minimize the volatility to cash‐in on the drift. When the edge is negative, it is best to be exposed as little as possible to the negative drift. The operator should optimize by taking as much risk as possible. Betting small would ensure a slow and certain death by letting central limit catch up on him. The mathematics and economic incentives that this advice is based on are certainly sound. It is advice that is known to every gambler (or ought to be) and is well founded in statistical theory. When the odds are in your favor, place many small bets; when the odds are against you, place one large bet. Essentially, when the odds are against you, you are attempting to minimize the length of time you are playing against the house since you are paying a tax, in the form of an expected loss, for the privilege of playing. However, although this makes perfect economic sense from the viewpoint of the individual trader, it is hardly the strategy the firm employing these traders would want to see them follow. The firm, whose P&L will be the sum of the results of many traders, would like to see traders with a negative expected return not take any positions at all rather than have these be the traders taking on the most risk. To the extent the firm’s management can figure out which traders have a negative edge, it will restrict their risk taking through limits and the replacement of personnel. However, the individual traders have the information advantage in knowing more than the firm about their expected returns. They also have the asymmetrical incentive to take larger risks in this case, even though doing so will probably hurt the firm. The traders will not derive much benefit from the firm doing well if they do not contribute to that result, but they will benefit if they do increase their risk and win against the odds. Moral hazard helps to explain the valuation that investors place on the earnings volatility of financial firms. You could argue that firms should 16 FINANCIAL RISK MANAGEMENT worry just about the expected value and not about volatility, since the market should place a risk premium only on risk that it cannot hedge away (an investor who wants less risk will just take the stock with the highest expected return and diversify by mixing with government bonds). However, empirical evidence shows that the market places a stiff discount on variable trading earnings. The reason may be information asymmetry. It is hard for outsiders to tell whether a firm is taking sound gambles to maximize expected value or is maximizing its insiders’ option on one‐way bets. Perold (1998) states: I view financial intermediaries as being special in several ways: First, these firms are in credit‐sensitive businesses, meaning that their customers are strongly risk‐averse with respect to issuer default on contractually promised payoffs. (For example, policyholders are averse to having their insurance claims be subject to the economic performance of the issuing firm, and strictly prefer to do business with a highly rated insurer.) The creditworthiness of the intermediary is crucial to its ability to write many types of contracts, and contract guarantees feature importantly in its capital structure. Second, financial firms are opaque to outsiders. They tend to be in businesses that depend vitally on proprietary financial technology and that cannot be operated transparently. In addition, the balance sheets of financial firms tend to be very liquid, and are subject to rapid change. Financial firms, thus, are difficult to monitor, and bear significant deadweight costs of capital. Guarantors face costs related to adverse selection and moral hazard. . . . Third, financial firms are also internally opaque. Information tends to be private at the business unit level, or even at the level of individual employees such as traders. Efficient management of these firms thus involves significant use of performance‐related compensation to mitigate against monitoring difficulty. Moral hazard can create a battleground over information between insiders and outsiders. Insiders are fearful that any information obtained by outsiders will be used as a tool to tighten controls over insiders’ actions. Insiders can be expected to have an inherent bias against tighter controls, partly because narrowing the range of actions available leads to suboptimal solutions and partly because incentive asymmetry makes riskier action more rewarding to insiders than to outsiders. One of the most common ways in which insiders can mislead outsiders about the need for controls is termed a Ponzi scheme. Institutional Background 2.2 17 PONZI SCHEMES In its original meaning, a Ponzi scheme is a criminal enterprise in which investors are tricked into believing that they will receive very high returns on their investments, but the early investors are paid out at high rates of return only with the payments coming from the cash invested by later investors. The illusion of high returns can be pretty convincing. After all, you can actually see the early investors receiving their high returns in cash, and the con men running these schemes can produce very plausible lies about the purported source of the returns. As a result, the pace of new investment can be intense, enabling the illusion of profit to be maintained over a fairly long time period. It’s a vicious cycle—the eagerness of new investors to place money in the scheme leads to the heightened ability to make investments appear highly profitable, which leads to even greater eagerness of new investors. However, ultimately, any Ponzi scheme must collapse, as there is no ultimate source of investment return (in fact, investment return is quite negative, as the flow of new investment must also be partially diverted to the criminals profiting from it). Ponzi schemes are also sometimes called pyramid schemes and bear a close resemblance to chain letter frauds. When I wrote the immediately preceding paragraph for the first edition of this book in 2003, I felt the need to thoroughly explain what a Ponzi scheme is. Today, it is probably not necessary, as Bernie Madoff has regrettably given us all an exhaustive lesson in how a Ponzi scheme is run. The original meaning of Ponzi schemes has been broadened by risk managers to include situations in which firms are misled as to the profitability of a business line by the inadequate segregation of profits on newly acquired assets and returns on older assets. Let’s consider a typical example. Suppose a trading desk has entered into marketing a new type of path‐dependent option. The desk expects substantially more customer demand for buying these options than for selling them. They intend to manage the resulting risk with dynamic hedging using forwards and more standard options. As we will see when discussing path‐dependent options in Section 12.3, it is very difficult to try to estimate in advance how successful a dynamic hedging strategy for path‐dependent options will be. In such circumstances, the pricing of the option to the client must be based on an estimate of the future cost of the dynamic hedging, applying some conservatism to try to cover the uncertainty. Let’s assume that a typical trade has a seven‐year maturity, and that the customer pays $8 million and the firm pays $5 million to purchase the initial hedge. Of the remaining $3 million, we’ll assume that the desk is estimating dynamic hedging costs of $1 million over the two years, but the uncertainty of these costs leads to 18 FINANCIAL RISK MANAGEMENT setting up a $2 million initial allowance (or reserve) to cover the hedging costs, leaving $1 million to be booked as up‐front profit. Suppose the trading desk has made a serious error in predicting the hedging costs, and the hedging costs actually end up around $5 million, leading to a net loss of $2 million on every transaction booked. You may not be able to do anything about deals already contracted, but you would at least hope to get feedback from the losses encountered on these deals in time to stop booking new deals or else raise your price to a more sustainable level. This should happen if P&L reporting is adequately detailed, so you can see the losses mounting up on the hedging of these trades (this is called hedge slippage). However, it is often difficult to keep track of exactly how to allocate a day’s trading gains and losses to the book of deals being hedged. You want to at least know that trading losses are occurring so you can investigate the causes. The most severe problem would be if you didn’t realize that trades were losing money. How could this happen? If P&L reporting is not adequately differentiated between the existing business and new business, then the overall trading operation can continue to look profitable by just doing enough new business. Every time a new deal is booked, $1 million goes immediately into P&L. Of course, the more deals that are booked, the larger the hedging losses that must be overcome, so even more new trades are needed to swamp the hedging losses. The resemblance to a Ponzi scheme should now be obvious. One key difference is that in its original meaning, the Ponzi scheme is a deliberate scam. The financial situation described is far more likely to arise without any deliberate intent. However, those in the front office, based on their close knowledge of the trading book, will often suspect that this situation exists before any outsiders do, but may not want to upset the apple cart. They would be jeopardizing bonuses that can be collected up front on presumed earnings. They may also be willing to take the risk that they can find a way to turn the situation around based on their greater participation in future upside than future downside. They may choose to hide the situation from outsiders who they suspect would not give them the latitude to take such risks. So moral hazard can turn an accidentally originated Ponzi scheme into one that is very close to deliberate. As a historical footnote, the Ponzi scheme derives its name from Charles Ponzi, a Boston‐based swindler of the 1920s (though it was not the first Ponzi scheme—William “520 Percent” Miller ran one in Brooklyn around 1900; an excellent 1905 play by Harley Granville‐Barker, The Vosey Inheritance, which has been revived frequently over the past decade, revolves around a lawyer specializing in trusts and estates trying to train his son to take over Institutional Background 19 the management of his Ponzi scheme). The following account of Charles Ponzi is drawn from Sifakis (1982): [Ponzi] discovered he could buy up international postal‐union reply coupons at depressed prices and sell them in the United States at a profit up to 50 percent. It was, in fact, a classic get‐rich‐slowly operation, and as such, it bored Ponzi. So he figured out a better gimmick. Ponzi figured out that telling people he was making the money and how he could make it was just as good as actually making it. He advertised a rate of return of 50 percent in three months. It was an offer people couldn’t refuse, and money started to come rolling in. When Ponzi actually started paying out interest, a deluge followed. On one monumental day in 1920, Ponzi’s offices took in an incredible $2 million from America’s newest gamblers, the little people who squeezed money out of bank accounts, mattresses, piggy banks, and cookie jars. There were days when Ponzi’s office looked like a hurricane had hit it. Incoming cash had to be stuffed in closets, desk drawers and even wastebaskets. Of course, the more that came in, the more Ponzi paid out. As long as new funds were coming in, Ponzi could continue to make payments. However, as with all pyramid schemes, the bubble had to burst. A newspaper published some damaging material about his past, including time spent in prison. New investors started to hesitate. Ponzi’s fragile scheme collapsed, since it required an unending flow of cash. His books, such as they were, showed a deficit of somewhere between $5 and $10 million, or perhaps even more. No one ever knew for sure. 2.3 ADVERSE SELECTION Let’s return to the situation described previously. Suppose our accounting is good enough to catch the hedge slippage before it does too much damage. We stop booking new deals of this type, but we may find we have booked a disturbingly large number of these deals before the cutoff. If our customers have figured out the degree to which we are underpricing the structure before we do, then they may try to complete as many deals as they can before we wise up. This pattern has frequently been seen in the financial markets. For 20 FINANCIAL RISK MANAGEMENT example, the last firms that figured out how to correctly price volatility skew into barrier options found that their customers had loaded up on trades that the less correct models were underpricing. A common convention is to label this situation as adverse selection as a parallel to a similar concern among insurance firms, which worry that those customers with failing health will be more eager to purchase insurance than those with better health, taking advantage of the fact that a person knows more about his own health than an insurance company can learn (Wilson 1989). So adverse selection is like moral hazard since it is based on information asymmetry; the difference is that moral hazard is concerned with the degree of risk that might be taken based on this asymmetry, whereas adverse selection is concerned with a difference in purchasing behavior. In 2001, George Ackerlof, Michael Spence, and Joseph Stiglitz won the Nobel Prize in economics for their work on adverse selection and its application to a broad class of economic issues. Concern about the risk from adverse selection motivates risk managers’ concern about the composition of a trading desk’s customer base. The key question is: What proportion of trades is with counterparties who are likely to possess an informational advantage relative to the firm’s traders? As a general rule, you prefer to see a higher proportion of trades with individuals and nonfinancial corporations that are likely trading to meet hedging or investment needs rather than seeking to exploit informational advantage. Alarm is raised when an overwhelming proportion of trades is with other professional traders, particularly ones who are likely to see greater deal flow or have a greater proportion of trades with individuals and nonfinancial corporations than your firm’s traders. Seeing greater deal flow can give a firm an informational advantage by having a more accurate sense of supply‐and‐demand pressures on the market. A greater proportion of customers who are not professional traders yields two further potential informational advantages: 1. At times you work with such customers over a long period of time to structure a large transaction. This gives the traders advance knowledge of supply and demand that has not been seen in the market yet. 2. Working on complex structures with customers gives traders a more intimate knowledge of the structure’s risks. They can choose to retain those risks that this knowledge shows them are more easily manageable and attempt to pass less manageable risks on to other traders. Traders may tend to underestimate the degree to which their profitability is due to customer deal flow and overestimate the degree to which it is due to anticipating market movements. This can be dangerous if it encourages them to aggressively take risks in markets in which they do not possess this customer flow advantage. A striking example I once observed was a foreign Institutional Background 21 exchange (FX) trader who had a phenomenally successful track record of producing profits at a large market‐making firm. Convinced of his prowess in predicting market movements, he accepted a lucrative offer to move to a far smaller firm. He was back at his old job in less than year, confessing he simply had not realized how much of his success was due to the advantages of customer deal flow. A pithy, if inelegant, statement of this principle was attributed to the head of mortgage‐backed trading at Kidder Peabody: “We don’t want to make money trading against smart traders; we want to make money selling to stupid customers.” Of course, stupid needs to be understood here as macho Wall Street lingo for informationally disadvantaged. It’s the sort of talk that is meant to be heard only in locker rooms and on trading floors. An unfriendly leak resulted in his quote appearing on the front page of the Wall Street Journal. It is delightful to imagine the dialogue of some of his subsequent conversations with the firm’s customers. 2.4 THE WINNER’S CURSE In response to the risks of adverse selection, traders may exhibit confidence that this is not something they need to worry about. After all, adverse selection impacts only those with less knowledge than the market. It is a rare trader who is not convinced that she possesses far more knowledge than the rest of the market—belief in one’s judgment is virtually a necessity for succeeding in this demanding profession. Whether the firm’s management shares the trader’s confidence may be another story. However, even if it does, the trader must still overcome another hurdle—the winner’s curse, the economic anomaly that says that in an auction, even those possessing (insider) knowledge tend to overpay. The winner’s curse was first identified in conjunction with bidding for oil leases, but has since been applied to many other situations, such as corporate takeovers. My favorite explanation of the mechanism that leads to the winner’s curse comes from Thaler (1992): Next time you find yourself a little short of cash for a night on the town, try the following experiment in your neighborhood tavern. Take a jar and fill it with coins, noting the total value of the coins. Now auction off the jar to the assembled masses at the bar (offering to pay the winning bidder in bills to control for penny aversion). Chances are very high that the following results will be obtained: 1. The average bid will be significantly less than the value of the coins. (Bidders are risk averse.) 22 FINANCIAL RISK MANAGEMENT 2. The winning bid will exceed the value of the jar. In conducting this demonstration, you will have simultaneously obtained the funding necessary for your evening’s entertainment and enlightened the patrons of the tavern about the perils of the winner’s curse. When applied to trading, the winner’s curse is most often seen in market making for less liquid products, where opinions on the true value of a transaction may vary more widely. Market makers are in competition with one another in pricing these products. The firm that evaluates a particular product as having a higher value than its competition is most likely to be winning the lion’s share of these deals. Consider a market for options on stock baskets. As we will discuss in Section 12.4, a liquid market rarely exists for these instruments, so pricing depends on different estimates of correlation between stocks in a basket. The firm that has the lowest estimate for correlation between technology stocks will wind up with the most aggressive bids for baskets of technology stocks and will book a large share of these deals. Another firm that has the lowest estimate for correlation between financial industry stocks will book the largest share of those deals. An anecdotal illustration comes from Neil Chriss. When Chriss was trading volatility swaps at Goldman Sachs, they would line up five or six dealers to give them quotes and would always hit the highest bid or lift the lowest offer. The dealers knew they were doing this and were very uneasy about it, limiting the size of trades they would accommodate. One dealer, on winning a bid, told Chriss, “I am always uncomfortable when I win a trade with you, as I know I was the best bid on top of five other smart guys. What did I do wrong?” Adverse selection can be controlled by gaining expertise and increasing the proportion of business done with ultimate users rather than with other market makers. However, the winner’s curse can be controlled only by either avoiding auction environments or adequately factoring in a further pricing conservatism beyond risk aversion. It provides a powerful motivation for conservatism in pricing and recognizing profits for those situations such as one‐way markets (see Section 6.1.3) in which it is difficult to find prices at which risks can be exited. We demonstrate the mechanism of the winner’s curse with a simple numerical example involving a market with only three firms, two buyers, and one seller. The results are shown in Table 2.1. We consider two different situations. In the first, direct negotiation occurs on the price between the seller and a single buyer. In the second, both buyers participate in an auction. 23 Institutional Background TABLE 2.1 Deal The Winner’s Curse Buyer 1 Buyer 2 Auction P&L Actual Value Seller’s Offer Bid P&L Bid P&L Buyer 1 Buyer 2 1 1.56 2.10 2.00 0.00 2.40 −0.69 0.00 −0.84 2 2.66 1.40 1.50 1.21 1.80 1.06 0.00 0.86 3 3.16 3.20 2.10 0.00 3.10 0.00 0.00 0.06 4 1.96 3.40 1.60 0.00 2.20 0.00 0.00 −0.24 5 1.36 1.90 1.70 0.00 1.20 0.00 −0.34 0.00 6 4.46 4.30 4.80 −0.09 3.00 0.00 −0.34 0.00 7 3.16 2.60 3.50 0.11 2.90 0.41 −0.34 0.00 8 1.96 1.80 1.40 0.00 2.10 0.01 0.00 −0.14 9 1.56 2.70 1.40 0.00 1.10 0.00 0.16 0.00 10 2.16 2.10 2.50 −0.14 2.70 −0.24 0.00 −0.54 Average 2.40 2.55 2.25 0.55 −0.86 −0.84 Total Correlation with actual value 2.25 1.09 63.2% 83.3% 72.2% Notes The column headed Actual Value shows what the deals are really worth. The columns headed Buyer 1 Bid and Buyer 2 Bid are the bid price of buyers of deals. The seller’s asking prices, in the column headed Seller’s Offer, are conservative (higher by .15 on average than true prices). The buyers are posting conservative bids (lower by .15 on average than true prices). The columns headed Buyer 1 P&L and Buyer 2 P&L are the buyers’ profits if they negotiate to the average of the seller’s asked and buyer’s bid. The columns headed Auction P&L Buyer 1 and Auction P&L Buyer 2 show the profits of the buyers in an auction. There are 10 transactions that the seller might sell to the buyers. Neither the buyers nor the seller is certain of the true value of these transactions (for example, they might depend on future dynamic hedging costs, which depend on the evolution of future prices, which different firms estimate using different probability distributions). After the fact, we know the true realized value of each transaction, as shown in column 2 of the 24 FINANCIAL RISK MANAGEMENT table. Buyer 1’s knowledge of this market is superior to buyer 2’s, and both have superior knowledge compared to the seller. This can be seen by the correlations between realized value and each party’s estimate of transaction value (83.3% for buyer 1, 72.2% for buyer 2, and 63.2% for the seller). The consequences of this informational advantage are that both buyer 1 and buyer 2 make a profit at the expense of the seller in direct negotiations, and that buyer 1’s profit in this situation is higher than buyer 2’s profit. In the direct negotiation situation, we assume that the buyer, being risk averse, has successfully biased his bids down to be on average lower than the realized value, and the seller, being risk averse, has successfully biased his asked prices up to be on average higher than realized value. We assume no transaction takes place if the buyer’s bid is lower than the seller’s asked. If the buyer’s bid exceeds the seller’s asked, we assume the transaction takes place at the average price between these two prices. As a result, buyer 1 has a total P&L of +1.09, and buyer 2 has a total P&L of +0.55. Now consider what happens in the auction when the buyers have to compete for the seller’s business, a situation very typical for market making firms that must offer competitive price quotations to try to win customer business from other market makers. The seller no longer relies on his own estimate of value, but simply does business at the better bid price between the two firms. Even though both firms continue to successfully bias their bids down on average from realized values, both wind up losing money in total, with buyer 1 having a P&L of –0.86 and buyer 2 having a P&L of –0.84. This is because they no longer have gains on trades that they seriously undervalued to balance out losses on trades that they seriously overvalued, since they tend to lose trades that they undervalue to the other bidder. This illustrates the winner’s curse. The spreadsheet WinnersCurse on the course website shows the consequences of changing some of the assumptions in this example. 2.5 MARKET MAKING VERSUS POSITION TAKING An important institutional distinction between participants in the financial markets that we will refer to on several occasions throughout this book is between market making and position taking: ■ Market making (also called book running or the sell side) consists of making two‐way markets by engaging in (nearly) simultaneous buying and selling of the same instruments, attempting to keep position Institutional Background ■ 25 holdings to a minimum and to profit primarily through the difference between (nearly) simultaneous buy and sell prices. Position taking (also called market using, price taking, speculation, or the buy side) consists of deliberately taking positions on one side or the other of a market, hoping to profit by the market moving in your favor between the time of purchase and the time of sale. Positions may be taken on behalf of a firm (in which case it is often labeled proprietary trading) or on behalf of an individual client or a group of clients, such as a mutual fund, hedge fund, or managed investment account. Some time lag nearly always occurs between the purchase and sale involved in market making. Depending on the length of time and degree of deliberate choice of the resulting positions, these may be labeled position‐taking aspects of market making. Market making almost always involves risk because you cannot often buy and sell exactly simultaneously. The market maker makes a guess on market direction by its posted price, but the bid‐ask spread can outweigh even a persistent error in directional guess as long as the error is small. (In Exercise 9.1, you’ll be asked to build a simulation to test out the degree to which this is true.) The experience and information gained from seeing so much flow means you most likely will develop the ability to be right on direction on average. However, the position taker has the advantage over the market maker of not needing to be in the market every day. Therefore, the position taker can stay away from the market except when possessed of a strong opinion. The market maker cannot do this; staying away from the market would jeopardize the franchise. The different objectives of market makers and position takers tend to be reflected in different attitudes toward the use of models and valuation techniques. A position taker generally uses models as forecasting tools to arrive at a best estimate of what a position will be worth at the conclusion of a time period tied to an anticipated event. The position taker will pay attention to the market price of the position during that time period to determine the best time to exit the position and to check whether new information is coming into the market. However, a position taker will generally not be overly concerned by prices moving against the position. Since the position taker is usually waiting for an event to occur, price movements prior to the time the event is expected are not that relevant. A frequently heard statement among position takers is: “If I liked the position at the price I bought it, I like it even better at a lower price.” By contrast, a market maker generally uses models to perform risk decomposition in order to evaluate alternative current prices at which a position can be exited. The market maker will pay close attention to current market prices as the key indicator of how quickly inventory can be reduced. 26 FINANCIAL RISK MANAGEMENT The direction in which prices will move over the longer term is of little concern compared to determining what price will currently balance supply and demand. An amusing analogy can be made to gambling on sports. Position takers correspond to the gamblers who place their bets based on an analysis of which team is going to win and by what margin. Market makers correspond to the bookmakers whose sole concern is to move the odds quoted to a point that will even out the amount bet on each side. The bookmaker’s concern is not over which team wins or loses, but over the evenness of the amounts wagered. Close to even amounts let the bookmakers come out ahead based on the spread or vigorish in the odds, regardless of the outcome of the game. Uneven amounts turn the bookmaker into just another gambler who will win or lose depending on the outcome of the game. As explained in Section 1.1, the focus of this book is on the active use of trading in liquid markets to manage risk. This view is more obviously aligned with market making than with position taking. In fact, the arbitrage‐based models that are so prominent in mathematical finance have been developed largely to support market making. Position takers tend more toward the use of econometric forecasting models. In Section 6.1.7, we will further discuss the issue of the extent to which position takers should adopt the risk management discipline that has been developed for market makers. Some authors distinguish a third type of financial market participant besides market makers and position takers—the arbitrageurs. I believe it is more useful to classify arbitrage trading as a subcategory of position taking. Pure arbitrage, in its original meaning of taking offsetting positions in closely related markets that generate a riskless profit, is rarely encountered in current financial markets, given the speed and efficiency with which liquid prices are disseminated. What is now labeled arbitrage is almost always a trade that offers a low but relatively certain return. The motivations and uses of models by those seeking to benefit from such positions are usually closely aligned with other position takers. A good example is merger arbitrage (sometimes misleadingly called risk arbitrage). Suppose that Company A and Company B have announced a forthcoming merger in which two shares of A’s stock will be traded for one share of B’s stock. If the current forward prices of these stocks to the announced merger date are $50 for A and $102 for B, an arbitrage position would consist of a forward purchase of two shares of A for $100 and a forward sale of one share of B for $102. On the merger date, the two shares of A purchased will be traded for one share of B, which will be delivered into the forward sale. This nets a sure $2, but only if the merger goes through as announced. If the merger fails, this trade could show a substantial loss. Institutional Background 27 Merger arbitrageurs are position takers who evaluate the probability of mergers breaking apart and study the size of loss that might result. They are prototypical forecasters of events with generally little concern for market price swings prior to the occurrence of the event. For further reading on the economics and institutional structure of market making and position taking, a book I would recommend very highly is Harris (2003). Anyone involved in risk management should attempt to gain insight into how risk management is viewed by traders. While friendship and conversation are the best way to approach this, it is also helpful to read about risk management from a trader’s perspective. The best book of this type I have encountered is Brown (2012). CHAPTER 3 Operational Risk perational risk is usually defined in the negative—it includes all of the risks that are not categorized as either market or credit risk. The industry does not yet have consensus on this terminology. Some firms use the term operational risk to cover a subset of the risks other than market and credit risk. For further discussion, see Jameson (1998a). Broadly speaking, these risks are the most difficult to quantify. One attempt at a more positive definition that has been gaining some currency has been made by the Basel Committee on Banking Supervision: “the risk of direct or indirect loss resulting from inadequate or failed internal processes, people, or systems, or from external events.” Another attempt would be to break apart risk into three pieces. View a financial firm as the sum total of all the contracts it enters into. The firm can suffer losses on the contracts in one of three ways: O 1. Obligations in contracts may be performed exactly as expected, but changes in economic conditions might make the sum of all contracted actions an undesired outcome. This is market risk. 2. The other parties to some of the contracts may fail to perform as specified. This is credit risk. 3. The firm may be misled about what the contracted actions are or the consequences of these actions. This is operational risk. Operational risk is virtually all risk that cannot be managed through the use of liquid markets, so, as argued in Chapter 1, it does not fall within the scope of financial risk management. In this way, it is very much like the risks traditionally managed by insurance companies. Indeed, one of the primary tools for managing operational risk is to try to buy protection from insurance companies, as we’ll discuss in Section 3.8. But, even though the financial risk management approach does not apply, these are risks that arise as 29 30 FINANCIAL RISK MANAGEMENT a result of trading and so are intertwined with financial risk management, justifying a quick survey of these issues in this book. Operational risk can be subdivided into the following categories: ■ ■ ■ ■ ■ ■ Operations risk is the risk that deficiencies in information systems or internal controls will result in unexpected loss. Operations risk can be further subdivided into the risk of fraud, risk of nondeliberate incorrect information, disaster risk, and personnel risk. Legal risk is the risk that the terms or conditions of a contract or agreement will prove unenforceable due to legal defects in the contract or in related documentation and procedures. Another type of legal risk is the risk that actions of the firm’s employees will have been found to be illegal and subject the firm to substantial penalties. Legal risk includes regulatory risk. Reputational risk is the risk that the enforcement of contract provisions will prove too costly in terms of damage to the firm’s reputation as a desirable firm for customers to do future business with. Accounting risk is the risk that an error in accounting practice will necessitate a restatement of earnings, which adversely affects the investors’ or customers’ perception of the firm. Funding liquidity risk is the risk that an institution will have to pay higher than prevailing market rates for its funding due to either the investors’ perception that the credit quality of the institution is impaired (possibly due to earnings problems or capital structure problems) or the overly heavy use of particular funding sources within a given time period, with the large size of transactions impacting funding cost. Enterprise risk is the risk of loss due to change in the overall business climate, such as the needs of customers, actions of competitors, and pace of technological innovation. This chapter briefly discusses each of these risks and possible controls, and then presents an overview of how these risks can be identified and the extent to which they can be quantified. A valuable source of ideas on operational risk and control procedures is the Trading and Capital‐Markets Activities Manual of the Federal Reserve System. I have used it as a foundation for several of the points in this chapter and recommend that readers interested in this topic look closely at the following sections: 2050.1 and 2060.1 (Operations and Systems Risk), 2070.1 (Legal Risk), 2150.1 (Ethics), 3005.1 (Funding Liquidity Risk), and 2040.1 (the subsection on New Products). Operational Risk 3.1 31 OPERATIONS RISK Operations risk can be further subdivided into the risk of fraud, risk of nondeliberate incorrect information, disaster risk, and personnel risk. 3.1.1 The Risk of Fraud The actual diversion of cash can take the form of creating unauthorized payments, conducting transactions at prices that are not the best available in return for bribes, or utilizing one’s position to engage in profitable personal trading at the expense of the firm’s profits. Deception about earnings, in order to generate unearned bonuses or further one’s career (or simply avoid being fired), can take the form of recording trades at incorrect prices or misreporting the current value of positions. We’ll encounter examples of such deceptions that occurred at Kidder Peabody, Barings, Allied Irish Bank (AIB), and Société Générale in Section 4.1. Section 4.1, covering financial disasters that were due to misleading reporting, should be read in conjunction with this section. Deception about positions, in order to appear to be operating within limits when an individual is actually outside them or to mislead management about the size of positions being taken, is done in order to preserve freedom of action—avoiding requirements to close down positions. This can be because a trader has a different belief about market movements than management or a different view toward risk than management (the moral hazard issue discussed in Section 2.1). Deception about positions can entail the outright misreporting of positions through the failure to enter transactions (tickets in the drawer) or manipulation of management reporting, or hiding positions by arranging for them to be temporarily held by another party with an unrecorded promise to take the position back (parking). Going back 30 years or so, the oral tradition within control functions was to worry about position falsification primarily by traders simply not entering some of their trades onto the firm’s books and records (tickets in the drawer). This approach seems to be on the decline, presumably because the possibility of the fraud being exposed through an inquiry from a counterparty to an unrecorded trade is too great. What seems to have replaced it is the entry of fictitious trades designed either to offset the risk position of actual trades, making the net risk look small, or to create bogus profit and loss (P&L) to disguise actual earnings. Since fictitious trades lack a real counterparty, they cannot be exposed through action of a counterparty but can be uncovered only by internal controls. The creator of fraud is in an ongoing battle with control personnel—the control personnel have the 32 FINANCIAL RISK MANAGEMENT advantage that uncovering only one clear‐cut case of falsification is enough to uncover the fraud; the advantage of the creator of the fraud is the multiplicity of methods the creator can employ to discourage this discovery. The most fundamental control for preventing fraud is by separating the responsibilities between the front office and the support staff (middle office, back office, and controllers), making sure that all entries of transactions and management reporting systems are under the complete control of the support staff. To make this separation of responsibilities work, the support staff must have a separate line of reporting from the front‐office staff and compensation that is reasonably independent of the reported earnings of the business area being supported. As much as possible, the reporting lines and compensation structure should align support staff interests with those of management rather than with those of the front office. However, even the best‐designed structures of this type are subject to pressures in the direction of alignment of support staff interests with front‐office interests. Constant vigilance is required to fight against this. These pressures include: ■ ■ ■ Support staff compensation cannot be completely independent from trading performance. At a minimum, unsuccessful results for trading may lead to the shrinking or elimination of a trading operation along with associated support staff positions. Since trading profits are the ultimate source from which expenses get paid, it is difficult to avoid some linkage between the trading performance and level of compensation. Section 4.1.4 presents a vivid example of how this pressure was felt in practice at AIB. Front‐office personnel almost always command higher compensation and prestige than members of the support staff, usually considerably higher. Often, support staff members are hoping to eventually move into front‐office positions. Front‐office staff can afford to offer informal incentives to the support staff for cooperation such as helping them seek front‐office jobs, giving access to perks such as lavish meals and free tickets to otherwise unavailable sports events, and even offering outright cash bribes. The higher prestige of front‐office positions and the reality of the greater market experience of front‐office personnel relative to support personnel can be utilized to place tremendous pressure on the support staff to adopt front‐office views. Since the support staff has responsibilities for supporting the front office as well as for supporting management, their ratings for job performance are often heavily dependent on the views of front‐office personnel, who are likely to be working far more closely with them than management personnel. Operational Risk 33 In addition to the separation of responsibilities, controls include: ■ ■ ■ ■ ■ ■ ■ ■ ■ ■ ■ ■ Support staff procedures should be thoroughly documented. Making these as unambiguous as possible lessens the scope for front‐office influence. Trader lines should be recorded to create a potential source for spotting evidence of collusion with brokers or traders at other firms. Make sure that trades are entered into the firm’s systems as close to execution time as possible. The further away from execution time you get, the greater the possibility that subsequent market movements will create a temptation to hide or otherwise misrepresent the transaction. Review all trades to look for prices that appear off‐market, and perform a thorough investigation of any trades identified as such. Make sure that all market quotes used to value positions come into support staff, not front‐office personnel, and are polled from as large a universe of sources as possible. Provide daily explanations of profit and loss (P&L) change and cash needs produced by the support staff. Incorrect reporting of positions can often be identified by the inability to explain P&L and cash movements based on the reported positions. Every customer confirmation of a new trade or a payment required by a previous trade should be reviewed by the support staff for consistency with transactions and positions being reported. All customer complaints should be reviewed by the support staff, not just front‐office personnel. The confirmation process should be conducted only with support personnel at other firms, not with front‐office personnel at other firms. Have clear policies about unacceptable practices that are consistently enforced. Deliberate actions to hide a position must entail strong penalties, without exceptions for star traders. Personal trading of both front‐office and support personnel should be closely monitored. Tight controls should be placed on after‐hours and off‐premises trading to ensure that transactions cannot be omitted from the firm’s records. Broker usage should be monitored for suspicious patterns—undue concentrations of business that might be compensation for supplying off‐ market quotes or direct bribery. Firms should insist on performing thorough background checks of a potential customer’s creditworthiness and business reputation before entering into transactions. In other words, they should refuse to deal with customers they do not know, even on a fully collateralized basis. Unknown customers could be in collusion with the firm’s personnel for off‐market trading or parking. 34 ■ ■ ■ ■ ■ FINANCIAL RISK MANAGEMENT Systems security measures should be in place to ensure that no one other than authorized support personnel can make entries or changes to management information systems. In particular, no front‐office personnel should have such access. The firm’s auditors should perform a periodic review of all operating procedures. Control functions must budget some spare capacity for investigative work. The advantage of the control functions relative to an attempt at concealing positions is that only one clear‐cut instance of concealment needs to be uncovered to expose a fraud and that any significant attempt at concealment will create a large number of warning indicators that can potentially trigger an investigation. But the disadvantage of control functions relative to an attempt at concealing positions is that a skillful perpetrator of fraud can be expected to be adroit at offering superficially plausible explanations of unusual patterns. If an investigation is being done in a little spare time of control personnel with a full plate of daily responsibilities, it will be too easy for them to try to wrap it up quickly by accepting a plausible explanation. If control personnel have some budgeted time for conducting such investigations and know that the thoroughness of performance of this task will be part of their job evaluations, it is far more likely that they will perform the extra work needed to uncover the true situation. This will lead to other benefits as well, since thorough investigation of unusual patterns may turn up other gaps in the control system, such as the need for new risk measures or accidental errors in recording positions. Control functions should maintain some central registry of investigations they have conducted (along with outcomes). Even if a perpetrator of fraud has been successful in fooling control personnel in several investigations, the unusual number of investigations that the perpetrator’s activity is engendering may itself be a clue that leads to a more thorough investigation. An overriding concern must be to protect control personnel against bullying. To have a good chance of uncovering frauds, investigations need to be launched based on warning indicators that will create many false positives. This means that the control personnel will go into the investigation knowing that the most likely outcome will be a finding that nothing is wrong. The very fact that they are conducting an investigation is likely to be resented by traders (as a waste of time and as an indicator that their honesty is being questioned). If control personnel feel they are going to be berated by the traders when their investigation finds nothing wrong, then they are likely to conduct fewer and more superficial investigations. Trading management needs to make sure that they give Operational Risk 35 control personnel the proper backing and try to explain to traders the motivation for such investigations. The careful documentation of major incidents of fraud and the difficulty in detecting them can provide trading managers with tools to use in making this case. 3.1.2 The Risk of Nondeliberate Incorrect Information It is far more common to have incorrect P&L and position information due to human or systems error than incorrect P&L and position information due to fraud. Many of the controls for nondeliberate incorrect information are similar to the controls for fraud. The separation of responsibilities is effective in having several sets of eyes looking at the entry of a trade, reducing the chance that a single individual’s error will impact positions. Checking confirmations and payment instructions against position entries, P&L and cash reconciliation, and the investigation of off‐market trades are just as effective in spotting inadvertent errors as they are in spotting fraudulent entries. Equally close attention needs to be paid to making sure customers have posted collateral required by contracts to avoid inadvertently taking unauthorized credit risk. (For further discussions of the role of collateral in managing credit risk, see Sections 4.1.1, 10.1.4, 14.2, and 14.3.3.) It is every bit as important to have front‐office personnel involved in reconciliation (to take advantage of their superior market knowledge and intuitive feel for the size of their P&L and positions) as it is to have support personnel involved (to take advantage of their independence). Front‐office personnel must be held responsible for the accuracy of the records of their P&L and positions, and cannot be allowed to place all the blame for incorrect reports on support personnel, in order to ensure that they will place sufficient importance on this reconciliation. Front offices should be required to produce daily projections of closing positions and P&L moves based on their own informal records, prior to seeing the official reports of positions and P&L, and should reconcile significant differences between the two. To prevent incorrect P&L and position information, it is important to ensure that adequate support personnel and system resources are available, both in quantity and in quality, relative to the size and complexity of trading. Careful attention needs to be paid to planning staff and system upgrades to anticipate growth in trading volume. Management needs to be ready to resist premature approval of a new business if support resources cannot keep pace with front‐office development. Should model risk be regarded as an operations risk issue? The viewpoint of this book is that model risk is primarily a market risk issue, since the proper selection and calibration to market prices of models and the 36 FINANCIAL RISK MANAGEMENT provision for adequate reserves against model uncertainty are best dealt with by the market risk discipline. Chapter 6 will elucidate this view. However, the proper implementation of models and the assurance that system changes are undertaken with the proper controls are best dealt with by the operations risk discipline. An area independent of model and system developers and the front office should be established to perform quality assurance testing of system implementation and modifications, and to review the adequacy of system documentation. 3.1.3 Disaster Risk The adequacy of support personnel and system resources for reporting P&L and positions must also be ensured in the event of a physical disaster. Examples of such disasters would include a power failure, fire, or explosion that closes down a trading facility and/or its supporting systems. Another example would be a computer system problem, such as a virus or error with consequences far‐reaching enough to jeopardize the entire support structure (the most famous example is the Y2K crisis). Resource adequacy cannot be limited to just the ability to keep track of existing positions. It is also necessary to allow continued trading in a sufficiently controlled environment, at least at a level that will permit the ongoing management of existing positions. The steps to deal with disaster risk begin with the development of a detailed contingency plan, which includes plans for backup computer systems, frequently updated backup data sets, backup power sources, and a backup trading floor. The adequacy of contingency plans must be judged against the likelihood that both the primary and backup facilities will be impacted by the same event. This concern was sharpened by the tragic events of September 11, 2001, when Bank of New York had both its primary and secondary trading systems, which were located in separate but nearby buildings, knocked out at the same time. This has caused many financial firms to rethink the degree of geographic separation that should be required between alternative sites. Widespread computer errors that cut across all systems of the firm (backup as well as primary) are particularly worrisome. For example, the only way around the Y2K bug was to get a complete fix in place and thoroughly tested prior to the onset of the potential problem. 3.1.4 Personnel Risk Investment banking firms have a history of raiding a competitor’s personnel and hiring, en masse, an entire group of traders along with key support Operational Risk 37 staff. This can have the same impact on the raided firm as a physical disaster, but it has a longer recovery time, since replacement personnel must be identified, hired, and trained. Protective steps are to utilize cross‐training and occasional backup duties as widely as possible to ensure that personnel are available to at least temporarily take over the duties of departed personnel. The requirements for thorough documentation of systems and procedures are also important. 3.2 LEGAL RISK There are two types of legal risk: (1) the risk that contracts will prove unenforceable and (2) the risk that actions of the firm’s employees will be found to be illegal, subjecting the firm to substantial penalties. We will examine both in turn. 3.2.1 The Risk of Unenforceable Contracts The legal risk that the terms or conditions of a contract will prove unenforceable due to legal defects can prove a more serious problem than the credit risk that a counterparty does not have the financial capacity to perform on a contract. If a contract is found to be unenforceable, it may simultaneously impact a large number of contracts and have exactly the same impact on a trading firm as if a large number of counterparties defaulted simultaneously. A classic case of this was the finding by British courts that derivative contracts with British municipalities were ultra vires; that is, they were not contracts that the municipalities were legally authorized to enter into. This simultaneously canceled all outstanding derivatives contracts that financial firms had with British municipalities. For more detail, see Malcolm, Sharma, and Tanega (1999, 149–150). Another reason why legal risk can be more serious than credit risk is that it suffers more from adverse selection. Counterparty default is generally unrelated to whether the counterparty owes money or is owed money. However, lawsuits occur only when counterparties owe money. The major mitigants to legal risk are: ■ ■ ■ ■ Thoroughly reviewing contract terms by experienced lawyers to ensure that language is properly drafted and that the contracted activities are authorized for the contracting parties. Thoroughly documenting what terms have been agreed to. Restricting dealings to reputable counterparties (know your customer). Placing limits on exposure to legal interpretations. 38 ■ FINANCIAL RISK MANAGEMENT Ensuring that contracts specify that legal jurisdiction resides with court systems that have experience in dealing with the particular issues involved and have previously demonstrated fairness in dealing with such cases. A thorough review of contract terms may require lawyers with specialized legal knowledge of particular subject areas of law and legal jurisdiction (such as laws of particular countries, states, and districts), including knowledge of how courts and juries in a jurisdiction tend to interpret the law as well as applicable precedents. This often requires that legal work be contracted to outside counsel who specialize in certain areas and jurisdictions. However, care must be exercised to prevent front‐office areas, which have a vested interest in seeing that a transaction gets done, from using this process to shop for a legal opinion, hiring a legal firm that can be counted on to provide a favorable opinion. The process of outside contracting of legal opinions must be controlled by an in‐house legal department or a single trusted outside legal firm that can be counted on to offer independent judgments in the interest of the trading firm when this conflicts with the interest of individual front‐office areas within the firm. Adequate and clear legal language may prove useless if sufficient documentation has not been obtained showing customer agreement to the language. The most important measure in this regard is a strong commitment to following up verbal trade agreements with well‐documented confirmations and signed legal agreements. This requires adequate documentation staff within trade support functions and the discipline to turn down potentially profitable business from counterparties that do not follow through on the required documentation. The enforcement of these rules is often placed within the credit risk function. Documentation should include written confirmation that a counterparty’s board of directors and senior management have knowledge of the activities being contracted and have authorized the officers of the counterparty firm with which the trading firm is dealing to enter into such contracts on behalf of the counterparty firm. It is also useful to record all conversations between the counterparties and trading firm personnel so that disputes as to what terms were verbally agreed to can be settled equitably, without resorting to costly legal proceedings. Firms have started to worry about what may be termed legal‐basis risk. This arises when a firm treats transactions with two different customers as offsetting and hence without market risk (although not without credit risk). However, it may turn out that slightly different wording in the two contracts means that they are not truly offsetting in all circumstances. Although carefully vetting contractual language is a necessary countermeasure, an even better preventative is to use standardized contractual language as much as Operational Risk 39 possible to make it easier to spot differences. The International Swaps and Derivatives Association (ISDA) has been working to develop standardized language that can be used in derivatives contracts. See Section 13.1.1.2 for more details. In addition to enforcing documentation rules, the credit risk function also needs to restrict the extension of credit to reputable counterparties. It is necessary to recognize that the willingness of a counterparty to meet contractual obligations is every bit as important as its financial ability to meet those obligations. A counterparty that does not have a good business reputation to protect may feel free to look for the slightest pretext to enter a legal challenge to meet its contractual obligations. Even if a firm has legal right strongly on its side, dealing with such a client may be very costly due to the expense of litigation and the threat of using a lawsuit as an excuse for a fishing expedition discovery process designed to uncover internal corporate information that can cause public embarrassment. The threat of such costs may incline a firm to settle for less than the full amount contractually owed, which serves as an incentive for unscrupulous firms to delay the settlement of legitimate claims. By contrast, a firm or an individual whose reputation for ethical business dealings is one of its assets will actually lean in the direction of making payments that meet its understanding of its obligations, even when the formal contract has been imperfectly drawn. Because it is extremely difficult to quantify legal risk, firms may overlook the usefulness of quantitative limits to control exposure. Consider an example of a particular legal interpretation that has the potential to void all contracts of a specific type. The firm’s legal consultants can issue opinions on the degree of likelihood that such an interpretation will be issued in the future by a court or regulatory body. Ultimately, business management must make a judgment on whether the economic benefits of the contract, relative to alternative ways of achieving the desired financial result, outweigh this risk. On a single deal, this is a binary decision—either you enter into the contract or you don’t. There are few circumstances under which protection against an unfavorable contract interpretation can be purchased, making legal risk very different from market or credit risk. However, this is all the more reason to place a quantitative limit on the total size of contracts subject to all being voided by a single interpretation, where the size of the contract can be quantified by the potential loss from being voided. Quantitative limits place a control on risk, can be sized based on the degree of economic benefit relative to the perceived degree of legal uncertainty, and provide a framework for ensuring that individual deal approval is limited to those with the greatest potential benefit relative to potential loss. One particular issue of legal risk that often causes concern is how bankruptcy courts will treat contractual obligations. When a counterparty goes 40 FINANCIAL RISK MANAGEMENT into default, the counterparty’s reputation and desire to deal fairly no longer serve as a bulwark against litigation risk. In bankruptcy, all of the bankrupt firm’s creditors become competitors in legal actions to gain as much of a share of the remaining assets as possible. Even when legal documents have been well drawn to provide specific collateral against an obligation or specific netting arrangements between derivative contracts on which the bankrupt firm owes and is owed money, other creditors may try to convince bankruptcy courts that it is only fair that they receive a share of the collateral or derivatives on which the bankrupt firm is owed money. Bankruptcy courts have been known to issue some very surprising rulings in these circumstances. Contractual intention can be voided not only by courts, but also by regulatory authorities or legislatures, which may issue rules that make certain contractual provisions unenforceable. Financial institutions can and do mount lobbying campaigns against such changes, but other parties may be as effective or more effective in lobbying on the other side. Financial firms often need to analyze what they believe is the prospect for future regulatory actions in order to determine whether certain current business will prove to be worthwhile. More detail on legal risk and how to control it can be found in Chapter 7 of Malcolm et al. (1999). 3.2.2 The Risk of Illegal Actions The possibility of a firm’s employees engaging in actions found to be illegal bears a very close relationship to reputational risk, which is examined in the next section. Any legal proceedings against a firm have the potential to damage the firm’s reputation and the willingness of clients to engage its services. Even when legal proceedings don’t result in a judgment against the firm, the publicity about the allegations and embarrassing disclosures in the legal discovery process can still impair reputation. And actions that can generate negative press, even if not rising to the level of illegality, can have a similar effect on reputation. One of the most effective screens for acceptable behavior remains the classic “Would you be comfortable seeing a description of this practice on the front page of the Wall Street Journal?” The primary focus of legal and reputational risk has always been on the fiduciary responsibilities owed by a firm to its clients, particularly its less sophisticated clients. But recent cases have extended concern to damages that the client may inflict on others that the firm may be seen as having abetted. Section 4.3.2 on the losses in lawsuits of JPMorgan Chase and Citigroup for having been party to the Enron deception of investors is a good case study in this respect. Operational Risk 3.3 41 REPUTATIONAL RISK Firms need to be sure not only that contract provisions are legally enforceable, but also that the process of enforcing their legal rights will not damage their business reputation. Even if a contract is strictly legal and enforceable, if its terms seem palpably unfair or can be portrayed as taking advantage of a client, the enforcement of the legal claims may be as damaging to the firm (or more so) as the inability to enforce the claims would have been. All transactions need to be reviewed by business managers from the viewpoint of whether the transaction is one that the client fully understands and it can reasonably be interpreted as a sensible action for the client to take. Ever since the Bankers Trust (BT) fiasco with Procter & Gamble (P&G) and Gibson Greetings, described in Section 4.3.1, all firms have placed increased emphasis on processes to ensure that transactions are appropriate or suitable for the client. The following processes are included: ■ ■ ■ ■ Conduct a careful review of all marketing materials to make sure that transactions have been fully explained, no misleading claims have been made, and no ambiguity exists as to whether the financial firm is simply acting as deal structurer or is also acting as an adviser to the client with fiduciary responsibility for the soundness of its advice. A full explanation of transactions may need to include simulations of possible outcomes, including stress situations. Make certain that any request from a client for a mark‐to‐market valuation of an existing transaction is supplied by support personnel using objective standards and not by marketing personnel who may have motivations to mislead the client as to the true performance of the transaction. Further, all valuations supplied need to be clearly labeled as to whether they are actual prices at which the trading firm is prepared to deal or simply indications of the general market level. Rank clients by their degree of financial sophistication and transactions by their degree of complexity, and ensure that a proper fit exists between the two. In cases where complex transactions are negotiated with less sophisticated clients, extra care needs to be taken to ensure that any advice given to the client by marketing personnel is consistent with their knowledge of the client’s needs. Verify that clients fully understand the nature of the transactions they are undertaking, including written confirmation of such assurances from senior managers in some cases, based on the size and complexity of deals. These steps to ensure appropriateness and suitability are important not only to guard a trading firm’s reputation, but, in extreme 42 FINANCIAL RISK MANAGEMENT cases, to also serve as protection against litigation. Note that the need to ensure the suitability of transactions to clients and the need to provide clients with evaluations that the trading firm can certify as reliable limit a trading firm’s ability to simply serve as a credit intermediary between two counterparties using back‐to‐back derivatives. 3.4 ACCOUNTING RISK Accounting risk can be viewed as a form of reputational risk. When a firm makes serious accounting errors, requiring the restatement of past earnings, it does not lead to any net loss of cash to the firm, as in cases of fraud, operations errors, or incorrectly drawn contracts. However, it can damage investor, creditor, and regulator confidence in the accuracy of information that the firm supplies about its financial health. This loss of confidence can be so severe that it threatens the firm’s continued existence, as the Kidder Peabody financial disaster, discussed in Section 4.1.2, illustrates. Measures to control accounting risk are similar in nature to those needed to control legal risk. Instead of needing knowledge of legal issues and precedents and how courts tend to interpret the law, knowledge of generally accepted accounting principles (GAAP) and how accounting boards of standards and regulatory authorities tend to interpret these principles is needed. The need for specialized knowledge by accounting jurisdiction is similar to the need for specialized knowledge by legal jurisdiction. The need to obtain independent accounting opinions and avoid opinion shopping parallels those considerations for legal risk. The need for thorough documentation showing that accounting rules are being followed parallels the need for thorough documentation of contractual understandings. The need for limits on exposure to accounting policies open to interpretation parallels the need for limits on exposure to legal interpretation. 3.5 FUNDING LIQUIDITY RISK Funding liquidity risk should be clearly differentiated from the liquidity risk we discussed as part of market risk in Section 1.2, which is sometimes called asset liquidity risk. Funding liquidity risk has two fundamental components: 1. The risk that investors’ perception of the firm’s credit quality will become impaired, thereby raising the firm’s funding costs relative to the costs of competitors across all funding sources utilized. Operational Risk 43 2. The overly heavy use of a particular funding source in a given time period, raising the firm’s funding cost relative to that of competitors for that particular funding source only. Controlling the cost of the firm’s liabilities by managing investors’ perceptions of the firm’s credit quality is the flip side of the coin of credit risk’s management of the credit quality of the firm’s assets. Crises in investor confidence are usually triggered by problems with earnings or the inadequacy of capital. As a result, they are functions of the overall management of the firm’s business. The chief financial officer of the firm has particular responsibility for controlling funding liquidity risk by explaining the earnings situation to financial analysts and rating agencies and ensuring that capital levels are maintained to meet both regulatory guidelines and the expectations of financial analysts and rating agencies. Specific funding liquidity responsibilities of the treasury function of the firm include ensuring that any such crisis is not exacerbated by having to raise too much funding from the market at a time of crisis. Preferably, the firm should be able to reduce to a bare minimum its funding during a crisis period to gain time for the firm to improve its fundamental financial condition and tell its side of the story effectively to financial analysts, rating agencies, and individual investors. The ability to avoid too much market funding in these circumstances requires: ■ ■ ■ Long‐term plans to get more funding from stable sources less sensitive to a firm’s credit rating (such as retail deposits and transaction balances), to lengthen the maturity of market funding, to create cushions of market funding to tap in emergencies by raising less than the full amount of potential funds available, and to arrange backup lines of credit. Information systems to project periods of large funding needs in order to spread out the period of time over which such funding is raised. Of particular importance is the use of funding needs projections to avoid having funding requirements over a short period being so heavy that they trigger a crisis of investor confidence. Well‐developed contingency plans for handling a funding crisis, which could include steps such as selling liquid assets, unwinding liquid derivatives positions that tie up collateral, and utilizing untapped cushions of funding and backup lines of credit. The treasury function’s management of particular funding sources to avoid overuse is also tied to information systems that can project future funding needs. It may be necessary to restrict particular types of investment 44 FINANCIAL RISK MANAGEMENT or derivative transactions that depend on access to particular funding sources to be profitable. For example, some transactions are profitable only if off‐balance‐sheet commercial paper funding can be obtained, bypassing the need for capital to be held against on‐balance‐sheet assets. However, the treasury function may need to limit the total amount of commercial paper being rolled over in any particular period to reduce the risk of having to pay a premium for such funding. 3.6 ENTERPRISE RISK Enterprise risk can be tied to the fixed nature of many of the costs of engaging in a particular line of business. Even heavily personnel‐intensive businesses, such as trading, still have fixed cost components such as buildings, computer and communications equipment, and some base level of employee compensation below which a firm loses its ability to remain in the business line through downturns in activity. However, these fixed costs entail the risk of losses to the extent that the amount of business that can be attracted in a downturn cannot cover the fixed costs. By its nature, the management of enterprise risk belongs more naturally to individual business managers than to a corporate‐wide risk function. Usually, the corporate‐wide operational risk function will restrict itself to attempting to include some measurement of enterprise risk in the risk‐adjusted return on capital (RAROC) or shareholder value added (SVA) measures. 3.7 IDENTIFICATION OF RISKS In Damon Runyon’s short story on which the musical Guys and Dolls is based, a gambler named Sky Masterson relates the following advice he received from his father: Son, no matter how far you travel or how smart you get, always remember this: Someday, somewhere, a guy is going to come to you and show you a nice brand-new deck of cards on which the seal is never broken, and this guy is going to offer to bet you that the jack of spades will jump out of the deck and squirt cider in your ear. But, son, do not bet him, for as sure as you do you are going to get an ear full of cider. The equivalent of this story for a risk manager is the trader or marketer who informs you that “There is absolutely no risk of loss on this product.” Operational Risk 45 As my experience with markets has grown, I have come to recognize this assertion as a sure harbinger of painful losses to come, either sooner or later. However, my first encounter with the statement came well before I was involved with the financial side of banking, when I was working in Chase Manhattan’s operations research department on projects like the simulation of the truck routes that delivered checks from branches to the head office and the sorting machines that then processed the checks. One day on the subway, I ran into someone I had worked with on these simulations, but had not seen in a few years. He told me about the wonderful new job he had heading up a unit of the bank that matched firms that wanted to borrow securities with those that wanted to lend them. The bank received a nice fee for the service and he was aggressively growing the business. The key to profitability was operational efficiency, at which he was an expert. He told me that since the bank was not a principal to any of these transactions, there was absolutely no risk of loss on the product. The losses came a few years later. When Drysdale Securities, a large borrower of government securities, could not repay its borrowings, it turned out that considerable ambiguity existed about whether the lenders of the securities understood they were being borrowed by Chase or by Drysdale with Chase merely arranging the borrowing. The legal contracts under which the transactions had been executed were open to the interpretation that Chase was the principal. Chase lost $285 million in settling these claims (see Section 4.1.1 for a more detailed discussion). My acquaintance, needless to say, lost his job. Before a risk can be controlled, it must first be recognized. Often, the management team that is involved with the introduction of a new product may lack the experience to perceive a possibility of risk and as a result may fail to call in the expertise needed to control the risk. For example, if a new legal risk is not recognized, the firm’s legal experts may never thoroughly review the existing contracts. This is why it has become the accepted best practice in the financial industry to establish a new‐product review process in which more experienced business managers and experts in risk disciplines (such as market risk, credit risk, reputational risk, legal, finance, and audit) vet proposals for products to make sure risks are identified and controls are instituted. 3.8 OPERATIONAL RISK CAPITAL We started this chapter by stating that operational risks do not fall under the scope of financial risk management, since they could not be managed using liquid markets. This is not to say that quantitative measures cannot 46 FINANCIAL RISK MANAGEMENT be developed for operational risk, just that the tools to do so will come from the traditional insurance industry and will be close to the tools used to manage exposure to physical disasters, such as hurricanes and nuclear plant breakdowns. There are certain common items in the tool kit of financial risk and insurance risk, given that they are both trying to measure exposures in the extreme tail of events, as discussed in Section 1.3. For example, you will see use of simulation, extreme value theory, and stress scenarios in both. But the specific techniques discussed in this book, very closely tied to relating loss estimation and control to liquid market price movements, cannot be applied. The primary impetus for developing quantitative measures of operational risk has been a desire to develop a methodology for operational risk capital to complement the measures of capital allocated for market risk and credit risk. In particular, the push by the Basel Committee on Banking Supervision to promulgate international standards requiring all banks to allocate capital against operational risk has spurred much work on how to quantify this capital requirement. Operational risk capital can be approached in two ways—from the bottom up and from the top down. The bottom‐up approach emphasizes quantitative measures of factors that contribute to operational risk. Some possibilities are: ■ ■ ■ Audit scores as a measure of operations risk. Counts of unreconciled items or error rates as a measure of operations risk. Measures of delay in obtaining signed confirmations as a measure of legal risk. Although these measures provide good incentives, tying reduction in capital to desirable improvements in controls, it is very difficult to establish links between these measures and the possible sizes of losses. Some firms are pursuing research on this, but supporting data is scarce. The top‐down approach emphasizes the historical volatility of earnings. This measure provides a direct link to the size of losses and can include all operational risks, even enterprise risk. But what incentive does this measure provide to reducing operational risk? No credit is given to a program that clears up back‐office problems or places new controls on suitability. Neither of these approaches bears much resemblance to the use of actual market prices for reduction of risk, which we discussed in Chapter 1 as the hallmark of financial risk management. To the extent market prices are available for some operational risks, it would come from the insurance Operational Risk 47 market, since it is possible to purchase insurance against some types of risk of fraud, operations errors, disasters, loss of personnel, legal liability, and accounting errors. An up‐to‐date and thorough introduction to methodological approaches to quantifying operational risk capital and the regulatory background of the Basel Committee initiatives can be found in the closely related books Moosa (2007), particularly Chapters 5, 6, and 7, and Moosa (2008), particularly Chapters 4, 5, and 6. CHAPTER 4 Financial Disasters ne of the fundamental goals of financial risk management is to avoid the types of disasters that can threaten the viability of a firm. So we should expect that a study of such events that have occurred in the past will prove instructive. A complete catalog of all such incidents is beyond the scope of this book, but I have tried to include the most enlightening examples that relate to the operation of financial markets, as this is the book’s primary focus. A broad categorization of financial disasters involves a three‐part division: O 1. Cases in which the firm or its investors and lenders were seriously misled about the size and nature of the positions it had. 2. Cases in which the firm and its investors and lenders had reasonable knowledge of its positions, but had losses resulting from unexpectedly large market moves. 3. Cases in which losses did not result from positions held by the firm, but instead resulted from fiduciary or reputational exposure to positions held by the firm’s customers. 4.1 DISASTERS DUE TO MISLEADING REPORTING A striking feature of all the financial disasters we will study involving cases in which a firm or its investors and lenders have been misled about the size and nature of its positions is that they all involve a significant degree of deliberation on the part of some individuals to create or exploit incorrect information. This is not to say situations do not exist in which firms are misled without any deliberation on the part of any individual. Everyone who has been in the financial industry for some time knows of many instances when everyone at the firm was misled about the nature of positions because a ticket was entered into a system incorrectly. Most typically, 49 50 FINANCIAL RISK MANAGEMENT this will represent a purchase entered as a sale, or vice versa. However, although the size of such errors and the time it takes to detect them can sometimes lead to substantial losses, I am not aware of any such incident that has resulted in losses that were large enough to threaten the viability of a firm. An error in legal interpretation can also seriously mislead a firm about its positions without any deliberate exploitation of the situation. However, such cases, although they can result in large losses, tend to be spread across many firms rather than concentrated at a single firm, perhaps because lawyers tend to check potentially controversial legal opinions with one another. The best‐known case of this type was when derivatives contracted by British municipalities were voided. See Section 3.2. If we accept that all cases of financial disaster due to firms being misled about their positions involve some degree of complicity on the part of some individuals, we cannot regard them completely as cases of incorrectly reported positions. Some of the individuals involved know the correct positions, at least approximately, whereas others are thoroughly misinformed. Understanding such cases therefore requires examining two different questions: 1. Why does the first group persist in taking large positions they know can lead to large losses for the firm despite their knowledge of the positions? 2. How do they succeed in keeping this knowledge from the second group, who we can presume would put a stop to the position taking if they were fully informed? I will suggest that the answer to the first question tends to be fairly uniform across disasters, while the answer to the second question varies. The willingness to take large risky positions is driven by moral hazard. As we saw in our discussion of moral hazard in Section 2.1, it represents an asymmetry in reward structure and an asymmetry in information; in other words, the group with the best information on the nature of the risk of a position has a greater participation in potential upside than potential downside. This often leads insiders to desire large risky positions that offer them commensurately large potential gains. The idea is that traders own an option on their profits; therefore, they will gain from increasing volatility, as we discussed in Section 2.1. The normal counterweights against this are the attempts by representatives of senior management, stockholders, creditors, and government regulators, who all own a larger share of the potential downside than the traders, to place controls on the amount of risk taken. However, when those who could exercise this control substantially Financial Disasters 51 lack knowledge of the positions, the temptation exists for traders to exploit the control weakness to run inflated positions. This action often leads to another motivation spurring the growth of risky positions—the Ponzi scheme, as discussed in Section 2.2. Some traders who take risky positions that are unauthorized but disguised by a control weakness will make profits on these positions. These positions are then possibly closed down without anyone being the wiser. However, some unauthorized positions will lead to losses, and traders will be strongly tempted to take on even larger, riskier positions in an attempt to cover up unauthorized losses. This is where the Ponzi scheme comes in. I think it helps to explain how losses from unauthorized positions can grow to be so overwhelmingly large. Stigum (1989) quotes an “astute trader” with regard to the losses in the Chase/Drysdale financial disaster: “I find it puzzling that Drysdale could lose so much so fast. If you charged me to lose one‐fourth of a billion, I think it would be hard to do; I would probably end up making money some of the time because I would buy something going down and it would go up. They must have been extraordinarily good at losing money.” I would suggest that the reason traders whose positions are unauthorized can be so “extraordinarily good at losing money” is that normal constraints that force them to justify positions to outsiders are lacking and small unauthorized losses already put them at risk of their jobs and reputations. With no significant downside left, truly reckless positions are undertaken in an attempt to make enough money to cover the previous losses. This is closely related to double‐or‐nothing betting strategies, which can start with very small stakes and quickly mushroom to extraordinary levels in an effort to get back to even. This snowballing pattern can be seen in many financial disasters. Nick Leeson’s losses on behalf of Barings were just $21 million in 1993, $185 million in 1994, and $619 million in just the first two months of 1995 (Chew 1996, Table 10.2). John Rusnak’s unauthorized trading at Allied Irish Bank (AIB) accumulated losses of $90 million in its first five years through 1999, $210 million in 2000, and $374 million in 2001 (Ludwig 2002, Section H). Joseph Jett’s phantom trades at Kidder Peabody started off small and ended with booked trades in excess of the quantity of all bonds the U.S. Treasury had issued. The key to preventing financial disasters based on misrepresented positions is therefore the ability to spot unauthorized position taking in a timely enough fashion to prevent this explosive growth in position size. The lessons we can learn from these cases primarily center on why it took so long for knowledge of the misreported positions to spread from an insider group to the firm’s management. We will examine each case by providing a brief summary of how the unauthorized position arose, how it failed to come 52 FINANCIAL RISK MANAGEMENT to management’s attention, and what lessons can be learned. In each instance, I provide references for those seeking more detailed knowledge of the case. General conclusions based on the cases in this section can be found in Section 3.1.1. 4.1.1 Chase Manhattan Bank/Drysdale Securities 4.1.1.1 Incident In three months of 1976, Drysdale Government Securities, a newly founded subsidiary of an established firm, succeeded in obtaining unsecured borrowing of about $300 million by exploiting a flaw in the market practices for computing the value of U.S. government bond collateral. This unsecured borrowing exceeded any amount Drysdale would have been approved for, given that the firm had only $20 million in capital. Drysdale used the borrowed money to take outright positions in bond markets. When the traders lost money on the positions they put on, they lacked cash with which to pay back their borrowings. Drysdale went bankrupt, losing virtually all of the $300 million in unsecured borrowings. Chase Manhattan absorbed almost all of these losses because it had brokered most of Drysdale’s securities borrowings. Although Chase employees believed they were only acting as agents on these transactions and were not taking any direct risk on behalf of Chase, the legal documentation of the securities borrowings did not support their claim. 4.1.1.2 Result Chase’s financial viability was not threatened by losses of this size, but the losses were large enough to severely damage its reputation and stock valuation for several years. 4.1.1.3 How the Unauthorized Positions Arose Misrepresentation in obtaining loans is unfortunately not that uncommon in bank lending. A classic example would be Anthony De Angelis, the “Salad Oil King,” who, in 1963, obtained $175 million in loans supposedly secured by large salad oil holdings, which turned out to be vast drums filled with water with a thin layer of salad oil floating on top. Lending officers who came to check on their collateral were bamboozled into only looking at a sample from the top of each tank. The following are some reasons for featuring the Drysdale shenanigans in this section rather than discussing any number of other cases of misrepresentation: ■ ■ ■ Drysdale utilized a weakness in trading markets to obtain its funds. Drysdale lost the borrowed money in the financial markets. It is highly unusual for a single firm to bear this large a proportion of this large a borrowing sting. Financial Disasters 53 There is not much question as to how Drysdale managed to obtain the unsecured funds. The firm took systematic advantage of a computational shortcut in determining the value of borrowed securities. To save time and effort, borrowed securities were routinely valued as collateral without accounting for accrued coupon interest. By seeking to borrow large amounts of securities with high coupons and a short time left until the next coupon date, Drysdale could take maximum advantage of the difference in the amount of cash the borrowed security could be sold for (which included accrued interest) and the amount of cash collateral that needed to be posted against the borrowed security (which did not include accrued interest). 4.1.1.4 How the Unauthorized Positions Failed to Be Detected Chase Manhattan allowed such a sizable position to be built up largely because it believed that the firm’s capital was not at risk. The relatively inexperienced managers running the securities borrowing and lending operation were convinced they were simply acting as intermediaries between Drysdale and a large group of bond lenders. Through their inexperience, they failed both to realize that the wording in the borrowing agreements would most likely be found by a court to indicate that Chase was taking full responsibility for payments due against the securities borrowings and to realize the need for experienced legal counsel to review the contracts. 4.1.1.5 How the Unauthorized Positions Were Eventually Detected There was some limit to the size of bond positions Drysdale could borrow, even given the assumption that the borrowings were fully collateralized. At some point, the size of the losses exceeded the amount of unauthorized borrowings Drysdale could raise and the firm had to declare bankruptcy. 4.1.1.6 Lessons Learned The securities industry as a whole learned that it needed to make its methods for computing collateral value on bond borrowings more precise. Chase, and other firms that may have had similar control deficiencies, learned the need for a process that forced areas contemplating new product offerings to receive prior approval from representatives of the principal risk control functions within the firm (see Section 3.7). 4.1.1.7 Further Reading Chapter 14 of Stigum (1989) gives a detailed description of the Chase/Drysdale incident, some prior misadventures in bond borrowing collateralization, and the subsequent market reforms. 4.1.2 Kidder Peabody 4.1.2.1 Incident Between 1992 and 1994, Joseph Jett, head of the government bond trading desk at Kidder Peabody, entered into a series of trades 54 FINANCIAL RISK MANAGEMENT that were incorrectly reported in the firm’s accounting system, artificially inflating reported profits. When this was ultimately corrected in April 1994, $350 million in previously reported gains had to be reversed. 4.1.2.2 Result Although Jett’s trades had not resulted in any actual loss of cash for Kidder, the announcement of such a massive misreporting of earnings triggered a substantial loss of confidence in the competence of the firm’s management by customers and General Electric, which owned Kidder. In October 1994, General Electric sold Kidder to PaineWebber, which dismantled the firm. 4.1.2.3 How the Unauthorized Positions Arose A flaw in accounting for forward transactions in the computer system for government bond trading failed to take into account the present valuing of the forward. This enabled a trader purchasing a cash bond and delivering it at a forward price to book an instant profit. Over the period between booking and delivery, the profit would inevitably dissipate, since the cash position had a financing cost that was unmatched by any financing gain on the forward position. Had the computer system been used as it was originally intended (for a handful of forward trades with only a few days to forward delivery), the size of error would have been small. However, the system permitted entry not only of contracted forward trades, but also of intended forward delivery of bonds to the U.S. Treasury, which did not actually need to be acted on, but could be rolled forward into further intentions to deliver in the future. Both the size of the forward positions and the length of the forward delivery period were constantly increased to magnify the accounting error. This permitted a classic Ponzi scheme of ever‐mounting hypothetical profits covering the fact that previously promised profits never materialized. Although it has never been completely clear how thoroughly Jett understood the full mechanics of the illusion, he had certainly worked out the link between his entry of forward trades and the recording of profit, and increasingly exploited the opportunity. 4.1.2.4 How the Unauthorized Positions Failed to Be Detected Suspicions regarding the source of Jett’s extraordinary profit performance were widespread throughout the episode. It was broadly perceived that no plausible account was being offered of a successful trading strategy that would explain the size of reported earnings. On several occasions, accusations were made that spelled out exactly the mechanism behind the inflated reporting. Jett seemed to have had a talent for developing explanations that succeeded in totally confusing everyone (including, perhaps, himself) as to what was going on. However, he was clearly aided and abetted by a management satisfied Financial Disasters 55 enough not to take too close a look at what seemed like a magical source of profits. 4.1.2.5 How the Unauthorized Positions Were Eventually Detected Large increases in the size of his reported positions and earnings eventually triggered a more thorough investigation of Jett’s operation. 4.1.2.6 Lessons to Be Learned Two lessons can be drawn from this: Always investigate a stream of large unexpected profits thoroughly and make sure you completely understand the source. Periodically review models and systems to see if changes in the way they are being used require changes in simplifying assumptions (see Section 8.2.8). 4.1.2.7 Further Reading Jett has written a detailed account of the whole affair (see Jett 1999). However, his talent for obscurity remains and it is not possible to tell from his account just what he believes generated either his large profits or the subsequent losses. For an account of the mechanics of the deception, one must rely on the investigation conducted by Gary Lynch on behalf of Kidder. Summaries of this investigation can be found in Hansell (1997), Mayer (1995), and Weiss (1994). 4.1.3 Barings Bank 4.1.3.1 Incident The incident involved the loss of roughly $1.25 billion due to the unauthorized trading activities during 1993 to 1995 of a single, relatively junior trader named Nick Leeson. 4.1.3.2 Result The size of the losses relative to Barings Bank’s capital along with potential additional losses on outstanding trades forced Barings into bankruptcy in February 1995. 4.1.3.3 How the Unauthorized Positions Arose Leeson, who was supposed to be running a low‐risk, limited return arbitrage business for Barings in Singapore, was actually taking increasingly large speculative positions in Japanese stocks and interest rate futures and options. He disguised his speculative position taking by reporting that he was taking the positions on behalf of fictitious customers. By booking the losses to these nonexistent customer accounts, he was able to manufacture fairly substantial reported profits for his own accounts, enabling him to earn a $720,000 bonus in 1994. 4.1.3.4 How the Unauthorized Positions Failed to Be Detected A certain amount of credit must be given to Leeson’s industriousness in perpetrating a 56 FINANCIAL RISK MANAGEMENT deliberate fraud. He worked hard at creating false accounts and was able to exploit his knowledge of weaknesses in the firm’s controls. However, anyone reading an account of the incident will have to give primary credit to the stupendous incompetence on the part of Barings’ management, which ignored every known control rule and failed to act on myriad obvious indications of something being wrong. What is particularly amazing is that all those trades were carried out in exchange‐traded markets that require immediate cash settlement of all positions, thereby severely limiting the ability to hide positions (although Leeson did even manage to get some false reporting past the futures exchange to reduce the amount of cash required). The most blatant of management failures was an attempt to save money by allowing Leeson to function as head of trading and the back office at an isolated branch. Even when auditors’ reports warned about the danger of allowing Leeson to settle his own trades, thereby depriving the firm of an independent check on his activities, Barings’ management persisted in their folly. Equally damning was management’s failure to inquire how a low‐risk trading strategy was supposedly generating such a large profit. Even when covering these supposed customer losses on the exchanges required Barings to send massive amounts of cash to the Singapore branch, no inquires were launched as to the cause. A large part of this failure can be attributed to the very poor structuring of management information so that different risk control areas could be looking at reports that did not tie together. The funding area would see a report indicating that cash was required to cover losses of a customer, not the firm, thereby avoiding alarm bells about the trading losses. A logical consequence is that credit exposure to customers must be large since the supposed covering of customer losses would entail a loan from Barings to the customer. However, information provided to the credit risk area was not integrated with information provided to funding and showed no such credit extension. 4.1.3.5 How the Unauthorized Positions Were Eventually Detected The size of losses Leeson was trying to cover up eventually got too overwhelming and he took flight, leaving behind an admission of irregularities. 4.1.3.6 Lessons to Be Learned One might be tempted to say that the primary lesson is that there are limits to how incompetent you can be and still hope to manage a major financial firm. However, to try to take away something positive, the major lessons would be the absolute necessity of an independent trading back office, the need to make thorough inquiries about unexpected sources of profit (or loss), and the need to make thorough inquiries about any large unanticipated movement of cash. Financial Disasters 57 4.1.3.7 Further Reading A concise and excellent summary of the Barings case constitutes Chapter 10 of Chew (1996). Chapter 11 of Mayer (1997) contains less insight on the causes, but is strong on the financial and political maneuvers required to avoid serious damage to the financial system from the Barings failure. Leeson has written a full‐length book that appears to be reasonably honest as to how he evaded detection (Leeson 1996). Fay (1996) and Rawnsley (1995) are also full‐length accounts. 4.1.4 Allied Irish Bank (AIB) 4.1.4.1 Incident John Rusnak, a currency option trader in charge of a very small trading book in AIB’s Allfirst First Maryland Bancorp subsidiary, entered into massive unauthorized trades during the period 1997 through 2002, ultimately resulting in $691 million in losses. 4.1.4.2 Result price. This resulted in a major blow to AIB’s reputation and stock 4.1.4.3 How the Unauthorized Positions Arose Rusnak was supposed to be running a small arbitrage between foreign exchange (FX) options and FX spot and forward markets. He was actually running large outright positions and disguising them from management. 4.1.4.4 How the Unauthorized Positions Failed to Be Detected To quote the investigating report, “Mr. Rusnak was unusually clever and devious.” He invented imaginary trades that offset his real trades, making his trading positions appear small. He persuaded back‐office personnel not to check these bogus trades. He obtained cash to cover his losses by selling deep‐in‐the‐money options, which provided cash up front in exchange for a high probability of needing to pay out even more cash at a later date, and covered up his position by offsetting these real trades with further imaginary trades. He entered false positions into the firm’s system for calculating value at risk (VaR) to mislead managers about the size of his positions. In many ways, Rusnak’s pattern of behavior was a close copy of Nick Leeson’s at Barings, using similar imaginary transactions to cover up real ones. Rusnak operated without Leeson’s advantage of running his own back office, but had the offsetting advantage that he was operating in an over‐the‐ counter market in which there was not an immediate need to put up cash against losses. He also was extremely modest in the amount of false profit he claimed so he did not set off the warning flags of large unexplained profits from small operations that Leeson and Jett at Kidder Peabody had triggered in their desire to collect large bonuses. 58 FINANCIAL RISK MANAGEMENT Like Barings, AIB’s management and risk control units demonstrated a fairly startling level of incompetence in failing to figure out that something was amiss. AIB at least has the excuse that Rusnak’s business continued to look small and insignificant, so it never drew much management attention. However, the scope and length of time over which Rusnak’s deception continued provided ample opportunity for even the most minimal level of controls to catch up with him. The most egregious was the back office’s failure to confirm all trades. Rusnak succeeded in convincing back‐office personnel that not all of these trades needed to be confirmed. He relied partly on an argument that trades whose initial payments offset one another didn’t really need to be checked since they did not give rise to net immediate cash flow, ignoring the fact that the purported trades had different terms and hence significant impact on future cash flows. He relied partly on booking imaginary trades with counterparties in the Asian time zone, making confirmation for U.S.‐based back‐office staff a potentially unpleasant task involving middle‐of‐the‐night phone calls, perhaps making it easier to persuade them that this work was not really necessary. He also relied on arguments that costs should be cut by weakening or eliminating key controls. Once this outside control was missing, the way was opened for the ongoing manipulation of trading records. Auditors could have caught this, but the spot audits performed used far too small a sample. Suspicious movements in cash balances, daily trading profit and loss (P&L), sizes of gross positions, and levels of daily turnover were all ignored by Rusnak’s managers through a combination of inexperience in FX options and overreliance on trust in Rusnak’s supposedly excellent character as a substitute for vigilant supervision. His management was too willing to withhold information from control functions and too compliant with Rusnak’s bullying of operations personnel as part of a general culture of hostility toward control staff. This is precisely the sort of front‐office pressure that reduces support staff independence, which was referred to in Section 3.1.1. 4.1.4.5 How the Unauthorized Positions Were Eventually Detected In December 2001, a back‐office supervisor noticed trade tickets that did not have confirmations attached. When informed that the back‐office personnel did not believe all trades required confirmations, he insisted that confirmation be sought for existing unconfirmed trades. Although it took some time for the instructions to be carried out, when they finally were carried out in early February 2002, despite some efforts by Rusnak to forge written confirmations and bully the back office into not seeking verbal confirmations, his fraud was brought to light within a few days. Financial Disasters 59 4.1.4.6 Lessons to Be Learned This incident does not provide many new lessons beyond the lessons that should already have been learned from Barings. This case does emphasize the need to avoid engaging in small ventures in which the firm lacks any depth of expertise—there is simply too much reliance on the knowledge and probity of a single individual. On the positive side, the investigative report on this fraud has provided risk control units throughout the financial industry with a set of delicious quotes that are sure to be trotted out anytime they feel threatened by cost‐ cutting measures or front‐office bullying and lack of cooperation. The following are a few choice samples from Ludwig (2002): ■ ■ ■ ■ ■ When one risk control analyst questioned why a risk measurement system was taking market inputs from a front‐office‐controlled system rather than from an independent source, she was told that AIB “would not pay for a $10,000 data feed from Reuters to the back office.” When questioned about confirmations, “Mr. Rusnak became angry. He said he was making money for the bank, and that if the back office continued to question everything he did, they would drive him to quit. . . . Mr. Rusnak’s supervisor warned that if Mr. Rusnak left the bank, the loss of his profitable trading would force job cuts in the back office.” “When required, Mr. Rusnak was able to use a strong personality to bully those who questioned him, particularly in Operations.” His supervisors “tolerated numerous instances of severe friction between Mr. Rusnak and the back‐office staff.” Rusnak’s supervisor “discouraged outside control groups from gaining access to information in his area and reflexively supported Mr. Rusnak whenever questions about his trading arose.” “[I]n response to general efforts to reduce expense and increase revenues, the Allfirst treasurer permitted the weakening or elimination of key controls for which he was responsible. . . . Mr. Rusnak was able to manipulate this concern for additional cost cutting into his fraud.” 4.1.4.7 Further Reading I have relied heavily on the very thorough report issued by Ludwig (2002). 4.1.5 Union Bank of Switzerland (UBS) 4.1.5.1 Incident This incident involves losses of between $400 million and $700 million in equity derivatives during 1997, which appear to have been exacerbated by lack of internal controls. A loss of $700 million during 1998 was due to a large position in Long‐Term Capital Management (LTCM). 60 FINANCIAL RISK MANAGEMENT 4.1.5.2 Result The 1997 losses forced UBS into a merger on unfavorable terms with Swiss Bank Corporation (SBC) at the end of 1997. The 1998 losses came after that merger. 4.1.5.3 Were the Positions Unauthorized? Less is known about the UBS disaster than the other incidents discussed in this chapter. Even the size of the losses has never been fully disclosed. Considerable controversy exists about whether the 1997 losses just reflected poor decision making or unlucky outcomes or whether an improper control structure led to positions that management would not have authorized. The 1998 losses were the result of a position that certainly had been approved by the UBS management, but evidence suggests that it failed to receive adequate scrutiny from the firm’s risk controllers and that it was not adequately disclosed to the SBC management that took over the firm. What seems uncontroversial is that the equity derivatives business was being run without the degree of management oversight that would be normally expected in a firm of the size and sophistication of UBS, but there is disagreement about how much this situation contributed to the losses. The equity derivatives department was given an unusual degree of independence within the firm with little oversight by, or sharing of information with, the corporate risk managers. The person with senior risk management authority for the department doubled as head of quantitative analytics. As head of analytics, he was both a contributor to the business decisions he was responsible for reviewing and had his compensation tied to trading results, which are both violations of the fundamental principles of independent oversight. The equity derivative losses appear to have been primarily due to four factors: 1. A change in British tax laws, which impacted the value of some long‐ dated stock options. 2. A large position in Japanese bank warrants, which was inadequately hedged against a significant drop in the underlying stocks (see the fuller description in Section 11.4). 3. An overly aggressive valuation of long‐dated options on equity baskets, utilizing correlation assumptions that were out of line with those used by competitors. 4. Losses on other long‐dated basket options, which may have been due to modeling deficiencies. The first two transactions were ones where UBS had similar positions to many of its competitors so it would be difficult to accuse the firm of excessive risk taking, although its Japanese warrant positions appear to have Financial Disasters 61 been unreasonably large relative to competitors. The last two problems appear to have been more unique to UBS. Many competitors made accusations that its prices for trades were off the market. The losses related to LTCM came as the result of a position personally approved by Mathis Cabiallavetta, the UBS CEO, so they were certainly authorized in one sense. However, accusations have been made that the trades were approved without adequate review by risk control areas and were never properly represented in the firm’s risk management systems. Although about 40 percent of the exposure represented a direct investment in LTCM that had large potential profits to weigh against the risk, about 60 percent of the exposure was an option written on the value of LTCM shares. However, there was no effective way in which such an option could be risk managed given the illiquidity of LTCM shares and restrictions that LTCM placed on UBS delta hedging the position (see the next‐to‐last paragraph in Section 11.1). The imbalance in risk/reward trade‐off for an option that was that difficult to risk manage had caused other investment banks to reject the proposed trade. UBS appears to have entered into the option because of its desire for a direct investment in LTCM, which LTCM tied to agreement to the option. Agreeing to this type of bundled transaction can certainly be a legitimate business decision, but it is unclear whether the full risk of the option had been analyzed by UBS or whether stress tests of the two positions taken together had been performed. 4.1.5.4 Lessons Learned risk oversight. This incident emphasizes the need for independent 4.1.5.5 Further Reading The fullest account of the equity derivative losses is contained in a book by Schutz (2000), which contains many lurid accusations about improper dealings between the equity derivatives department and senior management of the firm. Schutz has been accused of inaccuracy in some of these charges—see Derivative Strategies (1998) for details. There is also a good summary in the January 31, 1998, issue of the Economist. A good account of the LTCM transaction is Shirreff (1998). Lowenstein (2000), an account of the LTCM collapse, also covers the UBS story in some detail. 4.1.6 Société Générale 4.1.6.1 Incident In January 2008, Société Générale reported trading losses of $7.1 billion that the firm attributed to unauthorized activity by a junior trader, Jérôme Kerviel. 62 FINANCIAL RISK MANAGEMENT 4.1.6.2 Result The large loss severely damaged Société Générale’s reputation and required it to raise a large amount of new capital. 4.1.6.3 How the Unauthorized Positions Arose In this section and the next, I am drawing primarily on the Société Générale Special Committee of the Board of Directors Report to Shareholders of May 22, 2008 (I’ll abbreviate references to it as SpecComm) and its accompanying Mission Green Report of the Société Générale General Inspection Department (I’ll abbreviate it as MG). Kerviel took very large unauthorized positions in equities and exchange‐traded futures, beginning in July 2005 and ending when his concealment of positions was uncovered in January 2008. His primary method for concealing these unauthorized positions was to enter fictitious transactions that offset the risk and P&L of his true trades. The fictitious nature of these transactions was hidden mostly by creating transactions with forward start dates and then, relying on his knowledge of when control personnel would seek confirmation of a forward‐dated trade, canceling the trade prior to the date that confirmation would be sought (Kerviel had previously worked in the middle office of the firm, which may have provided him with particular insight into the actions of control personnel). Not surprisingly, given his need to constantly replace canceled fictitious transactions with new ones, there were a large number of these trades, 947 transactions according to MG Focus 4. How was Kerviel able to enter this many fictitious trades before discovery of his fraud? 4.1.6.4 How the Unauthorized Positions Failed to Be Detected Trade Cancellation There was no procedure in place that required control functions to confirm information entered for a trade that was then canceled and Kerviel knew this, nor was there a system in place for red‐flagging an unusual level of trade cancellations. SpecComm, point 10, notes that the back and middle office gave “priority to the correct execution of trades” and showed “an absence of an adequate degree of sensitivity to fraud risks.” The head of equity derivatives at a European bank is quoted as saying, “If he was able to cancel a trade and book a new one before the confirm was sent out, the clock [for obtaining confirmation] would start again. But at our bank, we actively monitor cancel‐and‐correct activity for each trader, which is standard practice at most institutions. It would stick out like a sore thumb if you had one trader who was perpetually cancelling and correcting trades” (Davies 2008). Hugo Banziger, chief risk officer of Deutsche Bank, is quoted as saying, “Routine IT controls can monitor unusual trades put on and cancelled—this is a particularly effective control mechanism” (Davies 2008). It Financial Disasters 63 certainly appears from the account in MG that no such procedures were in place at Société Générale, and even the inquiry to confirm the counterparty on a canceled trade that eventually led to Kerviel’s downfall in January 2008 appears to have been a matter of chance (MG Focus 6). Supervision Kerviel’s immediate manager resigned in January 2007. For two and a half months, until the manager was replaced, Kerviel’s positions were validated by his desk’s senior trader. Day‐to‐day supervision of Kerviel by the new manager, who started in April 2007, was weak (SpecComm, point 9; MG, page 6). While Kerviel had begun his fraudulent activities prior to January 2007, the size of his unauthorized positions increased explosively at this time (MG Focus 10). Trading Assistant The trading assistant who worked with Kerviel in entering trades, who would have the most immediate potential access to seeing how he was manipulating the trading system, may have been operating in collusion with Kerviel. This has not been confirmed (MG, page 3, notes that this is an allegation under investigation by the courts), but, in any case, the trading assistant appears to have accepted Kerviel’s directions without questioning. This would have helped Kerviel’s credibility with control functions, since the trading assistant reported to a control function and was the primary point of contact of other control functions regarding Kerviel’s positions (MG, page 4). Vacation Policy The normal precaution of forcing a trader to take two consecutive weeks of vacation in a year, during which time his positions would be managed by another trader, was not followed (MG, page 7). This control could easily have caused the collapse of a scheme based on constant rolling forward of fraudulent trading entries. Gross Positions There were no limits or other monitoring of Kerviel’s gross positions, only his net positions (SpecComm, point 10, notes the “lack of certain controls liable to identify the fraudulent mechanisms, such as the control of the positions’ nominal value”). Had gross positions been monitored, this would have revealed the abnormally large size of his activities and might have raised suspicions as to what the purpose was of such large positions. Henning Giescke, chief risk officer of the UniCredit Group, is quoted as saying, “In high‐volume businesses, banks have to look at gross as well as net position. This allows an institution to look at each trader’s book to see whether they are taking too much risk, regardless of whether the net position is neutral” (Davies 2008). The chief risk officer of a UK bank is quoted as saying, “To effectively manage basis risk, you have to be able to 64 FINANCIAL RISK MANAGEMENT see how the outright position—the notional—performs against the hedge. It is inconceivable such a sophisticated institution could have missed this. Modern systems are able to stress‐test positions, and to do this you automatically need the notional amount” (Davies 2008). Kerviel’s unusually high amount of brokerage commissions (MG, page 6), related to his high level of gross positions, could also have provided a warning sign. Cash and Collateral The use of fictitious transactions to conceal positions will often create positions of unusual size in cash and required collateral— since the fictitious trades do not generate any cash or collateral movements, there is nothing to balance out the cash and collateral needs of the real trades. This provides good opportunities for fraud detection. The reason that Société Générale’s control functions did not respond to these clues is that cash and collateral reports and inquiry procedures lacked sufficient granularity to detect unusual movements at the level of a single trader (MG Focus 13). P&L Concealment of trading positions will not always lead to unusual earnings patterns. A trader who is trying to conceal losses may be satisfied simply to show a small positive P&L. But some fraudulent traders will show unusual profits, either because their unauthorized positions have resulted in large gains for which they want to be compensated or because their success in hiding losing positions encourages them to also claim some phantom gains to fund bonuses. Kerviel was reporting trading gains in excess of levels his authorized position taking could have accounted for, and this should have given his management and the control functions a warning sign to investigate closely the source of his earnings (MG Focus 12). These warning signs were apparently not pursued. 4.1.6.5 How the Unauthorized Positions Were Eventually Detected One of Kerviel’s fictitious trades was identified as fabricated by control personnel as part of routine monitoring of positions, leading to a thorough investigation. Kerviel’s attempts to deflect the inquiry by forging confirmations proved fruitless. It appears that it was just a matter of chance that this particular inquiry led to identification of the fraud. 4.1.6.6 Lessons to Be Learned What new lessons can we draw from this control failure? From one point of view, the answer is not much—Kerviel’s methods for eluding scrutiny of his positions were very close to those used in previous incidents such as those of Kidder Peabody, Barings, and Allied Irish Bank. But, from another viewpoint, we can learn quite a bit, since clear patterns are emerging when we look across episodes. Financial Disasters 65 The obvious lessons for control personnel are to tighten procedures that may lead to detection of fictitious trade entries. Corresponding to the points raised in Section 4.1.6.4, the specific lessons follow. Trade Cancellation Institute systems for monitoring patterns of trade cancellation. Flag any trader who appears to be using an unusually high number of such cancellations. Any trader flagged should have a reasonably large sample of the cancellations checked to make sure that they represent real trades by checking details of the transaction with the counterparty. Supervision Control personnel should be aware of situations in which traders are being supervised by temporary or new managers. Tightened control procedures should be employed. Trading Assistant Control personnel must remember that even in situations where there is no suspicion of dishonesty, trading assistants are often under intense pressure from the traders with whom they work closely. Their job performance ratings and future career paths often depend on the trader, regardless of official reporting lines. The greater prestige, experience, and possible bullying tactics of a trader can often convince a trading assistant to see things from the trader’s viewpoint. Other control personnel must be cognizant of these realities and not place too much reliance on the presumed independence of the trading assistant. Vacation Policy enforced. Rules for mandatory time away from work should be Gross Positions Gross positions must be monitored and highlighted in control reports. This is particularly important since unusually high ratios of gross to net positions are a warning sign of potentially inadequately measured basis risk as well as a possible flag for unauthorized activities. The Kidder Peabody and Allied Irish Bank frauds could also have been uncovered by investigating unusually high ratios of gross to net trading. Cash and Collateral Cash and collateral requirements should be monitored down to the individual trader level. Better monitoring of cash and credit flows would have also been instrumental in uncovering the Barings and Allied Irish Bank frauds. P&L Any patterns of P&L that are unusual relative to expectations need to be identified and investigated by both management and the control functions. Identification of unusual patterns can be comparisons to historical 66 FINANCIAL RISK MANAGEMENT experience, to budgeted targets, and to the performance of traders with similar levels of authority. Investigation of suspicious earnings patterns could also have led to earlier discovery of the Kidder Peabody and Barings frauds. 4.1.6.7 Further Reading I have relied primarily on the Société Générale Special Committee of the Board of Directors Report to Shareholders (2008) and its accompanying Société Générale Mission Green Report (2008). 4.1.7 Other Cases Other disasters involving unauthorized positions are covered more briefly, because they had less of an impact on the firm involved, because it is harder to uncover details on what occurred, or because they do not have any lessons to teach beyond those of the cases already discussed: ■ ■ ■ Toshihida Iguchi of Daiwa Bank’s New York office lost $1.1 billion trading Treasury bonds between 1984 and 1995. He hid his losses and made his operation appear to be quite profitable by forging trading slips, which enabled him to sell without authorization bonds held in customer accounts to produce funds he could claim were part of his trading profit. His fraud was aided by a situation similar to Nick Leeson’s at Barings—Iguchi was head of both trading and the back‐office support function. In addition to the losses, Daiwa lost its license to trade in the United States, but this was primarily due to its failure to promptly disclose the fraud once senior executives of the firm learned of it. A more detailed account of this by Rob Jameson of ERisk can be found on their website, www.erisk.com. The Sumitomo Corporation of Japan lost $2.6 billion in a failed attempt by Yasuo Hamanaka, a senior trader, to corner the world’s copper market—that is, to drive up prices by controlling a large portion of the available supply. Sumitomo management claimed that Hamanaka had employed fraudulent means in hiding the size of his positions from them. Hamanaka claimed that he had disclosed the positions to senior management. Hamanaka was sent to jail for his actions. The available details are sketchy, but some can be found in Dwyer (1996), Asiaweek (1996), Kooi (1996), and McKay (1999). Askin Capital Management and Granite Capital, hedge funds that invested in mortgage securities, went bankrupt in 1994 with losses of $600 million. It was revealed that David Askin, the manager of the funds, was valuing positions with his own marks substituted for dealer quotes and using these position values in reports to investors in the Financial Disasters ■ ■ ■ 67 funds and in marketing materials to attract new clients. For a brief discussion, see Mayer (1997). Merrill Lynch reportedly lost $350 million in trading mortgage securities in 1987, due to risk reporting that used a 13‐year duration for all securities created from a pool of 30‐year mortgages. Although this duration is roughly correct for an undivided pool of 30‐year mortgages, the correct duration is 30 years when the interest‐only (IO) part is sold and the principal‐only (PO) part is kept, as Merrill was doing. See Crouhy, Galai, and Mark (2001). National Westminster Bank in 1997 reported a loss on interest rate caps and swaptions of about $140 million. The losses were attributed to trades dating back to 1994 and had been masked by deliberate use by traders of incorrect volatility inputs for less liquid maturities. The loss of confidence in management caused by this incident may have contributed to NatWest’s sale to the Royal Bank of Scotland. I have heard from market sources that the traders were taking advantage of the middle‐ office saving costs by checking only a sample of volatility marks against market sources, although it is unclear how the traders were able to determine in advance which quotes would be checked. A more detailed account is Wolfe (2001). The large Swiss bank UBS in 2011 reported a loss of $2.3 billion due to unauthorized trading by Kweku Adoboli, a relatively junior equity trader. This incident cost the CEO of UBS his job. Adoboli’s ability to enter into unauthorized trades appears to have been engineered by means very similar to those of Kerviel in the Société Générale incident discussed in Section 4.1.6. He took advantage of intimate knowledge of back‐office control procedures to identify a loophole. Trades with forward settlement greater than 15 days were not being immediately confirmed with counterparties; confirmation was delayed until closer to the settlement date. If the trade was canceled prior to the date on which the confirmation would have been confirmed, no confirmation ever took place. Adoboli appears to have been able to utilize this loophole to disguise his real positions by entering bogus offsetting forward positions and then canceling the fictitious positions prior to the date they would have been confirmed, replacing them with new fictitious forwards. For this to have gone on for any period of time, there must have also been flaws in UBS’s monitoring of excessive cancellations. Due to an ongoing criminal prosecution against Adoboli at the time of my writing, not many public details are available. Wilson (2011) is a good summary of what is known about the mechanics of the unauthorized trades, and Broom (2011) summarizes the devastating impact the revelation of this faulty control environment had on UBS. 68 4.2 FINANCIAL RISK MANAGEMENT DISASTERS DUE TO LARGE MARKET MOVES We will now look at financial disasters that were not caused by incorrect position information, but were caused by unanticipated market moves. The first question that should be asked is: How is a disaster possible if positions are known? No matter what strategy is chosen, as losses start to mount beyond acceptable bounds, why aren’t the positions closed out? The answer is lack of liquidity. We will focus on this aspect of these disasters. 4.2.1 Long‐Term Capital Management (LTCM) The case we will consider at greatest length is that of the large hedge fund managed by Long‐Term Capital Management (LTCM), which came close to bankruptcy in 1998. In many ways, it represents an ideal example for this type of case since all of its positions were marked to a market value daily, the market values were supplied by the dealers on the other end of each trade, no accusations have been made of anyone at LTCM providing misleading information about positions taken, and the near failure came in the midst of some of the largest market moves in recent memory. To review the facts, LTCM had been formed in 1994 by about a dozen partners. Many of these partners had previously worked together at Salomon Brothers in a highly successful proprietary trading group. Over the period from 1994 until early 1998, the LTCM fund produced quite spectacular returns for its investors. From the beginning, the partners made clear that they would be highly secretive about the particulars of their investment portfolio, even by the standard of other hedge funds. (Since hedge funds are open only to wealthy investors and cannot be publicly offered the way mutual funds are, they are not subject to legal requirements to disclose their holdings.) Within the firm, however, the management style favored sharing information openly, and essentially every investment decision was made by all the partners acting together, an approach that virtually eliminates the possibility of a rogue trader making decisions based on information concealed from other members of the firm. Although it is true that outside investors in the fund did not have access to much information beyond the month‐end valuation of its assets and the track record of its performance, it is equally true that the investors knew these rules prior to their decision to invest. Since the partners who managed the fund were such strong believers in the fund that they had invested most of their net worth in it (several even borrowed to invest more than their net worth), their incentives were closely aligned with investors (in other words, there was little room for moral hazard). If anything, the concentration of partner assets in the fund should Financial Disasters 69 have led to more risk‐averse decision making than might have been optimal for outside investors, who invested only a small portion of their wealth in the fund, with the exception of UBS, discussed in Section 4.1.5. In fact, even if investors had been given access to more information, there is little they could have done with it, since they were locked into their investments for extended time periods (generally, three years). This reflected the basic investment philosophy of LTCM, which was to locate trading opportunities that represented what the partners believed were temporary disruptions in price relationships due to short‐term market pressures, which were almost certain to be reversed over longer time periods. To take advantage of such opportunities, they needed to know they had access to patient capital that would not be withdrawn if markets seemed to be temporarily going against them. This also helped to explain why LTCM was so secretive about its holdings. These were not quick in‐and‐out trades, but long‐term holdings, and they needed to prevent other firms from learning the positions and trading against them. The following are two examples of the types of positions the LTCM fund was taking: 1. LTCM was long U.S. interest rate swaps and short U.S. government bonds at a time when these spreads were at historically high levels. Over the life of the trade, this position will make money as long as the average spread between the London Interbank Offered Rate (LIBOR) at which swaps are reset (see Section 10.1.6) and the repurchase agreement (RP) rates at which government bonds are funded (see Section 10.1.2) is not higher than the spread at which the trade was entered into. Over longer time periods, the range for the average of LIBOR‐RP spreads is not that wide, but in the short run, swap spreads can show large swings based on relative investor demand for the safety of governments versus the higher yield of corporate bonds (with corporate bond issuers then demanding interest rate swaps to convert fixed debt to floating debt). 2. LTCM sold equity options at historically high implied volatilities. Over the life of the trade, this position will make money if the actual volatility is lower than the implied volatility, but in the short run, investor demand for protection against stock market crashes can raise implied volatilities to very high levels. Perold (1999a) presents further analysis of why LTCM viewed these trades as excellent long‐term investments and presents several other examples of positions it entered into. One additional element was needed to obtain the potential returns LTCM was looking for. LTCM needed to be able to finance positions for longer terms in order to be able to ensure there was no pressure on them to 70 FINANCIAL RISK MANAGEMENT sell positions before they reached the price relationships LTCM was waiting for. However, the banks and investment banks who financed hedge fund positions were the very competitors that they least wanted to share information on holdings with. How were they to persuade firms to take credit risk without knowing much about the trading positions of the hedge fund? To understand why the lenders were comfortable in doing this, we need to digress a moment into how credit works in a futures exchange. A futures exchange (see Section 14.2) represents the extreme of being willing to lend without knowledge of the borrower. Someone who purchases, for example, a bond for future delivery needs to deposit only a small percentage of the agreed purchase price as margin and does not need to disclose anything about one’s financial condition. The futures exchange is counting on the nature of the transaction itself to provide confidence that money will not be lost in the transaction. This is because anytime the value of the bond falls, the purchaser is required to immediately provide added margin to fully cover the decline in value. If the purchaser does not do so, the position is closed out without delay. Loss is possible only if the price has declined so much since the last time the price fell and margin was added that the incremental price drop exceeds the amount of initial margin or if closing out the option results in a large price move. The probability of this occurring is kept low by setting initial margins high enough, restricting the size of position that can be taken by any one investor, and designing futures contracts to cover sufficiently standardized products to ensure enough liquidity that the closing out of a trade will not cause a big price jump. LTCM wanted to deal in over‐the‐counter markets as well as on futures exchanges partly because it wanted to deal in some contracts more individually tailored than those available on futures exchanges and partly because of the position size restrictions of exchanges. However, the mechanism used to assure lenders in over‐the‐counter markets is similar—there is a requirement to cover declines in market value by immediately putting up cash. If a firm fails to put up the cash, then positions are closed out. LTCM almost always negotiated terms that avoided posting the initial margin. Lenders were satisfied with the lack of initial margin based on the size of the LTCM fund’s equity, the track record of its excellent returns, and the firm’s recognized investment management skills. Lenders retained the option of demanding initial margin if fund equity fell too much. This dependence on short‐term swings in valuation represented a potential Achilles’ heel for LTCM’s long‐term focused investment strategy. Because the firm was seeking opportunities where market pressures were causing deviation from long‐run relationships, a strong possibility always existed that these same market pressures would push the deviation even further. LTCM would then immediately need to come up with cash to fund the change in Financial Disasters 71 market valuation. This would not be a problem if some of the trades were moving in its favor at the same time as others were moving against it, since LTCM would receive cash on upswings in value to balance having to put up cash on downswings (again, the same structure as exchange‐traded futures). However, if many of its trades were to move against it in tandem, LTCM would need to raise cash quickly, either from investors or by cutting positions. In the actual events of August and September 1998, this is exactly what led to LTCM’s rapid downfall. The initial trigger was a combination of the Russian debt default of August, which unsettled the markets, and the June 1998 decision by Salomon Brothers to liquidate proprietary positions it was holding, which were similar to many of those held by LTCM. The LTCM fund’s equity began to decline precipitously from $4.1 billion as of the end of July 1998, and it was very reluctant to cut positions in a turbulent market in which any large position sale could easily move the valuations even further against it. This left the option of seeking new equity from investors. LTCM pursued this path vigorously, but the very act of doing so created two perverse effects. First, rumors of LTCM’s predicament caused competitors to drive market prices even further against what they guessed were LTCM’s positions, in anticipation of LTCM being forced to unload the positions at distressed prices. Second, to persuade potential investors to provide new money in the midst of volatile markets, LTCM was forced to disclose information about the actual positions it held. As competitors learned more about the actual positions, their pressure on market prices in the direction unfavorable to LTCM intensified. As market valuations continued to move against LTCM and the lack of liquidity made it even more unlikely that reducing positions would be a viable plan, it became increasingly probable that in the absence of a truly large infusion of new equity, the LTCM fund would be bankrupt. Its creditors started to prepare to close out LTCM’s positions, but quickly came to fear that they were so large and the markets were so illiquid that the creditors would suffer serious losses in the course of doing so. The lenders were also concerned that the impact of closing out these positions would depress values in the already fragile markets and thereby cause considerable damage to other positions held by the creditors and other investment firms they were financing. Ultimately, 14 of the largest creditors, all major investment banks or commercial banks with large investment banking operations, contributed a fresh $3.65 billion in equity investment into the LTCM fund to permit the firm to keep operating and allow for a substantial time period in which to close out positions. In return, the creditors received substantial control over fund management. The existing investors had their investments valued at 72 FINANCIAL RISK MANAGEMENT the then current market value of $400 million, so they had only a 10 percent share in the positions of the fund. Although some of the partners remained employed to help wind down investments, it was the consortium of 14 creditors who now exercised control and insisted on winding down all positions. As a result, the markets calmed down. By 2000, the fund had been wound down with the 14 creditors having recovered all of the equity they had invested and having avoided any losses on the LTCM positions they had held at the time of the bailout. This outcome lends support to two propositions: LTCM was largely right about the long‐term values underlying its positions, and the creditors were right to see the primary problem as one of liquidity, which required patience to ride out. Please note that the bailout was not primarily a rescue of LTCM’s investors or management, but a rescue of LTCM’s creditors by a concerted action of these creditors. Even recently, I continue to encounter the view that the bailout involved the use of U.S. government funds, helped the LTCM investors and management avoid the consequences of their mistakes, and therefore contributed to an attitude that some firms are “too big to fail” and so can afford to take extra risks because they can count on the government absorbing some of their losses. I do not think evidence is available to support any of these claims. Interested readers can form their own conclusions by looking at the detailed account of the negotiations on the rescue package in Lowenstein (2000). An opposing viewpoint can be found in Shirreff (2000). The only government involvement was some coordination by the Federal Reserve, acting out of legitimate concern for the potential impact on the financial markets. The LTCM creditors took a risk by investing money in the fund, but did so in their own self‐interest, believing (correctly, as it turns out) that they were thereby lowering their total risk of loss. LTCM’s investors and managers had little left to lose at the point of the bailout since they could not lose more than their initial investment. It is true that, without a rescue, the fund would have been liquidated, which would have almost certainly wiped out the remaining $400 million market value of the investors. However, in exchange for this rescue, they were able to retain only a 10 percent interest in the fund’s positions, since the $3.65 billion in new investment was explicitly not being used to enable new trades, but only to wind down the existing positions. LTCM management was certainly aware of the potential for short‐term market movements to disrupt the fund’s fundamental trading strategy of focusing on longer‐term relationships. The firm tried to limit this risk by insisting that its positions pass value at risk (VaR) tests based on whether potential losses over one month due to adverse market moves would reduce equity to unacceptable levels. Where LTCM seems to have fallen short of Financial Disasters 73 best practices was a failure to supplement VaR measures with a full set of stress test scenarios (see Section 11.2). It did run stress versions of VaR based on a higher than historical level of correlations, but it is doubtful that this offers the same degree of conservatism as a set of fully worked‐through scenarios. A lesson that all market participants have learned from the LTCM incident is that a stress scenario is needed to look at the impact of a competitor holding similar positions exiting the market, as when Salomon decided to cut back on proprietary trading. However, even by best practice standards of the time, LTCM should have constructed a stress test based on common economic factors that could cause impacts across its positions, such as a flight to quality by investors, which would widen all credit spreads, including swap spreads, and increase premiums on buying protection against stock market crashes, hence increasing option volatility. Another point on which LTCM’s risk management could be criticized is a failure to account for the illiquidity of its largest positions in its VaR or stress runs. LTCM knew that the position valuations it was receiving from dealers did not reflect the concentration of some of LTCM’s positions, either because dealers were not taking liquidity into account in their marks or because each dealer knew only a small part of LTCM’s total size in its largest positions. Two other criticisms have been made of LTCM’s management of risk with which I disagree. One is that a simple computation of leverage would have shown that LTCM’s positions were too risky. However, as will be seen in Section 13.2.4, leverage by itself is not an adequate measure of risk of default. It must be multiplied by volatility of the firm’s assets. But this just gets us back to testing through VaR or stress scenarios. The second criticism is that LTCM showed unreasonable faith in the outcome of models. I see no evidence to support this claim. Major positions LTCM entered into—U.S. swap spreads to narrow, equity volatilities to decline—were ones that many proprietary position takers had entered into. For example, the bias in equity implied volatilities due to demand for downside protection by shareholders had long been widely recognized as a fairly certain profit opportunity for investors with long‐enough time horizons. That some firms made more use of models to inform their trading judgments while others relied more on trader experience tells me nothing about the relative quality of their decision making. Most of the focus of LTCM studies has been on the decision making of LTCM management and the losses of the investors. I believe this emphasis is misplaced. It is a fairly common occurrence, and to be expected, that investment funds will have severe drops in valuation. The bankruptcy of an investment fund does not ordinarily threaten the stability of the 74 FINANCIAL RISK MANAGEMENT financial system the way the bankruptcy of a firm that makes markets or is a critical part of the payments system would. It just represents the losses of a small number of investors. Nor is there a major difference in consequences between bankruptcy and a large loss short of bankruptcy for an investment fund. It shouldn’t matter to investors whether a fund in which they have invested $10 million goes bankrupt or a fund in which they have invested $30 million loses a third of its value. By contrast, losses short of bankruptcy hurt only the stockholders of a bank, whereas bankruptcy of a bank could hurt depositors and lead to loss of confidence in the banking system. The reason that an LTCM failure came close to disrupting the financial markets and required a major rescue operation was its potential impact on the creditors to LTCM, so we need to take a closer look at their role in the story. In retrospect, the creditors to LTCM believed they had been too lax in their credit standards, and the incident triggered a major industry study of credit practices relating to trading counterparties (Counterparty Risk Management Policy Group 1999). Some suggestions for improved practices, many of which are extensively addressed in this study, have been: ■ ■ ■ A greater reluctance to allow trading without initial margin for counterparties whose principal business is investing and trading. A counterparty that has other substantial business lines—for example, auto manufacturing or retail banking—is unlikely to have all of its economic resources threatened by a large move in financial markets. However, a firm that is primarily engaged in these markets is vulnerable to illiquidity spreading from one market to another as firms close out positions in one market to meet margin calls in another market. For such firms, initial margin is needed as a cushion against market volatility. Factoring the potential costs of liquidating positions in an adverse market environment into estimates of the price at which trades can be unwound. These estimates should be based on the size of positions as well as the general liquidity of the market (see Section 6.1.2). These potential liquidation costs should impact estimates of the amount of credit being extended and requirements for initial margin. A push for greater disclosure by counterparties of their trading strategies and positions. Reliance on historical records of return as an indicator of the volatility of a portfolio can be very misleading because it cannot capture the impact of changes in trading style (see Section 7.1). Increased allowance for liquidation costs of positions will be very inexact if the creditor only knows the positions that a counterparty holds with the creditor without knowing the impact of other positions held. Financial Disasters ■ 75 To try to deal with counterparties’ legitimate fears that disclosure of their positions will lead to taking advantage of this knowledge, creditors are implementing more stringent internal policies against the sharing of information between the firm’s credit officers and the firm’s traders. Better use of stress tests in assessing credit risk. To some extent, this involves using more extreme stresses than were previously used in measuring risk to reflect the increased market volatility that has been experienced in recent years. However, a major emphasis is also on more integration of market risk and credit risk stress testing to take into account overlap in risks. In the LTCM case, this would have required recognition by a creditor to LTCM that many of the largest positions being held by LTCM were also being held by other investment funds to which the firm had counterparty credit exposure, as well as by the firm’s own proprietary traders. A full stress test would then look at the losses that would be incurred by a large market move and subsequent decrease in liquidity across all of these similar positions. A complete account of the LTCM case covering all aspects of the history of the fund and its managers, the involvement of creditors, and the negotiations over its rescue can be found in Lowenstein (2000). The Harvard Business School case studies of Perold (1999a, 1999b) provide a detailed but concise analysis of the fund’s investment strategy and the dilemma that it faced in August 1998. 4.2.2 Metallgesellschaft (MG) The disaster at Metallgesellschaft (MG) reveals another aspect of liquidity management. In 1992, an American subsidiary of MG, Metallgesellschaft Refining and Marketing (MGRM), began a program of entering into long‐ term contracts to supply customers with gas and oil products at fixed costs and to hedge these contracts with short‐term gas and oil futures. Although some controversy exists about how effective this hedging strategy was from a P&L standpoint, as we’ll discuss in just a moment, the fundamental consequence of this strategy for liquidity management is certain. The futures being used to hedge were exchange‐traded instruments requiring daily cash settlement, as explained in Section 10.1.4. The long‐term contracts with customers involved no such cash settlement. So no matter how effective the hedging strategy was, the consequence of a large downward move in gas and oil prices would be to require MGRM to pay cash against its futures positions that would be offset by money owed to MGRM by customers who would be paid in the future. 76 FINANCIAL RISK MANAGEMENT A properly designed hedge will reflect both the cash paid and the financing cost of that cash during the period until the customer payment is due and hence will be effective from a P&L standpoint. However, the funding must still be obtained, which can lead to funding liquidity risk (see Section 3.5). As we will discuss in Section 6.1.6, such cash needs must be planned in advance. Limits need to be set on positions based on the amount of cash shortfall that can be funded. It appears that MGRM did not communicate to its parent company the possible need for such funding. In 1993, when a large decrease in gas and oil prices had resulted in funding needs of around $900 million, the MG parent responded by closing down the futures positions, leaving unhedged exposure to gas and oil price increases through the customer contracts. Faced with this open exposure, MG negotiated unwinds of these contracts at unfavorable terms. It may be that MG, with lack of advance warning as to possible cash needs, responded to the demand for cash as a sign that the trading strategy was deeply flawed; if only Barings’ management had reacted similarly. As mentioned earlier, the MG incident spurred considerable debate as to whether MGRM’s trading strategy was reasonable or fundamentally flawed. Most notably, Culp and Miller (1995a) wrote an article defending the reasonableness of the strategy, and Mello and Parsons (1995) wrote an article attacking the Culp and Miller conclusions, which were then defended by Culp and Miller (1995b). Although it is difficult to settle the factual arguments about the particular events in the MG case, I believe the following lessons can be drawn: ■ ■ ■ It is often a key component of a market maker’s business strategy to extend available liquidity in a market (see Section 10.2.2). This requires the use of shorter‐term hedges against longer‐term contracts. Experience shows that this can be successfully carried out when proper risk controls are applied. The uncertainty of roll cost is a key risk for strategies involving shorter‐ term hedges against longer‐term risk. As explained in Section 10.2.2, this requires the use of valuation reserves based on conservative assumptions of future roll cost. MGRM does not appear to have utilized valuation reserves; it just based its valuation on the historical averages of roll costs. A firm running short‐term hedges against longer‐term risk requires the flexibility to choose the shorter‐term hedge that offers the best trade‐ off between risk and reward. It may legitimately choose to follow a hedging strategy other than a theoretical minimum variance hedge, or choose not to hedge with the longest future available, based on liquidity Financial Disasters 77 considerations, or take into account the expectation of positive roll cost as part of potential return. It is not reasonable to conclude, as Mello and Parsons (1995) do, that these choices indicate that the firm is engaged in pure speculation rather than hedging. At the same time, regardless of a firm’s conclusions about probable return, its assessment of risk should include valuation reserves, as in the previous point, and volume limits based on reasonable stress testing of assumptions. 4.3 DISASTERS DUE TO THE CONDUCT OF CUSTOMER BUSINESS In this section, we focus on disasters that did not involve any direct financial loss to the firm, but were completely a matter of reputational risk due to the conduct of customer business. 4.3.1 Bankers Trust (BT) The classic case of this type is the Bankers Trust (BT) incident of 1994, when BT was sued by Procter & Gamble (P&G) and Gibson Greetings. Both P&G and Gibson claimed that they had suffered large losses in derivatives trades they had entered into with BT due to being misled by BT as to the nature of the positions. These were trades on which BT had little market or credit risk, since it had hedged the market risk on them with other derivatives and there was no credit issue of P&G or Gibson being unable to pay the amount they owed. However, the evidence uncovered in the course of legal discovery for these lawsuits was severely damaging to BT’s reputation for fair business dealing, led to the resignation of the firm’s CEO, and ultimately had such negative consequences for the bank’s ability to do business that it was forced into an acquisition by Deutsche Bank, which essentially amounted to a dismemberment of the firm. The exact terms of these derivative trades were quite complex and are not essential to understanding the incident. Interested readers are referred to Chew (1996, Chapter 2) for details. The key point is that the trades offered P&G and Gibson a reasonably probable but small reduction in funding expenses in exchange for a potentially large loss under some less probable circumstances. P&G and Gibson had been entering into such trades for several years prior to 1994 with good results. The derivatives were not tailored to any particular needs of P&G or Gibson in the sense that the circumstances under which the derivatives would lose them money were not designed to coincide with cases in which other P&G or Gibson positions would be making money. Their objective was just to reduce expected funding costs. Since 78 FINANCIAL RISK MANAGEMENT the only way to reduce costs in some cases is to raise them in others, P&G and Gibson can be presumed to have understood that they could lose money under some economic circumstances. On what basis could they claim that BT had misled them? One element that established some prima facie suspicion of BT was the sheer complexity of the structures. It was hard to believe that BT’s clients started out with any particular belief about whether there was a small enough probability of loss in such a structure to be comfortable entering into it. BT would have had to carefully explain all the intricacies of the payoffs to the clients for them to be fully informed. Since it was quite clear that the exact nature of the structures hadn’t been tailored to meet client needs, why had BT utilized so complex a design? The most probable reason was that the structures were designed to be complex enough to make it difficult for clients to comparison shop the pricing to competitor firms. However, this also made the clients highly dependent on BT on an ongoing basis. If they wanted to unwind the position, they couldn’t count on getting a competitive quote from another firm. BT claimed that it had adequately explained all the payoffs and risks to P&G and Gibson. But then came the discovery phase of the lawsuit. BT, like all trading firms, recorded all phone lines of traders and marketers as a means of resolving disputes about verbal contracts (see Sections 3.1.1 and 3.2). However, this recording also picked up internal conversations among BT personnel. When subpoenaed, they produced evidence of BT staff boasting of how thoroughly they had fooled the clients as to the true value of the trades and how little the clients understood the true risks. Further, the internal BT recordings showed that price quotes to P&G and Gibson were being manipulated to mislead them. At first, they were given valuations of the trades that were much too high to mask the degree of profit BT was able to book up front. Later, they were given valuations that were too low because this was BT’s bid at which to buy back the trade or swap it into a new trade offering even more profit to BT. For more details on what was revealed in the recordings, see Holland and Himelstein (1995). The BT scandal caused all financial firms to tighten up their procedures for dealing with customers, both in better controls on matching the degree of complexity of trades to the degree of financial sophistication of customers and in providing for customers to obtain price quotes from an area independent of the front office. These measures were detailed in Section 3.3. Another lesson that you would think would be learned is to be cautious about how you use any form of communication that can later be made public. BT’s reputation was certainly hurt by the objective facts about its conduct, but it was even further damaged by the arrogant and insulting tone some of its employees used in referring to clients, which could be Financial Disasters 79 documented through recorded conversations. However, even with such an instructive example, we have seen Merrill Lynch’s reputation being damaged in 2002 by remarks its stock analysts made in e‐mails and tape‐recorded conversations (see the article “Value of Trust” in the June 6, 2002, Economist) along with a number of similar incidents surrounding Wall Street’s relations with Enron (see the article “Banks on Trial” in the July 25, 2002, Economist). 4.3.2 JPMorgan, Citigroup, and Enron Following the Bankers Trust incident, investment banks put in controls to guard against exploitation of customers. But it was not seen as part of a bank’s responsibility to safeguard others from actions by the customer. This has changed as part of the fallout from Enron’s 2001 bankruptcy. As part of the process leading up to the bankruptcy, it was revealed that Enron had for years been engaging in dubious accounting practices to hide the size of its borrowings from investors and lenders (it was their part in these shenanigans that brought an end to the major accounting firm Arthur Andersen). One of the ploys that Enron had used was to disguise a borrowing as an oil futures contract. As a major player in the energy markets, it was to be expected that Enron would be heavily engaged in futures contracts on oil. But these particular futures contracts did not involve taking any position on oil price movements. Enron sold oil for future delivery, getting cash, and then agreed to buy back the oil that it delivered for a fixed price. So, in effect, no oil was ever delivered. When you canceled out the oil part of the trades, what was left was just an agreement for Enron to pay cash later for cash it had received up front—in practice, if not in legal terms, a loan. The advantage to Enron was that it did not have to report this in its public statements as a loan, making the firm appear more desirable as an investment and as a borrower. When this was finally disclosed, JPMorgan Chase and Citigroup, Enron’s principal counterparties on these trades, justified their activities by saying that they had not harmed Enron, their client, in any way, and that they had no part in determining how Enron had accounted for the transactions on its books; that was an issue between Enron and Arthur Andersen. JPMorgan and Citigroup had treated these transaction as loans in their own accounting and reporting to regulators, so they had not deceived their own investors or lenders. But both JPMorgan and Citigroup clearly knew what Enron’s intent was in entering into the transaction. In the end, they agreed to pay a combined $286 million for “helping to commit a fraud” on Enron’s shareholders. They 80 FINANCIAL RISK MANAGEMENT also agreed to put new controls in place to ascertain that their clients were accounting for derivative transactions with them in ways that were transparent to investors. The precedent of this successful legal action caused other investment banks to commit to similar new controls. And yet we have recently witnessed charges against Goldman Sachs for helping Greece hide its level of indebtedness from its European Union partners by disguising debt as an interest rate swap, a mechanism very similar to that in the Enron case. The details here are that the swap was deliberately done at an off‐market rate, creating an up‐front payment to Greece that would of course need to be paid back by Greece, with suitable interest, over the course of the swap’s life. The only reason for creating the swap at an off‐market rate would appear to be letting Greece take out a loan that didn’t need to show up on its books. Details on the Enron case can be found in McLean and Elkind (2003, 159–160, 407–408). Details on the Greek case can be found in Dunbar and Martinuzzi (2012). 4.3.3 Other Cases The following are some examples of other cases in which firms damaged their reputations by the manner in which they dealt with customers: ■ ■ ■ ■ Prudential‐Bache Securities was found to have seriously misled thousands of customers concerning the risk of proposed investments in limited partnerships. In addition to damaging its reputation, Prudential‐ Bache had to pay more than $1 billion in fines and settlements. An account of this incident can be found in Eichenwald (1995). In 1995, a fund manager at Morgan Grenfell Asset Management directed mutual fund investments into highly speculative stocks, utilizing shell companies to evade legal restrictions on the percentage of a firm’s stock that could be owned by a single fund. In addition to damage to its reputation, Morgan Grenfell had to pay roughly $600 million to compensate investors for resulting losses. A brief case account can be found in Garfield (1998). JPMorgan’s reputation was damaged by allegations that it misled a group of South Korean corporate investors as to the risk in derivative trades that lost hundreds of millions of dollars based on the precipitous decline in the Thai baht exchange rate against the dollar in 1997. An account of these trades and the ensuing lawsuits can be found in Gillen, Lee, and Austin (1999). Many investment banks had their reputations damaged in the events leading up to the large fall in value of technology stocks in 2001 and Financial Disasters 81 2002. Evidence showed that some widely followed stock market analysts working at investment banks had issued favorable recommendations for companies as a quid pro quo for underwriting business, with analyst bonuses tied to underwriting business generated. Regulators responded with fines for firms, bans from the industry for some analysts, and requirements for separation of the stock analysis function from the underwriting business. A summary account with references can be found in Lowenstein (2004, 212–213). Reputational risk incidents that arose in connection with the 2007– 2008 crisis are covered in Sections 5.2.1, 5.2.2, and 5.2.3. CHAPTER 5 The Systemic Disaster of 2007–2008 5.1 OVERVIEW There can be little question that the global financial disaster of 2007–2008 stemmed fundamentally from events in the market for collateralized debt obligations (CDOs) backed by subprime mortgages. Firms that failed or needed government rescue either had large losses in these CDOs or else got caught up in events triggered by the difficulties of firms that did have large losses on these CDOs. In examining the crisis, this chapter therefore begins with a section (5.2) focusing on CDOs backed by subprime mortgages. Section 5.3 looks at how this crisis then spread from the institutions with heavy losses in the CDO market to other institutions—by contagion through credit exposure and by contagion through impact on markets. Then, Sections 5.4 and 5.5 examine lessons from the crisis for, respectively, risk managers and government regulators. Section 5.6 takes a brief look at lessons from the crisis that go beyond the scope of risk managers and government regulators. Just to attempt to clear up one confusing bit of nomenclature at the beginning—CDOs on subprime mortgages were termed asset‐backed securities (ABSs), so what were called ABS CDOs were actually CDO‐ squared products (see Section 13.4.2 for explanation of a CDO‐squared). In fact, as documented in Cordell, Huang, and Williams (2012), a very substantial portion of subprime mortgage securities were CDO‐squared products. But since the economic and analytic characteristics of CDO‐squared products do not differ materially from primary CDO products, as discussed in Section 13.4.2, I will ignore this distinction in the remainder of this chapter. CDOs were the genesis of this crisis, and they were also at the root of what made it so damaging to the world economy. Large losses at banks due to lending in boom times that later goes sour under more challenging economic circumstances are a part of a fairly predictable cycle. Despite these periodic large losses, lending tends to be a profitable business over time, a conclusion that the studies on the excess of credit spreads over long‐term 83 84 FINANCIAL RISK MANAGEMENT loss rates would tend to support (see Hull 2012, Section 23.5, and Amato and Remolona 2003). Nor were losses on mortgage lending over this crisis period confined to holders of CDOs; large banks like Countrywide and Washington Mutual and government‐sponsored agencies like Fannie Mae and Freddie Mac managed to be big losers without much participation in CDOs. See the Financial Crisis Inquiry Report (2011, 106–109, 248–250, 305–307) on Countrywide and Washington Mutual; see Acharya et al. (2011) on the government‐sponsored agencies. What was different about credit losses that resulted from CDOs rather than from loans? The illusion that CDOs were bringing more liquidity to the mortgage lending market resulted in an exacerbation of what might otherwise have been a far more manageable downturn. As we’ll see in Section 5.2, treating the CDOs as if they were liquid securities rather than illiquid loans helped to fuel an expansion in lending far beyond what probably would have occurred without it. Then, when it became clear that the alleged liquidity wasn’t really there, the commitment of the investment banks to accounting for CDOs as if they were liquid assets turned what would have been longer‐term losses to be dealt with over the length of a credit cycle into immediate requirements for raising new capital. This quickly led to contagion in which markets, securities, and firms not originally involved in the CDO market got heavily impacted as well. We’ll follow this aspect of the story in Section 5.3. The focus in this chapter is on those aspects of the crisis that are most directly relevant to risk management. For those looking for a broader view of the crisis, the Financial Crisis Inquiry Report (2011), referenced as FCIR (2011), and the guide to the literature in Lo (2012) are good starting points. In my narrative and analysis and my lessons for risk managers section, I acknowledge that I cannot draw upon the firsthand detailed familiarity that would come from working in risk management at one of the affected institutions during the crisis period—I retired from JPMorgan Chase in 2004 and was working during the crisis period primarily as an educator. I have based my account on a combination of what is in the public record, what I have gleaned from conversations with people who were on the inside during the crisis, and what I have seen as an independent consultant to some of the impacted firms in the aftermath of the crisis. Balancing this, my lack of participation in the crisis leaves me relatively free of any axes to grind, positions to defend, or constraints due to confidentiality (though I can’t identify, either explicitly or by implication, specific clients I worked for after the crisis). In my analysis of lessons for regulators, I summarize the major proposals that have been offered, but restrict my own suggestions to those where my experience and judgment as a risk manager offer some direct benefit. The Systemic Disaster of 2007–2008 85 This means that I need to leave to others analysis of critical issues where my knowledge is less germane. To take one typical example, a good deal of policy discussion following the crisis has to do with issues surrounding how narrow you want to make the role of commercial banks—proposals like the Volcker rule banning proprietary trading by any firm with implicit government guarantees or suggestions about reimposition of Glass‐Steagall‐like restrictions on the mixing of commercial banking and investment banking. These proposals involve trade‐offs between reducing the probability of future disasters versus the possible negative impacts on a country’s economic growth by reducing financial innovation or by hurting the lending capacity of commercial banks by limiting their sources of revenue. As a risk manager, I have been trained in analyzing risks that arise within a given institutional structure and not in evaluating the economic impact of different institutional structures. On that which I cannot speak with insight, I will remain silent. 5.2 THE CRISIS IN CDOs OF SUBPRIME MORTGAGES It is not surprising that a disaster of the magnitude of the 2007–2008 CDO crisis had many causes and has led to a sizable literature of exposition. While I draw on many books and articles in what follows, I would like to particularly draw the reader’s attention to four relatively short articles that I find especially incisive in their analysis: Davidson (2007), Ashcroft and Schuermann (2008), Hull (2009), and Brunnermeier (2009). For those interested in further reading about the causes of the crisis and implications for the future, Lo (2012) is a concise guide to the best of the academic and journalistic literature on the topic, while Oyama (2010, Chapter 3) provides an excellent summary of reports that have been issued by various regulatory agencies and industry groups. Also recommended is the FCIR (2011) compiled by the Financial Crisis Inquiry Commission that was authorized by the U.S. Congress. In trying to look at all the causes, I have divided up the narrative into separate sections on the institutions with different roles in the CDO process. Section 5.2.1 covers the originators of subprime mortgages, Section 5.2.2 the issuers of the CDOs backed by these mortgages, Section 5.2.3 the rating agencies whose input was critical in the decision making of investors buying the mortgages, and Section 5.2.4 investors who suffered the actual losses when the CDOs lost value. Section 5.2.5 looks at those investment banks that had substantial direct exposure to the CDO losses, a subset of the CDO issuers studied in Section 5.2.2. In many ways the investment banks that, as we shall see, had by far the most catastrophic losses in the crisis are the most puzzling of the institutional 86 FINANCIAL RISK MANAGEMENT groupings. First, the “originate to distribute” paradigm of investment banking would call for only temporary use of a firm’s balance sheet, yet the CDO positions on which they were exposed were long‐standing and seemingly permanent. Second, the sophistication that should have resulted from origination of the CDOs should have made the investment banks far less vulnerable to being misled as to their value and riskiness than ordinary investors would be. And third, their well‐established risk management processes should have served as a check on such large and reckless exposures. We will spend some time in understanding how all these barriers were breeched. Finally, in Section 5.2.6, we will consider the AAA‐rated insurance companies, American International Group (AIG) and the monoline insurers, who became entangled in the crisis and who wrecked valuable franchises in pursuit of business completely tangential to their core competencies. 5.2.1 Subprime Mortgage Originators One point on which everyone examining the crisis can agree is that a significant contributor was the lax standards and misaligned incentives of the originators of subprime mortgages. Since the originators knew the mortgages were going to be bundled for purchase by an investor, the originators had no direct financial stake in the ultimate value of the mortgages. But the originators had a strong incentive to originate as many loans as possible, given that they were being paid a fee for originations and given the heavy demand by CDO creators for new product. To take just a few excerpts from the postmortems: ■ ■ Brunnermeier (2009): Mortgage brokers offered teaser rates, no‐ documentation mortgages, piggyback mortgages (a combination of two mortgages that eliminates the need for a down payment), and “no income, no job or assets” (NINJA) loans. Hull (2009): “Mortgage brokers started to increase their lending standards in about 2000. . . . How could mortgage brokers and mortgage lenders keep increasing their profits? Their problem was that as house prices rose it was more difficult for first‐time buyers to afford a house. In order to attract new entrants into the housing market, they had to find ways to relax their lending standards even more—and that is exactly what they did. The amount lent as a percentage of the house price increased. Adjustable rate mortgages (ARMs) were developed where there was a low ‘teaser’ rate of interest that would last for two or three years and be followed by a rate that was much higher. . . . Lenders also became much more cavalier in the way they reviewed mortgage applications. Indeed, the applicant’s income and other information reported The Systemic Disaster of 2007–2008 ■ 87 on the mortgage form were frequently not checked.” Even loan‐to‐value ratios and FICO scores (the credit score of the home buyer) reported to investors became suspect as “the property assessors who determined the value of a house at the time of mortgage application sometimes succumbed to pressure from lenders to come up with high values” and “potential borrowers were sometimes counseled to take certain actions that would improve their FICO scores.” Michael Youngblood, head of asset‐backed securities research at Friedman, Billings, Ramsey, is quoted by Peter Coy in the March 2, 2007, issue of BusinessWeek as stating that there was “a sudden but little‐noticed shift in lenders’ strategy that occurred at the end of 2005: Lenders went from competing for customers on price (by lowering rates) to competing for customers on easy terms (by lowering lending standards).” The incentives and the results seem clear. What is less clear is why other parties didn’t perceive this incentive structure and begin to exercise caution as evidence of lax standards started to mount. To cite just one example of concerns expressed at the time, a July 15, 2005, New York Times article by Edmund Andrews states that the areas that bank regulators find most worrisome “include granting loans equal to 100 percent of the value of homes; granting large loans without due attention to the likelihood of higher monthly payments in the future; and granting ‘no‐doc’ (no documentation) or ‘low‐doc’ loans that require little or no proof of income or assets.” This article quotes Barbara Grunkemeyer, deputy controller for credit risk at the Office of the Comptroller of the Currency: “You have to ask yourself, why would [a borrower] be willing to pay a quarter‐percent more when he could have gotten a lower rate by giving a copy of his pay stub and a W‐2 form. There’s a reason they’ve been called ‘liar’s loans.’” According to Davidson (2007), “mortgage market participants have long recognized that there is substantial risk in acquiring loans originated by someone else” and so require representations and warrants from the originator. If loans sold are later found not to meet the guidelines of the purchaser, the originator must repurchase the loans. But as the push for more new product to feed CDO issuance intensified, more marginal originators became part of the pipeline. The thin capitalization of these newer originators decreased the value of any promises to repurchase mortgages that were not as represented. But the CDO creators, the rating agencies, and the more sophisticated investors should all have been aware of this thinner capital cushion. Why didn’t this lead to more caution? We look at some specific reasons in the sections that follow. One general possibility, suggested by Brunnermeier (2009), is that the assumption that “background checks are unnecessary because house prices could only rise, and a borrower could thus 88 FINANCIAL RISK MANAGEMENT always refinance a loan using the increased value of the house” may have caused both originators and potential watchdogs to relax their vigilance. 5.2.2 CDO Creators Much of the blame for the problems with subprime mortgage CDOs must be allocated to the investment banks that created the CDOs. They certainly possessed the greatest amount of expertise, with highly compensated and very experienced structurers, marketers, traders, researchers, and risk managers specializing in mortgage markets and securitization. If any party was well positioned to be aware of the shortcuts that were being taken by the mortgage originators and to spot the potential dangers, the CDO creators were it. Certainly, the CDO creators cannot claim that they were misled by the rating agencies. As we will see in the next section, the investment banks had full access to the models the rating agencies used in determining ratings. In the process of playing with those models to determine how to optimally structure new issues, the CDO creators probably gained more intimate knowledge of these models than the people within the rating agencies who built them. And the investment banks could bring far more resources than the rating agencies into play, in terms of ability to pay high compensation to attract the best modeling talent (see Tett 2009, 100). The easy answer is just to focus on incentives. Since the CDO creators were operating on an originate to distribute business model in which all the CDO risk would eventually end up elsewhere, their incentives, like those of the mortgage originators, were to create as much product as possible, since fees earned were tied to volume sold, and to do their best to minimize anyone’s perception of possible loss. One could argue that this is failing to give the investment banks sufficient credit for concern for their longer‐term reputations with investors and future losses through possible lawsuits, but after their collectively dismal record in hyping technology initial public offerings (IPOs) in the late 1990s, it would be hard to take that argument very seriously. But incentives can’t be the whole story, for two reasons. One is that the investors should have been aware of these incentives and the track record the investment banks had shown when faced with these temptations in the past, and so should have exercised their own due diligence. The second is that many of these investment banks failed to execute their desired originate to distribute strategy so egregiously that they wound up being the largest losers when CDO values started to decline. We’ll look at how this occurred in Section 5.2.5; first, let’s see why the investors were willing to trust the CDO creators to the degree they did. The Systemic Disaster of 2007–2008 89 Part of the reason for this trust was undoubtedly comfort that came from the supposedly independent review role of the rating agencies. Why that trust was misplaced we’ll examine in the next section. Part came from the skillful marketing of the investment banks, which did their best to convince investors that gains and losses on CDOs would all be about esoteric issues like correlation assumptions, on which the investment banks would be happy to give tutorials to investors, ignoring issues like quality of underlying loans on which the CDO creators possessed insider knowledge that investors could not hope to obtain. And part of the reason for this trust was a structural feature that was supposed to align the interests of the CDO creator with the interests of investors: the retention of the first‐loss piece of the CDO by the creator, the so‐called equity tranche. The retention of this first‐loss piece meant that this part of the CDO would absorb all of the losses up to some given point and that investors could suffer losses only if the equity tranche was wiped out. The theory was that the investment banks had to closely monitor the quality of assets going into the CDO to avoid large losses on this first‐loss piece. The problem was that profits from the tranches that were sold to investors became so lucrative that the CDO creators stopped caring about how much they lost on the equity tranche. According to Hull (2009), “the equity tranche was often regarded as a ‘free good.’ The originators had obtained adequate compensation for the mortgages from the sales of the other tranches to investors.” So much profit had been generated that they could afford to take a full loss on the equity tranche and still come out ahead, or they could afford to purchase protection on the equity tranche from a hedge fund. 5.2.3 Rating Agencies The rating agencies—Standard & Poor’s (S&P), Moody’s Investors Service, and Fitch—all badly damaged their reputations by the role they played in providing ratings on CDOs backed by subprime mortgages. They have been the subjects of major investigations, and their role in CDO ratings has led to questions being raised about the role they play in all ratings, including their long‐established core business of rating corporate debt. In some ways, their story resembles that of the insurers we will look at later in Section 5.2.6 who jeopardized their core franchise in pursuit of new business. And yet, the rating agencies had a more plausible case than the insurers that this new business line was related to existing competency. Unlike the insurers, who entered a market that could have survived without them, the rating agencies had a role that was critical to the existence of the CDO business. Most debt investors, from long habit, would have been extremely uncomfortable investing without an agency rating; many were legally 90 FINANCIAL RISK MANAGEMENT prohibited from investing in debt that did not have a particular minimum rating—it was considered too risky. Ratings tied to probability of repayment were the rating agencies’ bread and butter. And they did have several decades’ worth of successful experience in rating structured debt that related to mortgages, credit cards, auto loans, and CDOs based on corporate debt. But in 2007 and 2008, the ratings on existing CDOs were downgraded far more violently than any other class of rated securities ever had been, sowing widespread distrust in the agencies. Where did this business model break down? Many critics in the wake of the CDO crisis point to conflict of interest as the main flaw in the rating agency structure: the rating agencies were being paid by the firms whose bonds they were rating. But that flaw has always existed for all agency ratings, including the core business of rating corporate debt. What is probably more germane is the very close relationship developed between the rating agencies and the investment bank structurers creating the bonds. Structurers had full access to the agency ratings models and a great deal of freedom in deciding what mortgages would go into a CDO. They could play with the structure until they optimized the disconnect between the risk represented by the rating and the true risk, maximizing their profits (see Brunnermeier 2009, 82). There is no comparable freedom to easily change corporate structure. Furthermore, a corporation that does not get the rating it wants will still continue in business and so may choose to pay for the rating anyway. A CDO not getting the rating it wants will not come to market, so the only way rating agencies could get paid is if CDOs did come to market; for further discussion of this point, see Davidson (2007, 4). There is considerable evidence that has come to light since the crisis that rating agencies did succumb to the pressure to find ways to give CDO structures the ratings they needed (see, for example, McLean and Nocera 2010, Chapter 8, and Lowenstein 2010, Chapter 4). Another significant flaw in the analogy between traditional agency ratings of corporate debt and agency ratings of CDOs was that the ratings methodology for CDOs required the agencies to make forecasts about the state of the economy whereas corporate debt ratings did not. This point is made well by Ashcroft and Schuermann (2008, Section 5.5): CDO ratings “rely heavily on a forecast of economic conditions. Note that a corporate credit rating is based on the agency’s assessment that a firm will default during neutral economic conditions (i.e., full employment at the national and industry level).” (This corresponds to the point made in Section 13.2.1.1, about agency ratings being through‐the‐cycle and not point‐in‐the‐cycle.) In CDO modeling, by contrast, “uncertainty about the level of loss in the mortgage pool is driven completely by changes in economic conditions” (such as the expected default rates of mortgages, which are closely tied to forecasts of real estate prices). Furthermore, CDO ratings “depend heavily The Systemic Disaster of 2007–2008 91 on quantitative models while corporate debt ratings rely heavily on analyst judgment.” This meant that rating agency senior management, experienced in corporate debt ratings, had little intuition for what was going on in the CDO ratings. And neither rating agency management nor investors had been warned about the precipitous decline in ratings a change in economic outlook could entail, in contrast to the far more steady corporate debt ratings. (CDO ratings are more volatile than corporate debt ratings both because they depend on economic forecasts and because the CDO tranching process concentrates sensitivity to the economy in the higher‐rated tranches—see Section 13.4.4.) While these are probably the two most important factors in the rating agency failure, other issues of some weight were: ■ ■ The failure of rating agencies to monitor the deteriorating credit standards of the subprime mortgage originators. There was certainly enough publicity about this issue that rating agencies should have been aware of a need to perform some due diligence. Former Moody’s managing director Jerome Fons in testimony to the Financial Crisis Inquiry stated that “never once was it raised to this group [Moody’s high‐level Structured Credit committee] or put on our agenda that the decline in quality that was going into pools, the impact possibly on ratings . . .” (FCIR 2011, 121). Even if the rating agencies didn’t believe it was their responsibility to check on the mortgage originators, they should at least have been questioning the relevance of historical default data to a rapidly changing situation. It wasn’t just issues being raised about slipping credit standards that should have triggered such questioning, but the sheer explosive growth of the market, which should have been enough to make the relevance of data from prior eras doubtful (compare with Section 8.2.8.2). The rating agencies’ response to these criticisms was to claim the transparency of their CDO ratings models as a virtue (see, for example, Tett 2009, 100). Anyone could see exactly what the model was doing, the agencies implied, so why blame us if you were later disappointed in the results? This was disingenuous in two directions. First, as emphasized by Tett, it was the very openness and transparency of the models that made them so easy for sophisticated structurers to manipulate. And second, the vast majority of investors certainly lacked the sophistication to understand the workings of the models and had far less capability than the rating agencies for checking loan quality and relevance of historical data. Why hadn’t these issues surfaced in the reasonably long history of agency ratings of other structured securities? I haven’t seen an analysis of this, but I suspect that while some of these issues were present for other 92 FINANCIAL RISK MANAGEMENT structured securities, they did not have as strong an impact as they did on the subprime mortgage CDOs. For example, subprime mortgage CDOs are particularly dependent on the state of the national real estate market. 5.2.4 Investors In many ways, the investors in CDOs can be regarded as the key players in the whole structure. It was the large appetite of investors to own CDO tranches that drove the growth of the market and set the incentives for all the other players. There was a large and diverse universe of these investors, including mutual funds, pension funds, insurance companies, hedge funds, high net worth individuals, and smaller banks (those not involved in the creation of CDOs). It was the CDO investors who were the claimed victims of fraud and misrepresentation by the other players. And it was the CDO investors who, in theory, should have been the ones to suffer the bulk of losses that occurred in the market meltdown. But somehow, it did not work out that way. The major institutions that suffered the greatest reverses and either went bankrupt or required government bailouts were not primarily the investors but rather the investment banks that created the CDOs. The best overall summary that is available of losses due to the crisis is the International Monetary Fund analysis of April 2009 (International Monetary Fund 2009, Table 1.3) that concluded that out of roughly $1 trillion in losses on U.S.‐originated mortgage CDOs, 60 percent was lost by banks, 25 percent by U.S. government‐sponsored enterprises (GSEs), 10 percent by insurers, and only 5 percent by hedge funds, pension funds, and other nonbank financial institutions. Still, the investors did suffer substantial losses, as can be seen by just looking at the damage claims in lawsuits filed against mortgage originators and CDO creators. The FCIR (2011, 225) asserts that “as of mid‐2010, court actions embroiled almost all major loan originators and underwriters— there were more than 400 lawsuits related to breaches of representations and warranties, by one estimate”; for an updated account of the many lawsuits that have been filed, see the Structured Finance Litigation blog: www .structuredfinancelitigation.com). The theory of these lawsuits and of many articles that have been written on the crisis is that deliberately misleading action was taken to entice investors to buy these securities. The previous three sections contain much evidence to support such claims, so there is at least a significant extent to which investors were misled. The question I want to ask here is: To what degree was that the entire story and to what extent were investors knowingly taking on significant risk? This question is one that has much relevance for risk managers in trying to learn lessons from the crisis. If there were clear signs of riskiness The Systemic Disaster of 2007–2008 93 that investors failed to understand or chose not to focus on, then we have material that can be used in designing better risk management procedures for the future. The principal arguments that investors were to at least some significant degree aware of the risk they were taking on are first that CDO tranches were yielding considerably higher returns than corporate bonds with comparable credit ratings and second that the very illiquidity of the tranches should have been a warning sign against placing too much faith in what they were being told. The historical data I have been able to access shows a steady yield advantage of about 80 basis points for AAA‐rated subprime CDO tranches over AAA‐rated corporate bonds throughout the period from 2000 to 2006. 5.2.5 Investment Banks As already noted in Section 5.2.4, investment banks that were among the major creators of CDOs were also the group that suffered the heaviest losses in the 2007–2008 meltdown. This can be seen from the previously cited International Monetary Fund (IMF) analysis that found that 60 percent of the $1 trillion in losses on U.S. originated mortgage CDOs came from banks while only 5 percent came from the mutual funds, pension funds, hedge funds, and other nonbank financial institutions that were the primary clients to which the investment banks marketed the CDO tranches. It is true that 10 percent of the losses came from insurance companies and many insurance companies were among the primary clients to whom CDO tranches were marketed. But a good portion of the insurance company loss is attributable to AIG and the monoline insurers, and, as we detail in Section 5.2.6, AIG and the monoline insurers can more reasonably be viewed as partners of the investment banks in CDO creation than they can be viewed as clients. It is also true that the IMF analysis does not distinguish how much of the $600 billion in losses came from investment banks that were CDO creators and how much was due to smaller banks that may have been clients. But an analysis by the Federal Reserve Bank of Philadelphia (Cordell et al. 2012, Table 11) shows losses of 72% on the $223 billion of mortgage‐ backed CDOs originated in 2006 and 84% on the $163 billion of mortgage‐backed CDOs originated in 2007. With loss levels this high, a substantial portion of the losses had to be going to the super‐senior tranches that were primarily held by the investment banks that were CDO creators, and Table 12(b) from the same report shows 67% losses on senior AAA tranches originated in 2006 and 76% losses on senior AAA tranches originated in 2007. At the level of an individual investment bank, UBS, which made a public and thorough report to shareholders in April 2008 of the fallout of the crisis, reported 2007 losses related to the U.S. residential mortgage 94 FINANCIAL RISK MANAGEMENT market of $18.7 billion, with about $12 billion due to CDO positions. By early 2009, estimates of total write‐downs and credit losses on U.S. financial assets were $48.6 billion for UBS, $67.2 billion for Citigroup, and $55.9 billion for Merrill Lynch (see Zandi 2009, Table 11.2). This is both unfortunate and surprising: unfortunate, because concentrated losses by large banks are far more damaging to the economy than the same amount of losses spread out over smaller banks and investors; surprising, because the sophistication, intimate familiarity with the product, and originate to distribute business model should all have worked to protect the investment banks. How then did investment banks wind up with so much mortgage CDO exposure? The initial mechanics of the situation are fairly straightforward. Clients were eager to purchase CDO tranches, thereby selling protection against mortgage defaults, but they were interested only in the mezzanine tranches that carried intermediate expected loss. The highest expected loss tranches, the so‐called equity tranches, attracting the first losses, could not have achieved investment‐grade ratings and were not considered suitable investment vehicles for most clients (though some hedge funds did take on this risk, mostly through derivatives). Also, it was considered appropriate that the CDO creator hold the equity tranche, as explained in Section 5.2.2. The tranches with the lowest expected loss, termed super‐senior because they supposedly had a statistical probability of loss even lower than AAA‐ rated corporate bonds, did not have a strong client demand. Because of their very low loss expectation, they carried very low credit spreads, just a few basis points, and it was virtually impossible to find a client that wanted to use valuable balance sheet room to earn such a meager return. (It might be thought that super‐senior tranches would be a possibly attractive investment as an alternative to Treasury securities that had similarly low returns, but Treasury securities had many advantages in terms of liquidity and attractive repurchase rate funding opportunities that super‐seniors lacked.) Here was a dilemma for the investment banks. To create more mezzanine tranches for which there was high demand, they also needed to create super‐senior tranches for which there was virtually no demand. Of course, one alternative would have been to substantially raise the yield on the super‐seniors to the point that demand was created, but this would have so severely cut into the profitability of the overall transaction that it wasn’t seriously considered. Their only alternatives were to stop the flow of lucrative new business or to pile up super‐senior tranches on their own balance sheets. They almost all chose the latter option. As Chuck Prince, the soon‐ to‐be‐ex‐CEO of Citigroup, infamously said in July 2007, “As long as the music is playing, you’ve got to get up and dance. We’re still dancing” (FCIR 2011, 175). It was this continuous buildup of super‐seniors, totally lacking The Systemic Disaster of 2007–2008 95 a liquid market, that was the source of almost all of the large CDO losses suffered by the investment banks. For example, the UBS report to shareholders showed that about $9 billion of its $12 billion 2007 losses on CDOs were due to super‐senior tranches. Other large investment banks that followed this pattern included Citigroup, Merrill Lynch, Morgan Stanley, and Bear Stearns (see Tett 2009, Chapter 9). Writing generally about investment banks that experienced large losses in 2007, the Senior Supervisors Group report of March 2008 on page 8 states that “some firms continued to underwrite or increase their exposures until the summer of 2007 despite an array of data indicating rising stress in the subprime mortgage market and worsening credit market conditions.” If management of these banks had placed sensible limits on the size of super‐senior holdings or had insisted on mark‐to‐market valuations of the holdings that reflected their total lack of liquidity (thereby lowering the profit that could be recognized on new CDO issuance and shrinking bonus pools), the entire mortgage CDO creation process would have come to a halt at a fairly early stage and the damage to the financial industry and the world economy would not have been nearly as severe. As Richard Bookstaber, an experienced senior risk manager, put it in his testimony before the Financial Crisis Inquiry Commission, “As everybody in any business knows, if inventory is growing, that means you’re not pricing it correctly. . . . It was a hidden subsidy to the CDO business by mispricing” (FCIR 2011, 196). What stopped reasonable action from being taken? The banks seemed to be operating as if they possessed a split personality. In one part of the firm, the CDO creation teams were behaving as if all risk was being taken on by clients, as if the originate to distribute mechanism was operating smoothly. This left them free to ignore warning signs about the increasingly poor quality of the mortgages being originated and about the potential impact on losses if the housing price bubble burst. In another part of the firm, super‐ senior tranche holdings were growing by leaps and bounds. One possible answer is that the traders and structurers who had the greatest degree of knowledge of the situation just didn’t care about the health of the firm and so did their best to mislead senior managers and risk managers about what was really going on. All they cared about was generating one more round of spectacular bonuses. They treated risk managers and senior management as just another set of clients to whom product needed to be sold—in this case, super‐senior tranches. While this no doubt contains an element of truth, it can’t be the entire story. Risk management of investment banks has always been built upon a healthy skepticism about the motivations of front‐office personnel, as we saw in Section 2.1 and as we will consider at greater length in Chapter 6. So let’s try looking at some other possible explanations. We’ll look at supporting evidence for them in 96 FINANCIAL RISK MANAGEMENT this section, and then draw on this material to examine risk management lessons in Section 5.4.5, using the same headings in both sections. Before going any further, let me clear up one possible source of confusion. In the midst of the crisis, there were many news reports concerning disputes over the marking‐to‐market of distressed securities—should firms holding securities experiencing what was hoped to be a temporary bout of illiquidity show losses based on the fire‐sale prices at which these securities were trading in the market? Since these disputes occurred during the same period that large losses were being recorded by the investment banks on their super‐senior tranches, it might have seemed that the super‐senior losses were at least partially an accounting fiction. But as we’ve just recounted, the super‐senior tranches never had a liquid market at any time, so their marks were always based just on the best judgment as to ultimate losses. Whatever the merits of the accounting debate over other securities that were caught up in the crisis (we’ll have more to say about this in Section 5.3.2), the losses reported on super‐seniors always represented best estimates of true ultimate cost. 5.2.5.1 Reliance on Inadequate Derivatives Protection One fairly common response to the inability to find clients to buy super‐senior tranches was to hold on to the super‐senior tranches but hedge the risk with derivatives. This should clearly have been viewed with suspicion by risk managers—if you couldn’t find clients willing to buy super‐seniors, why were you able to find clients willing to take on the risk through derivatives? Wasn’t there some substantial difference in the amount of risk being shed in the two different transactions? By saying that suspicion should have been aroused, I do not mean that it was obvious that the risk was not being fully hedged, just that thorough analysis should have been initiated. I have seen cases in which firms were willing to fully absorb risk but had limitations on balance sheet usage, perhaps because of lack of access to good funding sources or perhaps due to statutory restrictions. Analysis in these cases showed that the sellers of derivative protection were providing sufficient collateral and margining to keep risk very close to what would have been achieved with an outright sale, though with different funding requirements. (Discussion of collateral and margining will be found in Section 14.3.3.) Had thorough analysis been performed in the case of the derivatives hedging super‐seniors, a very different picture would have emerged. Many collateralization and margining agreements were either nonexistent or of very limited value. For example, Lowenstein (2010, Chapter 9) reports that Vikrim Pandit, on taking over as CEO of Citigroup, was “stunned to hear” from New York State’s top insurance regulator that “Citigroup’s insurance The Systemic Disaster of 2007–2008 97 did not entitle it to payments as the prices of CDOs declined.” Citi had “insurance on defaults, not on market value.” Given the long‐dated nature of the CDOs, “Citi (and every other bank with insurance) would have to wait years to file claims, at which point the insurers could be out of business.” This was very typical of insurance purchased (whether through insurance contracts or through derivatives) from the major suppliers of super‐senior insurance, AIG and the monoline insurers (whose role we will look at more closely in the next section). AIG did offer some collateralization, partly to gain a competitive advantage on the monoline insurers, which offered none (McLean and Nocera 2010, 190–191). But some of this was weak collateralization that would be triggered only under extreme circumstances, by which time AIG might already be facing difficulties (as proved to be the case). When we look at risk management lessons in Section 5.4, we’ll do a detailed analysis of all the alarm bells this arrangement should have sounded. But, as I will detail there, the risk management methodology for identifying the large gap in risk reduction between outright sale and insurance protection was well known and thoroughly disseminated well before these deals were booked. If this was not highlighted to senior management and regulators, it constituted a major breach of risk managers’ responsibilities. 5.2.5.2 Reliance on Off‐Balance‐Sheet Vehicles If you couldn’t find clients interested in holding super‐seniors because of their very thin spreads over funding costs, there was one more trick that could be used: If you set up an entity that could hold the super‐seniors and issue short‐dated AAA‐rated debt against them, the normal upward slope of the yield curve would provide enough cushion to generate some extra spread to entice investors in short‐dated AAA debt (Tett 2009, Chapter 6). The primary practioner of this bit of financial legerdemain was Citigroup, which began placing its super‐seniors into structured investment vehicles (SIVs). SIVs were officially independent enterprises whose commitments Citi had no legal responsibility for and so did not have to be consolidated onto Citi’s balance sheet (leading to their classification as off‐balance‐sheet vehicles). But SIVs were funded by commercial paper (CP), and commercial paper investors would invest only in AAA‐rated entities. Even if the rating agencies regarded the super‐senior tranches as AAA, the short‐dated funding and long‐dated assets of the SIVs raised the issue of what would happen to the CP holders if new CP investors could not be found when the old CP matured. To obtain an AAA rating for the SIV, Citi needed to offer liquidity puts that would allow the SIV to sell the super‐seniors back to Citi at par if the SIV ran into problems funding them (McLean and Nocera 2010, 240–241). Citi wrote about $25 billion of these liquidity puts. 98 FINANCIAL RISK MANAGEMENT The key risk management question would now be what probability of loss to assign to these liquidity puts. The attentive reader will not be surprised that Citi’s internal risk models estimated so remote a possibility of the liquidity puts being triggered that they only needed to hold 0.16% in capital against the put (FCIR 2011, 138). Hence only $40 million in capital would be required against Citi’s $25 billion in liquidity puts. And there seems no evidence that Citi continued to view the super‐seniors placed into the SIVs as still being part of its risk book. But clearly, placing the super‐seniors into an SIV made practically no difference to Citi’s risk position. In the event that there would be losses on the super‐seniors, it would be virtually certain that the liquidity put would be exercised. This is another clear case of violation of one of the well‐established rules of risk management, the need to account for wrong‐way risk (see Section 14.3.4 for more explanation of wrong‐way risk). 5.2.5.3 Use of Faulty CDO Models Felix Salmon’s February 2009 story for Wired magazine, “Recipe for Disaster: The Formula That Killed Wall Street” (Salmon 2009) brought David Li’s version of the Vasicek model to the attention of a wider audience than financial industry quants (see Section 13.3.3 for a description of the model). The article led off with statements such as: “One result of the [2008 financial system] collapse has been the end of financial economics as something to be celebrated rather than feared. And Li’s Gaussian copula formula will go down in history as instrumental in causing the unfathomable losses that brought the world financial system to its knees.” With this as background, I was somewhat surprised in my survey of the principal book‐length writings and journal articles on the crisis to see scant mention of either the Li model or the Vasicek model. Did faulty CDO modeling play a significant role in the crisis? The case for faulty CDO models playing, at best, a minor role in the crisis would go as follows: ■ ■ The Li model was primarily being used as an interpolation tool from more common tranches for which price quotes could be obtained to less common tranches. As such, its use was very similar to that of the Black‐Scholes model in interpolation of options prices and the use of fitting to a correlation skew implied by the market (see Section 13.4.2) as part of the interpolation shows that the Gaussian copula assumptions of the Li model were not being taken very seriously by the traders using it. The Li model was also being used as an aid to intuition (see Section 13.3.3) and as such it did its job admirably. In fact, it was particularly valuable in letting users see the degree of systematic risk embedded The Systemic Disaster of 2007–2008 ■ ■ 99 in different tranches, which should have directed attention to the riskiness of super‐senior tranches (see Section 13.4.4). The emphasis on the correct estimation of correlation levels and the shape of the correlation copula was very important for traders making decisions on the value of tranches. Had the tranches been liquid, this would also have been important for risk managers, in estimating where liquid positions could be exited. But given the illiquidity of super‐senior tranches, stress testing large changes in the common factor, closely linked to real estate prices, was overwhelmingly more important for risk managers than stress testing of either correlation level or copula shape. When investment banks wanted to perform more fundamental analysis of tranche pricing and risk, they were hardly lacking for more sophisticated versions of CDO models, as the discussion in Sections 13.3.3 and 13.4.2 clearly show. Many of the models cited in these sections date from the first half of the 2000s decade and were widely available—often referenced and explained in papers published by investment bank research teams, in the well‐known book by Schonbucher (2003), and in many issues of Risk magazine from that period. And yet there is one key way in which CDO models utilized by investment banks in this period were misleading. Too much emphasis was placed on fitting model parameters to observed market prices without an adequate consideration of the degree of illiquidity that pervaded many sectors of this market, including the entire super‐senior sector. This may have helped encourage the definitively faulty analysis we discuss in Sections 5.2.5.6 and 5.2.5.8. 5.2.5.4 Reliance on External Ratings It is uncontroversial that the rating agencies played a significant role in fueling the demand for CDO tranches by investors. But could they have also played a role in the willingness of investment banks to tolerate so large an exposure to super‐seniors? At first glance, this seems preposterous. As we noted in Section 5.2.2, the investment banks in their role as CDO creators had intimate knowledge of the rating agency models and knew the extent to which they had manipulated those models. How could they then rely on those models to take comfort with their exposure? And yet one finds in the March 2008 UBS report to shareholders (UBS 2008, Section 5.3.2) that the UBS market risk control group’s “VaR methodologies relied on the AAA ratings of the Super Senior positions. The AAA ratings determined the relevant product‐type time series to be used in calculating VaR. . . . As a consequence, even unhedged Super Senior positions contributed little to VaR utilization.” Tett (2009, 139) quotes Peter Kurer, a member of UBS’s board, as saying, “Frankly most of us had not even heard 100 FINANCIAL RISK MANAGEMENT the word ‘super‐senior’ until the summer of 2007. We were just told by our risk people that these instruments are Triple‐A, like Treasury bonds.” Anecdotal accounts I have heard from other investment bank risk managers indicate that UBS was not alone in utilizing AAA ratings of tranches as an invitation to calculate risk statistics for them based on time series of price moves of AAA‐rated corporate bonds. The March 2008 Senior Supervisors Group report on the risk management practices of investment banks leading up to the crisis states that at some firms “internal risk capital measures that relied too much on agency ratings underestimated the true price of the risk of such positions” and that some firms “tended to assume that they could apply the low historical return volatility of corporate credits rated Aaa to super‐senior tranches of CDOs” (p. 5). It further states, “Given that the firms surveyed for this review are major participants in credit markets, some firms’ dependence on external assessments such as ratings agencies’ views of the risk inherent in these securities contrasts with more sophisticated internal processes they already maintained to assess credit risk in other business lines” (p. 3). The impression left is consistent with the picture of front‐office personnel not sharing their knowledge of rating agency model limitations with risk managers. We address the lessons for risk managers in Section 5.4.1. 5.2.5.5 Overreliance on VaR Measures As we have just seen, UBS (and, anecdotally, some other investment banks) used the AAA ratings of super‐seniors as a shortcut in VaR calculations, essentially treating any AAA‐rated security as if its price movements could be represented by a time series drawn from AAA corporate bond prices. This was clearly an error—as discussed in Section 13.4.4, the volatility of tranche prices is expected to be quite different from the volatility of corporate bonds of the same rating. But an even more important question is: Why were firms even bothering to calculate VaR, a measure of vulnerability to short‐term price fluctuations, for an instrument as illiquid as super‐seniors? Now, perhaps this was just a calculation of VaR for a liquid proxy hedge of the super‐seniors, and the bulk of the risk was going to be evaluated elsewhere (a measure I will strongly advocate, in Sections 6.1.2 and 8.4, for highly illiquid instruments). If this was the case, then even the use of the computational shortcut might be justified—you would be choosing a portfolio of AAA corporate bonds as your liquid proxy hedge. It may not be the best choice, but as long as you are calculating the long‐term risk of the hedge separately no great harm will be done. But this does not appear to be the way UBS (or, anecdotally, some other investment banks) were operating. VaR was intended to be the primary risk measure for the super‐seniors. Quoting UBS (2008, Section 6.3.2), “Investment bank business planning The Systemic Disaster of 2007–2008 101 relied upon VaR, which appears to be the key risk parameter in the planning process. When the market dislocation unfolded, it became apparent that the risk measure methodology had not appropriately captured the risk inherent in the businesses having Subprime exposure.” Dash and Creswell (2008) relate that “when examiners from the Securities and Exchange Commission began scrutinizing Citigroup’s subprime mortgage holdings after Bear Stearns’s problems surfaced, the bank told them that the probability of these mortgages defaulting was so tiny that they excluded them from their risk analysis.” This brings us to the broader question of the extent to which the illiquidity of the super‐seniors was being factored into risk measurement. 5.2.5.6 Failing to Account for the Illiquidity of Super‐Senior Tranches The illiquidity of super‐senior tranches should have been evident to anyone involved in investment banking, even those most remote from direct trading and marketing of CDOs, just by the fact that it was such a problem to find willing buyers. But the Senior Supervisors Group report of March 2008 finds that “firms that faced more significant challenges in late 2007 . . . continued to price the super‐senior tranches of CDOs at or close to par despite observable deterioration in the performance of the underlying . . . collateral and declining market liquidity” (p. 3). The UBS report to shareholders Section 6.3.6.4 states that “The Super Senior notes were always treated as trading book (i.e., the book for assets intended for resale in the short term), notwithstanding the fact that there does not appear to have been a liquid secondary market and that the business tended to retain the Super Senior tranche.” Why were firms treating such clearly illiquid instruments as liquid? One clear motivation is alluded to in the same section of the UBS report: “Treatment under the ‘banking book’ would have significantly changed the economics of the CDO desk business as this would have increased the required regulatory capital charges.” Classifying assets in the trading book, available for resale in the short term, attracted more favorable capital treatment than the same assets placed in the banking book, intended to be held. Note that this is just a statement of intention—nothing stops you from selling assets in the banking book; loan sales occur all the time. But this statement of intention was allowed to impact required regulatory capital, a major driver of the economics of a product. This loophole was closed after the CDO‐fueled crisis revealed its shortsightedness; the Bank for International Settlements (BIS) Incremental Risk Capital Guidelines of July 2009 made capital requirements for credit products held in the trading book and banking book essentially equivalent—see PricewaterhouseCoopers (2011, Section 4.6.3.5). It also impacted the balance sheet reporting that might impact public perception of the degree of liquidity of the firm’s assets. 102 FINANCIAL RISK MANAGEMENT My guess, and it’s only a guess, is that the mechanism that operated at some firms was that the potential liquidity of CDOs, including super‐seniors, had been emphasized in order to obtain the favorable capital treatment— securities are, in general, more liquid and likely to be sold than loans are. While this accounting decision should not have forced a similar classification by risk managers, it is not uncommon for this kind of distinction between accounting principles and risk management principles to get blurred. 5.2.5.7 Inadequate Stress Tests Another possibility is that there was widespread conviction that risks that threatened mezzanine tranches could not spread to super‐senior tranches. I find this difficult to accept, since the simplest possible CDO model could easily show the vulnerability of even super‐ senior tranches to a large downturn in housing prices, the sort of economic stress scenario that risk management groups are supposed to run routinely (see Section 13.4.4 on the usefulness of the Vasicek model in analyzing vulnerability of senior tranches to systematic risk). One viewpoint I have frequently encountered in conversations with risk managers who were caught up in the crisis goes something like this: “Place yourself back in 2006 and suppose you were to stress test your CDO portfolio. Suppose that you chose to shock U.S. house prices down 30 percent to evaluate the impact on the prices of super‐senior tranches that you held. You would have been laughed out of the room—no one would have found this a plausible stress test scenario.” A published account of a closely related incident can be found in Lewis (2011, 211–212). With all due deference to the fact that I was not actively involved in risk management of any of the impacted firms during this critical period, I must respectfully but strongly dissent from this view. First, looking at the history of super‐senior tranches at many of the impacted firms, you find an active interest in purchasing protection on these tranches from AIG (see Tett 2009, 134–136; FCIR 2011, 139–142, 202– 204). It is only when AIG’s appetite for selling protection dried up that firms turned to either absorbing the risk completely or utilizing clearly inadequate substitutes, such as buying uncollateralized protection from inadequately capitalized monoline insurers. If losses on super‐seniors weren’t going to occur under any plausible shock, why spend money and effort on buying protection? The rejoinder might be that this was “just to keep the risk managers (or the accountants or the regulators) happy.” But keeping risk managers or accountants or regulators happy means addressing a shock that they would find plausible; what made them stop finding it plausible at just the moment the protection could no longer be purchased? Second, it is not difficult to find mainstream economic analysis that viewed a large drop in housing prices as not just plausible but reasonably The Systemic Disaster of 2007–2008 103 probable. Just using the Economist magazine as a representative voice, one finds articles in the issues of December 9, 2004 (“Flimsy Foundations”); December 8, 2005 (“Hear That Hissing Sound?”); and September 7, 2006 (“Checking the Thermostat”), all talking about U.S. house prices being overvalued by amounts ranging from 20 percent to 50 percent and all talking about the serious possibility of the “bubble bursting.” This was not some then‐unknown junior economist crying in the wilderness; this was in a prominent mainstream publication that is required weekly reading for virtually everyone in the financial industry. And the opinions were backed by detailed statistical analysis of historical relationships of housing prices to rental prices and to incomes. At the same time, the Yale economist Robert Shiller, already prominent for the timely concerns he had expressed about the Internet bubble and noted for his expertise in the field of housing prices, was quoted by David Leonhardt in the New York Times on August 21, 2005, as “arguing that the housing craze is another bubble destined to end badly, just as every other real‐estate boom on record has. . . . He predicts that prices could fall 40 percent in inflation‐adjusted terms over the next generation.” Now certainly there is room for disagreement among economists and financial analysts. Someone making a strong and detailed argument for a given viewpoint is no reason it can’t be rejected as a most likely or even reasonably probable view. But to reject it as a plausible view I find disingenuous. My guess would be that it is far more likely that risk managers were buying into a wholly unsupportable view of the liquidity of the super‐ seniors, as documented in the preceding subsection. And if you are treating super‐seniors as liquid, then of course a drop of 40 percent in housing prices over the next generation is none of your concern since you only need to be worried about what might be reflected in the market over a period of a few weeks. 5.2.5.8 Inadequate Analysis of Statistical Hedging Faced with the inability to fully eliminate super‐senior exposure, some investment banks very sensibly began seeking more liquid hedges that would eliminate at least some of the exposure. The question is not whether this was a prudent strategy (it was), but whether risk managers adequately analyzed the resulting risk. One case in which they conspicuously did not do so is at UBS. Section 4.2.3 of UBS (2008) states that the Amplified Mortgage Portfolio (AMPS) consisted of super‐senior positions “where the risk of loss was initially hedged through the purchase of protection on a portion of the nominal position. . . . This level of hedging was based on statistical analyses of historical price movements that indicated that such protection was sufficient to protect UBS from any losses on the position.” In Section 6.2.3, the report states that once 104 FINANCIAL RISK MANAGEMENT hedged through AMPS trades, the super‐senior positions were considered fully hedged and therefore did not appear in either VaR or stress test reports. The report further notes, in Section 6.3.6.1, that even though an internal audit had “identified certain risks in the Subprime trading books, senior risk control did not appear to take those issues into account when concluding that positions were hedged.” To get a better understanding of statistical hedging, we need to add just a bit of complexity to the basic picture we have painted of trading in mortgage CDOs. In addition to the tranches that were based on dividing up actual pools of mortgages, some synthetic tranches based on reference portfolios began to trade (see Section 13.4.1 for details on synthetic tranches). To some extent, these synthetic tranches were just side bets between investors who wanted to sell protection on mezzanine tranches (the vast majority) and a few investors looking to buy protection, either as an offset to previous sales or because of a belief that mortgage defaults were going to exceed market expectations. The investment banks’ involvement with these side bets would have been that of market maker in a reasonably liquid market. But to some extent, these synthetic tranches offered an opportunity to investment banks looking to reduce their exposures to mortgage tranches. An entertaining and informative book focused on the market for synthetic tranches of subprime mortgages is Lewis (2011). Lewis provides a detailed narrative of the role these synthetic tranches played in generating large profits by hedge fund managers, such as John Paulson and Steve Eisman, as well as traders for investment banks, such as Deutsche Bank’s Greg Lippmann, on bets that mortgage defaults would exceed expectations. The same market fundamentals drove this market as drove the market for pool tranches, namely the strong investor demand for selling protection on mezzanine tranches and little interest in either equity or super‐senior tranches. So the synthetic tranches did not offer a direct offset to warehoused super‐senior exposure. But synthetic tranches did make it possible for investment banks to buy more protection on mezzanine tranches than they had created through the pool tranching process. So they could consider offsetting some of their super‐senior position in a particular portfolio by buying mezzanine protection on a reference portfolio either identical to or closely related to the portfolio the super‐seniors were exposed to. Here’s where the statistical analysis came in: What was the best dollar volume of a mezzanine tranche to buy protection on to hedge a given volume of super‐ senior tranche, and just how large was the risk offset? As you would expect from the large difference between hedging against changes in credit spread and hedging against changes in default exposure, illustrated in Section 13.1.2.2, there was going to be a large residual risk in some direction. And given that it would have been prohibitively expensive to purchase The Systemic Disaster of 2007–2008 105 true default protection for super‐seniors using mezzanine tranches, you can be certain that the hedges actually employed were primarily hedges against credit spread movement, not against default. This highlights just how misleading it was, and how easy it should have been to spot the error of UBS treating statistical hedges as fully eliminating risk. Even in the far more liquid vanilla options market, no one treats positions that are “neutral in the Greeks” as having no residual risk (see Section 11.4, particularly the discussion of Table 11.6). 5.2.5.9 Too Big to Fail Finally, there is the question of why, leaving aside any probabilistic analysis of risk, the sheer size of the positions didn’t trigger alarms. Let me offer an analogy directly from my own experience. During the late 1990s, I was in charge of risk management for Chase’s derivatives business. A very conspicuous part of that business was the new, rapidly growing, and very profitable CDO business, based on commercial loans, not residential mortgages. But like the residential mortgage CDOs of the mid‐2000s decade, the commercial loan CDOs of the late 1990s were starting to run into an accumulation of super‐senior risk that the bank was finding difficult to buy protection on. While I was, whether correctly or incorrectly, quite convinced that the probability of loss on this super‐senior risk was extremely low, making presentations to the firm’s risk committee supporting this view, I was just as strong in my opposition to the continued buildup of super‐senior risk on the firm’s books. Even though limitations on the growth of super‐seniors ultimately meant limitations on the growth of the very profitable CDO business as a whole (for reasons similar to those discussed earlier for mortgage CDOs), the skeptical views of me and my similarly minded risk management colleagues prevailed. Super‐seniors were piling on exposure to what was already the firm’s largest vulnerability as a major commercial lender, exposure to a drastic economic downturn. No matter how remote a possibility we might have regarded such a downturn, it was not a scenario we could completely dismiss. Tett (2009, 65–66) reports a similar decision‐making process around the same time at JPMorgan, prior to the merger with Chase. While all anecdotal recollections of past risk management triumphs, perhaps including my own, should be taken with a grain of salt, what I saw of JPMorgan’s exposures going into the merger were consistent with Tett’s account. Arguments were offered by a few front‐office people at the time of this decision that “in the case of that drastic an economic downturn, the firm will need to be rescued by the Federal Reserve anyway, so what difference does the size of the rescue make?” These arguments were considered wholly without merit by both risk managers and senior management. But one wonders if perhaps this kind of view was behind some of the decision making in 2005–2007. 106 FINANCIAL RISK MANAGEMENT In the wake of the 2007–2008 collapse, suspicions have certainly been expressed that this confidence that regulators and the government owned the downside on big bets was explicitly or implicitly part of the calculation that drove the CDO‐creation machine past reasonable limits. “Moral hazard,” “Greenspan put,” and “too big to fail” have all become part of the common vocabulary used in the postmortem analyses of these decisions (see, for example, FCIR 2011, 57, 61, 341, 356). It is certainly in line with the moral hazard story we told in Section 2.1. And the greater the belief that your firm’s outrageous positions are not out of line with the outrageous positions of your competitors, the greater the tendency for arguments based on ultimate regulator rescue to gain traction. The usual counter of those who find these arguments specious, leaving aside considerations of morality that might not be shared by all discussants, can be summed up in a well‐circulated, but presumably apocryphal, story that goes back at least to the 1970s. In this story, the CEO of a large commercial bank attending an industry conference finds himself at a men’s room urinal next to the crotchety and brusque chairman of the Federal Reserve (in those days, all Fed chairmen were expected to be crotchety and brusque— and male). Looking around and seeing no one else in the room, he whispers to the chairman, “Just between us, would the Fed come to our rescue in a crisis?” The chairman, without looking up, responds, “That is a question I would need to discuss with your successor.” The moral of the story is supposed to be that the penalties for putting your firm in the position of being rescued are personally severe. And certainly one sees evidence of the regulators attempting to enforce this, going out of their way to demand that the price JPMorgan paid for Bear Stearns was punitive to the Bear Stearns stockholders, which included most of the firm’s longtime employees (see McLean and Nocera 2010, 347). And those of us who fought against a “too big to fail” mentality can point to the benefits to firms like Goldman Sachs, JPMorgan, and Deutsche Bank, whose need for government assistance was much less pronounced than Citigroup or Merrill Lynch or UBS. But in the modern era of outsized compensation for senior executives and star traders, which may include so‐called golden parachutes protecting them against the personal consequences of failure, is the government ownership of the downside becoming too great a temptation for risk takers? 5.2.6 Insurers Compared to the voluminous literature about the investment banks in the CDO meltdown, far less has been written about the insurance companies whose sale of protection for super‐senior tranches led to the destruction of The Systemic Disaster of 2007–2008 107 valuable business franchises. And what has been written about the insurance companies is mostly from the standpoint of the errors investment banks made in their reliance on this insurance. My primary source for what follows is FCIR (2011), which addresses AIG on pages 139–142, 200–202, 243–244, 265–274, 344–352, and 376–379 and the monoline insurance companies on pages 204–206 and 276–278. For the most part, these insurance companies appear to have regarded the super‐senior tranches of subprime mortgage CDOs as being virtually without risk of loss. Their analysis can therefore be subject to the same critical examination we have just been through in the previous section for the investment banks. But there is one major difference: The investment banks possessed some genuine expertise in evaluation and modeling of subprime mortgages and of CDO structures. The insurance companies possessed none of this expertise and just relied on analysis by the investment banks and rating agencies for their assurance that risk of loss was practically nonexistent. A telling quote comes from Alan Roseman, CEO of ACA Insurance, one of the monoline insurers: “We were providing hedges on market volatility to institutional counterparties. . . . We were positioned, we believed, to take the volatility because we didn’t have to post collateral against the changes in market value to our counterparty . . . [and] we were told by the ratings agencies that rated us that mark‐to‐market variations [were] not important to our rating, from a financial strength point of view at the insurance company” (FCIR 2011, 276). If this attitude was typical, then the insurers were operating on the premise that there was no genuine risk of loss on the super‐seniors, just annoying fluctuations in mark‐to‐market accounting, presumably due to technical liquidity factors. This view would see the insurers collecting a fee for providing an accounting arbitrage as opposed to being paid for absorbing risk (the accounting arbitrage would arise from an uninsured super‐senior holding at an investment bank being subject to mark‐to‐market earnings fluctuations; once insured, it would no longer need to be marked to market and the insurers did not have mark‐to‐market accounting). If you are just being paid for an accounting arbitrage, then you don’t require any expertise in assessing risk, just a knowledge of accounting rules. While the evidence for how typical this view was is not clear, it certainly does appear that little concern was shown by any of the insurers involved for making their own assessment of credit risk. The only one of these insurers that did begin to show some concern about the volume of exposure they were taking on was AIG (FCIR 2011, 200–201), but its slowdown in taking on CDO risk still left it holding $79 billion in CDO exposure. MBIA, Inc., another of the monoline insurers, stated, according to Norris (2009), that “‘the due diligence standard for a monoline insurer, which MBIA followed,’ did not involve looking into the quality of the securities underlying 108 FINANCIAL RISK MANAGEMENT the securities being insured . . . it primarily relied on the assurances by Merrill Lynch and the credit ratings of Moody’s and Standard & Poor’s.” While this was part of an MBIA suit brought against Merrill Lynch and so might be expected to exaggerate MBIA’s lack of sophistication, it is still revealing that such a claim would be even plausible relative to a business line in which the insurers had bet their entire franchises. 5.3 THE SPREAD OF THE CRISIS The crisis that began in the subprime mortgage CDO market spread to markets, instruments, and institutions that had no direct involvement with either mortgages or CDOs. There were two primary paths through which this spread: contagion through credit exposure to impacted firms, which we examine in Section 5.3.1, and contagion through market impact, which we examine in Section 5.3.2. 5.3.1 Credit Contagion The most direct way for a crisis to spread is through credit exposure to impacted firms. This was certainly a prime ingredient in the 2007–2008 crisis. One of the major paths for credit contagion was the great extent to which financial firms had large counterparty credit exposure to one another through the derivatives markets. I am not including in this the CDO‐related counterparty exposure of many firms to AIG and the monoline insurers, since this was part of the fundamental process creating the crisis. But many firms that may have had no dealings in CDOs had heavy exposure to firms that did have large CDO losses on other derivative contracts such as interest rate and foreign exchange swaps. And this was a decided worry for regulators, as they had to decide on how to handle firms approaching bankruptcy. One can see in the reporting on regulators’ decisions during this period just how big a worry this was (see, for example, FCIR 2011, 291, 329). Some contracts would not be backed by collateral and would result in outright loss; even where there was collateral, there would still be losses resulting from the market impact of so many counterparties simultaneously rushing to sell the collateral and to replace the defaulted derivatives positions. Not only did regulators need to worry about the direct impact on derivatives counterparties of a default, but they also had to be concerned about the potential freezing of derivatives markets as worry about defaults would cause reluctance to enter into new contracts. This in turn could worsen the situation for The Systemic Disaster of 2007–2008 109 counterparties of a defaulting firm, since they might have difficulty finding a replacement for a defaulted derivative contract, exacerbating the original loss. The bankruptcy of one firm might then drive other firms into bankruptcy in an ever widening circle. It is easy to understand the frustration of regulators at being placed in this position. Derivatives trading had originally been almost exclusively conducted on exchanges that had well‐developed procedures for minimizing credit exposure. A major argument of large investment banks in setting up over‐the‐counter derivatives markets as alternatives to exchange‐traded derivatives was that they had the credit systems and expertise that were capable of managing the extra credit risk that would arise. But now they had apparently done so poor a job of managing this that they needed to be bailed out by regulators, and this only a decade after the Long‐Term Capital Management crisis had supposedly led to reforms in counterparty credit management (see Section 4.2.1). Another major path for credit contagion was through financial firms that had made direct loans to firms whose CDO positions threatened them with bankruptcy. This direct lending was primarily in very short maturity instruments, such as commercial paper. Because of the short maturity and the previous sound financial status of major financial firms, this paper was very highly rated by the rating agencies and was supposed to be a very safe investment. Money market mutual funds that bought a diverse portfolio of this paper were considered nearly as sound as government‐guaranteed bank deposits, and regulators worried about the impact on small investors if defaults on commercial paper drove big money market funds to the point of “breaking the buck,” that is, not having sufficient funds to pay back investors’ principal. This, too, was a major concern for regulators as they considered how to deal with firms close to bankruptcy (see Sorkin 2010, Chapter 17). 5.3.2 Market Contagion If you look at Table 13.6, you will see the normal cyclic pattern of defaults of corporate borrowers. Lending institutions have survived this cyclic pattern for decades, building up reserves and capital during times of low defaults that can be used as a buffer against times of higher defaults. But providing credit is a business that requires patience—on the part of bank management, of bank regulators, and of those who invest in banks. When defaults start occurring at an accelerated pace, banks will start cutting back on the volume of new loans, but they won’t start panicking and trying to sell off large blocks of their remaining loans. Credit derivatives, such as credit default swaps and CDOs, brought the promise of increased liquidity to the business of bank lending. When used 110 FINANCIAL RISK MANAGEMENT reasonably, these instruments can be part of a blended strategy, in which some portions of the loan portfolio are judged liquid and managed accordingly while other portions continue to be viewed as illiquid, with a management approach that matches their lack of liquidity. Chapter 13 of this book, and particularly Section 13.3, outlines what I consider an appropriate blend of tools for managing a portfolio that contains both liquid and illiquid credit exposure. By falsely labeling all of the subprime mortgage CDOs as liquid in their desire to obtain more favorable regulatory capital treatment, the investment banks created a dilemma. When falling housing prices started to threaten widening defaults on these CDOs, there was no cushion of reserves or capital to allow for patience, as would have been the case if a large portion had properly been labeled illiquid. And the accounting for liquid instruments meant that banks had to recognize earnings losses immediately through mark‐to‐market accounting and therefore needed to immediately take action to get capital ratios back to allowable levels, since earnings gains and losses immediately impact capital. When a truly liquid position suffers a mark‐to‐market loss that requires an increase in capital, there is a readily available remedy: sell some of the liquid position to reduce the need for capital and also reduce a possible source of further losses requiring capital. This is why mark‐to‐market accounting, stop‐loss limits, and capital allocations designed to allow liquidation of positions over a temporary period of illiquidity fit so well with liquid positions (discussed at greater length in Section 6.1.1). But if you have been only pretending that a position is liquid, you don’t have this option. Since large losses usually occur in periods of economic stress, when raising new capital from investors is difficult and costly, your only remaining choice is selling other positions that truly are liquid to reduce the need for capital. But that doesn’t get the illiquid positions off your books, and if they continue to lose money, you will need to go through this cycle all over again. This is a sketch of how losses on illiquid positions treated as liquid can spread a crisis to other markets by continued forced selling of positions that were not related to the illiquid positions. This is roughly what occurred during the 2007–2008 period, but it was exacerbated by the realization that positions that had been labeled as liquid and as virtually immune to losses were in fact very vulnerable. This raised the level of suspicion in the markets about any asset or derivative that might conceivably have some type of hidden risk. This was another factor in driving down prices and drying up liquidity in other markets (see, for example, Greenlaw et al. 2008, Section 2.1). The full mechanics through which depressed asset values lead to market contagion, with many illustrations from the 2007–2008 crisis, are The Systemic Disaster of 2007–2008 111 covered in more detail in Duffie (2011, Chapter 3). The discussion of this in Brunnermeier (2009, 92–94) is also useful. 5.4 LESSONS FROM THE CRISIS FOR RISK MANAGERS My numbering of subsections is designed to allow easy reference to the discussion of the mechanics of the crisis in Sections 5.2 and 5.3. Sections 5.4.1 through 5.4.6 correspond to sections 5.2.1 through 5.2.6, respectively, while Section 5.4.7 corresponds to Section 5.3.1, and Section 5.4.8 corresponds to Section 5.3.2. 5.4.1 Subprime Mortgage Originators From the viewpoint of internal risk management, the lessons of the crisis regarding both subprime mortgage originators and CDO creators center on issues of legal and reputational risk. These nonquantitative areas of risk management do not align with the focus of this book, which is on risks that can be managed through liquid markets. The comments that I do have on legal and reputational risk can be found in Chapters 3 and 4, particularly Sections 3.2.2 and 3.3. 5.4.2 CDO Creators My comments for CDO creators are identical to those for subprime mortgage originators in Section 5.4.1. 5.4.3 Rating Agencies Key risk management lessons that are reinforced by the rating agency experience leading up to the crisis are the need for a strong separation between models used for risk management and input from traders and structurers (see Section 8.4.3) and the need to have data analysis be responsive to large changes in the market environment (see Section 8.2.8.2). 5.4.4 Investors The key risk management lesson we can draw from the experience of investors is the need for extreme skepticism in looking at marketing claims that you are getting superior returns without taking on additional risk, particularly when your ability to exit trades is limited by illiquidity. This point is discussed at greater length when looking at the experience of insurers in Section 5.4.6—everything said there can be applied here. 112 FINANCIAL RISK MANAGEMENT 5.4.5 Investment Banks The numbering of these subsections has been designed to correspond to the related discussion in Sections 5.2.5.1 through 5.2.5.9. 5.4.5.1 Reliance on Inadequate Derivatives Protection The tools needed to analyze the risk of uncollateralized and weakly collateralized derivatives protection were well known both in the academic literature and in common practice well before these transactions were booked. First, even well‐collateralized protection of illiquid transactions leaves a great deal of remaining risk, since in the event of counterparty default you may have great difficulty in finding a substitute insurance provider (see the bullet point regarding derivatives with actuarial risk in Section 14.3.3). Second, lack of collateralization or weak collateralization needs to be part of the calculation of counterparty credit risk, as emphasized throughout Section 14.3.3. Third, these trades were classical examples of wrong‐way risk, since the circumstances in which the counterparties would need to make insurance payments would be major economic downturns likely to impact the creditworthiness of the counterparties themselves. These trades were also wrong‐way because it was well known that the insurance firms entering into them were entering into many billions of dollars of similar trades with other investment banks; the circumstances that would cause them to have to pay on one trade were highly likely to make them pay on similar trades, and they clearly did not have the financial resources to make payments under all of these contracts (this point is further elaborated in Section 5.2.6). See Section 14.3.4 for a discussion of wrong‐way risk and how to account for it in calculations of counterparty credit risk. Indeed, some of these trades border on being the types of transactions Section 14.3.4 discusses as being so wrong‐way that they should be counted as offering no protection at all—look at the discussion of “end of the world” trades and extreme collateralization triggers. 5.4.5.2 Reliance on Off‐Balance‐Sheet Vehicles The proper risk measurement of liquidity puts is very similar to the measurement of wrong‐way counterparty risk and is addressed in Section 14.3.4. 5.4.5.3 Use of Faulty CDO Models The only point on which I would fault the use of CDO models was that there was too much emphasis on fitting market input and not enough emphasis on modeling that took into account the illiquidity of certain sectors, particularly the super‐senior sector. Section 8.4 addresses model risk for illiquid instruments in general, and Section 13.4 addresses this issue specifically as it relates to CDO tranches. The Systemic Disaster of 2007–2008 113 5.4.5.4 Reliance on External Ratings In Section 13.2.1.1 we discuss the proper use of rating agency input in the risk evaluations of a bank. Rating agency evaluations should always be used as a check on internal assessments, but never as a replacement for them. This should apply just as much to credit risk arising through securities holdings as it does to credit risk arising through traditional bank loans. By similar reasoning, risk managers should always rely on their own internal models of credit portfolio risk and not on rating agency models. The credit portfolio models developed to assess bank loans, discussed in Section 13.3, are exactly the same models that are used to evaluate CDOs, as is made clear in Section 13.4. In fact, the CDO models were simply adapted from preexisting loan portfolio models. 5.4.5.5 Overreliance on VaR Measures As is made clear in Sections 6.1.2 and 8.2.6, VaR can play a proper role in the risk management of illiquid instruments, as long as it is clearly understood that what is being represented in the VaR is a liquid proxy and that a separate analysis of the hedging risk of the illiquid instrument by the liquid proxy is an absolute necessity. 5.4.5.6 Failing to Account for the Illiquidity of Super‐Senior Tranches One point that is stressed several times in this book is that risk management measures must be arrived at independently, without deference to the way accounting is done for internal business decisions or for reporting to the public. Risk managers need to confirm claims of liquidity for an instrument by looking at trading history (both purchases and sales), as emphasized in the last paragraph of Section 6.1.2. Risk calculations and stress test scenarios for super‐seniors required long‐term (life of the security) thinking based on the lack of a liquid market. As is emphasized in Section 8.4, illiquid assets require long‐term risk measures, even when accounting principles insist on mark‐to‐market treatment. 5.4.5.7 Inadequate Stress Tests The key point here is that the illiquidity of the positions required longer‐term stress tests of the type discussed in Sections 8.4.3, 13.3.2, and 13.4.3. 5.4.5.8 Inadequate Analysis of Statistical Hedging As discussed in detail in Sections 8.2.6 and 8.4, when dealing with illiquid instruments it is vital that risk measures and reserves utilize detailed simulations of potential hedging costs over the life of the instrument to arrive at a conservative estimate. The specific case of hedging illiquid CDO tranches with more liquid CDO tranches is discussed in Section 13.4.3. 114 FINANCIAL RISK MANAGEMENT 5.4.5.9 Too Big to Fail Risk managers should be vigilant in arguing against reasoning that relies on indifference to the size of losses in the event of serious economic downturns. However, many people motivated by this reasoning will not articulate it but will look for other arguments that will disguise their true incentives. One outcome of the crisis is to recognize that the design of trader and executive compensation schemes has an important risk management component. See, for example, Turner Review (2009, 79): “In the past neither the FSA [the British bank regulator] nor bank regulators in other countries paid significant attention to remunerations structures. And within firms, little attention was paid to the implications of incentive structures for risk taking, as against the implications for firm competitiveness in the labour market and for firm profitability. In retrospect this lack of focus, by both firms and regulators, was a mistake. There is a strong prima facie case that inappropriate incentive structures played a role in encouraging behaviour which contributed to the financial crisis.” See also Financial Stability Forum (2008, Recommendation II.19). To what degree this is workable in practice remains to be seen. Much of the burden for proper controls on asymmetric incentives will probably rest with government legislation and regulation, as we will investigate more closely in Section 5.5.5. 5.4.6 Insurers The key risk management lesson we can gain from the experience of AIG and the monoline insurers is similar to the lesson we can take from the experience of investors—the need for extreme caution in taking on illiquid risks in an area in which you lack expertise. This lesson is even more pointed for the insurers, since they took on levels of risk that destroyed their franchises, something that few investors did. As we emphasized in Section 2.3 on adverse selection, risk managers should always be especially vigilant when traders are dealing in transactions for which they do not possess an informational advantage. The temptation to get involved may be great when a plausible case has been made that returns are high relative to risk, but even when this case seems overwhelming, the size of risk must be kept proportional to liquidity. The risk manager’s greatest friend is always the stop‐loss limit, as discussed in Section 6.1.1. Even for risks that the firm does not understand well, a limit can be placed on tolerable losses and an exit strategy planned. But when lack of liquidity means you won’t be able to exit as losses mount, no degree of promised return should be allowed to jeopardize a firm’s franchise. If an opportunity just seems too good to pass up, then invest in the expertise to manage it knowledgeably. Relying on regulatory constraints or advisory services, such The Systemic Disaster of 2007–2008 115 as rating agencies, cannot be considered in any way a substitute for this expertise when illiquidity prevents easy exit. 5.4.7 Credit Contagion Most of the ideas for reducing the risk of credit contagion are being addressed at the regulatory level and are covered in Section 5.5.7. Chapter 14 addresses counterparty credit risk management at the level of the firm. 5.4.8 Market Contagion The 2007–2008 experience on the degree to which market illiquidity spread and the length and depth of this illiquidity will need to impact the historical measures of VaR risk (Section 7.1) and the severity of stress tests (Section 7.2) that will be utilized going forward. 5.5 LESSONS FROM THE CRISIS FOR REGULATORS The numbering of subsections, as in Section 5.4, is designed to allow easy reference to the discussion of the mechanics of the crisis in Sections 5.2 and 5.3. Sections 5.5.1 through 5.5.6 correspond to Sections 5.2.1 through 5.2.6, respectively, while Section 5.5.7 corresponds to Section 5.3.1, and Section 5.5.8 corresponds to Section 5.3.2. Throughout this section, I have relied as much as possible on recommendations from the Financial Stability Board and its predecessor organization, the Financial Stability Forum. This organization is a joint effort of finance ministers and central bankers from the G‐20 countries, as well as major global public institutions such as the International Monetary Fund, the World Bank, and the Bank for International Settlements. It was established to coordinate financial regulation and standards setting globally. Many of its recommendations carry the endorsement of the G‐20, the group of 20 leading economies that account for over 80 percent of global gross domestic product (GDP). The G‐20 fosters cooperation and consultation on matters relating to the international financial system. As such, I believe it represents the broadest consensus views of the regulatory community. The Financial Stability Forum’s 2008 recommendations for enhancing market and institutional stability will be referred to as FSF (2008). I will bring in views of other regulatory bodies and of academics where there are significant disagreements or where a particular document has expressed a view with particular clarity. Two sources that this section utilizes often are the Turner Review of 2009, a broad review of regulatory policy authorized 116 FINANCIAL RISK MANAGEMENT by the British government, and the 2009 report on financial reform by the Group of Thirty, the same private, nonprofit group of leading representatives of the international business, regulatory, and academic communities whose influential report on derivatives risks I make heavy use of in Section 6.1.1. 5.5.1 Mortgage Originators There has not been as much focus on mortgage originators in the recommendations arising from the crisis as there has been on CDO creators and rating agencies. Davidson (2007) does have a persuasive suggestion: “There needs to be capital at the origination end of the process. Without capital, representations and warranties have no value. To achieve this, brokers (or whoever has direct contact with the borrower) should be licensed and bonded and firms in the chain of reps and warrants need to maintain sufficient reserves to support their financial promises. This capital would be available to assess damages in the case of fraudulent or predatory practices that hurt borrowers and homeowners.” One other interesting recommendation is to restructure mortgages to avoid the impact of negative home equity on homeowner defaults. Shiller (2008, Chapter 6) has several interesting suggestions along this line, including real estate derivatives, home equity insurance, and continuous‐ workout mortgages. A mortgage market that builds in this protection exists in Denmark. George Soros, in a Wall Street Journal article on October 10, 2008, explained the Danish mortgages as follows: “Every mortgage is instantly converted into a security of the same amount and the two remain interchangeable at all times. Homeowners can retire mortgages not only by paying them off, but also by buying an equivalent face amount of bonds at market price. Because the value of homes and the associated mortgage bonds tend to move in the same direction, homeowners should not end up with negative equity in their homes. To state it more clearly, as home prices decline, the amount that a homeowner must spend to retire his mortgage decreases because he can buy the bonds at lower prices.” 5.5.2 CDO Creators Recommendations have focused on trying to better align incentives between CDO creators and investors. For example, the Group of Thirty (2009, Recommendation 13) states that “regulators should require regulated financial institutions to retain a meaningful portion of the credit risk they are packaging into securitized and other structured credit products.” A requirement of this type is a part of both the Dodd‐Frank legislation in the United States and rules proposed by European authorities (see Global Legal Group The Systemic Disaster of 2007–2008 117 2011, Chapter 3: “EU and US Securitization Risk Retention and Disclosure Rules—A Comparison”). All such proposals face three large challenges: 1. How to measure credit risk retention given all of the ways that are now available for offsetting risk through credit derivatives, including the use of credit indexes. 2. How to avoid the situation discussed in Section 5.2.2 in which CDO creators viewed the total package as so lucrative that they could regard the retained equity as a “free good” to whose credit performance they were indifferent. Proposals to deal with this are combinations of raising the portion of risk retained and of requiring that a portion of risk be retained in all tranches sold, not just in a single tranche whose losses might not be well correlated with the tranches sold. 3. How to avoid making retention requirements so onerous that they discourage or raise the costs of securitization that is considered beneficial to the general public (e.g., homeowners). 5.5.3 Rating Agencies FSF (2008, Section IV) contains several proposals to deal with the rating agency issues that were raised by the crisis. I would highlight the following: ■ ■ ■ ■ ■ Rating agencies “should clearly differentiate, either with a different rating scale or with additional symbols, the ratings used for structured products from those for corporate bonds.” This would clarify the greater reliance of structured product ratings on models and economic assumptions and their “potential for significantly higher ratings volatility.” Rating agencies “should enhance their review of the quality of the data input and of the due diligence performed on underlying assets by originators, arrangers and issuers involved in structured products.” Regulatory authorities should “review their use of ratings in the regulatory and supervisory framework” to encourage investors to “make independent judgment of risks and perform their own due diligence” and reduce “uncritical reliance on credit ratings as a substitute for that independent evaluation.” “Investor associations should consider developing standards of due diligence and credit analysis for investing in structured products.” Rating agencies should revise their codes of conduct to better deal with conflict of interest issues and “demonstrate that they have the ability to maintain the quality of their service in the face of rapid expansion of their activities.” Rating agencies “should disclose past ratings in a more systematic way, and improve the comparability of their track records.” 118 FINANCIAL RISK MANAGEMENT Many of these points are echoed in the Group of Thirty (2009, Recommendation 14) and Richardson and White (2009). Richardson and White also suggest as an alternative that financial regulations be changed to deemphasize the use of rating agencies. In this approach, “regulated financial institutions would thus be free to take advice from sources that they considered to be most reliable—based on the track record of the advisor, the business model of the advisor (including the possibilities of conflicts of interest), the other activities of the advisor (which might pose potential conflicts), and anything else the institution considered relevant.” But “the institution would have to justify its choice of advisor to its regulator.” This alternative could lead to more competition and new approaches in the ratings advisory market. 5.5.4 Investors Most regulatory discussion concerning investors has been tied to protecting investors from CDO creators and rating agencies or to controlling the spread of financial crises through credit contagion or market contagion. These issues are addressed in other sections—5.5.2 for CDO creators, 5.5.3 for rating agencies, 5.5.7 for credit contagion, and 5.5.8 for market contagion. One attempt to address regulatory concerns for investors directly is in the Squam Lake Report (Squam Lake Group 2010, Chapter 4: “Regulation of Retirement Savings”). Given that many of the investors who had losses in the 2007–2008 disaster were pension funds reaching for excess returns in exchange for risks that may have been very poorly understood by the employees who were the ultimate recipients of these losses, this is a timely concern. Among the recommendations in this chapter are: ■ ■ ■ Requiring simple standardized disclosure for products offered in defined contribution retirement plans. Requiring simple and meaningful standardized disclosure of measures of long‐term risk and of investment costs. Any advertisement of average prior returns should also include a standardized measure of uncertainty. The standard part of a defined contribution plan should be restricted to well‐diversified products with low fees. 5.5.5 Investment Banks The regulatory responses to the default and near default of several major investment banks in the 2007–2008 crisis can be roughly divided into four categories. The first consists of measures to require tightening of internal risk management procedures, combined with greater scrutiny of these The Systemic Disaster of 2007–2008 119 procedures by regulatory authorities. The second is a new focus on regulatory oversight of compensation policy. The third is significantly higher requirements for bank capital to serve as a buffer against losses before they impact depositors, and therefore governments and taxpayers. And the fourth consists of proposed restrictions on the size and range of activities of investment banks. We consider each in turn. 5.5.5.1 Tightened Internal Risk Management Procedures It is natural for part of the regulatory response to the crisis to be to call for stronger internal risk management procedures within investment banks. But without either very specific guidance on how procedures should change or new ongoing regulatory scrutiny to ensure compliance, this will just be empty exhortation. This book addresses specific regulatory guidance on particular risk management issues at those points where it is most relevant: new guidance on model review in Chapter 8; new guidance on oversight of compensation policy later in this section; new guidance on stress tests as part of the discussion of increased capital later in this section; new guidance on counterparty risk in Section 5.5.7 on stemming credit contagion; new guidance on valuation in Section 5.5.8 on stemming market contagion. This subsection addresses proposals for increased regulatory scrutiny to ensure compliance. The most comprehensive set of recommendations for changes in regulatory scrutiny of investment bank internal controls is in the Group of Thirty (2009) Recommendations 1, 2b, 6, 7, and 8. Some of the key points in these recommendations are: ■ ■ ■ At a national level, countries should “eliminate unnecessary overlaps and gaps in coverage . . . removing the potential for regulatory arbitrage, and improving regulatory coordination.” The Squam Lake Report (Squam Lake Group 2010, Chapter 2) goes further, calling for a single regulatory authority in each country to be “responsible for overseeing the health and stability of the overall financial system.” At the international level, national regulatory authorities should “better coordinate oversight of the largest international banking organizations” and “move beyond coordinated rule making and standard setting” to convergence in application and enforcement of standards and closing of regulatory gaps. In countries where the central bank is not the primary regulator of banks, the central bank needs to become more involved in regulation, particularly with regard to the largest systemically significant firms and critical payment and clearing systems. 120 FINANCIAL RISK MANAGEMENT Saunders, Smith, and Walter (2009) call for a dedicated regulator in each country for “large complex financial institutions (LCFIs),” arguing that these firms are different in character and pose a greater threat to the global financial system than smaller and more specialized firms. “Most importantly, the regulator would have the power and the obligation to ensure that LCFIs operate consistently with priority attention to the institution’s safety and soundness, even if this can only be achieved at the cost of reduced growth and profitability.” 5.5.5.2 Compensation Policy The quotation from the Turner Review (2009) in Section 5.4.5.9 indicates the change in regulatory attitude toward the compensation structures of investment banks. The 2007–2008 crisis has led regulatory authorities to switch from regarding compensation as purely an internal matter for banks to regarding it as a key component of risk control. The Financial Stability Board issued a separate report on sound compensation practices, FSF (2009b), which states that the “perverse incentives” of generous bonus payments for high short‐term profits “without adequate regard to the longer term risks they imposed on their firms” “amplified the excessive risk‐taking that severely threatened the global financial system and left firms with fewer resources to absorb losses as risks materialized.” Some key principles elucidated in FSF (2009b) are: ■ ■ ■ ■ ■ Compensation must take into account both profit generated and risk entailed. The firm’s board of directors must actively oversee the design and operation of compensation policy and must ensure that the compensation policy addresses the balance between profit and risk. “Compensation must be adjusted for all types of risk,” including difficult‐to‐measure risks such as liquidity risk and reputational risk. This necessitates that both quantitative measures and human judgment play a role in determining risk adjustments. (This is consistent with this book’s emphasis on the need for subjective judgment in risk management; see Sections 1.3 and 6.1.1.) Compensation outcomes should be symmetric with risk outcomes, with bonuses diminishing or disappearing in the event of poor firm, divisional, or business unit performance. “Compensation payout schedules must be sensitive to the time horizon of risks,” with compensation deferred when risks are realized over long periods. (This is consistent with the distinction made on the differing risk management approaches for liquid and illiquid positions in Section 1.2 and 6.1.1. The mix of cash and equity in compensation also needs to be consistent with the nature and time horizon of risks generated.) The Systemic Disaster of 2007–2008 ■ ■ 121 “Firms should disclose clear, consistent and timely information about their compensation practices” to make sure that all stakeholders, including customers, creditors, and regulators as well as stockholders, can make informed decisions. Regulators must include review of compensation practices as part of their supervisory role and be prepared to take prompt action when compensation practices are deemed deficient. “Compensation is an incentive system, not simply a market wage” and so must be subject to regulatory review. Given the competitive nature of the labor market for financial institutions, “Market participants are pessimistic about the effectiveness of change unless it is industry‐wide and global. . . . Changing compensation practice will be challenging, time‐consuming and involve material costs. Therefore, in the absence of sustained external pressure, firms may fail to carry through on originally good intentions. Although some market participants are wary of regulatory pressure, many believe that a widespread change in practice can be achieved only with the help of supervisory and regulatory agencies, which should coordinate at the global level.” Other regulatory publications in response to the crisis are quite consistent with FSF (2009b). See, for example, Turner Review (2009, Section 2.5 (ii)). Clementi et al. (2009) provide an academic analysis very supportive of this approach. The FSF (2009b) compensation proposals are primarily aimed at those directly involved in the creation and management of risk positions—traders, marketers, and structurers. The Squam Lake Report (Squam Lake Group 2010, Chapter 6) suggests an interesting approach aimed at the compensation of senior management of financial institutions. Since “governments will bail out financial firms during a crisis,” “the stakeholders in financial firms—executives, creditors, and shareholders—do not face the full cost of their failure. This in turn increases the likelihood of bank failures, the potential for systematic risk, and expected taxpayer costs.” Along with other measures, a “mechanism for inducing financial firms to internalize the costs of their actions” would be holdbacks of a fixed dollar amount of compensation that would be forfeited “if the firm goes bankrupt or receives extraordinary government assistance” over some defined future time period. Some further points raised about this proposal are: ■ ■ “More familiar forms of deferred compensation, such as stock awards and options” help align manager incentives with stockholders’ interests but do not align them with taxpayers’ interests. “Resignation from the firm should not accelerate payment of an employee’s holdbacks,” since this would “weaken their concern about the 122 ■ FINANCIAL RISK MANAGEMENT long‐term consequences of their actions. . . . In the same spirit, managers should not be rewarded for taking their firms into bankruptcy. If a firm declares bankruptcy, its managers should receive their holdbacks only after its other creditors have been made whole.” “[D]eferred compensation leans against management’s incentive to pursue risky strategies that might result in government bailouts. Similarly, rather than wait for a bailout during a financial crisis, the management of a troubled firm would have a powerful incentive to find a private solution, perhaps by boosting the firm’s liquidity to prevent a run, raising new capital, or facilitating a takeover by another firm.” Rajan (2010) Chapter 7 and the section on “Reducing the Search for Tail Risk” in Chapter 8 offer strong supporting arguments for both the FSF and the Squam Lake proposals. 5.5.5.3 Capital Requirements Regulators have certainly taken many steps since the crisis to raise the levels of capital required. They have taken steps both to increase the level of risk‐weighted assets that will be calculated against trading positions and to increase the capital required for a given level of risk‐weighted assets. A good summary of the steps taken by the BIS is PricewaterhouseCoopers (2011), in which Chapter 4 covers increases in risk‐weighted asset calculations and Chapter 3 covers increases in capital. While the direction of the regulatory response is clearly correct, the specifics of the approach are troubling. Though regulators have enhanced stress‐testing requirements (PricewaterhouseCoopers 2011, Sections 11.1 and 11.3), there is no direct tie between these stress tests and capital requirements. It is true that there is now a stressed VaR calculation that impacts capital (PricewaterhouseCoopers 2011, Sections 4.6.3.3 and 11.2.3), but this just stresses VaR parameters, a form of stress testing that has been shown to be inadequate (see Section 7.2.1 and the discussion of the use of stress tests by Long‐Term Capital Management (LTCM) in Section 4.2.1; as I state there, LTCM “did run stress versions of VaR based on a higher than historical level of correlations, but it is doubtful that this offers the same degree of conservatism as a set of fully worked‐through scenarios”). What I find most troubling is that the degree of complexity of capital computations has grown to the point that there is great danger of risk managers and regulators losing sight of the largest risk exposures through the distraction of enlarged reporting requirements. PricewaterhouseCoopers (2011) expresses a similar apprehension in Section 4.7: “A particular area of concern is the introduction of many systems within the market risk process. Previously market risk departments have been reliant on one regulatory risk system, but they can now have up to five systems to manage. This in itself The Systemic Disaster of 2007–2008 123 is likely to increase the operational risk associated with market risk. Even though these measures are being introduced to ensure more comprehensive measurement, their complexity may cause banks to miss positions and, as always, there will be loopholes in the systems, harder to find but also harder to catch.” By contrast, capital requirements directly tied to stress‐test scenarios would focus management and regulatory attention in exactly the right place: on the impact of large moves in major economic variables, exactly the types of events that have led to the events that challenge the health of financial firms and of the financial system. It is this type of event, significant drops in price of important asset classes, for which capital cushions are needed. My arguments supporting this approach can be found in Sections 7.2.2 and 7.3, particularly toward the end of 7.3 where I discuss the reasons that Chase Manhattan moved to basing internal capital requirements on stress tests in the late 1990s. There may be two objections to basing regulatory capital on stress‐test scenarios. The first is that the limited number of individually tailored scenarios that can be considered may allow some risks that avoid attracting capital. This can be dealt with either by having some part of the capital requirement based on VaR or by utilizing statistically driven stress tests as supplements to individually tailored ones, as described in Section 7.2.3. The second objection is the inevitable subjectivity of stress‐test scenarios. Some element of subjectivity is unavoidable and, in fact, welcome, as emphasized in Sections 6.1.1 and 7.2.2. U.S. and European regulators have had no trouble specifying stress‐test scenarios in the capital adequacy tests mandated in the wake of the crisis (see, for example, Federal Reserve Board 2009). I advocate utilizing these crisis tests as a precedent for an ongoing process, for the following reasons: ■ ■ Whatever level of possible stress market move corresponds to the capital requirement, there will always be some possibility that an extreme market move will exceed this level and require some absorption of loss by taxpayers on behalf of depositors. Since it is the regulatory authorities that represent the taxpayers’ interests, they should be the ones to determine the level of protection. There are inevitable trade‐offs between capital requirements that are too high and hurt economic activity and capital requirements that are too low and create too high a risk of potential crisis. It is the regulatory authorities acting on behalf of government that should be weighing these consequences and deciding on the correct balance. Regulatory authorities could signal their willingness to support certain markets in a liquidity crunch by differentiating by instrument between 124 ■ ■ FINANCIAL RISK MANAGEMENT the time periods over which stress tests need to be run. For example, a two‐week stress event might be considered adequate for government bond and spot foreign exchange markets, signaling government readiness to intervene to quickly restore liquidity in these markets, but a three‐month stress event might be required for structured securities in which the government wished to indicate less urgency to intervene to restore liquidity. The risk managers of a firm should possess specialized knowledge regarding the trading positions and activities of that firm. There is no reason to think they would possess any specialized expertise about the probability of macroeconomic events, such as large moves in a stock indexes, government bond rates, or housing prices. So I do not see any comparative advantage argument in favor of having these stress levels be set by firm risk management as opposed to government regulators. When firm risk managers set the stress levels, there is an inevitable competitive pressure to set levels lower to free up capital and improve returns. A common level set by regulatory authorities would eliminate the competitive advantage a firm could get by hiring more optimistic risk managers. If political pressures prevent regulatory authorities from setting these levels on a regular basis, I would urge financial institutions to seek a way that a common level could be set by an industry association. 5.5.5.4 Limitations on Size and Allowable Activities The other regulatory proposals considered in this section, regarding tightened risk management procedures, capital requirements, and compensation, do not seek to fundamentally change the structure of the financial industry. Some suggested actions do address fundamental structure directly, trying to eliminate a “too big to fail” mentality either by placing limits on the size of financial firms or by creating a strong separation between firms that can engage in certain types of activities and firms that receive any kind of government support. No proposals of this type were part of the FSF (2008) recommendations, and the Turner Review explicitly rejected proposals for separation of activities, stating in Section 2.9 that “It does not therefore seem practical to work on the assumption that we can or should achieve the complete institutional separation of ‘utility banks’ from ‘investment banks.’ . . . Large complex banks spanning a wide range of activities are likely to remain a feature of the world’s financial system.” Points in support of this view offered by the Turner Review (2009) are: ■ A reimposition of Glass‐Steagall type separation between commercial and investment banking is impractical, given that many activities that used to be conducted solely by investment banks, such as the The Systemic Disaster of 2007–2008 ■ ■ 125 underwriting of corporate bonds, are now “core elements within an integrated service to corporate customers in a world where a significant element of debt is securitized.” Many so‐called narrow banks that focused almost entirely on traditional commercial and retail banking activities, such as Northern Rock, Washington Mutual, and IndyMac, also failed during the crisis. The international integration of financial markets would make it difficult to achieve such a separation without a broad consensus among governments, which is unlikely to be achieved. Another point in support of this view comes from Rajan (2010, 173): “Proprietary trading . . . is another activity that has come in for censure. . . . Critics argue that proprietary trading is risky. It is hard to see this as an important cause of the crisis: banks did not get into trouble because of large losses made on trading positions. They failed because they held mortgage‐ backed securities to maturity, not because they traded them.” Rajan’s analysis of the causes of bank failure in the crisis is certainly supported by Section 5.2.5 of this book. A major proponent of at least considering fundamental changes to industry structure is the Group of Thirty (2009), which in its Recommendation 1 proposes that “Large, systemically important banking institutions should be restricted in undertaking proprietary activities that present high risks and serious conflict of interest” and states that “nationwide limits on deposit concentration should be considered.” It is perhaps not coincidental that the steering committee for this report was chaired by Paul Volcker, whose “Volcker rule” (see McLean and Nocera 2010, 366) for the ban on much proprietary trading activity by deposit‐taking banks has been one of the principal legislative efforts in this direction. In support of its proposal for restrictions on proprietary trading, the Group of Thirty report states that “What is at issue is the extent to which these approaches can sensibly be combined in a single institution, and particularly in those highly protected banking institutions at the core of the financial system. Almost inevitably, the complexity of much proprietary capital market activity, and the perceived need for confidentiality in such activities, limits transparency for investors and creditors alike. In concept, the risks involved might be reduced by limiting leverage and attaching high capital standards and exceptionally close supervision. Some members of the G30 feel such an approach could be sufficient to deal with these risks. . . . Experience demonstrates that under stress, capital and credit resources will be diverted to cover losses, weakening protection of client interests. . . . Moreover, to the extent that these proprietary activities are carried out by firms supervised by government and protected 126 FINANCIAL RISK MANAGEMENT from the full force of potential failure, there is a strong element of unfair competition with ‘free‐standing’ institutions.” Other proponents of limits on industry structure are Roubini and Mihm (2011), who advocate both limits on size (223–230) and a reimposition of a (greatly expanded) Glass‐Steagall (230–233), and Stiglitz (2010, 164–168), who quotes former Bank of England governor Mervyn King: “If some banks are thought to be too big to fail . . . then they are too big.” Rajan (2010, 169–176) provides a very incisive analysis of these proposals. I would highly recommend this to anyone interested in this topic. While Rajan is skeptical of most of the value of most of these suggestions, he is sympathetic to the idea of limiting proprietary trading, not because it will reduce risk of bank failure, but because of the inherent conflict of interest between banks’ proprietary trading and the interests of their customers. Rajan argues that “Banks that are involved in many businesses obtain an enormous amount of private information from them. This information should be used to help clients, not trade against them.” But Rajan does clarify that he supports limiting bank proprietary trading, not eliminating it, because “some legitimate activities, including hedging and market making, could be hard to distinguish from proprietary trading.” My own experience supports Rajan on this point; see my account of market making in Section 9.1. 5.5.6 Insurers FSF (2008, Recommendation II.8) calls for insurance regulators to strengthen the regulatory and capital framework for monoline insurers in relation to structured credit. 5.5.7 Credit Contagion In response to the large role that counterparty credit risk on over‐the‐counter (OTC) derivatives played in credit contagion in the 2007–2008 crisis, it is natural that a major focus of regulatory concern has been to attempt to minimize future use of OTC derivatives and increase the use of exchange‐traded derivatives. Section 14.2 will review many of the advantages that exchange‐traded derivatives have relative to OTC derivatives in minimizing credit exposure: ■ ■ The elimination of credit exposure between counterparties, with all credit exposure centralized with the exchange (or associated clearinghouse). The relatively automatic mechanisms for margining, posting of collateral, and closing out of positions that minimize the credit exposure of the exchange. The Systemic Disaster of 2007–2008 ■ ■ ■ 127 The mutualized sharing of the residual counterparty risk among all members of the exchange. The ease with which counterparties can extinguish existing positions, reducing credit exposure levels. The greater transparency and information‐sharing that are encouraged by the exchange’s lack of any market exposure. The Squam Lake Report (Squam Lake Group 2010, Chapter 9) does an excellent job of laying out these arguments concisely in the context of reducing the risk of credit contagion in a crisis. A particular point the Squam Lake Report raises relative to crises is that the ease with which counterparties can extinguish existing positions also reduces demand for collateral, “a precious resource, especially during a financial crisis.” In response to these arguments, regulatory bodies have been highly motivated to push regulated institutions in the direction of reducing their use of OTC derivatives relative to exchange‐traded derivatives. While almost all observers agree that this is a move in the right direction, some cautions have been sounded on two grounds: (1) In the process, some of the advantages to customers of OTC derivatives relative to exchange‐traded derivatives, detailed in Section 14.3, will be lost, and (2) as more trading volume is funneled to exchanges, the exchanges may grow to the point where they will become a potential source of systemic risk that could trigger or exacerbate a crisis. Before looking at these warnings, let’s first summarize the actions being contemplated. In November 2009, the G‐20 summit issued a recommendation that “All standardised OTC derivative contracts should be traded on exchanges or electronic trading platforms, where appropriate, and cleared through central counterparties by end‐2012 at the latest” (see Financial Stability Board 2010). Obviously much of the force of this recommendation will turn on exactly how the word “standardised” is interpreted. In addition to those contracts that are being mandated to be traded on exchanges, powerful incentives are being put in place to encourage the replacement of OTC derivatives by exchange‐traded derivatives, by mandating more stringent capital requirements on OTC derivatives. These actions are covered in PricewaterhouseCoopers (2011, Chapter 5). Section 5.3.1.7 of PricewaterhouseCoopers explains that “new rules provide banks with strong incentives to move trades to a central counterparty clearing house (‘CCP’) with exposures to CCPs assigned fairly low risk weights. To complement this, the [Basel] Committee supports enhanced capital standards and rigorous risk management for CCPs. It has therefore specified that the favourable treatment of exposures to CCPs applies only where the CCP complies” with regulatory standards. 128 FINANCIAL RISK MANAGEMENT Now let’s turn to possible objections. The first is the possible loss of the advantages of OTC derivatives over exchange‐traded derivatives for some contracts. As detailed in Section 14.3, these are principally the ability to more closely customize OTC derivatives to client needs, less stringent operational requirements, and the willingness of OTC market makers to extend credit beyond what exchanges offer, along with occasional restrictions on trading that disadvantage some customers, mentioned in Section 14.2. How these concerns will be dealt with depends very much on implementation. For example, if the term “standardised” in the G‐20 recommendation of the previous paragraph is interpreted narrowly, it will not hamper customization much, but will leave a substantial portion of OTC derivatives outside clearinghouses. The second possible objection is that concentrating more derivatives trading in exchanges will increase the risk that the exchanges themselves will become a potential source of systemic risk. The clearest exposition of this argument is contained in Pirrong (2011). While exchanges have well‐developed mechanisms for containing credit risk, these are not perfect. As explained in Section 14.2, exchanges are exposed to counterparty risk in between margin calls, and their protection against this is much the same type of VaR and stress‐test calculations that have failed to prevent banks from being a source of systemic risk. While exchanges have avoided exposure to the illiquid instruments that have frequently been the source of problems for banks, this has been achieved by limiting exchange trading to the most liquid contracts; the price of concentrating more derivatives trading in exchanges may be to expose exchanges to more illiquid instruments. A balanced approach to the trade‐off between reduction of credit risk on OTC derivatives and avoiding the potential for systemic risk at exchanges is the Federal Reserve Bank of New York staff report, Duffie, Li, and Lubke (2010). While calling for measures that will increase the use of exchange trading for more liquid derivatives, a number of measures short of exchange trading are proposed to reduce the systemic risk of less liquid OTC derivatives. These include: ■ ■ ■ Increased capital requirements reflecting not just a bank’s exposure to counterparty default but also “the risks that it imposes on others” by its own risk of default. Increased public transparency of aggregate price and volume information and “going prices,” closer to the level of transparency available for exchange‐traded derivatives. Aggressive trade compression, along the lines discussed in Section 14.3.5 of this book. The Systemic Disaster of 2007–2008 129 One further area that regulators have considered for containing credit contagion is regulation of money market funds. The Group of Thirty (2009, Recommendation 3) calls for “Money market mutual funds wishing to continue to offer bank‐like services, such as transaction‐account services, withdrawal on demand at par, and assurances of maintaining a net asset value (NAV) at par . . . to reorganize as special‐purpose banks, with appropriate prudential regulation and supervision, government insurance, and access to central bank lender‐of‐last‐resort facilities.” Any money market fund not willing to subject itself to these requirements would not be permitted to offer “explicit or implicit assurances to investors that funds can be withdrawn on demand at a stable NAV.” 5.5.8 Market Contagion Three types of measures have been proposed to limit the spread of problems for any one firm to other firms through market contagion. The first is to limit the pressures on financial firms facing difficulties to quickly shrink balance sheets, thereby reducing downward pressure on markets from distressed selling. These measures are classified as ones to reduce procyclicality. The second is to provide for a more orderly process for placing a firm in bankruptcy, allowing more time for positions to be unwound. The third is to provide regulatory oversight for financial entities that might be impacted by financial contagion, to provide regulators with greater knowledge about positions that could be impacted through market contagion. We’ll consider each in turn. 5.5.8.1 Reducing Procyclicality The primary regulatory effort in this direction has been to require capital buffers that should be built up in periods of good profitability and drawn down in periods of stress. By having some portion of required capital that it is permissible to draw upon in a crisis, the intention is to relieve the pressure on banks to sell off assets in response to a sharp fall in market valuation. FSF (2009a, Section III) calls for “the capital framework . . . [to] be enhanced to provide stronger capital buffers during strong economic conditions that can be drawn down to a credible minimum requirement during periods of economic and financial stress.” Group of Thirty (2009, Recommendation 10) calls for mandated capital ratios to “be expressed as a broad range . . . with the expectation that as part of supervisory guidance, firms will operate in the upper end of such a range in periods when the market is exuberant and tendencies for underestimating and underpricing risk are great.” These recommendations have been acted on by the Basel regulators through requirements for capital buffers that can be drawn down during periods of economic stress. Details 130 FINANCIAL RISK MANAGEMENT of these requirements can be found in PricewaterhouseCoopers (2011, Sections 10.3.3 and 10.3.4). While capital buffers have been the focus of the regulatory response to procyclicality, some thought has also been given to reducing the cyclicality of accounting rules. With regard to provisions for loan losses, the Group of Thirty (2009, Recommendation 12(c)) calls for accounting principles that are “more flexible in regard to the prudential need for regulated institutions to maintain adequate credit‐loss reserves sufficient to cover expected losses across their portfolios over the life of the assets in those portfolios,” while maintaining transparent disclosure of reserve methodology. This recommendation runs counter to much of the past decade’s tendencies in accounting for loan loss provisions, which have emphasized provisioning only when loss potential on specific loans starts to become apparent (the “incurred loss” model). FSF (2009a, Section IV) also recommends reconsideration of the “incurred loss model by analyzing alternative approaches for recognizing and measuring loan losses that incorporate a broader range of available credit information.” The FSF states that such alternative approaches might have identified loan losses earlier in the credit cycle and potentially reduced procyclicality. The Basel regulators have begun promoting a longer‐run approach toward accounting for loan loss provisions, based on long‐term data series for default probabilities and historically conservative assumptions for loss given default. These actions are detailed in PricewaterhouseCoopers (2011, Sections 10.3.1 and 10.3.2). It would be consistent with this longer‐term approach to loan‐loss provisioning to move to a longer‐term approach to valuation of illiquid securities and derivative positions. This would have the same impact of building up reserves during buoyant markets that would reduce the pressure to liquidate assets in times of stress (for a more detailed discussion, see Section 8.4.4). The Group of Thirty (2009, Recommendation 12) calls for a move in this direction. In the supporting discussion, the Group of Thirty argues for “more realistic guidelines for addressing valuation issues for illiquid investments.” FSF (2008, Section III.3) also contemplates changes in this direction. Zandi (2009, 258–259) makes a similar suggestion: that to keep banks’ survival from being threatened in financial crises, “mark‐to‐market accounting rules could be tweaked most importantly for securities that financial institutions don’t ever plan on selling. . . . It is reasonable for institutions to value these securities based on expectations of any losses they might eventually suffer, but it isn’t reasonable to value these securities using prices they would get if they sold them today.” Where my proposal would differ from Zandi’s is that my criteria would be liquidity and not intention of sale, and for illiquid securities I would replace the prices at which they could be sold today with very conservative estimates of losses, as opposed to expected The Systemic Disaster of 2007–2008 131 losses. I believe large reserves are needed against illiquid instruments, but conservatism should make for relatively stable reserve levels that would only rarely need to be increased in a crisis. 5.5.8.2 More Orderly Bankruptcy The Group of Thirty (2009, Recommendation 16) calls for legislation to give regulators greater authority to provide for “orderly closings of regulated banking organizations, and other systemically significant regulated financial institutions.” The reasoning behind this recommendation states that “Market discipline works best in a system in which failures can happen without being a source of major disruption and contagion.” “To be fully effective, the legal regimes that operate once a failure is triggered should be modified, with a view to placing primary importance on the capacity of the authorities to take actions to protect the health of the system.” PricewaterhouseCoopers (2011, Chapter 15) provides a summary of actions that have been taken by international regulators along these lines, particularly with regard to requiring each large financial institution to prepare a “resolution plan” for the firm’s orderly liquidation in the event of insolvency. The Squam Lake Report (Squam Lake Group 2010, Chapter 8) provides specific recommendations for preparing resolution plans. Huertas (2011, Chapter 7) provides detailed analysis by a senior member of Britain’s financial regulatory agency of how to improve the resolution process. 5.5.8.3 Broader Regulatory Oversight The Group of Thirty (2009, Recommendation 4) calls for “managers of private pools of capital that employ substantial borrowed funds” (i.e., hedge funds and private equity funds) above some minimum size to register with and provide periodic reports to banking regulators. These reports should include information on “size, investment style, borrowing, and performance of the funds under management.” The regulators should also “have authority to establish appropriate standards for capital, liquidity, and risk management” for those funds “above a size judged to be potentially systemically significant.” This is recognized as being a clear break from the prevailing approach to fund regulation, which has primarily focused on regulation of lenders to hedge funds, an approach that has been justified by the fact that hedge funds do not employ any sort of government guarantee, whereas their creditors do. But the Group of Thirty notes that “the increased emphasis on financial stability in the mandates” of regulators “points to the need for greater, more systemic access to information crucial to understanding the growing risk imbalances in the system.” Strong academic arguments supporting the approach of this recommendation have been supplied by Lo (2008). 132 5.6 FINANCIAL RISK MANAGEMENT BROADER LESSONS FROM THE CRISIS When as widely respected a figure in the financial markets as Paul Volcker is moved to say that the single most important contribution of the financial industry in the past 25 years was the automatic teller machine, which at least had proven useful, there is something wrong with the industry that needs to be addressed at levels beyond risk managers and government regulators. (Volcker made this remark addressing an audience of senior finance industry figures in December 2009. It was widely reported—for example, in a Daily Telegraph article by Louise Armitstead on December 9.) I don’t doubt that comments like Volcker’s overstate the case—many of the innovations in markets, derivatives, and securitization of the past 25 years have genuinely made easier financing, broader investment opportunities, and valuable risk management tools available to firms and people who were worthy recipients; good narratives of these advances, from different perspectives, can be found in Shiller (2003) and Brown (2012). But clearly many reasonable people are starting to feel there is an imbalance— too many innovations that just provide tax and accounting gimmicks or introduce unnecessary complications relative to too few innovations addressing real economic issues. Some suggestive ideas for new directions are: ■ ■ The prominent economist Robert Shiller has been focusing on the question of identifying financial innovations that will more closely match genuine social needs (in his words, address “risks that really matter”). Shiller (2008) gives a brief account of these ideas in the context of the 2007–2008 crisis; Shiller (2003) provides a more thorough explication. Some of the innovations he advocates would be ways of hedging the cost of housing, providing home equity insurance, being able to insure against the economic risk of career choice, and hedging against the economic performance of a country. Richard Bookstaber, an experienced risk manager, in Bookstaber (2007) advocates a redesign of financial products in the direction of greater simplicity and greater tolerance for survival of disruptions. CHAPTER 6 Managing Financial Risk he management of financial risk can be divided into two parts: risk measurement and risk control. In general, the industry agrees more on how risk should be measured than on how it should be controlled. T 6.1 RISK MEASUREMENT 6.1.1 General Principles As stated in Chapter 1, the key characteristic that distinguishes financial risk management from other types of risk management is that financial risk management can take advantage of liquid markets as part of a risk management strategy. In this chapter we examine the structure of financial risk management in more detail, and a good starting point is to consider the hypothetical case in which a market is so liquid that any position can be liquidated instantaneously. While this is obviously an extreme that does not exist in reality, it will still provide an instructive background against which to consider more realistic cases. With such perfect liquidity, risk management could, in principle, just consist of setting loss limits for each trader and each trading group (the industry jargon for this is a stop‐loss limit). As soon as a trader reached the limit for a position, the entire position could be liquidated with no further loss. Or if management decided that its risk tolerance had changed because of changes in their view of the economy or the institutional environment, positions could be liquidated with no further losses. Even in such an extreme case, the following rules would be needed. ■ Careful and continuous tracking of market prices of existing positions. Otherwise, you would not know when a trader was through a stop‐loss limit. Traders may be tempted to hide the size of their losses, knowing that being through a limit will cause the position to be closed out and 133 134 ■ ■ FINANCIAL RISK MANAGEMENT eliminate their chance of making future gains. An optimistic mark of the position could delay the recognition of losses. And traders who know they are through a limit when management does not are very dangerous, since they will be tempted to swing for the fences (as discussed in Section 2.1). So, no matter how liquid the market, correct and independent valuation of current positions is at the heart of all good risk management. Sensible choices of limit size relative to trader expertise and trading strategy. A good example would be a position being taken that will benefit from a policy change, such as the lifting of governmental foreign exchange (FX) controls. Such positions often have predictable daily losses for as long as the current policy remains in place but have a large profit potential if the policy is changed. If management is convinced that this is a sensible gamble, or has sufficient trust in a given trader’s judgment to allow her to make that decision, it would be self‐ defeating to implement it with a very small stop‐loss limit that would, with high probability, cause the position to be closed out before the policy change occurs. Positions that require patience to make money should be undertaken only if the firm has the risk appetite to allow for that patience. Good procedures for review of request to exceed limits. When a trader reaches (or is approaching) a stop‐loss limit, there is an excellent chance that he will want to make a case to his management for a temporary expansion of the limit. He may believe that a market shift in his favor is “just around the corner.” A strict and firm policy to close out all positions that reach a stop‐loss limit with no possibility for review would be foolish—the trader may have excellent information and research to back up his belief and automatic closing of the position would mean passing on a profit opportunity without the ability to review the limit in the light of the latest information. I’ve rarely seen trading managers make this type of error, but an equally serious error in the other direction is unfortunately more common. Requests for temporary stop‐loss limit increases by a trader reaching or approaching the limit may get approved without serious thought by a busy manager. They may be treated as bureaucratic box‐checking exercises, particularly when the request comes from a respected trader with a good track record, rather than a genuine decision point. But this renders the stop‐loss limit useless, as it will never actually be enforced (of course, there may be some limit beyond which even the most blasé manager will stop approving increases, but then it would be better to acknowledge this as the true limit in advance, since this will lead to better recognition of the actual maximum losses the firm faces). Managing Financial Risk ■ ■ 135 A genuinely productive stop‐loss limit review requires thorough discussion between traders and their managers of the factors that have led to the existing loss and the latest information on prospects for the position. Sometimes even an experienced trader with a great track record needs time away from the market to consider whether new factors have come into play that require a change in approach. Considerations of moral hazard, as discussed in Section 2.1, will certainly influence the discussion. Traders own more of the upside than the downside of their positions and so have an incentive to argue for raising the limit, and they can take advantage of their intimate knowledge of the market to cherry‐pick data and arguments with which to make a persuasive case. Managers need to be aware of this informational asymmetry and employ a reasonable degree of skepticism while drawing on their experience of similar past situations and their outcomes. It also helps if the manager has been getting regular independent analyses of the causes of large gains and losses in the trading positions. This brings us to our next point. Analysis of reasons for large losses and large gains to put the manager in a good position to understand the logic of the trading strategy and to be able to review extension requests intelligently. In Sections 3.1, 3.2, and 4.1.6, we have already discussed the advantages for control of fraud and reporting errors of having control personnel develop thorough explanations of large moves in profit and loss (P&L), whether gains or losses. Here I want to emphasize how a robust P&L explanation process can also serve as excellent input for a manager who may need to review requests for stop‐loss limit extensions. Since decision making on stop‐loss limit extension requests must often be done under tight time limits in a stressful market environment, time that can be devoted beforehand to giving management deeper insight about the drivers of P&L in a trading book can have significant return on invested effort. Financing plans. Even when trading losses are well within stop‐loss limits, management still needs to be concerned that it has adequate financing for the cash needs of maintaining positions, as the Metallgesellschaft case illustrates (see Section 4.2.2). There is thus a need to understand and forecast funding needs and plan for their financing. It is now time to drop our unrealistic assumption that positions can be liquidated instantaneously. In virtually every case, when positions need to be liquidated, there will be some lapse of time between the decision to liquidate being made and the execution of the liquidation during which market prices can move. Stop‐loss limits need to be set in light of the knowledge of such possible market moves. For example, if you want to be sure that you 136 FINANCIAL RISK MANAGEMENT don’t lose more than $100 million on a given position and you estimate that you could lose $20 million in the course of liquidation, you need to set $80 million as the trigger point for the stop‐loss limit. All five of the points just made about stop‐loss limits under conditions of instantaneous liquidation continue to apply, perhaps even more strongly, but other risk control measures will be needed as well. The points made already are still needed to make the stop‐loss limits effective, but, with less liquidity, failure to know current market prices can be even more damaging. To deal with the additional costs of liquidation, an estimate of liquidation costs will need to be available to managers, in the form of both a statistical probability analysis of likely market moves during a period of liquidation— called value at risk (VaR) analysis—and of stress scenarios to measure potential liquidation costs during periods of unusual illiquidity. The risk control requirements we have outlined here are very close to the recommendations for managing derivatives risk that were issued by the Group of Thirty (G‐30) in July 1993. These recommendations have proved very influential, not just for the management of derivatives risk, but for all trading risk. The Group of Thirty is a private, nonprofit organization that studies international economic and financial issues and is headed by 30 senior representatives of the international business, regulatory, and academic communities. The recommendations that relate most directly to the measurement of trading risk are shown in the box, with the original numbering they had in the G‐30 report. While the G‐30 requirements and the approach being outlined here were developed in conjunction with market‐making trading operations, they have much wider scope and should be used for any type of financial risk management—that is, any type of risk management that relies on liquid instruments to help manage risk. If you are planning to use liquid instruments to limit your losses, you need to estimate the likelihood that (and degree to which) the liquidity will be there when needed. So if you are managing credit exposure by having counterparties post margin, you need to make estimates of how effective that margin will be in limiting losses (see Section 14.3.3 for details). If you run a hedge fund and hedge positions with liquid instruments or you run a pension fund and are counting on the ability to liquidate positions to assure not dipping below funding requirements for future payouts (a contingent immunization strategy), you need to take possible limitations on liquidity into account (see Section 6.1.7 for details). The rest of this chapter deals with how these recommendations should be put into practice, with many references to detailed discussion in subsequent chapters. But before getting to these specifics, I want to first lay out what I think are the essential components of any risk management Managing Financial Risk 137 GROUP OF 30 RECOMMENDATIONS RELATING TO THE MEASUREMENT OF TRADING RISK Here we review select recommendations by the Group of 30 on trading risk. Recommendation 2: Marking to Market Dealers should mark their derivatives positions to market, on at least a daily basis, for risk management purposes. Recommendation 3: Market Valuation Methods Derivatives portfolios of dealers should be valued based on mid‐ market levels less specific adjustments, or on appropriate bid or offer levels. Mid‐market valuation adjustments should allow for expected future costs such as unearned credit spread, close‐out costs, investing and funding costs, and administrative costs. Recommendation 4: Identifying Revenue Sources Dealers should measure the components of revenue regularly and in sufficient detail to understand the sources of risk. Recommendation 5: Measuring Market Risk Dealers should use a consistent measure to calculate daily the market risk of their derivatives positions and compare it to market risk limits. ■ Market risk is best measured as “value at risk” using probability analysis based upon a common confidence interval (e.g., two standard deviations) and time horizon (e.g., a one‐day exposure). ■ Components of market risk that should be considered across the term structure include: absolute price or rate change (delta); convexity (gamma); volatility (vega); time decay (theta); basis or correlation; and discount rate (rho). Recommendation 6: Stress Simulations Dealers should regularly perform simulations to determine how their portfolios would perform under stress conditions. Recommendation 7: Investing and Funding Forecasts Dealers should periodically forecast the cash investing and funding requirements arising from their derivative portfolios. Source: Group of Thirty, Global Derivatives and Principles (1993). 138 FINANCIAL RISK MANAGEMENT framework that will meet the needs identified earlier. I believe there are seven key principles that need to be considered: 1. Recognition of the nonnormal distribution of financial variables. It is an empirical fact that nearly every financial data series exhibits fat tails (see the Ratios worksheet of the VaR spreadsheet on the website and Exercise 7.3 based on this worksheet for illustrative examples). Part of the explanation for this is the psychology of markets—a tendency for a big move to create panic that exacerbates the size of the move. Part of the explanation is that financial variables are mostly human creations rather than natural phenomena. As Nassim Taleb says in The Black Swan, “Money in a bank account is something important, but certainly not physical. As such it can take any value without necessitating the expenditure of energy. It is just a number!” (Taleb 2010, 33). To put it another way, when the world’s tallest man walks into a room full of many people, he will change the average height of the people in the room by only a small amount. But when Bill Gates walks into a room full of many people, he will change the average income of the people in the room by a large amount. Whatever the explanation, risk managers need to recognize that financial data series are most likely fat tailed. They also need to recognize that large market moves in one financial variable often occur at the same time as large market moves in other financial variables, probably because investors will spread panic in one market to other markets. Therefore linear correlations are often very poor representations of the relationship between financial data series. Any risk management process chosen must allow for handling fat‐tailed series that have clustering of large moves. 2. The need for simulation. The need to handle fat‐tailed series and clustering of large moves, as emphasized in the previous point, virtually dictates the need for using computer simulation to generate estimates of potential liquidation costs. More detail can be found in Section 7.1, but the basic argument is that simulation handles fat tails and clustering of large moves in a simple and transparent fashion, while other statistical estimation techniques are far clumsier and more opaque in how they handle these features of financial series. Simulation consists of an initial specification of the distribution of underlying financial variables, followed by a calculation of the earnings impact of each instance of the distribution. The distribution of liquidations costs is then simply computed as the aggregation across individual cases. Since the step in which the distribution of the variables is specified is separate from the step in which earnings impact is calculated, there is Managing Financial Risk 139 complete freedom to specify the distribution in the most accurate way possible. Furthermore, simulation offers many other advantages, some of which will be elaborated on in Sections 7.1 and 7.3: ■ Many financial products, such as options, involve nonlinear returns. Simulation can handle this easily, since each path of the simulation computes the earnings impact independently from the computations on other paths and separately from the initial specification of the distribution of variables. Statistical techniques that mix together the specification of variables distribution and the calculation of earnings impact are much more vulnerable to error. ■ Simulation makes it easy to generate a rich set of statistics on the distribution of liquidation losses, by aggregation of results across paths. ■ Simulation makes it easy to attribute risk of potential liquidation losses to individual trading desks and individual positions. ■ Simulation methodology can easily handle a range of desired calculations in addition to the basic calculation of liquidation costs. For example, consider the point made earlier in this section concerning the desirability of fitting stop‐loss limits to trading strategies. In advance of setting a stop‐loss limit, a manager should get some idea of the probability that the stop‐loss limit will be activated by a particular trading strategy. This is straightforward for a simulation, since each individual case can be followed over a simulated time period, keeping track of whether a stop‐loss limit has been hit along the way. ■ Simulation methodology makes design of computations easy. Since each individual case of the simulation calculates earnings based on a single specification of the underlying variables, earnings calculations could be performed on each individual transaction by the exact same production models the firm uses for its official mark‐to‐market computations. Where this is computationally infeasible, due to a large number of individual simulation cases, approximations to production models are relatively easy to design and check against the official calculations. It is also easy to break up earnings calculations in each individual case by trading desk or product type. Since the earnings distribution is just a simple summation across individual cases, it is now easy to calculate the risk contributions of individual trading desks and product types. ■ Checking is made easy by the separation of specification of distributions and the calculation of earnings into two separate stages. Control personnel and front‐office personnel who may not be knowledgeable about probability distributions can focus on checking the earnings calculations, using the firm’s mark‐to‐market models for each 140 FINANCIAL RISK MANAGEMENT individual transaction, as discussed in the preceding bullet point. By parallel reasoning, the specification of probability distributions can be easily checked by economists and statisticians who may not be knowledgeable about earnings calculations. All of the advantages of simulation apply not just to the value‐at‐ risk computations for relatively liquid positions discussed in Chapter 7, but also to the modeling of relatively nonliquid positions discussed in Chapter 8. This point will be elaborated in Section 8.4.2. 3. The need to consider subjective probabilities as well as objective frequencies. As was discussed in Section 1.2, assessments of risk cannot afford to rely solely on historical frequencies. Subjective assessments of probabilities by the risk managers must be allowed to play a role. Even in computing historical frequencies, the risk managers must rely on some degree of subjective judgment regarding the length of historical period to use and the weight that should be placed on more recent historical experience relative to a longer period of history. These issues will be discussed in more detail in Sections 7.1.1 and 7.2.1. The need to utilize subjective judgment causes concern for many risk managers. Without anything objective such as a historical data set to point to, how can they count on their recommendations carrying conviction? Why will they be accepted as having expert opinions on a subjective matter? These are questions that must be confronted—when subjective judgment is required, it is best to be frank about it. The only way for a risk manager’s subjective judgments to be accepted is to have well‐researched and well‐reasoned arguments backing them up. For a good example of what such an argument looks like, see Section 5.2.5.7 for the articles in which to find the arguments presented by the Economist magazine in 2004 through 2006 to support a belief that there was a good case to be made for a large drop in real estate prices. It is very important in presenting such an argument to explain carefully that a belief that there is a significant probability that an event will occur is very different from, and requires a very different type of evidence from, a belief that an event is highly probable to occur or is the most likely outcome. Particularly when it comes to subjective judgments, ultimate decisions will rest with management. It is extremely rare for risk managers to carry enough political clout to be able to force acceptance of their subjective views. But it is surely not acceptable for risk managers to just state their views, have management disagree, and then shrug their shoulders and say nothing more. Risk managers must make their arguments forcefully and, if they believe that management is being unreasonable in its judgments, consider options such as taking their concerns to the risk committee of the board of directors or to regulators. A firm that does Managing Financial Risk 141 not allow a senior risk manager freedom to do this (on occasional significant points of disagreement) without damaging her career prospects is not a healthy working environment. And, when large disagreements occur more than occasionally or where the freedom to appeal is not part of the culture, risk managers must consider “voting with their feet,” to protect their reputations and integrity. I will relay one anecdote with respect to voting with your feet, though it preceded the days of dedicated risk management departments. An economist with whom I had worked closely at Chase in connection with the introduction of options products had left to take a good offer at a smaller firm. He later relayed his experience in joining that firm: When he asked for some orientation on how they measured their options risk, management responded by saying they had no need for such measures and could not be persuaded that such measures were needed—they just made “holistic” judgments about the positions they wanted to take. My former colleague reported thinking to himself, “I see—have a hunch, bet a bunch,” and immediately decided to start seeking other employment. 4. The distinction between diversifiable and nondiversifiable risk. The difference between systematic, diversifiable risks and idiosyncratic, nondiversifiable risks is one of the cornerstones of modern finance theory, as developed by Harry Markowitz, William Sharpe, Stephen Ross, and others. Discussion of this critical distinction can be found—in the context of expositions of portfolio theory, the capital asset pricing model (CAPM), and arbitrage pricing theory (APT)—in any textbook on investment theory (see, for example, Bodie, Kane, and Marcus 2009, Chapters 8 and 9); on corporate finance (see, for example, Brealey, Myers, and Allen 2011, Chapter 8); or on asset pricing theory (see, for example, Cochrane 2001, Section 1.4). A diversifiable risk position can be reduced in several ways, by direct hedging but also by diversification through investing in other positions that have low correlation with it. A nondiversifiable risk position, such as exposure to the Standard & Poor’s S&P 500 index or to interest rate levels, needs to rely almost entirely on direct hedging to reduce risk. Finance theory emphasizes the resulting market demand for high returns on nondiversifiable risks. Risk managers need to ensure that trading management is especially aware of sizable exposures to nondiversifiable risk, both because it may be more difficult to reduce such positions and because management will want to ensure it is receiving adequate returns for taking on these risks. This was a particularly important issue in the 2007–2008 financial crisis (see Sections 13.4.4, 5.2.4, and 5.2.5.3). 142 FINANCIAL RISK MANAGEMENT Diversifiable risk can be eliminated through hedging; nondiversifiable risk cannot be eliminated but can only be transferred to someone else. Risk managers need to be very sure they understand this risk transfer process, to make certain that the risk is truly being transferred and not just reappearing elsewhere on the firm’s books. We will discuss this further in Section 14.3.4 on wrong‐way counterparty risk. 5. The use of arbitrage theory to decompose risks. Suppose that you have some exposure to euro interest rates through interest rate derivatives and also some exposure to euro rates through forward U.S. dollar‐euro foreign exchange contracts. If these two positions were treated as completely different types of exposure, you might miss offsetting exposures to euro interest rates or, even worse, fail to measure a dangerous risk concentration by not adding together the euro interest rate exposures in the two positions. (This is not simply a hypothetical example—this treatment of interest rate exposures from interest rate swaps and foreign exchange forwards as separate exposures was often encountered in the 1980s in the risk management of major institutions.) Arbitrage theory for derivatives, which has been well developed over the past 40 years, as exemplified in the material in Hull (2012) and similar textbooks, has provided a valuable tool kit for unifying such positions. I will be drawing heavily on arbitrage theory for unifying positions throughout Chapters 10 through 14. 6. The need to consider periods of reduced liquidity. All of us who participate in financial markets have experienced several periods of severely reduced liquidity when the ability to trade at anything other than fire sale prices dries up for a prolonged period of time. Estimation of potential liquidation costs associated with stop‐loss limits must account for the possibility that liquidation will be required during such a stressful period. In fact, it is often during such periods that stop‐loss limits are breached, since lowered liquidity is often associated with sharp price swings and managements may need to cut risk limits in a crisis period. I argue in Section 7.2.1 that detailed scenarios based on subjective judgment must play the key role in this analysis but that there is still room for using simulation based on historical data as a supplement. 7. The need to distinguish degrees of illiquidity with different tools to handle each type. Given that projecting possible liquidation costs of a position are such an important part of risk management, it is natural that different tools are required based on the degree of liquidity of a position. In teaching classes on this topic, I like to use a variation of a quotation from Shakespeare, who said, “Some men are born great, some achieve greatness, and some have greatness thrust upon them” (Twelfth Night, Act 2, Managing Financial Risk 143 Scene 5). My variant is “Some positions are born illiquid, some achieve illiquidity, and some have illiquidity thrust upon them,” and each of these three types of positions requires a different type of risk management. Positions that have illiquidity thrust upon them are positions in instruments for which frequent liquid market quotations are available, and they are not of such large size that liquidation of the position will significantly impact market price. These positions will become illiquid only under conditions of extreme and unusual market stress. These are the type of positions well handled by normal mark‐to‐market pricing (Section 6.1.3) and VaR calculations (Section 7.1), supplemented by stress tests (Section 7.2) to consider the possible impact of an unusual market stress that causes a normally liquid position to become illiquid over a period of a few weeks during a large market move. Those positions that achieve illiquidity are also positions in instruments for which frequent liquid market quotations are available, but where position size has grown to the point that liquidation will significantly impact the price. In such cases, the risk tools just referred to must be supplemented by a way of measuring the impact of this larger position size. Note that an illiquid position size should impact both VaR calculations (since liquidation even in normal market conditions will come with added cost) and stress‐test calculations (since once a period of unusual market stress has been weathered and more normal liquidity has returned to the market there will still be added costs to liquidating a large position). My suggested approach for handling large positions is a separate and supplementary simulation of the distribution of possible costs to be incurred, explained in Section 6.1.4. Positions that are born illiquid are positions in instruments that lack liquidity even during the best of market conditions. They are the positions referred to in Section 1.1 as having actuarial risk. This could be a transaction that completely lacks a market component; we’ll discuss an example at the beginning of the next section. Or it could be an instrument with very limited liquidity; a good example is a position in a one‐way market, as discussed in Section 6.1.3. It could be a position that is so large relative to the size of daily trading that it cannot be liquidated even over an extended period; an example would be the loan books of most banks, as discussed in the introduction to Chapter 13. It could be a position that can only be sold under certain conditions, such as restricted stock. It could be an instrument that is so complex that liquidation in any reasonable time period is unlikely. (One of the first consulting assignments I had in the wake of the 2007–2008 crisis was for a large bank looking to find a methodology for valuation of collateralized debt obligation (CDO) tranches it was holding. When I asked whether the valuation was just for accounting purposes or was meant to drive decision making on possible sales, the somewhat testy 144 FINANCIAL RISK MANAGEMENT response was: “I couldn’t possibly sell any of these tranches—it would take me six months just to explain all the cash flows to a potential buyer.” Even allowing for hyperbole born of frustration, there is likely to enough truth in this comment to serve as a warning about assuming liquidity on highly complex positions.) It is the management of positions in instruments that lack liquidity that presents one of the greatest challenges to financial risk management, as has been confirmed by the 2007–2008 crisis that was largely due to inadequate risk management of illiquid CDO tranches (see the discussion in Section 5.2.5). I will therefore address a recommended approach to this issue in a separate section. 6.1.2 Risk Management of Instruments That Lack Liquidity When a market component is completely lacking, financial risk management techniques may be wholly inappropriate and it may be proper to manage risk utilizing the type of actuarial techniques we discussed in Section 1.1. A good example of how to identify instruments for which a market component is completely lacking and how to manage this type of risk can be found in the excellent discussion of weather derivative options in Jewson, Brix, and Ziehmann (2005, Section 1.4). They declare that “for locations where the [weather] swap is not traded, and which are not highly correlated with locations on which swaps are traded, actuarial valuation of the options is the only choice.” They specify that actuarial valuation is “fundamental analysis” of the type used in pricing insurance contracts, based on “historical meteorological data and meteorological forecasts to predict the distribution of possible outcomes.” When weather swaps are traded for the location to which the option is tied, or on locations whose weather is highly correlated with this location, then they advocate valuation based on market prices and arbitrage pricing models (e.g., Black‐Scholes). This is a good illustration of the approach I support. Positions that have no liquidly traded instruments that can meaningfully be used as hedges should be evaluated and managed just as if they were positions of a conventional insurance company. The general principles of risk management referred to in Chapter 1 apply; the financial risk management principles that are the subject of this book are irrelevant. But when liquidly traded instruments can be used to hedge a meaningful portion of the risk, then we can utilize financial risk management. For this latter case, the approach I strongly favor is to (1) set up a liquid proxy that allows the total risk to be split into liquid risk and illiquid risk; (2) use the liquid proxy in all standard risk reports and limits (e.g., position reports, VaR, stress tests); and (3) use a separate simulation to manage the Managing Financial Risk 145 risk of the mismatch. As an illustration, consider the example discussed in Section 10.2.2, a 40‐year interest rate swap in a market that has interest rate swap liquidity out to only 30 years. A 30‐year swap would be assigned as the liquid proxy and used in all standard risk reports, while a separate simulation would be used to assess the risk of using a 30‐year swap as a hedge against a 40‐year swap. The reasons I favor the use of a liquid proxy to represent positions in illiquid instruments are: ■ ■ Some of the risk in an illiquid instrument can be managed by liquid instruments, and the use of the liquid proxy ensures that this possibility can be exploited. To continue with our example of the 40‐year swap, the booking of 40‐year swaps certainly exposes the firm to interest rate risk over the first 30 years of the swap in addition to the exposure for years 31 through 40. Use of the liquid proxy assures that this exposure to the first 30 years shows up in position reports and limit calculations properly added in to liquid instrument exposures taken in the same direction. This will alert management to concentrated exposures and give traders and marketers proper incentives for hedging that portion of the risk that can be hedged through liquid instruments. If 40‐year swaps were treated as a completely separate category from swaps of 30 years or less, these goals might still be accomplished, but it would require the building of a completely separate reporting structure and there is always the possibility of gaps occurring in the design of new reporting structures. The simple act of insisting on a liquid proxy takes advantage of all existing reporting structures, such as mark‐to‐market, VaR, stress tests, and position limits (e.g., maturity bucket limits), with no further effort beyond the calculations currently in place for these reports. Basing reserves and limits for the illiquid risk just on the variability of the mismatch between the illiquid position and its liquid proxy should often lead to reducing the need for reserves and limits. Continuing with our example, managing the risk on the difference between a 40‐year swap and a 30‐year liquid proxy over the 10‐year period that you must wait until the 40‐year swap becomes liquid (after 10 years it has only 30 years left to maturity) will almost certainly be computed as significantly lower than the variability of return on an unhedged 40‐year swap. But this latter computation would be an overstatement of the risk of the trade, since use of a hedge involving the liquid proxy is always a choice the trading desk can make. Computation using the liquid proxy does not in any way require trading desks to make use of this actual hedge—but, if they don’t, the additional risk will show up as a use of their trading limits for VaR, stress tests, and positions. 146 FINANCIAL RISK MANAGEMENT ■ Less compelling, but still of some weight, is that making sure that even illiquid positions make an appearance in standard reports, such as VaR and stress tests, makes it less likely that managers will forget about these risk positions. It serves as a reminder. But true measurement of the risk of an illiquid position cannot be accomplished solely through standard reports designed for liquid positions. There must also be a separate and well‐thought‐through report on the potential cost of the mismatch between the illiquid position and the liquid proxy. This will be our next point of discussion. Modeling the potential impact of the difference between the actual trade and the liquid proxy should use simulation for the similar reasons as given for using simulations in Section 6.1.1—to reflect a full range of possible outcomes and to generate a statistical distribution that can be used in assessing issues such as capital adequacy. But simulations of the differences between the actual trade and its liquid proxy cannot just be for the short periods used in VaR calculations; they must go all the way to final payout or to when the trade becomes liquid. Simulations must reflect the possibility that the model used for pricing and trading the product may be wrong. All of these issues in simulation of the difference between the actual trade and the liquid proxy will be discussed in detail in Section 8.4. More controversially, I do not believe in using mark‐to‐market pricing on the difference between the actual illiquid trade and the liquid proxy. Reserve levels should be adequate to protect against extreme events, and it is extremely rare that short‐term market changes reveal new information about the potential depth of an extreme event. Mark‐to‐market pricing is designed to measure prices at which risk can be exited in the near future. Since an illiquid trade cannot be exited in the near future, mark‐to‐market pricing is not truly reflecting changes that impact the position. I know that to many people this will seem as if I am trying to go easy on illiquid positions, which would be particularly foolish in light of all the havoc overindulgence in illiquid products by financial firms caused in the recent crisis. But I do not believe this proposal is moving in the direction of easier treatment of illiquid trades. The size of reserves I want to keep is quite substantial, and my experience with this reserve methodology leads me to believe it would lead to less use and more cautious use of illiquid products than the wholly inadequate reserving processes that appear to have been operating in the run‐up to the recent crisis. I will ask readers to withhold judgment until they can see my argument in detail in Section 8.4.4. I discuss how my proposal might have mitigated the spread of the crisis in the discussion of reducing procyclicality in Section 5.5.8.1. Managing Financial Risk 147 Given the very different treatment I am advocating for illiquid instruments relative to liquid instruments, it is vital to have good tests available to distinguish between illiquid and liquid instruments. When trading desks make suspect claims of liquidity, independent risk managers need to insist on evidence from reliable external sources or from a history of actual trading tickets. Trading history that is overwhelmingly in one direction, extremely sporadic, or concentrated with just one or two counterparties, or that has been executed at prices substantially different from internal valuations needs to be regarded with extreme wariness. Trading desks that want to overcome such suspicion should be prepared to demonstrate the ability to liquidate significant blocks of inventory at prices close to internal valuations. 6.1.3 Market Valuation The policy of marking to market all trading positions, at least as often as the close of business each day, as per the G‐30’s Recommendation 2, constitutes the essential foundation for measuring trading risk because of three primary reasons. First, without a nearly continuous marking to market, it would be possible that ineffective hedging strategies would not be recognized until long after being put in place. Second, the analysis of revenue will yield insight only if the revenue figures being analyzed are tied to genuine changes in value. Third, in measuring the risk exposure to market moves, it is far easier to make good judgments about possible short‐term moves than it is about longer‐term moves. But if trades are not revalued frequently, it becomes necessary to measure risk exposure over longer periods. When highly liquid external prices are available for marking a position to market, then the issues involved in performing the mark are largely operational. An example might be a position in spot foreign exchange (FX) for the dollar versus Japanese yen. This is a market for which quotations are readily available on trading screens, with market conventions that ensure that firms posting prices are prepared to actually deal in reasonable size at these prices. Quotations for mark‐to‐market purposes can be captured electronically from trading screens or entered by hand and later checked against printouts from screens—the choice should be based on the operational cost versus error rate and the cost of correcting errors. Another example would be a position in a well‐traded stock or exchange‐traded futures option for which the last price at which an actual trade occurred is readily available from an exchange ticker. For many positions, mark‐to‐market pricing is not this straightforward. Either the market itself does not have this type of liquid quote available or the size of the position held is so large that closing it out 148 FINANCIAL RISK MANAGEMENT might impact the market. The price at which the position can be exited will be uncertain to some degree. In such cases, two interrelated questions must be asked: 1. How should a most likely exit price be arrived at? 2. Should some markdown of the price be used to account for the uncertainty and, if so, how should the amount of reserve be determined? Establishing the most likely exit price may require a model to create a mark based on more readily available prices of other instruments. Models can range from very simple computations, such as the interpolation of an illiquid two‐and‐a‐half‐year bond from prices on more liquid two‐ and three‐year bonds, to complex theoretical constructions. A discussion of how to use models in the marking process and how to establish reserves against the associated uncertainty can be found in Chapter 8. What if price quotes are available, but are not sufficiently liquid for a readily agreed‐upon external valuation? This implies that deriving the most likely exit price from these quotes requires an understanding of the relative quality of available quotes. For each quote, questions like the following need to be answered: Is the quote one at which the firm or broker providing the quote is offering to do business, or is the quote just provided as a service to indicate where the market is believed to be today? If the quote is an offer to do business, how large a transaction is it good for? What is the track record of the quotation provider in supplying reliable information? Are there possible motivations to provide misleading information in an attempt to influence pricing to move in a direction that favors a quote provider’s position? How frequently are quotes updated? With such a multiplicity of information bearing on the issue, there is no doubt that traders of an instrument have the best judgment on determining this valuation. Their continuous contact with other firms’ traders and brokers enables them to build the experience to make these judgments. The ability to make such judgments is a major factor in determining a trader’s success, so traders who have built a successful earnings track record can make a strong claim of having the expertise to determine most likely exit prices. Unfortunately, reliance on traders’ judgment raises moral hazard concerns. As discussed in Section 2.1, traders are often tempted to mislead management about position exit prices in order to inflate reported profits or to increase flexibility in the positions they are allowed to hold. Outsiders, from corporate risk management, corporate finance, or the middle office, need to be involved in making these judgments to preserve independence. However, Managing Financial Risk 149 designing mechanisms for resolving disputes between traders and control personnel raises many difficult issues: ■ ■ ■ ■ How can control personnel obtain a sufficient knowledge base to challenge traders’ judgments? At a minimum, traders should be required to make public the information on which judgments are made. This can be accomplished by insisting that quotes be sent to the firm in writing (whether through trading screens, e‐mail, or fax). Alternatively, control personnel should have the right to selectively listen in on phone conversations in which quotes are made. Ideally, control personnel should have a range of experience that enables them to arrive at independent conclusions regarding quotations, perhaps even prior trading experience. Records should be kept of prior experience with the reliability of a particular trader’s valuations by tracking the path of internal marks leading up to an actual purchase or sale price and noting suspicious patterns. Control personnel should adjust their deference to trader valuations by the degree of proven reliability. A trader’s ultimate weapon for bringing credibility to a valuation is to actually exit part of the position. A recorded price narrows disputes down to the single question of whether the size of the trade relative to the retained position is large enough to be a reliable indicator of the exit price for the remainder. Despite best efforts to design dispute resolution processes that balance power between traders and control personnel, traders inevitably retain a strong advantage based on information asymmetry. They can utilize their knowledge of a wide variety of sources of price quotes to selectively present only those that are the most advantageous to their case. They sometimes use friendships and exchanges of favors to influence other market participants to provide quotes biased toward their valuations. Traders also often rely on an aggressive personal style and internal political power based on their profitability to prevail through intimidation. To remedy this power asymmetry, some firms prefer to rely on more objective computations for determining valuations, even where this reduces accuracy by lessening the role of judgment. A typical approach would be to average the quotes obtained from a set panel of other firms or brokers, perhaps discarding outliers before averaging (discarding outliers is a possible protection against a few quotes that have been biased by friendship or favor). Changes in panel membership should be difficult to make and require agreement between traders and control personnel. 150 FINANCIAL RISK MANAGEMENT A promising development toward more objective valuations for less liquid instruments is the Totem Market Valuations service. This service is designed to share information among firms making markets in less liquid products. Firms can obtain access to quotes on only those products for which they are willing to provide quotes. Their access to quotes can be cut off if the quotes they provide are frequently outliers, indicating either a lack of expertise or an attempt to bias quotations. Although the extensive machinery of this process means it can make quotes available only once a month and with a lag of a few days, it still provides a valuable check on the valuations of a firm’s traders. The following are some pitfalls to be wary of when setting up a procedure for deriving valuations from less than fully liquid market quotes: ■ ■ Model‐derived quotes. Here is an illustration of a frequently encountered problem. You need a valuation for a particular bond and you have a choice: either use a model to compute the value based on observed prices of more liquid bonds with similar maturities and credit ratings or use price quotes for the particular bond obtained from brokers. Before choosing the latter, ask the following question: Are the brokers providing a quote specific to this bond or are they just providing the output of their own model based on prices of more liquid bonds? If your external source is model based, might you be better off using your own model? The following are some advantages to using a model‐based external quote: ■ You may be able to get model‐based quotes from several sources with the hope that errors will average out. ■ The external models are being tested by the use of the quotes by many different firms, so it is more likely that objections will be raised if the model is missing something. ■ It is less likely that traders will influence the outcome when an external source is being used. ■ The quotes may become so widely used as to be a good indicator of where the market is trading. ■ The primary disadvantage to using a model‐based external quote is that you may not be able to obtain details of the model used, so it is harder to estimate potential error and build adequate reserves for uncertainty than when using your own model. Revealing positions. When quotes are not available on regularly displayed screens or reports, firms seeking quotes may need to make specific inquiries to obtain quotes. Their inquiries reveal information about the positions the firm holds that can be used to the firm’s disadvantage by other market participants. This is particularly true if the Managing Financial Risk ■ 151 conventions of the market require an indication of either buy or sell interest to obtain a quotation, as opposed to obtaining a bid‐ask quote. Even when you do not need to reveal the direction of your interest, in some markets the direction of a firm’s position is well known to other participants and the expression of interest in a particular instrument is highly revealing of holdings. It is always possible to disguise positions by requesting quotes for a range of instruments, including instruments held and not held. However, the quality of the response may suffer as efforts to provide quotes get diffused over too many instruments. Market conventions concerning the tolerated ratio of inquiries to actual transactions also limit the amount of information that can be obtained. If trader reluctance to reveal positions limits the extent of the external quotes obtained, models may need to be relied on more heavily to infer valuations. One‐way markets. You not only need to worry whether the size of transaction for which an obtained quote is valid, but you must also worry about whether the quote is valid for your firm. Markets that tend to be one‐way, with customer demand strongly on one side and market maker supply on the other, may lead to quotations that are good for customers only. A typical example would be an options market in which almost all customer interest in options beyond five years was to sell options, not buy them. A market maker, in such circumstances, might supply reasonably liquid quotes for the purchase of long‐term options to customers, but be unwilling to buy on the same terms from other makers. The principle is to reserve the limited capacity to take on risk to encourage customer relationships, not to help competitors for this customer business by allowing them to distribute some of their risk. This is not to say that market makers will never buy longer‐term options from one another in such circumstances, but they may do so only on a negotiated basis, with no actionable quotes available, even through brokers. A market maker may still succeed in finding out the prices that other market makers are paying customers for longer‐term options, since customers often let them know what bids they are seeing from other firms. It will be a definite source of comfort to know that the firm’s prices are in the same range as their competitors’ prices, since this is an indication that the firm’s models and trading strategies are not suffering from some major error, such as overlooking a source of risk. Equivalently, a firm derives comfort from seeing that it wins its fair share of deals in a given category, neither too many nor too few. Although this comfort is genuine, it should not be confused with obtaining a price at which the firm can exit its risk positions. In the absence of 152 FINANCIAL RISK MANAGEMENT quoted prices at which the firm itself can transact, it is prudent to anticipate the need to hold risk longer and to utilize models to estimate longer‐term profit and loss (P&L) and reserves, and limits to control the associated risk, as discussed in Section 8.4. 6.1.4 Valuation Reserves When there is substantial doubt about the price at which a position can be exited, a safety margin can be provided by calculating a valuation reserve that can be subtracted from the most likely exit price. The issue of how large reserves should be for valuation uncertainty is probably the single issue that leads to the greatest conflicts between traders and corporate risk managers. Based on their experience and knowledge of the motivations of the creators of market quotes, traders tend to believe they know the price at which positions can be exited with a fair degree of certainty. With some justice, they will point out that the uncertainty is mostly on the part of the outsiders, such as corporate risk managers and the corporate finance function, who do not have the traders’ access to information. Reserves lower the reported P&L, which is the ultimate scorecard for the traders, determining bonuses, promotions, the size of positions management will allow, and, ultimately, continued employment. Understandably, traders will push to minimize reserves. (The one universal exception to this tendency is a trader who inherits a book from another trader. Invariably, the new trader will want to increase reserves for the inherited positions. I call this the principle that no profit should fund only a single bonus.) Occasionally, though, one encounters a trader who claims to be a proponent of large reserves. I came across one when a trading book of exotic options was being established for which I was to be responsible for the risk management. The head trader expounded on his philosophy of avoiding any appearance of claiming too much P&L before achieving certainty of the results. He wanted reserve levels to be generously high. Here, I thought, was someone I could get along with well. And so I did, through many months in which both P&L and reserve levels were high, with easy agreement between the two of us on the reserves. Then came the unfortunate day when an operations error in booking a trade was discovered several months after the trade had been booked. Rebooking the trade correctly would lead to a large loss, large enough that the trading desk would show its first negative P&L for a month. The head trader, although duly upset by the operations failure, was unfazed by the P&L consequence. Now, he informed me, was the time to release some of that reserve that had been accumulating—just enough to make P&L for the month come out positive. I protested. First of all, the reserves had been Managing Financial Risk 153 created for valuation uncertainty, not as a hedge against possible operating errors. Second, the amount of uncertainty in the valuation was exactly the same on the day after the error was discovered as it was on the day before it was discovered. So how could a lowering of reserves be justified? The era of good feelings had come to an abrupt end. This experience illustrates why a great deal of suspicion exists around valuation reserves, which is often expressed by regulators, such as the Securities and Exchange Commission (SEC) and auditors. Aren’t reserves just a cushion to allow reported earnings of a trading book to be smoothed, creating an illusion of less uncertainty of return than actually exists? To avoid this, a definite principle must be in place that reserves are strictly for uncertainty concerning current valuation and never for uncertainty concerning future market variation. As an example, take a position in a liquid instrument, such as the dollar versus Japanese yen spot FX we previously cited. The future movements of this highly volatile exchange rate (and hence the P&L) may be surrounded by great uncertainty, but there should be no reserve, since the position can be exited at short notice at a known price. A reserve should be considered only if the position reaches a size that places a limit on this freedom of exit and therefore calls into question the valuations of the current position. To make sure that this principle of using reserves only for the uncertainty of current valuation and not for the uncertainty of future market variation is followed, clear independence of reserve determination from the control of insiders with a motivation to show smoothed earnings must be demonstrated. This requires the final decision authority to be with an independent business unit, such as corporate risk management or corporate finance, and relatively objective standards for determining reserves to be utilized. The uncertainty of current valuation could be due to the illiquidity of available price quotes or it could be due to reliance on a model to obtain a valuation. Section 8.4 discusses how to establish objective standards for reserves against model uncertainty. We now focus on how to establish objective standards for reserves against positions for which only illiquid price quotations are available. The most direct method for reserving against an illiquid position is to estimate the degree to which exiting this position in the market might cause prices to move. This can accommodate fairly objective standards by using haircut tables on valuation. These have set percentage discounts tied to the size of the position held relative to some measure of the market size, such as the average amount of daily trades. This method takes proper account of both types of possible illiquidity, since this ratio could be high based on either a small denominator (indicating an illiquid market) or a large 154 FINANCIAL RISK MANAGEMENT numerator (indicating a big position in a liquid market). The downside to this method is that it may be difficult to establish reasonable haircut percentages to use. Rarely do firms keep good historical records of the impact of exiting large positions, and it will, in any case, be very difficult to sort out such impacts from other effects on market prices. This leaves the determination of haircut percentages to a subjective debate in which the traders’ greater experience will be difficult for outsiders to question. A method that lends itself to a more evenly matched debate is to first estimate the amount of time it will take to exit a position without substantially moving prices and then reserve against a possible market move over this time period. This exit time estimate will also be based on a ratio of size of position held to daily trading volume. It thus shares the previous method’s advantages of taking proper account of both types of possible illiquidity and also the previous method’s disadvantage of making it difficult for outsiders to debate trader judgment. However, the potential price move estimate allows for outsider objectivity, since it is very similar to the sort of calculation that goes into VaR. It also enables reserve levels to be calibrated to management‐ determined levels of uncertainty that should be reserved against. A uniform uncertainty level used for different trading desks can help to ensure the comparability of results across the firm. For example, consider a $500 million position in a stock in which the amount that can be transacted in one day without adversely impacting prices is estimated to be $50 million. So $500 million/$50 million = 10 days of price moves should be reserved against, which implies that on average there will be 10/2 = 5 days of price moves prior to sale. If the daily standard deviation of price moves is 1.5%, and if management decides on a reserve to a 95% confidence level, which is equivalent to 1.65 standard deviations of a normal distribution, then the reserve level should be: $500 million × 1.65 × 1.5% × 10 / 2 = $27.7 million (6.1) It should be reiterated that despite the appearance of a term that is tied to the uncertainty of future market variation, this remains a reserve methodology based on current valuation uncertainty. Future market variation is being reserved only to the extent it is outside management control, due to a large position size preventing exit at the desired time. A third method, which can be used as a complement to the other two, is to create a reserve against aged positions. This method establishes a formula that marks a position down by a certain percentage the longer it is held. This can only be used as a complement to one of the other two methods, since it will not establish any reserve against a large illiquid position recently entered into. Managing Financial Risk 155 Why should there be uncertainty about position valuation just because a position has been held for a long time? It is based on the observation that traders may delay exiting a position when they suspect that it will cause a decline in value from the level they are currently marking it at. Although I have heard much anecdotal evidence supporting this observation, it would be intriguing to perform a statistical study on the correlation between the length that positions are held and the size and direction of the price move between the last mark and actual sale. An aging reserve policy can also be justified on the pragmatic grounds that it is providing traders with the right incentives—to realize profits and cut losses in a reasonably short time period. As was stated at the beginning of this section, reserving against valuation uncertainty is probably the leading cause of the greatest conflicts between traders and corporate risk managers. The risk managers need to provide a degree of conservatism that will assure investors, lenders, and government regulators that P&L is not being overstated and must provide a degree of independence to allay suspicions that reserves are being used as a means for smoothing earnings results. However, this leads traders to suspect that too much conservatism is being applied to protect risk managers against any possibility of criticism. A reserve that is too conservative hurts not only the trader, but also the ultimate profitability of the firm by limiting the amount of business that can be transacted. In my experience, traders often misunderstand the need for conservatism and independence. One argument I’ve frequently encountered when specifying the reserve I think needs to be placed on an illiquid position goes something like this: “If you want the firm to value the asset at that low a value, then you would be happy if I went massively short the asset at that price.” However, this mistakes conservatism for a view on fair price—if the trader was to go short this illiquid asset, I would want reserves to establish a conservatively high value for the short position. In other words, reserves are used to establish a bid‐ask spread on an illiquid position, and the greater the illiquidity, the wider the spread. I’ve also encountered the argument from traders that they have excellent inside information as to where a position will trade, but they don’t currently want to enter into the trade at that price. I need to point out that unless they can find some means of translating inside information into something publicly verifiable, we cannot ask the firm’s shareholders and depositors to bear the risk that they are wrong. Of course, my dialogue with traders is far from a one‐way street. Often it is a case of their educating me on sources of information or aspects of hedges that cause me to change my initial view. Over time, with almost all of the traders I’ve dealt with, we’ve come to an accommodation of mutual respect, but with a realization that our interests sometimes differ. However, 156 FINANCIAL RISK MANAGEMENT I still wonder at times whether other risk managers have found better ways to avoid initial contentiousness. I was therefore a bit amused at some dialogue I overheard. I was meeting with the head of market risk at a major investment bank, one of the most respected individuals in the industry. Our conversation was interrupted by an urgent phone call from one of his staff. I heard only his side of the phone conversation, which went something like this: “Well, certainly you need to put a reserve on a trade like that. . . . I don’t care whether the trader likes it. If he doesn’t, let him sell some of the position and show us where it should be priced. . . . You can’t accept a statement like that from him. The fact that the reserve you’ve calculated would make him book an up‐front loss doesn’t prove that your reserve is stupid. Tell him that your reserve calculation shows that his price is stupid.” 6.1.5 Analysis of Revenue The G‐30 study states, in support of Recommendation 4 to identify revenue sources, that “measuring the components of profit helps participants to understand the profitability of various activities over time relative to the risk undertaken, as well as to gain insight into the performance of hedges.” A basic justification of using mark‐to‐market valuation in the management of risk is that it will lead to an early identification of ineffective hedging strategies, which can trigger experimentation with alternative hedges or changes in the mix of products being offered. This can happen only if an effective and frequent analysis is made of what is causing changes in P&L. In particular: ■ ■ ■ P&L must be segregated by product line to identify which products may be encountering hedging difficulties. P&L must be broken out into that part attributable to newly booked business versus that part attributable to hedging activity on existing business. This ensures that hedging problems will not be masked by the offset of profits from new business, leading to a Ponzi scheme, as discussed in Section 2.2. A persistent pattern of profitable new business offset by hedging losses is an indication that either traders have chosen to take positions that (at least temporarily) have had bad results or valuation reserves have been inadequate. To distinguish between these two cases, it is important to identify what portion of hedging profits is due to movements against specific risk factors, such as delta, gamma, vega, and theta. In this way, losses stemming from deliberately taken positions can be distinguished from those that arise from risks such as correlation exposure, which the trader Managing Financial Risk ■ ■ 157 cannot completely hedge. This analysis is also important in confirming that risk positions are reported correctly. If daily P&L swings cannot be accounted for by the reported size of risk positions and the daily changes in market variables, it is a warning that the reported risk measurements may be incorrect. This should lead to investigations of whether some transactions have been misrepresented in the reporting systems or whether additional or more detailed risk measures are required. Particular attention should be paid to unexplained P&L swings that take place around a date on which a payment is made or determined. If a model is not properly valuing a payment that has already been determined or is very close to determination, the probability is very high that the trade has been misrepresented. More detail on this point will be found in Section 8.2.7.1. It is extremely important to highlight any P&L changes due to changes in those assumptions that cannot be directly tested against available market prices or changes in models. This eliminates the possibility that P&L due to such changes will mask the results of ineffective hedging strategies. Significant differences between official P&L changes and the informal trading desk estimation of these changes should be investigated. These differences can be indicators of hedges that are not performing as expected. 6.1.6 Exposure to Changes in Market Prices The need for measuring exposure to market changes is emphasized in G‐30 Recommendations 5, 6, and 7. Proper daily mark‐to‐market valuation, as discussed in Section 6.1.3, is the key to properly measuring the exposure to changes in market prices. The correct daily valuation ensures that exposure is being evaluated from the correct starting point and also serves as a basis for translating changes in observable market prices into changes in portfolio valuation. Since the daily mark‐to‐market needs to relate valuation to some observable external prices, possibly through the use of models, this same relationship can be used to take a change in market price and convert it into a change in instrument value. To take a concrete example, consider an option position on the Standard & Poor’s S&P 500 index with an expiry in five months. When considering how to value it, decisions must be made about what model to use and what the inputs to the model should be. Let us say a Black‐Scholes model is chosen, requiring input for the price of the underlying and an implied volatility. For the underlying price, we might decide to use an average of one‐third of the closing three‐month S&P futures price and two‐thirds of the closing 158 FINANCIAL RISK MANAGEMENT six‐month S&P futures price. For the implied volatility, we might decide to use an average consisting of one‐third of the implied volatility of the closing three‐month S&P option price and two‐thirds of the implied volatility of the closing six‐month S&P option price. These choices will be made based on trade‐offs between basis risk and liquidity risk and could include reserve adjustments for lack of liquidity. However, once the choices are made for valuation, they become simple recipes for translating changes in market prices of the three‐month S&P futures, six‐month S&P futures, three‐month S&P implied volatility, and six‐month S&P implied volatility into a change in the five‐month option price, utilizing the Black‐Scholes model. Once these pieces have been established, the remaining task is to decide on the market price shifts on which to calculate exposure. Three primary types of shifts are used: 1. Standard shifts such as a 1 basis point interest rate move, a 1 percent stock price move, or a 1 percent implied volatility move. The advantage of standard shifts is that they easily convey a precise meaning to a wide group of users. The main issue to be decided when using standard shifts is which market prices to group together—do you want to report exposure to each individual stock price moving, all stock prices moving together, or a particular industry shifting relative to all others? These detailed decisions are best examined in the context of specific risks. We address these decisions more closely in subsequent chapters, particularly Section 7.1, Section 8.4, and Section 9.4. 2. Shifts based on the statistical analysis of the probability of the size of the change. The advantage of statistically based shifts is that they make it easier to compare the size of exposures in different risk classes. For example, it’s hard to say whether a $5 million loss for a 1 percent change in stock prices is more of a danger or less of a danger than a $2 million loss for a 1 percent change in implied interest rate volatilities. However, a $5 million loss for a stock price change that has a 5 percent probability of occurring is clearly more worrisome than a $2 million loss for an implied interest rate volatility change that has a 5 percent probability. Probability distributions also make it possible to combine shifts in unrelated asset classes into a single measure, such as the 95th percentile VaR, defined as the amount of loss that will be exceeded only 5 percent of the time, based on all of the positions within a portfolio. The difficult issue with statistically based measures of risk is how to determine the probability distributions. These measures and the means of deciding on distributions are discussed in Section 11.1. 3. Shifts based on scenarios determined by economic insight into the potential size of different shifts and the relation between them. An example Managing Financial Risk 159 would be a stress scenario for the impact of the debt default of a particular large developing economy, which might be judged to result in, say, a 5 percent decline in all stock prices, a larger decline in the stock of companies with large investment in that economy, a 10 percent decline in all emerging market FX rates, a 15 percent increase in the credit spread of all emerging market debt, and so on. We study alternative approaches to defining such shifts in Section 11.2. Scenario analysis is needed for cash flow as well as for P&L to anticipate funding liquidity problems, which is consistent with the G‐30 Recommendation 7, as discussed in Sections 3.5 and 4.2.2. 6.1.7 Risk Measurement for Position Taking It can be argued that the G‐30 recommendations should apply to the market‐making function of trading with an emphasis on keeping position holdings to a minimum, but not to the position‐taking function of trading, where positions may be held for very long time periods based on fundamental views of where market prices are headed (refer to Section 2.5 for the distinction between position taking and market making). Is it really important to measure short‐term price fluctuations in positions being held for the long term? In this context, it is interesting to note an SEC letter (December 8, 1999) that emphasized the obligation of mutual funds to value assets based on fair value, the amount an arm’s‐length buyer would currently pay for a security. The SEC letter specifically states that fair value cannot be based on “what a buyer might pay at some later time, such as when the market ultimately recognizes the security’s true value as currently perceived by the portfolio manager” or “prices that are not achievable on a current basis on the belief that the fund would not currently need to sell these securities.” These views reflect the G‐30 principles. Arguments for applying current market valuation and short‐term price exposure measures to positions being held for the long term include: ■ The desire to hold positions for the long term may reflect the motivation of fund or proprietary position managers, but they may not be the only constituency for valuation information on the fund. Fund investors, lenders to the fund, senior managers of the firm of which the proprietary position managers are a part, and regulators may all have an interest in knowing prices at which the positions may be exited in the near future. Investors may want to exit the fund. Lenders may need assurance that margin calls can be met. Senior managers could decide that they want to reduce the amount of risk‐taking authority being allocated to the position takers. Senior firm management will also want to view 160 ■ FINANCIAL RISK MANAGEMENT integrated risk reports for the entire firm, which will cover both market making and proprietary positioning functions. Regulators may be seeking assurance that fund withdrawals can take place in an orderly manner. All these points were particularly emphasized by the Long‐Term Capital Management (LTCM) experience discussed in Section 4.2.1. It is possible to find anecdotal evidence of successful fund managers and proprietary traders who do not desire any feedback from market price changes. They view themselves as investing for the long run, and they see short‐term price changes as distracting noise that does not reflect changes in fundamental values, but only short‐lived shifts caused by supply and demand imbalances. However, it is possible to counter this with anecdotal evidence of successful fund managers and proprietary traders who want to receive constant feedback from the market. Even though they are investing for the long term, they want to be constantly aware of the price at which risk positions can be unwound. They attempt to make money by having a few positions on which they are right and earn a large amount, and avoid having any positions on which they lose a large amount. The constant feedback of market prices at which positions can be exited provides both a means to ensure that a limit is placed on the amount that may be lost on any one position and a signal that markets are moving in ways they do not fully understand. In such circumstances, they seek to exit the market and wait until they can gain a better understanding before reentry. An alternative but related argument would be that fund managers do not need to be concerned with tail risk but only with the trade‐off between expected return and standard deviation of return (the Sharpe ratio), since prudent investors will utilize a fund as just one small part of their overall investments and that it is up to each investor to manage individual tail risk. This is essentially the argument we considered in Section 4.2.1 on the LTCM disaster: “Nor is there a major difference in consequences between bankruptcy and a large loss short of bankruptcy for an investment fund. It shouldn’t matter to investors whether a fund in which they have invested $10 million goes bankrupt or a fund in which they have invested $30 million loses a third of its value.” And, as we saw in Section 1.3, if all we need to be concerned about is the Sharpe ratio, many of the elements of financial risk management, such as the inclusion of subjective judgment, are not as strongly needed. While there is some truth to this argument, it also has some deficiencies. Firms that have credit risk to the fund, often through counterparty risk on derivatives, may care very much about a fund’s tail risk. Regulators are showing increasing concern about potential destabilizing effects of Managing Financial Risk 161 investment fund bankruptcy. And investors may be concerned that they are not receiving adequate return for tail risks the fund is taking, if these tail risks are unmonitored. There is a strong argument for at least measuring a fund’s tail risks, even if it is a greater tolerance for an investment fund taking recognized and adequately compensated tail risks. Thus we are seeing more use of financial risk management techniques such as VaR and stress testing for investment funds (for example, Duc and Schorderet 2008). Investment funds in which investors are expected to have concentrated risk, most particularly pension funds, are certainly expected to be concerned with tail risk. This is especially true when they pursue a strategy that explicitly depends on liquidity. A good example is contingent immunization, a strategy developed by Leibowitz and Weinberger (1981, 1982, 1983). In contingent immunization, you constantly monitor how much excess you have in fund assets relative to the amount needed to invest in perfectly safe assets, such as zero coupon Treasury bonds, to meet the minimum payout requirements of the fund. As long as an excess exists, fund managers are free to engage in any investment strategy they wish and still be able to assure meeting the minimum payout requirements—as soon as the surplus reduces to zero, they would switch the fund assets into the safe portfolio. But calculation of this surplus requires constant monitoring, both as to the amount of safe assets needed to meet the minimum payout, which will change as zero coupon Treasury rates fluctuate, and as to the liquidation value of the current portfolio in the event this switch needs to be made. Constant calculation of liquidation value requires all of the machinery of financial risk management: accurate, independent, and continuous marking to market; VaR and stress test computations to assess potential loss in liquidation; and valuation reserves against illiquid positions. O’Kane (2008, Section 22.2) has a good discussion of the gap risk in constant proportional portfolio insurance, a product whose design is very similar to a contingent immunization strategy. 6.2 RISK CONTROL Once an adequate measurement of risk is available, the next logical question is how to control it. Two fundamental and complementary approaches are available. The first is for higher levels of management to place detailed limits on the amount and type of risk that lower levels of management can take—limits on VaR, position size, vega, gamma, and so on. The second is for higher levels of management to provide incentives to lower levels of management to optimize the trade‐off between return and risk. The latter approach, based on incentives, gives lower levels of management, which are 162 FINANCIAL RISK MANAGEMENT closer to the information required to make informed trade‐off decisions, the flexibility to find combinations of risks that can maximize the return for a total risk level approved by senior management. However, the incentive approach can also lead to unacceptable risks in the aggregate if too many traders decide to take a similar position, pointing toward a mixed use of both approaches. This is the pattern that can be found at almost all investment banks and will be the approach followed in this book for discussions of control techniques for specific risk classes. The most extreme form of an incentive‐based approach is to restrict controls to assigning to each trading desk a maximum amount of trading losses the traders will be allowed to take before their positions are closed out. This gives the trading desk maximum flexibility in deciding what positions to put on and gives complete freedom as long as unacceptable losses are avoided. Everyone concerned—the traders, senior managers, and risk managers—can agree on such stop‐loss limits as a bare minimum for risk control. If all positions could be instantaneously liquidated at any time at the values reflected on the firm’s books, it could be argued that this is an adequate limit structure. However, there have been too many instances where a trader has built up a large risk exposure that proved costly to exit when management decided to stop out losses. The time that traders exceed loss limits is often also the time when markets are moving wildly, decreasing liquidity and subjecting positions to large P&L moves even when closeout can be accomplished in the relatively short time of a day or two. At least some additional form of risk control is needed. Historically, added risk controls have most often been quite detailed limits on the sizes of specific exposures that could be taken, with limit sizes closely tied to both the liquidity of exiting the exposure and the degree of management confidence in the trader. When the VaR measure was first introduced, it was initially seen as a possible supplement for limiting risk. However, soon traders came to see it as a tool for gaining added flexibility, since it treats all risk as fungible in arriving at a single risk number. Since this risk number is a statistical estimate of the loss that could occur during the period in which a position is being closed down, an argument could then be made for using this as the only supplement to a stop‐loss limit, allowing control on the loss that can occur after management has decided to close out a risk position without the need to place detailed controls on particular exposures. This control can take the form of a limit on the total VaR exposure that a trading desk can take and/ or a measure of risk in a calculation of return on risk or risk‐adjusted return that can be used to compare the performance of different trading desks against targets and against one another to decide on compensation, promotion, and continued employment. Managing Financial Risk 163 The following are arguments favoring an incentive approach, with senior management input reduced to broad measures such as stop‐loss limits and VaR limits or risk‐return targets, giving great flexibility in deciding on the risk‐taking profile to the business: ■ ■ ■ ■ An incentive approach enables trading desks to respond quickly to new opportunities without slowing down decision making by needing to make their case to senior management. By not restricting a given trading desk to positions in a particular asset class, an incentive approach encourages broad thinking across asset classes, searching for interrelationships. When a trading desk is confined to a particular market at a time when there is a shortage of good trading opportunities in that market, traders are often tempted to pursue riskier opportunities in that market as the only hope for earning a bonus. Giving trading desks the flexibility to trade in other markets when the one they specialize in is less promising is a way to avoid this temptation. It is less risky to have many traders with the ability to take positions in a given market than to restrict position taking to a single trading desk. In most circumstances, positions taken by one desk will be offset by positions taken by a desk with a different opinion. When enough desks all line up in the same direction to create a sizable net position, it is a good indication of particularly favorable return‐on‐risk circumstances. The following are arguments favoring a more detailed limit approach: ■ ■ ■ A more detailed limit approach enables management to restrict position taking in a particular market to only those trading desks possessing sufficient knowledge and expertise in the market to be able to make reasoned judgments. As a corollary to the previous argument, it forces trading desks to focus their attention on those areas in which they are expert without having this focus distracted by trying to find opportunities in other markets. The real danger is that a trading desk that does not have a successful strategy in its primary market can obscure management recognition of this fact by trying to build a profitable trading record in another market. This can be particularly harmful if it helps to perpetuate a Ponzi scheme in which the firm is delayed in recognizing the mispricing of a transaction with long‐term consequences. If a desk is allowed to play in another market, it is important to make sure that P&L attribution firmly separates the results for different products for the same reason we have seen that it is important to separate P&L in newly booked 164 ■ ■ FINANCIAL RISK MANAGEMENT deals from P&L on management of existing deals. As a particular case, if an options trading desk is allowed to take substantial outright positions in the underlying asset, the P&L from underlying positions must be clearly separated from the P&L on management of volatility and convexity risk. Likewise, if an exotic options trading desk is not forced to lay off the substantial part of the vanilla options risk it generates, then the P&L from vanilla options risk must be clearly separated from that on the residual exotic options risk. The final argument given in favor of more flexible position taking is actually quite misleading in two directions. First, it underestimates the degree to which opinions can be infectious and create bandwagon effects, particularly among traders who are not experts in a particular market. The risk is that when the trading desk with the most expertise in a particular market puts on a position, other trading desks will pile on to get a piece of the action. As a result, the firm as a whole will wind up with a much larger position than the trading desk with the expertise would have thought prudent. Second, when situations with less certainty arise and trading desks put on offsetting positions, the firm as a whole winds up being arbitraged—it has flat P&L if the market moves in either direction, but must pay a bonus to the winning trading desk in either case. This points to the need for trading management to insist on trading desks utilizing diverse styles to avoid this form of arbitrage, and detailed limits can play an important role in enforcing the diversity of trading style. Management may distrust the excessive reliance on statistical measures of risk. Statistics are based on history and may not reflect management judgment about risk. This may be particularly true in markets that tend toward infrequent but large jumps, such as pegged FX rates, which, due to government intervention, may show a long history of very little movement followed by one sharp break. When the government resources are no longer adequate to hold the desired peg, the tendency is for the resulting move to be very large to reflect the market pressures whose reflection in the price has been suppressed by government intervention. In a period when the peg still holds, historically based VaR will show very little risk, but this will not adequately reflect the possibility of a jump move. Instead of historical relationships, VaR can be based on implied volatilities, which reflect a market judgment of future uncertainty, or on management estimates of risk. A more direct approach is to explicitly limit exposures to management‐designed stress scenarios. The issue of whether to permit a trading desk to take positions in instruments outside its primary expertise is not just a question of whether a desk Managing Financial Risk 165 should be allowed to actively seek such positions. This issue also comes up as a question of whether a desk should be forced to close out positions that result as by‐products of its primary product focus. Consider an FX options market‐making desk. Their primary expertise would be on issues such as the proper management of volatility risk. However, outright FX positions arise naturally in the course of its business, as changes in exchange rate levels lead to changing deltas on its option positions. Should the options desk be forced to close out these outright FX positions, leaving the firm’s positioning of outright FX to the spot‐and‐ forward FX market makers, the firm’s experts at managing these positions? Or should the options desk be allowed to take its own view on these positions? The same arguments, pro and con, that we have presented previously apply here as well, with a particular emphasis on the second argument pro flexibility and the third argument against flexibility. Those who favor flexibility point to the broader view of economics and the markets that will come from the trading desk looking at its options positions as a whole, rather than trying to break them apart into a position in the underlying and a position in volatility. They will point out that this encourages thinking about correlations between underlying prices and volatility levels that can best be taken advantage of by being able to manage positions in both the underlying and volatility. These are powerful arguments, as discussed in Chapter 11. Those suspicious of the consequences of flexibility point to cases in which poor pricing of volatility risks and poor management of options positions were delayed in being recognized by profits that came from taking positions in the underlying (perhaps just by copying positions that the primary underlying desk was putting on). This certainly indicates the need to have, at a minimum, risk reporting that clearly breaks out P&L attributable to the underlying position from P&L attributable to volatility positions. Even if management decides in favor of the less flexible approach with specific limits on options traders taking positions in the underlying, some degree of flexibility should be retained from a pure transactional efficiency viewpoint. For example, if an options trading desk is never allowed any position in underlying assets, it will need to spend too much of its time writing tickets to close out delta shifts arising from underlying price changes and will lose too much of its P&L in bid‐ask spreads. (Even if these are only internal and hence not lost to the firm, it will still be demotivating to the traders.) The arguments we have presented here for options traders and their positions in the underlying apply equally to forward traders and their positions in the spot market, basis traders and their position in legs of the basis, and exotic options traders and their positions in vanilla options that can hedge part of the exotics’ risk. 166 FINANCIAL RISK MANAGEMENT This discussion on risk controls has important implications for the use of risk decomposition techniques throughout the remainder of this book. It explains why I place such a strong emphasis on utilizing risk decomposition to break apart less liquid transactions into constituent parts—usually a more liquid piece and a less liquid residual. Identifying the more liquid constituents enables the separation of P&L attribution and encourages closing out positions with the desk that can create the maximum liquidity for the firm. It also avoids the booking of phantom P&L by having a different valuation technique used for the same position depending on whether it was created directly or created as part of a more complex transaction. Finally, it also avoids the firm’s unknowingly building a large position in a particular product. For example, this motivates the use of a formula for vanilla options that does all the pricing and representation of risk in terms of forward prices derived from the trading desk that is the primary market maker in that product (see Chapter 11) and motivates the attempt to price and represent the risk of exotic options to the greatest extent possible as a combination of vanilla options prices derived from the trading desk that is the primary market maker in that product (see Chapter 12). A closely related question is whether trading books that take positions in a product in which they are not a primary market maker should be forced to do all their transactions through the firm’s primary market‐making desk for the product. As a concrete example, consider a trading desk specializing in FX options, which will certainly need to transact hedges in underlying spot and forward FX. Should the traders be forced to transact all such hedges with the firm’s spot and forward FX trading book, or should they be given the choice of dealing directly in the market? Note that this issue arises regardless of whether trading limits are used to force the options desk to restrict its outright FX positions to a small size. In either case, the desk will be transacting at least some hedges either internally or with the market. The arguments for requiring internal hedging are powerful: ■ ■ It enables the desk with the greatest expertise and advantage in trading a product to be the one initiating all external transactions. It reduces the amount of transaction costs the firm must pay by encouraging trades in opposite directions to be closed out within the firm and enabling internal trades to be crossed with customer transactions. Nothing pleases traders more than to be able to boast of the profits they have made by standing in the middle of trades in opposite directions put on by different desks of a rival firm. Even if positions are not completely offsetting or exactly simultaneous, funneling the trades through a single desk enables that desk to see the total flow of the firm’s dealings in the Managing Financial Risk ■ 167 product. This desk can build on observed patterns of usage to forecast and anticipate flows and minimize transaction costs. The use of a common central trading desk forces all desks within the firm to value positions in the same product at a common price. This avoids phantom profits arising from the internal arbitrage that can occur if two desks value their positions in the same trade using different broker quotes or different models. Proper valuation discipline can eliminate this even if a policy of forcing all trades through a single desk is not employed, but this is the easiest mechanism for enforcing this rule. The argument for permitting several desks to trade the same product directly with other firms is that competition for business will create enough efficiencies to overcome these strong advantages of a common central trading desk. The fear is that creating an internal monopoly in a product will permit the monopolist to try to collect monopoly rents from the other desks trading in the product—that is, to price at excessive bid‐ask spreads that will increase profits of the central desk, but decrease the firm’s overall profit by discouraging optimal use of the product by other desks. Avoiding this situation may require a difficult internal policing effort (it’s not always easy to measure the size of the bid‐ask spreads being used, since trades in different directions do not come in simultaneously). CHAPTER 7 VaR and Stress Testing n the statement of requirements for robust risk management in Section 6.1.1, the estimation of losses that could result from liquidation of positions figured very prominently. This showed up under the headings “The need for simulation” and “The need to consider periods of reduced liquidity.” The need for simulation, which closely corresponds to the G‐30 Recommendation 5, “Measuring Market Risk,” is discussed in detail in this chapter as value at risk (VaR). The need to consider periods of reduced liquidity, which closely corresponds to the G‐30 Recommendation 6, “Stress Simulations,” is discussed in this chapter as stress testing. These two methods for measuring the total risk exposure of a portfolio still need to be supplemented by more detailed nonstatistical risk measures, such as the value of the basis point, delta, or vega, for reasons given in Section 6.2. But measures of total portfolio risk do offer advantages that detailed nonstatistical risk measures do not: I ■ ■ Nonstatistical measures do not allow senior managers to form conclusions as to which are the largest risks currently facing the firm. It is not possible to meaningfully compare the value of a basis point in two different currencies, since this comparison does not reflect the relative size of potential interest rate moves in the two currencies. Both VaR and stress testing give a measure that combines the size of position and size of potential market move into a potential impact on firm profit and loss (P&L). Moreover, both produce a measure that can compare risks between disparate businesses, such as interest rates and equities. Nonstatistical measures do not interact with one another. Should you add up the risks under different measures into some total risk? Clearly this would be wrong because it would ignore the effect of correlation between market factors. Both VaR and stress testing account directly for correlation between market factors. 169 170 FINANCIAL RISK MANAGEMENT We will first discuss the methodology of statistical measurement, VaR, and then discuss the methodology for nonstatistical measurement, stress testing. A book‐length treatment of the topics discussed in this chapter is Dowd (2005), which offers a wealth of detail and covers all the methods that I consider best practices in this area. This is a book I recommend highly for those working on implementation of VaR methodology. You will see many references to it in this chapter. What I offer here are the aspects of VaR that are most important for everyone involved with risk management to know, and those methodological considerations for implementation that my experience in the field has shown to be of greatest consequence. 7.1 VAR METHODOLOGY Strictly speaking, VaR is a measure of the worst loss that can occur at a given confidence level. But the statistical methodology used to determine VaR can also be used to calculate broader measures of the distribution of potential losses. In Section 7.1.1 we’ll first look at the methodology for calculating the distribution and in Section 7.1.2 we’ll turn to the question of how best to summarize it. Since statistical risk measures first began to be calculated by financial firms (about 20 years ago), three methods have dominated: 1. Direct measurement of P&L distribution. 2. Calculation of P&L distribution based on historical statistics representing the variance and covariance of market variables and the current size of position exposures to each of these market variables. So if si represents the firm’s exposure to each market variable, σi represents the volatility of each market variable, and ρii,j represents the correlation coefficient between each pair of market variables, the volatility of overall firm P&L is calculated as: ∑ si s j σ i σ j ρi j i j The P&L distribution can now be calculated from this volatility. 3. Simulation of P&L distributions based on a selected set of possible moves of market variables and the current size of position exposure to each of those market variables. So if si represents the firm’s exposure to each market variable, mii,j represents the size of move of each market 171 VaR and Stress Testing variable in each considered scenario, and pj represents the probability assigned to each scenario, with: ∑ pj = 1 j Then the P&L movement in each scenario is calculated by: ∑ si mi, j i And the P&L distribution is calculated by multiplying each of these terms by its respective pj. We will consider the advantages and disadvantages of each of these three methods. The direct measurement of P&L distribution is still widely used, as can be seen from the frequent use of histograms of daily P&L distributions published in annual reports of financial firms, of the type illustrated in Figure 7.1. It has the advantage of simplicity of calculation, not having to make any use of models or statistical assumptions. It also has the ability to capture effects of the trading culture, which the other methods do not. For example, does management respond to periods of greater market volatility Daily IB and Other VaR less market risk–related losses Daily IB and Other Market Risk–Related Gains and Losses (95% Confidence Level VaR) V 8 Number of trading days Year ended December 31, 2010 80 Average daily revenue: $87 millioo 60 50 4 2 20 10 $ in millions FIGURE 7.1 P&L Histogram from JPMorgan 2011 Annual > 240 210 > < 240 180> < 210 150 > < 180 120 > < 150 90 > < 120 60 > < 90 30 > < 60 0 > < 30 (30) > < 0 < (30) 0 > 120 80 > < 100 $ in millions 30 100 > < 120 40 > < 60 40 60 > < 80 <0 0 20 > < 40 Number of trading days 70 6 172 FINANCIAL RISK MANAGEMENT by reducing position size? If it does, this will mitigate some of the earnings volatility resulting from market volatility. Direct measurement of P&L distribution is also the only method that is available for measuring risk when access to details of trading positions is not available. For example, a hedge fund investor probably does not have any access to details of the investment holdings of the hedge fund. To estimate its risk, the investor may need to rely on historical P&L distribution of the fund (for more on risk management of investments in hedge funds, see Section 8.4.1). However, direct measurement of P&L distributions cannot take into account the possibility that current position taking may be radically different from historical position taking (in the fund management world, this is known as style drift). Corporate risk managers and regulators will insist on risk measures that fully reflect current portfolio composition, whenever available. This renders direct measurement of the P&L distribution close to useless as a stand‐alone risk measure, though it is still valuable as a complement to other measures. The use of the variance‐covariance method has now been virtually abandoned by sophisticated financial firms in favor of simulation methods. The primary reason for this is that relative to the simulation method, the variance‐covariance method provides very little flexibility in evaluating the contribution of nonlinear positions, notably options positions, to P&L distributions. As we will see, simulation gives the flexibility to tailor the degree of detail used in calculating nonlinear positions to the degree of accuracy required for particular portfolios. Detail can range from simple factor approximations (using delta, gamma, vega, etc.) to full valuation of each individual option, with several gradations available in between. By contrast, variance‐ covariance can’t go beyond factor approximation. Secondary reasons are: ■ ■ ■ The greater difficulty that the variance‐covariance method has in dealing with the fat‐tailed distributions normally encountered in financial markets. The inability of variance‐covariance to pick up the phenomenon, often observed in financial markets, that the largest changes in variables often cluster together (e.g., the high correlation between stock markets in different countries in the 1987 stock crash) to a greater degree than will be indicated by correlation coefficients (i.e., the joint distribution is not bivariate normal). The realization that almost all the benefits of simplicity and speed of computation claimed for variance‐covariance relative to simulation were based on fallacious comparisons. As will be seen in our discussion of simulation methodology, the degree of simplicity and speed of computation are largely determined by the choice of the user. Achieving 173 VaR and Stress Testing a level of accuracy similar to that obtained by variance‐covariance, simulation is at least as simple and fast to compute as variance‐ covariance. Simulation offers the flexibility, which variance‐covariance does not, of increasing accuracy as a trade‐off against simplicity and computation time, but having more flexibility can surely not count as a disadvantage. Currently, the primary users of variance‐covariance are smaller firms that do not hold significant options positions and that wish to outsource the market data component of their VaR computations. For such firms, variance‐covariance does offer the distinct advantage that they only need to obtain volatilities and correlations rather than the day‐by‐day pricing histories required for simulation, a considerable savings in the amount of data to be transferred. In Exercise 7.1, you will have a chance to see an example of how variance‐covariance computes VaR and why a simulation calculation that is as simple computationally and is superior in flexibility is always available. I will therefore not spend any more time on variance‐covariance or the various tricks that have been devised to provide capability to approximate option positions and incorporate fat tails within it. For readers who wish to pursue this approach, I recommend Chapters 6 and 10 of Dowd (2005). 7.1.1 Simulation of the P&L Distribution Remember that the simulation approach consists of determining a number of possible scenarios, to be indexed by j, determining the size of move of each market variable in each scenario mii,j, and then calculating: ∑ si mi, j i as the firm’s total P&L movement in each scenario. The steps in a P&L simulation consist of (1) determining a set of scenarios specified by the size of move in each of a set of underlying market variables and a probability to be assigned to each set and (2) translation from the size of move of underlying market variables to size of move for all market variables. For example, the underlying market variables for a set of bond positions could be interest rates for 10 key tenors, and the full set of market variables could be prices for individual bonds. There are two alternative approaches to the first step—historical simulation and Monte Carlo simulation. The decisions to be made for the second step do not depend on the choice made for the first step. We will discuss each step in some detail. 174 FINANCIAL RISK MANAGEMENT 7.1.1.1 Step 1: Determine Underlying Market Probabilities The historical simulation approach is quite simple; a group of historical periods is chosen and the observed sizes of market moves in each of these historical periods constitute the scenarios. So, for example, you could choose 1,200 scenarios consisting of all the most recent one‐business‐day changes in market variables—the changes in market variables from 6/7/99 to 6/8/99 would be one scenario, the change from 6/8/99 to 6/9/99 another scenario, and so forth. Or one could choose all the 10‐business‐day changes. The most commonly used method for historical simulation assigns equal‐ probability weights to all of these possible market moves. This makes calculation of VaR very simple, since it is just equivalent to one particular scenario (for example, if you wanted the 99th percentile VaR and you are working with 1,200 scenarios, the 12th worst loss in any of these scenarios is the 99th percentile VaR). When we explain the calculation of measures of P&L distribution in Section 7.1.2, we will discuss the possible advantages of and methodology for assigning unequal probability weights to these market moves. Historical simulation offers a large advantage in terms of simplicity— simplicity of implementation, simplicity of assumptions, simplicity of explanation. The advantage in terms of assumptions is that no modeling assumption needs to be made beyond the assumption that the immediate future will resemble the past. There is no parameterization of either variance or correlation and no assumptions about distribution shape (e.g., normality). If fat tails or clustering of large moves between variables are present in the historical data, they will be reflected in the simulation. The advantage in terms of explanation is that any questions raised by traders or managers concerning a VaR that seems too high can be easily traced to a subset of specific historical dates that would show large losses against the current firm holdings. Disagreement can be quickly focused on accuracy of data for a few specific dates or on arguments about the probabilities to be assigned to repetition of particular historical events. By contrast, both the variance‐covariance approach and the Monte Carlo simulation approach make it far more difficult to resolve such questions. This advantage of simplicity of historical simulation also underlies its primary disadvantage—the VaR produced is dominated by market moves on a few specific historical days. If a particular combination of market events did not occur in the historical period being considered, it cannot contribute to VaR. It is difficult to overcome this problem by just expanding the historical period you are considering. Data availability tends to get sparse once you go back more than a few years, because of failure to retain data, because data becomes more difficult to clean the further back you go in time, and because some currently traded instruments may not have histories that go back that far. VaR and Stress Testing 175 This disadvantage of generating scenarios utilizing the historical method is the primary argument in favor of the Monte Carlo method. The Monte Carlo method starts with a specification of the underlying market variables that is similar to that of the variance‐covariance approach, but may have a richer specification of each single variable than just a volatility—for example, a multiparameter specification that allows the generation of distributions that are skewed or fat‐tailed. Monte Carlo generation of distributions that fit specified parameters can be achieved in several ways: ■ ■ ■ By mixing together normal distributions, distributions that are skewed and fat‐tailed can be generated. This can be done using the MixtureOfNormals spreadsheet we encountered in Chapter 1. Mixing normal distributions with the same mean and different volatilities produces fat tails but no skew. The larger the difference in volatilities, the greater the kurtosis, a measure of how fat‐tailed the distribution is. Mixing normal distributions with different means and different volatilities produces both fat tails and skew. More detail can be found in Dowd (2005, Section 6.5.3) and Wang (2001). By using stochastic volatility and jump process specifications, similar to those we discuss in Sections 11.6.2 and 12.3.2. See also Hull (2012, Sections 26.1 and 26.2) and Dowd (2005, Sections 6.5.4 and 6.5.5). By using processes specially designed to generate Monte Carlo distributions that match given skew and kurotsis parameters, such as those discussed in Shaw (1997). Monte Carlo techniques are then used to generate a set of scenarios that fit the desired statistical specifications. Shaw (1997) discusses building Monte Carlo simulations following the algorithm of Ramberg et al. (1979). An implementation of this algorithm can be found on the website for this book (this algorithm on the website is called “Quasifit”). Usually, users of Monte Carlo simulation want to take advantage of the flexibility it offers to generate many more scenarios than can be practically generated with historical simulation. This has led to the incorrect assertion that Monte Carlo simulation requires more scenarios than historical simulation does. Rather, Monte Carlo simulation offers the flexibility of achieving greater accuracy if the greater expense of running more scenarios is justified by the increase in accuracy. Standard computerized techniques for improving the trade‐off between accuracy and speed for Monte Carlo simulation can also be employed (e.g., stratified sampling, importance sampling, low‐discrepancy sequences; see Hull 2012, Section 20.7; Dowd 2005, Section 8.4; and Jackel 2002, Chapters 8 and 10). 176 FINANCIAL RISK MANAGEMENT Advantages that Monte Carlo simulation offers are: ■ ■ ■ ■ Ability to select the most suitable technique to estimate each parameter. Volatilities and correlations can be forecast using statistical techniques such as weighted moving averages and generalized autoregressive conditional heteroscedasticity (GARCH). For a discussion of the most common statistical methods used in forecasting volatilities and correlations, see Hull (2012, Chapter 22), Dowd (2005, Chapter 5), and Jorion (2007, Chapter 9). Statistical methods for adjusting parameters derived from historical data to be more robust, including random matrix theory and shrinkage estimation, can be found in Fabozzi, Focardi, and Kolm (2006, Chapters 8 and 9). Valuable discussion of this topic can also be found in Meucci (2005, Chapter 4). Swensen (2000, Chapter 5) is a valuable approach with less emphasis on statistical methodology and more on economic insight. Where implied volatilities are available, they can be substituted for or blended with statistical measures. (Should implied volatility always be used when available? We’ll examine this question at the end of this subsection.) The choice can be separately made for each variable, though you do need to be careful not to generate impossible or implausible combinations of correlation coefficients; for discussion of how to avoid creating impossible correlation matrices, see Dowd (2005, Section 5.3). Ability to select the most relevant data set for estimating each parameter. You might have 10 years of good historical data for one variable and only two years for another. Historical simulation would force you to use only two years’ worth of data for both. Monte Carlo simulation lets you choose the data set individually for each variable. You can also choose the most appropriate weights to assign to different historical periods for each variable, with more discounting of older historical data for some variables than for others. Historical simulation can only utilize a single weighting scheme that applies equally to all variables (see the discussion of this weighting of historical simulation in Section 7.1.2). Ability to select the most relevant data set for estimating different aspects of a single variable. For example, volatility could be based on recent data or derived from an implied volatility while higher‐order parameters of the distribution are estimated from longer data periods. Recent data is often considered a better predictor of near‐term future volatility, but shape parameters, such as fatness of tails, are hard to discern from a small data set. Greater flexibility in handling missing data. Data for individual dates can be missing because a particular market was closed for a holiday or because of errors in data gathering. In fact, all sources of market data, VaR and Stress Testing ■ ■ 177 whether data vendors, brokers, or databases internal to the firm, are notoriously poor in quality and require major data scrubbing efforts. But some data will not have sufficient duplication of sources to scrub successfully and must be regarded as unavailable. Monte Carlo simulation can exclude periods for which a particular data series is missing from the calculation of each individual variable without excluding this period from the calculation of other variables for which the data are available. Historical simulation lacks this flexibility—it must either completely include or completely exclude a particular day’s data. Greater flexibility in handling nonsynchronous data. Correlations observed between variables that are sampled at different times of the day can be highly misleading and result in significant misstatements of risk. Monte Carlo simulation has the flexibility to measure correlation for each individual pair of variables based on quotations from the best time of day to represent that particular pair, or by basing the correlation on a multiday time interval, which will tend to smooth out nonsynchronous effects. For more detail on statistical methods that can be used in estimating correlations between nonsynchronous data, see RiskMetrics Group (1996, Section 8.5) and Holton (2003, Section 6.3). Ability to combine histories. Consider a corporate bond held in the firm’s portfolio. By historical experience, one knows that some of these bonds may suffer a ratings downgrade and subsequent large fall in price. But it may be that none of the bonds currently held has suffered such a downgrade since the firm avoids holding such bonds. Historical simulation would show no ratings downgrade events for these bonds. But Monte Carlo simulation could be used to combine ratings downgrade possibilities based on the history of a large pool of bonds with specific pricing history of actual bonds held. Another example would be a foreign exchange (FX) position held in a currency that has been pegged at a fixed exchange rate to the dollar by government intervention. You may have no historical example of this particular currency devaluing, yet want to include some probability of devaluation. Monte Carlo simulation could incorporate a devaluation event, possibly parameterized by devaluation experience in other currencies, as a jump process superimposed on the specific history of this FX rate. Still another example would be two stocks that have begun trading in a very tightly related fashion since a merger announcement. You would not want to reflect their previous more volatile arrangement as part of the history that determines VaR. So you must generate the price of one stock as a function of the other, but with a random element introduced to represent the risk of a sharp break in the price relationship 178 FINANCIAL RISK MANAGEMENT if the merger fails to go through. This random element should be based on the price history of a large pool of stock pairs following a merger announcement. Finally, Monte Carlo simulation allows users great flexibility in deciding on the most effective approach to specifying each individual variable. Consider as an example specifying parameters for credit default swaps (CDSs) (see Section 13.1.1.2 for more details on CDSs). CDS prices for some corporations may have sufficient liquidity that you want to estimate the parameters for this price based solely on the price history of this particular CDS. For other corporations with less liquidity in CDS prices, you can choose to break the CDS price up into a credit spread on a bond issued by that corporation plus a spread between the CDS spread and the bond’s credit spread (we’ll call this the CDS basis). The parameters for the credit spread on the bond might be based solely on the history of credit spreads for this corporation’s bonds, while the parameters for the CDS basis might be better estimated from observations of CDS basis history drawn from a larger universe of similar corporations. Even though you are estimating the CDS basis for several different corporations from the same data source, you don’t expect them to be perfectly correlated, but you can estimate a correlation coefficient from historical observations of how changes in CDS basis differ between corporations. It is straightforward to reproduce any desired correlation matrix in a Monte Carlo simulation using the Cholesky decomposition method described in Dowd (2005, Section 8.3). But covariance matrices employ correlations based on multivariate normal distributions and therefore do not capture any relationships that are extremely unlikely under this hypothesis (e.g., the clustering of large changes in variables). Addressing these concerns requires more refined data analyses. For example, different correlation matrices could be used depending on the size of price moves (see Kim and Finger 2000). Days in which price moves are larger would use a correlation matrix derived from a sample of days with large moves. The MixtureOfNormals spreadsheet can produce correlations with different degrees of clustering, as shown in Figures 7.2 and 7.3. Figure 7.2 shows the correlation between two normally distributed variables with 25% correlation, both with mean 2%, standard deviation 5%. Figure 7.3 shows a mixture of 95% of the first distribution and 5% of two normally distributed variables with 60% correlation, both with mean 0, standard deviation 10%. Note the clustering of points with large losses in both variables and large gains in both variables in Figure 7.3. This clustering does not appear in Figure 7.2, which displays a multivariate normal distribution. 179 VaR and Stress Testing 14 00% 14.00% 12 00% 12.00% 10 00% 10.00% 8 00% 8.00% 6 00% 6.00% 6.00 4 00% 4.00% 0% 2 00% 2.00% % 0 00% 0.00% % -4.00% -2.00% 0 % -2 -2.00% 2 00% 0 %0.00% 2.00% .00% 4 4.00% 6.00% 8.00% 10.00% 12.00% 14.00% -4 00% -4.00% -6 00% -6.00% FIGURE 7.2 Correlation Between Two Normally Distributed Variables 30 00% 30.00% 20 00 20.00% 00% 10.00% 0 00 0.00% -30.00% -20.00% % -10.00% 00.00% 0 00% % -10.00% 10.00% 20.00% 30.00% 40.00% 20.00% 00% -20 -30.00% 30 00% -40.00% 40 00% FIGURE 7.3 A Mixture of Two Normal Distributions Shows a Clustering of Points with Large Gains and Large Losses 180 FINANCIAL RISK MANAGEMENT More general methods for analyzing nonlinear correlations and generating Monte Carlo distributions based on this analysis have been widely studied in recent years. This is known as copula methodology; see Dowd (2005, Section 6.8) for details. Given all these advantages to Monte Carlo simulation in its flexibility to handle data and estimation issues, it is preferable, and sometimes even unavoidable, to still employ some Monte Carlo simulation techniques when you have chosen historical simulation as your primary methodology. Consider these examples: ■ ■ A certain stock held in your portfolio has only recently been issued. To develop a past history for the price of this stock for use in historical simulation, you may represent it by some formula based on a selected stock index. But if you are long this stock and short this index, you would measure your position as having no risk during the period when it is represented by the index. To avoid this, you need to introduce a random element into your generation of the stock’s back price history, basing the size of the random element on observed changes during the period since the stock began trading. But this is precisely the Monte Carlo approach. The ratings downgrade risk case, the FX devaluation risk case, the merger arbitrage risk case, and the CDS basis case discussed in the bullet point headed “Ability to combine histories” under advantages of Monte Carlo simulation are good examples of where a random element needs to be introduced into the historical series. In cases like these, how should a random element be introduced into historical simulation? One method that is sometimes used is to randomly assign the distribution of the random element among the days of historical data. For example, if there is a 1 in 250 chance of a ratings downgrade for a bond, the price drop that would result from a downgrade would be randomly assigned to four days out of a 1,000‐day historical simulation. But this has a large chance of having no impact on the VaR measurement since the four days randomly selected would likely miss the days of largest losses that determine the VaR measure, but a small probability of having a large effect if it happens that one of the four days selected at random corresponds to one of the days of largest loss that determine the VaR measure. I believe that this randomness in contributing to VaR contributes nothing to the accurate measurement of risk. It is far better to simply accept that this part of the VaR measurement mustt be performed by Monte Carlo simulation, even if you have chosen to do the bulk of your VaR measurement by historical simulation. The historical simulation results for the main body of the portfolio are treated as a single series as input to the Monte Carlo VaR and Stress Testing 181 simulation, with a uniform distribution assigning an equal probability to each day’s simulated result (for example, if there are 1,000 simulated days in the historical simulation, each path in the Monte Carlo simulation has a 1 in 1,000 chance of picking each of these days). The elements that cannot be treated by historical simulation would be the remaining series in the Monte Carlo simulation, parameterized as discussed in the section on Monte Carlo simulation. Correlations between factors will be chosen based on best historical evidence and economic intuition. There are other areas in which historical simulation can usefully borrow Monte Carlo simulation techniques. For example, historical simulations can be modified to choose a volatility for a particular instrument based on any of the techniques mentioned in the first bullet point under advantages of Monte Carlo simulation. All that is required is to multiply each historical observation for the instrument by the ratio between the desired volatility and the volatility over the historical period. This transformation leaves all shape characteristics of the historical distribution, such as fatness of tails and correlation structure, intact. This approach is illustrated in the VaR spreadsheet, using the volatility override input explained in the documentation for the historical simulation portion of the spreadsheet. Dowd (2005, Section 4.4.2) outlines a similar idea. When it comes to dealing with missing or nonsynchronous data, the options for historical simulation are very limited. Some way needs to be identified for modifying data before it is input into the simulation. For missing data, some type of statistical inference must be used to arrive at a most likely value for the missing data based on the last prior good data point, the next following good data point, and good data points for related data series (for example, if data is missing for an interest rate for a two‐year tenor, related data series would be interest rates for the one‐year tenor and the three‐year tenor). The simplest methods involve averaging between the last prior good point and next following good point, but overlook valuable information from other data series. At a minimum, one should modify simple averaging to follow the pattern of change that took place in related data series between the last prior good data point and next following good data point. Possibly, more advanced statistical models could be used. For nonsynchronous data, a new data series should be generated of “most likely” synchronous values. For example, suppose that you have available data for closing prices for some stock issues as of Tokyo close of business (COB) and other stock issues as of New York COB. You need to find some series that can bridge the time gap—perhaps a futures contract on a Japanese index that trades in the New York time zone. Then all of the stock quotes as of Tokyo COB can be adjusted for the movement that took place in the Japanese stock index between Tokyo COB and New York 182 FINANCIAL RISK MANAGEMENT COB, generating a series that approximates what the quotes for these stocks would be as of New York COB. This obviously leaves room for error in estimating where true liquidation of positions will take place, but it is the best you can do if you are not utilizing Monte Carlo simulation. Just as we can modify historical simulation to include some of the advantages of Monte Carlo, we might want to modify Monte Carlo to include some of the advantages of historical simulation. Beyond simplicity, the primary advantage of historical simulation is the more refined way in which it handles multivariate correlation. By utilizing actual daily simultaneous price moves across the set of all relevant market variables, nonlinear impacts of arbitrarily great complexity are directly incorporated. This points toward a modification of Monte Carlo in which all individual variables are generated by standard Monte Carlo techniques and all correlations between variables are based on historical simulation. This approach, roughly following Shaw (1997), works as follows. First you perform a standard historical simulation, with equal probabilities assigned to each day’s history. Then, each individual variable is regenerated using a Monte Carlo method based on whatever estimation technique is considered most appropriate (e.g., GARCH, implied volatilities, multiparameter specification). Different methods can be individually tailored to different variables. The use of the historical simulation values is to determine which values of the variables occur simultaneously, based on rank order. On the website for this book, you will find my implementation of this procedure in MATLAB, titled “Reorder.” For example, suppose you have a historical simulation with 850 days. Monte Carlo simulation is used to generate 850 values of each variable. If a particular historical data set consisted of the fourth highest value of variable 1, 38th highest value of variable 2, 625th highest value of variable 3, and so on, then you would create a simulation instance with the fourth highest value of the 850 Monte Carlo simulations of variable 1, 38th highest value of the 850 Monte Carlo simulations of variable 2, 625th highest value of the Monte Carlo simulations of variable 3, and so on. While this approach retains many of the advantages of Monte Carlo simulation, it cannot incorporate them all. It lacks the flexibility to base some correlations on one data set and other correlations on another data set. It requires complete data for every variable in every day to be included in the dates determining the correlation structure. And it has the same problems as historical simulation with nonsynchronous data. Finally, let’s examine the question of whether implied volatility should always be preferred to historical volatility when it is available. In Chapters 11 and 12 of this book, “Managing Vanilla Options Risk” and “Managing Exotic Options Risk,” I argue strongly for always valuing VaR and Stress Testing 183 options at volatilities implied from liquid market prices. But this is a pricing argument—we need to determine prices at which longer‐term volatility risk can be exited in order to convert longer‐term risks into shorter‐term risks. VaR is already dealing with shorter‐term risks, usually overnight. It is also doubtful that there are liquid prices for options to manage risk over such short time periods. Implied volatilities can be used as indicators of overnight volatility, and there may be arguments for believing they carry a great deal of information. But there are also arguments against giving much weight to implied volatilities—they sometimes have more to do with supply and demand factors than forecasts of price variation, as discussed in Section 11.6.2. The decision must be based on belief about their predictive value, as there is no pricing argument for using them. 7.1.1.2 Step 2: Determine All Market Variables This section discusses various approaches to representation of the firm’s exposure to market variables. More details for specific positions can be found in Sections 9.2, 9.3, 9.4, 10.4, 11.4, and 13.1.3. Dowd (2005, Chapter 12) is also a good source for recommendations on this point. Whatever choices are made for how positions should be stored and represented, the most important point with regard to representation of firm position in VaR and stress test calculations is the need for basing all position inputs on data that is entered and controlled by support staff independent of the front office (as per Section 3.1.1). The VaR and stress test reports are key elements for controlling and managing the firm’s risk, and it is just as important for position information feeding these systems to be immune from front‐office manipulation as it is to have independent P&L reporting. Indeed, some of the most recent major frauds by rogue traders have been perpetrated primarily through manipulation of risk reporting rather than through the manipulation of P&L reporting (as per Section 3.1.1). Falsification of P&L reporting can lead to stop‐loss limits being missed, but the equally important limits on buildup of large positions that can be very costly to liquidate depend on VaR and stress test reports. Nearly as important is to have checks in place to ensure that the position information that feeds P&L calculations and desk‐level risk reports is identical to the position information that feeds VaR and stress test reports. Not only is this a vital check on accuracy of risk reporting, but it is also needed to maintain good dialogue between risk managers and front‐office personnel. Nothing is as destructive of good dialogue as VaR or stress limit violations that make no sense to traders because they contradict desk‐level risk reports. When computing VaR for spot positions, the translation from underlying market variables to the full set of market variables that you want to 184 FINANCIAL RISK MANAGEMENT multiply by the firm’s positions is quite direct. Spot positions such as spot FX or the holding of an individual stock or stock index or spot gold or spot oil is just directly multiplied by the generated price change from Step 1. Computation for forward positions is less straightforward. If you are currently holding a Treasury bill maturing one month from now, you don’t want to apply to it the price move you observed for that Treasury bill on a date six months ago, since at that point the Treasury bill had seven months to maturity, and you expect seven‐month instruments to demonstrate much larger price changes than one‐month instruments. So you want to utilize yield curve parameters as underlying market variables and then multiply those yield curve parameters by the appropriate value of a basis point measure of forward position. This has the important added advantage of not having to separately price each interest rate instrument but instead working with a summary description of the entire position. Issues are most complex for option positions (in which we include any nonlinear payoff positions). The conceptually simplest and most accurate approach would be to value each individual option separately based on the changes in the underlying market variables of forward price and implied volatility. Even such a simple approach has complications, since for each scenario it is necessary to choose a volatility at which to evaluate the option. This requires deciding which point on the implied volatility surface is the right one to apply. Suppose you are repricing an option with one year to expiry, a strike price of 100, and current underlying price of 80. Which implied volatility shift do you use when sampling from a period six months ago when the underlying price was 100? Most practitioners would opt for looking at the shift in options with a one‐year expiry and a strike of 125, since that would give the same “moneyness” (i.e., a strike 25 percent above current spot). But this is clearly open to interpretation and a variety of theories on what drives options pricing (see Derman 1999). Very similar considerations apply to option‐adjusted spreads on mortgage and mortgage‐backed securities, which should be related to the security that had a comparable relationship to the prevailing new mortgage rate. The reasoning is similar, since option‐adjusted spreads represent the market pricing of uncertainty in options exercised by homeowners. While the simplest approach is the most accurate, it is clearly also the most costly, and the heavy expense of doing full individual revaluation of each option position is what was primarily responsible for incorrect claims that simulation methodology for VaR was inherently expensive to perform. In fact, simulation methodology can achieve better accuracy than variance‐covariance at no greater cost by the easy trick of representing option portfolios by summary statistics of deltas, gammas, and vegas and multiplying these by the appropriate price change, half the square of change VaR and Stress Testing 185 in price, and the change in implied volatility, respectively. This simplified representation makes options positions no more computationally difficult for simulation than linear positions. So it is a matter of trade‐off in desired accuracy versus cost to be determined for each options position. There are also intermediate approaches. One that can provide quite accurate approximations is to interpolate results based on a price‐vol matrix representation of the options portfolio, as per Section 11.4. If a reasonably detailed price‐vol matrix is already being calculated as part of the trading desk’s own risk reporting, this is a good way of taking advantage of a large number of full revaluation runs that are already being made (since each bucket of the matrix requires all options in the portfolio to receive a full revaluation) without needless duplication of effort. As we note in Section 11.4, the price‐vol matrix can potentially capture all higher‐order terms in the Taylor series of both the underlying price and the volatility, as well as cross‐terms between them. It will not capture impacts such as nonparallel shifts in volatility surface, so these sensitivities will need to be separately accounted for. Whatever approximations are used should be tested occasionally against a full revaluation by individual option to see if a finer degree of detail is needed. The scenarios involving the very largest shifts should probably always be evaluated by full revaluation by individual option. This is a form of importance sampling (see Dowd 2005, Section 8.4.3). One possible implementation would be to first use a selected approximation technique to simulate all possible shifts, then focus on the ones that produce the highest P&L changes, which will have the greatest influence on the VaR measure, and recalculate these using full revaluation of each option. Choices as to whether to work with full revaluation of individual option positions, a price‐vol matrix, or summary sensitivity statistics should be solely motivated by trade‐offs between computation time and expense versus accuracy. In all cases, the ultimate accuracy of P&L simulations rests on the accuracy of the models the firm uses to value transactions. This is true whether the models are used directly in full revaluation or indirectly in supplying the deltas, gammas, vegas, and price‐vol matrices, which are multiplied by positions in simulations or in variance‐covariance calculations. Reviews of accuracy of the firm’s models should always consider their impact on risk calculations such as VaR and stress tests along with their impact on valuations and limit calculations (see Section 8.2.3). Another important determinant of the cost of calculating simulations and the cost of storing the data needed as input to these simulations is the degree of detail with which positions and market prices are recorded. At one extreme, it would be foolish not to keep separate prices and positions for each different currency for spot FX—there just are not that many 186 FINANCIAL RISK MANAGEMENT different currencies, and movements between them can be significant. At the other extreme, it would be equally foolish to store market data on forward rates for all possible tenors (i.e., 365 days × 30 years). Most of these rates are just being produced by interpolation anyway, so you might as well store just the 20 to 50 liquid rates on the curve that all the others are calculated from. In between, there are trade‐off decisions to be made. For example, do you want to track individual histories on every stock you hold, or do you want to keep track of just indexes with individual stocks represented through their betas relative to the index? If you choose the latter approach, then a separate estimate needs to be made of the VaR due to idiosyncratic stock risk. Finally, we note that some of the determinants of exotic derivative prices are not market variables whose price history can be observed and so are not suitable for inclusion in a VaR analysis. Consider an option on a basket of stocks. The impact of changes in the prices of the stocks and in the implied volatilities of each stock in the basket can be computed and included in the VaR. But there will probably be no liquid market quotations for the implied correlations impacting this option. Analysts are occasionally tempted to substitute changes in historical correlation for unobservable changes in implied correlation. I would argue that this is an error. If the basket option has three years remaining, you should presumably look at the change from one business day to the next of a change in the three‐year historical correlation. But since these two three‐year periods will share all but one day at the beginning and end in common, the change in correlation that you will measure must be tiny. We know from experience that implied volatility can change far more rapidly than a similarly computed change in historical volatility, and I do not know of any reason why correlations should behave differently. If, on the other hand, you decided to choose a much shorter period for computing the historical correlation in order to increase the potential size of the change from day to day, how would the choice of period be justified? I believe it is better to acknowledge that such nonmarket observables cannot be included in VaR analyses and that their risks should be accounted for separately through reserves and stress tests, as discussed in detail in Section 6.1.2. Another factor that some risk managers have been trying to incorporate into VaR is liquidity considerations (see Dowd 2005, Chapter 14). Rather than using overnight price moves to represent each instrument, price moves over a longer period will be used to represent less liquid instruments. If this is not handled carefully, it can result in underrepresentation of illiquid risks. For example, you might have a short position in a very liquid government bond and a smaller long position in a less liquid corporate bond. If you compute VaR based on a one‐day move for the government bond and VaR and Stress Testing 187 a two‐day move for the corporate bond, this could show less risk than a one‐day move for both, since the larger moves for the corporate bond have the same effect in the computation as increasing the size of the position. A better approach is to separately calculate a liquidity penalty, as an add‐on to VaR, for the cost of exiting less liquid positions, using a formula similar to that proposed for liquidity reserves in Section 6.1.4. 7.1.2 Measures of the P&L Distribution Simulation is ideally suited to producing full P&L distributions, since individual cases are simulated and probabilities assigned to each case. While the full distribution can be represented graphically, for example by a histogram like that in Figure 7.1, some type of summary statistics are desirable to convey information succinctly. In practice, the primary focus has been on producing a single summary measure, the percentile loss. For example, the VaR at the 99th percentile would be the amount of loss that will be equaled or exceeded only 1 percent of the time. While less well known, another summary measure that is very useful is the shortfall VaR, which is the average loss conditional on being beyond a given percentile. For example, the shortfall VaR at the 99th percentile is the probability‐weighted average of all losses greater than the VaR at the 99th percentile. Computation of both VaR and shortfall VaR at any selected percentile is very direct from a simulation. If we have simulated 1,000 equally probable P&Ls, we only need to sort them. The 990th P&L in the sort is the 99th percentile VaR. the average of the 991st P&L through 1,000th P&L in the sort is the 99th percentile VaR shortfall. The VaR spreadsheet on the book’s website demonstrates this calculation for both historical and Monte Carlo simulation. Despite the VaR measure being better known than the shortfall VaR measure, the latter has several advantages that recommend it as a superior summary statistic. The advantages are: ■ ■ Shortfall VaR is sensitive to the entire tail of the distribution, whereas VaR will not change even if there are large increases in some of the losses beyond the cutoff percentile at which the VaR is being measured. This can be quite dangerous if it encourages businesses to tailor products to produce risks that escape the VaR measure by being too far out in the tail. In practice, shortfall VaR has proved a more stable measure than VaR in showing less sensitivity to data errors and less day‐to‐day movement due to seemingly irrelevant changes in input data. Presumably, this is due to a greater tendency to average out the noise in the data. 188 FINANCIAL RISK MANAGEMENT TABLE 7.1 Negative Portfolio Effects Third worst case for A Second worst case for A First worst case for A Third worst case for B Second worst case for B First worst case for B 99th percentile (third worst case) ■ Portfolio A Portfolio B Combined Portfolio A and B –20 million –25 million –30 million –7 million –10 million +5 million +10 million –17 million –10 million –20 million –40 million –60 million –10 million –42 million –40 million –27 million –50 million –55 million –20 million –20 million –42 million With VaR, apparently negative diversification effects can arise, as shown in Table 7.1, in which the 99th percentile of the combined portfolios, a loss of 42 million, is greater than the sum of the 99th percentile losses in the two separate portfolios, 20 million + 20 million = 40 million. Shortfall VaR never displays negative diversification effects. Negative portfolio effects are undesirable both from the standpoint of clarity of exposition, when explaining risk measures to managers, and from the standpoint of control structure; even if all units of the firm are within allocated VaR risk limits, the firm itself may be outside its risk limits. Negative portfolio effects are associated with risk measures that have been termed incoherentt in the terminology of Artzner et al. (1997). By contrast, shortfall VaR and stress scenario measures are coherent and so cannot have negative diversification effects. Dowd (2005, Section 2.3) has a good discussion of coherent risk measures in general and shortfall VaR in particular, though Dowd uses the terminology expected shortfalll (ES) instead of shortfall VaR. Given these drawbacks of VaR, why has it been so widely adopted as a risk measure? The real question senior managers and regulators would like to ask is “What is the worst loss that can possibly occur?” But this is a question that does not admit a concrete answer, so a confidence interval needs to be specified, which presumably leads to questions like “What is the worst loss that will happen no more than 1 percent of the time?” This is the question to which VaR is the answer. But it seems doubtful that management really wishes to convey indifference to the size of the losses beyond this threshold. And my experience confirms that there is a very real danger that traders and product structurers will interpret a fixed VaR threshold as an invitation to hide risk in the tails—deliberately create positions or design products that result in low‐probability risks that are just beyond the VaR and Stress Testing 189 threshold and so show up in VaR reports as having no risk. No risk translates to no risk capital charge, and even a small return on a position that attracts no capital charge can look attractive to some front‐office personnel. Such extreme tail risks are often quite illiquid and should, in any case, attract a capital charge on grounds of illiquidity, but sending the right signal through VaR is also constructive. Based on these considerations, I would recommend shortfall VaR as a more desirable summary statistic. If management or regulators still wish to know the VaR, then I would recommend estimating it by a properly selected shortfall VaR. For example, a good estimate of the 99th percentile VaR is the 97.6th percentile shortfall VaR. The two measures are almost exactly equal for normal distributions, and using the 97.6th percentile shortfall VaR as an estimator provides greater stability, avoids negative diversification effects, and eliminates incentives to hide risk in the tails. The VaR spreadsheet illustrates the estimation of VaR by a properly selected shortfall VaR, as detailed in the documentation for the calculation of the historical simulation VaR. When Monte Carlo simulation is utilized, all simulation runs are assigned equal probability weights, since any differences in weightings of historical data has already been taken into account in the estimation of input parameters to the simulation. But for historical simulations, if you want to assign different weights to different historical periods, you need to do it at the point at which VaR and shortfall VaR are computed, by considering the probabilities that are assigned to each simulation run. Utilizing different weights for different historical periods in historical simulation can help to overcome one of its least attractive features, the way in which the VaR calculation can shift suddenly when a particularly volatile day leaves the data set. For example, if you are using the past 1,000 days of data for your VaR calculations and June 20, 2010, was a very volatile day, VaR calculations on June 20, 2014, might include that day, and VaR calculations on June 21, 2014, and subsequent days might exclude it. If you assign weights to historical periods with a gradual drop in weights as a date becomes more distant, this shift will take place far more smoothly. Dowd (2005, Section 4.4) has a good discussion of this issue and of a variety of reasonable weighting schemes to consider. If you want to use simulation results to project possible extreme results (i.e., at very large percentiles), then you need to extrapolate beyond the historical data set. For example, if you want to produce a VaR or shortfall VaR at 99.99 percent, you need to forecast what will happen 1 out of every 10,000 days. But you will almost certainly be working with far less than 10,000 days of historical data. We will discuss later, in Section 7.3, the reasonableness of calculating such extreme measures, but for now, let’s see how it can be done if needed. 190 FINANCIAL RISK MANAGEMENT Extrapolation beyond the historical data set requires statistical tools from extreme value theory (EVT). A very brief summary of the principal EVT techniques most often used in VaR analysis appears in the box. KEY RESULTS FROM EVT The results from EVT that are most often used in portfolio risk measurement are estimates for VaR and shortfall VaR at percentiles far out on the tail of the distribution. For example, you can find the formulas for these estimates along with derivations as numbers (6) and (10), respectively, in McNeil (2000). I will state them in slightly altered notation, which is designed to make them easier to utilize in a standard VaR framework. Let VaRp and ESp stand for the VaR and shortfall VaR at any given percentile p. Let u be a percentile at which we can directly measure VaRu by standard simulation. The formulas are: VaRp = VaRu + (β/ξ){[(1 – p)/(1– u)]−ξ – 1} ESp = (VaRp + β – ξVaRu)/(1 – ξ) The estimation procedure requires a choice of a base percentile u as well as a choice of the parameters β and ξ. A good discussion of the most frequently used methods for determining these parameters and how much confidence may be placed in the estimation procedure can be found in Diebold, Schuermann, and Stroughair (2000). An example using these formulas can be found in the EVT spreadsheet. Dowd (2005) has a good discussion of the application of EVT methods to portfolio risk measurement in Chapter 7, with derivation of these formulas and examination of parameter estimation in Section 7.2. Dowd, in Section 7.1.2, also provides a shortcut version of EVT that can be used as a first approximation. Schachter (2001) also has a good presentation of this material. There are many issues with the use of EVT, such as the need to make assumptions that are nearly impossible to test and the difficulty in estimating parameters. But its virtue is that, if such data extrapolations need to be made, it provides a smooth and consistent methodology that is superior to the alternative of extrapolating based on empirical curve fitting. A brief and lively discussion of these issues with plentiful references can be found in Embrechts (2000). As Embrechts indicates in this article, EVT is even more VaR and Stress Testing 191 problematic when used with high‐dimensional data, which combines in a nonlinear fashion. This is a good description of VaR of a large firm’s portfolio, with options valuation providing the nonlinearity. So direct application of EVT to the VaR measure for the portfolio is highly questionable; for a similar critique of applying EVT to VaR, see the section in Schachter (2001) titled “EVT Is No Panacea Either.” More reasonable is application of EVT to each individual input variable in a Monte Carlo simulation, combined with as much structural modeling of correlation as possible. This approach will be discussed in Section 7.2.3. As with any model, a VaR model needs to have its predictions tested against real results to see if it is sufficiently accurate. This process is someg, since you are looking back to see how the times known as back‐testing model would have performed in the recent past. It has been particularly emphasized for VaR models, owing to insistence by regulators that if firms are to be allowed to use internally built models for calculation of regulatory capital, they must be able to demonstrate that the models fit real results. The suggested regulatory back‐test is a straightforward comparison between the 99th percentile produced by a VaR model on each day during a specified period (since it is this percentile that determines regulatory capital) and the actual P&L on each day. The model is considered satisfactory (or at least erring acceptably on the side of too much capital) if the number of days on which P&L exceeds the predicted 99th percentile is not statistically significantly greater than 1 percent. While this approach has the virtue of simplicity, it is statistically quite a blunt instrument. Much more information can be extracted by comparing VaR projections to actual results at many different percentiles. More sophisticated methods for back‐testing are very well presented in Chapter 15 of Dowd (2005). Chapter 6 of Jorion (2007) also covers some alternative back‐testing methods, with particular emphasis on how VaR interacts with the Basel capital rules. A methodological question is whether to back‐test against actual reported P&L or against P&L that has been adjusted for components that the VaR cannot reasonably be expected to pick up. Such components are revenue from newly booked transactions, revenue from intraday or (when running VaR for periods longer than a day) intraperiod trading, and gains or losses due to operational error (e.g., trades incorrectly booked). The argument in favor or using unadjusted P&L in the comparison, besides simplicity of computation, is that these are all real components of P&L that can be quite difficult to identify, so it is better to be aware of the extent to which your model is underpredicting actual reported loss events. An argument in favor of making at least the largest adjustments is that without getting the target data to line up with the forecasting process, you are working with a suboptimal diagnostic tool. 192 7.2 FINANCIAL RISK MANAGEMENT STRESS TESTING 7.2.1 Overview As stated in Section 6.1.1, risk assessment must include an evaluation of the potential impact of a period of severely reduced liquidity, stress tests for short. There are two fundamental approaches that have been proposed to performing stress tests: reliance on historical data and reliance on economic insight. I will argue that strict reliance on historical data is not a viable option—economic insight must be utilized. But I will also argue that economic insight can usefully be supplemented by historical data. From a computational standpoint, stress testing is simply another variant of simulation; it just uses a different method to generate the scenarios of underlying market variables. But after that, the other two steps in simulation analysis—translation to all market variables and calculation of firm P&L—can be carried out exactly as per simulation VaR; indeed, the exact same system can be used for both. As we will see, the use of economic insight requires a great deal of extra effort and introduces a substantial amount of subjective judgment. So why bother departing from statistics? Couldn’t we just rely on Monte Carlo simulation based on historical data to generate highly unlikely but still plausible scenarios? The answer is clearly no, for several reasons: ■ ■ ■ The distribution of market moves in a crisis event may not resemble the distribution of market moves in normal market circumstances. Experience indicates that you cannot safely assume that market moves in a crisis event simply represent extreme values of ordinary market distributions. In particular, correlations often swing to extreme values in a crisis (Dowd 2005, introduction to Chapter 13). For example, in a flight to quality triggered by a major credit scare, otherwise uncorrelated asset prices may move sharply down at the same time. Some scenarios represent such sharp breaks with history that no analysis of past experience can offer a complete story. Economic forecasting based on hard‐to‐quantify judgment is required. When firms were worried in 1999 about the potential impact on the financial markets of the Y2K systems bug, no purely historical analysis could offer any guidance. When many of the nations of Europe adopted a common currency, a scenario based on the possible collapse of that currency could not be based on any clear historical precedents. Some scenarios do not relate to public price observations at all, so cannot be based on historical records of price changes. If a firm has an inventory of options on stock baskets whose pricing depends on VaR and Stress Testing ■ ■ ■ ■ 193 long‐term correlations for which no liquid public prices exist, a scenario for a market event that would cause the portfolio to be revalued must be formed based on market knowledge. For example, a wave of mergers might drive up the input level of correlations used in valuations. Both the judgments about how plausible a given level of merger activity might be and how much this might impact the firm’s internal valuation policies must be based on the knowledge and experience of individuals. Many scenarios require judgment about the impact of large declines in market liquidity that often accompany extreme price moves. Record keeping on price liquidity is extremely sparse relative to record keeping on price levels, so it is doubtful that any such scenario could be constructed based on historical statistics. To deal with this limitation, it is generally necessary to estimate the length of time it will take to liquidate a position in a crisis. Since this bears little resemblance to the time it takes to liquidate a position in normal circumstances, it requires an analysis completely independent from that which goes into VaR calculations. Some scenarios focus on the plausibility of contagion (chain reactions of changes in one market spilling over into other markets through investor behavior). An example may be fear that a stock market crash will spur sales of bonds by firms needing to meet margin calls. Refer back to the discussion in connection with Long‐Term Capital Management in Section 4.2.1. Such scenarios must be constructed based on knowledge of the current composition of investor portfolios. Historical statistical analysis is likely to be of limited value. Some scenarios need to emphasize the interactions among market risk, credit risk, funding liquidity risk, and reputational risk (see Basel Committee on Banking Supervision 2009a, Principles for Banks 10 and 14). Historical data will be of little use here. What is required is economic insight based on thorough examination of previous stress periods and creative thinking about similarities between what has occurred in the past and current economic and institutional circumstances. As emphasized in Sections 1.3 and 6.1.1, when attempting to estimate low‐probability events, it is important to include subjective judgment. Estimating the impact of infrequent episodes of diminished liquidity is a paradigm of estimating low‐probability events. Use of scenarios based on economic insight is a systematic way to ensure that subjective judgment is utilized. 7.2.2 Economic Scenario Stress Tests The use of economic insight may be necessary for stress testing, but it does pose difficulties. 194 FINANCIAL RISK MANAGEMENT Working out plausible combinations of the entire set of underlying variables that can impact a large firm’s trading position is hard work and requires a lot of attention to detail. While in principle subjective probability judgments could be used to specify probabilities for scenarios, once we leave the realm of historical distributions, different people are likely to have wide differences in subjective probabilities that are difficult to reconcile. In practice, a standard of plausibility is substituted for one of probability, and plausibility is a very subjective notion. But, however subjective, plausibility must still be insisted upon. Without such a standard, stress testing becomes equivalent to the child’s (and childish) game, “Who can name the largest number?” No one ever wins, because one can always be added to the last number. And you can always specify a stress test that is one shade more extreme than the last one specified. Here are some points that should be considered in scenario generation to try to deal with both the amount of effort involved and the degree of subjectivity. ■ ■ One aid is to split the work up between a senior group that determines a global scenario for the most important variables and specialist groups that work out the consequences of that global scenario for less important variables. Global scenarios generally reflect major shifts in economic conditions: a stock market crash, an oil embargo, a series of large credit defaults. It is important to be sure that splitting the work among specialist groups does not allow inconsistent relationships to develop in the overall scenario. For example, if one group develops the government bond yield curve and another group develops the AAA‐rated corporate bond curve, you don’t want there to be any tenors at which the government bond yield is higher than the AAA corporate bond yield. This can be avoided by having the second group develop a curve for the spreads between AAA corporate bonds and government bonds, rather than developing a curve for the absolute level of AAA corporate bond yields. Schachter (2001), in the section on “Implementing Useful Stress Tests,” has many valuable suggestions along these lines, including: ■ Using proportional shocks rather than absolute shocks for volatilities, to avoid the possibility of specifying negative volatilities. ■ Specifying shocks to yield curve shape and to volatility surface shape, rather than individual shocks to each interest rate and volatility, to avoid unreasonable shapes. ■ Checking that arbitrage relationships, such as cost of carry relationships between cash and futures prices, are maintained. VaR and Stress Testing ■ ■ ■ ■ ■ 195 Given the difficulty of developing hypothetical scenarios, it is unreasonable to think that more than a handful (say between 10 and 20) can be in use at any one time. Given all the potential combinations of events in markets, it is important to focus on those possibilities that are most significant to the types of positions your firm generally holds. Anchoring the assumptions for the move of a particular variable to the largest move previously observed historically is a good preventative against playing the “Who can name the largest number?” game and overcoming some of the inherent subjectivity. But care should be taken to consider a broad enough range of evidence. For example, if the largest previous daily decline in one country’s broad stock market index has been 10 percent and that of the stock index in another country with a similar level of economic development has been 15 percent, there is a presumption in favor of using 15 percent as a historical worst case for both. Acknowledging the need for subjectivity and plausibility rather than probability must never be used as an excuse for just utilizing the opinions of a narrowly drawn group. In fact, subjectivity and plausibility are strong markers of the desirability and necessity of considering a wide range of viewpoints. When you encounter (or, even better, seek out) a view with which you strongly disagree but that is backed by reasonable arguments, you need to take it into account. If you were just producing a most likely scenario or deciding on expected return, you would need to finally rely on your best judgment and not on views you strongly disagreed with. But a search for plausibility must cast a wider net, and you can easily include views you don’t agree with as being improbable but still having a small probability of occurring, and so worthy of consideration when degree of protection is being measured. See Section 5.2.5.7 for a specific illustration. The most important choices are always about which variables can plausibly move together, not about the size of moves. History can be some guide, particularly experience in prior large moves; history of statistical correlations is virtually worthless. It is important to consider linkages that are caused by investors as well as linkages caused by economics. For example, consider the correlations experienced between seemingly unrelated markets when Long‐Term Capital Management was forced to begin liquidating its holdings. Building in such correlations requires market intelligence on the type of holdings that large institutional players may have accumulated. Large moves in variables are closely associated with market illiquidity. The size of variable moves chosen should correspond to moves that occur from the time a liquidity crisis begins to the time it ends; prices 196 ■ ■ FINANCIAL RISK MANAGEMENT recorded in between these times often have little meaning, since you can’t really do any significant size of business at those prices. Since record keeping related to market liquidity is usually sparse, choice of the starting and ending points for a liquidity crisis usually depends on the institutional memory of people involved in the trading business. One point of contention between traders on one side and risk managers and regulators on the other side is the assumption that no delta rehedging of options positions will take place during the unfolding of a stress scenario (there is a parallel contention about the same assumption when used for the largest moves seen in VaR simulation). Traders rightly point out that they often have firm rules and limits that would require them to perform a delta rehedge when underlying prices move sufficiently. However, the reason that risk managers and regulators often insist on assuming no rehedging is the fear that leak of market liquidity in a crisis will prevent rehedging from being executed successfully. Creating linkages between large market moves and related losses due to credit risk, funding liquidity risk, and reputational risk is difficult; for some guidance, Basel Committee on Banking Supervision (2009a) is a good source. A starting point could be an internal database of difficult‐ to‐quantify risk factors, as discussed in Section 8.2.6.5. Particular focus should be on situations in which: ■ Credit exposure (most usually counterparty credit exposure) is highly correlated with market prices, such as stock market levels, interest rate levels, or foreign exchange rates; or credit exposure will be impacted by a change in a counterparty’s credit rating (see further discussion in Section 14.3.4). ■ The firm’s ability to hold positions through a liquidity crisis may be impacted by actions of the firm’s creditors or by changes in accounting treatment. ■ Reputational concerns combined with large market moves may cause the firm to voluntarily take losses on positions for which the firm has no legal responsibility. It has been my experience that some of the time and effort that goes into the generation of a scenario produces little benefit and may even decrease the value of the results. Too much attention to trying to produce values for every market variable that comprises a particular scenario can be self‐ defeating. For example, suppose you start with an assumption that there will be a big drop in stock markets globally. Both historical experience with previous stock market crashes and economic insight about the responses of central banks to such events may lead to incorporating a large drop in short‐term government bond rates with this event. But historical experience VaR and Stress Testing 197 with previous crashes may show mixed results about the direction of foreign exchange rate changes, and economic insight may not offer clear guidance. To spend a lot of time arguing over which direction of exchange rate move is more plausible given the main characteristics of the scenario is unproductive. It may actually reduce the value of the scenario by choosing a direction that, given the firm’s portfolio, reduces the size of the overall P&L impact when in fact it is just as likely that exchange rates would move in the opposite direction and exacerbate the firm’s losses. One possible remedy would be to split the scenario in two: one that has exchange rates going up and one with exchange rates going down. But there may be several such choices to make, and multiplication of scenarios may quickly get out of hand. A better solution is to utilize Monte Carlo simulation on some variables to supplement the scenario analysis of the major variables. For example, the decision could be made that the stress loss would be considered the worst 16th percentile loss (roughly one standard deviation) or all cases that consist of the specified scenario levels for the major variables and a normal VaR‐type Monte Carlo simulation of the other variables. 7.2.3 Stress Tests Relying on Historical Data Supplementing hypothetical scenarios with those developed primarily on historical data is desirable for a few reasons. The intensity of effort that goes into developing a hypothetical scenario limits the number that can be used at any given time, which leaves open the possibility that some plausible large risks have been ignored. While exposures to systematic risk factors, such as a large change in stock market prices or a large shift in interest rate levels, will be captured, large exposures to idiosyncratic risk factors, such as a long position in one set of stocks and a short position in another set of stocks, are likely to show no stress exposure in generated scenarios (review Section 6.1.1 for the definition of systematic and idiosyncratic risk as used here). But such positions are subject to losses in some periods of extreme reduction in liquidity. Also, having a more methodical process in place for searching for plausible extreme events may lessen some of the concern about the subjective nature of scenario generation. We can distinguish two general approaches to forming hypothetical scenarios based on historical data: 1. A complete replay of a previous stressful event, like the 1987 stock market crash or the 1997 Asian crisis. The fact that such an event has actually occurred is a strong argument for the plausibility of a similar event occurring in the future. While there are always some arguments 198 FINANCIAL RISK MANAGEMENT along the lines of circumstances having changed so much since the time of the event to make a similar event unlikely, it should be remembered that the standard is plausibility, not probability, so arguments against reoccurrence should be fairly overwhelming in order to rule it out. The simulation process for a prior event is pretty simple: select the proper start and end dates based on when market liquidity was restored, make sure you’ve stored or have researched the historical values of the market variables, and do some artful creation of values for variables for which you don’t have historical values. For example, there was no significant liquid emerging market debt in 1987, so you have to create prices based on how emerging market debt fared in subsequent large stock market downturns. But even utilizing specific past historical events is very resource intensive in researching the needed historical data, determining appropriate start and end dates, and creating values for some variables, so the number of separate scenarios that can be considered will not be large. Idiosyncratic risk positions, such as the long‐short stock position described earlier, will still probably not have their vulnerability to liquidity crises properly measured. This shows the need for some reliance on computation methods, which will be our next topic. 2. Use of a computational approach in which a large number of scenarios is generated. This approach is much closer in spirit to VaR calculations, but focuses on trying to determine large moves outside the range of standard VaR. The rest of this section is devoted to different ideas for implementing this computational approach. It is not difficult to specify plausible large moves for individual parameters. Often these have already been specified as part of stress scenarios based on economic insight. Even when they haven’t, similar techniques to those recommended for economic scenarios can be used, looking at a long run of past historical data, but alert to larger moves that may have occurred for similar variables. This is also a good place to apply the extreme value theory (EVT) techniques outlined in Section 7.1.2, since EVT is most appropriate when applied to individual parameters. The difficult question is how to combine plausible large moves for individual variables into plausible large moves for combinations of variables. One approach is to use historical data to determine a correlation matrix, apply Monte Carlo simulations to generate a distribution of returns, and establish some probability threshold as a quantitative measure of plausibility. Another approach is to find a more mechanical rule for determining which combinations will be considered plausible. The most popular of these mechanical rules is the “factor‐push” methodology, which starts by defining VaR and Stress Testing 199 any possible combination of plausible large moves of individual variables as a plausible large move for the combination of factors. The major drawback for the Monte Carlo approach is the discomfort many risk managers feel for translating the notion of plausibility into a specific probability threshold. The major drawback of the factor‐push methodology is that assuming that all variables make a worst‐case type of move simultaneously may strain the limits of what is legitimately considered plausible. And both approaches entail significant computational challenges. In the remainder of this section we look at the specifics of these two approaches, along with some suggested variants, and see how these drawbacks might be mitigated and how the computational challenges might be met. We consider the more mechanical factor‐push approach first. 7.2.3.1 Factor‐Push Stress Tests Factor‐push stress testing involves determining a plausible maximum up move and down move for each variable, and then evaluating all possible combinations of these up and down moves. Those that produce the largest negative P&Ls become plausible stress scenarios. The advantage of this approach is that it investigates a large number of possible scenarios (2f where f is the number of factors) while requiring decision making or statistical analysis around a small number of inputs, the plausibility ranges for each factor. Dowd (2005, Section 13.3.1) provides a useful analysis of factor‐push stress testing. Two principal criticisms of factor‐push methodology have been offered. The first is that it does not follow from each individual factor move being plausible that each combination of these factor moves is plausible. This would be particularly true for closely related factors—it would be totally implausible for the two‐year Treasury rate to make its largest plausible up move while the three‐year Treasury rate is making its largest plausible down move. The second criticism of factor‐push methodology is that it assumes that worst‐case P&L always occurs at the extremes of the factor range. While true for linear products, it may not be true once options are involved (e.g., Dowd’s example of a long straddle option position where the greater the move, up or down, the greater the gain, so maximum loss occurs far from the extremes). The second criticism is easier to overcome than the first. Mechanically, it would be easy to design a Monte Carlo simulation that uniformly takes samples from all possible moves of the individual variables between the agreed plausible up and down extremes. Since all possible combinations of plausible individual moves are regarded as plausible, whatever combination shows up with the worst P&L of all these runs is regarded as a plausible worst case. In practice, this involves a very large number of runs, so a number of methods have been proposed for finding the worst case 200 FINANCIAL RISK MANAGEMENT with fewer runs under certain conditions; see Dowd (2005, Sections 13.3.2 and 13.3.3) for an introduction to maximum loss optimization and crash metrics and Breuer and Krenn (2000, Sections 2.3.2 and 2.3.3) for implementation details. Attempts to deal with the criticism that not all combinations of plausible individual moves are plausible combinations has fostered a variety of suggested approaches for selecting some combinations as plausible without relying on probabilities. For example, a simple approach would be to sum up the severity (measured by the percentage of the largest plausible moves) of all individual variable moves and create a boundary on this total beyond which a combination is considered implausible. Approaches along this line are discussed in Breuer and Csiszar (2010). 7.2.3.2 Monte Carlo Stress Tests The alternative approach is to accept the identification of plausibility with some type of probability measure. Part of the resistance to this identification is the idea that correlation relations are totally destroyed in crisis events. But, as pointed out by Kim and Finger (2000), “The well‐known tendency of correlations to change abruptly in stress events is no valid argument against the inclusion of correlations in the formulation of plausibility standards. For the plausibility standards can be based on crisis correlations as well as correlations in calmer periods.” The greater resistance is to how to identify plausibility with a specific probability level. Risk managers have good reason to resist attempts to identify numerical probability estimates with a standard of plausibility; given the lack of historical data to support estimation of such low‐probability events, it would be easy to try to override sensible caution by ridiculing the low probabilities of the events that are being guarded against. So I would suggest substituting a standard of “relative plausibility.” For example, suppose that risk managers have agreed to some plausible economic scenarios that, judged by historical data, have a 0.005% chance of occurring (since many economic scenarios are tied to large moves in a single key variable, such as a stock market crash or a spike in oil prices, it is not an unreasonable task to estimate such a probability). Then accept as plausible any losses generated by Monte Carlo simulation that have that degree of probability or greater. No one needs to concede the accuracy of the probability estimate; it is quite probable that historical data leads to severe underestimates of the true probability given the number of times markets are hit with what were declared “once in 10,000 years” events. But we are hypothesizing that events that come out with the same measure of probability based on historical data have roughly similar degrees of plausibility. Operationally, this methodology works similarly to a Monte Carlo simulation of VaR, with the exception that the parameters for individual VaR and Stress Testing 201 variables has been specified so as to include large plausible moves within the probability range that has been agreed on as the cutoff for plausibility (for example, this is easy to accomplish with a mixture of normal approaches). Correlation matrices are specified based on historical data, probably weighted toward data from crisis periods. Many cases will need to be run in order to be able to make a reasonable estimate of losses at an extreme probability level, so some form of importance sampling will be needed to keep to a reasonable usage of resources, possibly by first using quick estimates for P&L and then making more detailed estimates for only cases that have the highest preliminary loss estimate. A paper by Andrea Rafael that illustrates this methodology can be found on the website for this book. The advantage of this approach is that it does not have any of the absurd combinations of the factor‐push methodology (use of correlation matrices, even ones drawn from crisis periods, won’t allow extreme up moves in the two‐year Treasury rate along with extreme down moves in the three‐ year Treasury rate). But all positions will get stressed, including positions like one long in some stocks and short in others, at roughly similar levels of severity. It should produce a loss level as severe as or greater than most of the economic stress scenarios, since the plausibility level is directly derived from these scenarios. 7.3 USES OF OVERALL MEASURES OF FIRM POSITION RISK In an excellent article, Wilson (1998) distinguishes several possible uses of VaR: preventing embarrassing losses, setting operational risk limits, risk comparability, determination of capital adequacy, and performance measurement (see Section 3.2 of Wilson’s article). I will use Wilson’s framework, stating my own opinions on the usefulness of both VaR and stress testing for these purposes, and comparing my views to his. Certainly a major concern that firms have looked to VaR and stress testing to help mitigate is the risk of embarrassing losses such as those discussed in Chapter 4, “Financial Disasters.” I would agree with Wilson that many of those disasters are due to issues of improper controls (e.g., Barings, Allied Irish Bank) or improper valuation (e.g., Kidder Peabody, UBS) that cannot be controlled by VaR or stress testing. Improper controls and valuation lead to positions being incorrectly reported, and VaR and stress testing cannot overcome issues of deliberate or inadvertent errors in input. If you look at the disasters covered in Chapter 4, only two resulted from unexpectedly large market moves interacting with correctly reported positions: Long‐Term Capital Management and Metallgesellschaft. Even for cases like these, I share Wilson’s skepticism about the usefulness of standard VaR as 202 FINANCIAL RISK MANAGEMENT a controlling mechanism since market moves that cause losses of sufficient size to threaten a firm’s stability are generally radical departures from recent historical experience. This still leaves the possibility of using stress testing or an extreme value version of VaR as a good controlling mechanism for those embarrassing losses that are based on large market moves. For the reasons I have given in Section 7.2, I believe stress tests based on economic insight are far more likely than statistical methods to produce useful measures for controlling extreme market moves. When it comes to risk comparability, both VaR and stress offer the advantages I emphasized at the beginning of this chapter—allowing meaningful comparison and aggregation between different businesses. As Wilson states, traditional risk measures, such as value of a basis point or vega, “provide little guidance when trying to interpret the relative importance of each individual risk factor to the portfolio’s bottom line or for aggregating the different risk categories to a business unit or institution level.” The ability that VaR and stress provide to make such comparisons and aggregation, Wilson says, correctly allows an institution to gain a deeper understanding of the relative importance of its different risk positions and to gauge better its aggregate risk exposure relative to its aggregate risk appetite. VaR accomplishes these objectives by defining a common metric that can be applied universally across all risk positions or portfolios: the maximum possible loss within a known confidence interval over a given holding period. Besides being able to be applied universally across all risk categories, including market, credit, operational, and insurance risks, this metric is also expressed in units that are (or should be) meaningful at all levels of management: dollars (or pounds, francs, etc.). It therefore serves as a relevant focal point for discussing risks at all levels within the institution, creating a risk dialogue and culture that is otherwise difficult to achieve given the otherwise technical nature of the issues. Wilson’s words on this issue square very closely with my own experience. From the very first VaR runs and stress test runs our risk management group performed for Chase Manhattan, management interest was as strong or stronger in what they revealed about the relative risk of individual positions as it was in the measurement of total firm risk. Of particular interest were positions that management had regarded as relatively insignificant contributors to the firm’s risk that showed up as among the largest contributors to VaR and stress tests—small absolute position size was outweighed by large price volatility. It’s the ability of VaR and stress tests to combine position size, price volatility, and correlation with the rest of the firm’s portfolio VaR and Stress Testing 203 into a single measure, comparable across all business lines, that makes them valuable tools in conveying risk information to management. This information on relative risk of positions has many potential uses. It can provide input for management discussions with trading desks on the proper size of stop‐loss limits. It identifies business lines and positions that require extra management attention. It can be used in calculations of risk versus return in performance measurement. When there is a need to reduce risk because limits are being breached, it helps identify actions that will have the quickest impact. Given the importance of reports on the contributions of risk positions to VaR and stress tests, careful attention to the design of these reports will have large payoffs in better management processes and in appreciation of the value of the risk function. The classic work in this area remains the Goldman Sachs report “Hot Spots and Hedges” by Litterman (1997a, 1997b). Section 11.2.2 of Dowd (2005) provides a succinct précis of these ideas. Here is my take on the main points to consider: ■ ■ ■ ■ ■ Reporting is needed for several different types of decomposition— business lines and trading desks for performance measurement, trading positions that may go across trading desks for understanding of the firm’s risk structure, and to identify targets for risk reduction. Reporting needs to be able to accommodate both organization structure and highlighting of critical risks. Some reports will need to be organized in a hierarchal fashion, so that reporting matches the way management is used to thinking of the businesses. But other reports should be organized in a largest to smallest risk fashion to be sure that there is sufficient awareness of the largest risks and to facilitate risk reduction. All reports should be designed with drill‐down capability, so that risks that need extra attention can be further broken down. Ability to take quick actions to reduce risk and management understanding of risk are both enhanced by reporting risks using categorization that is meaningful to businesses and to management. The same guidance that will be given in Sections 9.2, 9.3, 9.4, 10.4, 11.4, and 13.1 for informative reporting of nonstatistical positions should be followed here. For example, VaR and stress test risk of interest rate positions should be reported by exposure to parallel shifts of the yield curve and exposure to changes in steepness of the curve as in Section 11.4. Design of optimization procedures to identify small portfolios of a few instruments that can replicate a large portion of the VaR or stress test risk is useful both as a design for a quick hedge and as a way to convey an intuitive understanding of the major components of the firm’s position. 204 FINANCIAL RISK MANAGEMENT In reporting the contribution of product lines, trading desks, and risk components to overall firm risk, several approaches must be considered: ■ ■ ■ Each component can be represented by the scenario risk measure it would have as a stand‐alone portfolio. This is the easiest approach to implement and certainly gives a good indicator of relative risk, but fails to capture any correlation effects with other risk components that contribute to overall firm risk. Each component can be represented by the impact on total firm risk the full elimination of that risk component would have. This captures correlation effects, but may be unrealistic in that full elimination of a business line may not be a feasible alternative. Each component can be represented by its marginal impact on total firm risk. This captures correlation effects and gives a good measure of the immediate impact on firm risk of adding to or offsetting some of a component’s risk, but it is very dependent on the current mixture of risk components. A very risky business line may get represented as having a small contribution to risk just because it has low correlation with the current mix of risk for the firm. It may be best to use a stand‐alone risk measure in conjunction with a marginal impact measure to make sure that components that can potentially make large contributions to risk receive timely management focus. The marginal impact measure has a nice side benefit—when you take the weighted sum of marginal impact, weighted by current positions, you get the total risk measure for the firm. Compare the discussion here with Dowd (2005, Section 11.2.1)—note that Dowd uses the terminology component VaR for what I am calling marginal impact. This makes the marginal impact a convenient tool for exercises such as allocation to business line of firm capital where you need the sum of the parts to equal the whole. In order to have this property, a risk measure need only satisfy the condition that it scales directly with position size; that is, a position that has the same composition but is k times as large has a risk measure k times as large as the original position. This homogeneity condition is clearly met by both VaR and stress testing measures. To see that the weighted by position sum of marginal impacts equals total risk, first write the risk measure of the portfolio as R(x1, x2, . . ., xn) where xi is a component of the portfolio. By hypothesis, R(kx1, kx2, . . ., kxn) = kR(x1, x2, . . ., xn). Taking the derivative of both sides with respect to k, the left‐hand side by the chain rule, we obtain: ∑ xi i ∂R(kx kx1 , kx2 , ∂kxi , kx k n) = R(x1 , x2 ,. . ., n) 205 VaR and Stress Testing Setting k = 1, ∑ xi i ∂R(x1 , x2 , ∂xi , xn ) = R(x1 , x2 , , xn ) which states that the sum of the marginal impacts weighted by position equals total risk. Given this ability to place different risks on a common footing, it is quite natural to want to place limits on businesses based on VaR and stress scenario losses. Stress scenario losses offer the added benefit of controlling against at least some forms of financial disaster. However, this does not provide a complete solution to control of a trading business, and other (nonstatistical) limits are needed as well. Wilson emphasizes speed of calculation and ease of understanding and communication as the reasons for needing other limits besides VaR. I would emphasize, as in Section 6.2, the need to match position taking to expertise and to assure adequate diversity of trading style. A supplement to the use of limits to control risk is the provision of an adequate capital cushion against potential losses. This cushion is required for both earnings volatility and market moves. Earnings volatility measurement aligns well with VaR, while the impact of large market moves is a risk better measured by stress scenarios. While I believe this to be a sound argument for basing internal measures of capital adequacy on both VaR and stress loss, regulators have strongly favored VaR as the measure on which to base capital required for regulatory purposes. Since capital required for regulatory purposes can have a direct impact on the firm’s stock price performance, regulators have been wary of any tie to a measure such as stress, which directly relies on human judgment, for fear that management will manipulate it. VaR has been viewed as preferable based on the relative difficulty of manipulating a statistical measure. VaR is viewed as at least capturing relative differences in level of risk. Translation into a required capital cushion against large, unexpected moves is then approximated through multiplication by an essentially arbitrary constant. For a more detailed discussion of the regulatory capital standards revolving around VaR, see Chapter 3 of Jorion (2007). For performance measurement, the critical objective is to have a means of adjusting the P&L performance of the firm and of business units for the level of risk taken in achieving this performance. As with the capital cushion, the risk taken is both a function of earnings volatility and of vulnerability to unexpectedly large market moves, arguing for using a mix of VaR and stress loss in developing this measure. But the subjectivity of stress scenarios, combined with the sole reliance of regulatory capital on VaR, has led almost all firms to the decision to base this risk measure completely on VaR. The 206 FINANCIAL RISK MANAGEMENT firm where I have worked for the past several years, Chase Manhattan (now JPMorgan Chase), has been very unusual in utilizing both VaR and stress in this measure. I will relate some of the history that led Chase management to conclude that stress loss was worth utilizing despite the disputes between the central risk management group and business units, which are inevitable when experience and judgment are significant determinants of a performance measure. When the Asian credit crisis of the fall of 1997 started to spread to other emerging market economies, we noticed that the losses being experienced by Chase trading desks very closely matched the projections of the hypothetical flight‐to‐quality stress scenario we had constructed. The match was not just for the firm as a whole but for individual business units. This experience persuaded management to experiment with tying the risk adjustment of business units relative to stress losses, as an incentive to reduce vulnerability to large market shocks. As business adjusted to the new performance measure in early 1998, we noticed a significant impact in terms of strategies to continue to meet P&L targets with less reliance on positions that were vulnerable to these shocks. The result was that Chase weathered the fall 1998 market shock due to the Russian default and the unraveling of Long‐Term Capital Management with much smaller losses than in the fall 1997 crisis and smaller losses than almost all of our largest competitors (see O’Brien 1999). Continued experience with the impact of this decision since then has continued to confirm its value. The mechanisms for adjusting P&L return for risk, which include calculating risk‐adjusted return on capital (RAROC) and shareholder value added (SVA), are not topics addressed in this book. Interested readers are referred to Chapters 20 and 21 of Culp (2001) and Chapter 16 of Jorion (2007). EXERCISES 7.1 Vaule‐at‐risk computations Using the data in the VaR spreadsheet (with equal weights on all days) and a 10 percent position in each of the 10 variables, calculate the 99th percentile VaR using the following five methods: 1. Variance‐covariance. 2. Historical simulation using a single‐point estimate of the 99th percentile. VaR and Stress Testing 207 3. Historical simulation using 2.33 × the standard deviation of the daily total portfolio valuations as the 99th percentile. 4. Historical simulation using a single point estimator of the 99th percentile and substituting the historical volatility over the most recent 100 business days for the historical volatility over the full data set, but using the full data set to simulate results. 5. A Monte Carlo simulation. Your answers to 1, 3, and 5 should be very close to equal. Why? What does this tell you about the relative ease of implementation of the three methods? 7.2 Maximizing diversification Try the same exercise as in 7.1 with a combination of investment percentages that you choose yourself. Can you find a combination without any short positions (all investment percentages positive) that gives a high diversification benefit (cell D24 of the Var‐Cov VaR worksheet in the VaR spreadsheet)? 7.3 Measuring fat tails in historical data Look at the Ratios worksheet in the VaR spreadsheet. What does it tell you about how fat tailed the time series used in these calculations is? At what percentile level do you begin to see a significant impact of the fat tails? 7.4 Generating fat tails in Monte Carlo simulations Experiment with the MixtureOfNormals spreadsheet and see how different selections of input parameters produce different degrees of kurtosis and clustering of large changes. CHAPTER 8 Model Risk ny book on financial risk management needs to address the subject of model risk, the risk that theoretical models used in pricing, trading, hedging, and estimating risk will turn out to produce misleading results. This book, which emphasizes quantitative reasoning in risk management, pays particularly close attention to how models can be used and misused in the risk management process. Since the publication of the first edition of this book, the financial risk management focus on model risk has intensified. In the wake of the 2007– 2008 crisis, as we discuss in Sections 5.1 and 5.2.5.3, there have been accusations that model failure was one of the root causes of the meltdown. When a widely discussed article has the title “Recipe for Disaster: The Formula That Killed Wall Street” (Salmon 2009), it is clear that model risk needs to be addressed with a sense of urgency. Fortunately, in addition to this sense of crisis surrounding model risk, the past several years have witnessed greater attention to analysis of how model risk can be controlled. Concise, excellent articles by Derman (2001) and Rebonato (2003) are now recognized as touchstones for the analysis of model risk. Morini (2011) is the first thorough book‐length treatment of model risk. The Federal Reserve and Office of the Comptroller of the Currency joint document for “Supervisory Guidance on Model Risk Management,” which I will reference as FRB (2011), and the Basel Committee on Banking Supervision’s “Supervisory Guidance for Assessing Banks’ Financial Instrument Fair Value Practices,” which I will reference as Basel (2009b), provide regulatory responses to the lesson of the 2008 events for model risk. I find the joint Federal Reserve/Comptroller of the Currency document to be particularly thorough and persuasive in its analysis of the many aspects of model risk. This chapter begins, in Section 8.1, with an overview focusing on the variety of opinions that have been expressed about the importance of models, or their unimportance, in managing financial risk. Section 8.2 examines A 209 210 FINANCIAL RISK MANAGEMENT the procedures that ought to be used for risk evaluation and control for models of all types. The following three sections give a more detailed analysis of model review standards, distinguishing among three types of models: those used for valuation and risk measurement of liquid instruments in Section 8.3, those used for valuation and risk measurement of illiquid instruments in Section 8.4, and those used for making trading decisions in Section 8.5. 8.1 HOW IMPORTANT IS MODEL RISK? When examining model risk, one immediately encounters a very wide range of views on the role that models can play in controlling risk and creating new risks. These vary all the way from viewing model error as the primary cause of financial risk to viewing models as largely irrelevant to risk. The view that models are largely irrelevant to risk can often be encountered among traders who view models as just convenient mathematical shorthand with no real meaning. All that really matters are the prices the shorthand stands for. A good example is the yield of bonds as calculated by Securities Industry Association standards. This includes many detailed calculations that have no theoretical justification, but can only be explained historically (for example, some parts of the calculation use linear approximations, which made sense before calculations were done on computers). No one would claim that this yield has a precise meaning—you don’t necessarily prefer owning a bond yielding 7 percent to one yielding 6.90 percent. However, you can translate between yield and precise price given the industry standard rules. It is convenient shorthand to convey approximate values. The degree to which these calculations give misleading yields hurts intuitive understanding, but does not result in mispricing. Those who view models as playing no real role in pricing and risk management view almost all models used in financial firms as playing a similar role to that of bond yield calculation. A typical claim would be that the Black‐Scholes option model, probably the model most frequently used in the financial industry, is just a mathematical convenience that provides shorthand for quoting options prices as implied volatilities rather than as cash prices. In this view, implied volatilities are an attractive way of providing quotations, both because of common usage and because they provide more intuitive comparisons than a cash price, but they should not be regarded as having any meaning beyond representing the price that they translate to using the Black‐Scholes formula. If this viewpoint is correct, models would play an extremely minimal role in controlling risk, and model testing would consist of little more Model Risk 211 than rote checking to see if industry‐standard formulas have been properly implemented. However, this extreme a view cannot explain all the ways in which trading firms use models such as Black‐Scholes. The valuation of unquoted options is derived by interpolating the implied volatilities of quoted options. The Black‐Scholes model is used to translate prices to implied volatilities for the quoted options and implied volatilities to prices for the unquoted options. The risk reports of position exposures use the Black‐Scholes model to compute the expected impact of changes in underlying prices on option prices. Scenario analyses presented to senior management quantify the impact of changes to the implied volatility surface. For more details, see Chapter 11 on managing vanilla options risk. This behavior is inconsistent with a claim that the model is being used purely to provide convenient terminology. By contrast, the industry standard bond yield formulas are not used in comparable calculations— interpolations and risk reports are based on a more sophisticated model of separately discounting the individual cash flows that constitute a bond, with a different yield applied to each cash flow. In this computation, none of the linear approximations of the industry standard formulas are utilized. For more details on these calculations, see Chapter 10 on managing forward risk. The view that models are the primary cause of financial risk is often encountered in articles describing major trading losses, which are frequently ascribed to the firm having the wrong model. What is often unclear in these claims is whether “having the wrong model” just means making incorrect forecasts about the future direction of market prices or if it means misleading the firm’s traders and managers about the nature of positions being taken. A good illustration is the discussion in Section 4.2.1 of whether the reliance by Long‐Term Capital Management (LTCM) on models should be viewed as a primary cause of the collapse of the fund. And after the 2007–2008 crisis in mortgage collateralized debt obligations (CDOs), one began to encounter claims such as “[David] Li’s Gaussian copula formula will go down in history as instrumental in causing the unfathomable losses that brought the world financial system to its knees” (Salmon 2009). Of course, once products start encountering losses, modelers who had been promoting a view of the importance of models may now wish to take the opposing view. Morini (2011, Preface) quotes modelers, speaking after the crisis, telling him “Models were not a problem. The problem was in the data and the parameters! The problem was in the application!” Morini’s response is that “Models in finance are tools to quantify prices or risks. This includes mathematical relations, a way to use data or judgment to compute the parameters, and indications on how to apply them to practical issues. 212 FINANCIAL RISK MANAGEMENT Only by taking all these things together can we talk of ‘a model.’ Modellers should stay away from the temptation to reduce models to a set of mathematical functions that can be thought of separately from the way they are specified and from the way they are applied. If this were the case, models would really be only blank mathematical boxes and people would be right to consider them useless, when not outright dangerous.” I would add that any modelers who want to separate their work from choices on data or parameters are basically saying that they are programmers. There’s nothing wrong with being a programmer—it’s a highly demanding profession. But with rare exceptions (a few people who are able to pioneer an extraordinary speedup of existing calculations), programmers are not compensated at the level modelers are and do not have the degree of influence in making decisions about innovations in products that modelers do. In the final analysis, whether model builders take the responsibility or the traders and risk managers who use them take the responsibility, models play a key role in managing risk and we must develop clear guidelines to see that the role they play is to clarify issues rather than to obscure them. This is the task to which we now turn. 8.2 MODEL RISK EVALUATION AND CONTROL In this section, we look at those procedures that ought to be used for risk evaluation and control for all types of models—those used for making trading decisions as well as those used for valuation and risk measurement. In Section 8.2.1, we discuss the scope of model review and in 8.2.2 the proper roles and responsibilities that need to be established around model review and control. In Section 8.2.3, we look at those procedures that check whether the model selected has been correctly implemented—whether the model actually performs as specified; Morini (2011) calls this model verification. In Sections 8.2.4 and 8.2.5, we examine two particularly important pieces of model verification, the verification that contractual arrangements have been correctly specified in the model and the evaluation of approximations. In Section 8.2.6, we turn to procedures that check whether the model selected is appropriate for the product or trading strategy being modeled; Morini (2011) calls this model validation. The procedures in Sections 8.2.3, 8.2.4, 8.2.5, and 8.2.6 are primarily designed for the initial evaluation of models leading up to the decision whether the model should be approved for use, and what restrictions, if any, should be placed on its use. In 8.2.7 and 8.2.8, we look at those aspects of model evaluation and control that should take place continuously or periodically during the life of the model’s use to see if any new information is Model Risk 213 available to change the initial conclusions about the model approval or to suggest model modification or replacement. 8.2.1 Scope of Model Review and Control The first point that needs to be established is what determines that something is a model that requires review and control. FRB (2011, Section III) casts the net very wide, stating that “For the purposes of this document, the term model refers to a quantitative method, system, or approach that applies statistical, economic, financial, or mathematical theories, techniques, and assumptions to process input data into quantitative estimates. . . . Models meeting this definition might be used for analyzing business strategies, informing business decisions, identifying and measuring risks, valuing exposures, instruments or positions, conducting stress testing, assessing adequacy of capital, managing client assets, measuring compliance with internal limits, maintaining the formal control apparatus of the bank, or meeting financial or regulatory reporting requirements and issuing public disclosures.” It is important that a definition this broad be used. A computation may be made by a simple formula in a spreadsheet and still give rise to as great a danger of incorrect estimation as a computation requiring a complex mathematical derivation and a supercomputer churning away for hours to produce the result. Simply averaging observed two‐ and three‐year interest rates to obtain a two‐and‐a‐half‐year interest rate already entails an assumption that requires review and control (we’ll discuss this example further in Section 8.2.6.1). The mental image many of us have of a model as a complex piece of mathematics and computer engineering can create blinders when we are looking for potential sources of model risk. A second point made in FRB (2011, Section V) is that “Vendor products should nevertheless be incorporated into a bank’s broader model risk management framework following the same principle as applied to in‐house models, although the process may be somewhat modified.” Whether a model has been created in‐house or by a vendor, the consequences of the model being incorrect still affect the profit and loss (P&L) of the firm using the model, so there should be no variation in the standards applied for model review and control. The Federal Reserve goes on to point out the challenges of reviewing vendor models since they “may not allow full access to computer coding and implementation detail” and there is a need for “contingency plans for instances when the vendor model is no longer available or cannot be supported by the vendor.” The model review procedures of this chapter can be used for vendor models, but I have encountered instances where a vendor model is so opaque that I have needed to insist that it be replaced 214 FINANCIAL RISK MANAGEMENT by an in‐house model or by another vendor model that permitted more transparency. After establishing the scope of the definition of a model, the next step is to agree on what needs to be included in a review. Some key points from FRB (2011, Section III): ■ ■ ■ ■ ■ “Models are of necessity simplified representations of real‐world relationships and so can never be perfect.” As a result, model use invariably results in model risk, which can be defined as “financial loss, poor business and strategic decision making, or damage to a bank’s reputation” based on “incorrect or misused model outputs and reports.” Model risk can result from either fundamental errors in the model or inappropriate use of a model, particularly the use of a model outside the environment for which it was designed. “Model risk should be managed like other types of risk. Banks should identify the source of risk and assess the magnitude. Model risk increases with greater model complexity, higher uncertainty about inputs and assumptions, broader use, and larger potential impact.” The intensity and rigor of model reviews need to be matched to the degree of model risk identified. Model risk cannot be eliminated, so it needs to be controlled through limits on model use, monitoring of model performance, adjusting or revising models over time, and informed conservatism in inputs, design, and outputs. But while conservatism “can be an effective tool” it cannot be “an excuse to avoid improving models.” 8.2.2 Roles and Responsibilities for Model Review and Control I have been involved with the design and approval of several firmwide model review policies. In every case, I have insisted on a prominent statement that “Risk management serves as a second set of eyes for model review.” This means that the business unit that develops and utilizes the model has first responsibility for reviewing the model and assessing its risks. The role of the risk management function is very important in ensuring that an independent unit without insider incentives reviews the model and in creating a uniform model review environment throughout the firm. But the knowledge that an independent review will be performed by risk management cannot be used by the business unit as an excuse for not performing its own thorough review. Having two sets of eyes reviewing the model is important both for providing an extra layer of security and in obtaining the benefit of insider Model Risk 215 product expertise to complement outsider independence and model review process expertise. FRB (2011) supports this viewpoint on business unit responsibility. In Section VI it states: “Business units are generally responsible for the model risk associated with their business strategies. The role of model owner involves ultimate accountability for model use and performance. . . . Model owners should be responsible for ensuring that models are properly developed, implemented, and used . . . [and] have undergone appropriate validation and approval processes.” Just stating that business units have responsibility for model review is only the first step. Incentives need to be properly aligned to make sure this responsibility is taken seriously. Steps to assure this include: ■ ■ ■ ■ The business unit responsibility for model review is just as much about clear communication as it is about clear thinking. As Morini (2011, Section 1.4.1) emphasizes strongly: “The choice of a valuation model must be based on an analysis . . . reported to senior management in an aggregated and understandable form.” “Quants, traders, and other technically strong practitioners” must find ways to communicate technical ideas in nontechnical language “comprehensible for senior management.” Technically strong practitioners who have difficulty in doing this should seek help from colleagues who have stronger communication skills or from corporate risk management personnel who have more experience in this aspect of model review. But no one should be under the illusion that they will escape responsibility for consequences because “the senior guys just weren’t capable of understanding what we were doing.” If you truly can’t get senior managers to understand the potential consequences, even with renewed effort at clever communications, then this is a product your business unit should not be trading. A clear distinction should be made between losses due to market uncertainties that were clearly identified and advertised as part of the business unit’s model review and losses due to market uncertainties that were ignored in the model review. The latter should have more serious consequences for performance review and compensation than the former, and this policy should be widely advertised within the firm. To make sure that the policy in the previous bullet point is successfully implemented, an analysis of significant trading losses needs to be conducted by control personnel independent of the business unit to determine how losses are related to model reviews. The independent model review conducted by the risk management area should include identification of weaknesses in the business unit 216 FINANCIAL RISK MANAGEMENT model review. Patterns of weaknesses need to be addressed by corrective action, as well as consequences for performance review and compensation. I have not addressed here the issue of how business unit model responsibility should be divided between model builders and traders. This will have different solutions for different business units and different models. I will just note again my comments in Section 8.1 that any function seeking to shun responsibility for model error should accept that reduced responsibility and reduced compensation opportunities go hand in hand. This emphasis on business unit accountability for model review is consistent with placing the main responsibility for model development with the business unit and allowing them as much freedom in structuring models as possible. Models need to take advantage of as much inside information, in the form of trader beliefs about the future, as possible. Firms must try to be open to as many trading ideas as possible and not dismiss ideas on the grounds that they do not line up with some approved theory (for example, rational expectations or marketplace efficiency). However, a culling process must also be available for measuring the success of trading ideas and eliminating those ideas that are not proving successful. Insiders should be given latitude in the theories used in deciding how to trade, but not in the theories used in deciding when to recognize P&L. Profits should not be booked and bonuses not be paid out until the forecasts of the trading models have proven correct. For decisions on when to book P&L, it is better to rely on outsiders to avoid bias. You may lose accuracy by not having access to the insiders’ market knowledge, but this will only result in delays in recognizing earnings, which is not as serious a problem as taking the wrong positions. Insiders may object that this delay in recognizing P&L will cause them to turn away good business, but they have two alternatives: find others in the market who share their opinions and sell off the risk recognizing the profits, or, if they are sufficiently confident, wait to recognize the P&L until after the risk position has matured. The risk management units that are part of control functions and that constitute the independent “second set of eyes” in model review do not have the business units’ incentive issues. It is very clear to them that unidentified model risks that lead to trading losses will cost them in compensation and may cost them their jobs and even their careers. The incentive problem here runs in the opposite direction: the negative consequences of approving a model that later proves defective are so clear that there is a danger of playing it safe by creating unreasonably high barriers to model approval. After all, a rejected model will never have a chance to show how it would Model Risk 217 have performed, so there might appear to be little danger in being proved wrong by being overly cautious. Fortunately, in most firms, business units will press their case with sufficient passion and sound analysis to overcome such unreasonable barriers to new business. But managers of independent model reviews need to be alert to this temptation toward caution, and need to be constantly challenging model reviewers to make sure the right balance is being struck. The problems for independent risk managers are more likely to rest with issues of expertise and access than they are with issues of incentive. As outsiders, they have less chance to build up the thorough knowledge of models and markets that business units possess. This can sometimes be addressed by employing former model builders as independent reviewers, but this is a career move that appeals to only a small subset of model builders. Other techniques for trying to overcome this gap in expertise will be addressed throughout the remainder of this chapter and this book. For issues of access, rules need to be put in place and enforced to see that business units are forced to share model code, documentation, and supporting data with independent reviewers. Claims of need for secrecy to protect proprietary model features must be viewed with suspicion—these are often just excuses to try to avoid independent scrutiny. When found legitimate, such claims need to be addressed by controls that restrict access to only those actually directly involved in the independent review; they must never be used as a reason to limit the scope of the independent review. Independent model reviewers need to clearly identify steps that need to be taken when they find issues with models. As FRB (2011, Section VI) states, “Control staff should have the authority to restrict the use of models and monitor any limits on model usage. While they may grant exceptions to typical procedures of model validation on a temporary basis, that authority should be subject to other control mechanisms, such as timelines for completing validation work and limits on model use.” In all cases where follow‐up action is called for, there should be definite dates for further review established and a well‐organized procedure for making certain that a follow‐up review is performed evaluating these follow‐up actions. The role just specified for independent model review is consistent with the guiding principle of FRB (2011, Section III) of “effective challenge” of models—“critical analysis by objective, informed parties who can identify model limitations and assumptions and produce appropriate changes.” Requirements for effective challenge outlined by the Federal Reserve are separation of the challenge from the model development process, knowledge and modeling skills adequate to conduct appropriate analysis and critique, and sufficient “influence to ensure that actions are taken to address model issues.” 218 FINANCIAL RISK MANAGEMENT To what extent should external resources (i.e., consultants) be used as part of the independent model review process? There are many reasons for wanting to minimize the use of external resources: the desire for confidentiality of proprietary models, the desire to build up in‐house expertise through the experience gained by conducting model reviews, and the fears of discontinuity if an external resource becomes unavailable or proves unsatisfactory. There still may be times when use of external resources is desirable, either because of a lack of in‐house expertise within the independent model review group or because manpower available is not sufficient to meet the demand of new models needing review. When external resources are utilized, care should be taken that a designated in‐house reviewer becomes as familiar as possible with the work of the consultant. This serves the function of acquiring some in‐house expertise that can be utilized in subsequent model reviews, as well as having someone in‐house who can monitor and coordinate the work of the consultant, provide a point of contact for subsequent discussion of the consultant’s work, and be able to ramp up involvement in case of discontinuity. (Compare the discussion in this paragraph with the segment headed “External Resources” in FRB [2011, Section VI]). Finally, all model review activities, both those of business units and those of independent reviewers, must be properly documented. This covers both documentation of the model itself and of the model review process. Model developers are often anxious to get on to the next project, and business units are anxious to develop the next model; model documentation can get shortchanged in the process. Poorly documented models are likely to cost money in the long run, by making model revisions more difficult and time‐consuming and by increasing the likelihood that model errors will be missed. If necessary, business units may need to establish separate model documentation teams to complete documentation based on interviews with model developers. Standards for documentation of models and model reviews should include the following: ■ ■ ■ A review of the adequacy of business unit documentation by the independent model reviewer should be included, with recommendations for gaps that need to be remedied. “Documentation and tracking of activities surrounding model development, implementation, use and validation are needed to provide a record that makes compliance with process transparent” (FRB 2011, Section VI). An inventory of all models that require review should be maintained, both by business unit and firmwide. This inventory can serve as a central control point for scheduling model reviews, keeping track of Model Risk ■ 219 documentation, providing information on contacts, and scheduling updates of model reviews (FRB 2011, Section VI, “Model Inventory”). Documentation is also required for the policies governing model review and the roles and responsibilities of business units and independent reviewers (FRB 2011, Section VI, “Policies and Procedures”). What is the role for the senior management of the firm in model review? FRB (2011, Section VI) emphasizes senior management responsibility for assuring that a process is in place that meets the standards outlined in this section and in Section 8.2.1. This certainly includes providing adequate funding for these functions. Basel (2009b) has a similar statement in its Principle 1 and, in Principle 2, emphasizes that the review capacity has to be adequate to handle conditions of stress. This is obviously a response to the stressed conditions of the 2008 crisis. In addition to these more formal requirements, senior management must be prepared to understand the aggregate level of model risk that the firm is exposed to and to set limits on this risk. Having a requirement that business units communicate model risk in an aggregated and comprehensible form to senior managers, as we have at the start of this section, entails a corresponding responsibility of senior managers to make use of this information. There will, of course, be cases of conflicting presentation to senior management—often a more sanguine view of risk from the business unit and a more cautious view from the risk management groups. Senior managers must insist that both sides get a fair hearing, preferably in the same room at the same time, and that arguments be presented in a comprehensible manner. At the end of the process, it is senior management that owns the risk and must reach a decision. Boards of directors in principle exercise an oversight role over senior management in controlling model risk, as in all other critical aspects of the business. In practice, it is very difficult for directors, whose involvement is only for a small part of each month, to have much impact, particularly since their access to information is often tightly controlled by senior management. The one course of action I would recommend is that directors on the risk committee of the board insist on private meetings with senior risk management personnel. This will at least provide a forum for concerns to be expressed that could give directors enough information to pose questions to senior management. 8.2.3 Model Verification Most model problems are related to the fit between the product or trading strategy being modeled and the model selected. This issue of model validation 220 FINANCIAL RISK MANAGEMENT we will address in Section 8.2.6. Here we deal with the simpler question of whether the model selected has actually been properly implemented—does the model actually do what it claims to do? This can be controlled by adequate model documentation and thorough checking by competent reviewers before the model is put into production. Here are a few rules that should be borne in mind to make sure this gets done properly (more detail on these points can be found in FRB [2011, Section V] and Morini [2011, Section 1.5.1, “Model Verification”]): ■ ■ ■ ■ ■ Thorough documentation of what the model is trying to achieve, model assumptions, and derivation of formulas must be insisted on. All formula derivations should receive an independent check. Useful advice from Morini (2011, Section 1.5.1) is: “When a model is used for the first time the passages from dynamics to closed‐form formulas or the other way around should be verified. This should be the case for any new model developed by a front office quant, also for any new model that simply appears on the Internet—or in a journal. There are errors even in published literature, never be too trusty.” Systems implementation of the model should be subject to rigorous standards of documentation, change control procedures, and systems testing. The best check on an implementation is to perform an independent implementation and see if the results agree. It is tempting to cut costs by confining checking to having an independent analyst read through the documentation, equations, and code of the model builder and confirm it is correct. But it is much easier to miss an error in reading through someone else’s equations or programming code; it is much more unlikely for two analysts working independently of one another to make the same error. Whenever possible, the independent implementation used as a check should employ a different solution methodology than the implementation being tested. For example, if the implementation being tested has used a Monte Carlo simulation, the test should be made solving backwards on a tree, where this is feasible. Using different implementation methodologies reduces further the chances that the two implementations will have the same flaw. Models should be tested on degenerate cases that have known solutions. For example, a down‐and‐out call with a barrier of zero is equivalent to a vanilla call, so setting the barrier to zero in a down‐and‐out call model should produce the standard Black‐Scholes result. Other examples would be: (1) to always check that put‐call parity for European‐style options is preserved for any model used to price options, and (2) to Model Risk ■ ■ ■ ■ ■ ■ 221 always check exotic option models against known analytic solutions for flat volatility surfaces (see the introduction to Section 12.1 and Section 12.3.1). Models should be tested for their impact on VaR and stress test calculations as well as on valuation and limit calculations. Models should be tested on extreme inputs to see that they handle these cases properly. For example, interest rates much lower and much higher than have recently been experienced should be input to see that the model can produce reasonable results. For more details on this model stress testing, see Morini (2011, Section 3.1). Produce graphs of model output plotted against model inputs and explore any instances where they do not make intuitive sense. This is another good check on the model’s ability to handle extreme inputs, as per the previous bullet point. The impact of varying several inputs simultaneously should be compared to the sum of the individual impacts to check if the interactions of variable changes produce reasonable results. When a new model is replacing an existing model, a thorough benchmarking process should be used to compare results of the two models for an identical set of inputs. Model differences should be checked for reasonableness and unreasonable differences investigated. The same benchmarking standards should be utilized whenever one systems implementation of an existing model is being replaced or supplemented by another systems implementation of that model. Model error due to incorrect representation of transactions is just as worrisome as model error due to incorrect equations. This will be addressed in Section 8.2.4. A particular point of concern is approximation error introduced by the need for fast response time in a production environment. This will be addressed in Section 8.2.5. Be careful about the degree of complexity introduced into models. Is there sufficient gain in accuracy to justify the reduction in intuitive understanding that results from added complexity? To illustrate with an example from my own experience: I had recently taken a new job and found that my most pressing problem was widespread user dissatisfaction with a model upgrade that had recently been introduced. The old model had been easy for traders and risk managers to understand; the new one was supposed to be more accurate, but could be understood only by the model development group. My initial examination showed that, on theoretical grounds, the difference between answers from the two models should be too small to make an actual difference to decision making, so I tried to persuade the model builder to switch 222 FINANCIAL RISK MANAGEMENT back to the original, simpler model. Finding him adamant on the need for what he viewed as theoretical correctness, I examined the new model more closely and found a major implementation error—a factor of 2 had been dropped in the equation derivation. This is the sort of mechanical error that would certainly have been picked up as soon as a formal model review was performed. But a similar error in a less complex model would have been caught long before, by the people using it on a day‐to‐day basis. 8.2.4 Model Verification of Deal Representation Verification of transaction details that serve as input to models can be just as important in avoiding valuation and risk measurement errors as verification of the model itself. The quote from FRB (2011) in Section 8.2.1 that “Models are of necessity simplified representations of real‐world relationships and so can never be perfect” applies just as much to the representation of transactions in models as it does to the models themselves. A single transaction confirmation document often runs to tens of pages and its representation in the model is just a few numbers, so inevitably some simplification and approximation are being utilized. In some ways, this is a more difficult issue to deal with than verification of the model itself, because of the large number of transactions that are often input to a single model. Reconciling transaction details between confirmations and position entries to models is an important middle‐office control function, as emphasized in Sections 3.1.1 and 3.1.2, and will certainly be expected to catch numerical errors in data entry and details such as correct day count convention. But middle‐office personnel lack the intimate knowledge of the model that might allow them to identify an important contract detail that is not being captured in the way the transaction is being represented in the model. Some steps that should be taken to control this risk are: ■ While model builders and independent model reviewers who do have intimate model knowledge won’t have the time to review every transaction confirmation, they should review all of the very largest transactions and a sample of the remainder, to look for both individual errors and patterns of errors. Samples should be selected at random, but with some weighting scheme that makes it more probable that large‐impact transactions have more of a chance of being part of the reviewed sample than those of lesser impact. Review should consist of a thorough reading of the confirmation, comparison with how the confirmation has been represented in the model input, and consideration of any possible gaps in the representation. Model Risk ■ ■ ■ 223 Middle‐office personnel should be strongly encouraged to immediately raise any question they have about adequate representation with an independent model reviewer. In the daily P&L verification process, discussed in Section 8.2.7.1, any transaction that makes a payment significantly out of line with the payment projected by the model should be investigated. This may uncover an outright error in data entry, but may also identify a facet of the contract that has not been adequately represented in the model input. Some wording differences between different variants of contracts may be very subtle (for example, see the discussion of legal basis risk on credit default swaps in Section 3.2 and Section 13.1.1.2). The best approach in this type of case may not be to try to capture all these variants in the model input. It may be better to have a separate offline calculation of the risks arising from wording differences and to establish limits and reserves against this risk on the basis of this offline calculation. This can be regarded as a type of liquid proxy, per our discussion in Section 8.4, with the most common contract type serving as the liquid proxy in all standard risk calculations, such as VaR and stress tests, but with the separate offline calculation capturing the nonliquid risk. 8.2.5 Model Verification of Approximations The quote from FRB (2011) in Section 8.2.1 that “Models are of necessity simplified representations of real‐world relationships and so can never be perfect” could be paraphrased as saying that “All models are approximations.” But for model review it is very important to distinguish between two different types of approximations: 1. Approximations in which some source of risk or driver of value has been omitted to reduce model complexity. 2. Approximations in which a computational approximation is being used in order to speed calculations and reduce cost. Approximations involving the omission of a risk factor pose greater challenges for model review, since it can be very difficult to estimate the potential impact on earnings and risk. This issue dominates our discussion of model validation in Section 8.2.6 and in Sections 8.3, 8.4, and 8.5. Approximations involving computational approximation are much easier to control, since the model review process can create a detailed comparison between the production model and a more thorough model that is run less often or only on a selected sample of transactions. This section focuses on techniques for dealing with computational approximations. 224 FINANCIAL RISK MANAGEMENT We begin by looking at a set of suggested controls for computational approximation and then illustrate with a detailed example. Suggested controls are: ■ ■ ■ ■ ■ ■ Model reviews should explicitly recognize the trade‐offs between model accuracy and investment of resources. Models used in production must be sufficiently fast to produce answers within the time frame required for providing quotes to customers and providing risk analysis to the trading desk and senior management or they will prove useless. Their development cost must be reasonably related to the revenue that can be realized on the products they support. The time required for development must be consistent with overall business plans. Evaluations of the inaccuracy of a production model need to be made by comparison to a more thorough model. Since model testing can be performed over a period of days or weeks, as compared to the minutes or seconds required of a production model, there is ample room to develop much more thorough models in testing environments. Comparison of results to the production model will show just how much accuracy is being lost. Comparisons of the production model to a more thorough model need to be performed not just for current market conditions. Tests should be performed to anticipate the impact on approximation of potential future market conditions. Where this test shows significant loss of accuracy, this identifies a good target for improved approximations. Until such improvements can be implemented, remedies can include valuation reserves against inaccuracy along with periodic revaluations with a slower but more accurate model, as well as traders exercising a degree of conservatism in pricing and hedging. Improved approximations can be achieved by “throwing more money” at the problem—buying more hardware to increase the number of calculations that can be performed in a given period of time. But more can usually be accomplished by the design of clever approximation algorithms. Indeed, one of the dirty little secrets of industry quants is just how much of the effort of people with PhDs goes into applying advanced mathematics to creating better approximation algorithms, rather than to the creation of new ideas for financial modeling. For example, see Section 13.3.3. In some cases, the thorough model can be so computationally intensive that it can be evaluated on only a sample of transactions. This is clearly a less desirable test of accuracy than one that looks at the full portfolio, but when this is necessary the reviewed sample should include all of the Model Risk ■ ■ 225 very largest transactions and a random selection of the remainder. The random selection should be chosen with some weighting scheme that makes it more probable that large‐impact transactions have more of a chance of being part of the reviewed sample than those of lesser impact. Approximations need to be reevaluated periodically, since approximation inaccuracy can be strongly related to portfolio size and composition. This point will be discussed in more detail in Section 8.2.8.1. A clear distinction needs to be made between the degree of accuracy needed to specify initial market conditions and the degree of accuracy needed to specify the evolution of market conditions. For example, as we will see in Section 12.5.2, a multifactor model for the evolution of interest rates does not offer much added accuracy over a single‐factor model for the valuation of Bermudan swaptions. But a specification of the shape of the initial yield curve that does not utilize a full set of liquid points on the yield curve can have a very significant impact on this valuation. As an illustrative example, I want to consider a situation I was involved with as a model reviewer. It involved a large portfolio of illiquid interest rate derivatives, all of which required Monte Carlo simulation for valuation calculations and for calculations of risk parameters. The number of simulation paths being run had been very clearly selected by the traders as the number that would allow all of the needed calculations to be performed between the close of business and the opening of trading the next day. To test for the impact of this choice on accuracy, I first set up a simulation for the whole portfolio that would run continuously for about a week in an offline environment. The results from this very much larger number of runs allowed me to estimate the number of runs needed to determine accurate valuation to within the tolerance required for financial significance (see, for example, Hull 2012, Section 20.6, “Number of Trials”). I was also able to see how much variance from this accurate valuation resulted from the smaller sample being used in production runs. I next used the larger sample to estimate the impact of the selection of different sets of Monte Carlo simulation paths on the production run. I determined that there was reasonable stability over a sufficiently small time frame; a set of paths that produced accurate values when comparing results from the smaller number of paths in the production run to the larger number of paths used in the offline run would also be fairly accurate over the next few days. But as the time period lengthened from days to weeks, a set of paths that had previously produced accurate values lost accuracy, both as a result of shifts in composition of the portfolio as new transactions were added and older transactions had less time left to expiry, and as a result of changes in market parameters. 226 FINANCIAL RISK MANAGEMENT I therefore set up the following process. Once a month, a new offline run would take place and would be used to determine the set of paths that was going to be used in the production runs for the next month. These production runs determined valuations and sensitivities reported for P&L and risk management purposes. Each month, the shift from one set of paths used in the production runs to a new set of paths would cause a change in P&L. These changes were randomly distributed, as likely to be increases in P&L as decreases, with a standard deviation that could be estimated from comparisons between valuations of different subsets of paths in the offline run. I argued that while these changes were randomly distributed, the firm should have a reserve against negative changes so that we were reporting to shareholders only valuations we could be fairly sure were actually achievable in the long run (since the portfolio consisted of illiquid positions, it was not possible to realize a current value just by liquidation—we were committed to holding the portfolio for a longer time period). Therefore, each time new transactions were added, the reserve would be increased to reflect added uncertainty, while as older transactions got closer to expiry, the reserve would be reduced to reflect less uncertainty. Each month, when the change in paths took place, the resulting gain or loss would impact the reserve and would not impact P&L (the reserve would have had to be exhausted to impact P&L, but this never occurred in practice). I was able to get agreement from the firm’s finance function and accounting firm to make this process part of the official books and records of the firm, and not just the risk management reports. Note that this process had a built‐in set of controls against changes in the market environment or portfolio composition, since the monthly offline run would automatically pick up any such impacts; the size of required reserve was recomputed each month by recomputation of the standard deviation between valuations of different sets of paths. The issue of computational approximations is particularly important for credit portfolio models and the closely related collateralized debt obligation (CDO) models. These will be covered in some detail in Sections 13.3.3 and 13.4.2. Another area in which computational approximation plays a major role is in multifactor interest rate models. This is discussed in great depth in Morini (2011, Section 6.2), which includes extensive examination of the accuracy of these approximations. 8.2.6 Model Validation We must now move beyond the tests of internal model consistency we have focused on in Sections 8.2.3, 8.2.4, and 8.2.5 to look at the fit between the model and the product or trading strategy being modeled, what Morini Model Risk 227 (2011) calls model validation. It is not surprising that this more challenging task does not lend itself to the easy consensus and process‐oriented approach of these last three sections. We will distinguish between three basic approaches to model validation: one focused on interpolation, one focused on the long‐term cost of hedging, and one focused on discovering the prevailing market model. These are not completely competitive approaches— there is some overlap among them—but they do have distinctly different emphases. As we will see, the appropriate approach has much to do with the purpose the model will be used for and the liquidity of the product being modeled. We will also look at the issue of how to deal with risks that may not be evident to model reviewers. I would strongly recommend comparing my approach in this section to Morini (2011, Sections 1.5.1 and 1.5.2). 8.2.6.1 Interpolation Approach In the interpolation approach, models are viewed as primarily serving as interpolation tools from observable to unobservable prices. This is closely related to the view discussed in Section 8.1 that downplays the importance of models. Viewing models as interpolation tools provides valuable insight into why certain models have been able to achieve a high degree of acceptance in financial management. It is much easier to agree on an interpolation methodology than it is to agree on a fundamental method for pricing an instrument. The danger is that this view leads to unwarranted complacency, since model builders often regard interpolation as being a mathematically trivial or uninteresting task. The result can be uncritical acceptance of what seems a plausible interpolation method or a view that the choice of interpolation methods is somehow a matter of taste. A closer examination will show that every choice of interpolation method entails significant financial assumptions. The interpolation of an unobservable price based on a set of observable prices amounts to the theory that the instrument with the unobservable price can be well hedged by the set of instruments with observable prices. As with any theory, this should be subjected to empirical testing and competition with alternative hedging proposals. Even the simplest‐sounding interpolation proposal (for example, calculating the two‐and‐a‐half‐year rate as a 50–50 average of the two‐ and three‐year rates) should be regarded as a model subject to the same tests as more mathematically complex models. We examine this in more detail in Sections 8.3 and 10.2.1. Models rarely cost firms money because modelers have made an error in complex mathematics; they frequently cost firms money because they embody financial assumptions that are not borne out by future events. 8.2.6.2 Cost of Hedging Approach Basing model validation on an examination of the possible costs of hedging a transaction over the long term is closely 228 FINANCIAL RISK MANAGEMENT related to the approach advocated in Section 6.1.2 of establishing a liquid proxy for an illiquid instrument and then simulating the difference between the liquid proxy and the actual trade. This viewpoint has been laid out very eloquently in Derman (2001): It’s never clear what profit and loss will result from hedging a derivative security to its expiration. Markets will move in unexpected ways, sometimes intensifying transactions costs and often dismantling what seemed a reasonable hedging strategy. These effects are rarely captured by the conventional models used in front‐office valuation systems. . . . Therefore, for illiquid positions, it is important to estimate the adjustments to conventional marked values that can occur as a result of long‐term hedging. One should build Monte Carlo models that simulate both underlyer behavior and a trader’s hedging strategy to create distributions of the resultant profit or loss of the whole portfolio. These distributions can be used to determine a realistic adjustment to the trading desk’s conventional marks that can be withheld until the trade is unwound and their realized profit or loss determined. . . . Monte Carlo analysis provides a good sense of the variation in portfolio value that will be exhibited over the life of the trade due to transactions costs, hedging error and model risk. Ultimately, such analyses should be part of the desk’s own front‐office valuation system. Note: There is a rough equivalence between Derman’s use of “underlyer” liquid instruments being used to hedge the illiquid instrument and my emphasis on a representative liquid hedge. As Derman says: “Derivative models work best when they use as their constituents underlying securities that are one level simpler and one level more liquid than the derivative itself.” 8.2.6.3 Prevailing Market Model Approach Rebonato (2003) emphasizes model validation based on the model that prevails in the marketplace and anticipation of directions in which the prevailing market model might evolve: “Model risk is the risk of occurrence of a significant difference between the mark‐to‐model value of a complex and/or illiquid instrument, and the price at which the same instrument is revealed to have traded in the market.” ■ “From the perspective of the risk manager the first and foremost task in model risk management is the identification of the model (‘right’ or ‘wrong’ as it may be) currently used by the market to arrive at traded prices.” Model Risk 229 “[M]arket intelligence and contacts with the trader community at other institutions are invaluable.” ■ Requires a variety of models to reverse‐engineer observed prices. ■ Requires information about as many observed prices as possible. ■ “No matter how good or convincing a theoretical model might be, few states of affairs should worry a risk manager more than the trader who, using this model, consistently beats all competing banks in a competitive‐tender situation.” “The next important task of the risk manager is to surmise how today’s accepted pricing methodology might change in the future” (including changes to model, changes to calibration, and changes to numerical implementation). “Being aware of the latest market developments, and of academic papers can be very useful in guessing which direction the market might evolve tomorrow.” “To a large extent, the model risk management task can be described as an interpolation and extrapolation exercise that simply cannot be carried out in an informational vacuum . . . without at least some anchor points of solid knowledge about the levels and nature of actual market transactions.” ■ ■ ■ Hull and Suo (2001) present an approach to model validation closely related to Rebonato’s. They quantify the risk of a model being used by a trading desk by estimating how much of a loss the trading desk would suffer if a different model turned out to be correct. 8.2.6.4 Matching Model Validation to Model Purpose There are two dimensions to matching model validation to model purpose. The first relates to the degree of liquidity of the instrument being modeled. The second relates to differences between models being used for managing the firm’s overall risk, the models of valuation and position measurement described in Section 6.1, and models being used to make trading decisions. The key to designing a proper model valuation procedure for models being used to manage the firm’s risk is to fit the model review to the degree of liquidity of the instrument for which the model is being used. Essentially, we need to work out the model review implications of the liquidity differences discussed in Section 6.1.1. Please note that the distinction here is between liquid and illiquid instruments, not liquid and illiquid positions. If a position in a liquid instrument is so large as to create an illiquid position, this needs to be dealt with by modifying VaR and stress test calculations, as discussed in Section 6.1.4, but does not require a different model or model review than would be needed for a smaller position in the same instrument. 230 FINANCIAL RISK MANAGEMENT The interpolation approach to model validation is usually very reasonable for liquid instruments. We will look at the details of applying the interpolation approach to liquid instruments in Section 8.3. For illiquid instruments, the interpolation approach has little relevance; the kind of frequent checks of interpolation methodology with the market recommended in Section 8.3 are not possible because of illiquidity. Given that interpolation will be of little use, model verification for illiquid instruments must therefore rely on either the cost of hedging approach or the prevailing market model approach or some combination of the two. We explore this issue in detail in Section 8.4. For models used in making trading decisions, discussed further in Section 8.5, the prevailing market model approach is the most salient, as the next example illustrates. To better understand the implications of different types of model use, consider the case of the 1998 breakdown of the historical relationship between the pricing of interest rate caps and interest rate swaptions, discussed in detail in Morini (2011, Sections 11.3 and 11.4). For models utilized for managing the firm’s overall risk on liquid instruments, this breakdown was probably a nonevent. Since both caps and swaptions had adequate liquidity in external price quotes, most firms would be using some form of interpolation model to value and compute risk statistics for their caps just using market cap prices, and a separate interpolation model to value and compute risk statistics for their swaptions just using market swaption prices. There would have been no interaction between the two models. (Some possible exceptions where liquid prices in one market would be allowed to override liquid prices in another market will be examined in Section 8.3.) There may have been a different story for models used in managing the firm’s overall risk on illiquid products. Some products, such as Bermudan swaptions, knock‐out caps, and forward‐start interest rate options (see Section 12.5.2 for details), may have been priced using models that incorporated both market cap prices and market swaption prices as inputs. Would the breakdown in the historical relationship have caused problems for these models? Using the approach we discuss in Section 8.4, there should be a very significant degree of conservatism relative to historical relationships built into the reserves kept against model risk and, in any case, the relationship that is important is what will happen over the lives of the deals. Could a temporary period of breakdown in historical relationships be enough to call into question the adequacy of these reserves? We discuss this further in Section 8.4. But models being used by the trading desk to determine trading strategies involving caps and swaptions would definitely have been impacted. Here’s the description by Morini (2011): “No matter whether or not the long term equilibrium was going to come back, the market had gone too Model Risk 231 far from it for too long for a bank or a fund with a risk management unit to stand it.” Here it is clearly the case that the prevailing market model approach must govern. 8.2.6.5 Capturing Risks That Are Difficult to Identify I once heard a senior risk manager for an investment bank say, “I don’t stay awake at night worrying about the risks I know, but about the risks I don’t know.” This is a sentiment with which I could readily identify. In performing model validation, the great fear is that there will be some exposure that is not being captured by the model and that you, as the independent model reviewer, don’t even know about. For example, there might be some potential piece of legislation or judicial decision that would have a big impact of the transaction being modeled, but it has never been discussed in the literature you have access to. This type of potential exposure is probably known to the front‐office personnel who specialize in the product. Ideally, they should consider it in their internal model review and share their concerns with the independent model reviewer. But the usual moral hazard concerns come into play, with the incentives discussed in Section 2.1, motivating front‐office personnel to be reluctant to share information that might lead to tightened controls. JPMorgan in the late 1990s instituted an internal system called Risk Identification for Large Exposures (RIFLE) to try to address this issue. It is still in operation, as can be seen from the following quote from JPMorgan’s 2010 annual report: Individuals who manage risk positions in the Investment Bank are responsible for identifying potential losses that could result from specific, unusual events, such as a potential change in tax legislation, or a particular combination of unusual market events. This information is aggregated centrally for the Investment Bank. Trading businesses are responsible for RIFLEs, thereby permitting the Firm to monitor further earnings vulnerability not adequately covered by standard risk measures. (p. 145) But even with a mechanism like this in place, the incentive issue remains. Traders who do an honest job of reporting these risks may thereby lower their return on risk measures and attract added scrutiny of position sizes. To attempt to overcome this, a firm needs to make clear distinctions in performance evaluation between losses that occurred due to an event that the traders had made certain received adequate firmwide attention in advance of the trade being approved and losses that occurred due to an event where this type of advance notice was not provided. 232 FINANCIAL RISK MANAGEMENT 8.2.7 Continuous Review FRB (2011, Section V) calls for “validation activities [that] continue on an ongoing basis after a model goes into use.” Three major components of ongoing validation activities are (1) daily P&L reconciliation for models being used for valuation and risk‐reporting purposes, (2) back‐testing for statistical forecasting models, and (3) analysis of overrides for cases where model output needs to be altered based on the expert judgment of model users. We consider each in turn. 8.2.7.1 Daily P&L Reconciliation In Sections 3.1.1 and 3.1.2 we stressed the importance of a daily explanation of P&L produced by independent support staff as a control measure against both fraud and nondeliberate incorrect information. Here, we want to stress its equal importance as a tool for identifying model weaknesses. The basic approach of P&L reconciliation is to take position reports from the close of business (COB) of the prior day and combine them with actual market movements from the previous day’s COB to the current day’s COB to estimate P&L for the day. The idea is that since position reports show sensitivities to changes in market variables (e.g., option “greeks”), multiplication of these sensitivities by actual price changes should produce a reasonable estimate of P&L. Of course, this applies only to those positions that were in place as of the COB of the previous day, so it is first necessary to identify and segregate the P&L due to trades booked during the day (this includes hedges that may have been put on during the day in response to market moves). This segregation of P&L between that due to previous COB positions and that due to trades booked during the day is already valuable as a tool in avoiding inadvertent Ponzi schemes in which profits on newly booked trades cover up hedge slippage on existing trades (compare with the discussion in Section 2.2). If the initial estimate of P&L is significantly different than actual P&L, what can be the possible causes? Incorrectly recorded positions are certainly a possibility; this is why P&L reconciliation is a valuable tool in uncovering fraud and incorrect information. It’s true that an incorrectly reported position might impact both the COB position report and the daily P&L record, but at some point there will be an actual payment on the position, and at this point, when the payment becomes part of the daily P&L, a discrepancy between projected P&L and actual P&L will show up. Another possibility is that the COB position has not been reported with sufficient detail. For example, an option position might be reported using just the first‐order greeks, such as delta and vega, and not the second‐ order greeks, such as DdelV and the price‐vol matrix (see Section 11.4 for a Model Risk 233 detailed discussion of option sensitivity measures and how they can be used in P&L reconciliation). In this case, the P&L reconciliation will identify the need for more detail in the position report, which will enhance management’s ability to accurately measure exposure. A third possibility is a deficiency in the model. For example, it could be that even though the contract details of a position are correctly entered, the model is not using these details correctly in computing P&L or in computing the position. Again, we might worry that the model flaw will impact both P&L and position reporting and so will not be spotted in reconciliation. But when a payment date is reached and actual payment becomes part of the daily P&L, a discrepancy should appear. More likely, the model is missing or mishandling some key factor of risk, and a position that appears hedged is actually suffering some hedge slippage. An example might be a Bermudan swaption model that fails to identify some circumstances in which early exercise becomes more profitable for the counterparty. So daily P&L reconciliation should form an important part of identifying model problems that may have been missed in initial model validation. 8.2.7.2 Back‐Testing Any statistical forecasting model needs to have continuous monitoring of actual performance. The best‐known example in risk management is back‐testing of VaR models, discussed in detail in Section 7.1.2. Since VaR models produce statistical distributions of the size of losses that can be expected to occur at different percentiles (e.g., 1 percent of the time, 2 percent of the time), these projected distributions need to be continuously compared to actual experience. Statistical analysis should be applied to results, and when this analysis indicates a strong probability that the actual distribution differs from the projected distribution, corrections to the model need to be considered. Until corrections can be made, an extra layer of conservatism may be necessary in utilizing limits and reports based on the existing model. Similar back‐testing is called for in any statistical model being used to suggest a trading strategy. In hedge funds and on trading desks, one frequently finds trading strategies employed based on statistical studies of how the strategy would have performed historically (see Fabozzi, Focardi, and Kolm 2010, Chapter 7). Continuous back‐testing is needed to update evaluation of the model’s performance and identify changes in market environment that require alteration or even abandonment of the model. 8.2.7.3 Analysis of Overrides When hedge funds and trading desks employ statistical models for trading, there will occasionally be the need for a trader to override the trading strategy recommended by the model because of economic insights that cause the trader to doubt the advisability of the model’s 234 FINANCIAL RISK MANAGEMENT recommendation. It is important that all such overrides be recorded and analyzed, with performance of the model strategy not pursued compared to success of the override, to spot possible needs for model modification. It is less common for models used for valuation and risk reporting to be overridden, but there are examples. One typical case is the representation of binary options in option greeks. When trading desks do not use the liquid proxy representation of binary options by a call spread recommended in Section 12.1.4, it often happens that the trading desk must come to the risk managers and ask for an override on the large delta and gamma positions produced by a binary option nearing expiry at a price close to the strike. Risk managers will be willing to grant these requests, since the trading action that would be required to get back within the limit would be a foolish position to take, as discussed in Section 12.1.4. But any such overrides should be recorded and analyzed. It is just such analysis that has persuaded some firms to move to the use of the liquid proxy representation, since a model that is producing position reports that would be foolish to act on is clearly flawed. 8.2.8 Periodic Review Once a model has been approved and is in production, ongoing validation is still required. FRB (2011, Section V) reiterates what has been a long‐standing regulatory requirement that existing models should be reviewed at least once a year, but also calls for continuous monitoring of model performance. We’ll examine periodic review in this section and ongoing monitoring in the next section. To be productive, the periodic (generally, annual) review of existing models must be carefully designed. Merely replicating previous tests is likely to be both unlikely to produce new insights and wasteful of resources. Reviews need to be focused on changes to the environment in which the model is being used that should trigger new testing and possibly new conclusions. We’ll focus on four types of environment changes that should be investigated: (1) changes in the population of transactions the model is being applied to, (2) changes in the market environment, (3) changes in the academic literature or in market practices, and (4) changes in technology. In addition, the periodic review should examine any patterns that have been revealed by the ongoing monitoring we describe in the next section. 8.2.8.1 Changes in the Population of Transactions Consider the Kidder Peabody disaster, discussed in Section 4.1.2, as an illustration. Whatever your opinion about whether Joe Jett was deliberately gaming the system, there is no doubt that the firm was ill‐served by having a model that computed the value of Model Risk 235 forward transactions without proper discounting. But Kidder Peabody was hardly alone in the industry in using a model that omitted discounting of forwards. This is not due to a widespread ignorance of this fundamental principle of finance. What does often happen—and this is a pattern I have seen over and over again—is that a sensible decision is made at the time a model is built but is not subjected to adequate review as circumstances change. So a model might be set up for valuing forwards that at the time of implementation is being used to evaluate trades that are of moderate size and no more than a few days forward. The added accuracy that comes from correct forward discounting might be quite small and thus easily justify a decision not to devote the added programming time and computational resources to include this factor. The situation changes as larger transactions with longer forward periods are added. As the situation changes through time, there comes a point at which the proper decision would be to change to a more accurate model. But the decision to invest the resources needed to improve accuracy can be a difficult one, involving considerable expense, diversion of resources from important new ventures, and perhaps a limitation on trading volume until the change is made. The environment may be changing gradually, so that no single point in time stands out as the time at which to switch. This is the kind of situation in which a periodic review of the impact of changes in the population of transactions being valued by a model can be of tremendous value. Note that the change in population of transactions could be due to changes in number of transactions (an approximation that had little impact when the model was being used to value just a few deals causes more concern when the model is being used to value many deals); size of transactions (maybe the model is being used to value just a few deals, but the average size of deals has grown to the point that approximations have become worrisome); terms of transactions (for example, the large increase in the length of the forward period in the case discussed previously, or an increase in time to maturity, or the more frequent use of features that are difficult to evaluate); or a combination of all three. Morini (2011, Section 1.4.1) explains the reasoning behind this policy very well: A bank cannot expend big resources for a small exposure; and additionally banks and traders learn by trial and error, a new model needs to be tested for a while to really know its risks. When the exposure starts growing, a previous model validation must not automatically be considered valid: a surplus of effort can be spent on the model used, an effort that was not economically meaningful in the past but is crucial in the face of the increased exposure. 236 FINANCIAL RISK MANAGEMENT When faced with a large change in deal population, an independent model review group must think very carefully about what is behind the change. Is it just a new market taking off to meet a customer need, or is it a structuring group looking to arbitrage a deficiency in the way a transaction is being valued or a risk is being measured? If traders and structurers are hired because of their skills in uncovering complex arbitrage opportunities in markets, one shouldn’t be too shocked if they sometimes use the same skill set to try to find arbitrage opportunities in regulations, whether external (government) regulations or internal (risk management) regulations. When independent reviewers see signs that such an opportunity is possibly being exploited, they must expend extra effort on trying to uncover the motivation and possible consequences. If a reviewer does spot a loophole being exploited, there should be no hesitancy in quickly improving the valuation procedure or risk measure. There will inevitably be cries of foul play from structurers who can no longer take advantage of the old system and complaints that “The rules of the game have been changed without warning.” Those complaining need to be reminded that risk management is not a game but a serious endeavor to protect shareholders, depositors, and taxpayers. Keeping a fixed set of rules and allowing structurers to experiment and see where the weaknesses are is a recipe for disaster, as the rating agencies amply demonstrated by publishing fixed models for evaluating the risk of CDO tranches and letting bank structurers play with the models until they had designed trades that optimized the degree to which the models underreported deal risk (see Section 5.2.3 for further discussion of this example). 8.2.8.2 Changes in the Market Environment An example of a change in market environment would be a risk factor that previously could not be priced based on market observation but now has liquid prices available. This could change previous conclusions about which model inputs need to be derived from market prices. In the other direction, deterioration in the liquidity of a pricing source might prompt the need for new reserves or limits. Another example would be changes in levels of prevailing market prices that might prompt reruns of sensitivity analyses and model stress tests. FRB (2011, Section V) states, “Sensitivity analysis and other checks for robustness and stability should likewise be repeated periodically. . . . If models only work well for certain ranges of input values, market conditions, or other factors, they should be monitored to identify situations where these constraints are approached or exceeded.” Another instance of change in market environment would be rapid growth in the size of a market. This should prompt reexamination of the relevance of historical data, since rapid growth may be the result of major Model Risk 237 changes in the nature of the market. A recent example was the explosive growth in U.S. subprime mortgages. The use of historical data on default rates for subprime mortgages should have then been treated with extreme caution. As it soon became clear, underwriting standards for approving these mortgages had become drastically more lax than in previous eras, which contributed to steeply rising default rates (see Section 5.2.1). A prior example was the precipitous growth in non‐investment‐grade bonds in the late 1970s and early 1980s. This growth was largely due to the efforts of Michael Milken at Drexel Burnham Lambert, and a major contributor was studies by Milken and others showing the very favorable historical returns of these bonds after adjusting for default losses. But as soon as there was a large increase in the issuance of these bonds, it should have been suspected (as turned out to be the case) that the growth was largely being fueled by types of transactions that had rarely been done previously and for which the historical data was of dubious relevance. Bruck (1988) is a good account of the Milken story; see particularly page 28–29 on historical return studies, and pages 266–270 on skepticism about their continued relevance. 8.2.8.3 Changes in the Academic Literature or in Market Practices Periodic reviews offer a good opportunity to consider any new approaches to modeling a particular type of transaction that have appeared in the academic literature, have been discussed at conferences, or have begun to be used by other market participants. This is where the emphasis in Rebonato (2003) on “market intelligence and contacts with the trader community at other institutions” and the reverse engineering of observed prices from other firms can be of particular value. 8.2.8.4 Changes in Technology Increased computational capacity may change the conditions on which previous decisions about approximation techniques have been made. Increased computational capacity could be due to newly purchased or upgraded hardware or to advances in computational theory. New conditions should lead to a reassessment of prior decisions—replacing existing approximation techniques either with more accurate ones or with full computations. 8.3 LIQUID INSTRUMENTS Models for liquid instruments are robust and easy to test, since they can constantly be checked against actual liquid market quotes. This is why they lend themselves so readily to the interpolation approach to model validation outlined in Section 8.2.6.1. Risk reports only need to look at exposures, 238 FINANCIAL RISK MANAGEMENT such as delta and vega, measured against current market prices. If changes in price levels lead to new exposure levels that concern senior managers, the liquidity of the instrument will allow for reduction in positions at the time the exposure exceeds desired levels. We illustrate with an example of a model review for a very liquid instrument. Consider a portfolio of U.S. dollar interest rate instruments (e.g., interest rate futures, forward rate agreements, interest rate swaps, government bonds) with no option component (see Chapter 10 for a detailed discussion of how the risk on such a portfolio is managed). There will be liquid market quotes available throughout the day for trades on a large subsection of these positions. But many instruments will need some form of modeling for valuation, since even a very liquid (“on‐the‐run”) instrument at the time of original transaction may soon become less liquid (“off‐the‐run”) through the passage of time (e.g., a five‐year swap, for which direct market quotes are readily available, soon becomes a four‐year 11‐month swap, whose liquidation price needs to be inferred from market quotes for on‐the‐run instruments). The models used for this off‐the‐run valuation will also be needed for computing the change of the portfolio’s value in VaR and stress test simulations. The models needed for these computations are quite standard throughout the financial industry by now, but there are still choices in interpolation methodology that need to be made that constitute forecasts of relative movements between instruments (Section 10.2.1 provides details). These modeling choices are best made by the front‐office personnel who have the product expertise and superior data access. In addition, it is the front office to whom the profits from correct forecasting decisions (and losses from poor forecasting decisions) properly belong. Model validation by outside reviewers only requires periodic checking of model valuations against actual market prices. Close agreement shows model adequacy; significant differences indicate the need to establish limits and/or valuation reserves and may serve as clues for model revision. The most robust price checks come when there is an actual transaction in an off‐the‐run instrument, but price checks can also be performed by polling brokers, dealers, and other independent sources of pricing information (issues involved in obtaining such quotes are addressed in Section 6.1.3). While actual conduct of the price checks may be performed by support staff, model reviewers and other senior control personnel should be involved in the design of the price check procedures, with regard to frequency and standards for confirmation of model adequacy. The type of price check just discussed should be complemented by the daily P&L explanation exercises discussed in Section 8.2.7.1. As observed there, the P&L explanation process often identifies model deficiencies when Model Risk 239 there are unexplained P&L changes, particularly around transaction dates and dates for scheduled payments and resets. But even a thorough daily P&L explanation process should not be regarded as a full substitute for price checking; it may be that a model performs very well in handling on‐ the‐run transactions from inception through maturity and is rarely tested on off‐the‐run transactions because the trading desk almost always transacts on on‐the‐run dates. But what happens if the desk goes through a stop‐loss limit or if the firm’s appetite for risk decreases? A reduction in positions may need to take place by reversing previously booked transactions that are now off‐the‐run. Risk managers will want to have prepared for this eventuality by testing model pricing of off‐the‐run positions. If the disagreement between an observed market price and a model value represents a clear difference between where a risk can be sold at the current time and a theory as to the value of the asset over a longer period of time, then no matter how sound the reasoning behind the theory, I would recommend holding to the mark‐to‐market principle. If a firm deviates from this principle and values based on longer‐term values that it believes can be realized, rather than on prices at which risks can currently be exited, it is turning short‐term risks into much harder‐to‐evaluate long‐term risks. Morini (2011, Section 1.2.1) supports this view colorfully: “on intuitive grounds, anyone who claims that arbitrage opportunities are abundant in the market should always be asked if he is fabulously rich. If he is not, it is not clear why he knows of so many free lunches and yet, rather than exploiting them, goes around passing preposterous judgments about market functioning.” However, sometimes the difference between an observed market price and a model value represents two different ways in which a risk can be sold at the current time. Although this would seem to violate several important axioms of finance theory—the no‐arbitrage principle and the law of one price—these are just models and cannot expect any absolute deference in the face of empirical exceptions. However, there needs to be careful evaluation of what lies behind an observed difference between a market price and a model price before an intelligent decision can be made as to which is the best of two different ways to represent the risk. Let’s focus on a concrete illustration. You have observable market prices for a European call option, a European put option, and a forward to the same expiry date, with the same underlying and, in the case of the put and the call, the same strike price. The combined prices, however, do not agree with put‐call parity. This would imply, for example, that a position in the put that you have sold can be offset in two different ways—you could buy a put, or you could synthetically create a put by buying a call and entering into a forward. It also implies that the call‐forward combination will offset the position at a cheaper price than the direct purchase of the put. 240 FINANCIAL RISK MANAGEMENT What should a risk manager recommend in such circumstances? Since the main argument behind a no‐arbitrage principle such as a put‐call parity is that the lack of parity will be quickly eliminated by profit‐seekers taking advantage of a riskless opportunity to make money, any persistence of parity violation is suggestive of some liquidity difficulties preventing the opportunity from being exploited. We’ll consider some possibilities: ■ ■ ■ ■ This is an arbitrage of which very few market participants can take advantage, but your firm is one that can. This could be because the market for the put is in some way restricted to only a few firms. It could be an arbitrage that is difficult to identify computationally and your firm has a computational advantage. It could be a diversified basket of assets that is difficult to accumulate and your firm has an advantage in its market access (see the discussion in Section 12.4.1.1). In such cases, it is right to base valuation on the model‐derived price (in this instance, the call‐forward combination), since this represents a liquid external price at which risk can actually be extinguished in the short term. One of the prices is less liquid than the others. For example, the amount of trading for that strike and date could be much more active in calls than in puts. This would be a strong indication of the desirability of using a model (put‐call parity) to supply a price based on more liquid quotations rather than utilizing a less liquid price. The same reasoning would apply if the call and put markets are significantly more active than the forward market, in which case I would recommend replacing an illiquid forward price with a put‐call parity–derived price based on liquid put and call prices. A timing difference exists in price quotations. Perhaps the options market posts closing prices at an earlier time of day than the forwards market. It is certainly legitimate to use a model to update both call and put quotes to adjust for changes in the forward since the time the options market closed. Some contract features make the model not completely applicable. Sometimes, on closer examination, contract provisions call into question the applicability of a model. In this case, it might be an allowance for early option exercise in certain circumstances, whereas put‐call parity applies only to options without early exercise provisions. This last type of case has led to a considerable number of disputes between risk managers and trading desks. One example that has arisen at several firms is traders’ desire to unlock stock option values contained in convertible bonds. Option models applied to convertible bond prices frequently indicate implied volatilities that are quite low compared with the Model Risk 241 implied volatilities that can be derived from plain equity options on the same stock, leading traders to conclude that buying the convertible is a good value trade. Trading desks hungry to book immediate profits have pressed for overriding reasonably liquid convertible price quotes with a model‐driven quote based on the implied volatility from the equity options market. But a convertible bond contains the option to exchange a bond obligation for a stock obligation rather than to exchange cash for a stock obligation, so it cannot be completely reduced to the value of an equity option (see the discussion in Section 12.4.4). When turned down on their first attempt, some trading desks have shown good enterprise in marketing total return swaps on the bond portion of the convertible in an attempt to isolate the equity option portion. So long as the swap has been properly engineered to cover all contingencies, such as canceling the swap without penalty in the event that the bond is converted for equity, a complete decomposition can be achieved and it is legitimate to value the resulting position as an equity option. Risk managers have, however, been very careful to check that no uncovered contingencies are present before allowing this valuation change. 8.4 ILLIQUID INSTRUMENTS 8.4.1 Choice of Model Validation Approach Model use for illiquid instruments is much more critical than it is for liquid instruments and, unfortunately, model validation is also much more challenging. There may be a complete absence of actual market prices at which positions can be unwound, so modeling assumptions and inputs for unverifiable model parameters now become a key driver of model valuation. Both Derman (2001) and Rebonato (2003) have strong statements as to the difficulty these risks can entail. ■ ■ Derman: “Because of their illiquidity, many of these positions [in long‐ term or exotic over‐the‐counter derivative securities that have been designed to satisfy the risk preferences of their customers] will be held for years. Despite their long‐term nature, their daily values affect the short‐term profit and loss of the banks that trade them.” Rebonato: “What differentiates trading in opaque instruments from other trading activities is the possibility that the bank might accumulate a large volume of aggressively‐marked opaque instruments. When, eventually, the true market prices are discovered, the book‐value re‐adjustment is sudden, and can be very large. Stop‐loss limits are ineffective to deal with this situation, since the gates can only be shut once the horse has well and truly bolted.” 242 FINANCIAL RISK MANAGEMENT Let’s consider which of the model validation methodologies of Section 8.2.6—cost of hedging or prevailing market model—is more appropriate for illiquid instruments. My own view is that the cost of hedging approach is the more relevant for independent model reviewers for two reasons: 1. Even if a given model prevails in the market place, so long as the trading desk can’t actually extinguish positions at the prices implied by the model, owing to illiquidity, it is actually the hedging costs that will determine the firm’s P&L on the product. The model might continue to prevail in the marketplace for many years, and all the while the firm loses money on its hedging strategy. An advocate of the prevailing market model approach might respond that if, in fact, the model leads to hedging losses, then firms will eventually replace the model, so this is just a case of anticipating the direction in which the prevailing market model may evolve, in line with Rebonato’s proposed criteria. But then I would still want to utilize the cost of hedging approach as a key tool in anticipating prevailing market model evolution. 2. I am wary of the ability of risk managers to anticipate prevailing market model evolution using any other tool besides cost of hedging simulation. Some of the tools Rebonato recommends—market intelligence and contacts with the trader community at other institutions—seem much easier for traders to utilize than independent reviewers. If independent reviewers do rely on the cost of hedging approach, it would still be valuable for the front‐office reviewers to utilize the prevailing market model approach as a supplement. This is particularly true when mark‐to‐market policies require marking to the prevailing market model, so that even if an instrument is being priced and hedged in a way that will virtually guarantee long‐term profit, accounting losses may need to be booked in the shorter term. Rebonato (2003) makes this clear in Section 2.1, saying that “model risk arises . . . because of a discrepancy between the model value and the value that must be recorded for accounting purposes.” This would not be the case for the mark‐to‐market policy I advocate in Section 8.4.4. In a thorough review of cost of hedging, Monte Carlo simulation allows systematic consideration of many possible future paths of relevant liquid market prices and other economic variables. The soundness of the model can be judged only over longer time periods, when longer‐term unobservable prices transform into shorter‐term observable prices, when there is enough time to observe the impact of required rehedges, or when trades reach maturity and require contractual payments. Over a short time period, almost any model chosen will appear to perform well by a type of circular Model Risk 243 reasoning: The instrument with unobservable prices will be valued using the model and the observable price inputs. Therefore, the movement of the unobservable prices relative to the observable prices will seem stable since the same model is being used for valuation throughout the time period (see the example of this discussed in Section 10.2.1). Proper design of a model review of an illiquid instrument utilizing Monte Carlo simulation has two parts: (1) choice of the liquid proxy, which will be analyzed in Section 8.4.2, and (2) design of the simulation, discussed in Section 8.4.3. I will then look at issues this approach raises for mark‐to‐ market policies in 8.4.4 and for risk measurement in 8.4.5. Another application of this approach to creation of liquid proxies and simulations of hedge slippage involves investments in hedge funds for which you lack data on current holdings of the hedge funds. Trying to draw conclusions from the historical pattern of the returns for the hedge fund and historical correlation of these returns with other positions is dubious, given the possibility that current holdings of the hedge fund may not resemble historical holdings. The case for treating these investments as illiquid rests both on limitations on hedge fund withdrawals and on lack of information on true exposure. A liquid proxy can be built by making reasonable inferences about the current style of the hedge fund, based on whatever public and private information you have available from the fund. Ineichen (2003) is an excellent starting point for explanations of the differing hedge fund styles (Chapter 5) and detailed examination of each style in relation to indexes of liquid investments (Chapters 6 through 8). A statistical approach to creating liquid proxies for each hedge fund style can be gleaned from Hasanhodzic and Lo (2007). The liquid proxy can be used for representing hedge fund investments in VaR and stress test calculations. Statistical analysis can then be performed on deviations between the liquid proxy and historical returns on hedge funds. 8.4.2 Choice of Liquid Proxy A choice of liquid proxy is equivalent to a choice of what liquid market prices are utilized in modeling the illiquid instrument. Every model choice implies a liquid proxy, and every liquid proxy choice implies a model. In evaluating whether a liquid proxy choice is correct, it is necessary to ask whether the implied model makes adequate use of available liquid market prices. This is closely related to one of the key questions in Derman (2001): “Has the model been appropriately calibrated to the observed behavior, parameters and prices of the simpler, liquid constituents that comprise the derivative?” This point can be most clearly made using a concrete example, which is discussed more fully in Section 12.4.2. 244 FINANCIAL RISK MANAGEMENT Consider an option written on a basket consisting of two stocks. You could choose two different ways to model this: (1) have a complete model of the price evolution of each of the two stocks individually and assume a correlation between them, or (2) directly model the price evolution of the basket. We’ll call these the correlation model and the direct model, respectively. Assume that there are liquid market prices for options on the individual stocks but no liquid market prices for options on the basket or on the correlation, which is a fairly standard situation. It can be argued that either model is a reasonable choice. In either case you will need input for a variable that cannot be observed in the market. In both cases, you have included all the sources of risk in your model. But, as will be shown in our more detailed discussion in Section 12.4.2, where correlation is not expected to be too negative, the first model offers definite advantages in terms of making better use of liquid market prices. Options on the individual stocks will serve as effective partial hedges for the basket option, so utilizing the first model, which can be calibrated to current market quotes for these options, offers the following advantages over the second model: ■ ■ ■ The correlation model implies a liquid proxy that represents the basket trade in the exposure reports for options positions on the two individual stocks. This encourages the use of liquid hedges. The correlation model will require valuation changes in the basket when there are changes in the implied volatility of the two individual stock options. The direct model does not require such valuation changes and so can result in stale valuations not fully reflecting the cost of unwinding some of the risk in the trade. The correlation model exhibits significantly lower statistical uncertainty of results compared with the direct model. This should permit lower required reserve levels and larger limits than could be allowed if the direct model was used. Note that these advantages of the correlation model over the direct model are based on empirical, not theoretical, findings. As can be seen in the fuller discussion, if correlation levels are expected to be very negative or if the product were structured differently (for example, an option on the difference between the stock prices rather than on the basket), the advantages of the first model over the second would diminish to the point of indifference between the models. In some cases, the liquid proxy used could consist of an instrument that is not itself liquid, but for which modeling in terms of a liquid proxy and simulation of the remaining risk have already been incorporated into Model Risk 245 the firm’s risk management system. This is reminiscent of the quote from Derman in Section 8.2.6.2: “Derivative models work best when they use as their constituents underlying securities that are one level simpler and one level more liquid than the derivative itself.” For example, in Section 12.3.3, in examining a liquid proxy for barrier options based on Peter Carr’s approach, I advocate the use of illiquid binary options as part of the liquid proxy, noting that “techniques we have already developed for managing pin risk on binaries” in Section 12.1.4 “can now easily be brought into play.” 8.4.3 Design of Monte Carlo Simulation Modeling the differences between the actual trade and its liquid proxy must go all the way to final payout or to when the trade becomes liquid. Modeling must reflect the possibility that the model used for pricing and trading the product may be wrong. Modeling should be by Monte Carlo simulation to reflect a full range of possible outcomes and to generate a statistical distribution that can be used in assessing issues such as capital adequacy. Let us take these points in more detail: ■ ■ ■ Don’t assume that an illiquid instrument will become liquid—it may happen but it shouldn’t be assumed. Another way of saying this: It is important that statistical analysis of the distribution of parameters be based on actual market observations and not on derived values, since the derived values often themselves contain modeling assumptions subject to error. For example, if a given market is currently liquid only out to seven years, use only quotations out to seven years in your hedging simulations; 10‐year quotations derived by extrapolation should not be used. This is analogous to the point made earlier in Section 8.4.1 about avoiding circular reasoning in model validation. Statistical assumptions used in determining distributions should not be constrained by any assumptions made within the valuation model. For example, the valuation model may assume a normal distribution of a factor because it is computationally simple and the increase in accuracy from using a different distribution can be shown not to be worth the added investment. This would not in any way justify assuming that the corresponding input variable is normally distributed in a model‐testing simulation, since the computational trade‐offs motivating the model‐ building decision do not apply to the model‐testing calculation. Independent reviewers must be careful not to rely on statistical analysis prepared by traders. It is notoriously easy to employ data‐mining techniques to find statistical proofs of nearly any relationship by selecting the right historical data set. Statistical controls, such as careful 246 ■ ■ ■ FINANCIAL RISK MANAGEMENT discipline about segregating historical data into sample periods to fit parameters and out‐of‐sample periods to test results, are useful, but can still be defeated by sufficiently industrious data mining. It is better to have truly independent analysis, even at the risk of inaccuracy (on the side of conservatism) from lack of insider information. Use of Monte Carlo simulation allows for generation of a full statistical distribution of results, which can be very useful for issues such as determining capital adequacy on illiquid positions. This is a necessity if the capital adequacy proposal of Section 8.4.4 is to be followed. It must be emphasized that any statistical distribution involving tail risks requires subjective probability judgments (as discussed in Section 1.3). Still, the basic approach of insisting that simulation be of hedge trades involving liquid instruments, and that simulation go all the way to the point at which the original position becomes liquid, means that there will be a lot of historical liquid pricing data that can be utilized in forming these probability judgments. In essence, while illiquid instruments cannot be fully evaluated based on current liquid prices, they can be evaluated based on the future evolution of liquid prices. For illustrative examples, see Sections 10.2.2, 12.1.4, 12.3.3, 12.4.2, 12.5.2, and 13.4.3. Use of Monte Carlo simulation avoids the overstatement of risk that can result from more formulaic risk calculations. For example, if the desire is to reserve to a 90th percentile degree of certainty, using 90th percentile values of the distribution of two or more input parameters will likely result in a far greater than 90th percentile degree of certainty in the reserve. In a Monte Carlo simulation, many reruns of the valuation model are made based on sample points chosen randomly from the assumed distribution of each nonliquid variable, and with explicitly assumed correlations between variables. The 90th percentile of model outputs can then be estimated. Derman recommends a full simulation that includes both underlying behavior and trader hedging strategy. Section 11.3 contains an example that comes close to Derman’s proposed full simulation: a Monte Carlo simulation of dynamic hedging of a less liquid option (less liquid because of a nonstandard strike). Sampling over the simulation paths yields a statistical distribution of the differences between the payout on the option and the costs of the hedge. Derman’s recommendation of a full simulation including trader hedging strategy represents an ideal that may sometimes be difficult to achieve in practice. In the simulation in Section 11.3, a full simulation is possible because the assumed trader strategy is very simple, just varying the delta Model Risk 247 hedge of the underlying forward. Trader strategies that involve changes in options positions are much more difficult to simulate, because a full specification of the volatility surface is required at each node of the simulation. An illustration of this point can be found in the discussion of barrier options in Sections 12.3.2 and 12.3.3. When a full simulation is not practical, then I still believe that a simulation should be done, but computation can be simplified by restricting hedging strategies. Easier implementation comes at a cost of greater conservatism, since the full range of possible trader hedging strategies will not be captured. The simulations that I refer to in the next‐to‐last bullet point of the preceding list can serve as helpful paradigms. 8.4.4 Implications for Marking to Market Choosing a good liquid proxy, following the guidelines of Section 8.4.2, should assure that illiquid positions are marked to market to reflect changes in liquid market prices. To illustrate with the example used in that section, when there is a change in the implied volatility of one of the two stocks in the basket, it will be immediately reflected in the marking to market (MTM) of the liquid proxy and hence in the MTM of the basket option, which consists of the sum of the MTM of the liquid proxy and the reserve for the difference between the basket option and the liquid proxy. But should there also be an adjustment to the MTM of the illiquid position based on new information about parameters that cannot be sourced from a liquid market? Continuing with the same example, the question would be whether to change the MTM of the basket option based on new information about correlation between the two stocks. My answer would be that this should be done only very rarely. We have classified the correlation parameter as one that has no liquid market pricing source, so where would frequent updates be coming from? There are two possible sources: 1. Analysis of historical price data has led to a change in estimates of the correlation to be used. But this will only occur infrequently—if the correlation has been estimated from a long data history, then it will usually take months of new data before conclusions will change significantly. 2. There is evidence that the price being charged customers for correlation has changed. But since this is not a liquid market at which risk can be exited, the argument for making immediate use of such new data is not nearly as strong as it is for liquid instruments. In both cases, new information on correlation might ultimately impact the reserve for the difference between the basket option and the liquid 248 FINANCIAL RISK MANAGEMENT proxy, and thereby impact the total MTM of the basket option. But in both cases, you would expect to see this impact take place infrequently. In fact, I would argue for designing reserve calculations in a way that would make such changes extremely infrequent. For example, in this case, calculate the reserve based on an extremely unlikely level of correlation as opposed to an extremely unlikely change in correlation from the long‐term average. That makes it less likely that new information about a shift in the long‐term average will require a change in reserve level. The reason I want to make reserve changes infrequently is that I don’t think reserve level changes provide good incentives to traders and marketers. Changes in MTM of liquid instruments provide good incentives for exiting positions—either because stop‐loss limits are being breached or because accumulating losses cause traders to rethink the desirability of positions (this includes changes in MTM of liquid proxies, which can trigger hedging actions in liquid markets). But changes in reserve levels won’t provide much incentive to exit existing positions, since the illiquidity of the instrument makes such exits very difficult. It is true that raising reserve levels may send a signal to marketers to be more reluctant to book new trades, which might argue for raising reserve levels on new trades but not on existing ones. Even if you are convinced this policy makes good risk management sense, you still might be reluctant to have it guide the MTM reporting of the firm. Financial controllers, independent accounting firms, and regulators all tend to be suspicious of policies that involve high reserve levels that shield reported earnings from fluctuation; it looks like an attempt to smooth reported earnings. Let me make the following points concerning this: ■ ■ ■ I believe that the policies I am advocating here represent an accurate picture of what is known about earnings. The true earnings on illiquid positions are often not known until the trade matures. A highly conservative reserve level is therefore justified, and it is unreasonable to expect much new information to arrive from outside sources; the real information will come over time as the trade matures. There are exceptions—new information that would change your outlook for the whole distribution of an illiquid input. An example would be long‐term default rates on home mortgages in 2007 when new information on deteriorating underwriting standards would have impacted reserve levels that were previously viewed as prudently conservative. Reserving policies can be designed to assure independence and shielding from manipulation that attempts to use reserve levels to smooth earnings. See Section 6.1.4. These policies could help to deal with some of the concerns being expressed about the harmful impact MTM policies are having on bank Model Risk ■ ■ ■ 249 management (see the reducing procyclicality discussion in Section 5.5.8.1). MTM losses for liquid instruments encourage banks to shed volatile assets, but MTM losses for illiquid instruments, since the banks can’t shed the assets, result in a need to raise new capital, often in economic environments that are the most challenging for raising capital, leading to paralysis of the banking system. (This is discussed in greater detail, in the context of the 2007–2008 crisis, in Section 5.3.2.) My proposal causes large reserves to be taken up front, when the environment is still favorable for raising capital, and then releases the reserves, and hence frees capital for new investments, as the existing investments unwind. I would definitely advocate strong controls on the use of this accounting policy, only permitting it for positions the firm designates at the time of creation as illiquid. I have experience with a policy close to the one described working in practice over a several‐year period, from 1996 to 2003, at Chase and JPMorgan Chase, with the full knowledge of risk managers, financial controllers, independent accountants, and regulators. Reserve levels established were sufficiently conservative that they almost always proved adequate at an individual product level, and always proved more than adequate at an aggregated firm level. In the current environment, following the debacle of 2007–2008, it may no longer be possible to get independent accountants and regulators to go along with a policy like this; it requires more trust of the motivations of firm risk management than may now be achievable. In that case, I think risk managers should argue for keeping an internal set of accounts that most accurately reflects the economics of a business, even where this diverges from external reporting. 8.4.5 Implications for Risk Reporting In Section 8.3 we noted that for liquid instruments risk reports only need to look at exposures measured against current market prices, since future exposures due to changes in price levels can always be reduced utilizing the liquidity of the instrument. This approach will not work for illiquid instruments. To take an example, discussed at greater length in Section 12.1.4, a binary option might currently show very little gamma exposure but might have an unacceptably large gamma in the future if prices are close to the strike level when little time is left until option expiry. You can’t just wait to see if this will happen, since if it does you can’t count on being able to extinguish the risk by selling the digital option. You need to deal with this contingency at the time you are considering creating the option. 250 FINANCIAL RISK MANAGEMENT One way of handling this is to run risk reports at the time you are considering creating the position that look at a range of future possible price levels for future dates. Acceptability of possible future risk exposures are evaluated as part of the decision‐making process for taking on the position. Another way of handling this is to make sure that the liquid proxy and simulation methodology of Sections 8.4.2 and 8.4.3 adequately control possible future exposures. Continuing with the binary option example, you would make sure that the call spread liquid proxy chosen can only give rise to reasonable future gammas, by making sure that there is a sufficiently wide gap between the strikes of the call spread. As you will see in Section 12.1.4, widening the gap between the strikes will lead to more uncertainty in the simulation and hence higher reserve levels and tighter limits, but this should be viewed as a necessity for controlling future gamma exposure. 8.5 TRADING MODELS When a model is being used as part of a trading desk’s decision‐making process, it clearly requires internal model review by the model creators and users. For the model validation part of this process, it is particularly important to review how the model relates to the prevailing model being used in the market and to try to anticipate evolution of the prevailing market model, as argued in Section 8.2.6.3. The question I want to examine here is whether such models also require an external review by an independent group if the model is to be used only for trading decisions and not for the firm’s official valuations and measurement of risk. Major trading losses are frequently ascribed to the firm having the wrong model. What is often unclear in these claims is whether “having the wrong model” just means making incorrect forecasts about the future direction of market prices or if it means misleading the firm’s traders and managers about the nature of positions being taken. A good illustration is the discussion in Section 4.2.1 of whether the reliance by Long‐Term Capital Management (LTCM) on models should be viewed as a primary cause of the collapse of the fund. Any firm engaged in making markets or investing funds must take positions whose profit or loss will depend on the correctness of forecasts of moves in market prices. Different strategies will be tied to different price relationships. Some depend on overall market direction, whereas others depend on the relative price of related assets; some depend on getting a long‐ term trend right, whereas others depend on correctly anticipating short‐ term moves. However, traders will always need to make judgments about an uncertain future, and firm managers in turn will always need to make Model Risk 251 judgments about how much of a risk of loss they will allow a trader to take in exchange for a possible gain. When making this assessment, management will be guided by evidence of prior accuracy of the trader’s forecasts. Nothing in the last paragraph will be altered by whether a trader uses a model as a computational aid in forecasting, unless perhaps management is lulled into a false sense of security by believing that the use of a model lessens the chance of errors in trading judgment. However, if a model results, either purposely or inadvertently, in misleading traders and managers about the relationship between positions being taken and the size of possible losses, then the accusation that model error resulted in the loss is far more plausible. For example, a spot foreign exchange (FX) trader could be using a very complex model when deciding which positions to take. This could even extend as far as program trading, in which a computer actually issues the buy and sell instructions based on model output. However, spot FX positions can easily be valued based on external quotes, and position size is extremely easy to understand without the aid of models (see the discussion in Sections 9.1 and 9.2). So it is easy for management to see what the profit and loss (P&L) is every day and to cut the risk if P&L performance has been poor. Thus, the modeling does not have any of the dangers of hidden risk, such as Ponzi schemes (see Section 2.2). No FX trader would dream of asking to report more profits this year because he can “prove” that his model (or trading style) will work better next year than this year. When I was in the position of managing the independent model reviews for a firm, I argued strongly against my group reviewing the validity of models that were being used only for trading decisions. Partly, this was an attempt to conserve resources for what I viewed as the more important task of validating models used for valuation and risk measurement. But even more, I was concerned that traders would use model validation by my independent reviewers as a stamp of approval that would discourage critical review of trading strategies by senior management. I argued that since we weren’t being asked to review trading strategies that didn’t involve models, the use of a model did not transform us into experts on trading strategy. In particular, how could we obtain the insider knowledge that could allow us to anticipate evolution of the prevailing market model? This is the position I advocated in the first edition of this book, but on reflection, I would reconsider my previous stance. When position limits are being set and when actions following a stop‐ loss limit overage are being reviewed, there is no question that traders will utilize results from their trading models to make their case to senior managers. Since senior managers will not have the time or, usually, the skill set to form their own judgment of these models, it is only by having independent 252 FINANCIAL RISK MANAGEMENT reviewers look at the models that an effective challenge to trader claims can be prepared. Independent reviewers must make clear the limited scope of their review, but can certainly raise issues concerning possible cherry‐picking of historical data or reasons why shifts in the economic environment might bring conclusions based on historical data into question. These challenges may prove of value to traders as well as senior managers. And certainly, independent review of model mechanics—the model verification of Sections 8.2.3, 8.2.4, and 8.2.5—can add value. FRB (2011) seems quite clearly to endorse independent review of trading models. Its Section III, which examines the criteria for which models need to be subject to the review standards of the document, states, “Models meeting this definition might be used for analyzing business strategies, [and] informing business decisions” and “The definition of model also covers quantitative approaches whose inputs are partially or wholly qualitative or based on expert judgment, provided the output is quantitative in nature.” CHAPTER 9 Managing Spot Risk pot trades are trades that involve an immediate exchange. This includes trades such as purchases of stock, purchases of gold, and exchanges of one currency for another. It excludes trades that involve a promise to deliver at some future time. Most of our study of risk involves future promises to deliver—unconditional promises constitute forward transactions, and promises whose payments are predicated on some future condition constitute options transactions. The mathematical modeling and risk management of forwards and options are far more complex than the corresponding elements of spot transactions, and far more space in this book is devoted to forwards and options than to spot positions. However, positions in spot trades often constitute the largest portion of a firm’s risk. Spot transactions are also the fundamental building blocks for valuing and risk managing forward and option positions. We can find the present value equivalent of a set of forward cash flows or the delta equivalent of an options position, but we then need to be able to value and risk manage these resulting spot positions. So a brief survey of the management of spot risk is in order. S 9.1 OVERVIEW All instruments traded by financial firms are commodities in the sense of not being individually identifiable. (If I borrow—that is, rent—a house from you, you expect me to return that exact same house, so houses are not a commodity; this is not true for dollar bills, bars of gold, barrels of oil, shares of IBM stock, specified amounts of a given bond, and so on.) This commodity feature means that traders are free to sell before they buy, since they can always borrow the instrument in order to make delivery. In this way, financial markets are more symmetrical than noncommodity markets such as houses, where you must build up an inventory by buying before you can sell. 253 254 FINANCIAL RISK MANAGEMENT Commodities can be divided into physical commodities, such as gold and oil, and financial commodities, such as stocks, bonds, and currencies. We do not study any trading in bonds in this chapter. Since bonds represent a fixed obligation to deliver an amount of currency, they are studied in Chapter 10 on managing forward risk. A general convention in the market is to use the term commodities to mean physical commodities only. Financial commodities are now almost universally transferable from one location to another in electronic form, so they have negligible transportation and storage costs per unit. Physical commodities have nonnegligible transportation and storage costs, which will have consequences we will study shortly. Let us begin by looking at the hedging activities of a market maker in the dollar versus yen spot foreign exchange (or to adopt the terminology of that market, USD–JPY FX). In terms of instruments used, this represents the simplest type of trading possible—it is completely one‐dimensional. The trader’s position at any point in time can be represented as either long or short a certain quantity of JPY (or, completely equivalently, short or long a certain quantity of USD). In a more complex spot market, such as the commodities market for wheat, a trader’s position would need to reflect being long or short different grades of wheat. However, currencies do not have grades—$1 million is $1 million, whether it is made up of 10,000 $100 bills, 100,000 $10 bills, 1,000,000 $1 bills, or 100,000,000 pennies. Our market maker will receive orders throughout the day from customers who are either looking to sell JPY and buy USD or looking to sell USD and buy JPY. Each customer will state the quantity of USD she wishes to sell and ask for a bid of the quantity of JPY that the market maker will exchange for it, or state the quantity of JPY she wishes to sell and ask for a bid of the quantity of USD the market maker will exchange for it. Trading screens are available at all times that show the best bids currently available from other market makers for selling JPY in exchange for USD and for selling USD in exchange for JPY. Market makers are constantly submitting their own bids for these two trades for the consideration of other market makers. When a customer’s inquiry is for a small enough quantity, the market maker can guarantee a profit by quoting a bid just slightly higher than the best bid currently quoted on the trading screen, and if the customer accepts the bid, the market maker will immediately be able to close out the position created by hitting the bid quoted on the trading screen and making the small differences between the two as profit. The market maker is only required to decide how much of a margin to build into the quote to the customer. The higher the margin, the higher the profit, but the greater the chance that the customer will turn down the quote and seek a quote from another market maker. The size of margin quoted must depend on the market maker’s knowledge of the customer—how likely Managing Spot Risk 255 is this customer to be polling a large number of market makers simultaneously rather than just coming to a single firm seeking a quote? In practice, the decision making at a firm will probably be divided up between a trader and a salesperson. The salesperson, who has a close knowledge of and continuing relationship with the customer, will bear the primary responsibility for determining the size of margin quoted. The trader will be credited, in the internal record keeping of the firm, with only a small portion of this margin. A trader who followed this risk‐averse a strategy would be unlikely to retain a job for long. The firm would probably judge that the profit the trader was making for the firm was not worth the opportunity cost of the trading seat. Higher profits would likely come from giving the seat to a more aggressive trader who would choose to take some risk by not closing positions out immediately. It is true that more aggressive traders are running the risk that prices will move against them, but, assuming that the firm sees a decent flow of customer orders, it is likely that a customer order will soon come in on the other side, and, on average, over time, the spread between the bid on each side of the market will be greater than losses from price movement through time. When a large customer order comes in, then the market maker has no choice but to take some risk—the only choice is how to divide the risk between the liquidity risk of trying to offset the position immediately and the basis risk of offsetting the position over time. With a large order, the trader can no longer count on being able to close the position out at the price posted on the trading screen since this quote will only be for a reasonably small transaction. Of course, the customer will be charged a premium for the liquidity risk posed by the size of the order, which will provide some cushion to the trader against the risk that must be taken. The trader needs to make a judgment as to the relationship of this large customer order to overall market conditions. Is it an order that simply reflects the idiosyncratic circumstances of this customer, perhaps a payment that needs to be made in the customer’s business? In this case, it is unlikely that a relationship exists between the order and any price trend in the market. Unless the trader has some other reason to believe that the market will be trending in a direction that will cause losses to this position, it will be better to close the position slowly, relying on customer orders and small trades with other market makers, minimizing liquidity risk. However, if the large customer order is likely to be part of a large movement, such as a customer wanting protection against the announcement of economic data that may impact the market, it may be better to close the position more quickly, bearing some liquidity cost in order to reduce the exposure to market trend. Almgren and Chriss (2001) show how to calculate the efficient frontier of strategies that have the optimal trade‐off between the liquidity costs of 256 FINANCIAL RISK MANAGEMENT offsetting the position in large blocks and the volatility risk (which we call basis risk) that the price at which the offset occurs differs from the price at which the position was put on. In the absence of price drift, the strategy that minimizes liquidity cost is one in which position covering is spread out over as long a period of time as possible, minimizing transaction size, and the strategy that minimizes volatility risk is one in which the entire position is offset at once, with as little chance for prices to change as possible. Thus far, we have pointed out two advantages of seeing good customer order flow to a market‐making firm: the increased likelihood of closing out positions at the favorable side of the bid spread and knowledge about the motives behind large orders. There are other advantages as well. Working with customers closely enables a firm to anticipate a large order and allows positions to accumulate through customer flow to meet part of the order in advance, thereby further lowering liquidity risk. When a firm’s traders have a market view and want to put on a position, customer order flow enables them to put positions on and close out the positions more cheaply than if all positioning had to be done by aggressively seeking bids from other market makers. All of these advantages of customer order flow and the trade‐offs of liquidity versus basis risk are present in all market‐making activities, but can be observed in their purest form in spot risk market making, where other complicating factors do not intrude. Even for the simplest spot product, FX spot, positions can be closed over time in other possible ways. For example, another source of liquidity is to spread out the closing of the position between the spot FX market and forward FX markets. This introduces a new basis risk in the form of the risk of unfavorable interest rate movements between the time the forward position is put on and the time it is closed out, but lowers the time basis risk. The trader must judge which is the most favorable risk mix. A trader in the currency of a smaller economy, let us say one trading the Danish krone against the dollar, might choose to temporarily hedge some of a position by a euro‐ USD trade that will eventually be closed out by a krone‐euro trade. Adding a leg to the trade adds transaction costs, but euro‐USD has more liquidity than krone‐dollar and the trader’s judgment may be that the basis risk of a krone‐euro position is considerably smaller than that of a krone‐USD position, given the closer tie of the Danish economy to the economy of the euro bloc countries than to the U.S. economy. When we move to more complex spot products such as commodities or equities, the potential avenues for redirecting basis risk multiply enormously. A position in IBM stock could be temporarily hedged by a Standard & Poor’s (S&P) index future, judging this basis risk to be smaller than an outright IBM stock position. A position in one grade of wheat could be temporarily hedged with a position in another grade of wheat that trades with greater liquidity. Managing Spot Risk 257 Firm‐level risk management for spot risk is relatively straightforward. The more liquid spot positions can be valued by directly obtaining market prices. As a result, it is not necessary to utilize models for valuation and to establish reserves against possible model errors. Most spot markets are liquid enough that prices can be obtained from trading screens or closing prices on public exchanges, so it is not even necessary to arrange for a price collection from brokers. For market‐making trading desks with reasonable customer order flow, positions should be marked to midmarket, since the presumption is that, on average, most positions can be unwound without needing to aggressively seek bids from other market makers. The only adjustment that might arise with any frequency is a reserve against liquidity risk if a spot position grows sufficiently large relative to the size of customer order flow that significant liquidity costs may arise in closing the position. For proprietary trading desks, positions should generally be marked to the side of the bid‐ask spread that is least favorable for the position, since, in the absence of customer order flow, it should be presumed that closing out the position will require aggressively seeking bids from market makers. Less liquid spot markets may require some form of modeling for valuation purposes. For example, an over‐the‐counter stock that does not trade very often or a commodity grade that is thinly traded may not have readily available price quotes. A model may need to be established that relates this price to the price of a more liquid instrument. For example, the over‐the‐ counter stock price could be priced in relationship to a stock index, or a less liquid commodity grade could be priced as a spread to a more liquid commodity grade. In this way, the valuation can be updated daily based on quotes for the more liquid instruments. The relationship can be reestimated less frequently as reliable trading prices for the less liquid instrument are obtained. When models of this type are used, a reserve is needed against the statistical uncertainty of the relationship between liquid and less liquid prices being utilized. The issues of nonstatistical limits and risk reporting to senior management for spot positions center completely on issues of which positions should be grouped together, since the position in any particular spot instrument is a single number. We’ll discuss this issue for each of the spot markets: first FX, then equity, and finally physical commodities. 9.2 FOREIGN EXCHANGE SPOT RISK To consider a concrete example, a USD‐based firm will want to limit and report to senior management its net FX spot exposure to USD. This firm 258 FINANCIAL RISK MANAGEMENT will also want to have individual currency limits for FX spot exposure for every currency it trades. It will set limit sizes relative to the overall liquidity of the market for that currency and the firm’s degree of customer order flow in that currency to ensure that traders have explicit management approval to build up positions that will require large time periods to reverse. However, senior management would probably need to be informed only of the largest individual currency positions. The remaining decision is determining which currency groupings are the best to use in setting net FX spot exposure limits and reporting to senior management. For example, does a grouping of all‐Asian currencies make sense? A grouping of all‐Asian currencies excluding the yen, Australian dollar, and New Zealand dollar? Should Asian currencies be divided into groupings based on national gross domestic product (GDP) per person? Should all currencies of countries with lower GDP per person be grouped together as emerging market currencies? Each firm will reach its own conclusions based on economic theory, trading experience, and, perhaps, statistical analysis of which currency movements tend to occur together. 9.3 EQUITY SPOT RISK Equity reporting and limits can begin from a similar starting point as for FX. There should be reporting and limits for positions in individual stocks, for an overall long (or short) net position in all stocks, and for groupings by geographic region. Decisions on whether to group together stocks in all companies based in Europe or based in emerging markets is subject to the same type of analysis as the decisions for FX. But geography is just a starting point for stocks. There are several other considerations: industry and industry sectors, and style. Much research has been devoted to which factors play the largest role in explaining the performance of equity managers, an issue that is known as performance attribution; the classic article in this area is Sharpe (1992), which was highly influential in the recognition of the importance of stocks of smaller‐capitalization firms versus larger‐capitalization firms and growth stocks versus value stocks as important style attributes in explaining performance. Much of this analysis translates very directly to how to group stocks together for purposes of risk reporting and limits. Here are examples of some popular classifications for performance attribution: ■ Morningstar, in its evaluations of mutual funds that invest in equities, has created a very influential style box based on smaller‐capitalization Managing Spot Risk ■ 9.4 259 firms (less than $2 billion) versus larger‐capitalization firms (more than $10 billion) and growth stocks versus value stocks. Morningstar follows Sharpe (1992) in defining growth stocks as those with little or no dividend payout, high price‐to‐book and price‐to‐earnings ratios, and promising capital appreciation, and value stocks as those likely to pay high dividends but with low price‐to‐book and price‐to‐earnings ratios. The Global Industry Classification Standard (GICS) developed by Morgan Stanley Capital International (MSCI) and Standard & Poor’s (S&P) consists of 10 sectors, 24 industry groups, 68 industries, and 154 subindustries. The 10 sectors are energy, materials, industrials, consumer discretionary, consumer staples, health care, financials, information technology, telecommunication services, and utilities. PHYSICAL COMMODITIES SPOT RISK Physical commodities are further complicated by the presence of transportation costs, which leads to different markets for the same commodity in different locations (for example, oil for delivery in Seattle is a different product from oil for delivery in El Paso). This plays a role in valuation, since delivery at a location where liquid prices are not available could be priced using a model based on a more liquid price for delivery at another location and estimated transportation cost between the two locations. It also plays a role in the design of limits and reporting. Locations that are reasonably closely related in price, by having low transportation costs between them, should have their positions summed into a net position for reporting and perhaps limits. An interesting analogy can be made between location relationships based on transportation costs and relationships between forward prices for different time periods. In Section 10.3.2, we will see that some commodities have forward prices for different times tightly linked by the possibility of cash‐and‐carry arbitrage. It is instructive to think of this as a form of location relationship, with the storage and financing costs as the cost of “transporting” the commodity from one time period to a later one. Just as transportation can be so expensive between some locations that they virtually form independent markets, storage can be so expensive for some commodities, such as electricity, as to virtually eliminate the possibility of cash‐and‐carry arbitrage. However, although transportation costs are almost always symmetrical (it costs just as much to ship from A to B as from B to A), a commodity cannot be transported from a later period to a former period, so cash‐and‐carry arbitrage works only in the forward direction. 260 FINANCIAL RISK MANAGEMENT Other types of potential transformations besides location play a role in physical commodities. To take two examples from McDonald (2006, Chapter 6): ■ ■ Soybeans can be crushed to produce soybean meal and soybean oil. A trader with a position in soybean futures and an opposite position in equivalent quantities of soybean meal and soybean oil futures is trading the crush spread. The trader is taking a position not on what will happen to the cost of soybeans but on what will happen to the cost of processing soybeans into soybean meal and soybean oil. To the extent that positions in soybeans and soybean meal and soybean oil offset, the resulting position should be reported and limits set on the crush spread and not on the individual legs. Crude oil can be separated into different petroleum products such as heating oil and gasoline by a refining process known as cracking. A trader with a position in crude oil futures and an opposite position in equivalent quantities in heating oil and gasoline futures is trading the crack spread. The trader is taking a position not on what will happen to the cost of crude oil but on what will happen to the cost of processing crude oil into heating oil and gasoline. To the extent that positions in crude oil and heating oil and gasoline offset, the resulting position should be reported and limits set on the crack spread and not on the individual legs. Reports should also be designed and limits set on aggregated positions across physical commodities whose prices tend to be highly correlated. So there might be an overall limit on total net long (or short) exposure to all energy products, summed over crude oil, heating oil, gasoline, natural gas, and electricity. Which products get grouped together may differ by firm, based on economic theory, trading experience, and statistical analysis. EXERCISE 9.1 Simulation of the Impact of Trading Rules on Expected Return and Risk A market maker in a spot market is operating under the constraint that she must close out her position by the end of each trading day. We want to see the impact of different possible trading limits on the size of the position that can be built up. Managing Spot Risk 261 Divide the trading day into 100 time segments. In each time segment except the last, there is a 50 percent chance of receiving a customer order for one unit. A customer order has a 50 percent chance of being a buy and a 50 percent chance of being a sell. Customers pay $0.10 per trade in transaction costs. So if the midmarket price is $100.00, a customer will purchase at $100.10 and sell at $99.90. The market maker cannot close out a trade without waiting at least one period. Midmarket price changes from one period to the next are normally distributed with a standard deviation of $0.10 (assume a starting midmarket price of $100.00). The market maker must close out her open position by the last trading period. She pays $0.05 per trade in transaction costs to close positions with another market maker. So if the midmarket price is $100.00, she sells positions at $99.95 and purchases at $100.05. It is to the market maker’s advantage if she can wait until a customer order comes in to close out her position, since she will make a $0.10 transaction spread on each side of the trade, for a total of $0.20, rather than making only $0.10 minus $0.05, for a total of $0.05 in transaction spread by closing out with another market maker. However, the longer she waits for a customer order, the greater her risk of prices moving against her. Simulate a set of trading rules to see the trade‐off between expected return and risk. Use 1,000 paths for each simulation. The measure of expected return should be simply the average over these paths. You can choose any reasonable measure of risk, such as the 95th percentile loss or the standard deviation. One trading rule should be to never close out until the last period. Another should be to always close out in the period immediately after the customer trade. Intermediate rules can be based on a limit of how large the absolute size of a position is allowed to grow— when the position gets larger than this limit, the excess must be closed out. 1. Determine the impact on the risk/return trade‐off of a lower standard deviation of the midmarket price of $0.05 per period. 2. For a more extended exercise, you could experiment with more complex trading rules, such as having the transaction cost for closing a position be an increasing function of the absolute size of position to be closed, or allowing the market maker to influence the probability of customer trades being buys or sells by shifting her quoted price away from the midmarket price. CHAPTER 10 Managing Forward Risk anaging forward risk is considerably more complex than managing spot risk due to the large number of dates on which forward payments can take place. With some forward markets going out to 30 years and even beyond, even if we restrict deliveries to take place on the 250 business days of a year, it still leaves 30 × 250 = 7,500 days on which future flows can occur, each of which requires a mark‐to‐market valuation and risk measurement. It is clearly impractical to have liquid market quotations for each possible forward, so modeling needs to be heavily relied upon. Having a spot versus forward position is an interest rate differential position, not a price view. If I believe the market will get a surprise announcement that will raise the stock price, even if I think it will not come for three months, I don’t want to be long the forward and short the spot. When the announcement comes, both will be roughly equally impacted. I want this position only if the announcement I expect is something like a one‐shot dividend that will impact the relative value of the spot and forward. If I put on a long forward and short spot position, I’m taking a view on the interest rate. Let me cite a real example. On June 24, 1998, a trader was holding a long forward position in Telecom stock against which he was short the stock. AT&T announced plans to purchase Telecom at a sizable premium, but the trader wound up with a sizable loss. Why? His outright position in Telecom stock was even, so he didn’t gain from the rise in the stock price. Telecom had never paid a dividend, so the forward traded at a large premium to the cash. As soon as the market anticipated that the stock could be traded for a dividend‐bearing AT&T stock, this forward‐to‐cash premium shrunk significantly since it was now less expensive to hold a cash position in the stock for delivery into a forward sale. The difference between an outright position and a borrowing or lending position is the difference between wanting to hold an asset as a good investment (you expect it to gain value) versus wanting to make use of an asset. Consider a house. When you buy it, you get a combination of an investment M 263 264 FINANCIAL RISK MANAGEMENT and a place to live. You might want to split the two. If you like it as an investment but don’t want to live there, you can buy it and lend it to someone (rent it out). If you want to live in it but don’t like it as an investment, you should borrow it (rent it) rather than buy it. Similarly, a firm that is in the business of milling wheat and is running short of wheat supply to keep its production process going but does not like wheat as an investment (does not believe it will go up in price) will seek to borrow wheat rather than buy it (although borrowing may take the form of buying spot wheat while selling forward wheat). Likewise, a firm that likes wheat as an investment but does not need it for any production process will buy wheat and then lend it out (possibly combining the two steps into one by buying forward wheat). Although a clear distinction can be made between an outright spot position and a borrowing or lending position, they also share close relationships. As we saw in the discussion of spot risk management in Chapter 9, maintaining a spot risk position over a longer period than a single trading day requires some form of borrowing or lending. In some markets, the use of borrowing or lending to maintain outright spot risk positions becomes such a dominant force that it is the principal driver of interest rate movements in the market. In many trades, such as forward purchases and sales, spot and forward risk are bound together, so it will be necessary to study the interactions between these two risks to fully understand the dynamics of forward risk management. It is important for the risk management function to clearly separate spot risk from forward risk in transactions in which they are bundled to ensure that all the firm’s spot risk in a given asset is reported and managed in a unified fashion. The borrowing and lending markets in currencies and gold started as a means for businesses and individuals to adjust the timing between when income is earned and when purchases are made. Borrowing and lending in other commodities started with users and suppliers of the commodity satisfying short‐term needs, as in the previous milling example. Borrowing in stocks and bonds started with the need for short sellers, who want to act on the view that an asset will decline in value, needing to first borrow what they wanted to sell short. Borrowing to support short selling is also a feature of all the other borrowing markets. Once borrowing and lending markets are established, they begin to attract investors, speculators, and hedgers who have views on the borrowing rate rather than on the asset price. So one trader who believes that a particular borrowing rate will soon decline will lend at that rate solely in hopes that he can match that lending with a borrowing at a lower rate when the rate declines. Another trader might believe that the borrowing rate for May 2015 is too high relative to the borrowing rate for April 2015 so she Managing Forward Risk 265 will borrow for April and lend for May, hoping to reverse the transactions when rates return to a more normal relationship. Another trader might believe that borrowing rates for a particular corporation will decline relative to those of another corporation or the government, so he will lend to the former by buying its bond and borrow to support a short sale of the latter’s bond. A business firm worried about the possible impact of high borrowing costs on its financial health in 2017 will borrow funds now that do not become available until 2017. The emphasis I am placing on borrowing and lending rates as the foundation of forward risk is somewhat nonstandard; but see Williams (1986) for an incisive economic analysis of forward, futures, and lending markets for commodities using this approach; also see Brown (2012, Chapter 10) for an excellent discussion along similar lines. A more conventional exposition, such as Hull (2012, Chapter 5), would focus on borrowing rates only for currencies and would analyze forward risk on commodities and securities through forward contracts that involve exchanging the commodity or security for currency. The borrowing rate on the commodity or security still comes into play as one of the inputs determining the price of the forward or implied by the price of the forward. The two methods are mathematically equivalent, so choosing between them is a matter of deciding which is the most convenient and supplies the greatest financial insight. My choice of emphasis is based on the following considerations: ■ ■ ■ Direct borrowing and lending markets exist for many assets—such as gold, stocks, and government bonds—that do not require any involvement with borrowing/lending risk on currencies. Let’s look at an example. Suppose that the rate for borrowing gold for three months is 2 percent annualized. If I want to borrow 1,000 ounces of gold today, I must be prepared to return 1,000 × (1 + 2% × 3/12) = 1,005 ounces of gold in three months. No mention has been made of any currency— there is simply an equivalence of a certain amount of gold on one date and some other amount of gold on another date. A uniform approach to all underlying instruments makes for easier exposition of some concepts. For example, Section 10.2 on mathematical models for forward risk is built around a single discount curve that could represent borrowing costs for a currency, but could represent borrowing costs for a security or commodity equally well. It is consistent with a risk management viewpoint in which, for example, it is natural for a gold trader to be taking risk with regard to gold borrowing rates, but not with regard to dollar borrowing rates. Gold borrowing costs are primarily impacted by economic factors unique to 266 FINANCIAL RISK MANAGEMENT the gold market, including the supply and demand for gold, so it would be a sound risk management practice for the same trading desk to run risks in the gold spot and borrowing rates. However, there is little linkage between gold and dollar borrowing rates. A gold trader running dollar borrowing risks through the vehicle of positions in gold/dollar forwards requires serious management scrutiny. At a minimum, dollar interest rate exposures taken in this way need to be reported and aggregated together with other dollar rate risks throughout the firm. Similar comments apply to borrowing risk on other commodities and securities. The primary argument against a borrowing rate focus is that for some assets, such as oil, no borrowing market exists, requiring forward risk to be managed through forward contracts. Even for some assets for which a borrowing market does exist, the borrowing market has considerably less liquidity than the comparable forward contract. However, it is always possible to take spot and forward prices and currency interest rates and derive implied asset borrowing rates that can then be used just as if they had been obtained by a direct quote. Indeed, even in some currency markets, the most liquid source for rate quotes is to combine forward foreign exchange (FX) prices with dollar rates to derive interest rates for the currency. This is no bar to developing discount curves for the currency or combining directly obtained rates that are the most liquid price source for some maturity segments with implied rates for other maturity segments and using them to form a single discount curve. Within the fixed‐income departments of investment banks, it is customary to find separate trading desks for interest rate and credit products, with interest rate trading focused exclusively on changes in credit risk‐free rates and credit trading focused exclusively on changes in the credit spread to risk‐free rates. Clearly, some products cut across this boundary—a fixed‐rate bond issued by a corporation will change in value because of both changes in risk‐free rates and changes in credit spreads. But interest rate swaps that convert fixed‐rate into floating‐rate payments can be used to transform a fixed‐rate corporate bond into an instrument that is almost totally dependent on credit spread, so trading desks can utilize internal transfers to almost completely separate the two types of exposure. We will want to follow this division in studying risk. While there is a certain amount of overlap between interest rate risk and credit risk measurement and modeling, particularly in extracting term structure from market prices, the differences are greater than the similarities: ■ Option products are very important instruments in interest rate trading, requiring the modification of traditional option models to more Managing Forward Risk ■ 267 complex multiple‐tenor environment. Option products are currently of negligible importance in credit trading. Credit modeling focuses on correlation between debt and equity within a firm and between debt of different firms. There are no comparable issues for interest rate models. Consequently, we will focus only on products free of credit risk in this chapter, reserving the study of credit risk management for Chapter 13. Strictly speaking, it is only bonds issued by the central government (for example, U.S. Treasury bills and bonds in the United States) that are (nearly) completely free of credit risks. It is only the central government that has unlimited power to issue its own currency and so can (nearly) certainly meet any obligations to pay that currency. But fixed‐income trading desks of investment banks generally also trade a variety of instruments whose credit risk is extremely low: bonds issued by agencies of the central government, mortgages guaranteed by such agencies, and derivatives tied to bank indexes such as the London Interbank Offered Rate (LIBOR). This latter case is a particularly important class for interest rate products; indeed, the largest interest rate risk exposures of financial institutions is usually to LIBOR products: LIBOR futures, forward rate agreements, swaps, caps, floors, and swaptions. So we ought to examine closely why credit risk on these products is considered negligible and why it predominates over Treasury rates as the basis for derivative products. First, we need to distinguish between the credit risk to the counterparty on a derivative and the credit risk on the derivative instrument itself. Consider the example of a typical derivative tied to LIBOR, a 10‐year interest rate swap of fixed coupon payments against three‐month US Dollar LIBOR reset each quarter. Certainly there is credit risk that the counterparty will default on its obligations under this contract, with the severity of risk tied to the creditworthiness of the counterparty. But this has nothing to do with credit risk of the swap itself. An equity option or foreign exchange swap or option on a Treasury bond would also entail counterparty credit risk but no underlying credit risk. By contrast, a default swap in which you must pay Company A an agreed amount based on the default of Company B against fixed payments to you from company A entails both counterparty credit risk of losing your fixed payments if company A defaults and underlying credit risk of Company B defaulting. So we need to see whether there is any underlying credit risk on a bank index product. Let us continue with our example of the 10‐year swap based on three‐month US Dollar LIBOR resets. There is clearly some credit risk—a severe economic downturn will raise concern about potential bank defaults and therefore raise the rate that banks need to pay on three‐month deposits 268 FINANCIAL RISK MANAGEMENT relative to three‐month Treasury bill rates. But to see just how small this element of credit risk is, let us contrast it with the credit risk on a 10‐year bond issued by one of the banks whose deposit rates form the LIBOR index, noting that the credit risk on this bond is very small to begin with, since all banks in the index are of very high credit quality, generally Aa. The credit spread on the 10‐year bond needs to reflect the probability of default over a 10‐year period, which includes scenarios in which the creditworthiness of the bank declines severely prior to the default. But these scenarios will have little impact on the average LIBOR index over the 10 years, since a bank that declines in creditworthiness will be replaced in the panel that determines the LIBOR index. For example, Moody’s data for a 20‐year period shows that 0.81% of Aa‐rated firms defaulted within 10 years of the rating, but only 0.02% of Aa firms defaulted within one year of an Aa rating. Furthermore, even in the event of default, the chances of depositors losing money are very small since bank regulators are primarily concerned with protecting depositors and take steps to ensure that losses will be absorbed by stockholders and bondholders but not depositors. Spreads between LIBOR‐based rates and Treasury rates therefore primarily reflect the superior liquidity of Treasuries and their value as collateral. Under the very extreme conditions of the global banking crisis of 2007–2008, there was a period in which a spread between LIBOR and Treasury rates based on credit concerns came to play a major role (see Tuckman and Serrat 2012, 431–432), but this is a very rare occurrence. Individual government issues have idiosyncratic characteristics (liquidity, borrowing rates, country of issue for euros) getting in the way of creation of a single discount curve against time. This is closely tied to government bonds being in fixed supply as opposed to swaps, which can be freely created. Government rates represent only investment rates for most firms and not borrowing rates (you can only borrow at government rates if you have government bonds available as collateral) while deposit rates are two‐sided, at least for the large financial institutions that serve as market makers. This has led to LIBOR being the risk‐free rate generally used to price derivatives, such as futures, forwards, interest rate swaps, interest rate options, and credit swaps, and as a target against which to measure borrowing rates. This explains why LIBOR is far more popular than government rates as a basis for derivatives used to hedge interest rate risk. As this book is going to press, news stories about manipulation of the LIBOR rate setting process are raising questions that could threaten the popularity of LIBOR as a basis for derivatives. As this story develops, its consequences will be addressed on this book’s website. A good summary of the issues raised by this manipulation of LIBOR rate setting is the article “The Rotten Heart of Finance” in the July 7, 2012, Managing Forward Risk 269 issue of the Economistt magazine. The website of the British Bankers Association (www.bbalibor.com), the organization in charge of determining LIBOR, has many articles giving details of the LIBOR‐setting process. When derivative contracts based on bank deposit rates were designed, a significant worry was that if a derivative referenced the rate set by a particular bank, that bank might manipulate the rates at which it bid for deposits in order to generate profits in its derivatives holdings. The decision was made to tie derivatives products to an index of bank deposit rates, which would be harder for any one bank to manipulate. A large panel of banks is selected (currently 16 for US Dollar LIBOR), based on criteria of expertise and prominence in the market and the highest degree of credit worthiness (any bank no longer meeting these requirements would be replaced in the panel). The highest and lowest quartiles of submitted rates are dropped, to minimize any potential for manipulation, and only the middle two quartiles averaged (in addition, any bank not operating within the spirit of the rules would be dropped from the panel). But when the global banking crisis of 2007–2008 caused a large decline in the use of interbank deposits, the lack of market liquidity may have opened the door to potential manipulation. Given the complexities of forward risk management, we will need to carefully organize our study into the following sections: ■ ■ ■ ■ Section 10.1. This is a study of the variety of instruments that entail forward risk and that can be used to manage forward risk. The large variety of structures in which spot and forward risk (and occasionally implicit options risk) are woven together means that an important part of risk analysis is often just making sure that all the risks of a particular trade have been properly identified. In addition to the market risks, slight variations in structure, which may result in virtually identical spot and forward risk, can have large differences in credit risk, legal risk, and funding liquidity risk. Section 10.2. This section provides a study of the mathematical models used to value and measure forward risks. Although these models have been used heavily for many years and a great deal of consensus has been built up around them, enough subtle issues remain to merit a careful understanding of the residual risks of model uncertainty. Section 10.3. This section takes a brief look at the factors that impact borrowing and lending costs. Although this is not primarily a book about economics, at least some familiarity with the determinants of forward prices is necessary to properly understand the requirements for designing a risk management structure for forward risks. Section 10.4. This section provides a study of how to design a risk management reporting system for forward risk. 270 FINANCIAL RISK MANAGEMENT 10.1 INSTRUMENTS The management of forward risk can involve a number of different instruments that can be used to take on the same market risk position. These instruments may differ in legal form, with different regulatory consequences and standing in bankruptcy proceedings, and have different implications for credit risk and funding liquidity risk. They also differ in the extent to which they bundle together spot and forward risk. We consider each of the following categories: ■ ■ ■ ■ ■ ■ ■ ■ Direct borrowing and lending Repurchase agreements. Forwards. Futures. Forward rate agreements (FRAs) Interest rate swaps Total return swaps Asset‐backed securities 10.1.1 Direct Borrowing and Lending Suppose a trader wants to sell a given asset short. In a number of asset markets—such as stocks, bonds, currencies, and gold—the asset can be borrowed directly in order to sell short. Other markets, such as most physical commodities, have not developed direct borrowing products. One drawback to using borrowing as the means of obtaining an asset to short is that it creates a sizable credit risk and funding liquidity risk for the asset lender, who could lose the entire value of the asset if the borrower defaults and who has to finance the asset that has been lent. The borrower may be paying for credit usage that is not really needed, since the cash raised by selling the asset short is incidental to the original objective of selling the asset short to position for a price drop. One solution is to use the cash raised as collateral against the borrowing. This reduces the credit risk for the asset lender, who can hold on to the cash collateral in case of borrower default, and reduces the funding liquidity risk, since the cash collateral received by the asset lender can be used to fund the asset purchase. Providing cash collateral to the asset lender creates credit risk for the asset borrower, even though this is mitigated by the value of the asset, which does not need to be returned if the recipient of the cash collateral defaults. Managing Forward Risk 271 10.1.2 Repurchase Agreements In the previous example, one party borrows the asset and provides cash collateral to the other party. An equivalent way of describing the same trade is to say that one party borrows the asset and lends cash, whereas the other party borrows cash and lends the asset. Yet another equivalent way of describing the same trade is a transaction in which the first party purchases the asset for cash and, at the same time, contracts to sell the asset back to the second party at an agreed forward date for an agreed cash price. Table 10.1 demonstrates that all three ways of describing this transaction are equivalent in terms of the flows of cash and asset. TABLE 10.1 Alternative Descriptions of an Asset Borrowing Collateralized by Cash Description 1 Today A borrows $1 million par amount of a Treasury bond from B. A sells the bond in the market and receives $980,000. A places the $980,000 as collateral with B. One month from today A buys the $1 million par bond in the market and returns it to B. B returns the $980,000 collateral to A. A pays $1,000 in interest for borrowing the bond to B. B pays $5,000 in interest for the use of the cash to A. Net effect A delivers $1 million in par amount of the Treasury bond to A. B pays $980,000 + $5,000 – $1,000 = $984,000 in cash to A. Description 2 Today A borrows $1 million par amount of a Treasury bond from B. B borrows $980,000 in cash from A. One month from today A repays the $1 million par Treasury bond loan to B plus $1,000 cash in interest on the loan. B repays the $980,000 in cash to A plus $5,000 in interest on the loan. Net effect A delivers $1 million par amount of the Treasury bond to A. B pays $980,000 + $5,000 – $1,000 = $984,000 in cash to A. Description 3 Today A purchases $1 million par amount of a Treasury bond from B for $980,000 in cash. One month from today B buys the $1 million par amount of the Treasury bond from A at the prearranged price of $984,000. 272 FINANCIAL RISK MANAGEMENT The third description, which is known as a repurchase agreementt, possesses some legal advantages in the event of default. If the party lending the asset defaults, the other party technically owns the asset, since it purchased it rather than just borrowing it, so it has fewer legal restrictions on its ability to use the asset. If the party borrowing the asset defaults, the party lending the asset, since it technically sold the asset and received cash as payment rather than just as collateral for the borrowing, has fewer legal restrictions in its ability to use the cash. 10.1.3 Forwards A forward contractt is an agreement to pay a fixed price on a set forward date for a specified amount of an asset. As such, it combines into a single transaction borrowing the asset and then selling the asset in the spot market. The seller of the forward needs to deliver the asset at a fixed forward date and price, exactly as a borrower of the asset must do. The seller of the forward is at risk for increases in the asset’s price and will gain from decreases in the asset’s price, just like a borrower of the asset who sells it in the spot market. The buyer of the forward is in the same position as a buyer of a spot who lends out the underlying asset but does not need to fund the currency to purchase the asset. Since no cash will change hands until the forward date, it does not have the credit and funding liquidity risks that an uncollateralized borrowing of the asset would have. From a credit risk standpoint, a forward transaction is very similar to a borrowing of the security that has been collateralized by cash. Credit risk on either a forward or a borrowing collateralized by cash starts close to zero, but can build up as the market price of the underlying asset goes up or down. Managing this counterparty credit risk can be quite complex, as the amount of credit exposure varies through time and is correlated with movements in a market price. We will not be fully prepared to address this issue until Chapter 14 on counterparty credit risk. For now, we will just note that a frequently used approach to mitigate this credit exposure is the collateral call, in which the borrower and lender agree in advance that upward moves in the asset price will require the asset borrower to increase the amount of collateral placed with the lender, and downward moves in the asset price will require the asset lender to increase the amount of collateral placed with the borrower. This cross‐collateralization agreement is an automatic feature of futures contracts, which are very closely related to forward transactions. 10.1.4 Futures Contracts Futures contracts are identical to forward contracts in their market price exposures. They also specify the payment of a fixed price on a set forward date Managing Forward Risk 273 for a specified amount of an asset. They differ from forward transactions in two primary dimensions: the management of counterparty credit risk and the degree to which they are tailored to trade off basis risk versus liquidity risk. We briefly discuss both aspects. Unlike forward transactions, which are direct transactions between two firms or individuals, futures contacts are always arranged to have a futures exchange as one of the counterparties to each contract. So if Firm A agrees to sell 100,000 barrels of oil for delivery on June 15, 2015, to Firm B in exchange for $3 million, it is technically broken up into an agreement for A to deliver 100,000 barrels of oil on June 15, 2015, to the futures exchange in exchange for $3 million and an agreement for B to pay $3 million to the futures exchange for the delivery of 100,000 barrels of oil on June 15, 2015. This greatly simplifies credit risk management for the firms, which do not need to worry about the creditworthiness of one another but only need to evaluate the creditworthiness of the futures exchange. This would involve enormous credit management problems for the futures exchange since it has credit exposure to every firm trading on the exchange, but it is managed by strict insistence on continuous cash payments to and from the futures exchange as the prices of the futures transactions rise and fall (details can be found in Section 14.2). This also requires sufficient initial collateral to reduce credit risk to a minimum. This has several significant implications. Because of the continuous collateral calls, a firm using futures contracts will have constant inflows and outflows of cash as asset prices rise and fall. This has important consequences for both funding liquidity risk and market risk. The funding liquidity risk consequence is that if a firm is using futures contracts to offset transactions that do not have this cash settlement feature, it may lead to substantial funding needs. For details, refer back to the Metallgesellschaft (MG) case in Section 4.2.2. The market risk implication is that the constant cash payments create risk to the extent that payment amounts are correlated with the time value of the payments—see the discussion on convexity risk in Section 10.2.4. This system of credit risk control requires that the terms of futures contracts be standardized with only a few possible delivery dates and assets that can be contracted for. This contrasts with forward transactions, which, as agreements between two firms, can be tailored to very specific forward dates and assets to be delivered. This freedom is permitted by the firms performing their own management of the credit considerations of the transaction. However, the futures exchange must have the ability to quickly close out any contract on which a counterparty cannot meet a collateral call. The ability to quickly close out contracts without a substantial risk of loss requires the liquidity derived from a few standardized contract terms. 274 FINANCIAL RISK MANAGEMENT The liquidity that results from this standardization can also be attractive to potential counterparties who may welcome the reduction in liquidity risk this offers. With only a few standardized contracts, it is easier to find good price valuations and close out positions that are no longer desired. The price of lower liquidity risk is, as always, heightened basis risk. A firm might desire to hedge flows on particular dates but need to accept a hedge with nearby standardized dates. It may also desire to sell short a particular asset—a particular grade of wheat, say—but need to accept a hedge with a related standardized grade. The maintenance of necessary liquidity may require the provision that several possible grades be deliverable, which requires formulas for determining how much of each grade must be delivered. However, changes in actual market conditions will differ from any set formula, resulting in profit opportunities and basis risks that may need quite complex modeling. For an example, see the discussion of the modeling of delivery options on Treasury bond futures in Hull (2012, Section 6.2). 10.1.5 Forward Rate Agreements A forward rate agreementt (FRA, pronounced “fra”) is a particular type of forward contract in which the asset to be delivered on the forward date is a borrowing with a specified maturity date, interest rate, and borrower. For example, it might be an agreement to deliver two years from today a $200 million one‐year deposit with Bank of America paying an interest rate of 6.50 percent. This means that in two years the buyer of the FRA will be able to deposit $200 million with Bank of America in exchange for receiving $213 million back at the end of the third year: $200 million × (1 + 6.50%) = $213 million. The standard practice for FRAs is to cash settle, meaning that no actual deposit of cash with Bank of America is expected; instead, a cash amount equal to the value of the deposit will change hands. In our example, if Bank of America is offering 5.00% on one‐year deposits at the end of two years, the right to place a deposit at 6.50% is worth 1.50% × $200 million = $3 million. Therefore, the FRA seller owes $3 million to the FRA buyer, which will be paid at the end of the one‐year deposit period. (In most, but not all, cases, the payment will be made when the FRA settles, which is the beginning of the deposit period, not the end. However, the settlement price will be determined by the present value of the payment due, using the prevailing discount rates at the time of settlement. Economically, this is no different in value from receiving the payment at the end of the deposit period, but it has the beneficial effect of reducing credit exposure.) If Bank of America is offering 7.50% on one‐year deposits at the end of two years, the requirement to place a deposit at 6.50% has a cost of 1.00% × $200 million = $2 million. Managing Forward Risk 275 Therefore, the FRA buyer owes $2 million to the FRA seller, which will be paid at the end of the one‐year deposit period. FRAs are valuable tools for managing forward risk since they can be used to lock in borrowing and lending costs for future time periods or take positions on rates rising or falling. They are almost wholly confined to rates offered on currency borrowings by very high‐credit‐grade banks, since they have developed primarily as tools for managing the cost of borrowing and lending currencies rather than tools for speculating on changes in credit quality. The most popular instruments are those tied to the deposit rates averaged over a panel of high‐grade banks, such as the London Interbank Offered Rate (LIBOR), thereby reducing the link to credit quality even further. Some interest rate futures, such as LIBOR futures, are essentially FRAs contracted using futures rather than forward structuring. 10.1.6 Interest Rate Swaps Standard interest rate swaps are equivalent to a package of FRAs. A very typical example would be a five‐year swap for $200 million settled quarterly with one party paying U.S. dollar LIBOR and the other party paying a fixed rate of 6.50%. This is equivalent to a package of 20 FRAs that are settled on each quarterly date for the next five years. Interest rate swaps are extremely flexible instruments that can be tailored to specific customer needs. Although it is typical that the terms of each FRA that constitutes the package will be the same on all terms except the forward date, it is quite possible for customers to arrange swaps with rates, deposit lengths, and notional amounts that differ by period. It is also quite common to combine FRAs in different currencies into a single swap and combine FX forwards along with FRAs into a single swap. To better understand the customer motivation for these features, it is important to understand the relationship between bonds and interest rate swaps. The primary initial demand for interest rate swaps, and much of the demand to this day, comes from issuers of and investors in corporate bonds. Most corporate bonds pay fixed coupons, as this represents the form popular with most investors. However, bond issuers may prefer to borrow at a floating interest rate rather than a fixed rate, either because they believe rates are likely to fall in the future or because they believe floating‐rate borrowings are a better match to their overall asset‐liability position. Some investors may prefer lending at a floating rate rather than at the fixed coupon on a bond, either because they believe rates are likely to rise or because they are primarily looking to take a position in the creditworthiness of the firm and don’t want to combine this with a position on whether risk‐free rates will rise or fall. For such clients, a fixed‐for‐floating single‐currency interest 276 FINANCIAL RISK MANAGEMENT rate swap, which is just a package of FRAs for a single currency, can transform a fixed‐rate bond position into a floating‐rate one. Another instance of interest rate swap demand arising out of the corporate bond market occurs when the currency a firm would prefer to owe debt in is different than the currency that is preferred by a segment of investors in the firm’s bonds. A typical example would be a European firm that wanted to tap into investor demand in the U.S. market. The firm might prefer to have all its debt in euros, but most U.S. investors prefer to invest in dollar‐denominated bonds. One solution is to have the firm issue a dollar‐ denominated bond, but then enter into a cross‐currency interest rate swap in which the firm receives fixed dollar payments equal to the coupon payments it owes on the dollar‐denominated bond and pays fixed euro cash flows. The firm would probably also want to combine this with an FX forward contract to exchange the euro principal it wants to pay on the maturity date of the bond for the dollar principal that it owes on the dollar‐denominated bond. This combination is a standard product, a cross‐currency interest rate swap with the exchange of fixed principal. The firm might also want to make floating‐rate euro coupon payments rather than fixed‐rate euro coupon payments for the reasons given in the previous discussion on single‐ currency swaps. Rather than execute two separate swaps, this can all be accomplished in a single fixed dollar for floating euro cross‐currency swap. A cross‐currency swap can therefore be a combination of a bundle of FRAs and a bundle of FX forwards. As such, it combines the spot FX risk of FX forwards, the forward risk of FX forwards, and the forward risk of FRAs. 10.1.7 Total Return Swaps We have already seen how a cross‐currency swap can combine FRA and forward positions in a foreign currency asset. Total return swaps are instruments that generalize this approach to enable forward positions to be taken in any asset. Instead of having an agreement to exchange a fixed amount of euros for a fixed dollar price on an agreed forward date, as might be the case in a cross‐currency swap, an agreement might be made to exchange a fixed amount of an asset such as a bond or stock for a fixed dollar price on an agreed forward date. The most common form of total return swap calls for the following. Party A makes a series of intermediate payments to Party B, usually tied to intermediate coupon payments or stock dividends. Party A delivers an asset to B on a fixed date for a fixed price. Finally, Party B makes a series of intermediate payments to A, usually tied to a funding index such as LIBOR. This form of total return swap is economically equivalent to a forward transaction. Like a forward, it combines into a single bundle the spot sale of an Managing Forward Risk 277 asset and the borrowing of that asset for a fixed term. However, although a forward bundles together the sale price and borrowing costs into a single final fixed price, the total return swap makes the intermediate borrowing costs more explicit. One major contractual difference between total return swaps and forwards is that a total return swap can be used by a party that might otherwise find it difficult, for legal or institutional reasons, to invest in a particular asset class. Although the forward contract generally calls for the actual delivery of the asset on the specified forward date, the total return swap will often call for a cash settlement based on the value of the asset on the specified date. This can be a necessity for a party that cannot legally own the asset (for example, the party may not have a subsidiary in the country in which the asset trades). This can still be a great convenience for a party that can legally own the asset but may not be well positioned to trade it. In effect, it is contracting out to the other party the actual sale, which makes sense if the other party is a major market player in this asset or if the asset is actually a participation in a portfolio of assets. The downside of this arrangement is that it can lead to disputes as to what the actual settlement price should be in cases where there is not a publicly available and reliable pricing source. So although it may be easy to agree that the settlement of a basket of stocks traded on the New York Stock Exchange (NYSE) will end up at the published closing exchange prices for the day, it may be necessary to build elaborate legal processes for the settlement of a bond of limited liquidity. (For example, the processes could involve an appeal to a panel of other market makers or the right of the party receiving the value of the bond to take physical delivery in the event of failing to agree on a cash price.) The primary initial impetus for the total return swap market came from parties wanting to purchase assets they would have difficulty holding. They were assets they either could not legally hold or would have difficulty trading, leading to the demand for a cash settlement discussed previously, or assets they would have difficulty funding. A firm wanting to purchase a bond with a higher credit grade than that of the firm could face the negative carry costs of having to fund at a higher credit spread than it can earn on the bond. To avoid this situation, one must find a way to borrow against the collateral of the security, as discussed previously in Sections 10.1.1 and 10.1.2. The total return swap offers the convenience of bundling purchases together with a locked‐in borrowing cost for a fixed period. Collateralization is not required since the asset will not be delivered until the end of the borrowing period. Another example of funding difficulty would be a firm with a sufficiently high credit grade that is under regulatory pressure to reduce the size of its 278 FINANCIAL RISK MANAGEMENT balance sheet. If it can find another high‐credit‐grade firm that is not under similar regulatory pressure, it can “rent the balance sheet” of the other firm by entering into a total return swap, although it must expect to pay for the service. Many of the suppliers of total return swaps to parties wanting to purchase assets they would have difficulty holding simply purchase the asset and hold it until the scheduled delivery or cash settlement. They are being paid to provide a service, as a market maker able to skillfully execute purchases and sales at good prices, an efficient provider of a desired portfolio of assets, a firm having the legal standing to hold assets of a desired country, or a firm with a higher credit standing and lower funding costs or more balance sheet room. However, as the market has evolved, many suppliers are also using this market as an efficient means of borrowing assets in which they want a short spot position. As with a forward, the total return swap provides convenient bundling of the asset borrowing and short sale into a single transaction. For example, a firm wanting to gain on a price decline of a specific bond, either because of a market view or because this offers a hedge against the credit exposure the firm has to the bond issuer, can enter into a total return swap in which it needs to deliver the bond forward (or equivalently cash settle) and then simply does nott hold a cash bond against this forward obligation. 10.1.8 Asset‐Backed Securities In general, an asset‐backed security can be viewed as an alternative instrument to total return swaps in taking on exposures to asset classes where the actual management of the exposure is desired to be left to another party. The reasons why this might be desirable could be copied almost verbatim from Section 10.1.7. The use of asset‐backed securities rather than total return swaps in a particular situation is largely a matter of how documentation and collateralization of the swap agreement are handled. When a particular total return exposure is expected to have a fairly broad appeal to a class of investors, it may be desirable to standardize the terms and offer the exposure through a security rather than a swap. This eliminates the need to individually negotiate swap terms since a single document covers the terms of the security, but the trade‐off is a loss of flexibility in fitting terms to an individual investor. The use of a security structure forces investors to invest cash up front, a convenient solution to collateralization concerns, particularly when the number of investors is potentially too large to make the negotiation of individual credit coverage attractive. Of course, the disadvantage is that investors must tie up cash in the transaction, Managing Forward Risk 279 but in return they get a standardized security that can be sold or pledged as collateral. By contrast, it is hard to transfer ownership of a swap position since your counterparty on the swap, which did not place cash up front, may object to the creditworthiness of the new party to which you want to transfer ownership. The cash nature of the investment protects the party managing the assets from credit concerns. Investors get credit protection by having the assets on which the exposure will be taken placed in some form of trust, thereby immunizing their payoff from default on the part of the party managing the assets. This leads to a potential problem in the flexibility of asset‐backed securities in relation to total return swaps. If assets need to be walled off in a trust, then how can a market maker use this as a vehicle for taking a short sale position in the asset, as we showed can be done with a total return swap? The solution is to have a total return swap as an asset placed with the trust and sufficiently collateralize or protect it by third‐party insurance. Asset‐backed securities lend themselves to pooling positions in a large number of similar assets, such as mortgages, credit card outstandings, loans to businesses, or bonds. The standardized nature of the documentation is well suited to the sharing of exposure by a large number of investors in a large pool of assets, thereby decreasing the event risk of each investor owning a particular block of assets. However, this is more a matter of convenience than necessity, and virtually any financial position that can be achieved through an asset‐backed security can also be achieved through a total return swap. When mortgages and loans are pooled to create an asset‐backed security, it can be structured so as to virtually eliminate credit exposure to the underlying mortgages and loans by investors in the asset‐backed security, or it can be structured to have this credit exposure be part of the risk borne by the investors. Risk management aspects of asset‐backed securities for which credit exposure has been eliminated will be discussed in Section 12.4.6. Risk management aspects of asset‐backed securities that involve credit exposure will be discussed in Section 13.4.3. Table 10.2 summarizes the difference in risk between the different types of instruments through which forward risk can be taken. Spot risk refers to the underlying asset. Forward risk is always present for the underlying asset, but may or may not also involve forward risk in a currency (in the case where the underlying asset is a currency, the question is whether forward risk in a second currency is involved). Credit risk refers only to credit risk to the counterparty on the instrument, not to any credit risk embedded in the underlying asset. The distinction between the lender and borrower refers to their position relative to the underlying asset. 280 Spot Risk No No Yes Yes, except for some interest rate futures No Direct borrowing and lending, including loans, bonds, and deposits Repurchase Forward contract Futures contract FRAs No Yes Yes Yes No Currency Forward Risk Starts at zero, but can build up if not mitigated by collateral. Credit exposure is to the futures exchange and is small due to cash settlement. Starts at zero, but can build up if not mitigated by collateral. Small, only to the extent collateral is inadequate. Yes, but can be mitigated by collateral. Credit Risk for Lender Comparison of Risks in Forward Transactions Instrument TABLE 10.2 Starts at zero, but can build up if not mitigated by collateral. Credit exposure is to the futures exchange and is small due to cash settlement. Starts at zero, but can build up if not mitigated by collateral. Small, only to the extent collateral is in excess of borrowing. Only if collateral needs to be posted. Credit Risk for Borrower May have delivery option. Convexity risk caused by cash settlement. Possible funding liquidity risk if hedging positions that do not have cash settlement. Funding liquidity risk for the lender, unless collateralized. Other Risks 281 No Typically yes Yes Yes Single‐ currency swaps Cross‐ currency swaps Total return swaps Asset‐ backed securities Interest Rate Swaps Currency Forward Risk Spot Risk (Continued) Instrument TABLE 10.2 Yes Yes Yes No Credit Risk for Lender Yes, but can be mitigated by collateral and by trust structure. Starts at zero, but can build up if not mitigated by collateral. Starts at zero, but can build up if not mitigated by collateral. Starts at zero, but can build up if not mitigated by collateral. Credit Risk for Borrower Typically no, fully collateralized. Starts at zero, but can build up if not mitigated by collateral. Starts at zero, but can build up if not mitigated by collateral. Starts at zero, but can build up if not mitigated by collateral. Other Risks 282 10.2 FINANCIAL RISK MANAGEMENT MATHEMATICAL MODELS OF FORWARD RISKS The most important fact about the mathematical models used to manage forward risk is that they rely on one very simple principle: a flow on a given date owed by a particular entity should be regarded as absolutely equivalent to any other flow of the same quantity on the same date owed by the same entity (the term flow is used rather than cash flow, since we want to consider more general cases than cash payments, such as an entity owing an amount of gold or oil). At first glance, this may look like a tautology, a statement true by definition. And it is close to one, which helps to explain why practitioners agree so widely on the models used to manage forward risk. However, it is not quite a tautology—a reminder that mathematical finance deals with market realities, not theoretical abstractions. When the product a trader is dealing with is actually a complicated bundle of flows on a large number of different dates, it is not immediately clear that breaking the valuation apart into many different pieces, few of which can independently be priced in the market, is the best way to proceed. Indeed, a few decades ago, objections to this practice were still being raised along the lines that it would be very expensive in terms of transaction costs for a trader to actually hedge the instrument in this way. By now, everyone involved has come to appreciate that the principle, far from causing traders to try to aggressively rehedge each piece of a deal, is actually a powerful analytical tool that enables very complex transactions to be managed in a way that permits a maximum amount of netting of risks before resorting to aggressive hedging. We need to examine where the complexities in this approach lie in order to see what residual risks still need to be managed. Before turning to the hard part, however, let’s first take a few moments to appreciate some of the benefits entailed by the simplicity of this approach. One benefit is the computational simplicity of the method. The actual bundles of forward transactions that trade in the market can have very complex structures. Our fundamental principle says to ignore all these complexities; just calculate the individual flows that have been bundled together, calculate the present value of each flow in isolation from the others, and then sum the present values. It is not necessary to devise special methods that apply to particular cases, a feature that hobbled many of the methods that were used before the fundamental principle was generally adopted. A second benefit is the generality of the principle. The same computational method can be used for cash flows, commodity flows, bonds, swaps, forwards, futures, risk‐free debt, risky debt, and obligations to deliver stock. (Please note carefully that this is not saying that this method Managing Forward Risk 283 can be used for valuing stocks, since stocks involve unknown future obligations rather than known flows; what is being said is that an obligation to deliver a fixed number of shares of a stock in the future can be translated into an equivalent amount of shares of the stock to be delivered today.) The same computational method can be used to value individual transactions or portfolios of transactions, since each can be reduced to the summation of a set of individual flows, and therefore can also be used for valuing total return swaps or asset‐backed securities tied to portfolios of transactions. A third benefit is that when all transactions are completely reduced to their respective constituent obligations, you are free to describe transactions in whatever manner is most convenient in a given context. When discussing a physical commodity such as gold, it is often convenient just to think in terms of equivalent quantities; for example, you are willing to trade 100 ounces of gold for delivery today for 102 ounces of gold for delivery in one year. When discussing a currency, you might prefer to talk about the interest rate to be paid—say, 6 percent for one year. Although this is just a different way of saying that you are willing to trade $100 for delivery today for $106 for delivery in one year, the interest rate view is often easier to understand using economic theory. When doing computations, it is usually best just to think of discount factors to be multiplied by each flow and then summed to get a net present value equation. When checking the reasonableness of a given set of input parameters to the model, it is often most convenient to think in terms of interest rates that apply to distinct forward time periods—the rate that applies to a particular month, week, or day. Improbable inputs can be more easily spotted if you can see that they imply that a rate of 20 percent on one day will occur in between a 7 percent rate on the immediately preceding and following days. Formulas for translating from discount factors to zero‐coupon interest rates, par‐coupon interest rates, or forward interest rates and back again are readily available, as should be expected from our general principle. You are probably already familiar with these formulas. If not, consult Hull (2012, Chapter 4). The Rates spreadsheet on the website for this book illustrates the techniques for valuing a portfolio of flows based on a given set of forward rates. The forward rates are translated into equivalent zero coupon rates, par rates, and discount factors. Each set of flows, which might correspond to a forward, a swap, a bond, or any of the other instruments discussed, is broken up into its individual flows. For each individual flow, a discount factor is determined based on interpolation from the discount factors derived from the given forward rate. Each individual flow is then multiplied by its discount factor, and these values are summed over the portfolio. 284 FINANCIAL RISK MANAGEMENT The interpolation methodology used in this spreadsheet for illustrative purposes is a simple linear interpolation. In practice, more complex interpolation methodologies are often used. Tuckman and Serrat (2012, Chapter 21) presents a good introduction. The same spreadsheet will be used elsewhere in this chapter to illustrate how to derive a set of forward rates that can match a given set of observed market prices and to demonstrate the calculation of risk statistics for a portfolio of flows. Having postponed looking at complexities, it’s time to face up to the task. Basically, this discussion can be divided into four topics: 1. Section 10.2.1. Models are needed to perform interpolation from flows for which market prices are available to other flows. 2. Section 10.2.2. Models are needed to extrapolate prices for longer‐ dated flows. 3. Section 10.2.3. In some cases, going from flow prices to bundle prices is not as simple as the general approach. This is because some products involve flows representing a promised delivery that is actually a promise to deliver a future flow (for example, a forward purchase of a bond). Untangling these flows involves some complexities. 4. Section 10.2.4. Although the method is designed to handle fixed obligations, it can be applied to a very important class of nonfixed obligations with just a bit of work—flows that will be determined by certain types of indexes. However, this extension must be performed with care; otherwise, a significant source of risk can slip in unidentified. 10.2.1 Pricing Illiquid Flows by Interpolation As was pointed out at the beginning of this chapter, the large number of days on which future flows can occur makes it almost certain that liquid quotations will be available from the market for only a small portion of possible flows. Creating price quotes for all possible flows will require some theory that enables the inference of prices of illiquid flows from prices of liquid flows. We will present two theories: 1. The interpolation of the price of an illiquid flow from prices of liquid flows that are both earlier and later than it. 2. A stack‐and‐roll methodology for pricing flows that have longer maturities than any flows with liquid prices. The mathematics of interpolation is so simple that it can be easy to lose sight of the fact that interpolation is a financial model to the same Managing Forward Risk 285 degree as more complex options models. It shares the same characteristics of being a methodology for predicting future financial events, requiring well‐thought‐out assumptions about the financial markets as grounds for choosing one possible methodology over another, and being a source of potential earnings loss to the extent future events diverge from predictions. When the modeling nature of interpolation is not kept clearly in mind, the choice of interpolation method can be made based on aesthetic criteria, as if it is just a matter of individual taste with no financial consequences. So let us be very specific about financial assumptions and the financial consequences of choices. Consider the following example, which is typical of the circumstances in which interpolation needs to be employed in pricing forward flows. You need to price a forward flow occurring in 6½ years in a market in which liquid prices can be obtained for 6‐ and 7‐year flows, but nothing in between. Let us suppose you choose to price the 6½‐year flow as the average of the prices of the 6‐ and 7‐year flows. If you put on a hedge that consists of 50 percent of the 6‐year flow and 50 percent of the 7‐year flow, you will be perfectly hedged in the short run, since at first changes in the daily mark of the 6½‐year flow will just reflect the average of changes in the daily mark of the 6‐ and 7‐year flows. The same would be true of any other interpolation method chosen (for example, 25 percent of the 6‐year flow and 75 percent of the 7‐year flow) as long as you match the hedge to the chosen interpolation method. The test of the hedge’s effectiveness will come through time. How well will it hold up as flows come closer to maturity, encountering the denser price quotations that exist (in all forward markets) for nearby flows? If, in this example, liquid prices are available for 2‐, 1½‐, and 1‐year flows, then the hedge will prove effective to the extent that the 1½‐year flow is priced at the average of the 1‐ and 2‐year flows at the time five years from today, when it will be possible to unwind the trade and its hedge at these liquid prices. To the extent the interpolated value differs from the actual value at unwind, an unexpected loss, or gain, will result. Note that the unwind values are determined by the relationship between 1‐, 1½‐, and 2‐year flow prices five years from now. The current relationship between 1‐, 1½‐, and 2‐year flow prices cannot be locked into and plays no role other than serving as a historical observation to use in forecasting future relationships. An interpolation methodology needs to be judged by the stability of the valuations it will lead to. Trading desks develop a feel over time for how stable the valuations produced by particular interpolation techniques are in a particular market. Historical simulation can be used as a quantitative check on these judgments. (Exercise 10.1 takes you through a test of some possible interpolation methods judged by their degree of instability around historical 286 FINANCIAL RISK MANAGEMENT price quotes.) The potential valuation errors determined by simulation can be controlled through limits and reserves. The most important lessons to be drawn are: ■ ■ ■ Interpolation, like any other model, represents a judgment about what is most likely to occur in the future. To the extent the judgment is wrong, unanticipated future losses and gains will result. The key event that needs to be projected by an interpolation model is determining the actual relations between prices for flows at a future date when more liquid unwinds are possible. Historical relationships between these liquid flows can be used as inputs to and tests of judgments about future relationships. Limits and reserves can be based on measured historical instability. The preference usually shown for interpolations that produce smooth pricing curves can be explained by two complementary facts: historical relationships between most liquid flows tend to show smooth pricing curves, and economic intuition about future events tends toward long‐term trends without a belief that at some specific future date a sharp change in conditions will occur. However, these are only general trends, not rules. If some specific dates may be believed to have forecastable effects, you should expect to see patterns, such as seasonal patterns, reflected in the interpolations. For a discussion of the impact of seasonal patterns on different forwards markets, see Section 10.3.4. The choice of which variables to interpolate, whether they are discount prices, zero rates, or forward rates, is in one sense arbitrary since we know that each way of representing prices of forward flows is mutually translatable. However, interpolation using one representation may turn out to be more natural than interpolation using a different representation based on the economic motivation supporting the interpolation method chosen. One approach that would follow naturally from our discussion would be to choose an interpolated value that minimizes a selected smoothness measure for forward rates or zero coupon rates. Methods that are utilized on many trading desks, such as cubic splines, have been justified on formal or informal arguments along these lines. Another approach that is fairly widely used is to interpolate the logarithm of the discount factor. Table 10.3 shows how this works, with the resulting zero coupon rates and forward rates. As shown in Table 10.3, the impact of this interpolation method is to use a constant forward rate in all subperiods of the period between two already determined discounts. This method is generally favored by traders with backgrounds in the forwards and futures markets who believe that “all you really know is the quoted forward.” So if you have a forward rate 287 Managing Forward Risk TABLE 10.3 Interpolation Based on the Logarithm of the Discount Factor Time Zero Rate T1 R1 eR T T2 R2 eR T T1 + K(T1 – T1) Discount Factor 1 2 e( Forward Rate 1 R2T2 − R1T1 T2 T1 2 K ) R1T1 KR2T1 [(1 − K)) 1 1] + K(( = 2 KR2T2 − R1T1 1) KR2T2 − KR1T1 K(T2 − T1) = R2T2 − R1T1 T2 − T1 agreement that runs from the end of month 9 to the end of month 12 of 7 percent and no other market observations in this vicinity, this method would assign forward rates of 7 percent to the subperiods from the end of month 9 to the end of month 10, the end of month 10 to the end of month 11, and the end of month 11 to the end of month 12. But what do you do if you have a 7 percent deposit maturing at the end of month 3 and 8 percent FRA from the end of month 3 to the end of month 6, and you are looking to price a FRA from the end of month 3 to the end of month 4? The methodology says use 8 percent, but most practitioners’ economic intuition says the rate should be lower than 8 percent, since it seems as if the market is anticipating rising rates over the period. Most traders make some kind of exception when rates are changing this sharply, but an interpolation methodology tied to a smoothness measure has the advantage of building on this approach in a more general setting. Computationally, it would be convenient if a definitive set of flows was available for which liquid prices could be obtained on the basis of which prices for all other flows could be interpolated. This is rarely true for two reasons: 1. Price liquidity is a matter of degree. Some instruments have prices that are less liquid than others but still show some liquidity. Therefore, these should be given less weight in determining the discount curve, but should not be completely ignored in setting the curve. 2. Prices are often not available for single flows, but are available for bundles of flows—for example, coupon‐paying bonds and fixed‐for‐floating swaps. If enough liquid flow prices are available to interpolate prices for 288 FINANCIAL RISK MANAGEMENT all but the last of the flows in a bundle, then the common technique of bootstrappingg (see Hull 2012, Section 4.5) can be used to first price all the flows except the last and then derive the price of the last flow from these prices and the price of the bundle. However, often not enough prices are available to value all but the last flow. For example, many bond markets have a liquid price for a 7‐ and 10‐year bond, but have no liquid prices in between. To derive a value for the flows occurring in the eighth, ninth, and tenth years, it is necessary to combine interpolation and price fitting into a single step. The Rates spreadsheet on the book’s website provides a sample discount curve‐fitting methodology that is very general in allowing the optimization of a weighted mixture of the accuracy of fitting known liquid prices and determining a forward rate curve that fits closely to an expected smoothness criterion. The optimization method simultaneously determines all the discount rates needed to match all of the market prices of instruments that can potentially be priced off a single discount curve. All these discount rates are taken as input variables in the optimization. The objective function of the optimization is a combination of two measures. The first is a measure of how closely the derived price comes to the market‐quoted price for each instrument, and the second is a measure of how smooth the discount curve is. The measure of closeness of fit of the derived price to the market quote can take several forms. The spreadsheet uses a very simple measure, a summation of the square of the differences between the derived price and the market quote summed over all instruments. Each is multiplied by a selected weight. Higher weights are assigned to more liquid prices, and lower weights are assigned to less liquid prices. This places a greater premium on coming close to the more reliable prices while still giving some influence to prices that have some degree of reliability. Greater complexity can be introduced, such as placing a higher weight on differences that are outside the bid‐ask spread. The most extreme form of this approach would be to introduce constraints that require that the fit be within the bid‐ask spread (this is equivalent to placing an extremely high weight on differences outside the bid‐ask spread). The desirability of putting such a high weight on the bid‐ask spread depends on your opinion of the quotations you are obtaining, how prone they are to error, and whether you really can count on being able to get trades done within the bid‐ask range. The measure of the smoothness of forward rates used in the spreadsheet is also a very simple one: to minimize the squares of second differences of the forward rates. This measures smoothness based on how close the forward rates come to a straight line, since a straight line has second Managing Forward Risk 289 differences equal to zero. (For example, the sequence 7, 7.5, and 8, which forms a straight line, has first differences of 7.5 – 7 = 0.5 and 8 – 7.5 = 0.5, and therefore a second difference of 0.5 – 0.5 = 0. The sequence of 7, 7.25, and 8, which is not linear, has first differences of 0.25 and 0.75 and therefore a nonzero second difference of 0.5.) Practitioners may use more complex measures of smoothness, such as minimizing second derivatives. Different weights can be specified for how important the closeness of price fit is relative to the smoothness of the discount curve. This is just one more appearance of the trade‐off between basis risk and liquidity risk. The lower the weight put on smoothness and the more even the weight put on fitting each of the instruments, the greater the assurance that the discount curve produced matches exactly the observed market prices of all instruments. This minimizes basis risk, but increases liquidity risk. If it turns out that you really cannot close out one of these positions at the price obtained from the market, you could have significant losses, since the price you used for valuation was based only on the assumed market price, even if this differed a great deal from the price that could be obtained by hedging with more liquid instruments. Conversely, the higher the weight put on smoothness and the more weight put on more liquid instruments, the greater the assurance that you are pricing off hedges that are based on liquid, achievable prices. This minimizes liquidity risk but increases basis risk, since you are now pricing off hedges that can be achieved with combinations of nearby instruments in the market. The same guidance we have given for testing the financial impact of interpolation rules carries over to testing the financial impact of procedures for extracting a discount curve from a set of liquid prices. Historical simulation should be used to estimate the stability of valuations that will result from a candidate procedure. Figure 10.1 illustrates the degree to which greater smoothness of forward curves can be achieved with the optimization procedure just discussed than with a simple version of the bootstrapping technique. This simple version is used in Hull (2012, Section 4.5) and is used on a number of trading desks; the Bootstrap spreadsheet gives details of this comparison. To repeat, the degree of importance of the greater smoothness resulting from the optimization is not to be found in aesthetic pleasure, but should be measured quantitatively in financial impact. We have been assuming that all the instrument prices can be completely be determined by discount prices. However, some instruments could have option features, such as callable bonds or futures, that have a nonlinear component to their price. This can be handled by subtracting the option component of the price, leaving a pure nonoption portion that can be priced off the discounts. A complexity is that the option component price may depend in part on the discount curve. An iterative process might be needed. 290 FINANCIAL RISK MANAGEMENT 8.600% 8.400% 8.200% 8.000% 7.800% 7.600% 7.400% 7.200% 7.000% 6.800% 1 2 3 4 5 Forward Rates from Fitting FIGURE 10.1 6 7 8 9 10 Forward Rates from Bootstrapping Comparison of Forward Rates from Bootstrapping and Optimal Fitting Option components based on a first approximation to discounts can be used to get the inputs to the optimization, which yields a discount curve. This is then used to reprice the option components. These can be used as inputs to a second round of optimization. This cycle can be repeated until the discount curves produced stabilize. When developing discount factors, it is important to remember that every obligor will have its own set of factors; a promise to deliver a flow on a given date will be worth something different, depending on how reliable the promise is. There even need to be multiple sets of discount factors for promises of the same obligor since some debts are senior to others and will more likely be paid in case of a default condition. Before this multiplicity of sets of discount factors seems too overwhelming, let’s introduce a note of simplification. It is rare that the type of flow owed plays any role in determining the probability of payment. If you can observe a set of discount factors for a firm relative to the discount factors for the assured delivery of one asset, you can infer the discount factors for that firm for delivery of any other asset. However, this rule has exceptions. If the government of Mexico owes you a debt denominated in its own currency, the peso, you would certainly apply a different discount factor than to its promise to pay a debt denominated in another currency. Mexico has control over the supply of its own currency and can create new currency to meet its Managing Forward Risk 291 payments. It has no such ability in another currency, and although it could create new pesos and exchange them for the currency owed, this might have a severe enough impact on the exchange rate between the currencies to call into question the country’s willingness, and even its ability, to do so. As the procedures for minimizing credit exposure to a counterparty become more complex, involving collateralization, netting, and margin calls, among other techniques, it becomes more difficult to represent the credit exposure in discounting procedures. Any oversimplification should definitely be avoided, such as discounting the flows owed by A to B on a swap at a discount rate appropriate for A’s obligations and the flows owed by B to A on the same swap at a discount rate appropriate to B’s obligations. This treats the gross amounts owed on the swap as if they were independent of one another, completely ignoring a primary motivation for structuring the transaction as a swap—the netting of obligations. As credit exposure mitigation techniques grow in sophistication, they demand a parallel sophistication in valuation technology. This consists of initially treating all flows on a transaction to which credit exposure mitigation has been applied as if they were flows certain to be received. The actual credit exposure must then be calculated separately, taking into account the correlation between the net amount owed and the creditworthiness of the obligor. This methodology is more complex than we can tackle at this point in the book. We will return to this topic in Section 14.3. 10.2.2 Pricing Long‐Dated Illiquid Flows by Stack and Roll An issue that arises frequently for market‐making firms is the need to provide value to customers by extending liquidity beyond the existing market. This need arises not only for bonds and single‐currency swaps and forwards, but also for FX forwards and commodity forwards. A concrete example would be a firm trying to meet customer demand for 40‐year swaps in a market that has liquidity only for swaps out to 30 years. To see the actual profit and loss (P&L) consequences of a methodology for pricing these longer‐term flows, we need to consider a well‐known trading strategy: the stack‐and‐roll hedge. In our example, a stack‐and‐roll hedge would call for putting on a 30‐year swap in the liquid market as a hedge against a 40‐year swap contracted with a customer. Then, at the end of 10 years, the 20‐year swap to which the 30‐year swap has evolved will be offset and a new 30‐year swap in the liquid market will be put on, which will completely offset the original 40‐year swap, which is, at this point, a 30‐year swap. This stack‐and‐roll strategy can be characterized as a quasistatic hedge in that it requires one future rehedge at the end of 10 years. The results 292 FINANCIAL RISK MANAGEMENT of this rehedge cannot currently be known with certainty, either as to the transaction costs (that is, bid‐ask spread) or as to the impact without transaction costs. However, the fact that only a single rehedge is required allows for great simplification in estimating the expected cost of the hedging strategy and its statistical uncertainty. These features recommend using the methodology to quantify the cost and risk of the longer‐term position. To carry out a numerical example, assume that today’s 30‐year yield is 6.25 percent. Since you are planning to roll at the end of 10 years from a 20‐ to a 30‐year yield, to the extent you expect yield curve shifts to be predominantly parallel, you should enter into a duration‐weighted hedge of 1.192 30‐year swaps for every 40‐year swap you are trying to create. The number 1.192 is the ratio of a 30‐year swap duration (13.40) to a 20‐year swap duration (11.24), assuming a 6.25 percent annual par swap rate. To estimate the impact of the roll, you should look at the history of the relationship between 20‐ and 30‐year swap rates. If 30‐year swap rates tend to be 5 basis points higher on average than 20‐year swap rates plus or minus a standard deviation of 7 basis points, and if you want to keep a reserve against a two‐standard‐deviation adverse move, you could mark the 40‐year swap to a rate of 6.25% + 0.05% = 6.30% and set up a 14‐basis‐point reserve. If historical analysis shows that 30‐year swap rates minus 105% of 20‐year swap rates have a lower standard deviation (say, 5 basis points) than an unweighted spread, because 20‐year rates are more volatile than 30‐year rates, you could set up a hedge ratio of 1.192/1.05 = 1.135 and set up a 10‐basis‐point reserve. The actual hedging practice on a trading desk might be to initiate a stack and roll, but it would probably be flexible as to the time at which the roll was actually carried out. The roll could take place at the end of 10 years, but the trading desk might, at that time, decide it was more favorable to defer the roll, since the roll could just as well be carried out at other times using equally liquid instruments—for example, roll at the end of 20 years from a 10‐year swap into a 20‐year swap. The trading desk might also decide, opportunistically, to roll into a less liquid instrument. For example, after two years, the opportunity might arise in which a bid is available for a 38‐year swap that would close out the remaining term of the 40‐year swap. In this case, the desk would also need to look for a 28‐year swap to close out the remaining term of the 30‐year swap it was using as a hedge. Although a trading desk will want to retain flexibility in managing a stack‐and‐roll strategy once it is entered into, modelers and risk managers can best achieve their aims by assuming a fixed‐roll strategy that involves liquid instruments. By considering a strategy that involves liquid instruments, it should be possible to get very reasonable data history that bears on the probable cost of the strategy. If 20‐ and 30‐year swaps have liquid Managing Forward Risk 293 market quotes available, it may be possible to obtain several years’ worth of daily data on the cost of rolling out of a 20‐year swap into a 30‐year swap. This data can be used not just to decide on an expected roll cost, but also to determine a probability distribution of roll costs. The probability distribution can give reasonable estimates of the uncertainty of results, which can serve as an objective basis for establishing limits and reserves. The advantages of this method for risk management are: ■ ■ ■ ■ ■ Appropriate hedge ratios can be based on historical data since different possible hedge ratios can be judged based on the relative degree of historical uncertainty of roll cost. The method makes a clear distinction between the portion of the expected cost of creating a long‐term instrument that can be locked into at current market prices versus the portion that requires projections. In this example, the portion that can be locked into is the current 30‐year rate, and the portion that requires projection is the spread between the 30‐ and 20‐year rates at the time of the roll (in 10 years). This approach gives a solid financial foundation for what is often a loose intuitive argument along the following lines: “The current 20‐ to 30‐year portion of the yield curve is flat to slightly upward sloping, so to price the 40‐year swap at the same yield as the current 30‐year swap is conservative relative to extrapolating the 20‐ to 30‐year upward slope out to 40 years.” This approach makes clear that what matters is not the current 20‐ to 30‐year relationship, but the projected one, which can probably best be estimated based on a longer history of this relationship. Estimates of uncertainty for establishing limits and reserves can be based on readily observable historical market data. Future liquidity costs, such as the potential payment of the bid‐ask spread, are confined to a single point in time. Exercise 10.2 takes you through some sample calculations using the stack‐and‐roll methodology. 10.2.3 Flows Representing Promised Deliveries Let us consider a typical example of a product involving a flow that represents a promised delivery of future flows. A market maker is asked to quote a price for a three‐year U.S. Treasury bond to be delivered in seven years (let’s assume we are working with zero coupon instruments for the sake of simplicity—the principles for coupon‐paying instruments are the same). If the U.S. Treasury were trying to create such a forward, it would be easy. The Treasury would value the forward as a reduction in its need for 10‐year 294 FINANCIAL RISK MANAGEMENT borrowing and an increase in its need for seven‐year borrowings, both of which can be valued off the standard U.S. Treasury discount curve. However, a market maker has a lower credit rating and hence higher borrowing costs than the U.S. Treasury has. If the market maker tries to create the forward by buying a 10‐year instrument, the price it would need to charge for the forward would be burdened by seven years’ worth of the credit spread between the Treasury and the market maker. To avoid this, the market maker needs to find a way to borrow for seven years at essentially a U.S. Treasury rate. Since it has the 10‐year Treasury purchased to put up as collateral against its seven‐year borrowing, this should be feasible. However, it is an institutional fact that a liquid market does not exist for borrowing against Treasury collateral at a fixed rate for seven years. It is certainly possible to borrow against Treasury collateral for short periods with great liquidity, and the market maker should feel no fear about the ability to continuously roll over this borrowing. However, this introduces a large variance in the possible funding costs due to uncertainty about the direction short‐term repurchase rates will take over a seven‐year period. The way around this impasse is for the market maker to buy a 10‐year Treasury, borrow a seven‐year Treasury, and sell the seven‐year Treasury. The 10‐year Treasury is financed for seven years by a series of overnight repurchase agreements (RPs). The borrowing of the seven‐year Treasury is financed by a series of overnight RPs. The market maker has succeeded in achieving the same cost of creating the forward that the Treasury would have, except for any net cost between the overnight RP rates at which the longer Treasury is financed and the overnight RP rate at which the borrowing of the shorter Treasury is financed. In general, these two RP rates should not differ; on any given day, each represents a borrowing rate for the same tenor (overnight) and with the same quality collateral (a U.S. government obligation). However, the RP market is influenced by supply‐and‐demand factors involving the collateral preferences of the investors. Some of these investors are just looking for an overnight investment without credit risk, so they don’t care which U.S. Treasury security they purchase as part of the RP. Other investors, however, are looking to receive a particular U.S. Treasury that they will then sell short—either as part of a strategy to create a particular forward Treasury or because they think this particular Treasury issue is overpriced and they want to take advantage of an anticipated downward price correction. The higher the demand by cash investors to borrow a particular security, the lower the interest rate they will be forced to accept on their cash. When RP rates on a particular Treasury issue decline due to the demand to borrow the issue, the RP for the issue is said to have gone on special. 295 Managing Forward Risk So the market maker in our example will not know in advance what the relative RP rates will be on the shorter security on which it is receiving the RP rate and the longer security on which it is paying the RP rate. To properly value the Treasury forward created by a market maker, it is necessary to make a projection based on past experience with RP rates for similar securities. This source of uncertainty calls for risk controls, which could be a combination of limits to the amount of exposure to the spread between the RP rates and reserves on forward Treasuries, with reserve levels tied to the uncertainty of RP spreads. Constant‐maturity Treasury (CMT) swaps (see Hull 2012, Section 32.4) are popular products with rate resets based on U.S. Treasury yields. They are therefore valued in the same way as U.S. Treasury forwards. Control of the risk for this product focuses on creating long and short cash Treasury positions and managing the risk of the resulting RP spreads. 10.2.4 Indexed Flows We will now examine how to extend our methods for handling fixed flows to handling nonfixed flows tied to certain types of indexes. Let’s start with a simple example. To keep this clear, let’s label all the times in our example with specific dates. Let’s say the current date is July 1, 2013. Bank XYZ is due to pay a single flow on July 1, 2015, with the amount of the flow to be determined on July 1, 2014, by the following formula: $100 million multiplied by the interest rate that Bank XYZ is offering on July 1, 2014, for $100 million deposits maturing on July 1, 2015. Since this interest rate will not be known for one year, we do not currently know the size of this flow. However, we can determine a completely equivalent set of fixed flows by the following argument and then value the fixed flows by the methodology already discussed. We write our single flow as the sum of two sets of flows as follows: July 1, 2014 July 1, 2015 Set 1 –$100 million +$100 million × (1 + Index rate) Set 2 +$100 $ million –$100 $ million Contracted flow 0 +$100 million × Index rate We will argue that the flows in set 1 should be valued at zero. If this is true, then the present value of our contracted flow must be equal to the present value of the second set of flows, which is a set of completely fixed flows. 296 FINANCIAL RISK MANAGEMENT It can be argued that the present value of the flows in set 1 should be zero because the very meaning of the interest rate that Bank XYZ will be offering on July 1, 2014, for $100 million deposits maturing on July 1, 2015, is the rate at which customers of XYZ are willing on July 1, 2014, to pay XYZ $100 million in order to receive a cash flow of $100 million × (1 + Rate) on July 1, 2015. So why would we currently value the right to enter into a transaction that will, by definition, be available on that date at anything other than zero? A second argument can be given for why the present value of the flows in set 1 should be zero. Mathematically, it is equivalent to the argument already given, but it differs in institutional detail and can deepen intuition, so I will provide it. Let’s say we are considering offering a FRA with the following flows: July 1, 2015 Set 3 +$100 million × Index rate –$100 million × Fixed rate At what fixed rate would you be willing to enter into this FRA at an up‐ front cost of zero (which is equivalent to saying it has a discounted present value of zero)? You should be willing to do this only if the fixed rate is one that you can lock into today at zero cost. The only such rate is the one that makes the following set of flows have a discounted present value of zero; see Hull (2012, Section 4.7) for a detailed example. Set 4 July 1, 2014 July 1, 2015 –$100 million +$100 million × (1 + Fixed rate) Since both sets 3 and 4 have discounted present values of zero, their sum must also have a discounted present value of zero. The fixed rate is the same in sets 3 and 4 by construction, so the sum is just: July 1, 2014 July 1, 2015 –$100 million +$100 million × (1 + Index rate) This is the set of flows we wanted to prove has a discounted present value of zero. One major caveat exists for this approach: it works only when the timing of the index payment corresponds exactly to the index tenor. If, in our example, the payment based on the one‐year index, set on July 1, 2014, Managing Forward Risk 297 had taken place on July 1, 2014, rather than July 1, 2015, the argument would not have worked in eliminating the index rate from the cash flows and we would have ended up with an additional term consisting of the receipt of $100 million × the index rate on July 1, 2014, and the payment of $100 million × the index rate on July 1, 2015. The value of this early receipt of payment depends on what the level of the one‐year interest rate will be on July 1, 2014, and the size of the early receipt also depends on what the level of the one‐year interest rate will be on July 1, 2014. This nonlinearity gives rise to convexity, which is very similar to an options position in that no static hedge is possible (a dynamic hedge is required), and the value of the position rises with higher rate volatility. Other situations also lead to convexity: ■ ■ Positions that have up‐front cash settlement without discounting, such as futures. The value of receiving gains up front is dependent on future rate levels. If changes in the value of the future correlate with changes in rate levels, as they certainly will for an interest rate future, the value will be a nonlinear function of rate levels. Positions where a payment is linearly based on the future rate, rather than the future price, of a bond or swap. The value of payments based on the future price can be determined by discounted cash flows. However, the future rate is a nonlinear function of the future price. Hull (2012, Section 6.3 and Chapter 29) discusses the issue of convexity adjustments in the valuation of forward risk. Although complete models of convexity adjustments require term structure interest rate options models, Hull offers some reasonable approximation formulas for convexity adjustment in these sections. We will examine a more precise technique for convexity adjustments in Section 12.1.3. Now that we have found the set of fixed cash flows that are equivalent to an indexed flow, it is important to remember that these fixed flows need to be identified with the same obligor as the indexed flows. Indexed flows are almost exclusively determined for a panel of highly creditworthy banks. For example, LIBOR is determined by a set formula from offering rates of a panel of banks determined by the British Bankers’ Association. Panels of firms are used because they minimize the danger of firms manipulating the index. If many contractual rates were tied to the rate at which a certain individual bank was offering to pay for deposits, the bank could set its rate a bit higher if it knew this would impact the amount it owed on a large number of contracts. By using a panel of banks and having rules that throw out high and low offers and average those in between, the impact of any one bank on the index rate is lessened. 298 FINANCIAL RISK MANAGEMENT So the index flows need to be translated into fixed flows representing the average credit discount of a panel of banks. This can lead to risk in four different directions, all of which need to be properly accounted for: 1. Different panels are used for different currencies within the same location. There are more Japanese banks in the panel that determines yen LIBOR than in the panel that determines dollar LIBOR; therefore, if Japanese banks are perceived to decline in creditworthiness, it will lead to a higher credit spread applied to the fixed flows that yen LIBOR is equivalent to than for the fixed flows that dollar LIBOR is equivalent to. 2. Different panels are used for the same currency within different locations. There are more Japanese banks in the panel that determines the yen Tokyo Interbank Offered Rate (TIBOR) than in the panel that determines the yen LIBOR. Fluctuations in the perceived creditworthiness of Japanese banks lead to fluctuations in yen LIBOR‐TIBOR spreads. Firms that have taken the shortcut of valuing all yen index flows the same have suffered significant losses from overlooking this exposure. 3. The panel of banks determining an index has a different credit rating than that of an individual obligor. It is important to discount indexed flows at a different set of discount factors than fixed flows of a specific obligor and make sure that exposure to changes in the relationship between these discount factors is kept under control. During the global banking crisis of 2007–2008, this became particularly important, as credit concerns caused wide gaps to appear between the funding costs of individual banks; see Tuckman and Serrat (2012, Chapter 13) for an analysis of how modeling of interest rate products needed to adjust to these events. 4. There can be differences in the pricing of index flows for different frequencies. For example, if there is an expectation that six‐month LIBOR will average 5 basis points more than three‐month LIBOR over a five‐ year period, you would expect a five‐year swap against six‐month LIBOR to have a 5‐basis‐point higher fixed rate than a five‐year swap against three‐month LIBOR quoted at the same time. There is a swap product, basis swaps, that trades LIBOR at one frequency against LIBOR at another frequency. Usually, basis swap pricing shows a slightly higher expectation for LIBOR that is reset less often (e.g., six‐month LIBOR would be greater than three‐month LIBOR). This is both because banks in raising funds prefer to lock in rates for a longer time period, guarding against temporary periods of illiquidity, and because swap investors receiving LIBOR have a slight preference for more frequent resets; it gives them an advantage if a bank in the LIBOR panel has to drop out due to a deteriorated credit outlook and is replaced Managing Forward Risk 299 on the panel by a bank with lower funding costs. This basis difference is normally quite small, but rose during the banking liquidity crisis of 2007–2008. For details, see Tuckman and Serrat (2012, 449–450). 10.3 FACTORS IMPACTING BORROWING COSTS When designing stress tests and setting limits for forward risk for a given product, risk managers must understand the economics of the borrowing costs for that product in order to gauge the severity of stresses the borrowing cost can be subject to. Four key characteristics, which differ from product to product, should be distinguished: 1. Section 10.3.1. How large and diversified is the borrowing demand for the product? 2. Section 10.3.2. To what extent does cash‐and‐carry arbitrage place a lower limit on borrowing costs? 3. Section 10.3.3. How variable are the storage costs that impact the cash‐ and‐carry arbitrage? 4. Section 10.3.4. To what extent are borrowing costs seasonal? We also discuss the relationship between borrowing costs and forward prices in Section 10.3.5. 10.3.1 The Nature of Borrowing Demand A source of borrowing demand that exists for all products comes from traders wanting to borrow in conjunction with short selling. For some products—such as stocks, bonds, and gold—this is the only significant source of borrowing demand. At the other extreme are currencies, where there is strong credit demand by businesses and households to finance purchases and investments. Intermediate cases include most physical commodities, such as oil or wheat, where borrowing demand exists to meet immediate consumption needs. Products for which borrowing demand comes almost exclusively from short sellers tend to have very low borrowing rates most of the time, since there is little competition for the borrowing. This may not be immediately obvious in market quotes if the quotes are made as forward prices rather than borrowing rates. For example, a one‐year gold forward might be quoted at 314.85, a 4.85 percent premium to a $300 spot price. However, when this is broken apart into a borrowing cost for cash and a borrowing cost for gold, it almost always consists of a relatively high borrowing cost for cash, 300 FINANCIAL RISK MANAGEMENT say 6 percent, and a relatively low borrowing cost for gold, say 1 percent. As a result, $300 today is worth $300 × 1.06 = $318 received in one year, and 1 ounce of gold received today is worth 1 × 1.01 = 1.01 ounces of gold received in one year, giving a forward price of gold of $318/1.01 ounces = $314.85 per ounce. Borrowing rates rise as short‐selling activity increases. The major risk for short sellers in these products is the short squeeze in which borrowing costs are driven sharply upward by a deliberate policy of a government or of holders of the assets seeking to support prices by restricting the supply of available borrowing. The resulting increase in borrowing rates pressures short sellers to abandon their strategy and close out their positions. Short squeezes are possible for any asset class, but are more difficult to achieve for assets where borrowing demand has a broader base. A government wanting to support the price of its currency may be tempted to tighten the money supply in order to place borrowing cost pressures on the short sellers of the currency, but it will be limited by the fact that these increased borrowing costs will also hurt business firms and consumers who borrow. Even so, a government faced with a run on its currency will still decide on occasion that the desire to pressure short sellers outweighs other considerations, and will either take steps to sharply raise rates or put in place legal measures that discriminate against certain classes of borrowers who are believed to be selling the currency short. An example of the former is the Irish central bank driving short‐term rates to 4,000 percent in 1992 in an attempt to teach speculators a lesson (Taleb 1997, 212). An example of the latter is Malaysia in 1997 closing its currency borrowing markets to foreign investors. The possibility of a short squeeze on borrowing rates acts as a brake on those who want to take a position on an asset declining in value, since they are faced with a risk to which those taking a position on the asset price increasing are not subject. Those wanting to position for price increases in a particular asset have the freedom to borrow any other asset (most probably, but not necessarily, a currency) relative to which they believe it will increase in price and exchange one for the other in the spot market. However, those wanting to position for a price decrease in a particular asset must borrow that particular asset. The consequence of this asymmetry for rate scenarios is that the possible short squeeze means the risk of very high borrowing rates needs to be guarded against. 10.3.2 The Possibility of Cash‐and‐Carry Arbitrage When available, the possibility of a cash‐and‐carry arbitrage position acts as a lower limit on borrowing costs. A cash‐and‐carry arbitrage is one in which an asset is either purchased or borrowed at one date and repaid or sold at a Managing Forward Risk 301 later date, being held or stored in between the two dates. The example most often cited is a limit on currency‐borrowing rates not to go below zero, since if they did, a trader could borrow the currency at a negative interest rate, hold the currency, and then use the currency held to pay back the borrowing, collecting the negative interest payment as guaranteed profit. A more general result says that a lower limit on negative borrowing costs is the storage cost of the asset. This generalization makes it clear that the specific result for a currency rests on the assumption that currency storage costs are zero (by contrast, a physical commodity such as gold has handling and insurance costs of storage that can lead to negative borrowing costs). Although currency storage costs are almost always zero for large amounts (retail depositors may be charged transactions fees), there have been a few historical exceptions. For example, governments wanting to slow the pace at which foreign deposits are driving up the value of their currency have imposed transaction fees or taxes on large deposits, permitting negative borrowing costs. Cash‐and‐carry arbitrage is not feasible for all asset classes. Perishable physical commodities, such as live steers or electricity, cannot be stored, so cash‐and‐carry arbitrage does not place a lower limit on borrowing costs in such markets. Although arbitrage is not available as a limit, some pressure on borrowing costs getting too low will still result from economic incentives for consumers to change the patterns of demand. So if current prices get too high relative to those six months forward, beef consumption will be postponed to the point that the spot price will start to decline relative to the forward price, thereby raising the borrowing rate. 10.3.3 The Variability of Storage Costs Storage costs on physical commodities tend to be reasonably stable, since they are the cost of physical processes such as handling and transportation. Coupon payments on bonds, a storage benefit, are also stable. However, the storage benefit on stocks, the receipt of dividends, can be quite unstable. A financial setback could lead to a sudden drop in dividends. A merger could lead to a sudden increase in dividends, as in the example at the beginning of this chapter. Changes in tax laws applied to dividends have also resulted in substantial sudden changes in stock‐borrowing costs. Note that it does not matter whether the contractual borrowing terms call for the stock borrower to receive the dividends or pass them through to the stock lender. If the borrower receives the dividend, then an increase in dividend will cause the lender to demand a higher borrowing rate. When the borrower must pass the dividend through to the lender, then a borrower who has sold the stock short (which is the only economic rationale for borrowing stock) must pay the increased dividend out of the borrower’s own pocket. 302 FINANCIAL RISK MANAGEMENT 10.3.4 The Seasonality of Borrowing Costs Interpolation methodology for discount factors and the evaluation of the risk of incorrect interpolation must take into account the seasonality of borrowing costs, which can lead to patterns that would be missed by simple interpolation from adjoining prices. To illustrate this with an extreme example, suppose a stock pays a dividend on exactly every July 15. The value of the dividend to be received on July 15, 2014, will be reflected in the borrowing cost to July 15, 2014, but not in the borrowing cost to July 14, 2014. Without knowing this, no conventional methodology for interpolating between borrowing costs to January 1, 2014, and January 1, 2015, will pick up the sharp difference between the borrowing costs to these two dates. Most borrowing markets do not hinge on such specific scheduling. However, markets for physical commodities such as oil and other energy products and agricultural products often reflect seasonal supply and demand factors such as a stronger demand for heating oil as winter approaches and stronger supply of wheat immediately following harvesting months. The seasonality of borrowing costs for physical commodities is closely tied to the possibility of cash‐and‐carry arbitrage. Commodities capable of storage that permit cash‐and‐carry arbitrage will have a smaller seasonal component since the storage of supply can be used as a response to seasonal demand. Perishable commodities that do not permit cash‐and‐carry arbitrage show a stronger seasonal component, since pricing differentials need to become large enough to start shifting demand. In the extreme case of electricity, which cannot be stored for even very short periods of time, seasonality effects can be seen within a single day, with different forward prices for different times of the day based on differing demand by the time of day. Borrowing rates for gold, stocks, bonds, and currencies generally show far less of a seasonal effect than borrowing rates for physical commodities, both because of the possibility of storage and because the seasonality of supply and demand is weaker than that for physical commodities. However, some seasonal effects can be observed—most prominently turn‐of‐the‐quarter effects in currency borrowing. This effect is a sharp spike in demand for borrowing currency on the last business day of each quarter and particularly the last business day of each year. A more detailed discussion can be found in Burghardt and Kirshner (1994). A particularly pronounced seasonal borrowing effect for currencies was experienced throughout 1999 as fears of computer operational problems starting on January 1, 2000—the Y2K problem—caused a large demand for liquidity over the first few weeks of January 2000. Since firms wanted to lock in the currency availability for this period, they were willing to pay much higher borrowing rates for this period than for any period preceding it or succeeding it. 303 Managing Forward Risk 10.3.5 Borrowing Costs and Forward Prices As emphasized in Section 10.2, every statement made about borrowing costs can be translated into an equivalent statement about forward prices, and vice versa. In market convention, statements about currencies are usually made in terms of borrowing costs, and statements about physical commodities are usually made in terms of forward prices. Since currencies generally have more widespread borrowing demands than physical commodities, as discussed in Section 10.3.1, the borrowing costs for physical commodities will usually be lower than the borrowing costs for a currency. This is usually expressed in forward price terms by saying that the forward price of a physical commodity is generally higher than its spot price—a condition known as contango. However, when a strong demand exists for the availability of a particular physical commodity, its borrowing cost may be driven above the borrowing cost of a currency, resulting in forward prices being lower than spot prices—a condition known as backwardation. An example of this relationship is shown in Table 10.4. (This terminology has considerable history behind it. In the 1893 Gilbert and Sullivan operetta Utopia, Limited, d a character is introduced as a financial wizard with the phrase “A Company Promoter this, with special education, Which teaches what Contango means and also Backwardation.”) A similar situation arises when the borrowing costs are quoted on a net basis in a situation where collateral is being lent to reduce the credit risk of the borrowing. For example, if a security is being borrowed and cash is being lent as collateral, there may be no explicit quote on the borrowing cost of the security. Instead, a net rate is quoted as an interest rate on the cash. If a short squeeze develops on the security, making it expensive to borrow, this will manifest itself as a low (possibly negative) interest rate to be paid TABLE 10.4 Examples of Contango and Backwardation Contango Example Currency Commodity Price Spot 1 month 2 months 3 months $100.00 $100.50 $101.00 $101.50 1 unit 1 unit 1 unit 1 unit $100.00/unit $100.50/unit $101.00/unit $101.50/unit Backwardation Example Currency Commodity Price Spot 1 month 2 months 3 months $100.00 $100.50 $101.00 $101.50 1 unit 1.05 units 1.10 units 1.15 units $100.00/unit $95.71/unit $91.82/unit $88.26/unit 304 FINANCIAL RISK MANAGEMENT for the loan of the cash. An identical trade, from an economic viewpoint, is a repurchase agreement. An expensive‐to‐borrow security will manifest itself through a low to negative rate being paid on the cash side of the transaction. 10.4 RISK MANAGEMENT REPORTING AND LIMITS FOR FORWARD RISK Risk management reports for forward risk must be more detailed than those for spot risk. Not only do the reports involve an extra dimension of time, but they also involve a dimension of credit quality, since the same flow owed to you on the same day has different risks depending on who owes it to you. We’ll examine the time dimension first and then the credit quality dimension. The basic principle of breaking all forward instrument exposures apart into individual flows has already done a lot of the necessary work for risk management reporting. A complete risk report would just show the amount of net flow exposure for each forward date. The remaining question is what types of date groupings make sense in giving a trading desk and then senior managers a more concise picture of this exposure. One issue that can lead to some confusion when designing and using risk management reports for forward risk is the overlap in the usage of many close‐to‐equivalent measures. This starts with disagreement over the simple convention of what is meant by a long position and a short position. In spot markets, longg clearly means to own an asset, benefit by a rise in the asset price, and lose from a decline in the asset price, while shortt means exactly the opposite in each respect. In forward markets, some practitioners who think about owning a bond use longg and shortt in the same way—the long position benefits from bond prices rising and therefore from interest rates falling, and the short position benefits from bond prices falling and therefore from interest rates rising. Other practitioners with backgrounds in instruments such as swaps and FRAs, where no natural concept of an asset being owned is available, use long to mean a position benefiting by interest rates rising and shortt to mean a position benefiting by interest rates falling. Often, all you can do is remind yourself which trading desk you’re talking to in order to know which way the term is being used, but insist that everyone must agree to use a firmwide convention, no matter how much they hate it, when talking to the chairman of the board. A similar set of differences in convention is at work when describing the size of a position. Some traders have grown up using the term value of a basis pointt (or equivalently value of an 01), whereas others refer to a 5‐year equivalent, t 10‐year equivalent, t or duration. Tuckman and Serrat (2012, Chapters 4 and 5) is highly recommended for a detailed and intuitive Managing Forward Risk 305 explanation of these concepts. Table 10.5 illustrates this with a numerical example in which we’ll consider a position with just two components: a 5‐year flow and a 10‐year flow. As shown in Table 10.5, the different position size measures differ only by a constant factor. The five‐year equivalent of a position is just the value of a basis point of that position divided by the value of a basis point of a five‐ year instrument. Any other instrument could be used as a similar common denominator (also known as a numeraire). Table 10.5 also shows that the weighted duration is essentially just the value of a basis point divided by minus 1 basis point (–0.01 percent). However, note that duration needs to be weighted by the price value of the position, whereas all the other measures are weighted by the par value of the position, reflecting the definition of duration as the price change per dollar of portfolio value. See Tuckman and Serrat (2012, 130); see also Tuckman and Serrat (2012, 145–147) for a proof that the duration of a cash flow is simply equal to its tenor. You can check that if the position held was +100 of the five‐year flow and –74.536562 of the 10‐year flow, using the ratio between values of a basis point, the five‐year equivalent, 10‐year equivalent, and duration measures for the portfolio would all come out equal to 0. However, the impact of a 100‐basis‐point increase would not be 0; it would be –3.613013 + (0.74536562 × 4.725634) = –0.090688. So a position that is completely hedged for a 1‐basis‐point rate move is not completely hedged for a 100‐basis‐point move. This nonlinearity stems from the fact that the formula for converting interest rates to prices is not a linear formula. Risk exposure to the size of a move in input variables is known as convexity risk. This is a risk that does not exist for spot exposures, which are linear, and is a major issue for options exposures; it will be a principal topic of the next chapter. The convexity of forwards is much less severe than for options, and it is rare for risk managers to focus much attention on it. In addition to not being a very large effect, it is directly tied to hedging longer positions with shorter positions (since the nonlinear effects grow with time to maturity), and risk reporting will already be directed at the degree of maturity mismatch. Convexity is an important issue for one type of forward risk—credit exposure. Because a credit event, such as the downgrade of a credit rating or, at the extreme, a default event, can cause credit spreads to jump by hundreds or even thousands of basis points, the degree of hedge exposure can be enormous. Reconsider our previous example with the hedge ratio of 100:74.536562, making the position neutral to a 1‐basis‐point change in the credit spread. In the event of default, there will no longer be any difference between a 5‐ and 10‐year flow—both will just represent claims in a bankruptcy proceeding. If a 30 percent recovery occurs on these claims, the hedged position will show a loss of 70% × (–100 + 74.536562) = –17.8244066. 306 TABLE 10.5 FINANCIAL RISK MANAGEMENT Sample Computation of Forward Risk Positions Amount Current zero coupon rate Current discount factor Current value of positions Discount factor given 1‐basis‐point increase in zero coupon rates 5‐Year Flow 10‐Year Flow +100 –100 6.00% 7.00% 1/exp(0.06 × 5) = 0.74081822 1/exp(0.07 × 10) = 0.49658530 74.081822 –49.658530 1/exp(0.0601 × 5) = 0.74044790 1/exp(0.0701 × 10) = 0.49608897 Portfolio 24.423292 Value of portfolio given 1‐basis‐point increase in zero coupon rates 74.044790 –49.608897 Impact of 1‐basis‐point increase in zero coupon rates 74.044790 – 74.081822 = –0.037032 −49.608897 − (−49.658530) = 0.049633 0.01260 5‐year equivalents +100 –100 × (0.049633/ 0.037032) = 134.0305 –34.0305 10‐year equivalents +100 × (0.037032/ 0.049633) = 74.536562 –100 –25.390117 5 years 10 years +74.081822 × 5 = 370.40911 −49.658530 × 10 = −496.58530 Discount factor given 100‐basis‐point increase in zero coupon rates 1/exp(0.07 × 5) = 0.70468809 1/exp(0.08 × 10) = 0.44932896 Value of portfolio given 100‐basis‐point increase in zero coupon rates 70.468809 –44.932896 25.535913 70.468809 – 74.081822 = –3.613013 −44.932846 − (−49.658530) = 4.725634 1.112621 Duration Weighted duration Impact of 100‐basis‐ point increase in zero coupon rates 24.435943 –126.17619 Managing Forward Risk 307 This is a risk that investors need to be aware of. It explains why investors in bonds issued by firms with high default risk (known as high‐yield debtt, or, less politely, as junk bonds) tend to deal directly with prices and avoid reference to interest rates. For a further discussion of the impact of convexity on credit exposure, see Section 13.1.2.2. Firm‐level risk management for forward risk requires decisions about the degree of detail with which exposure to changes in yield curve shape will be represented. Senior management almost certainly needs to be informed of only a few parameters that represent the rate exposure. Many studies have been performed on the historical changes in the shape of many different rate curves, and almost all have shown that about 80 to 90 percent of all changes can be explained by just two parameters, and close to 95 percent of all changes can be explained by just three parameters. Although statistical methods can be employed to determine the best two or three principal components, it makes for better intuitive understanding if parameters can be chosen that convey a concrete meaning. Fortunately, almost all studies of yield curve movement show that intuitively meaningful parameters perform almost as well as parameters selected by statistical means (see, for example, Litterman and Scheinkman 1988). The three parameters that explain most of the change, in order of importance, are: 1. A parallel shift parameter. 2. A parameter to measure the degree of linear tilt of the yield curve. 3. A parameter to measure yield curve twist, the degree to which the middle of the curve changes relative to the two ends of the curve. The Rates spreadsheet illustrates the calculation of the impact of parameter shifts on a portfolio. The parallel shift parameter certainly represents a nondiversifiable risk in the sense of Section 6.1.1, and a case could be made for considering the linear tilt parameter having an element of nondiversifiable risk as well. It is therefore particularly important that these exposures be highlighted to management. Nonstatistical limits on yield curve shape exposure also often start with such overall parameters, but it is usually found to be necessary to have more refined limit measures as well. The debate is often between bucket measures based on groupings of forward risks (for example, zero‐ to one‐year forwards, one‐ to two‐year forwards, two‐ to three‐year forwards, and so on) versus bucket measures that break the yield curve exposure down to exposure to yield changes in the most liquid hedging instruments (such as futures contracts out to five years and then 7‐, 10‐, and 30‐year swaps). The primary argument in favor of the latter 308 FINANCIAL RISK MANAGEMENT approach is that these are the actual hedging instruments most likely to be used; therefore, limits expressed in these terms are immediately operational (a trader knows what action needs to be taken to close a position) and can more easily be judged as to the viability of limit size relative to customer order flow and market liquidity for that instrument. The primary argument against this approach is that the translation of cash flow exposures into liquid hedging instrument equivalents is not completely determined, and very small changes in the choice of algorithm can lead to large changes in how a position is distributed between different instruments. For further discussion of this choice, see Tuckman and Serrat (2012, 158–159). The decision of which currencies, commodities, and equities should be grouped together rests on very similar considerations for yield curves as for spot risk (refer to the discussion in Chapter 9). Within a grouping, limits are needed by obligor. You would, at a minimum, want to have limits on the government’s curve and the interbank rate curve (also known as the swap curve or LIBOR curve), but would probably want to group together rate curves for other obligors, probably by credit rating and possibly by industry and country. EXERCISES 10.1 Interpolation For this exercise, make use of the RateData spreadsheet. Suppose you are making a market in 16‐, 17‐, 18‐, and 19‐year swaps. Liquid swaps are available at 15 and 20 years. Try out some different interpolation methods and test their effectiveness when using them to derive unwind values. Here are some suggestions: ■ There isn’t enough data on the spreadsheet to see what the impact of initiating a hedge at one point and unwinding in 10 years would be, so let’s make the reasonable assumption that the long‐term distribution of rate curve shapes is reasonably stable. So, for example, we’ll judge the effectiveness of interpolating the 18‐year rate from 40% × the 15‐year rate + 60% × the 20‐year rate by looking at the long‐term distribution of unwind costs of Managing Forward Risk ■ ■ ■ ■ ■ 309 an eight‐year rate relative to 40% × the five‐year rate + 60% × the 10‐year rate. Standard deviation can be used as a reasonable summary statistic for the uncertainty of unwind cost, although you should feel free to explore other possible measures such as the 99th percentile. To keep the math easier, ignore any compounding effects; that is, treat the par swap rates as if they were zero coupon rates. So the gain from buying an eight‐year swap at 6% and selling a five‐year swap at 5.70% + 60% of a 10‐year swap at 6.10 percent is just the following: (40% × 5.70% + 60% × 6.10%) – 6% = –0.06%. You can look at the impact of interpolating with different percentages than those suggested by maturity; for example, consider a 50% five‐year, 50% 10‐year interpolation for an eight‐year swap as well as the standard 40% five‐year, 60% 10‐year interpolation. You can consider the impact of factoring the 30‐year rate into the interpolation; this will lead to the use of a 20‐year rate in the unwind. Explore how much improvement in reducing hedge uncertainty comes about by interpolation rather than just assuming a flat curve by looking at the degree to which uncertainty is reduced by using both the five‐ and 10‐year rates in the unwind rather than just the five‐year rate (or just the 10‐year rate). 10.2 Stack and Roll Use the sample stack‐and‐roll computation in Section 10.2.2 and the rate data history from the RateData spreadsheet to calculate two standard deviation reserves for the following products: ■ 40‐year swap 35‐year swap 33‐year swap ■ 50‐year swap ■ ■ As in Exercise 10.1, assume that the par swap rates are actually zero coupon rates to keep the math simpler. 310 FINANCIAL RISK MANAGEMENT 10.3 Rates Use the Rates spreadsheet to calculate risk exposure for a portfolio of forward instruments: 1. Begin by creating a discount curve that can be used in subsequent calculations. Enter a set of benchmark instruments and market prices into the Instruments worksheet and solve for a discount curve that fits these prices, following the spreadsheet instructions. You might, for example, select a set of U.S. Treasury bonds with one‐, two‐, three‐, four‐, five‐, seven‐, and 10‐year maturities. A reasonable set of parameters is to put an equal weighting of 1 on each of your benchmark instruments and to place a weight of 90 percent on fitting prices and 10 percent on the smoothness of the resulting forward curve, but you are encouraged to try different parameters and see their impact on the resulting discount curve. 2. After creating the discount curve, select a portfolio of instruments for which to calculate risk exposure by placing weights on each instrument (you can also add other instruments beyond the benchmark instruments). Look at the resulting risk exposure by forward bucket and summary exposure to forward shifts, tilt shifts, and butterfly shifts, and try to make intuitive sense of them. 3. By trial and error (or by creating an optimization routine with the Solver), find modifications to your portfolio weights that make parallel shift exposure close to zero, but retain roughly the same tilt exposure and butterfly shift exposure as your original portfolio. 4. Follow the same instructions as for part 3, but make tilt exposure close to zero and leave parallel shift exposure and butterfly shift exposure roughly the same as in your original portfolio. CHAPTER 11 Managing Vanilla Options Risk very book should have a hero. The hero of this book is not a person but an equation: the Black‐Scholes formula for pricing European‐style options. Like every hero, it has its flaws and no shortage of detractors ready to point them out. But with help from some friends, it can recover to play a vital role in integrating all options risk into a unified, manageable framework. This is the theme of this chapter and the next. Options risk may be subdivided into two categories: the risk of relatively liquid options, termed plain‐vanilla or vanilla options, and the risk of less liquid options, termed exotic options. Managing options risk for vanilla options is quite different from managing options risk for exotic options, so we will discuss them in two separate chapters. Almost without exception, the only relatively liquid options are European‐style calls or puts, involving a single exercise date and a simple payoff function equal to the difference between the final price level of an asset and the strike price. As such, vanilla options can be priced using either the Black‐Scholes formula or one of its simple variants (see Hull 2012, Section 14.8, Chapter 16, and Sections 17.8 and 25.13). The only notable exception to the rule that all vanilla options are European style is that some American‐style options on futures are exchange traded and liquid. However, the early exercise value of such options—the difference between their value and that of the corresponding European option—is quite small (as discussed in Section 12.5.1). So treating all vanilla options as European‐style calls and puts is a reasonable first approximation. To simplify our discussion of European options, we will utilize the following three conventions: E 1. All options are treated as options to exchange one asset for another, which enables us to only consider call options. So, for example, we treat an option to put a share of stock at a fixed price of $50 as being a call option to exchange $50 for one share of stock. This is a more natural 311 312 FINANCIAL RISK MANAGEMENT way of treating foreign exchange (FX) options than the usual approach, since whether an FX option is a call or a put depends on which currency you use as your base. 2. Options prices and strikes will often be expressed as percentages of the current forward price, so a forward price of 100 (meaning 100 percent) will be assumed. 3. All interest rates and costs of carry are set equal to zero. This means that the volatilities quoted are volatilities of the forward, not the spot; the hedges calculated are for the forward, not the spot; and option payments calculated are for delivery at the option expiry date. Although almost all options traded are paid for at contract date rather than expiry, discount curves derived from market prices, as shown in Section 10.2, can always be used to find the current spot price equivalent to a given forward payment. With these three conventions, we can use the following formula for Black‐Scholes values: BS(K, T, σ) = N(d1) − KN(d2) where (11.1) K= strike as percentage of current forward to time T T = time to option expiry in years N = cumulative normal distribution σ = annualized volatility of the forward d1 = [ln(1/K) + 1/2 σ2T]/ T σ T d2 = d1 − σ T This is similar to Equation 25.5 in Hull (2012, Section 25.13). Technically, we are using a model in which the zero coupon bond price is the numeraire (see Hull 2012, Section 27.4). Stating the equation in terms of the forward price rather than the spot price is important for reasons other than formula simplification. First, it follows the principle stated and justified in Section 6.2 that all forward risk should be disaggregated from options risk. Second, this has the advantage of not assuming constant interest rates; the volatility of interest rates and their correlation with spot price are all imbedded in the volatility of the forward. The historical volatilities of forwards can often be measured directly. If they cannot be measured directly, they can easily be calculated from the spot volatility, interest rate volatilities, and correlations. Hedges with forwards are often the most liquid hedges available. If a spot hedge is used, then the appropriate interest rate hedges should be used as well, since interest rates Managing Vanilla Options Risk 313 and carry costs cannot be assumed to be constant. This combined hedge will be synthetically equivalent to a hedge with a forward. 11.1 OVERVIEW OF OPTIONS RISK MANAGEMENT Even when we limit our discussion to vanilla options, the vast variety of instruments available makes it unlikely that liquidity of any single instrument will be large. For the options on just a single asset, not only do we face the multiplicity of dates we encountered for forward risk products, but each date also has a multiplicity of possible strikes. Once we take into account that options involve an exchange between pairs of assets, the number of possible contracts expands even more rapidly. For example, if a desk trades 10 different currencies, the number of currency pairs of FX options is 10 × 9 = 90. In fact, the degree of liquidity available for option products is significantly smaller than that for spot or forward products. When options market trading first began and, to a more limited extent, as options markets continue to develop for new assets, initial market‐maker hedging strategies were often a choice between acting as a broker (attempting to find a structure for which a simultaneous buyer and seller could be found) or relying on an initial static hedge with the underlying instrument until a roughly matching option position could be found. The broker strategy is very limiting for business growth. The static hedge strategy can only convert call positions into put positions, or vice versa; it cannot reduce the nonlinear nature of the option position. As such, it can be used only by trading desks that are willing to severely limit the size of positions (thereby limiting business growth) or to take very large risks on being right about the maximum or minimum levels to which asset prices will move. Static hedging with limited position size remains a viable strategy for a proprietary desk, but not for a market‐making desk. The development of dynamic hedging strategies was therefore a major breakthrough for the management of options market making. Consider Table 11.1, which extends an example that Hull (2012, Tables 18.1 and 18.4) presents, using Monte Carlo simulation to evaluate the performance of dynamic hedging strategies. Table 11.1 shows that even a very naive dynamic hedging strategy, the stop‐loss strategy, which calls for a 100 percent hedge of a call whenever the forward price is above the strike and a 0 percent hedge whenever the forward price is below the strike, results in a large reduction in the standard deviation of results—76 percent of option cost relative to 130 percent of option cost for a static hedge. However, an increased frequency of rehedging can only improve stop‐loss results up to this point. By contrast, the dynamic 314 FINANCIAL RISK MANAGEMENT TABLE 11.1 Performance of Dynamic Hedging Strategies Price = $49, interest rate = 5 percent, dividend rate = 0, forward price = $50 Strike = $50 Volatility = 20 percent Time to maturity = 20 weeks (0.3846 years) Drift rate = 13 percent Option price = $240,000 for 100,000 shares Performance Measure (Ratio of Standard Deviation to Cost of Option) Delta Hedge Frequency of Rehedging 5 weeks 4 weeks 2 weeks 1 week ½ week ¼ week Limit as frequency goes to 0 Stop Loss No Vol. of Vol. 10% Vol. of Vol. 33% Vol. of Vol. 102% 93% 82% 77% 76% 76% 43% 39% 26% 19% 14% 9% 44% 41% 29% 22% 18% 14% 57% 52% 45% 47% 43% 38% 76% 0% 11% 40% With no hedging, the performance measure is 130 percent. hedging strategy corresponding to the Black‐Scholes analysis enables the standard deviation to get as close to zero as one wants by a suitable increase in the frequency of rehedging. You can see why the Black‐Scholes approach had such an impact on options risk management. But almost immediately, this was followed by a backlash, focusing on the unrealistic nature of the Black‐Scholes assumptions. Principally, these assumptions and the objections are: ■ ■ ■ Trading in the underlying asset can take place continuously. (In fact, a practical limit exists on how frequently trading can occur, which places a lower limit on the standard deviation that can be achieved.) No transaction costs are involved when trading in the underlying asset. (In practice, transaction costs place an even tighter limit on the frequency of rehedging.) The volatility of the underlying asset is a known constant. (If we make the more realistic assumption that volatility is uncertain, with a Managing Vanilla Options Risk ■ 315 standard deviation around a mean, we get results like those in the last two columns of Table 11.1, placing a lower limit on the standard deviation that can be achieved.) The underlying asset follows a Brownian motion with no jumps. (In practice, discontinuous jumps in asset prices can occur, even further limiting the degree to which standard deviation can be lowered.) Trading desks that have tried pure Black‐Scholes hedging strategies for large positions have generally found that unacceptably large risks are incurred. A related example is the portfolio insurance strategy. Many equity portfolio managers were using this strategy in the mid‐1980s to create desired options positions through dynamic hedging. In October 1987, the global stock market crash caused liquidity to dry up in the underlying stocks, leading to trading discontinuities that resulted in large deviations from planned option payoff profiles. As a result, vanilla options market makers have generally moved in the direction of a paradigm in which they attempt to match the options positions bought and sold reasonably closely, enabling basis risk to be taken both over time while waiting for offsetting trades to be available and with regard to strike and tenor mismatches. The Black‐Scholes model is relied on as an interpolation tool to relate observed market prices to prices needed for the residual risk positions left after offsetting closely related buys and sells. Black‐Scholes dynamic hedging is used to hedge these residual risk positions. Three key tools are needed for managing a vanilla options book using this paradigm: 1. A reporting mechanism must be available to measure the amount of basis risk exposure resulting from mismatches in the strike and tenor of options bought and sold. Although summary measures such as vega (exposure to a move in implied volatility levels) and gamma (the sensitivity of delta to a change in underlying price level) can be useful, the two‐dimensional (strike and tenor) nature of the exposure requires a two‐dimensional risk measure to be really effective. This measure is the price‐vol matrix that depicts portfolio valuation sensitivity to the joint distribution of two variables: underlying asset price and implied volatility. It therefore measures exposure to both jumps in underlying asset price and changes in implied volatility. It also measures simultaneous changes in both. We will examine illustrative examples and discuss the use of price‐vol matrices in Section 11.4. 2. Dynamic delta hedging of the portfolio of bought and sold options needs to be performed. Guidance for this process comes from the Black‐Scholes formula. The targeted hedge for the portfolio is a simple 316 FINANCIAL RISK MANAGEMENT summation of the targeted hedges of each individual option position, as determined by Black‐Scholes. However, given the reality of transaction costs for executing the delta hedges in the underlying, a set of guidelines about how often to hedge is necessary. It has been shown, both by theory and trader experience, that hedging guidelines based on the distance between the current delta hedge and the target delta hedge are more effective than guidelines tied to the frequency of hedging. The degree of tolerance for deviation from the target delta determines a trade‐off between higher transaction costs (for lower tolerances) and higher uncertainty of results (for higher tolerances). Section 11.5 discusses these delta‐hedging guidelines in more detail along with related issues such as what implied volatility to use to determine the target hedge. 3. Options for which liquid market prices are not available are valued based on interpolation from options that do have liquid market prices available. The interpolation methodology translates prices of liquid options into implied volatilities using the Black‐Scholes formula, interpolates these implied volatilities to implied volatilities for less liquid options (interpolation is based on both strike and tenor), and then translates implied volatilities to prices of the less liquid options, again using the Black‐Scholes formula. Limits and reserves are needed to control uncertainty in the interpolation process. Section 11.6 gives a detailed account of this interpolation method. Note how closely bound together the three operative legs of this paradigm are. The Black‐Scholes formula serves as the glue that binds them together: ■ ■ ■ The price‐vol matrix shows how the portfolio valuation will change based on a joint distribution of changes in underlying asset price and implied volatility. However, many (probably most) of the options in the portfolio lack liquid market prices, so their valuation depends on the interpolation step. Furthermore, the calculation of the change in option value for a change of asset price and implied volatility is calculated using the Black‐Scholes formula. As will be seen in the detailed discussion of the price‐vol matrix, all calculations are done under the assumption that exposure to small changes in underlying asset price have been delta hedged with a position in the underlying asset, so the validity of the price‐vol matrix depends on the execution of this dynamic delta hedging. The need for this approach to options risk management is based on the flat rejection of the key assumptions of the Black‐Scholes model: continuous rehedging, no transaction costs, no price jumps, and known and Managing Vanilla Options Risk 317 constant volatility. How, then, can we continue to rely on the Black‐ Scholes model to calculate the impact of changes in underlying asset price, calculate the target delta hedges, and play a critical role in value interpolation? The answer is that position limits based on the price‐vol matrix are being counted on to keep risk exposures low enough that deviations from the Black‐Scholes assumptions will not have that large an effect. Small risk exposures mean that the size of required delta hedges will be small enough that transaction costs will not be that significant. Small risk exposures mean that the differences between the Black‐Scholes model and the presumably much more complex true model (whatever that may be) are small enough to hold down the errors due to valuing and hedging based on a model that is only an approximation to reality. It is important to be aware of the degree to which this paradigm depends on the availability of market liquidity for hedging instruments. The paradigm works best when reasonable liquidity in vanilla options is available for at least some combinations of strike and tenor. This enables risks to be hedged by actively pursuing the purchase and sale of options to lower exposures as measured by the price‐vol matrix. As we will see in Exercise 11.1, price‐vol matrix exposures can be held reasonably flat even if only a small number of strike‐tenor combinations provide significant liquidity. The valuation of options with other strike‐tenor combinations can be interpolated from the liquid set. If a particular options market does not have liquidity, the paradigm can still work reasonably well as long as the underlying asset has liquidity. The price‐vol matrix now serves primarily as a measure of position imbalance. It can serve as a signal to marketers to encourage customer business at some strike‐tenor combinations and discourage it at others. It can be used to place limits on new customer business when this would cause risk to exceed management guidelines. It can be used as input to setting limits and determination of reserves against illiquid concentrations of risk. It can also be used as input to calculations of portfolio risks such as value at risk (VaR) and stress tests. Price interpolation, in the absence of liquid market quotations, becomes primarily a mechanism to enforce the consistency of valuations. Delta hedge calculations continue to serve the function of directing dynamic hedging and ensuring the proper representation of options positions in firmwide reports of spot and forward risk. It is far more questionable to employ this paradigm in the absence of liquidity in the underlying asset. In this case, it is doubtful that dynamic delta hedging can be carried out in any systematic way, and it probably becomes preferable to analyze positions based primarily on how they will behave under longer‐term scenarios, with limits and reserves calculated 318 FINANCIAL RISK MANAGEMENT from this scenario analysis. An example where this may apply is for options written on hedge fund results where there are restrictions on the ability to buy and sell the underlying, which is an investment in the hedge fund. A specific case to illustrate this point is the option Union Bank of Switzerland (UBS) wrote on Long‐Term Capital Management (LTCM) performance (see Section 4.1.5). How well does this paradigm work? Trading desks that have years of experience using it have generally been satisfied with the results. But this is insider knowledge and may be specific to conditions in particular markets. How can outsiders get comfortable with these assumptions, and how can these assumptions be tested in new options markets to which they might be applied? The best tool available is Monte Carlo simulation, in which all of the Black‐Scholes assumptions can be replaced with more realistic assumptions, including limits on hedge frequency, transaction costs, uncertain volatility, nonlognormal changes in the underlying price, and price jumps. In Section 11.3, we examine the results of a typical Monte Carlo simulation to see what it indicates about the feasibility of this risk management paradigm. 11.2 THE PATH DEPENDENCE OF DYNAMIC HEDGING To understand options pricing, an important distinction must be made between path‐independent and path‐dependent options. A path‐independent option’s payout depends only on what the price of some underlying asset will be at one particular point in time and does not depend on the actual path of price evolution between the current date and that future date. All European‐style options are path independent. Exotic options are divided between path‐independent and path‐dependent options. In Chapter 12 on managing exotic options risk, we will see that path‐independent options are generally much easier to risk manage than are path‐dependent options. Although, when considered in isolation, European‐style options are path independent, once we start to evaluate the impact of dynamic hedging, we find that dynamic hedging makes “every option become path dependent.” (This is quoted from Taleb [1997, Chapter 16]. I strongly recommend reading Taleb’s Chapter 16 along with this chapter.) This is a direct consequence of the limitations of the Black‐Scholes assumptions, since continuous hedging at a known constant volatility would result in a definite value with no variation (hence, you would achieve not just path independence, but independence of the final underlying asset value as well). Sporadic dynamic hedging and stochastic volatility make the realized value of a dynamic hedging strategy dependent on the full price history of the underlying asset. Let’s illustrate this with a few examples. Managing Vanilla Options Risk 319 The first example is based on one presented in Taleb (1997, 270). It is an out‐of‐the‐money call on $100 million par value of a stock with 30 days to expiration that is purchased for $19,000. If no dynamic hedging is attempted, then the option will expire either out‐of‐the‐money for a total loss of the $19,000 premium or in‐the‐money with upside potential. The amount of return will be completely dependent on where the underlying asset price finishes in 30 days. Suppose a trader wanting to reduce the uncertainty of this payoff attempts to dynamically hedge her position. Taleb demonstrates a plausible price path for the underlying asset that results in a loss of $439,000, not even counting any transaction costs. The NastyPath spreadsheet provided on the course website enables you to see the details of this path and experiment with the impact of other possible paths. What is it about the path that leads to a loss that is so large relative to the option’s cost? Try to reach your own conclusion. I will provide my answer at the end of Section 11.5. The second example is drawn from my own experience. In early 1987, I was part of a team at Chase Manhattan that introduced a new product—a term deposit for consumers that would guarantee a return of principal plus a small interest payment, but could make higher interest payments based on a formula tied to the closing price of the Standard & Poor’s (S&P) stock index on the maturity date of the deposit. Although the stock market had been showing very good returns in the mid‐1980s, stock market participation among smaller investors was still not well developed. Therefore, a product that would be Federal Deposit Insurance Corporation (FDIC) insured, would guarantee against loss, and would provide some upside stock participation quickly attracted a sizable amount of investment. Our hedging strategy for this product was to invest part of the proceeds in standard deposit products, ensuring the ability to return principal plus guaranteed minimum interest, and use the remainder to fund an S&P index call position. As might be anticipated by those who remember the financial events of 1987, this product suffered an untimely demise in the autumn of that year. After the stock market crash of October 19, consumer interest in possible stock market participation sharply diminished, so new funds stopped coming in. We also experienced severe losses on our hedging of the existing product, and the postmortem we conducted to determine the reason for these losses produced some interesting results. The equities options markets were at a very early stage of development in 1987, so there was virtually no liquidity for options with tenors beyond a few months. Since our market research had determined that there would be little interest in a deposit product with tenors shorter than a year or two, we had decided to initially rely entirely on a dynamic hedging strategy, using a Black‐Scholes–determined delta hedge. We were certainly aware of the 320 FINANCIAL RISK MANAGEMENT vulnerability of this approach to high volatility, but we had done extensive research on the historical patterns of stock market volatility and concluded that we could price the product at an implied volatility that allowed a margin for error that would result in hedging losses only in extremely rare cases. Not surprisingly, our postmortem showed significant losses due to our inability to carry out the delta‐hedging strategy during the period of October 19 and the following few days. The cash and futures equities markets during that period were highly illiquid in the face of panicky selling, and there were even some short periods in which the markets were closed in an attempt to restore stability to chaotic trading. Illiquid markets in the underlying during large price moves result in gapping losses to options sellers employing dynamic hedging strategies. We were not alone in this vulnerability. In October 1987, a substantial number of asset managers following portfolio insurance strategies in which they attempted to achieve the payoff profiles of an option through delta hedging experienced heavy losses as a result of this gapping. What was less expected, though, was our finding that a considerable part of our loss would have been experienced even if the markets had not gapped. Our loss was due to higher‐than‐anticipated volatility. This was despite the fact that when we looked over the tenor of our deposit product the average realized volatility was well within the range we had anticipated in pricing the product. Here’s where path dependence comes in. The average realized volatility consisted of very high volatility during a short period when the market was plunging sharply, which was preceded and followed by periods of much lower volatility. However, exposure to volatility depends on the relationship between the price level and strike. The higher‐than‐average volatility during the period when prices were falling sharply cost us much more than we saved from the lower‐than‐average volatility during the other periods. This phenomenon can be easily illustrated with some simple Black‐Scholes calculations. Suppose you have written a one‐year call option with a strike equal to the current forward price. You intend to delta hedge and expect volatility to average 20% over the year. If you are wrong and volatility averages 30%, your expected losses will be BS(100%, 1, 30%) − BS(100%, 1, 20%) = 11.923% − 7.966% = 3.957%. Suppose one‐tenth of a year goes by and the forward price is at the same level as when you wrote the option. Your remaining exposure to volatility averaging 30% is BS(100%, 0.9, 30%) − BS(100%, 0.9, 20%) = 11.315% − 7.558% = 3.757%. So 3.757%/3.957% = 94.9% of your volatility exposure comes in the last 90% of the option’s life and only 5.1% comes in the first 10% of the option’s life (a consequence of the fact that .9 / 1 = .949). However, if the price at the end of one‐tenth of a year has fallen by 30%, the remaining exposure to volatility averaging Managing Vanilla Options Risk 321 30% is BS(70%, 0.9, 30%) − BS(70%, 0.9, 20%) = 1.188 − 0.184 = 1.004. So (1.004%/3.957%) = 25.4% of your volatility exposure comes in the last 90% of the option’s life and 74.6% comes in the first 10% of the option’s life. A very similar effect will be seen for a large rise in underlying price. With the benefit of experience, we concluded that we had badly underestimated the risk of the product. First, we had not taken into account the potential losses from pricing gaps. Second, the chances of volatility being very high during a short time period are much larger than the chances of it being very high during a long time period, so we had not properly calculated our vulnerability to a short period of high volatility combined with a large price move. Third, we had not looked at the impact of other market participants pursuing strategies similar to ours, thereby decreasing liquidity by competing with us for hedges in the underlying when we most needed them. What would have been a more prudent way of managing this risk? We had been considering, but had not implemented, a proposal from a broker in exchange‐traded, shorter‐term S&P options for a hedge of our longer‐term options with these shorter‐term options. See Section 11.6.3 for a discussion of the risk characteristics of this hedge. 11.3 A SIMULATION OF DYNAMIC HEDGING In the immediately preceding section, we established that, under realistic economic assumptions, dynamically hedged options are path dependent. In the section before that, we observed the need for testing how well the paradigm of managing options risk using Black‐Scholes theory works. Both sections point toward using Monte Carlo simulation to see what the probability distribution of results can be for dynamically hedging an options portfolio. Using Monte Carlo simulation for dynamic hedging options is an invaluable tool for understanding how the management of an options trading book works in practice. When new options products or hedging strategies are proposed, traders and risk managers alike will want to look at simulation results to assess potential pitfalls. This is an example of the use of simulation in model testing recommended in Section 8.4.3. Simulation gives the flexibility to take into account the impact on hedging results of real‐life constraints such as liquidity constraints on the size of changes in hedges that can be performed in a given time period (or the impact of larger changes on the price at which the hedge can be executed). Simulation also provides a vital learning tool for people who are unfamiliar with the workings of options markets. Theoretical demonstrations 322 FINANCIAL RISK MANAGEMENT of the power of dynamic hedging rarely carry the conviction that can be provided by observing hundreds of simulation paths that, despite wild gyrations in underlying prices, produce almost identical hedging results. Nothing short of actually suffering through a losing options strategy can convey the pain of an unsuccessful hedge as will observing the losses pile up on a simulation path. In the course I teach, on which this book is based, I have always insisted that each student personally program and run simulations of a dynamic hedge. I lack a comparable power of persuasion over readers of this book, but I urge each of you to do as much of Exercise 11.2 as you can. Even if you lack the time to program your own simulation, you should at least do parts 4 and 5 of this exercise using the provided spreadsheets. What features do we want a Monte Carlo simulation of dynamic hedging to contain? ■ ■ ■ ■ ■ ■ ■ The simulation must be over a sufficiently large number of possible price paths to produce stable statistics. Prices for the underlying variable must be sampled at enough points on each path to allow for rehedging. Since volatility of the underlying price is not constant, but is a stochastic variable, a random process should drive it. Data to determine reasonable values of volatility can be obtained by looking at historical distributions of realized volatility for separate time periods. A separate volatility should be chosen for each path generated. The distribution of the underlying price does not necessarily need to be lognormal. Different mixtures of normal and lognormal processes should be tried. Rehedges should be allowed only at periodic intervals, and transaction costs of the hedge should be calculated explicitly. Different rules for determining hedge amounts, as discussed in Section 11.5, should be considered. When calculating Black‐Scholes deltas for rehedging, you generally do not want to take advantage of knowing what volatility is being used for the path, since this would not be available in making actual hedging decisions. Either you want to use the same implied volatility to calculate rehedges on all paths or you want to use some adaptive rule tying volatility used to the history of price moves on the path up to the time of the rehedge. A random process of significant price jumps, where no rehedging is permitted until after the jump is completed, can be used as a simulation of periods of illiquidity. When simulating a portfolio of options for one particular expiry date, it is usually convenient to assume that all hedges are performed with Managing Vanilla Options Risk 323 a forward with the same expiry to avoid needing to keep track of discounting rates. When simulating options with different expiry dates, some assumptions about discounting rates must be used to arrive at relative prices between forwards. In effect, we are testing the performance of the Black‐Scholes model as a hedging tool by running a Monte Carlo simulation based on a more complex, and presumably more accurate, model of underlying price behavior than Black‐Scholes utilizes. Why not just value and hedge options by directly using this more complex and complete model? For two reasons: 1. Computational complexity. The speed of the computation of the Black‐ Scholes model for valuation and the fast and direct computation of the target underlying hedge are enormous advantages in providing timely risk information on portfolios of options that may have many thousands of deals outstanding at any given time. By contrast, more complex models can be orders of magnitude slower when computing valuations and often lack a direct computation of target hedges, requiring multiple runs of the valuation algorithm to determine the appropriate hedge. This advantage can particularly be seen in Monte Carlo testing of hedge effectiveness. At each potential rehedge point, the Black‐Scholes target hedge is a simple equation; a more complex model may require full recalibration to compute each hedge (see Section 12.3.2 for a discussion of this point in conjunction with hedging barrier options). 2. Validity. We don’t necessarily know what the correct model is. For testing hedge performance with Monte Carlo, we can make different runs with alternative candidates for the correct model. As a first example of a simulation, let’s look at a comparison between hedging an option using a pure Black‐Scholes hedge and hedging using a combination of Black‐Scholes delta hedging and hedging with other options. We may suppose that an option has been sold at a strike for which no liquidity is readily available. We can either utilize a dynamic hedging strategy or buy some options at strikes for which liquidity is available and then utilize a dynamic hedging strategy for the residual risk. For this example, we will assume that a one‐year option has been sold at a strike 5 percent in‐the‐money and that one‐year options are available for purchase at strikes at‐the‐money and 10 percent in‐the‐money. For the second case, we will consider purchasing the same notional amount of options as has been sold, but split 50–50 between the at‐the‐money option and 10 percent in‐the‐money option. The reason for thinking that this might be a good hedge will be shown in Section 11.4. There we will see that the 324 FINANCIAL RISK MANAGEMENT TABLE 11.2 Monte Carlo Simulation Comparing Pure Dynamic Delta Hedging with Combined Static Option and Dynamic Delta Hedging Standard Deviation of Standard Deviation of P&L P&L Given 0% Standard Given 33% Standard Deviation of Volatility Deviation of Volatility Number of Rebalancing Unhedged Two‐Sided Hedge Unhedged Transaction Costs Two‐Sided Two‐Sided Hedge Unhedged Hedge 10 20 25.7% 19.8% 6.4% 5.6% 50.6% 41.5% 6.3% 6.7% 1.5% 2.2% 0.1% 0.2% 50 12.4% 4.6% 40.9% 5.5% 3.5% 0.4% 100 8.5% 3.6% 42.6% 4.9% 5.0% 0.6% 200 6.3% 2.5% 41.6% 4.8% 7.1% 0.9% 300 5.1% 1.9% 39.9% 3.8% 8.5% 1.1% 400 4.3% 1.8% 40.1% 4.1% 9.9% 1.2% 500 3.9% 1.6% 38.4% 3.9% 11.2% 1.4% 600 3.5% 1.4% 35.9% 3.4% 12.0% 1.5% 700 3.3% 1.3% 41.0% 3.5% 13.3% 1.6% 800 900 3.2% 2.9% 1.3% 1.5% 39.2% 40.0% 3.7% 3.8% 14.4% 15.0% 1.7% 1.9% All results are shown as a percentage of the cost of the option to be hedged. The option is a one‐year call struck 5 percent in‐the‐money. The expected volatility is 20 percent, and all hedges are calculated based on a 20 percent implied volatility. The two‐sided hedge has half a call struck at‐the‐money and half a call struck 10 percent in‐the‐money. Transaction costs are based on a bid‐ask spread of one‐fourth point per $100. price‐vol matrix for this portfolio (Table 11.9) shows very little sensitivity to changes in either the price level or implied volatility. This does not, by itself, prove that the hedge will work well over the life of the option, since it only shows a snapshot at one point in time. In fact, you will be able to see from Tables 11.10 and 11.11 in Section 11.4 that although this portfolio does continue to show low sensitivity to price on volatility shifts for a substantial time period, this sensitivity increases at some point in its evolution. So we need the Monte Carlo simulation to get a statistical measure of the sensitivity. Table 11.2 shows the results of the simulation. Managing Vanilla Options Risk 325 In the context of the discussion of model risk in Section 8.4, the 50−50 mixture of at‐the‐money option and 10 percent in‐the‐money option constitutes the liquid proxy that would be used to represent the 5 percent in‐the‐money option in standard risk reports, such as VaR and stress tests. The Monte Carlo simulation would be used to generate a probability distribution of how much extra risk there is in holding the 5 percent in‐ the‐money option than there is in holding the liquid proxy. The assumption that the 50−50 mixture will constitute a good hedge all the way to the expiration of the option is a simplifying assumption that makes the Monte Carlo simulation easier. In reality, a trading desk would change this mixture through time, particularly as time to option expiry was close. But while a Monte Carlo simulation that included changes in the mixture would be more realistic, it would also be far more difficult to perform. Changes in the volatility surface would need to be simulated, since changes in the mixture will require purchases and sales of options at future dates; transaction costs for purchases and sales of options would need to be included; behavioral rules for trading decisions would be needed on the trade‐off between these transaction costs and the desirability of changing the mixture. What conclusions can we reach? ■ ■ If the standard deviation of volatility is zero, then both the pure dynamic hedging and the mixed‐option/dynamic hedging strategies can achieve as low a standard deviation of results as you like by increasing the frequency of rebalancing the dynamic hedge, although the mixed strategy achieves a given level of standard deviation with far fewer rebalancings than the pure strategy. For either strategy, there is a trade‐off between higher expected transaction costs with more frequent rebalancing and lower standard deviations of results. (Standard deviations of total results, including transaction costs, don’t differ significantly from the standard deviations without transaction costs, which are shown in Table 11.2.) However, the mixed strategy can achieve a desired level of standard deviation at a far lower transaction cost level than the pure strategy. For example, achieving a 3% standard deviation with the pure strategy requires about 900 rebalancings with an associated transaction cost of 15.0%. Achieving a 3% standard deviation with the mixed strategy requires about 150 rebalancings with an associated transaction cost of about 0.8%. If the standard deviation of volatility is 33%, then there is a lower bound on how much the standard deviation of results can be decreased. For both the pure and mixed strategies, this lower bound is reached at about 250 rebalancings. The lowest level of standard deviation of 326 ■ FINANCIAL RISK MANAGEMENT results that can be achieved by the mixed strategy is about one‐tenth of what can be achieved by the pure strategy, roughly 4% compared to roughly 40%. The inability to reduce the standard deviation of results below a lower bound is due to both the uncertainty of volatility and the use of incorrect volatility inputs in forming hedge ratios. However, the first effect is many times larger than the second. A Monte Carlo run with 33% standard deviation of volatility, but with hedge ratios on each Monte Carlo path based on the actual volatility of that path, results in a lower bound on the standard deviation of results that is only reduced from 40 to 36% Please note that although we are using standard deviation as a convenient summary statistic to give a rough feel for relative levels of uncertainty, both in this example and others in this book, more detailed analysis would be needed before arriving at any precise conclusions. For example, if a measure was being developed for a risk versus return trade‐off as input to a decision on a trading strategy, a more complete set of measures of the probability distribution of returns should be used. The discussion of measures of portfolio risk in Section 7.1.2 gives more of a flavor for these considerations. These results will not be surprising when we examine the price‐vol matrix in Table 11.9 in Section 11.4. From the relative insensitivity of portfolio value to a shift in implied volatility we will see there, you would expect low sensitivity to the standard deviation of volatility. The small size of the portfolio’s convexity translates into small changes in the delta when prices move, so transaction costs should be low. A reasonable inference, which is supported by experience with Monte Carlo simulations, is that a trading desk can estimate its vulnerability to uncertain volatility and transaction costs by forecasting how large its price‐vol matrix positions are likely to be given the anticipated flows of customer business and the availability of hedges with liquid options. Management can keep these vulnerabilities under control by placing limits on the size of price‐vol matrix positions. It is important to recognize the distinction between the two aspects of dynamic hedging costs—transaction costs that arise from bid‐ask spreads and gamma hedging costs from buying high and selling low that would be present even if all trades were at midmarket. Transaction costs are a direct function of the frequency of rehedging, and a trade‐off occurs between higher transaction costs and lower variability of profit and loss (P&L) with less frequent rehedging. By contrast, there is no a priori reason to believe 327 Managing Vanilla Options Risk that the level of gamma hedging costs will vary in any systematic way with the frequency of rehedging. A good way to see this latter point is to look at how P&L is related to the gap between actual hedges held and the theoretical hedge called for by the Black‐Scholes formula. The expected value of this P&L under the standard Black‐Scholes assumption is given by the formula: ∑( small time periods actually held theoretical ) × expected price change of underlying forward (11.2) A full mathematical derivation of this formula can be found in Gupta (1997). I will give an alternative derivation using a simple financial argument. In the presence of the Black‐Scholes assumptions, use of the theoretical delta will lead to an expected return of zero, so any holdings above or below the theoretical delta can be regarded as proprietary positions that will lead to the same expected return as an outright position in the underlying forward. The consequence of this formula for the relationship between gamma hedging costs and the frequency of rehedging is that as rehedging becomes less frequent, it widens the gap between Δ actually heldd and Δ theoreticall. However, unless a correlation between the sign of this gap and the sign of the expected price change in the underlying forward is expected for some reason, the expected value of the incremental P&L should be zero. (Although this formula is strictly correct only in the case where the Black‐Scholes assumptions hold, Monte Carlo simulation with stochastic volatility shows similar results.) Are there cases where we might expect a relationship between the sign of the delta gap and the sign of expected price changes in the underlying forward? Let’s consider a case that will cast an interesting light on a long‐standing debate among practitioners. The debate is over what options pricing is appropriate for a market in which the underlying process shows mean reversion, resulting in a narrower dispersion of future price levels than would be implied by a pure random walk with the short‐term volatility of the underlying process. One group argues that delta‐hedging costs are completely a function of short‐term volatility, so mean reversion is irrelevant to pricing. The opposing group argues that risk‐neutral valuation principles should result in the same pricing of options as would be implied by the probability distribution of final prices; compare the discussion here to Rebonato (2004, Sections 4.7 and 4.8). Some of this dispute reflects a failure to distinguish between the short‐ term volatility of spot prices and forward prices. If the market is pricing 328 TABLE 11.3 FINANCIAL RISK MANAGEMENT Impact of Drift and Mean Reversion on Dynamic Hedging Results All Paths 20 rehedges −0.07% 100 rehedges +0.01% 1,000 rehedges −0.01% Upward Drift Downward Drift −0.33% −0.06% −0.45% −0.10% 0% 0% Mean Reversion +0.57% +0.20% −0.02% the mean reversion process into the forward prices, we should expect to see a lower historical short‐term volatility of forward prices than a historical short‐term volatility of spot prices. Equivalently, this can be viewed as a correlation between changes in spot prices and changes in the discount rate of the forwards, a pattern that can be seen in the market for seasonal commodities. When seasonal demand is high or seasonal supply is low, spot prices rise, but so does the discount rate, dampening the rise in forward prices. When seasonal demand is low or seasonal supply is high, spot prices fall, but so does the discount rate, dampening the fall in forward prices. Since the option can be delta hedged with the forward, replication costs will be tied to the volatility of the forward, so we should expect implied option volatilities to reflect the impact of mean reversion relative to the volatility of the spot price. Suppose that a trader believes that the market has not adequately priced in mean reversion, so he expects that forward prices will show mean reversion. In this case, we cannot resolve the controversy between the two differing views on options pricing by an appeal to the difference between short‐term volatility of spot and forward prices. Let us look at the results of a Monte Carlo simulation in which we ignore transaction costs and study the impact of rehedging at a fixed number of evenly spaced intervals. We will calculate statistics for the whole sample of paths, but also for three subsamples: 1. The third of paths having the highest finishing forward prices, which we can take as representing upward drift of the forward. 2. The third of paths having the lowest finishing forward prices, which we can take as representing downward drift. 3. The remaining third of the cases, which we can take as representing mean reversion. Table 11.3 shows the resulting expected values of a delta‐hedging strategy for a written (sold) option (for a purchased option, the signs would be reversed). Managing Vanilla Options Risk 329 What conclusions can we draw? ■ ■ ■ ■ As you increase the frequency of rehedging, you get the same expected results regardless of drift or mean reversion. This is consistent with the theoretical result that, under the Black‐Scholes assumptions, standard deviation of results goes to zero as the frequency of rehedging increases, so the P&L will be the same on every path. It is also consistent with Equation 11.2, since frequent rehedging drives the difference between the Δ actually heldd and Δ theoreticall terms to zero. As you decrease the frequency of rehedging, you increase the losses from a sold option with drift or a purchased option with mean reversion, and you increase the gains from a sold option with mean reversion on a purchased option with drift. All of these results are consistent with Equation 11.2. For example, here’s the reasoning for mean reversion on a sold option: It is likely that one period’s up move will be followed by the next period’s down move, and vice versa. After an up move, the Δ theoreticall on the sold option will increase, but if no rehedge is performed, due to the infrequency of rehedging, this will make the Δ actually held − Δ theoreticall for the next period be negative. Since the expected price change in the next period is negative, the expected P&L is the product of two negatives, and hence positive. The consequence of the last point for hedging strategies is that if you anticipate mean reversion, you should try to decrease hedging frequency for a sold option (which also saves transaction costs, but increases the uncertainty of return) and try to increase hedging frequency for a bought option (but this needs to be balanced against the increase in hedging costs and uncertainty of return). This is intuitively correct. As the option seller, you want to hold off on rehedging since you expect the market to rebound; as the option buyer, you want to take advantage of the market move with a rehedge prior to the expected rebound. Conversely, if you anticipate a drifting market, whether up or down, you should try to decrease hedging frequency for a bought option and increase hedging frequency for a sold option. If you cannot anticipate either drift or mean reversion, there is no difference in gamma hedging costs based on the frequency of rehedging, so the decision rests purely on the trade‐off between transaction costs and the uncertainty of return. 11.4 RISK REPORTING AND LIMITS The best tool for managing residual options risk on a trading desk is the price‐vol matrix, which depicts valuation sensitivity to joint 330 FINANCIAL RISK MANAGEMENT distributions of two variables: the asset price and implied volatility. The PriceVolMatrix spreadsheet on the website for this book calculates a price‐vol matrix for a small portfolio of options. See the accompanying documentation for details. We will note just three important points about the computation: 1. All boxes in the matrix represent full valuations using the Black‐Scholes model utilizing the shifted volatility level and underlying price level. No approximations are being used in the computation. 2. Each box assumes that an underlying position has been put on to neutralize the initial delta position of the options. 3. Only the initial delta position is neutralized; no delta rehedging is allowed during a price shift. Therefore, the price‐vol matrix represents the potential impact of price jumps that cannot be delta hedged. For those who respond better to visual presentations than to numerical information, the spreadsheet produces two graphical representations of the price‐vol matrix: 1. A three‐dimensional surface of the P&L consequences of changes in the underlying price and implied volatility. 2. A chart showing changes in valuation, delta, vega, and gamma as price levels change. The price‐vol matrix enables a trading desk manager to see at a glance the convexity (the nonlinear impact of large price changes), vega (sensitivity to a small change in implied volatility), nonlinearities in vega, and interactions between convexity and vega. The price‐vol matrix can pick up discontinuities caused by strikes in a portfolio clustering around certain levels. In order for the price‐vol matrix to highlight nonlinear effects, it is best to assume that any linear delta position has already been hedged. To the extent that the trading book chooses not to hedge delta risk, the resulting underlying position should be reported separately and be subject to limits separate from those on options positioning for the reasons given in Section 6.2 concerning the need for clear separation of linear and nonlinear risks. Traders have recently shown greater focus on the sensitivity of vega to changes in implied volatility and the sensitivity of vega to changes in spot. A sign of this increased focus is that these sensitivities have acquired their own mock Greek names, vomma, also known as wisoo, and vanna, also known as DdelV, V respectively. Note that the price‐vol matrix measures changes in P&L impact due to both vomma and vanna. Also note that the convexity measure goes well beyond a simple P&L impact of gamma, which is just the Managing Vanilla Options Risk 331 second derivative of price changes, and hence determines the second‐order term in the Taylor expansion of option price in terms of underlying price. Since the matrix is filled in by a full revaluation of the Black‐Scholes model for each box, the impact of as many terms in the Taylor series as desired can be picked up by a sufficient refinement of the underlying price grid. The price‐vol matrix is a valuable tool both in the daily P&L reconciliation needed to control model risk (compare with Section 8.2.7.1) and in P&L approximations used in VaR and stress test calculations (compare with Section 7.1.1.2). When the price‐vol matrix is used for making P&L approximations, it is often referred to as a heat map. For P&L reconciliation, it allows a quick first‐cut calculation of P&L change from the close of one business day to the close of the next business day due to the combined change in underlying price and overall volatility level if no delta hedging had been performed during the day. It can then be supplemented by more detailed calculations of P&L changes due to changes in the shape of the volatility surface and due to delta hedging performed during the day. P&L due to changes in volatility shape can be calculated from a matrix that breaks down vega exposure by strike and by time to expiry (the PriceVolMatrix spreadsheet contains a sample computation of a vega exposure matrix). Another valuable tool for P&L approximation is dollar gamma. Dollar gamma is calculated as one‐half the gamma multiplied by the square of the current price level. Its use in P&L approximation is that when you multiply a portfolio’s dollar gamma by the difference between the volatility at which positions have been marked and the actual price move for the day, you get a good first estimate of a delta‐hedged portfolio’s P&L for the day. A good explanation of dollar gamma and sample calculations can be found in Allen, Einchcomb, and Granger (2006, Section 4.1). We will now use the price‐vol matrix to examine some representative option positions as a way to learn about both risk characteristics of the positions and the analytic power of the price‐vol matrix: ■ ■ Short a call option. This is the simplest possible options portfolio. We are short one unit of a one‐year call struck at‐the‐money. Table 11.4 shows the price‐vol matrix. Naturally, vega and gamma are both negative, and vega remains negative at all price levels. Negative vega is largest at‐the‐money and declines as prices rise and fall, reflecting the decline in the time value of an option as it goes into or out of the money. The negative gamma is reflected in large losses from either up or down price jumps at the current volatility. Call spread. We are short one unit of a one‐year call option struck at the money and long 1.06 units of a one‐year call option struck at 110 percent of the forward price. Table 11.5 shows the price‐vol matrix. The 332 Delta Vega Gamma Theta −0.38% −2.0% 0.015% −1 0.76% −0.49% −2.07% −3.87% −0.80% 0.90% 2.01% 2.42% 2.14% 1.27% −0.08% −1.77% −3.67% −15 −10 −5 0 5 10 15 20 25 −4.13% −2.41% −0.93% 0.23% 0.97% 1.21% 0.89% 0.00% −1.43% −3.30% −5.51% −4% −4.43% −2.80% −1.40% −4.78% −3.22% −1.90% −0.88% −0.32% −5.16% −3.67% −2.42% −1.45% −0.83% −0.23% 0.37% −5.57% −4.14% −2.95% −2.03% −1.43% −1.21% 0.60% −0.60% −1.99% −3.02% −4.45% −6.24% 4% 0.00% −1.48% −2.59% −4.12% −6.01% 2% −1.38% −0.24% −0.97% −2.18% −3.82% −5.81% 0% −0.81% 0.33% −0.48% −1.79% −3.55% −5.64% −2% Implied Volatilities Price‐Vol Matrix for Being Short a Call Option 1.56% 1.81% 1.45% 0.46% −1.10% −3.10% −2.94% −5.40% −5.33% −20 −6% 0.015% −2.0% −0.38% −54.0% −7.58% 20.0% 1 100 100 call −25 −8% BS price −54.0% TABLE 11.4 Price Time Implied vol −7.58% 2% Spot-Vol Matrix Portfolio Strike Price Price 5 call/put Spacing Volatility 5.00% Volume Discount −6.00% −4.64% −3.50% −2.62% −2.04% −1.81% −1.96% −2.51% −3.46% −4.80% −6.50% 6% −6.46% −5.15% −4.05% −3.20% −2.65% −2.41% −2.53% −3.03% −3.91% −5.17% −6.77% 8% −0.18% −0.22% −0.25% −0.28% −0.30% −0.30% −0.28% −0.25% −0.20% −0.14% −0.09% Vega 0.01% 0.01% 0.01% 0.00% 0.00% 0.00% 0.00% 0.00% 0.01% 0.00% 0.00% Vega Convexity 333 Delta Vega Gamma Theta −16.5% 0.00% 0.0% 0.000% TABLE 11.5 Price −1 2.0% 0.08% −0.20% −0.41% −0.48% −0.33% 0.02% 0.36% 0.56% 0.46% 0.14% −0.27% −0.61% −0.76% −0.65% −0.28% −15 −10 −5 0 5 10 15 20 25 0.31% −0.04% −0.22% −0.24% −0.13% 0.03% 0.17% 0.19% 0.03% −0.33% −0.87% −4% Price‐Vol Matrix for a Call Spread 0.30% 0.36% 0.19% −0.21% −0.10% −20 −0.80% −0.74% −6% −0.014% 0.38% −2.0% 0.015% 37.4% −0.38% 4.33% 20.0% 1 110 100 call 1.06 −54.0% −7.58% 20.0% 1 100 100 call −25 −8% BS price vol Implied Time Strike −3.25% 2% Spot-Vol Matrix Portfolio Volatility Price Price call/put Spacing 5 5.00% Volume Discount 0.59% 0.22% 0.00% −0.08% −0.06% 0.01% 0.07% 0.04% −0.12% −0.46% −0.95% 0.85% 0.46% 0.21% 0.06% 0.01% 0.00% −0.01% −0.08% −0.26% −0.58% −1.04% 0% Implied Volatilities −2% 1.09% 0.69% 0.39% 0.20% 0.08% 0.00% −0.07% −0.19% −0.38% −0.70% −1.14% 2% 1.31% 0.89% 0.57% 0.32% 0.15% 0.01% −0.12% −0.27% −0.49% −0.81% −1.23% 4% 1.53% 1.09% 0.73% 0.44% 0.21% 0.03% −0.15% −0.34% −0.59% −0.91% −1.32% 6% 1.73% 1.27% 0.88% 0.55% 0.28% 0.05% −0.17% −0.40% −0.67% −1.00% −1.41% 8% 0.12% 0.12% 0.10% 0.07% 0.03% 0.00% −0.03% −0.06% −0.07% −0.06% −0.05% Vega 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% Convexity Vega 334 FINANCIAL RISK MANAGEMENT TABLE 11.6 Price Center Boxes of Price‐Vol Matrix Implied Volatility −2% −5 0 5 0% 2% −0.01% 0.01% 0.00% 0.00% 0.01% 1.06 units have been deliberately selected to create a portfolio with zero vega, gamma, and theta. However, as the price‐vol matrix shows, this is not the same as saying there is no options risk in the portfolio. Focus on the center five boxes in the price‐vol matrix of Table 11.5, representing the current price and implied volatility, as well as one shift up and down in price and implied volatility, as shown in Table 11.6. You can see that this is consistent with vega and gamma being zero, since vega and gamma measure the sensitivity to small changes in volatility and price. However, as you widen your view to the whole matrix, you see both convexity and volatility exposure. The convexity exposure is to a loss on downward price jumps for which the impact of the sold at‐the‐money option will outweigh the impact of the purchased option at a higher strike. The convexity impact of upward price jumps is a gain, since the effect of the purchased higher‐strike option will outweigh the effect of the sold at‐the‐money option. As prices rise, vega will be positive, reflecting the greater impact of the purchased higher‐strike option. As prices fall, vega will be negative, reflecting the greater impact of the sold lower‐strike options. Option positions that display these characteristics—acting like a bought option to some price levels, with positive vega and gains from convexity, and acting like a sold option at other price levels, with negative vega and losses from convexity—are known as risk reversals, since the direction of risk exposure reverses itself with changes in price level (for further discussion of risk reversals, see Taleb [1997, 135, 275–276]). Here are two stories that illustrate some of the characteristics of risk reversals. The first comes from the Japanese equity derivatives market in the mid‐1990s. Many Japanese banks were selling warrants on their stock that had the price‐vol profile of a risk reversal. The warrant buyer would have a positive vega and convexity at the stock price levels then prevailing, but would switch to a negative vega and convexity if stock prices were to fall significantly. Rumors in the market indicate that some trading desks purchased these warrants to provide a hedge against the negative vega and convexity exposure they had from other positions in Japanese equity derivatives, but Managing Vanilla Options Risk 335 did not adequately plan for what would happen if stock prices plummeted, causing the now negative vega and convexity on the warrant to exacerbate the overall negative vega and convexity of the desk. When Japanese stock prices did experience a sharp decline in 1996, it was accompanied by a rise in implied volatility and a decline in the liquidity of underlying stock positions, so negative vega and convexity positions resulted in large trading losses. Some reports indicate that this was one of the events that contributed to the large losses at UBS (refer to the discussion in Section 4.1.5). The second story goes back further in time to the early days of options trading. The business executive of a newly formed options business, for which I was in charge of analytics, came to me with a situation that was disturbing him. A recent series of large moves had occurred in this particular market, with large decreases in underlying prices and increases in implied volatility followed by large increases in underlying prices and decreases in implied volatility. The net effect was that prices and implied volatilities had pretty much finished up where they had started. Although the market had retained good trading liquidity throughout, the implied volatility moves were substantial enough to trigger material P&L swings. What was disturbing to the business head was that the trading book had been a loser in both the increase and decrease in implied volatility. The time period that was involved had been short enough that no significant change in the options position had taken place. So how could this pattern be explained? This trading desk did not yet have a regular price‐vol matrix, but my team was able to put one together, which quickly revealed a risk reversal pattern for the portfolio. At the price level that prevailed at the beginning of the period, the portfolio’s vega was negative, leading to losses from rising implied volatilities. At the level to which prices then fell, the portfolio’s vega was positive, leading to losses from falling implied volatilities. So far, so good. But underlying prices and implied volatilities ended where they began. In an unchanged portfolio, wouldn’t the Black‐Scholes valuation yield the same option prices at the end of the period as at the beginning of the period given that not enough time had elapsed to make a significant difference? It would be a good exercise to think this through yourself before seeing my answer. The key to understanding what happened is that the portfolio was not really unchanged since delta hedging had gone on throughout the period. Since the markets had retained liquidity throughout, this delta hedging had been smooth and no gains or losses due to price jumps had occurred. If price jumps had occurred rather than smooth delta hedging, then the portfolio would have come back to its original value. If this is not clear, follow the example in Table 11.7, which corresponds to being short one unit of a one‐year at‐the‐money call and long one unit of 336 FINANCIAL RISK MANAGEMENT TABLE 11.7 P&L Consequences of a Cycle in Prices and Volatilities Moves Volatilities up 8% (0, 0%) → (0, 8%) With Price Jumps With Smooth Delta Hedging −1.06% −1.06% Prices down 25% (0, 8%) → (−25, 8%) +0.84% [−0.22% − (−1.06%)] 0 Volatilities down 8% (−25, 8%) → (−25, 0%) −0.83% [−1.05% − (−0.22%)] Prices back up to original level (−25, 0%) → (0, 0%) Total −0.83% +1.05% 0 0 −1.89% a one‐year call at 80 percent of the current price. Assume that the following four moves take place in sequence: volatilities up 8 percent, prices down 25 percent, volatilities down 8 percent, and prices back up to the original level. Table 11.7 shows the P&L consequences, contrasting a case with price jumps and a case with smooth delta hedging. The computations for Table 11.7 can be found in the PriceVolMatrixCycle spreadsheet. This is the most extreme case in which implied volatility moves completely precede price moves. When implied volatility and price moves are mixed together, the effect is attenuated but not lost. Altogether, this constitutes another example of the maxim that delta hedging makes all options path dependent. ■ ■ Calendar spread. We are short one unit of a one‐year call option struck at‐the‐money and long one unit of a six‐month call option struck at‐ the‐money. The price‐vol matrix in Table 11.8 shows positive P&L from price jumps but negative P&L from an increase in implied volatility. This is also reflected in the positive gamma and negative vega measures for the portfolio. Shorter‐term options generally have a greater impact on sensitivity to price jumps than longer‐term options of the same size, but longer‐term options generally have greater exposure to implied volatility than shorter‐term options of the same size. Reduced risk portfolio. We are short one unit of a one‐year call option struck at 105 percent of the forward price and long 0.525 units of a one‐year call option struck at‐the‐money and 0.5 units of a one‐year call option struck at 110 percent of the forward price. The price‐vol matrix is shown in Table 11.9. These weights have been deliberately selected to make gamma and vega zero. However, unlike the call spread 337 Theta −0.007% TABLE 11.8 1.90% 1.59% 1.17% 0.80% 0.66% 0.83% 1.20% 1.65% 2.05% 2.37% −20 −15 −10 −5 0 5 10 15 20 25 2.21% 1.85% 1.43% 0.99% 0.64% 0.50% 0.62% 0.98% 1.41% 1.77% 1.99% 2.03% 1.64% 1.20% 0.78% 0.46% 0.33% 0.45% 0.78% 1.22% 1.62% 1.90% –4% 1.82% 1.41% 0.98% 0.58% 0.28% 0.17% 0.28% 0.60% 1.03% 1.46% 1.79% Price‐Vol Matrix for a Calendar Spread 2.05% −25 –6% −0.021% 2.8% 0.015% 0.27% −2.0% 1.61% 1.19% 0.76% 0.38% 0.10% 0.00% 0.11% 0.41% 0.84% 1.29% 1.66% 0% –2% Gamma 0.8% 52.8% 5.50% −0.38% −54.0% −7.58% 20.0% 0.5 100 100 Implied Volatilities Vega −0.10% –8% Delta −1.2% 20.0% 1 100 100 1 call Price BS price −2.08% Implied vol Time −1 call Spot-Vol Matrix Portfolio 2% Strike Price Price 5 call/put Spacing Volatility 5.00% Volume Discount 1.38% 0.96% 0.54% 1.15% 0.73% 0.33% −0.21% −0.01% 0.18% −0.42% −0.25% −0.07% 0.92% 0.51% 0.12% −0.49% −0.33% −0.16% −0.39% −0.13% 0.27% 0.74% 1.19% 6% −0.23% 0.05% 0.46% 0.92% 1.36% 4% −0.06% 0.23% 0.65% 1.11% 1.51% 2% 0.00% 0.00% 0.01% −0.11% −0.11% 0.69% 0.29% 0.00% −0.11% −0.09% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% Convexity Vega −0.10% −0.09% −0.08% −0.08% −0.09% −0.10% −0.09% −0.07% Vega −0.39% −0.60% −0.66% −0.56% −0.30% 0.09% 0.55% 1.03% 8% 338 Delta Vega Gamma Theta 0.00% 0.0% 0.000% TABLE 11.9 Price −1 0.5 1.0% −0.008% −2.0% 0.00% −0.02% −0.04% −0.05% −0.04% 0.00% 0.03% 0.04% 0.04% 0.01% −0.02% −0.05% −15 −10 −5 0 5 10 15 20 25 −0.03% −0.02% 0.00% 0.01% 0.01% 0.01% 0.00% −0.01% −0.01% 0.01% 0.05% −0.02% −0.02% −0.01% −0.01% −0.01% −0.01% 0.00% 0.00% 0.00% 0.00% 0.00% 0.01% 0.03% 0.07% 0% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.02% 0.06% −2% Implied Volatilities 0.18% 17.7% Price‐Vol Matrix for a Reduced Risk Portfolio 0.02% 0.02% 0.02% 0.00% −0.02% −0.03% −0.01% −0.03% −20 0.03% 0.02% −6% −4% 0.9% −0.007% 0.20% −0.38% 0.014% 28.3% −44.3% 2.04% 3.98% −5.62% 20.0% 1 110 100 call 20.0% 1 100 100 call 0.525 20.0% 1 105 100 call −25 −8% BS price 1.7% Implied vol Time 0.40% Spot‐Vol Matrix Portfolio 2% Strike Price Price 5 call/put Spacing Volatility 5.00% Volume Discount −0.01% −0.01% −0.01% 0.00% 0.00% 0.00% 0.00% 0.01% 0.02% 0.05% 0.08% 2% 0.00% −0.01% −0.01% 0.00% 0.00% 0.00% 0.01% 0.02% 0.03% 0.06% 0.09% 4% 0.00% 0.00% 0.00% 0.00% 0.00% 0.01% 0.01% 0.02% 0.04% 0.07% 0.10% 6% 0.01% 0.00% 0.00% 0.00% 0.01% 0.01% 0.02% 0.03% 0.05% 0.08% 0.11% 8% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.01% 0.01% 0.01% Vega 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% Convexity Vega Managing Vanilla Options Risk 339 case, the zero gamma and vega are reflected throughout the price‐vol matrix by low exposures at all combinations of price jump and volatility shift. This demonstrates the ability to achieve greater risk reduction by using positions that are symmetrical in strike price. Tables 11.10 and 11.11 show how this position evolves through time. We can see that at the end of 0.5 years (Table 11.10), there is still not much risk exposure, but at the end of 0.9 years, with only 0.1 year left until option expiration (Table 11.11), there is some convexity, with gains if prices jump upward and losses if prices jump downward. This shows that even a hedge of options against options that works very well at first cannot be maintained as a purely static hedge. We have already explored the implications of this for options risk management using Monte Carlo simulation in Section 11.3. The price‐vol matrix has the great advantage of looking at precise sensitivity to many different values of two variables, but this carries the disadvantage of only being able to consider two variables. This has two consequences: the choice of which two variables to look at is an important one, and the price‐vol matrix needs to be supplemented with risk measures that go beyond these two variables. The selection of the best variables to use in the price‐vol matrix can be based on economic insight or on statistical techniques, such as principal component analysis. On the side of asset prices, one question is whether to assume a parallel shift in forward prices. This is equivalent to assuming zero correlation between changes in the underlying asset price and changes in discount curves. Another question is whether to assume constant spreads between different variants of the asset, such as different grades for a physical commodity and different individual stocks relative to a stock market index. For volatilities, the question is whether to assume parallel changes in the volatility surface or whether to assume a statistical relationship based on historical experience. Looking more closely at the issue of whether to assume a parallel shift in the volatility surface, let’s break this down into a time‐to‐expiry component and a strike component. With regard to time to expiry, the first principal component of changes in volatility surfaces has less tendency to be flat than the first principal component of changes in interest rate curves. Longer‐term volatilities often tend to move substantially less than shorter‐term ones. Although a time‐differentiated shift conveys less immediate intuitive meaning in discussions with senior management than a flat 1 percent shift, the increase in likelihood may outweigh the communications disadvantage. A possible compromise that is reasonably easy to express and often reasonably close to historical experience is a proportional rather than an absolute shift. So if one‐year volatilities are currently 20 percent and five‐year volatilities 340 Delta Vega Gamma Theta −0.01% −0.1% 0.000% −1 0.5 0.02% −0.07% −0.16% −0.23% −0.07% −0.08% −0.02% 0.07% 0.10% 0.04% −0.06% −0.18% −0.26% −15 −10 −5 0 5 10 15 20 25 −0.21% −0.15% −0.20% −0.14% −0.08% −0.02% −0.01% −0.07% 0.01% 0.01% −0.01% −0.02% −0.01% 0.03% 0.11% −2% −0.19% −0.14% −0.08% −0.04% −0.18% −0.13% −0.09% −0.05% −0.02% −0.01% −0.17% −0.13% −0.09% −0.06% −0.03% −0.02% −0.01% 0.00% 0.01% 0.03% 0.07% 0.13% 4% −0.01% 0.00% 0.02% 0.06% 0.13% 2% −0.01% −0.01% −0.01% 0.00% 0.05% 0.12% 0% Implied Volatilities 0.03% 0.03% −0.01% −0.04% −0.03% 0.02% 0.11% −4% −0.16% −0.13% −0.10% −0.07% −0.04% −0.02% 0.00% 0.01% 0.04% 0.08% 0.14% 6% −0.16% −0.13% −0.10% −0.07% −0.04% −0.02% 0.00% 0.02% 0.05% 0.09% 0.15% 8% Price‐Vol Matrix for the Reduced Risk Portfolio of Table 11.9 After 0.5 Years Have Elapsed 0.06% 0.04% −0.01% −0.05% −0.05% 0.01% 0.00% −20 0.10% 0.10% −6% 1.2% −0.009% 0.020% −0.011% 0.12% 13.7% 1.08% 20.0% 0.5 110 100 call 1.5% −2.7% 0.14% 27.7% −39.2% −0.27% 2.89% −3.53% 0.5 100 100 call 0.525 20.0% 0.5 105 100 call 20.0% −25 TABLE 11.10 Price −8% BS price 2.2% Implied vol Time 0.44% Spot-Vol Matrix Portfolio 2% Strike Price Price 5 call/put Spacing Volatility 5.00% Volume Discount 0.01% 0.00% 0.00% −0.01% −0.01% −0.01% 0.00% 0.00% 0.01% 0.01% 0.00% Vega 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% Convexity Vega 341 Delta Vega Gamma Theta −0.01% −0.4% 0.003% −1 26.9% −23.0% 0.08% −0.47% −1.14% −1.55% −1.83% 0.51% 0.15% −0.12% 0.03% 0.16% −0.47% −1.18% −1.56% −1.83% −15 −10 −5 0 5 10 15 20 25 −1.82% −1.53% −1.10% −0.48% 0.00% 0.02% −0.05% 0.16% 0.51% 0.88% 1.25% −4% −1.81% −1.51% −1.07% −1.80% −1.48% −1.05% −0.52% −0.11% −0.06% −0.50% 0.00% 0.01% 0.19% 0.52% 0.88% 1.25% 0% 0.01% −0.02% 0.18% 0.51% 0.88% 1.25% −2% Implied Volatilities −1.79% −1.46% −1.03% −0.54% −1.78% −1.44% −1.02% −0.56% −0.19% −0.03% −0.01% −0.15% 0.05% 0.22% 0.53% 0.88% 1.25% 4% 0.03% 0.21% 0.52% 0.88% 1.25% 2% −1.77% −1.42% −1.01% −0.57% −0.23% −0.04% 0.06% 0.24% 0.53% 0.89% 1.25% 6% −1.75% −1.41% −1.01% −0.59% −0.26% −0.06% 0.07% 0.26% 0.54% 0.89% 1.25% 8% Price‐Vol Matrix for the Reduced Risk Portfolio of Table 11.9 After 0.9 Years Have Elapsed 0.03% −0.08% 0.15% 0.51% 0.88% 0.88% −20 1.25% 1.25% −6% 1.1% −0.008% 0.02% 3.5% 0.10% 20.0% 0.1 110 100 3.3% −4.8% 0.5 call 0.037% −0.026% 0.07% 1.32% −0.81% −0.10% 20.0% 0.1 100 100 call 0.525 20.0% 0.1 105 100 call −25 TABLE 11.11 Price −8% BS price 7.4% Implied vol Time 0.60% Spot-Vol Matrix Portfolio 2% Strike Price Price 5 call/put Spacing Volatility 5.00% Volume Discount 0.01% 0.01% 0.01% −0.01% −0.02% −0.01% 0.01% 0.01% 0.00% 0.00% 0.00% Vega 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% Convexity Vega 342 FINANCIAL RISK MANAGEMENT are 15 percent, a 5 percent proportional shift would move the one‐year volatility up 1 to 21 percent and the five‐year volatility up 0.75 to 15.75 percent. The PriceVolMatrix spreadsheet allows the user specification of either flat or proportional shifts. With respect to the strike component, a frequently used alternative to a flat shift by instrument is a flat shift by delta. For example, assume that an at‐the‐money option currently has a 20 percent implied volatility and an in‐the‐money option with a delta of 75 percent currently has a 19 percent implied volatility, and assume that we are dealing with a currently at‐the‐ money option. Then a volatility shift of down 2 percent combined with a price jump in the underlying asset that makes this option in‐the‐money with a 75 percent delta results in an implied volatility of 20% − 2% = 18% if we are assuming a flat shift by instrument. It results in a 19% − 2% = 17% implied volatility if we are assuming a flat shift by delta. The PriceVolMatrix spreadsheet allows the user specification of either flat shift by instrument or flat shift by delta. The driving force behind the use of a flat delta shift is that the factors that generate the shape of the volatility surface by the strike, such as stochastic volatility and the structure of jumps, tend to remain static across changes in the underlying price level. We discuss these factors in Section 11.6.2. Taleb (1997, 138–142) provides a detailed exposition of a flat delta shift methodology and its consequences for hedging. Derman (1999) contrasts flat instrument shifts with flat delta shifts (“sticky‐strike” versus “sticky‐delta” in Derman’s terminology) along with a third possibility, “sticky‐implied‐tree.” Derman presents empirical evidence that differing market environments over time can result in a change in which shift patterns provide the greatest explanatory power. No matter what selections are made for the price‐vol matrix variables, there is clearly enough residual risk to require traders to also look at more detailed risk reports as supplements to price‐vol matrices. Certainly, this will include exposure to changes in the shape of the volatility surface with respect to both time and strikes. The PriceVolMatrix spreadsheet includes a calculation of exposure to changes in the volatility surface. These more detailed reports usually focus only on the impact of small one‐at‐a‐time changes, although a particularly significant residual risk might justify a price‐vol matrix of its own. For example, an equity options trading desk might want to look at an overall price‐vol matrix that considers parallel shifts in all stock market indexes as well as price‐vol matrices for each individual country’s stock index, but would probably want only a simple delta and vega measure to reflect the exposure to each individual stock traded. Senior management will want to see much less detail than the trading desk regarding options. The primary concern of senior management is Managing Vanilla Options Risk 343 making sure that they are comfortable with large macro positions that may be an accumulation of the holdings of many trading desks. As such, the most important measure for senior management is outright exposure to spot positions (for example, JPY/USD FX, S&P index, and gold) or to forward positions (for example, exposure to a parallel shift in the USD interest rate curve). Since options desks hold delta‐equivalent positions in these spot and forward markets, including these positions in reports of the total spot exposure of the firm is necessary in order to ensure an accurate summary. So senior management will generally just be interested in a single outright position number for each product, along with some measure of vega. For many options positions, the delta will fit the need for an outright position measure. Control of convexity risk around this delta is then left to the trading desk level, probably prescribed by limits on convexity. However, the positions of some complex trading books may not be at all accurately represented by the delta. If a book will gain $100 million for the next 1‐point rise in the S&P but lose $2 million for each point rise after that, representing the position by a +$100 million per point delta will be totally misleading. For senior management purposes, the delta needs to be defined not mathematically, as the instantaneous derivative, but economically, as a finite difference over a selected economically meaningful price movement (a one‐standard‐deviation daily price move might be a reasonable choice). Limit‐setting detail for options books lies somewhere between the level needed for trading desk control and that needed for senior management. Some form of limits on price‐vol matrix positions is desirable, but separate limits for each matrix box would be overdetailed, whereas a single limit that no matrix box could exceed would be too broad. A limit set high enough to accommodate really unlikely combinations would be too liberal a limit for combinations close to the matrix center. A reasonable compromise is differentiated limits by groups of matrix boxes, where a similar likelihood of outcomes determines grouping. Limits on exposure to changes in the shape of the volatility surface can often be best expressed in terms of a few parameters that determine the shape. For details on possible parameters, see the discussion in Section 11.6.2. The management of options risk is an inherently dynamic process. Unlike spot or forward risk, you can rarely just put on a hedge once and for all; you must constantly make adjustments. So options traders need measures to show them how their P&L and positions should change as a result of the passage of time or changes in prices. This enables them to prepare for the trading actions they will need to take and serves as a check against actual changes in P&L and positions to highlight anything that is happening that they don’t understand. The best‐known measures of this type are theta (the change in option values with time) and gamma (the change in delta with a 344 FINANCIAL RISK MANAGEMENT change in price). However, many other examples are available: for instance, bleed d (see Taleb 1997, 191–199) and Ddeltadvoll (Taleb 1997, 200–201). By contrast, corporate risk managers are rarely interested in such measures. Theta cannot be a direct measure of risk since clearly you are not uncertain as to whether time will pass. It does measure the possibility of gain or loss if implied volatility fails to be realized over a given time period, but the same risk can be captured in a more comprehensive way by a time‐bucketed vega measure. Gamma is of interest only to the extent that it can be used to compute convexity, which is a genuine P&L exposure, but gamma is a reliable indicator of convexity only for very simple portfolios. In general, corporate risk managers expect that trading desk heads will be able to deal with the operational issues of evolving positions. The only exceptions might be changes so large as to make liquidity questionable, which might require limits to be set. 11.5 DELTA HEDGING In the presence of transaction costs, it is necessary to use optimization to determine a delta hedging strategy. A trade‐off exists between achieving a lower standard deviation of results utilizing more frequent hedging, and achieving a higher expected return utilizing less frequent hedging leading to lower transaction costs. Whaley and Wilmott (1994) have shown that the efficient frontier for this problem consists of hedging policies with the following characteristics: ■ ■ ■ Hedges will be triggered not by time intervals, but by the distance that the current delta hedge ratio differs from the theoretical delta hedge ratio required by the Black‐Scholes formula. If transaction costs are only a function of the number of hedge transactions and not the size of the hedge transactions, then whenever a hedge transaction is triggered, the amount will be exactly enough to bring the hedge ratio in line with the desired theoretical ratio. Since the transaction cost is the same no matter how large the amount, you should go to the hedge ratio you would use in the absence of transaction costs. If transaction costs are only a function of the size of the hedge transaction, then whenever a hedge transaction is triggered, the amount of the transaction is only large enough to bring the difference between the actual and theoretical hedge ratios down to the trigger point. Since you don’t care how many transactions you need to use, only the size of transactions, it makes sense that you will stay as close as possible to the point at which hedge inaccuracy exactly balances between the desire for low standard deviation of results and low transaction costs. Managing Vanilla Options Risk ■ 345 If transaction costs are a function of both the number and size of hedge transactions, then the optimal rule will be a combination of these two cases, with an outer trigger distance between current and theoretical delta that institutes a trade to bring the difference down to an inner trigger distance. Target delta hedges are determined by the Black‐Scholes formula as N(d1), whered1 [ ( / k) k) σ 2T / ] / σ T . What value of σ, the volatility of the underlying asset, should be used to determine this target hedge? Options should be valued at the implied volatility that corresponds to the market price at which the position could be exited, but this does not provide any reason for using this implied volatility to determine delta hedges of positions that are not exited. Given that any misestimation of true volatility while determining the hedge will result in unintended proprietary positions in the underlying asset, as per our discussion in Section 11.3, it is best to give traders reasonable latitude to make their best estimate of future volatility as input to the target hedge. This brings us to the suggested solution we promised to the question in Section 11.2. What causes the large losses from the nasty path? It is caused by the dramatic difference between actual realized volatility and implied volatility. You will see in the NastyPath spreadsheet that the option was priced at a 7 percent implied volatility, which was also used in creating the delta hedge. However, the actual price moves of 0.13 a day correspond to a realized volatility of 2 percent. Had the trader been able to foresee this and form the delta hedges based on a 2 percent volatility, P&L on the trade would have been close to 0 (try this out in the spreadsheet). Continuing the theme from Section 11.3, concerning what actions to take if a trader believes the underlying price is mean reverting, simulations similar to those reported in Table 11.3 indicate that gains will result from delta hedges based on overestimates of actual realized volatility. If underlying prices are trending (either up or down) rather than mean reverting, then gains will result from delta hedges based on an underestimate of the actual realized volatility. So traders should consider biasing their volatility estimates if they have a view on mean reversion. To get an intuitive understanding of this result, consider what happens if you overestimate volatility. The higher volatility in the denominator of the formula for d1 will cause the target delta to move less as price movements result in the option moving into or out of the money. If price moves tend to be followed by moves in the opposite direction, as they will be if the price process is mean reverting, then the difference between actual delta and theoretical delta will be in the right direction to create positive P&L. 346 11.6 FINANCIAL RISK MANAGEMENT BUILDING A VOLATILITY SURFACE Building a volatility surface for pricing European options is similar to building a discount curve, but it operates in two dimensions rather than one, since volatilities will vary by strike as well as by time. However, the general principle is the same: Build a surface that balances the fitting of known options prices with a smoothness criterion. The smoothness criterion is designed to minimize the risk of loss from hedging options for which market prices are not known with options for which prices are known. To build the surface in both dimensions simultaneously requires a stochastic volatility model to which you can fit parameters (for example, the Heston model—see Heston 1993). The more common approach is to build a volatility curve for at‐the‐money strikes by time period and separately build a volatility curve for a few selected time periods by strike. Arbitrary combinations of time and strike can then be interpolated from already determined points. We will look in turn at the issues of interpolating between time periods, interpolating between strikes, and extrapolating beyond the longest liquid time period. 11.6.1 Interpolating between Time Periods We have a problem that’s extremely similar to the one we faced for discount curves. We have a set of fitting conditions, wanting to choose underlying discount prices (implied volatilities), so that when they’re plugged into pricing formulas, they come out with bond prices (option prices) that closely match those observed in the market, and a set of smoothness conditions, wanting to choose discount prices (implied volatilities) that lead to maximum smoothness of forward interest rates (forward volatilities) across periods. The forward volatility, the amount of volatility expected to take place in some reasonably small time period in the future, is a natural analogy to the forward rate. With forward rates, we discussed whether to have an additional set of constraints stating that all forward rates must be nonnegative and examined economic arguments for and against this (refer to Section 10.3.2). With forward volatilities, there isn’t any doubt—a negative standard deviation is not a mathematical possibility, so the constraints are necessary. We can set up an optimization to solve for forward volatilities in a completely analogous manner to the optimization we set up to solve for forward rates, with different solutions corresponding to different trade‐offs between the tightness of the fitting constraints and tightness of the smoothness constraints and different weights on different fitting constraints based on the liquidity of the price quotes. (Note that it is a more viable possibility Managing Vanilla Options Risk 347 with options than with interest rates to just find forwards that exactly fit all available market prices and then interpolate between the forwards. Unlike bonds and swaps, options have no intermediate payments to require a bootstrap. However, optimization still might be desirable as a way of trading off between fitting and smoothness objectives.) When fitting forward interest rates, we had to preprocess to adjust for the lack of smoothness that we were anticipating based on our economic theories, such as turn‐of‐the‐quarter effects (see Section 10.3.4). In the same way, forward volatilities need preprocessing. Generally, the opinions of options traders regarding the patterns of forward volatility tend to be much more strongly held than the opinions of interest rate traders regarding forward rates. Opinions on forward volatility center on issues of the flow of information into the markets that will cause price fluctuations. If we look at daily forward volatilities (and traders of shorter‐term options often do work at this level of detail), you might find a trader anticipating nearly zero volatility on weekends and holidays (markets are closed so no new prices can be observed), higher volatility on Mondays and days after holidays than on other weekdays (governments sometimes like to make surprise announcements when markets are closed), lower than normal volatility on days when most traders can be expected to be on vacation or leaving work early (such as the day before a three‐day weekend), and higher than normal volatility on a day when a key economic statistic is scheduled to be announced. For more examples, see Taleb (1997, 98) and Burghardt and Hanweck (1993). The website for this book has two spreadsheets to illustrate fitting a forward volatility curve to observed options prices. The first, VolCurve, can be used for all European options other than interest rate caps and floors, and emphasizes the adjustment for anticipated volatility patterns. The second, CapFit, is designed for use only for interest rate caps and floors, which are packages of individual options (known as caplets and floorlets, respectively). Since liquid prices are generally available only for the options packages and not for the underlying options, an optimization is needed to fit the observed prices of packages with as smooth a forward volatility curve as possible. 11.6.2 Interpolating between Strikes—Smile and Skew Now let’s turn to building a volatility curve by strike for a given time period. Market prices will be available for certain strikes that we will want to fit. Which variable should play the corresponding role to forward interest rates and forward volatilities as the one for which we try to achieve smoothness? A natural choice is the risk‐neutral probability that the underlying variable finishes in a range between two prices. If these ranges are chosen small 348 FINANCIAL RISK MANAGEMENT enough, options at all strikes can be priced to as close a precision as you want based on such probabilities. If S is the strike and pi is the risk‐neutral probability that the underlying will finish between price Pi and price Pi+1, the option price must be bounded by ∑ max( i , 0)pi , )pi from below, and bounded by ∑ max( i i i from above. Like forward volatilities, probabilities must be constrained to be nonnegative. Using this formula allows translation among cumulative probability, probability frequency, and implied volatility by strike as alternative, mutually translatable ways of describing a probability distribution, in much the same way that par rate, zero coupon rate, forward rate, and discount price are alternatives for describing the discount curve. See the VolSurfaceStrike spreadsheet for an illustration of this principle. Jackwerth and Rubinstein (1996) illustrate an optimization setup to derive probability distributions based on a trade‐off between the tightness of fitting constraints and smoothness constraints. When choosing a smoothness criterion, an alternative to just looking at how smooth the changes in probability levels are is to look at how closely the probabilities fit a distribution selected on theoretical grounds (for example, normal or lognormal) as the most likely prior distribution (prior, that is, to any knowledge of the actual options prices). This use of prior distribution ties closely to Bayesian statistical methods. In Section III.A of their paper, Jackwerth and Rubinstein explore several such smoothness criteria. A fundamental problem often encountered when trying to derive volatility curves by strike is the relative paucity of market observations available by strike. It is not at all uncommon to find markets in which options prices are available for only three or four different strike levels at a given time period. In such circumstances, a smoothness criterion that does not utilize a prior distribution is of little use—you at least need to restrict your choice to some family of possible candidate distributions on theoretical grounds. Of course, any such choice is a model and should be analyzed for the degree of mispricing possible if the model is wrong by considering how different the volatility curve would be if another plausible model were chosen. Reserves and limits against model error should be considered. A good discussion of candidate distributions and the theoretical basis for selecting between them can be found in Hull (2012, Sections 26.1–26.3). Let us first state some general facts about the shape of volatility surfaces observed in the markets; these comments can be compared with those in Hull (2012, Sections 19.2 and 19.3) and Rebonato (2004, Chapter 7). In this discussion, we use the term smile to refer to a pattern of volatility by strike where volatility rises as strikes move away from at‐the‐money in the Managing Vanilla Options Risk 349 direction of either into or out of the money. We use skew to refer to a pattern of volatility by strike in which volatility either decreases or increases with increasing strike levels. So skew is primarily a linear relationship and smile is primarily a quadratic relationship. (Market practice from firm to firm, and even desk to desk within a firm, may differ in nomenclature. Sometimes skew is used to cover all aspects of volatility surface shape, and sometimes smile is used to cover all aspects of volatility shape.) Using these definitions, the observed patterns are: ■ ■ ■ ■ Smiles tend to appear in all options markets. Equity options markets almost always show a pronounced skew, with volatility decreasing with increasing strikes. The combination of this skew with the smile produces a pattern that can be described as a sharp skew at strikes below at‐the‐money and relatively flat volatilities at strikes above at‐the‐money. No general skew pattern exists in markets for FX options between strong currencies (for example, between the dollar, euro, yen, sterling, and Swiss franc). However, there does tend to be a strong skew pattern (volatility decreases with increased strikes) for FX options between a strong currency and a weaker currency such as an emerging market currency. Skew patterns in interest rate options markets tend to vary by currency, with the strongest patterns of volatilities decreasing with increasing strikes appearing for currencies with low interest rate levels, particularly in yen. What explanations have been offered for these observed patterns? ■ The prevalence of volatility smiles can be explained in two different ways: stochastic volatility and jump diffusion. Stochastic volatility utilizes a probability distribution for the volatility that determines the probability distribution of underlying prices, whereas jump diffusion assumes that some price uncertainty is expressed through price jumps as opposed to a smooth random walk. Both assumptions result in a distribution of final prices with fatter tails than the lognormal distribution used by Black‐Scholes. Fatter‐tailed distributions have little effect on options at close to at‐the‐money strikes, which are primarily affected by the center of the distribution; however, they have greater effects the more an option is in‐the‐money or out‐of‐the‐money, since these options are primarily affected by the size of the tail. The pricing formula for options using either stochastic volatility or jump diffusion (see the equations in Hull 2012, Sections 26.1 and 350 ■ ■ FINANCIAL RISK MANAGEMENT 26.2) consists of averages of option prices using the Black‐Scholes formula across a range of volatilities. The difference between the two models is the probability weight used in averaging across these volatilities. Stochastic volatility results in a more pronounced smile as the time to option expiry increases, whereas jump diffusion results in a more pronounced smile as the time to option expiry decreases. It may be necessary to combine the two to obtain actual smile patterns observed in market options prices. See Matytsin (1999). The Black‐Scholes model assumes a lognormal distribution of the underlying asset price. If the market is assuming a normal, rather than lognormal, price distribution, this will evidence itself as higher implied volatilities for lower‐strike options and lower implied volatilities for higher‐strike options when implied volatilities are computed using the Black‐Scholes formula. So if the market is assuming that price changes are independent of market level rather than proportional to market level, implying normal rather than lognormal price distributions, this will lead to a skew with volatilities decreasing with increasing strikes. If the market is assuming a distribution intermediate between normal and lognormal, this skew pattern will still exist, but it will be less pronounced. Historical evidence shows support for interest rate movements that are sometimes closer to being independent of the rate level and other times closer to being proportional to the rate level. The skew for implied volatilities of interest rate options is generally believed to be driven primarily by the expectation that rate movements are not completely proportional to the rate level, with the expectation in low‐rate environments that rate movements are close to independent of the rate level. The skew pattern in equity markets has sometimes been explained as the outcome of asymmetry of the value of investment in a corporation, which can suddenly collapse as a company approaches the bankruptcy point. Hull (2000, Section 17.7) discusses three alternative models based on this explanation—the compound option model, the displaced diffusion model, and the constant elasticity of variance model. A more general explanation of skew patterns can be found in analyzing the degree of asymmetry in the structure of a particular market. For a thorough exposition of this viewpoint, see Taleb (1997, 245–252), on which much of my discussion here is based. This asymmetry can be described in two complementary ways: one that focuses on investor behavior and the other that focuses on price behavior. From an investor behavior viewpoint, in some markets, investment has a structural bias toward one side of the market. Equity markets are a good Managing Vanilla Options Risk 351 example. There are far more investors long equity investments than there are investors who have shorted the market; hence, more investors are seeking protection from stock prices falling than are seeking protection from stock prices rising. The reason is that corporate issuance of stock is a major source of supply, and corporations are not seeking protection against their stock rising; in fact, they welcome it. So you expect to see greater demand to buy puts on stock at strikes below the current market level, sought by investors protecting their long equity positions, than the demand for calls on stock at strikes above the current market level sought by short sellers to protect their short equity positions. This imbalance in demand drives up implied volatilities on low‐strike options relative to high‐strike options. The complementary view from a price behavior viewpoint is that stock market crashes, in which large downward jumps occur in stock prices, are far more common than large upward jumps in stock prices. This can be seen as a consequence of the imbalance in investors who are long stocks relative to those who are short stocks. Falling prices can trigger a selling panic by investors faced with large losses forced to exit leveraged long positions supported by borrowings. There are fewer short sellers and leveraged short positions to cause a panic reaction when prices are rising. A bias toward downward jumps over upward jumps leads to a skew in the distribution of probabilities of price movements that will translate into higher implied volatilities at lower strikes. In addition, the anticipation of possible stock market crashes will exacerbate the demand for crash protection through puts at lower strikes. A similar structural analysis can be constructed for FX markets for an emerging market currency versus a strong currency. These exchange rates are often maintained at artificially high levels by governments defending the value of the emerging market currency through purchases of the currency, high interest rates, or currency controls. When breaks in the FX rate come, they tend to be large downward jumps in the value of the emerging market currency. There is no similar possibility of upward jumps. This price behavior directly leads to a probability distribution that translates to higher implied volatilities at lower strikes (lower in terms of the value of the emerging market currency). Indirectly, this price behavior encourages holders of the emerging market currency to buy puts at lower strikes, bidding up the implied volatility at these strikes. Other markets generally tend toward a more symmetrical structure. Exchange rates between two strong currencies are usually freer floating with less bottled‐up pressure. Thus, no bias exists toward large upward jumps or large downward jumps. Most interest rate markets and commodity markets tend to be roughly evenly divided between longs and shorts—investors who would benefit from upward movement and those who would benefit 352 FINANCIAL RISK MANAGEMENT from downward movement. However, some particular asymmetries can be observed—for example, the large demand by U.S. mortgage investors for protection against falling interest rates leading to accelerated prepayments or a temporary imbalance of the suppliers of a commodity seeking put protection against falling prices relative to the consumers of the commodity seeking call protection against rising prices. The VolSurfaceStrike spreadsheet illustrates both ways in which a probability distribution can be fit to a set of option prices at different strikes. With input on prices at a number of different strikes, it trades off the smoothness of the probability distribution and closeness of price fit. With input on prices at only a few strikes, it fits two parameters: one representing standard deviation of volatility and one representing the degree of proportional versus absolute price change to assume. 11.6.3 Extrapolating Based on Time Period When we were looking at forward risk, we saw how to create valuation and reserves for a forward that had a longer tenor than any liquid instrument (see Section 10.2.2). The technique was to assume you were going to hedge the longer‐term forward with a liquid shorter‐term forward and later roll the shorter‐term forward into a longer‐term forward. The expected cost of the roll needs to be added into the initial cost of the hedge to obtain a valuation, and a reserve can be based on the historical standard deviation of the roll cost. A similar approach suggests itself for valuing and reserving for long‐ term options that have a longer tenor than any liquid option. For example, if you want to create a 10‐year option in a market that has liquid quotes only out to seven years, you could begin by hedging with a seven‐year option and, at the end of five years, roll out of what will then be a two‐year option into the five‐year option you need to exactly match your actual position. Expected differences in implied volatility between five‐ and two‐year options determine expected roll costs. Reserves can be based on the historical standard deviation of differences in two‐ and five‐year implied volatilities. However, options are more complicated because they depend on strike level as well as the time to expiry. The price‐vol matrix in Table 11.12 shows that a ratio of seven‐year options to 10‐year options selected so as to minimize roll‐cost uncertainty when the prices are at 100 leaves large roll‐cost uncertainty when prices are above or below 100. To minimize roll‐cost uncertainty over a wide range of prices, you need to hedge with a package of options that differ by both tenor and strike. The price‐vol matrix in Table 11.13 shows the impact of selecting a hedge from a set of six‐ and seven‐year options at various strike levels, using the OptionRoll spreadsheet to select weightings of these options that will achieve 353 Delta Vega Gamma Theta 0.00% 1.0% −0.008% 2.89% 1.70% 0.78% 0.21% 0.02% 0.20% 0.71% 1.50% 2.48% 3.62% −20 −15 −10 −5 0 5 10 15 20 25 −1 74.5% 0.68% 1.9% −0.68% −0.9% 3.32% 2.26% 1.35% 0.64% 0.17% 0.01% 0.18% 0.70% 1.54% 2.66% 3.97% −6% 3.06% 2.07% 1.22% 0.57% 0.15% 0.01% 0.16% 0.63% 1.40% 2.45% 3.71% −4% 0.005% −0.013% 13.64% −58.8% 20.0% 2 100 100 call 1.34 −13.78% 20.0% 5 100 100 call 2.83% 1.90% 1.12% 0.52% 0.14% 0.00% 0.14% 0.57% 1.29% 2.27% 3.47% −2% 2.64% 1.76% 1.03% 0.48% 0.12% 0.00% 0.13% 0.53% 1.19% 2.11% 3.25% 0% Implied Volatilities 2.47% 1.65% 0.96% 0.44% 0.12% 0.00% 0.12% 0.49% 1.11% 1.97% 3.06% 2% 2.32% 1.54% 0.90% 0.42% 0.11% 0.00% 0.12% 0.46% 1.04% 1.86% 2.89% 4% Hedge of a 10‐Year Option with a Seven‐Year Option after Five Years 4.26% −25 TABLE 11.12 Price −8% BS price 15.7% Implied vol Time Strike −0.14% 2% Spot-Vol Matrix Portfolio Volatility Price Price call/put Spacing 5 5.00% Volume Discount 2.19% 1.46% 0.85% 0.40% 0.11% 0.01% 0.12% 0.44% 0.98% 1.75% 2.74% 6% 2.08% 1.39% 0.82% 0.39% 0.11% 0.02% 0.12% 0.42% 0.94% 1.67% 2.60% 8% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% −0.01% −0.02% −0.04% −0.06% −0.09% 0.00% 0.00% 0.00% 0.00% 0.00% Convexity Vega 0.00% −0.01% −0.02% −0.04% −0.07% −0.10% Vega 354 Vega −0.68% 0.005% 0.04% 0.19% 20 25 −0.06% −0.03% −0.05% −0.01% −0.04% −0.05% −0.03% −0.01% 0.00% −0.01% −0.04% −0.07% −0.07% 0.01% 0% −0.06% 0.04% 0.4465 0.6% 0.20% 13.7% 1.95% 20.0% 2 120 100 call 2% −0.04% −0.05% −0.03% −0.01% 0.01% 0.02% 0.00% −0.03% −0.06% −0.07% 0.8% 0.14% 33.7% 5.93% 20.0% 1 90 100 call 0.4588 4% −0.04% −0.03% −0.01% 0.02% 0.04% 0.04% 0.02% −0.01% −0.05% −0.07% 6% 8.09% −0.03% −0.01% 0.02% 0.05% 0.07% 0.07% 0.05% 0.01% −0.03% −0.06% −0.06% −3.91% –2.3471 3.0% 0.57% 56.5% 6.53% 20.0% 1 110 100 call 1.6000 8% 0.00% 0.02% 0.06% 0.08% 0.10% 0.10% 0.08% 0.04% −0.01% −0.05% −0.01% 0.00% 0.01% 0.01% 0.01% 0.01% 0.00% 0.00% 0.00% −0.00% −0.02% −8.22% 0.009% −1.3% −0.24% −18.2% −1.78% 20.0% 1 120 100 call –0.8726 0.00% 0.00% 0.00% 0.00% −0.01% −0.01% −0.01% 0.00% 0.00% 0.00% 0.00% Vega Convexity 18.60% Vega −33.65% 0.034% −0.022% −4.7% −0.89% −126.7% −17.78% 20.0% 1 100 100 call −0.06% 0.001% −0.006% −0.2% −0.03% −16.1% −3.65% 20.0% 1 80 100 call –0.1811 −0.05% 4.49% −0.02% 8.16% −0.06% Hedge to Roll Over into a 10‐Year Option 0.10% 0.00% −0.04% −0.02% 15 −0.04% −0.04% −0.01% −0.03% 10 0.00% 0.00% 0.07% 5 0.04% 0.03% −0.05% −0.08% −0.06% 0.05% −2% −0.02% 0.12% 0 −0.04% 0.9% 0.34% 28.4% 4.54% 20.0% 2 110 100 Implied Volatilities −9.75% −0.01% 0.10% −5 −0.02% −0.07% 0.6714 call 0.006% −0.006% −0.004% −0.9% −0.33% −36.9% −6.74% 20.0% 2 100 100 0.01% 0.00% −10 −0.07% −0.05% 0.10% −4% 46.04% –0.6627 call 0.00% −0.08% −15 −0.03% 0.17% −6% 3.2% 1.16% 180.5% 38.49% 20.0% 2 90 100 call 0.01% 0.24% −0.03% −20 −8% 5 2.5922 0.004% −0.022% −0.6% −0.23% −56.9% Current −15.05% −14.78% Theta −25 0.031% 0.001% Gamma −0.9% 0.01% −0.1% 20.0% 2 80 100 call –0.6906 Years BS price −13.78% −14.42% Delta −58.8% TABLE 11.13 Price 5 100 Implied vol 20.0% Time Strike 100 call −1 −0.8% −0.72% 2% Spot-Vol Matrix Portfolio Volatility Price Price call/put Spacing 5 5.00% Volume Discount Managing Vanilla Options Risk 355 minimal roll‐cost uncertainty in five years. This example only accounts for roll‐cost uncertainty due to shifts in volatility level; a more complete treatment would include shifts in the shape of the volatility surface. Expected roll costs and standard deviations of roll costs must now be computed relative to the weighted average of implied volatilities of the hedge package. 11.7 SUMMARY By way of summary, let us see how the paradigm for managing vanilla options risk deals with the criticisms of the Black‐Scholes analysis that have been offered. Compare the analysis here to Taleb (1997, 110–113). ■ ■ ■ ■ Black‐Scholes unrealistically assumes a constant risk‐free interest rate and drift rate of the forward. The way we have set up our Black‐Scholes model, directly incorporating rate and drift volatility into the volatility of the forward, shows that this criticism is not a serious one. Black‐Scholes assumes that asset prices are lognormally distributed. This has long ceased to be true in trading practice. With traders valuing positions at each strike at different market‐observed volatilities, any probability distribution believed by the marketplace can be accommodated. In part 2 of Exercise 11.2 you are asked to examine the success of hedging options at one strike with those at another strike, using a Monte Carlo simulation that does not assume asset prices to be lognormally distributed. You will find relatively small uncertainty of hedging results. Black‐Scholes assumes that hedging in the underlying asset can take place continuously and without transaction costs. These assumptions are closely linked since the presence of transaction costs will certainly force hedging to be less frequent, even if more frequent hedging is theoretically possible. Our Monte Carlo simulations have shown that, with the use of options to hedge other options, the resulting positions can be delta hedged at discrete times, resulting in relatively small uncertainty of hedging results and relatively low transaction costs. Any uncertainty and transaction costs that remain will contribute to wider bid‐ask spreads for options. Black‐Scholes assumes that underlying asset prices will follow a Brownian motion with no sudden jumps. In practice, sudden jumps do occur and these are unhedgeable other than by offsetting options positions. The price‐vol matrix reports exposure to price jumps. In part 1 of Exercise 11.2, you are asked to examine the success of hedging options at one strike with those at another strike, using a Monte Carlo simulation that assumes price jumps will take place. You will find relatively small uncertainty of hedging results. 356 FINANCIAL RISK MANAGEMENT ■ Black‐Scholes assumes that volatility is constant. This is obviously false. The implications of stochastic volatility for the standard deviation of hedging results have been noted. The price‐vol matrix reports exposure to changes in volatility, and positions that have small exposure as measured by the price‐vol matrix have been shown, using Monte Carlo simulation, to have a relatively small uncertainty of hedging results. Black‐Scholes assumes that volatility is known. This is also obviously false. Our Monte Carlo simulations were carried out under the assumption that actual volatility was not known when setting hedge ratios, and the resulting uncertainty of hedging results is small. ■ EXERCISES 11.1 Options Portfolio Risk Measures Start with a portfolio consisting of less liquid options as follows: Volume 1 Call/put Price Strike Time to expiry Implied volatility Call 100 84 1.3 25.9% −1 −1 Call 100 107 0.7 24.4% Call 100 114 1.7 23.0% 1. Calculate the risk exposure of this portfolio. 2. Use the Solver to minimize risk using more liquid options. The liquid options available are as follows: Call/Put Call Call Call Call Call Call Price 100 100 100 100 100 100 Strike 100 90 110 100 90 110 Time to expiry 1 1 1 2 2 2 Implied volatility 25.5% 26.0% 24.0% 24.5% 25.5% 23.5% 3. Compare the risk exposure of the risk‐minimized portfolio to that of the original portfolio. How much has the risk been reduced? How would you characterize the exposures that remain? Managing Vanilla Options Risk 357 4. Is this a static hedge or will it need to be rehedged through time? 5. Create your own portfolio of less liquid options and go through the same exercise. 11.2 Monte Carlo Simulation of Options Hedging Program a Monte Carlo simulation to compare the results of dynamic hedging on a single option position and on an option hedged by other options. To begin, try to match the results in Table 11.2. Start with eight simulations (2 × 2 × 2), corresponding to a pure dynamic hedge/two‐sided options hedge, 0 percent standard deviation of volatility/33 percent standard deviation of volatility, and 100 rebalancings/500 rebalancings. Use 1,000 paths for each simulation. When using a standard deviation for volatility, apply it at the point that a volatility is assigned to a path (if you let the volatility vary at each rebalancing along the path, the volatilities will average out along the path and little difference will exist between the results of your 0 percent standard deviation and 33 percent standard deviation cases). For all eight cases, initial price = strike = 100, time = 1 year, average volatility = 20 percent, rate = dividend = 0 percent, and transaction costs are based on one‐fourth point per $100 bid‐asked spread, so any transaction, either buy or sell, incurs a cost of $0.125 per $100 bought or sold (but don’t charge any transaction cost for establishing the initial delta hedge). Use the OptionMC and OptionMCHedged spreadsheets to check your simulation programs. To do this, run your simulation for just 20 time steps. You can then check a particular path by taking the random numbers drawn for that path and substituting them for the random numbers selected in the spreadsheets. You can then compare results. Once you match the results from Table 11.2, you should try to expand the runs in the following ways: 1. Four runs with 100 rebalancings/500 rebalancings for the pure dynamic hedge/two‐sided options hedge, a 33 percent standard deviation of volatility, and a jump process. Jumps should occur on average once on each path, there should be a 50–50 chance that a jump is up or down, and the average absolute jump size should be 10 percent of the current price with a 33 percent standard deviation around this 10 percent. So a one‐standard‐deviation range 358 FINANCIAL RISK MANAGEMENT 2. 3. 4. 5. would be from 10% × exp(−0.33) = 7.2% to 10% × exp(0.33) = 13.9%. The volatility of the underlying should be adjusted down from 20 percent to whatever level will leave the average pure dynamic hedging cost equal to what it had been without the jump (you will need to try out a few different volatility levels to determine this). A similar set of runs to test the impact of a volatility skew with a standard deviation of 20 percent. For the case of a pure dynamic hedge, 500 rebalancings, and 33 percent standard deviation of volatility, check the impact of imposing different threshold levels for rehedging on the trade‐off between the expected transaction cost and standard deviation of P&L. Examine a sample of 50 individual paths and observe the relationship between the final price of the underlying and the total hedge P&L. Does the observed relationship support the claim in Section 11.3 that “despite wild gyrations in underlying prices, [the simulation paths] produce almost identical hedging results”? What pattern do you observe of hedge ratios along the individual paths? For example, how quickly does the hedge ratio go to 100 percent for paths whose final price is above the strike and to 0 percent for paths whose final price is below the strike? For parts 4 and 5 of this exercise, you need to examine individual paths of the Monte Carlo simulation. Use paths taken from the simulation with 0 percent standard deviation of volatility and 500 rebalancings. If you do not have the time or programming background to create your own Monte Carlo simulation, then carry out these parts of the exercise using the OptionMC1000 and OptionMCHedged1000 spreadsheets. Use the following input settings: price = $100, strike = 100, time to expiry = 1, implied volatility = 20 percent, volatility = 20 percent, skew = 0 percent, and jump probability = 0 percent. CHAPTER 12 Managing Exotic Options Risk e need to first determine what we mean by an exotic option. Some articles on options emphasize complex formulas and difficult mathematical derivations as the hallmarks that distinguish exotics from vanillas. The criterion I am using in this book emphasizes market liquidity. If you can readily obtain prices at which the option can be bought and sold, then it counts as a vanilla option; if not, then it is an exotic option. To understand why I favor this definition, consider a forward‐start option as an illustrative example. This is an option priced now, but its strike is not set until some future date. Generally, it is set to be at‐the‐money on that future date. There is certainly no complexity about the formula or mathematical derivation of the formula for this product. It is the standard Black‐ Scholes formula with the strike and underlying price set equal. However, this product has no liquid market, and relating its valuation and hedging to the valuation and hedging of ordinary European options is not straightforward. Equivalently, we can say that no clear relationship exists between the volatility that is needed as input to the Black‐Scholes formula for the forward‐start option and the volatilities implied by the prices of standard European options. The two preceding chapters, on managing forward risk and vanilla options risk, emphasized the use of methods that maximize the degree to which all transactions can be viewed as being managed within a common risk‐measurement framework—a single discount curve for forwards and the price‐vol matrix for vanilla options. This common framework increases the chance that exposures on different transactions can be netted against one another and offset by transactions involving the forwards and vanilla options with the greatest liquidity. This paradigm does not work for exotic options since none of them have enough liquidity to provide confidence that risks can be offset at publicly available prices. Therefore, the emphasis throughout this chapter is on methodology that enables, as much as possible, the risks in an exotic option to be represented W 359 360 FINANCIAL RISK MANAGEMENT as an equivalent vanilla option and forwards position. The vanilla option and forwards position is the liquid proxy representation of the exotic discussed in Section 6.1.2 and in Section 8.4. My primary arguments for anchoring exotic option risk management to a liquid proxy are presented there. Additional reasons were discussed in the arguments favoring a more detailed limit approach in Section 6.2, in the broader setting of general risk decomposition. As applied specifically to exotic options, these reasons are: ■ ■ ■ ■ It permits the separation of exotic options risk into a part that can be managed with vanilla options and a residual that cannot. It is important to identify and quantify this residual risk so that adequate reserves can be held against it, and to facilitate the management recognition of pricing that is inadequate to support actual hedging costs. Without separating out the part of the risk that can be hedged with liquid vanilla options, it is quite possible that gains from ordinary vanilla risk positions will obscure losses from the truly illiquid residual. It encourages as much of the risk as possible to be managed as part of the far more liquid vanilla options position. It reduces the risk of having exotic options positions valued with a methodology that is inconsistent with that used for valuing the vanilla options positions. It consolidates exotic options positions into already well‐developed reporting mechanisms for vanilla options—price‐vol matrices, volatility surface exposures, deltas, and other “greeks.” This has the advantage of building on well‐understood reports, thus clarifying the explanation to senior managers. It also guards against large positions accumulating without being recognized by a common reporting mechanism, since all exotics for a given underlying will be consolidated into the same set of vanilla options risk reports. The use of liquid proxies in risk management closely parallels the use of the control variate technique in modeling. The control variate technique uses the best available model to value a particular exotic option, but it also uses the same model to value a related vanilla option (or basket of vanilla options). Since the vanilla options are liquid, they can then be valued directly from the market (or interpolated from direct market prices). The model is only used to value the difference between the exotic and related vanilla. Risk reporting and risk management are similarly divided between reporting and managing the risk of the related vanilla option as we would any other vanilla and creating separate risk reporting and management for the difference between the exotic and related vanilla, thereby reducing the model dependence of valuation. See Hull (2012, Section 20.3) for a discussion that Managing Exotic Options Risk 361 emphasizes the computational efficiency of this technique, which is strongly analogous to the risk management advantages stressed here. If the underlying assumptions of the Black‐Scholes framework were true—in particular, if volatility was known and constant—the choice of models for exotics would generally be easy. Most exotics can be valued using formulas derived from market assumptions similar to those used in the Black‐Scholes analysis of European options. However, when volatility is unknown and variable, there is seldom a direct way of translating a volatility surface used for valuing European options into a single volatility to be used in valuing an exotic. Usually, we will need to rely on more complex formulations tailored to a particular exotic to establish this relationship. Much of this chapter is devoted to developing these formulations for specific exotics. A distinction that will prove very important when analyzing these models can be made between those where the relationship between the exotic and the vanilla is static and those where the relationship between the exotic and the vanilla is dynamic. Static relationships mean that the same vanilla (or package of vanillas) can be used to represent the exotic in vanilla option risk reports throughout the life of the exotic. Dynamic relationships mean that the package of vanillas used may need to change in composition over the life of the exotic. Dynamic relationships correspond to the full simulation approach recommended by Derman (2001) that was discussed in Section 8.4. Static relationships, and the quasistatic relationships I will discuss in a moment, correspond to the simulation of a limited hedging strategy I propose as an alternative to Derman’s full simulation approach in Section 8.4.3. Static representations have obvious operational advantages. Once it is booked at the inception of a trade, the representation does not need to be updated. Even more important is the simplicity introduced when the potential cost of differences between the actual exotic and its vanilla option representation over the life of the transaction is estimated. As emphasized in Section 11.3, dynamic representation requires simulation to evaluate potential costs. However, the simulation of dynamic changes in vanilla option hedges can be far more computationally difficult than the simulation of dynamic changes in underlying forwards hedges studied in Section 11.3. The reasons for this are discussed in Section 12.3.2. The ideal of a static representation cannot always be achieved. It will be possible in Sections 12.1 and 12.4 when we are discussing options whose payout depends on prices at a single future time. However, when discussing options whose payout is a function of prices at different times, as is true in Sections 12.2, 12.3, and 12.5, static representation will not be possible. Our alternatives will be either dynamic representation or quasistatic representation, in which changes in the representation are minimized, often to only a single change, to simplify calculations of potential cost. We will use the 362 FINANCIAL RISK MANAGEMENT simpler term static for the remainder of this chapter, but this is shorthand for quasistatic representation, and will pay due attention to the estimation of the cost of the hedge changes. Table 12.1, which was taken from Smithson (2000), shows the principal forms of exotic products and how widely they are used in different markets. The study of exotic options in this chapter is divided into five sections, following the categories used in Table 12.1. ■ ■ ■ ■ ■ Section 12.1—single‐payout options. These are options whose payoffs are the function solely of the price of an underlying asset at a single future time. We will show how to replicate these options exactly using a basket of forwards and vanilla options. The resulting replication can be used both to value the exotic and represent it in risk reports. The only residual risk will be the liquidity of the resulting basket, particularly in the replication of binary options. A particular example of an important single‐payout exotic is a log contract, which makes payments based on the logarithm of the underlying price. Its importance is mostly due to its close linkage to a variance swap, an exotic product not in Table 12.1 but that shows increasing use. We also discuss the volatility swap in this section, a close cousin of the variance swap. Section 12.2—time‐dependent options. These are options whose payoffs are the function of the price of a vanilla option at a single future time. As in Section 12.1, we will show how to eliminate all risk of underlying price movement for these exotics by replication using forwards and vanilla options. The residual risk exposure to implied volatility at a future time can be quasistatically hedged with vanilla options. These exotics include forward‐ start options, cliquet options, chooser options, and compound options. Section 12.3—path‐dependent options. These are options whose payoffs depend on the price of a single underlying asset at several future times. We will focus on barrier options, but also use the lessons learned to apply to ladder, lookback, double barrier, and partial‐time barrier options. We will examine and contrast replication approaches that utilize dynamic hedging with vanilla options and approaches that permit quasistatic hedging with vanilla options. Section 12.4—correlation‐dependent options. These are options whose payoffs depend on the prices of several underlying asset securities and that therefore must be priced based on assumptions about correlations. We will examine several important cases: basket forwards and options, quanto forwards and options, diff swaps, mortgage‐backed securities, collateralized debt obligations (CDOs), and convertible bonds. Section 12.5—correlation‐dependent interest rate options. A particular subset of correlation‐dependent options are options whose payoffs 363 R R R A A R R R A A A A R R A A O A A A A A O A A R A R R O A Equity Options OTC Exchanges A A O A O O A A A A O R O O O A A O R A A A Commodity Options OTC Exchanges FX = foreign exchange; OTC = over the counter; A = actively used; O = occasionally used; R = rarely used; blank = not used. Source: Smithson (2000). O A A R A R A A A A A O R O O R A FX Options OTC Exchanges A A O R O O R A A A Interest Rate Options OTC Exchanges Intensity of Use of Option Structures in Various Markets First-generation options European style American style Bermuda style Second-generation options Path-dependent options Average price (rate) Barrier options Capped Lookback Ladder Ratchet Shout Correlation-dependent options Rainbow options Quanto options Basket options Time-dependent options Chooser options Forward-start options Cliquet options Single-payout options Binary options Contingent premium options TABLE 12.1 364 FINANCIAL RISK MANAGEMENT depend on multiple future interest rates. This includes the important special case of American and Bermudan swaptions. 12.1 SINGLE‐PAYOUT OPTIONS In continuous time finance, the Breeden‐Litzenberger theorem states that any option whose payout is a smooth function of a terminal forward price can be perfectly replicated by an infinite package of forwards and plain‐vanilla calls and puts (see Carr and Madan 2002, Section II.A). The discrete time version states that any option whose payout is a smooth function of a terminal forward price can be replicated as closely as desired by a finite package of forwards and plain‐vanilla calls and puts, with the tightness of fit of the replication dependent on the number of vanilla calls and puts in the package. In both cases, replication is static, meaning the forwards and vanilla calls and puts are purchased at the deal inception and then no further hedging is needed. The terminal payout on the replicating package will match the terminal payout of the exotic option. The discrete time result can be established in two stages: 1. Any smooth function can be approximated as closely as desired by a piecewise‐linear function. The tightness of fit depends on the number of pieces of the replication. 2. Each piece of a piecewise‐linear function can be replicated by adding another vanilla option to a package of options that replicates all of the pieces up to that point. This can be easily seen from an example. Consider a function that pays out nothing at prices $100 or below, pays out $2 for every $1 gain in price up to $102, pays out $3.5 for every $1 gain in price from $102 to $105, and pays out $2.3 for every $1 gain in price above $105. This payout can be replicated by buying 2 calls at $100 and 1.5 calls at $102, and selling 1.2 calls at $105, as shown in Table 12.2. The BasketHedge spreadsheet on the website for this book enables you to calculate the vanilla option hedges and the associated valuations based on this discrete time approach. The impact of smiles and skews in the volatility surface of the vanilla options on the valuation of the exotic options can be readily calculated using this spreadsheet. Even if this is not selected as a desirable hedge from a trading viewpoint, it still makes sense as a way to represent the trade from a risk management viewpoint for the following reasons: ■ It permits realistic valuation based on liquid, public prices. Alternative valuation procedures would utilize an analytic pricing model, which is 365 Managing Exotic Options Risk TABLE 12.2 Price Vanilla Options Replication of a Piecewise‐Linear Payout Payout +2 Calls at $100 +1.5 Calls at $102 –1.2 Calls at $105 $100 0.0 0.0 0.0 0 $102 4.0 4.0 0.0 0 $105 14.5 10.0 4.5 0 $110 26.0 20.0 12.0 –6.0 ■ ■ usually easily derivable, but a level of volatility needs to be assumed and no straightforward procedure is available for deriving this volatility from observed market volatilities of vanilla options at different strikes. The hedge package method will converge to this analytic solution as you increase the number of vanilla option hedges used, provided all vanilla options are priced at a flat volatility (it is recommended that this comparison always be made as a check on the accuracy of the implementation of the hedging package method). However, the hedging package method has the flexibility to price the exotic option based on any observed volatility surface (in fact, instead of using the volatility surface, the directly observed vanilla option prices are used, so the pricing is not dependent on any option model). The hedge package method gives an easy means of integrating exotic options into standard risk reports, such as price‐vol matrices and vega exposure by strike and maturity. Placing as much risk as possible within a single context also increases the chances that risks from one position may offset risks in another position. Only the net risks need to be managed. (See the arguments for requiring internal hedging in Section 6.2.) Although the representation will be incomplete due to the use of a finite package of vanilla options, the residual risk can be easily calculated by Monte Carlo simulation based on an assumed probability distribution of final forward prices multiplied by the amount of mishedge. This is an easier calculation of remaining risk than the analytic method, which requires a Monte Carlo simulation of dynamic hedging. One objection that is sometimes raised to the static hedging strategy for exotic options is that the required basket of vanilla options is unrealistic, in terms of using options at strikes that have little market liquidity, in terms of the number of different options in the basket, or in terms of the required odd lots of individual options. 366 FINANCIAL RISK MANAGEMENT Although this objection may have validity in the context of a proposed actual hedge to be placed against a particular deal, it does not carry much force in the context of risk management, in which hedging strategies are utilized as devices for representing risk in standard reports through liquid proxies. The tools for managing vanilla European options within a portfolio framework are well established. As was pointed out when discussing dynamic hedging in Section 11.3, good empirical evidence exists that vanilla options at less liquid strikes when statically hedged with vanilla options at more liquid strikes result in dynamic hedging strategies that achieve far greater stability than pure dynamic hedging strategies. As a result, we would argue that risk managers should not hesitate to represent exotic option trades as baskets of vanilla options in a vanilla options portfolio risk report. The advantages are parallel to those cited at the beginning of Section 10.2 for representing an illiquid forward as a static combination of liquid swaps: unified risk reporting increases risk transparency, maximizing liquidity and minimizing transaction costs. The one point of legitimate concern would be if the resulting representation would be a position too large to be managed with the existing market liquidity. This would be an argument against representing a binary option as a very large position in a very narrow call spread. Instead, liquidity considerations should limit the size of the call spread position that is used as a representation, which in turn limits the narrowness of the call spread used. The resulting residual risks must be managed by the exotics desk through a combination of limits and reserves. We discuss this approach in more detail in Section 12.1.4. Another example would be if the representation revealed heavy reliance on very high‐ or low‐strike vanilla options outside the range at which the firm’s vanilla option traders would be comfortable managing the residual risk against more liquid strikes. Note that in both cases, the method of representing exotics exposure as a basket of vanilla options has the advantage of highlighting the regions of illiquidity impacting the exotic, a focus that many analytic pricing methods lack. These points hold generally for the replication of exotic derivatives with vanilla options. By representing the exotic derivative as closely as possible with a hedge package of vanilla options, you can minimize the remaining basis risk that needs to be managed using techniques specific to the exotic derivative and maximize the amount of risk that can be combined with and managed as part of the vanilla options book, utilizing established risk management tools such as the price‐vol matrix. Examples of options that can be risk managed in this way are calls on the square, cube, square root, or other power of the excess above a strike, or the corresponding puts. Other mathematical functions, such as the logarithm 367 Managing Exotic Options Risk of the excess above or below a strike, are also possible. This style of option, sometimes collectively known as power options, has largely fallen out of favor following the Bankers Trust (BT)/Procter & Gamble (P&G)/Gibson Greetings blowup of 1994, which is discussed in Section 4.3.1. The lawsuits and allegations prompted by large losses on contracts with complex payoff formulas with no discernible tie to any of the end user’s economic motives led to a distrust of such derivatives. Currently, most market makers’ client appropriateness rules permit such contracts only in very limited circumstances. Nonetheless, some power options remain in active use. The most prominent are log contracts, which are of particular interest because of their link to valuing and hedging variance swaps, and a type of quanto option that is utilized in the foreign exchange (FX) and bullion markets. In addition, the convexity adjustments needed for valuing and hedging certain types of forward risk, which we discuss in Section 10.2.4, can usefully be viewed as a type of power option and managed by this technique. After examining each of these three cases, we will follow with an examination of the important case of binary options, which illustrates the issue of how to handle liquidity risk arising from static replication. Finally, we will show how binary options can be combined with vanilla options to create other exotics—a contingent premium option and an accrual swap. 12.1.1 Log Contracts and Variance Swaps A variance swap is a forward contract on annualized variance whose payout at expiry is: ( ρ 2 VAR ) × N (12.1) where σ ρ2 is the realized stock variance (quoted in annualized terms) over the life of the contract, KVAR is the delivery price for variance, and N is the notional amount of the swap in dollars per annualized volatility point squared. The holder of a variance swap at expiry receives N dollars for every point by which the stock’s realized variance, σ ρ2 , has exceeded the variance delivery price, KVAR, and pays N dollars for every point by which the stock’s realized variance, σ ρ2 , falls short of the variance delivery price, KVAR. This contract can be generalized to assets other than stocks and to amounts other than dollars. Variance swaps give their holders a vega exposure similar to what they would have by purchasing a vanilla option. However, variance swaps differ from vanilla options in that their vega exposure remains constant over time, 368 FINANCIAL RISK MANAGEMENT whereas vanilla options may go into or out of the money, reducing their vega exposure. This can be a significant advantage to a position taker whose main concern is to find an investment that expresses her economic view of future volatility. It also has the advantage of enabling her to avoid maintaining delta and gamma hedges, which will be seen as a distraction to the real intention, which is just to express a volatility view. The downside is the relative illiquidity of variance swaps versus vanilla options, leading to their being priced with wider bid‐ask spreads. The log contract offers a means to link the hedging and valuation of the illiquid variance swap to that of liquid vanilla options, using the basket hedge methodology. The link between the variance swap and the log contract comes from the following analytic formula for the value of a log contract: ln F T 2 ∫ σ 2 dT (12.2) 0 where ln is the natural logarithm function, F is the current price of the underlying forward to contract expiry T, and σ2 is actual realized variance over that time period. This formula is a direct consequence of Equations 10 and 11 in Demeterfli et al. (1999). A derivation can also be found in Neuberger (1996). Under the Black‐Scholes assumptions of known constant volatility, this implies that the log contract should be valued at ln F – ½ σ2T T, an analytic formula used in the BasketHedge spreadsheet to check the value derived for the log contract when the volatility surface is flat. Since we can use the spreadsheet to find a set of vanilla options to replicate the log contract, we now have a hedging strategy for a variance swap. Buy a replicating set of vanilla options for twice the volume of log contracts as the volume of variance swaps sold (twice the volume in order to counteract the ½ in front of the integral in the formula). Delta hedge these vanilla options. Since the log contract is losing value at exactly the rate of 1 T 2 ∫ σ 2 dT , the delta hedging should be producing profits at exactly the rate 0 needed to cover payments on the variance swap. In practice, this will not work exactly, due to jumps in underlying prices, as explained in Demeterfli et al. (1999, “Hedging Risks”). Monte Carlo simulation would be necessary to quantify the risk of this tracking error. However, the replication of the log contract still offers a good first‐order hedge and valuation for the variance swap. The section “The Difficulty with Volatility Contracts” in the same article discusses why this approach will not work for volatility swaps, which differ from variance swaps by having a payout of (σρ – KVOL) × N rather than (σρ2 – KVAR) × N. No static hedge for the volatility contract exists. In the Managing Exotic Options Risk 369 categorization we are using in this chapter, it is path dependent and needs to be risk managed using the techniques of Section 12.3, utilizing local volatility or stochastic volatility models to determine dynamic hedges. However, its close relationship to the variance swap, and thus to the log contract, suggests the use of a liquid proxy approach: use dynamic hedging just for the difference between the volatility swap and log contract while static hedging the log contract. For further reading on the modeling and risk management of variance and volatility swaps, I highly recommend Demeterfli et al. (1999) and Gatheral (2006, Chapter 11). Exercise 12.1 asks you to utilize the BasketHedge spreadsheet to look at the impact of changes in the volatility surface on the valuation of log contracts and hence on variance swaps. Demeterfli et al. (1999) also has an instructive section on the “Effects of the Volatility Skew” on variance swaps. Log contracts and variance swaps require hedges over a very wide range of strikes and should therefore show valuation sensitivity across the whole volatility surface. This seems reasonable from an intuitive standpoint since changes in volatility impact variance swaps even when the underlying forward price has moved very far away from the current price, leaving a currently at‐the‐money option very insensitive to vega. So high‐ and low‐strike vanilla options are needed to retain the vega sensitivity of the package. 12.1.2 Single‐Asset Quanto Options In Section 12.4.5, we discuss dual‐currency quanto derivatives in which the percentage change of an asset denominated in one currency is paid out in another currency. For example, a 10 percent increase in the yen price of a Japanese stock will be reflected by a 10 percent increase in a dollar payment at a fixed‐in‐advance dollar/yen exchange rate. We will see that the forward price of a quanto is the standard forward multiplied by exp(ρσSσF), where exp is the exponential to the base e, σS is the standard deviation of the asset price, σF is the standard deviation of the FX rate, and ρ is the correlation between them. A related product is a single‐currency quanto derivative in which the asset whose percentage change is to be calculated is also the asset whose exchange rate is fixed. Here are two examples: 1. A dollar/yen FX option, which, if the yen rises in value by 10 percent relative to the dollar, will be reflected by a 10 percent payout in yen. Since the yen has gone up in value by 10 percent versus the dollar, the payout in dollar terms is 110% × 10% = 11%. In general, for a p percent increase, the payout is (1 + p%) × p% = p% + p2%. 370 FINANCIAL RISK MANAGEMENT 2. A dollar/gold option struck at $300 per ounce. If gold rises in value by 10 percent to $330 per ounce, the payment is 1 ounce × 10 percent = 0.1 ounces of gold. The payout in dollars is therefore 0.1 × $330 = $33, which is $300 × 11%. In general, for a p percent increase in gold prices, the payout is p% + p2%. Since just a single asset is involved, the σS and σF in the quanto formula are the same and ρ is equal to 1, so the standard forward is multiplied by exp(σS2). The BasketHedge spreadsheet has a worksheet called Quanto that calculates the value of a single‐asset quanto using a static hedge basket of vanilla options. As you can see from the spreadsheet, the hedge consists of 101 percent of a standard call at the quanto strike plus calls of 2 percent of the notional at all strike levels above the quanto strike. This gives a payoff, if the asset rises by p percent, of: 101% × p% + 2% × p =1 ∑i i =1 101% × p% + 2% × p2 − p = p% p 2 p2 % When the static hedge cost is computed from a flat volatility surface, the results agree exactly with an analytic formula derived from the forward multiplied by exp(σS2). If higher volatilities are assumed for higher strikes, the cost of the basket hedge will exceed the cost derived from the analytic formula. If lower volatilities are assumed for higher strikes, the cost of the basket hedge will be less than the cost derived from the analytic formula. 12.1.3 Convexity In Section 10.2.4, on applying mathematical models of forward risk to indexed flows, we raised the issue of convexity or nonlinearity of some index flows and the complications this can entail for valuing and hedging these flows. We pointed out the availability of analytic formulas that approximate the convexity adjustments needed to account for the impact on valuation of the nonlinearity of these flows. These approximation formulas (see formulas 6.3, 29.1, 29.2, and 29.4 in Hull 2012) all require an interest rate volatility as a key input. However, in a world of nonflat volatility surfaces, which implied volatility should be used? Equivalently, what are the strikes of the options contracts that should be used to hedge this exposure? The basket‐hedging methodology we have developed in this section provides a more precise valuation for convexity adjustments, one that is sensitive to the shape of the volatility surface, and also provides details of the required hedge that can be used to represent the exposure in conventional vanilla option position reports, as shown in the Convexity worksheet within the BasketHedge spreadsheet. 371 Managing Exotic Options Risk TABLE 12.3 Payouts of a Binary Option Final Price = S S≤K S>K Vanilla Call, Strike = K Cash‐or‐Nothing Call, Strike = K, Payout = K Asset‐or‐Nothing Call, Strike = K 0 S–K 0 K 0 S 12.1.4 Binary Options European binary options (also known as digital options or bet options) have highly discontinuous payoffs. The basic form, the cash‐or‐nothing option, which we will focus on in this section, pays either zero if the price finishes below the strike or a set amount if the price finishes above the strike. A variant, the asset‐or‐nothing option, pays zero if the price finishes below the strike or the ending price if the price finishes above the strike. An asset‐or‐ nothing option is simply the sum of a standard vanilla option and a cash‐ or‐nothing option at the same strike that pays the strike price. Table 12.3 illustrates the payouts. European binary options fulfill the condition of having a payout that is a function of the price of an asset at one definite time. Therefore, it can be treated by the methodology just stated, using a basket of vanilla options to hedge it and using this hedge package to calculate valuation, including skew impact, to calculate remaining risk, and to be incorporated into standard risk reports. However, the discontinuous nature of the payment at the strike leads either to unrealistically large hedge positions in vanilla calls (liquidity risk, since market prices would be impacted by an attempt to transact so many calls) or to significant hedge slippage (basis risk) between the binary option and its hedge. For example, let’s say a customer approaches a trading desk wanting to buy a one‐year binary call that will pay $10 million if the Standard & Poor’s (S&P) index is above the current one‐year forward level at the end of one year and nothing otherwise. The vanilla option decomposition of a barrier option is particularly simple. It can be represented as a call spread between two vanilla options of equal notional size. Assume you buy a vanilla call at a strike just below the current forward level and sell a vanilla option at a strike just above this level with a spread of 0.01 percent of the price between the two options. You will need to receive $10 million if the index rises by 0.01 percent above the first strike, since for any index move above the second strike, you are paying as much on the second option as you are receiving on the first. So the notional amount of the call to be bought and sold is $10 million/0.01% = $100 billion. 372 FINANCIAL RISK MANAGEMENT Let us start by assuming that all vanilla calls are priced at a 20 percent flat implied volatility. The straight analytical formula for the value of the binary option is the amount to be paid × N(d2), the term in the Black‐Scholes equation for a vanilla option that gives the risk‐neutral probability that the price will finish above the strike. In this case, we have: $10 million N (d2 ), ) where h d ⎛ ⎝ ⎞ 1 price / strike) – σ 2t ⎟ / σ t ⎠ 2 ⎛ ⎞ 1 = ⎜ln(1) – 20%2 ⎟ / 20% = 10%N (–0 (–0.1) 1) = 0.46017, giving a price off the ⎝ ⎠ 2 binary of $4,601,700 (12.3) Replicating the binary option using a vanilla call spread, the exact choice of vanilla calls to be used makes virtually no difference to the price (as long as we assume a flat implied volatility), but it does make a significant difference to the mix between liquidity risk and basis risk. For example: ■ ■ ■ Buy a vanilla call on $100 billion at a strike of 99.995% of the forward level at a price of BS(99.995%, 1, 20%) = 7.9678802% for $100 billion × 7.9678802% = $7,967,880,200 and sell a vanilla call on $100 billion at a strike of 100.005% of the forward level at a price of BS(100.005%, 1, 20%) = 7.9632785% for $7,963,278,500, for a net cost of $7,967,880,200 – $7,963,278,500 = $4,601,700. Buy a vanilla call on $2 billion at a strike of 99.75% of the forward level at a price of BS(99.75%, 1, 20%) = 8.0812430% for $161,624,900 and sell a vanilla call on $2 billion at a strike of 100.25% of the forward level at a price of BS(100.25%, 1, 20%) = 7.8511554% for $157,023,100, for a net cost of $4,601,800. Buy a vanilla call on $500 million at a strike of 99% of the forward level at a price of BS(99%, 1, 20%) = 8.4357198% for $42,178,600 and sell a vanilla call on $500 million at a strike of 101% of the forward level at a price of BS(101%, 1, 20%) = 7.5152765% for $37,576,400, for a net cost of $4,602,200. Note the inverse relationship between the width of the call spread (0.01 percent, 0.50 percent, and 2 percent, respectively) and the size of the legs of the call spread ($100 billion, $2 billion, and $500 million, respectively). The first combination offers the smallest basis risk. It will replicate the binary option exactly as long as the S&P index at the end of one year is outside the range 99.995% to 100.005%—that is, as long as the S&P index does not finish within about one‐half basis point of its current forward level. However, liquidity risk is heavy; purchases and sales in the size of Managing Exotic Options Risk 373 $100 billion would be certain to move market prices if they could be accomplished at all. (Even if the trading desk does not expect to actually buy this call spread, its use in representing the risk profile of the trade will lead to illiquid dynamic hedging requirements.) At the other end of the spectrum, the third combination is of a size that could possibly be transacted without major market movement, but basis risk is now much larger. Exact replication of the binary option takes place only in a range outside 99% to 101% of the current forward, so there are about 100 basis points of market movement on either side of the current forward level in which replication would be inexact. And replication could be very inexact. If the index ended at 100.1% of the forward, for example, the customer would be owed $10 million, but the vanilla call at 99% would pay only $500 million × 1.1% = $5.5 million, a net loss of $4.5 million. Of course, the basis risk can be dynamically hedged with purchases and sales of S&P futures. However, the large payment discontinuity of the binary option can lead to unmanageable hedging situations. For example, suppose you are close to expiration and the S&P is 1 basis point below the forward level. If no further movement occurs, you will make about $4.95 million [(99.99% – 99%) × $500 million] on the vanilla call and owe nothing on the binary, but an uptick of just 2 basis points will lead to a loss of about $5 million. Should you put on a delta hedge of a size that will make $5 million for a 2‐basis‐point uptick? The problem is that a position of this size will cost you $10 million for a 4‐basis‐point downtick, and you do not gain anything from option payouts to offset this loss. In theory, in a world of complete liquidity and no transaction costs, you could put on this hedge only at the exact moment you approach the binary strike and take it off as soon as you move away from that strike; but in practice, such strategies are wholly implausible. The actual experience of trading desks caught needing to delta hedge a sizable binary position that happens to be near the strike as expiration approaches is excruciatingly painful. Traders have their choice of gambles, but they must decide on a large bet in one direction or another. In light of this, risk managers will always seek to place some sort of controls on binary positions. These controls, which may be complementary, come in the form of both limits and reserves. Limits are placed on the size of the loss that can occur for a certain size price move, the maximum delta position that can be required for a hedge, or the maximum gamma (the change in delta) that can be required for a given price move. Delta and gamma limits are based on the anticipated liquidity and transaction costs of the underlying market in which hedging is being done. Limits on loss size are designed to enable traders to take a purely insurance approach to binaries, hoping to come out ahead in the long run. This requires that 374 FINANCIAL RISK MANAGEMENT no one binary be too large. Such an approach needs to be combined with eliminating binaries close to a strike and expiration from delta and gamma reports, so that delta hedging is not attempted. It also requires decisions about how binaries should be combined for limit purposes. To operate like insurance, binaries need to be widely scattered as to maturity date and strike level, and limits need to bucket strikes and maturities in a manner that forces this scattering. However, bucketing should combine binaries in only one direction (bought or sold); it is dangerous to permit the netting of one binary with another except when date and strike (and any other contract terms, such as a precise definition of the index) exactly match. A valuation and reserve policy should also be consistent with the insurance approach to binaries—profit and loss (P&L) should be recognized only to the extent it can come close to being locked in. Gains that have great uncertainty attached to them should only be recognized when realized. This can be accomplished with several methods. I will provide a detailed example of a method that I consider particularly elegant in its capability to balance liquidity and basis risks, its maximal use of static hedge information, and its good fit with dynamic hedging risk reporting. In this approach, every binary has an internal liquid proxy representation assigned to it that is designed to be as close as possible to the binary in its payouts while still being capable of liquid hedging and conservative relative to the binary in that the internal representation will always produce a lower P&L for the firm than the binary. All risk reports for the firm are based on the internal representation, not the true representation of the binary. No special rules are required for eliminating binaries close to a strike and expiration from the firm’s delta and gamma reports, since the internal representation has been designed to be small enough not to require unreasonable hedges. The valuation difference between the true and internal representation, which by design must always be a positive value to the firm, is booked to a reserve account. Since the reserve is always positive, this policy sometimes results in the firm recognizing windfall profits, but never windfall losses. Let’s see how this policy would work in the case we have been considering. A call spread is selected as the internal representation of the binary by choosing the smallest spread that results in a position size that is considered to be small enough to be liquid, either by representing a real possibility for purchase in the market or by being representable in the firm’s risk reports by delta positions that can be achieved with reasonable liquidity. However, rather than choosing a call spread that straddles the binary, and therefore has payouts greater than the binary in some scenarios, we choose a call spread that is on one side of the binary and therefore always has payouts greater than the binary. If 2 percent is the width of the call spread Managing Exotic Options Risk 375 we select as the smallest consistent with a liquid position, then we use as an internal representation a call spread consisting of a sale of $500 million at a strike of 98 percent and a purchase of $500 million at a strike of 100 percent (notice that the internal representation has the opposite sign from the hedge that would extinguish it). The resulting valuation would be $500 million × BS(98%, 1, 20%) – $500 million × BS(100%, 1, 20%) = $500 million × 8.9259724% – $500 million × 7.9655791% = $44,629,900 – $39,827,900 = $4,802,000. This is the valuation of the internal representation. The actual binary continues to be valued at $4,601,700; the difference of $200,300 is placed into a reserve. If the actual sale price of the binary to a customer is $5 million, then only $200,000 of the profit from the difference between the price and valuation goes into immediate P&L recognition; the other $200,000 goes into a reserve against anticipated liquidity costs of managing the binary risk. What happens to this reserve? There are several possibilities: ■ ■ The firm might decide to actually buy the static overhedge, which costs $4,802,000. The internal hedge reports of the firm will not show the net position between the internal representation of the binary and the actual call spread hedge. If the S&P index ends up below 98 percent or above 100 percent, no difference will appear between the eventual payout under the binary and the pay‐in due to the call spread, and the reserve will end up at zero. If the S&P index ends up between 98 and 100 percent, the call spread will have a pay‐in while the binary has no payout. For example, if the S&P index ends at 99 percent, the call spread will pay $5 million, which will be the final value of the reserve. At expiry of the options, this $5 million will be recognized in P&L as a windfall gain. The firm might not do any static hedging and might just delta hedge based on the internal representation of the static overhedge. Since the static overhedge was selected to be of a size that enables liquid delta hedging, the results in this case should be close to the results in the case that the static overhedge is actually purchased, but with some relatively small variance. As an example, suppose that we are very close to expiry and the S&P index forward is at 99 percent. Based on the internal representation of the call spread overhedge, the appropriate delta will be a full $500 million long in the S&P index forward, and roughly $5 million in dynamic hedging profits should already have been realized but held in reserve. If the index ends at 99 percent, the $5 million in dynamic hedging profits will be taken from the reserve and recognized in P&L as a windfall gain. If the index ends just above 100 percent, the $5 million in dynamic hedging profits realized to date plus the 376 ■ FINANCIAL RISK MANAGEMENT $5 million gain from the 1 percent increase on the $500 million long in the S&P index will be exactly enough to pay the $10 million owed on the binary. Note that keeping the $5 million in dynamic hedging profits realized to date in reserve is necessary to avoid having to reverse a previously recognized gain in order to pay off on the binary. Other combinations are possible, such as static hedges that are not overhedges, but all produce similar results. In Exercise 12.2 you will run a Monte Carlo simulation of the potential differences between final payout on a portfolio of binary options and the overhedge liquid proxy, utilizing the spreadsheet BinaryMC. It will allow you to see a practical example of how a well‐diversified portfolio of binaries requires lower reserves than a more concentrated portfolio of binaries. This technique of representing a binary internally as a static overhedge is sometimes objected to by front‐office personnel as trading off a very probable gain in order to achieve security. In this view, the $400,000 that was originally realized on the transaction was real P&L, and $200,000 was sacrificed in order to achieve security in the very small minority of cases in which the index finishes very close to the strike. The idea that $200,000 has been thrown away is, in fact, an optical illusion caused by focusing only on those cases in which the index finishes outside the 99 to 101 percent range. The trade still has a $400,000 expected value—it just consists of a sure $200,000 in the vast majority of cases in which the index finishes outside 99% to 101% and a set of windfall profits up to $10 million when the index finishes within this range. The front‐office view would be correct if some means were available, such as dynamic hedging, of being almost sure of achieving this $400,000 result in all cases. But it was exactly the lack of such means—the fact that the use of dynamic hedging to try coming close to achieving $400,000 in all cases results in some cases with disastrous losses—that caused us to seek an alternative approach. This reserve methodology can be seen as being consistent with moving the front office away from viewing these trades as normal derivatives trades that can be approached in an isolated manner and toward viewing them as necessarily being part of a widely diversified portfolio of binaries. In this context, over a long enough time period, the sum of occasional windfall gains can become a steady source of income. If limits can ensure a wide enough diversification, then reserves may not be necessary. So far in the example we have assumed a lack of volatility skew. In the presence of skew, the binary will price quite differently. Let’s see the impact of using a 20.25 percent implied volatility for a strike of 99 percent and a 20 percent volatility for a strike of 101 percent. The cost of the 99 percent vanilla call is now BS(99%, 1, 20.25%) = 8.534331%, resulting in a net cost Managing Exotic Options Risk 377 of $5,095,274. Just as with the cases previously discussed, the reduction to a package of vanilla options lets us pick up the impact of volatility skew. We can see that binary options are highly sensitive to skew. Taleb (1997, Chapter 17) gives a lucid discussion of the practical aspects of hedging binary options. On page 286, Taleb says that “the best replication for a binary is a wide risk reversal (that would include any protection against skew). There will be a trade‐off between transaction costs and optimal hedges. The trader needs to shrink the difference between the strikes as time progresses until expiration, at a gradual pace. As such an optimal approach consumes transaction costs, there is a need for infrequent hedging.” Using a call spread (also known as a risk reversall) that is wide reduces the size of the vanilla options that are needed, reducing transaction costs and liquidity concerns, and also capturing the volatility skew more accurately, since a wide spread could utilize more liquid strikes. As we have seen, the width of the spread should not materially impact the total hedge cost. In many cases, the underlying price will finish nowhere near the strike and no further transactions are needed. However, in those cases where the underlying is threatening to finish close to the strike, the basis risk will get too large and the trader will need to roll from the original call spread into a tighter call spread, incurring transaction costs due to the need to purchase and sell options and because the sizes of the option transactions are growing as the spread narrows. Factoring this potential transaction cost into the valuation of binary options is an alternative method for establishing a valuation reserve on a binary. As Taleb (1997, 286) states, “when the bet option is away from expiration, the real risks are the skew. As it nears expiration, the risks transfer to the pin. In practice, the skew is hedgeable, the pin is not.” (We have been using basis risk for what Taleb terms the pin risk.) Gatheral (2006, Chapter 8) also has a good discussion of digital options, with a very clear demonstration of the dependence of digital option valuation on the skew of the volatility surface. 12.1.5 Contingent Premium Options A contingent premium option entails no initial payment by the option buyer, who pays only at option termination under the circumstances that the option finishes in‐the‐money. This type of option is popular with some clients because of the deferral of cash payment and because the client will not need to pay for an option that turns out to be useless, although it should be noted that an option that finishes just slightly in‐the‐money will still require a net payment by the option buyer, since the payment due from the option seller will be less than the option’s cost. It is easy to see that a contingent premium option is just a standard vanilla option plus a forward to defer payment of 378 FINANCIAL RISK MANAGEMENT the option premium plus a binary option to offset the option premium due in the event the price finishes below the strike of the vanilla option. 12.1.6 Accrual Swaps Accrual swaps are swaps where interest on one side accrues only when the reference rate is within a given range (see Hull 2012, Section 32.6). An accrual swap can be represented as a package of binary caps and floors since interest accruing is an all‐or‐nothing event. Being above the floor rate requires the payment and being above the cap rate cancels the payment, which can be represented by a payment with the opposite sign. 12.2 TIME‐DEPENDENT OPTIONS Now that we have provided a methodology for hedging and valuing the price of a linear underlying instrument at a single future point, we will extend that approach to exotic options whose payoffs depend on the price of a vanilla option at a single future point. This dependence on a vanilla option’s future price can be decomposed into dependence on the price of the underlying of the vanilla option and dependence on its implied volatility. We can hedge the first element of this decomposition by a direct application of the methodology of the preceding section, leaving only dependence on implied volatility. To see how to hedge this piece, let us first look at an exotic that is dependent only on implied volatility and has no dependence on the underlying price. 12.2.1 Forward‐Starting and Cliquet Options A forward‐start option is specifically constructed to have its price depend entirely on the at‐the‐money implied volatility of a vanilla option at a specified time. For example, a forward‐start option could be sold on April 1, 2013, for a one‐year at‐the‐money option to buy 1,000 shares of IBM that starts on November 1, 2013. The strike of the option will be set on November 1, 2013, at the then underlying price. Hence, no underlying price exposure exists prior to November 1, 2013, and the only exposure prior to that time is to what implied volatility the at‐the‐money option will sell at on November 1, 2013. A cliquet option is a package of forward‐start options, usually with one starting just as the previous one expires. For example, a cliquet might consist of three‐month forward‐start options beginning March 10, June 10, September 10, and December 10, 2013. Since the payoff on each option in Managing Exotic Options Risk 379 the package is determined independently of any other option in the package, a cliquet can be valued by valuing each forward‐start option separately and then summing. A natural approach would be to consider valuing a forward‐start option with an extension of the method we used to roll into a longer‐term option (see Section 11.6.3). The only difference is that we need to set up the target price‐vol profile that we want to achieve as that of an at‐the‐money option, regardless of the underlying price level. The ForwardStart spreadsheet on the website for this book shows the details. The essential point is that the difference in the price of the at‐the‐money option at two different implied volatility levels, σ1 and σ2, can just be represented as BS(100%, T, σ1) – BS(100%, T, σ2), where T is the tenor of the at‐the‐money option to be created. Optimal fitting can then find the combination of current options that has close to the desired profile of volatility exposure at the time the forward‐start option expires and the at‐the‐money option begins. If you look at the example given in Table 12.4, you will find that the package of current options that creates the desired profile has a significant weighting at many different strike levels, so it will vary in valuation based on both the current smile and the current skew. This is not surprising, given that we are creating an option that has flat exposure to future implied volatility levels at all strikes. The situation parallels that of the log contract, which has flat exposure to variance. 12.2.2 Compound Options It is now quite straightforward to extend this approach to exotics that depend on both underlying and implied volatility. A call‐on‐a‐call option is one example of a compound option, which gives the purchaser of the compound option the right to buy (or sell) a particular vanilla option at a given strike price. It is also known as a split‐fee option because a major selling point is that a customer who may want an option but is not willing to invest that much in one can put up a smaller down payment to defer the decision. Analytical formulas for compound options, assuming flat volatility surfaces and constant volatility, are well known (see Hull 2012, Section 25.6). We will make use of these formulas to work through an illustrative example. Let’s say that a customer wants to buy a one‐year at‐the‐money call on 100 million euros on April 1, 2013, expiring on April 1, 2014. Assuming 20 percent implied volatility, the cost would be 7.97 percent of the principal amount. We’ll assume the at‐the‐money euro exchange rate is $0.90. The customer might prefer to pay 4.45 percent to get an option that can be exercised on November 1, 2013. On that date, the customer can either pay 5 percent to get a call on 100 million euros at a strike of $0.90 expiring on 380 2% 5 TABLE 12.4 –8% 0.52% –0.14% –0.30% –0.12% 0.13% 0.23% 0.16% 0.02% –0.03% 0.07% 0.32% 5.00% Volume call/put Price Strike Time Implied vol BS price Delta Vega Gamma Theta Current Years 2 –1 –1.8054 6.9829 –5.9652 2.7291 call call call call call 100 100 100 100 100 100 80 90 100 110 1 2 2 2 2 20.0% 20.0% 20.0% 20.0% 20.0% –7.58% –37.71% 103.69% –60.70% 18.44% 0.0% – 148.7% 486.3% –331.8% 115.3% –0.38% –0.61% 3.13% –3.01% 1.37% 0.0% –1.7% 8.6% –8.3% 3.8% 0.000% 0.011% –0.060% 0.058% –0.026% –6.86% –34.12% 93.82% –54.93% 16.69% Implied Volatilities –6% –4% –2% 0% 2% 0.39% 0.28% 0.18% 0.07% –0.03% –0.06% –0.04% –0.04% –0.07% –0.10% –0.18% –0.13% –0.10% –0.09% –0.08% –0.10% –0.09% –0.07% –0.05% –0.03% 0.02% –0.03% –0.03% –0.01% 0.01% 0.06% –0.01% –0.02% 0.00% 0.04% 0.02% –0.03% –0.04% –0.01% 0.03% –0.04% –0.07% –0.07% –0.04% 0.00% –0.06% –0.09% –0.09% –0.07% –0.03% 0.00% –0.05% –0.08% –0.08% –0.07% 0.17% 0.05% –0.02% –0.07% –0.08% 4% –0.12% –0.14% –0.08% –0.01% 0.05% 0.08% 0.08% 0.05% 0.01% –0.03% –0.08% 0.3272 call 100 120 2 20.0% 1.43% 10.1% 0.15% 0.4% –0.003% 1.29% 6% –0.22% –0.18% –0.09% 0.01% 0.08% 0.12% 0.13% 0.11% 0.06% 0.01% –0.05% 8% –0.31% –0.22% –0.10% 0.02% 0.11% 0.16% 0.18% 0.16% 0.12% 0.06% –0.01% –0.9075 2.2714 call call 100 100 80 90 1 1 20.0% 20.0% –18.29% 29.36% –80.6% 166.9% –0.17% 0.71% –0.9% 3.7% 0.006% –0.027% –16.55% 26.57% Hedge at Rollover of a One‐Year Option with a Forward Start in Two Years 0.12% 30.3% 0.12% 0.0% 0.001% 0.105% Spot–Vol Matrix Price –25 –20 –15 –10 –5 0 5 10 15 20 25 Portfolio Volatility Discount Spacing Price Vega –0.05% –0.01% 0.00% 0.01% 0.01% 0.01% 0.02% 0.02% 0.01% 0.00% –0.01% –5.8165 call 100 100 1 20.0% –44.07% –314.0% –2.20% –11.5% 0.085% –39.88% 0.00% 0.01% 0.01% 0.00% –0.01% –0.02% –0.01% –0.01% 0.00% 0.00% –0.01% 4.9924 –2.3706 call call 100 100 110 120 1 1 20.0% 20.0% 20.38% –4.84% 176.4% –49.4% 1.77% –0.66% 9.3% –3.4% –0.068% 0.025% 18.44% –4.38% Vega Convexity Managing Exotic Options Risk 381 April 1, 2014, or choose to let the option expire. The attraction to the customer is that if the euro declines in value by November 1, 2013, the option will seem unattractive and he will have saved money by having paid only 4.45 percent rather than 7.97 percent for the original option. The customer will pay more than 4.45 percent only if the option turns out to be valuable. Of course, the downside is that if he does want the option, he will have paid a total of 4.45% + 5.00% = 9.45% for it rather than 7.97 percent. When the call‐on‐a‐call option expires on November 1, 2013, the value of the call option that the customer must now decide to purchase or let expire is determined by both the price of the underlying euro exchange rate (forward to April 4, 2014) and the implied volatility for a six‐month option on the euro struck at $0.90. The basket hedging procedure used in Section 12.1 can find a set of vanilla option hedges that eliminate the risk of the uncertainty of the underlying euro exchange rate. However, exposure to the uncertainty of the six‐month implied volatility on November 1, 2013, will remain. This implied volatility exposure can be hedged by the same option roll approach as used in Section 12.2.1. The Compound worksheet of the BasketHedge spreadsheet calculates the vanilla option hedge against the underlying price and also calculates the price‐vol matrix exposure of the resulting hedged position. This price‐vol matrix can then be used as input to the ForwardStartOption spreadsheet to compute a hedge on the residual forward‐starting volatility risk. Exercise 12.1 takes you through pricing this call‐on‐a‐call option in the BasketHedge spreadsheet. For a flat volatility surface, the basket hedge reproduces the analytical value, but different valuations are produced in the presence of smile and/or skew. Further steps in the exercise have you utilize the spreadsheet to calculate hedges and valuations for other compound options and choose options in which the decision on whether an option should be a call or a put can be deferred. 12.3 PATH‐DEPENDENT OPTIONS So far we’ve dealt strictly with exotic options whose payment is based on the price of an asset at a single time period—that is, European‐style options. Now we want to look at how an option that is based on the prices of a single asset at many time periods can be handled. Barrier options are a good example to focus on for the following reasons: ■ ■ They illustrate dependence on the entire volatility surface, in terms of both time and strike level. They have a large range of variants. 382 FINANCIAL RISK MANAGEMENT ■ They are overwhelmingly the most traded exotic options among FX options and are also used with equities, commodities, and interest rates. They can be used as building blocks to form static hedges for other exotic options, such as lookback and ladder options. ■ A barrier option is one whose payoff is equal to that of a standard call or put, but that pays off only under the condition that some price level (called the barrier) has been breached (or not) at some time period prior to the time the call or put payoff is determined. Options that pay only if a barrier has been breached are called knock‐in (down and in if the barrier is below the asset’s price at the time the option in written, and up and in otherwise). For example, a one‐year down‐and‐out call on the S&P index with a barrier of 1,050 will have no payout if the S&P index goes below 1,050 at any time during the year. Options that pay only if a barrier has not been breached are t Variations include called knock‐outt (either down and outt or up and out). double barrier options that either knock out if either a down‐and‐out or an up‐and‐out condition has been reached or knock in if either a down‐and‐in or up‐and‐in condition has been reached. Another variation is a partial‐time barrier, where the barrier condition can be activated only during a specified time period that begins after the option start date and/or ends before the option termination date. A variation that can be combined with all of these options is a fixed rebate to be paid if an option is knocked out. We will first show that standard analytic models for barrier options are inadequate, both for valuation and for risk representation, in the presence of nonflat volatility surfaces for vanilla options. We will therefore need to turn our attention to two alternative approaches to valuing and hedging barriers: dynamic hedging utilizing both vanilla options and the underlying and quasistatic hedging with vanilla options. One particular quasistatic hedging approach, developed by Peter Carr, is particularly useful for developing an intuitive understanding of the risk profile of barrier options. We will then demonstrate how to statically hedge lookback and ladder options with barrier options and how to handle rebates. Finally, we will briefly discuss how the methods developed for standard barrier options can be applied to the broader class of single‐asset exotic options, including double barriers and partial‐time barriers. One noticeable difference between this section and all of our previous discussions of options is that we are concerned with the drift, t which can be thought of either as the difference between the risk‐free rate and the dividend rate, or more generally as the discount rate between forward prices at different expiries. Up until now, we didn’t need to worry about drift because we were considering only options whose value would be determined by the Managing Exotic Options Risk 383 asset price at a single point in time; hence, all hedges could be based on a forward with a single expiry date. Since we are now considering options that depend on price behavior at several points in time, hedges may need to involve forwards for different expiry dates and the relationship between forward prices can no longer be ignored. 12.3.1 Standard Analytic Models for Barriers Good analytic models based on partial differential equations (PDEs) have been developed for barrier options; see Hull (2012, Section 25.8) for the equations. Analytic models have great advantages in terms of computational speed relative to Monte Carlo and tree‐based models. The ease of calculating a valuation by just plugging input variables into a formula explains much of the success of the Black‐Scholes equation. The formulas for barrier options require a bit more computation than Black‐Scholes, but they are still quite manageable. However, the analytic models for barriers have the drawback that they need to assume a single level of volatility, and there are no good rules for translating a volatility surface observed for European options into a single volatility to be used for a particular barrier option. In fact, cases can be shown where no single volatility assumption can be utilized with the standard analytic approach to give a reasonable price for the barrier option. We will illustrate this point with the following example. Consider an at‐the‐money three‐month up‐and‐out call that knocks out at a barrier 20 percent above the strike. Its valuation at different volatility levels, using the standard analytic formula shown in Hull (2012, Section 25.8) is shown in Table 12.5. Note that the analytic result has option values that first increase as the volatility level rises, since rising volatility causes the call value to increase. At higher volatility levels, the option values decrease as the volatility level rises, since rising volatility increases the probability of a knock‐out. Since the barrier level starts far away from the current price, it is only at high volatilities that the impact of rising volatility on the probability of a knock‐out dominates the impact of rising volatility on the value of the call. The methods for utilizing the full volatility surface, which we will discuss shortly, would agree with these analytical results for flat volatility surfaces. However, if we assume a nonflat volatility surface, with an implied volatility of 20 percent for a European call struck at 100 and 18 percent for a European call struck at 120, approaches that utilize the full volatility surface (either the Derman‐Kani dynamic hedging approach or the Carr static hedging approach) would price the barrier option at 3.10, which is 10 percent higher than the 2.81 maximum value the barrier option reaches at any volatility level using the analytic approach. The reason for this is 384 FINANCIAL RISK MANAGEMENT TABLE 12.5 Value of a Barrier Based on Analytic Formula Volatility Value of Up‐and‐Out Call 1.00% 2.00% 3.00% 4.00% 5.00% 6.00% 7.00% 8.00% 9.00% 10.00% 11.00% 12.00% 13.00% 14.00% 15.00% 16.00% 17.00% 18.00% 19.00% 20.00% 21.00% 22.00% 23.00% 24.00% 25.00% 26.00% 27.00% 28.00% 29.00% 30.00% 0.1995 0.3989 0.5984 0.7979 0.9973 1.1968 1.3962 1.5956 1.7942 1.9897 2.1772 2.3499 2.5008 2.6242 2.7166 2.7771 2.8070 2.8087 2.7858 2.7421 2.6816 2.6080 2.5245 2.4340 2.3390 2.2415 2.1432 2.0455 1.9492 1.8552 that the lower volatility as you approach the barrier decreases the chance of penetrating the barrier without simultaneously lowering the value of the call. This example also shows why the analytic method is inadequate for representing the risk in standard option reports. The analytic method does not give any breakdown of how much of the risk should be represented as sensitive to changes in the at‐the‐money vanilla options versus how much should be represented as sensitive to changes in the out‐of‐the‐money vanilla options. Managing Exotic Options Risk 385 12.3.2 Dynamic Hedging Models for Barriers Dynamic hedging models price barrier options (or any other exotic option whose payoff is a function of a single underlying asset) based on the cost of dynamically hedging the exotic with a portfolio of the underlying asset and vanilla European options. This is analogous to the Black‐Scholes model pricing of vanilla European options based on the cost of dynamically hedging with the underlying asset. These models utilize the full set of the current prices of vanilla European options, so they make use of the full volatility surface along with a theory of how these vanilla option prices can evolve with time. If you utilize an actual dynamic hedging strategy consistent with the model, you will be successful in replicating the model’s price for the exotic to the extent that the model’s theory about the evolution of the vanilla options prices is correct and that transaction costs are manageable. Two principal types of dynamic hedging models are used for exotics: 1. Local volatility models that assume that volatility is a known and unvarying function of time and the underlying price level. These models are natural extensions of the Black‐Scholes model, which assumes that volatility is known and unvarying, but which also assumes it is the same at all times and underlying price levels. Based on the assumption of the local volatility model, you can derive a definite price at any future time and the underlying price level of any vanilla or exotic option. The cost of the dynamic hedge therefore differs from the originally derived price only to the extent that future volatilities prove to follow a varying function of time and underlying price level (or that transaction costs are significant). 2. Stochastic volatility models that assume that volatilities will vary over time and that might include price jumps, based on some assumed model. The cost of the dynamic hedge differs from the derived price to the extent that the process of actual volatility variation differs from that assumed by the model (or to the extent that transaction costs are significant). A relatively straightforward implementation for a local volatility model is the trinomial tree approach of Derman and Kani (1994), which builds the unique trinomial tree for modeling the price diffusion of the underlying asset that meets the following two criteria: 1. Volatility is a known and unvarying function of time and the underlying price level. 2. The tree correctly prices alll European calls and puts on the underlying asset at different strike levels and times to expiry. 386 FINANCIAL RISK MANAGEMENT A thorough discussion of the Derman‐Kani approach and its application to barrier pricing can be found in Chriss (1997, Chapters 9 and 11). If any reader wants to implement this model, I strongly recommend reading Chapter 5 of Clewlow and Strickland (1998), which provides wonderfully detailed instructions and examples. A general introduction to stochastic models can be found in Derman and Kani (1998). A frequently used computationally tractable stochastic volatility model is that found in Heston (1993). A model that is attracting current interest is the variance gamma modell, which is explained in Madan, Carr, and Chang (1998). Gatherall (2006) and Lee (2001) contain insightful analysis on the differences between local volatility and stochastic volatility models in the pricing of exotic options. Matytsin (1999) suggests that a combination of stochastic volatility and jump processes is needed to explain observed volatility surfaces implied by the vanilla option prices. The jump processes are needed to explain the steepness of smile and skew observed at shorter‐term maturities, whereas stochastic volatility is needed to explain the steepness of smile and skew at longer‐term maturities. Dynamic hedging utilizes the full volatility surface in pricing barrier options. It can be readily employed for representing the barrier option in risk reports through its vanilla option hedges. Dynamic hedging can also be applied to any derivative based on a single underlying. Its drawback is its vulnerability to incorrect assumptions about volatility evolution and possible instability of the hedge representation. The most thorough discussion of the vulnerability of dynamic hedging models to incorrect assumptions about volatility evolution that I know of is in Gatheral (2006), a relatively short book that is long on elegance and insight. At the close of Chapter 4, Gatheral states that “From the results of our computation, we can see that the local volatility model and the stochastic volatility model price European options almost identically” and that “to value an option, it’s not enough just to fit all the European option prices, we also need to assume some specific dynamics for the underlying.” In Chapter 8, Gatheral then analyzes the difference in evolution of the volatility surface implied by local volatility models versus stochastic volatility models. He states, “If the payoff we are hedging depends (directly or indirectly) on the volatility skew, and our assumption [which is implied by a local volatility model] is that the . . . skew is independent of the volatility level, we could end up losing a lot of money if that’s not how the market actually behaves.” Once an exotic has been priced by a given model, the exotic can be hedged by a set of vanilla options that have the same sensitivity to the model’s input parameters as the exotic. As long as the model’s input parameters remain unchanged, the hedge does not require changing. However, changes Managing Exotic Options Risk 387 in observed vanilla option prices may require changes to input parameters to fit current prices, and once parameters change, the hedge may need adjustment. How stable is the resulting representation? To what degree does it require frequent and sizable adjustments in the options hedges that can result in hedge slippage as a result of both transaction costs (generally considerably higher for options than for the underlying) and the instability of the hedge against parameter changes? The more the price of a product is dependent on assumptions about volatility evolution, the greater the instability of hedges. Although trading desks may gain experience with the stability of particular models in particular markets through time, it is difficult to obtain a risk measure in advance. The projection of hedge changes through Monte Carlo simulation (as recommended by Derman [2001] as discussed in Section 8.2.6.2), which has proved very useful in establishing results for the hedging of vanilla options with other vanilla options, is orders of magnitude more difficult to achieve for exotics. This is because each step on each path of the Monte Carlo simulation requires recomputation of the hedge. When the only hedge change is in the underlying, this is a very simple calculation of the N(d1) in the Black‐Scholes formula. When the hedge change is in an option, a complete recalculation of the model being used to link the vanilla options and the exotic option together is required. 12.3.3 Static Hedging Models for Barriers The uncertainty surrounding the hedging costs of using dynamic hedging for barriers provides the motivation to search for static or near‐static hedging alternatives. Static hedging models price barrier options based on the cost of a replication strategy that calls for an almost unvarying hedge portfolio (at least of the vanilla options; it would be possible to use a dynamic hedge of the underlying, although the particular static hedging models we discuss only utilize vanilla options in the hedge portfolio). These models utilize nearly static hedge portfolios both as a way to reduce transaction costs and as a way to reduce dependence on assumptions about the evolution of volatility. Chapter 9 of Gatheral (2006) analyzes these nearly static hedges of barrier options from a different vantage point than mine, but with broadly similar conclusions. Three approaches to the static hedging of barriers can be distinguished: 1. The approach of Derman, Ergener, and Kani, which is broadly applicable to all exotic options whose payoff is a function of a single underlying asset, but has considerable exposure to being wrong about future volatility levels. 388 FINANCIAL RISK MANAGEMENT 2. The approach of Carr, which is more specifically tailored to barrier options, utilizing an analysis of the Black‐Scholes formula to form a hedge portfolio that is immune to changes in overall volatility level and volatility smile. However, the Carr approach is still vulnerable to changes in the volatility skew. It is easier to implement than the Derman‐Ergener‐Kani approach for barriers in the absence of drift (that is, forward equal to spot) and produces a very simple hedging portfolio that helps develop intuitive understanding of the risk profile of the barrier. 3. Approaches that utilize optimal fitting give solutions close to those provided by the Carr approach for single barriers in the absence of drift, but are more flexible in handling drift and are less vulnerable to changes in volatility skew. Optimal fitting can be generalized to broader classes of exotics, but with less ease than the Derman‐Ergener‐Kani approach. All three approaches are based on the idea of finding a basket of vanilla options that statically replicate the differences between the barrier option and a closely related vanilla option. To facilitate the discussion, we will confine ourselves to the case of a knock‐out call, since a knock‐in call can be handled as a vanilla call less a knock‐out call, and all options can be treated as call options to exchange one asset for another (refer back to the introductory section of Chapter 11). The idea is to purchase a vanilla call with the same strike and expiration date as the knock‐out being sold and then reduce the cost of creating the knock‐out by selling a basket of vanilla options (this basket may have purchases as well as sales, but the net initial cash flow on the basket is positive to the barrier option seller). The basket of vanilla options must be constructed so that: ■ ■ It has no payoff if the barrier is never hit. In this case, the payout on the barrier option, which has not been knocked out, is exactly offset by the pay‐in from the vanilla call that was purchased, so nothing is left over to make payments on the basket. Its value when the barrier is hit is an exact offset to the value of the vanilla call. When the barrier is hit, you know you will not need to make any payments on the barrier option, so you can afford now to sell the vanilla call you purchased. You do not want to later be vulnerable to payouts on the basket of vanilla options you sold, so you must purchase this basket. In order for cash flows to be zero, the basket purchase price must equal the vanilla call sale price. You can guarantee the first condition by only using calls struck at or above the barrier in the case of a barrier higher than the current price and by only using puts struck at or below the barrier in the case of a barrier lower Managing Exotic Options Risk 389 than the current price. If the barrier is never hit, then you certainly won’t be above the up barrier at expiration, so you won’t owe anything on a call, and you certainly won’t be below the down barrier at expiration, so you won’t owe anything on a put. All three static hedging techniques take advantage of knowing that at the time you are reversing your position in these vanilla options, the underlying must be at the barrier. A useful analogy can be made between these approaches to static hedging and the one we examined for forward‐start options in Section 12.2. For forward‐start options, we purchased an initial set of vanilla options and then had a fixed date on which we would make a single switch of selling our initial package of vanilla options and buying a new vanilla option. For barrier options, we cannot know in advance what the time of the switch will be, but we can know what the forward price of the underlying will be at the time of the switch. As with forward starts, we confine ourselves to one single switch out of the initial vanilla option hedge package. All of these approaches therefore share many of the advantages we saw for the static hedge technique for forward starts: ■ ■ ■ ■ A clear distinction between the portion of expected cost that can be locked in at current market prices of vanilla options (including current volatility surface shape) versus the portion that requires projections of what the volatility surface shape will be at the time of the switch. An estimate of uncertainty for establishing limits and reserves can be based on readily observable historical market data for possible volatility surface shapes. The impact of uncertainty is easy to calculate since it only needs to be computed at one particular point. Future liquidity costs, such as the potential payment of bid‐ask spread, are confined to a single switch. Although it is to be expected that trading desks will, in practice, adjust the static hedge as market circumstances evolve, it remains useful as a risk management technique to evaluate the consequences of an unadjusted hedge. The three approaches differ in how they attempt to ensure that the option package will be equal in value to the vanilla call at the time the barrier is hit. The Derman‐Ergener‐Kani approach (see Derman, Ergener, and Kani 1995) uses a package of vanilla options that expire at different times. The algorithm works backward, starting at a time close to the expiration of the barrier option. If the barrier is hit at this time, the only vanilla options still outstanding will be the vanilla call and the very last option to expire in the package. Since both the underlying price is known (namely, the barrier) and the time to expiry is known, the only remaining factor in determining the 390 FINANCIAL RISK MANAGEMENT values of the vanilla options is the implied volatility, which can be derived from a local or stochastic volatility model (if it is derived from a stochastic volatility model, it will be based on expected values over the probability distribution). Thus, the Derman‐Ergener‐Kani approach can be viewed as the static hedging analog of the dynamic hedging approaches we have been considering. Once the prices of the vanilla options at the time the barrier is hit are calculated, you can easily determine the amount of the option that is part of the basket that needs to be sold in order to exactly offset the sale of the vanilla call with the purchase of the option in the basket. You then work backward time period by time period, calculating the values of all vanilla options if the barrier is hit at this time period and calculating the volume of the new option in the basket that is needed to set the price of the entire basket equal to the price of the vanilla call. At each stage, you only need to consider unexpired options, so you only need to consider options for which you have already computed the volumes held. The following points about the Derman‐Ergener‐Kani approach should be noted: ■ ■ ■ If the barrier is hit in between two time periods for which vanilla options have been included in the package, the results are approximated by the nearest prior time period. The inaccuracy of this approximation can be reduced as much as you want by increasing the number of time periods used. The approach can easily accommodate the existence of drift (dividend rate unequal to risk‐free rate), since a separate computation is made for each time the barrier could potentially be hit. Since the approach relies on the results of a local or stochastic volatility model to forecast future volatility surface levels and shapes, it is vulnerable to the same issue as when these models are used for dynamic hedging—the hedge works only to the extent that the assumptions underlying the model prove to be true. As Derman, Ergener, and Kani state, “The hedge is only truly static if the yield curve, the dividend, and the volatility structures remain unchanged over time. Otherwise, the hedge must be readjusted.” This is illustrated in Table 12.6, which shows the potential mismatch in unwind cost at a period close to expiry based on differences between model‐assumed volatilities and actual volatilities at the time the barrier is hit. Note that the Derman‐Ergener‐Kani approach is vulnerable to model errors relating to both the level of volatility surface and the shape of volatility surface. 391 Managing Exotic Options Risk TABLE 12.6 Unwind Costs of Derman‐Ergener‐Kani Hedge of Barrier Option Strike at‐the‐money, barrier at 95 percent of forward, and three months to expiry. Down‐and‐out call value at initial 20 percent volatility is 3.1955. Unwind with one month to expiry. Volatility at Unwind Unwind Gain or Loss 10.00% 15.00% 20.00% 25.00% 30.00% 0.4479 0.2928 0.0000 –0.3595 –0.7549 The Carr approach (see Carr, Ellis, and Gupta 1998) avoids this dependence on projecting future volatility surfaces and is much simpler to implement, but at a price—it cannot handle volatility skews (though it can handle volatility smiles) and its simplicity depends on the absence of drift (dividend rate equals risk‐free rate). The Carr approach achieves a degree of model independence by using a framework that corresponds directly with the Black‐Scholes equation and determining a hedge package that will work, providing no drift or volatility skew is present. In these circumstances, one can calculate exactly a single vanilla put that will be selling at the same price as the vanilla call in the case that a down barrier is hit. It is based on the principle of put‐call symmetry. In the boxes, we first explain how the principle of put‐call symmetry can be derived from the Black‐Scholes equation and then show how the exact Carr hedges can be derived from put‐call symmetry. PUT‐CALL SYMMETRY The principle of put‐call symmetry says that if you have two strikes, K1 and K2, whose geometric average is the forward price, that is, K1K2 = F , then the current price of a call strike at K1 for expiry T, C(K1,T T), and the current price of a put struck at K2 for the same expiry T, P(K2,T), are related by the equation: C(K1 , T ) / K1 = P(K2 , T ) / K2 392 FINANCIAL RISK MANAGEMENT This formula is a direct and easy consequence of the Black‐Scholes formula. From Hull (2012, Section 17.8), the Black‐Scholes formula for the price of a call and put based on the forward price is: C(K1 , T ) = e− t (FN (( ((l (Fl (FlK llK1) + σ 2T / 2) / σ T ) −K1N((ln(FlK N ((ln(FlK llK1) − σ 2T / 2) / σ T )) P(K2 , T ) = e− t (K2N ((l (K2 / F) + σ 2T / ) / σ T ) −FN ((ln(K2 / F) − σ 2T / ) / σ T )) But since F K2 / F K1K2 , K2 / K1K2 K2 / K1 K1K2 / K1 = F / K1 So, C(K1 , T ) / K1 = e t( K2 N (((l (FlK1) − K1 N((ln(FlK N ((ln(FlK llK1) σ 2T / 2 / 2T / 2)) / T) T )) And substituting F/K1 for K2/F, P(K2 , T ) / K2 = e t ( K2 N (((l (FlK1) 2 T / 2)) / T) − K1 N((ln(FlK N ((ln(FlK llK1) σ 2T / 2) / σ T )) = C(K1 , T ) / K1 Since we have utilized the Black‐Scholes formula in our derivation, this result holds only under the Black‐Scholes assumption of a flat volatility surface for the expiry time T or if the deviation from flat volatility surface is exactly the same at strike K1 and K2. However, since the forward is the geometric average of these two strikes, this is equivalent to saying that one strike is the same percentage above the forward as the percentage the other strike is below the forward. For their volatilities to be equal, the volatility surface must have a smile shape, not a skew shape, using the terminology of Section 11.6.2. Managing Exotic Options Risk 393 DERIVING THE CARR HEDGE Since no drift is present, the forward price is equal to the spot price, which is the barrier level, H. Since the call is struck at K, we can find a reflection strike, R, such that KR = H and, by put‐call symmetry, RCal C l (k) KPut (R). Since K K 2 KR = H, H R = H / K , R H / K , you need to purchase = R H puts struck at H2 / K. For an up barrier, one must separately hedge the intrinsic value and the time value of the vanilla call at the time the barrier is hit. The intrinsic value can almost be perfectly offset by selling binary options that pay 2 × I, where I is the intrinsic value. Any time the barrier is hit, there will be nearly a 50–50 chance that the binary will finish in‐the‐money, so its value is close to 50% × 2 × I = I. In fact, the standard lognormal pricing of a binary results in assuming slightly less than a 50 percent chance of finishing above the barrier, so we need to supplement the binary with I / H of a plain‐vanilla call struck at the barrier. The exact ⎛ σ T⎞ value of the binary is 2 × I × N ⎜ − , and the value of the vanilla 2 ⎟⎠ ⎝ call struck at the barrier, and hence exactly at‐the‐money when the barrier is hit, is ⎛ ⎛σ T⎞ ⎛ σ T ⎞⎞ ( / H) × H ⎜ N ⎜ − N⎜− ⎟ ⎟ 2 ⎟⎠ ⎠ ⎝ ⎝ ⎝ 2 ⎠ ⎛ ⎛ σ T ⎞⎞ = I × ⎜1 − 2 × N ⎜ − ⎟ 2 ⎟⎠ ⎠ ⎝ ⎝ The sum of these two terms is then I. The Carr approach has several advantages: ■ ■ It shows that it is at least plausible to price the barrier based on options with tenor equal to the final tenor of the barrier, indicating that this is probably where most of the barrier’s risk exposure is coming from. Having a large binary component of the hedge is an excellent means of highlighting and isolating the pin risk contained in this barrier that dies in‐the‐money. Techniques we have already developed for managing pin risk on binaries can now easily be brought into play. For example, we could establish a reserve against the pin risk of the binary 394 ■ ■ ■ FINANCIAL RISK MANAGEMENT (see Section 12.1.4). This approach is quite independent of whether the trading desk actually sells a binary as a part of the hedge—the risk of the binary is present in any case. Because the Carr approach uses a small number of options in the hedge package, it is very well suited for developing intuition about how changes in the shape of the volatility surface impact barrier prices. Even if you choose to hedge and price using a dynamic hedging approach, the Carr methodology can still be useful as a liquid proxy. Dynamic hedging can be employed for the difference between the barrier and the static hedge determined by the Carr approach. By choosing an initial hedge that, on theoretical grounds, we expect to be close to a good static hedge, we expect to minimize the degree to which changes in option hedges are required. However, by using dynamic hedging, we allow for as much protection as the accuracy of the model provides against uncertainty in skew and drift. Neither the presence of volatility smiles nor the uncertainty of future volatility smiles impacts the Carr approach. Since it deals with options that are symmetrically placed relative to the at‐the‐money strike, all smile effects cancel out. The simplicity of the Carr approach is lost in the presence of drift or volatility skew. See the appendix to Carr and Chou (1996) for a method of using a large number of vanilla options to create a volatility‐independent static hedge of barrier options in the presence of drift. See Carr (2001) for a method of handling volatility skew. To appreciate how the Carr model performs and to gain the benefit of its insight into the risk structure of barriers, you should study the CarrBarrier spreadsheet provided on the website for this book. The spreadsheet shows the hedge structure for all eight possible simple barrier structures and the result of the barrier unwind for a specified scenario. Exercise 12.3 guides you through some sample runs. Here are some of the points you should be looking for: ■ ■ The one common element in all eight variants is the use of the reflection option—the one that utilizes the principle of put‐call symmetry. It captures the time value of the barrier option at the point the barrier is hit. The sample run displayed in Table 12.7 shows that on unwind, for the down call and up put cases, the reflection option exactly offsets the value of the option that needs to be purchased for the in cases and needs to be sold for the out cases. For the up call and down put cases, a binary piece also needs to be offset, but the reflection option offsets the entire time value. In Table 12.8, in which the only change from Table 12.7 is Managing Exotic Options Risk ■ ■ ■ ■ 395 that the volatility at unwind has been raised, the binary piece (the sum of the binary and binary correction) is unchanged from Table 12.7, but the time value has increased exactly equally for the vanilla option and the reflection option. The time value when the barrier is hit depends on how far the barrier is from the strike. In the Table 12.7 example, the up barrier of 110 is further from the 100 strike than the 95 down barrier is, so the up reflection options have far less value than the down reflection options. You can think of the reflection option as taking value away from the out option and transferring it to the in option. The up call and down put cases are ones with binary components, since these in options will begin life already in‐the‐money and these out options cause an in‐the‐money component to be extinguished. The size of the binary component at the time the barrier is hit is the exact difference between the strike and barrier. It is divided into two pieces: the principal piece is the binary option and the secondary piece is the vanilla option used to supplement the binary. The total value of these two components at initiation will be less than the potential value on hitting the barrier, precisely reflecting the (risk‐neutral) probability that the barrier will be hit. By trying different values for barrier‐hitting scenarios, you will see that as long as volatility skew and drift are both equal to zero, the total impact of buys and sells in all eight cases is always zero. That is, the hedge works perfectly regardless of the assumptions made as to the time remaining when the barrier is hit, the at‐the‐money volatility, the volatility smile, or the risk‐free rate. However, if either drift or volatility skew differs from zero, gains and losses will occur when the barrier is hit, varying by case. Examples are shown in Tables 12.9 and 12.10. It would clearly be a relatively easy task to calculate the size of potential losses based on assumptions about how adverse drift and skew could be at different possible times the barrier is hit. This could serve as input for the determination of reserves and limits. When the initial volatility skew, volatility smile, and drift are set equal to zero, pricing given by the standard analytic formula for barriers (shown on the top line in each column) exactly equals the total creation cost of the Carr hedges, as can be seen from the zero on the line labeled “difference.” When any of these values is different from zero, the Carr hedge gives a different value than the analytic formula. For example, Tabe 12.11 shows a case that corresponds to the one analyzed in Table 12.5, showing a 3.104 value for the up‐and‐out call in the presence of a volatility skew compared with a 2.7421 value using the analytic formula. Note that the presence of volatility skew (or drift) in the initial conditions does not imply that the Carr hedge will 396 FINANCIAL RISK MANAGEMENT not work. Only conditions at the time the barrier is hit determine the efficiency of the hedge. In Exercise 12.4 you will run a Monte Carlo simulation of the cost of a hedging strategy that hedges a barrier option with the Carr hedge, utilizing the spreadsheet CarrBarrierMC. A more general approach to static hedging that can handle all drift and volatility shape conditions is optimization, in which a set of vanilla options is chosen that fits as closely as possible the unwind of the barrier option at different possible times, drifts, volatility levels, and volatility surface shapes that may prevail when the barrier is hit. The optimization approach is discussed in Dembo (1994). Often no perfect static hedge can be found, but in these cases the optimization produces information on the distribution of possible hedge errors that can be useful input for determining a reasonable reserve. A similar approach can be taken to many different types of exotic structures. The OptBarrier spreadsheet illustrates how optimization can be used to find a static hedge for a barrier option. If the possible conditions when the barrier is hit are restricted to zero drift and volatility smile but no skew, then the Excel Solver will find a set of vanilla options that almost exactly matches the barrier unwind for all volatility levels and times to expiry (although the particular set of hedges chosen may lack the clarity of insight that the Carr hedges offer). Of course, this is not a surprise since we know from the Carr approach that a perfect static hedge is possible under these circumstances. When different nonzero drift and volatility skew conditions are allowed, the match of the barrier unwind is no longer as exact. The spreadsheet determines how much this slippage can be across all the specified cases of hitting time, skew, and drift. As with the Carr approach, this information can then be used to set reserves and limits. The difference from the Carr approach is the objective to find a hedge that minimizes the amount of this slippage. Exercise 12.3 guides you through some sample runs. As a concluding note, observe that there is a lower limit on the uncertainty of unwind costs for any static hedging approach. Any dynamic hedging model can be used to compute the unwind cost of a selected static hedging strategy. So any difference in the pricing of barrier options between different dynamic hedging models translates into uncertainty of unwind costs. Practical experience with dynamic hedging models shows that differences in assumptions (for example, stochastic volatility versus local volatility and the frequency of jumps) give rise to substantial differences in barrier options prices utilizing the same input for current vanilla options prices. So you can search for static hedges that minimize the uncertainty of unwind costs, but an irreducible uncertainty will always remain that can be controlled only through limits and reserves. Static hedging greatly simplifies the calculations needed for limits and reserves. 397 100 110 95 Forward Up forward Down forward CDO 3.1955 –3.9878 0 0 0.7923 –3.1955 0.0000 90.25 –1.8881 0 0 1.8881 0.0000 Vanilla Digital Correct dig Reflect Total Difference Reflect point At barrier Vanilla Digital Correct dig Reflect Total Carr Static Hedge 0.25 0.00% 0.00% 20.00% 0.00% 0.00% Time to expiry Rate Drift ATM volatility Vol smile Vol skew TABLE 12.7 100.00 100.00 110.00 95.00 0.25 0.00% 0.00% 20.00% 0.00% 0.00% Price Strike Up barrier Down barrier Time to expiry Rate Drift ATM volatility Vol smile Vol skew 1.8881 0 0 –1.8881 0.0000 CDI 0.7923 0 0 0 –0.7923 –0.7923 0.0000 90.25 –10.9539 9.6012 0.3988 0.9539 0.0000 CUO 0.6343 –3.9878 3.1581 0.0867 0.1087 –0.6343 0.0000 121 10.9539 –9.6012 –0.3988 –0.9539 0.0000 CUI 3.3535 0 –3.1581 –0.0867 –0.1087 –3.3535 0.0000 121 –6.8881 5.1994 –0.1994 1.8881 0.0000 PDO 0.0778 –3.9878 3.2171 –0.0994 0.7923 –0.0778 0.0000 90.25 6.8881 –5.1994 0.1994 –1.8881 0.0000 PDI 3.9100 0 –3.2171 0.0994 –0.7923 –3.9100 0.0000 90.25 –0.9539 0 0 0.9539 0.0000 PUO 3.8791 –3.9878 0 0 0.1087 –3.8791 0.0000 121 0.9539 0 0 –0.9539 0.0000 PUI 0.1087 0 0 0 –0.1087 –0.1087 0.0000 121 398 TABLE 12.8 Forward Up forward Down forward –5.5195 0 0 5.5195 0.0000 CDO 3.1955 –3.9878 0 0 0.7923 –3.1955 0.0000 90.25 5.5195 0 0 –5.5195 0.0000 CDI 0.7923 0 0 0 –0.7923 –0.7923 0.0000 90.25 Carr Static Hedge with Higher Volatility at Unwind 100 110 95 At barrier 0.25 Vanilla 0.00% Digital 0.00% Correct dig 40.00% Reflect 0.00% Total 0.00% Time to expiry Rate Drift ATM volatility Vol smile Vol skew Vanilla V Digital Correct dig Reflect T Total Difference Reflect point 100.00 100.00 110.00 95.00 0.25 0.00% 0.00% 20.00% 0.00% 0.00% Price Strike Up barrier Down barrier Time to expiry Rate Drift ATM volatility Vol smile Vol skew –14.2920 9.2034 0.7966 4.2920 0.0000 CUO 0.6343 –3.9878 3.1581 0.0867 0.1087 –0.6343 0.0000 121 14.2920 –9.2034 –0.7966 –4.2920 0.0000 CUI 3.3535 0 –3.1581 –0.0867 –0.1087 –3.3535 0.0000 121 –10.5195 5.3983 –0.3983 5.5195 0.0000 PDO 0.0778 –3.9878 3.2171 –0.0994 0.7923 –0.0778 0.0000 90.25 10.5195 –5.3983 0.3983 –5.5195 0.0000 PDI 3.9100 0 –3.2171 0.0994 –0.7923 –3.9100 0.0000 90.25 –4.2920 0 0 4.2920 0.0000 PUO 3.8791 –3.9878 0 0 0.1087 –3.8791 0.0000 121 4.2920 0 0 –4.2920 0.0000 PUI 0.1087 0 0 0 –0.1087 –0.1087 0.0000 121 399 100 110 95 Forward Up forward Down forward CDO 3.1955 –3.9878 0 0 0.7923 –3.1955 0.0000 90.25 –1.9757 0 0 1.8010 –0.1746 Vanilla V Digital Correct dig Reflect Total Difference Reflect point At barrier Vanilla Digital Correct dig Reflect Total 1.9757 0 0 –1.8010 0.1746 CDI 0.7923 0 0 0 –0.7923 –0.7923 0.0000 90.25 Carr Static Hedge with Nonzero Skew at Unwind 0.25 0.00% 0.00% 20.00% 0.00% 10.00% Time to expiry Rate Drift ATM volatility Vol smile Vol skew TABLE 12.9 100.00 100.00 110.00 95.00 0.25 0.00% 0.00% 20.00% 0.00% 0.00% Price Strike Up barrier Down barrier Time to expiry Rate Drift ATM volatility Vol smile Vol skew –10.8303 9.2028 0.3988 1.0830 –0.1457 CUO 0.6343 –3.9878 3.1581 0.0867 0.1087 –0.6343 0.0000 121 10.8303 –9.2028 –0.3988 –1.0830 0.1457 CUI 3.3535 0 –3.1581 –0.0867 –0.1087 –3.3535 0.0000 121 –6.9757 5.0002 –0.1994 1.8010 –0.3739 PDO 0.0778 –3.9878 3.2171 –0.0994 0.7923 –0.0778 0.0000 90.25 6.9757 –5.0002 0.1994 –1.8010 0.3739 PDI 3.9100 0 –3.2171 0.0994 –0.7923 –3.9100 0.0000 90.25 –0.8303 0 0 1.0830 0.2527 PUO 3.8791 –3.9878 0 0 0.1087 –3.8791 0.0000 121 0.8303 0 0 –1.0830 –0.2527 PUI 0.1087 0 0 0 –0.1087 –0.1087 0.0000 121 400 100 109.178086 94.2901652 Forward Up forward Down forward CDO 3.1955 –3.9878 0 0 0.7923 –3.1955 0.0000 90.25 –1.6691 0 0 2.1120 0.4429 Vanilla Digital Correct dig Reflect Total Difference Reflect point At barrier Vanilla Digital Correct dig Reflect Total 1.6691 0 0 –2.1120 –0.4429 CDI 0.7923 0 0 0 –0.7923 –0.7923 0.0000 90.25 Carr Static Hedge with Nonzero Drift at Unwind 0.25 0.00% –3.00% 20.00% 0.00% 0.00% Time to expiry Rate Drift ATM volatility Vol smile Vol skew TABLE 12.10 100.00 100.00 110.00 95.00 0.25 0.00% 0.00% 20.00% 0.00% 0.00% Price Strike Up barrier Down barrier Time to expiry Rate Drift ATM volatility Vol smile Vol skew –10.2694 9.0052 0.3610 0.8243 –0.0789 CUO 0.6343 –3.9878 3.1581 0.0867 0.1087 –0.6343 0.0000 121 10.2694 –9.0052 –0.3610 –0.8243 0.0789 CUI 3.3535 0 –3.1581 –0.0867 –0.1087 –3.3535 0.0000 121 –7.3790 5.4974 –0.2179 2.1120 0.0125 PDO 0.0778 –3.9878 3.2171 –0.0994 0.7923 –0.0778 0.0000 90.25 7.3790 –5.4974 0.2179 –2.1120 –0.0125 PDI 3.9100 0 –3.2171 0.0994 –0.7923 –3.9100 0.0000 90.25 –1.0913 0 0 0.8243 –0.2671 PUO 3.8791 –3.9878 0 0 0.1087 –3.8791 0.0000 121 1.0913 0 0 –0.8243 0.2671 PUI 0.1087 0 0 0 –0.1087 –0.1087 0.0000 121 401 100 120 90 Forward Up forward Down forward CDO 3.9244 –3.9878 0 0 0.1266 –3.8611 0.0633 81 –0.7124 0 0 0.7124 0.0000 Vanilla Digital Correct dig Reflect Total Difference Reflect point At barrier Vanilla Digital Correct dig Reflect Total 0.7124 0 0 –0.7124 0.0000 CDI 0.0633 0 0 0 –0.1266 –0.1266 –0.0633 81 Carr Static Hedge with Nonzero Skew at Initiation 0.25 0.00% 0.00% 20.00% 0.00% 0.00% Time to expiry Rate Drift ATM volatility Vol smile Vol skew TABLE 12.11 100.00 100.00 120.00 90.00 0.25 0.00% 0.00% 20.00% 0.00% –10.95% Price Strike Up barrier Down barrier Time to expiry Rate Drift ATM volatility Vol smile Vol skew –20.1473 19.2024 0.7976 0.1473 0.0000 CUO 2.7421 –3.9878 0.8708 0.0130 0.0000 –3.1040 –0.3619 144 20.1473 –19.2024 –0.7976 –0.1473 0.0000 CUI 1.2457 0 –0.8708 –0.0130 0.0000 –0.8838 0.3619 144 –10.7124 10.3988 –0.3988 0.7124 0.0000 PDO 0.8479 –3.9878 3.7355 –0.0935 0.1266 –0.2192 0.6287 81 10.7124 –10.3988 0.3988 –0.7124 0.0000 PDI 3.1399 0 –3.7355 0.0935 –0.1266 –3.7686 –0.6287 81 –0.1473 0 0 0.1473 0.0000 PUO 3.9874 –3.9878 0 0 0.0000 –3.9878 –0.0003 144 0.1473 0 0 –0.1473 0.0000 PUI 0.0003 0 0 0 0.0000 0.0000 0.0003 144 402 FINANCIAL RISK MANAGEMENT 12.3.4 Barrier Options with Rebates, Lookback, and Ladder Options We will show how to use barrier options to create a static hedge for barrier options with rebates, lookback, and ladder options. Thus, we can transfer the techniques we have studied for using vanilla options to represent and hedge barrier option positions to create vanilla option representations and hedges of barrier options with rebates, lookback, and ladder options. The use of a rebate feature in a barrier option can be regarded as a binary option triggered by a barrier. For example, suppose you have a down‐ and‐out call that pays a rebate of $2 million if the down barrier is hit and the call is canceled. This can be viewed as the sum of a down‐and‐out call with no rebate and a down‐and‐in binary option that pays $2 million if the barrier is hit. However, since a binary option can be represented by being long one vanilla call and short another vanilla call, as discussed in Section 12.1.4, a down‐and‐in binary can also be treated as being long one down‐ and‐in call and short another down‐and‐in call. So the rebate can be hedged and valued through the methodology we have already developed for barriers without rebates. Lookback options come in two varieties: those that pay the difference between the maximum price that an asset achieves during a selected period and the closing price and those that pay the difference between the maximum price that an asset achieves during a selected period and a fixed strike. Symbolically, the lookback pays either Smax – ST or max (0, Smax – K). We can reproduce the payoffs of a lookback of the first type exactly by buying a lookback of the second type with a strike equal to the current price of the asset (S0), selling the asset forward to time T, and buying a forward delivery of S0 dollars at time T. Since Smax is certainly ≥ S0, max(0, Smax – S0) = Smax – S0, the total payoff of this combination at time T is: max (0, Smax – S0) – ST + S0 = (Smax – S0) – ST + S0 = Smax – ST (12.4) So if we can hedge the second type of lookback option by static hedging with barriers, we can create the first type of lookback option by static hedging with barriers as well. Lookback options have a closely related product called ladder options that pay max (0, Smax – K) rounded down by a specified increment. For example, if K = 100 and Smax = 117.3, the lookback call of the second type would pay 17.3, a ladder with increments of 1 would pay 17, a ladder with increments of 5 would pay 15, and a ladder with increments of 10 would pay 10. Since a lookback call can be approximated as closely as we want by a ladder with a small enough increment, it is sufficient to show how to statically hedge a ladder with barriers. Managing Exotic Options Risk 403 It is easy to create a static hedge for a ladder option using up‐and‐in binary options. For each ladder rung, you buy an up‐and‐in binary option of the same tenor that pays the increment conditional on the rung being breached at some point during the life of the option. For example, if K = 100 and we have a ladder with increments of 5, we buy an up‐and‐in binary option having a payoff of 5 and a barrier of 105, another with a payoff of 5 and a barrier of 110, and so on. If the highest level the underlying reaches during the life of the ladder option is 12, then 10 will be owed on the ladder option, but the binary up‐and‐ins with barriers of 105 and 110 will both have been triggered for a payment of 5 + 5 = 10. 12.3.5 Broader Classes of Path‐Dependent Exotics Now that we have looked at several dynamic hedging and static hedging alternatives for managing risk on standard barrier options, we want to examine how these approaches can be generalized to the full universe of single‐asset exotics. We will focus most of our attention on double barriers and partial‐time barriers, since these are reasonably popular products and since any techniques that are flexible enough to handle these variants would be flexible enough to handle any product. Double barriers knock out (or knock in) if either a higher or a lower barrier is crossed. An example would be a one‐year call option struck at 100 that knocks out if the price during the year is ever either above 120 or below 80. Partial‐time barriers have a restricted time period during which the barrier provision applies. An example would be a one‐year call option struck at 100 that knocks out if the price is below 90 any time between the end of month 3 and the end of month 9. If the price goes below 90 prior to month 3 but then goes back above 90 by the end of month 3, no knock‐out occurs. Similarly, if the first time the price goes below 90 is after month 9, no knock‐out occurs. The greatest flexibility is offered by dynamic hedging, using either local volatility or stochastic volatility models, and by the Derman‐Ergener‐Kani approach to static hedging. Both can be easily generalized to double barriers and partial‐time barriers. Local volatility models that solve for the exotic option values on a tree constructed to fit vanilla option prices can be easily adapted to solve for virtually any set of payoffs. Stochastic volatility models, which may require Monte Carlo simulation solutions, can easily handle any deterministic payout. The Derman‐Ergener‐Kani static hedging algorithm can solve for hedge packages that give zero unwind costs for double barriers and partial‐time barriers just as easily as for standard barriers. The DermanErgenerKaniDoubleBarrier and DermanErgenerKaniPartialBarrier spreadsheets illustrate this computation. An interested reader could use 404 FINANCIAL RISK MANAGEMENT these spreadsheets as a guide to program a general calculator for applying the Derman‐Ergener‐Kani method to more complex barriers. The drawbacks of dynamic hedging and Derman‐Ergener‐Kani static hedging that we analyzed for standard barriers apply in a more general setting as well. It will still be difficult to project the potential effects of hedge slippage for dynamic hedging. This is a heightened concern for double barriers since they have a reputation among exotics traders as particularly treacherous to dynamically hedge since they are almost always threatening to cross one barrier or the other. The dependence of Derman‐Ergener‐Kani on the model used to calculate the hedge ratios, and hence its vulnerability to being wrong about future volatility levels, remains true for the expanded product set. Peter Carr and his collaborators have done a lot to expand the applicability of his static hedging approach beyond standard barriers. In particular, Carr, Ellis, and Gupta (1998, Section 3.1) have developed a static hedge for double barriers, and Carr and Chou (1997) have developed a static hedge for partial‐time barriers. Similar results are presented in Andersen and Andreasen (2000). These hedges offer one of the major advantages of the Carr hedge for standard barriers—protection against shifts in volatility levels. However, they do not offer another major advantage of the Carr hedge for standard barriers: They are not simple to compute and do not provide much intuitive insight into the risk structure of the exotic being hedged. The specialized nature of each construction does not offer significant guidance as to how to build hedges for other exotics. Optimal fitting would seem to offer the best hope for an easy‐to‐ generalize static hedge that will minimize sensitivity to model assumptions. However, unlike the Derman‐Ergener‐Kani method, which automates the selection of the vanilla options to be used in hedging a particular exotic, the optimal fitting approach relies on practitioner insight to generate a good set of hedge candidates. A poor choice of possible hedges results in a poorly performing static hedge. A possible solution is to try to generalize the Derman‐Ergener‐Kani approach to fit to a range of volatility surfaces rather than to a single one. Some promising results along these lines have been obtained by Allen and Padovani (2002, Section 6). A copy of this paper is on the book website. 12.4 CORRELATION‐DEPENDENT OPTIONS Valuation and hedging strategies for derivatives whose payoff is a function of more than one underlying asset are critically dependent on assumptions about correlation between the underlying assets. With only a few exceptions Managing Exotic Options Risk 405 (which are discussed in Section 12.4.3), there is an absence of sufficiently liquid market prices to enable implied correlations to be inferred in the way implied volatilities can be derived from reasonably liquid prices of vanilla options. So much of the focus of risk management for these derivatives revolves around controlling the degree of exposure to correlation assumptions and building reserves and limits against the differences between actual realized and estimated correlations. An important distinction within derivatives with multiasset payoffs should be made between those whose payoff is based on a linear combination of asset prices (for example, the average of a set of prices or the difference between two prices) and those whose payoff is based on a nonlinear combination of asset prices (for example, the maximum of a set of prices or the product of two prices). When the payoff is based on a linear combination of asset prices, risk management is considerably simpler, even if the payoff itself is a nonlinear function of the linear combination of asset prices, such as an option on the average of a set of prices. We therefore discuss these two types of derivatives in separate sections. A final section discusses options that depend on a different type of correlation—the correlation between underlying asset value and the probability of option exercise. 12.4.1 Linear Combinations of Asset Prices Derivatives whose payoff depends on a linear combination of asset prices share several important characteristics that simplify their risk management: ■ ■ ■ If the payoff function is a linear function of the linear combination of asset prices, then the derivative does not have any option characteristics and can be perfectly hedged with a static portfolio of the underlying assets. In such cases, the valuation of the derivative is independent of correlation assumptions. This is not true of derivatives whose payoff function is a linear function of a nonlinear combination of asset prices, such as a forward based on the product of an asset price and an FX rate (a so‐called quanto) that requires dynamic hedging. Even when the payoff function is a nonlinear function of the linear combination of asset prices, such as an option on the average of a set of prices, and therefore requires dynamic hedging, the rules for dynamic hedging are particularly simple to calculate. Even when dynamic hedging is required, it is often possible to make very good approximations of valuation and the risk of incorrect correlation assumptions using a standard Black‐Scholes model. 406 FINANCIAL RISK MANAGEMENT We will examine each of these characteristics more closely. We will then make use of the approximation technique discussed previously to answer questions about how the risk of these derivatives should be managed. 12.4.1.1 Derivatives Whose Payoffs Are Linear Functions of Linear Combinations of Asset Prices In principle, any derivative whose payoff is a linear function of a linear combination of asset prices, such as a forward on the average price of a basket of assets, can be statically hedged by buying the properly weighted basket of forwards. In practice, this could be operationally difficult for a basket composed of a very large number of assets, and a market maker may choose to hedge with a differently weighted basket selected to statistically track the derivative payoff closely, with a resulting possibility of tracking error. However, in either case, the performance of this hedging strategy will not be influenced by the level of correlations of assets within the basket. In particular, the valuation of a basket should not be influenced by whether the assets in the basket are well diversified or highly concentrated. Both well‐diversified and highly concentrated baskets should be valued as the weighted average of the valuations of the individual components. At first, this may seem to violate intuition, since firms devote considerable resources to calculations such as value at risk (VaR) that rate highly concentrated baskets as riskier than well‐diversified baskets. Shouldn’t some penalty in valuation be applied for an asset basket that carries more risk? The answer from capital market theory is that only systemic risk, which is not capable of being diversified away, should be penalized and that the role of tools such as VaR is to make certain that a firm has considered the proper hedges against risk that can be diversified away. So a trader entering into a forward on the average price of a basket will be charged a higher risk premium by his firm’s risk systems for running an open position (that is, not putting in place the basket hedge) in a highly concentrated basket than in a well‐diversified basket. But in either case, he has the ability to put on the hedge closing out the position, so concentration should only play a role in the evaluation of the risk of running an open position, not in the valuation of the derivative. A particularly clear discussion of this point can be found in Varian (1987, “Value Additivity Theorem”). As Varian emphasizes, this principle only applies as long as payoffs are linear and ceases to apply when payoffs are nonlinear. This is true both for nonlinearity of the payoff function, such as an option on the average price of a basket of stocks, and the nonlinearity of a combination of asset prices, such as a forward on the maximum price of a set of stocks. As soon as nonlinearity is introduced, considerations that only play a minor role in the risk assessment of linear products begin to play a role in valuation. For example, the probability of extreme tail events based on the correlation of Managing Exotic Options Risk 407 default probabilities plays no role in the valuation of a CDO based on a basket of loans and/or bonds so long as the CDO divides ownership of the basket proportionally. (A CDO is an example of an asset‐backed security; see Section 10.1.8.) However, CDOs often divide the ownership of the basket into tranches, with some tranches paying all credit losses up to a certain level and other tranches paying only losses above that level. This enables the investor market to be segregated more efficiently by creating some bonds that are tailored to investors seeking lower credit risk and other bonds that are tailored to investors willing to take on more credit risk in return for adequate compensation. Tranching CDOs introduces nonlinearity of payoffs. As a result, valuation is dependent on the probability of extreme tail events based on the correlation of default probabilities. For further discussion of this point, see Section 13.4.1. A second point to note is that the arbitrage principle only applies if the assets comprising the basket are sufficiently liquid. If not, investors who would have a hard time acquiring a diversified basket of assets may be willing to pay a premium to receive a payment on an index based on the average price of such a basket. This offers a profit opportunity to market makers who can efficiently acquire diverse baskets that other market participants would find difficult to replicate. The market maker can then offer to pay an index based on its earnings on the basket and build a premium into the index. This diversification premium has definitely been observed in the default swaps market. 12.4.1.2 Rules for Dynamic Hedging The required dynamic hedges for an option on a linear combination of asset prices are very easy to determine. Standard deltas can be derived from option pricing models, and the delta hedge can then be formed by multiplying this delta times the linear weights of each asset in the basket. This simplifies ongoing hedging calculations and the calculation of required hedges in Monte Carlo simulations of hedging strategies. Consider an at‐the‐money one‐year option on a 5,000‐share stock basket consisting of 20 percent IBM, 45 percent General Electric (GE), and 35 percent Merck. If the volatility of the basket is assumed to be 25 percent, the delta, using the Black‐Scholes formula, is 55 percent. The hedge should be: 5,000 × 55% × 20% = 550 shares of IBM 5,000 × 55% × 45% = 1,237.5 shares of GE 5,000 × 55% × 35% = 962.5 shares of Merck (12.5) 12.4.1.3 Approximation of Option Values The calculation of the value of an option on a linear combination of asset prices can be reasonably approximated by calculating the volatility of the underlying basket based on the weights 408 FINANCIAL RISK MANAGEMENT of each asset in the basket, the implied volatilities of each asset, and the assumed correlations between assets. This calculated volatility can then be used as input to the Black‐Scholes formula for the basket option. Continuing the previous example, assume that the volatility of IBM stock is 30 percent, the volatility of GE stock is 33 percent, and the volatility of Merck stock is 28 percent, with correlations between IBM and GE of 60 percent, between IBM and Merck of 50 percent, and between GE and Merck of 40 percent. Then the volatility of the basket can be estimated as: SquareRoot [(20% × 30%)2 + (45% × 33%)2 + (35% × 28%)2 + 2 × (20% × 30% × 45% × 33% × 60% + 20% × 30% × 35% × 28% × 50% + 45% × 33% × 35% × 28% × 40%)] = 25.2% (12.6) This is only an approximation for two reasons. The first reason is that the representation of an asset’s distribution by a single implied volatility is only accurate if the implied volatility surface for that option is flat, that is, the same at all strike prices. However, as discussed in Section 11.6.2, this is rarely the case. The second reason is that even if we had an example in which the implied volatility surfaces of the options on all the individual assets were flat, meaning that the market was pricing them all as if they were lognormally distributed, a linear combination of lognormal distributions is not lognormal, so the implied volatility surface for the basket option would not be flat and thus could not be represented by a single volatility. For assets with reasonably flat implied volatility surfaces, this approximation technique will give accurate enough results to be useful as a way of building intuition about the degree to which basket option prices depend on the implied volatilities of the individual assets and on the assumed correlations between them. This is how we will make use of this approximation in the remainder of this section. Actual valuations require more accurate numerical techniques. In practice, two are generally used. One technique is a Monte Carlo simulation in which each asset process is specified by a full distribution that corresponds to the implied volatility surface for that asset, following the approach discussed in Section 12.3.2. Assumed correlations between assets can be enforced by the technique discussed in Hull (2012, Section 26.7). This technique is flexible enough to support more complex assumptions, such as correlations that vary based on the price level or price movement of the component assets. Finally, the value of the basket can be computed along each sample path and the resulting value of the option can be calculated. The flexibility to have correlation vary with price level or price movement can be important since large downward price moves tend to be Managing Exotic Options Risk 409 accompanied by higher correlation than ordinary price moves. This can result in baskets being priced at higher volatility skews than individual components of the basket since it increases correlation and hence increases volatility at lower price levels. For further discussion of this point, see Derman and Zou (2001). The Monte Carlo approach affords great flexibility, including the incorporation of stochastic volatility and price jump assumptions. Its drawback is difficulty in valuing American‐style options that require the determination of optimal early exercise strategies. Further developments in Monte Carlo modeling do allow approximations of American option valuation; see, for example, Broadie, Glasserman, and Jain (1997). The alternative approach for American‐style options on baskets is the three‐dimensional tree approach described in Hull and White (1994). This approach enables the combination of two trinomial trees that have been fitted to the full implied volatility surface, using the techniques discussed in Section 12.3.2, to be combined into a single tree based on assumed correlations, which can vary by node. Basket values can then be computed on the combined tree and option values determined by working backwards on the tree. This approach has the advantage of greater precision in determining early exercise strategies. The disadvantages are that it is only computationally feasible for baskets involving two assets and it is restricted to using local volatility models to replicate the implied volatility surface, which lacks the flexibility to incorporate stochastic volatility or price jumps. A possible combination of the two methods for more than two assets would be to determine the option price for the final exercise using the more precise Monte Carlo method and estimating the extra value due to possible early exercise using the three‐dimensional tree technique using the first two principal components of the assets as the two variables to be modeled on the tree. 12.4.2 Risk Management of Options on Linear Combinations We will now take advantage of the simple formula available to approximate the value of an option on a linear combination of assets to examine how risks arising from positions in these options should be managed. One possible risk management technique is pure dynamic hedging of options positions in a particular linear combination. This is operationally straightforward, as discussed in Section 12.4.1.2. However, it encounters the same deficiencies of reliance on the delta‐hedging strategy that we discussed in Section 11.1. The same arguments favoring the use of other options in hedging that were given in Section 11.1 apply, but it is unusual to find any liquidity in options on asset combinations. This suggests the use of options on individual assets comprising the basket as part of the hedge. 410 TABLE 12.12 FINANCIAL RISK MANAGEMENT The Impact of Hedging Basket Options with Single‐Stock Options Dynamically hedge with underlying stocks only Purchase at‐the‐money options on stocks A and B and then dynamically hedge Standard Deviation Transaction Costs 28.7% 2.3% 14.0% 1.9% Consider the following simple example. An option has been written on the average of two assets, A and B. Compare the simulation results of a pure dynamic hedge with the underlying stocks with the simulation results of a hedge that involves first purchasing options on assets A and B and then dynamically hedging the resulting position with the underlying stocks. Suppose a one‐year at‐the‐money option has been written on the average of the prices of two stocks, A and B. Assume that both A and B have 20 percent volatility on average with a 33 percent standard deviation of volatility and that correlation between the two assets averages 0 percent with a 33 percent standard deviation. We will simulate two hedging strategies: Use a pure dynamic hedge with the underlying stocks, or first purchase an at‐the‐ money option on A and an at‐the‐money option on B and then dynamically hedge the resulting position with the underlying stocks. The ratio of the notional of purchased options on individual stocks to the notional of the sold basket option we will use is 70 percent, split equally between the option on A and the option on B. This 70 percent ratio is suggested by the average volatility of the basket option being ( % × 20%)2 + ( % × 20%)2 = 14. %, which is just a little bit more than 70 percent of the 20 percent average volatility of the individual stocks. Simulation starting with different ratios of individual stock options to the basket options confirms that 70 percent is the ratio that results in the lowest standard deviation of the dynamic hedging results. Table 12.12 compares the results between the two hedging strategies. Although a substantial reduction in uncertainty and transaction costs results from utilizing an option in the constituent stocks as a hedge, it is not as large a reduction as was shown for hedging vanilla options with vanilla options at other strikes in Table 11.2. Even if we were certain of the correlation, the static hedge utilizing the purchase of at‐the‐money options on stocks A and B can only reduce the standard deviation to 12.2 percent. The intuitive reason for this is that the relationship of one strike being located midway between two other strikes is obviously stable, whereas the underlying stock options can move into or out of the money without a Managing Exotic Options Risk 411 similar move on the part of the basket option. For example, if stock A’s price rises by 20 percent and stock B’s price falls by 20 percent, the previously at‐the‐money call options on stock A and B will now be substantially in‐the‐money and out‐of‐the‐money, respectively. In both cases, their sensitivity to volatility will be considerably reduced from the time of initiation. This is not true for the basket option, which will still have its same initial sensitivity to volatility since it is still at‐the‐money relative to the average price of A and B. A possible remedy would be to dynamically change the amount of single stock options being used to hedge in response to changes in relative volatility sensitivity of the basket option and single stock options. This has many similar virtues and drawbacks with the proposal to dynamically hedge barrier options with vanilla options that was considered in Section 12.3.2. One advantage in this case is that it is considerably easier to calculate the required option hedges in the Monte Carlo simulation, provided you are willing to accept the degree of approximation of the simple formula. Whether employing static hedging or dynamic hedging with single‐asset options, the following rules should apply: ■ ■ ■ ■ Any residual exposure to the uncertainty of correlation should be reflected in reserve policies and limits, since this is an exposure that cannot be hedged with liquid instruments. Residual unhedgeable exposure to the uncertainty of single‐asset volatility should be quantified, as shown in the Monte Carlo example in Table 12.12, and reflected in reserve policies and limits. Valuation procedures and risk measurement should be in agreement. If implied volatilities of individual assets are used as an input to the valuation of a basket option, then the exposure to changes in each constituent asset’s implied volatility should be reflected, either statically or dynamically, in price‐vol matrix reports and other volatility exposure measures computed for the individual asset. Similarly, delta exposure should be reflected in individual underlying asset position reports. If this principle is not followed, valuation exposure to changes in the price or volatility of an asset can grow without control by being included in more and more basket products. In some cases, individual asset volatility may be so slight a contribution to the risk of a basket option that it is not worth the effort of utilizing the implied volatility as an input to valuation or reflecting exposure to volatility changes in individual asset risk reports. The basket option will then effectively be managed as if it was an option on a separate underlying unrelated to the single‐asset options. Note that this does not change the use of the individual underlying to perform delta hedging. 412 FINANCIAL RISK MANAGEMENT TABLE 12.13 Sensitivities of Option on Basket Correlation Level 1% Shift in Volatilities 10% Shift in Correlation 90% 75% 50% 25% 0% –25% –50% –75% –90% –95% –98% 0.97% 0.94% 0.87% 0.79% 0.71% 0.61% 0.50% 0.35% 0.22% 0.16% 0.10% 0.51% 0.53% 0.57% 0.62% 0.69% 0.79% 0.95% 1.30% 1.85% 2.31% 2.90% The BasketOption spreadsheet on the website for this book shows the calculation of basket option exposures to changes in correlation and individual asset volatility under the approximation of the simple formula. Table 12.13 shows some sample results for an equally weighted two‐asset basket with both assets having a 20 percent volatility. The impacts shown are for a 1 percent shift in the volatilities of both assets (for example, 20% + 1% = 21%) and a 10 percent shift in correlation (for example, 75% + 10% = 85%). Note how the relative contribution of individual stock volatility relative to correlation declines sharply as correlation levels become negative. This is very relevant for options on the spread between two asset prices, since the hedge basket then consists of a positive position in one asset and a negative position in the other. If the assets are strongly correlated, their positions in the basket will show high negative correlation. In these cases, hedging the individual option volatilities is questionable. One reporting issue for all multiasset derivatives is whether to take correlation into account when reporting delta and vega exposure of the derivative. As a concrete example, consider a forward on the average of two stocks, A and B, whose prices are 90 percent correlated. If the overall basket position has an exposure of $1 million for a 10 percent rise in the average price, should you show the exposure to A as $500,000 or as something closer to $1 million to reflect the probability that a rise in the price of A will be accompanied by a rise in the price of B? Clearly, for purposes of the firm’s consolidated risk‐management reports, $500,000 is the right figure since the consolidated reports will also be showing a $500,000 exposure Managing Exotic Options Risk 413 to B and these two positions will contribute to the consolidated reporting of total exposure to a 10 percent increase in stock prices. If you used a position closer to $1 million for the A exposure, it would have the absurd result, when combined with exposure to B, of showing an exposure greater than $1 million to a 10 percent increase in stock prices. However, including a correlation may be appropriate for specially tailored reports for traders who want a quick rule of thumb about how much the basket price will move when stock A’s price moves (perhaps because A’s price is more liquid than B’s). A particular example that has attracted industry attention is the sensitivity of convertible bond prices to changes in the underlying stock price, which we discuss further in Section 12.4.4. A particular example of a basket option is an Asian option on a single asset. An Asian option is an option on the average price of the asset over a specified set of observations. This is equivalent to an option on a basket of forwards where all the forwards are for the same underlying asset. Obviously, one would expect correlations on such forwards to be quite high. In fact, the conventional Asian option pricing formula assumes a correlation of 100 percent (see Hull 2012, Section 25.12), which is equivalent to assuming constant interest rates, which is slightly inaccurate. Note that the time period over which each forward will contribute volatility to the basket is different, which is a key element to be taken into account in the pricing of the option. 12.4.3 Index Options As a generalization, we have stated that most multiasset derivatives are illiquid. But this rule has clear exceptions—most prominently, options on interest rate swaps and options on equity indexes. Options on interest rate swaps, also known as swaptions, are mathematically and financially equivalent to options on a basket of forwards so they reflect an implied correlation. This special case is treated at length in Section 12.5. Options on stock indexes, such as the S&P 500, NASDAQ, FTSE, and Nikkei, are among the most widely traded of all options. Comparing implied volatilities of stock index options with implied volatilities of options on single stocks that are constituents of the index will therefore yield implied correlation levels. We look at the risk management consequences, which can also be applied to other liquid index options such as options on commodity baskets and FX baskets. The first principle is that the valuation of a reasonably liquid index option should always be directly based on market prices for the index option and not derived from prices for options on individual stocks in the index and a correlation assumption. Correlation assumptions, no matter how 414 FINANCIAL RISK MANAGEMENT well based in historical analysis and economic reasoning, should never be allowed to replace a market‐derived implied correlation to assess the price at which risk can be exited. This is just an application of the same reasoning that says that reasonably liquid options need to be valued using implied volatilities, not volatility assumptions based on history. This does not mean that room is not available for models that analyze the index option price in terms of its constituent parts. Traders frequently employ trading strategies based on how rich or cheap the implied correlation is relative to correlations based on historical and economic analysis. When they conclude that implied correlations are too low, they buy the index option and sell options on individual stocks in the index, hoping to gain if realized correlation is higher than implied. This is called a convergence position. When they conclude that implied correlations are too high, they buy options on individual stocks and sell the index option. This is called a divergence position. Corporate risk managers need to make a judgment about how high or low realized correlation can go in measuring the riskiness of these positions. Index options are also potentially useful in hedging illiquid basket options. For example, if a market maker has written an option on an average of 50 stocks, all of which are components of the S&P index, hedging the volatility risk of the basket option by buying an option on the S&P 500 index is likely to leave less residual risk than buying options on the 50 individual stocks and it will certainly be far more efficient from an operational risk viewpoint (an error is more likely tracking 50 options positions in single stocks than 1 options position in the index). Also in favor of the index option hedge is that index options are almost always more liquid than single stock options. However, if the option written was on the average of two stocks that are components of the S&P 500 index, hedging the volatility risk of the basket option by buying options on the two single stocks is likely to leave less residual risk than buying an option on the S&P 500 index. At some point between two and 50 stocks, the index hedge is less uncertain than the individual stock hedge, but it needs to be found empirically through simulation. Simulation is also necessary to measure the residual uncertainty of the index stock hedge for purposes of calculating reserves and limits. The most accurate means of simulation is a Monte Carlo with dynamic hedging in an underlying asset package for which the deltas on individual stocks are computed as the net of the delta on the basket option and the delta on the index option. An approximation that is much easier to compute and reasonably accurate for large baskets is to assume no delta hedging and just compute the tracking error between the two options that occurs at the final payoff. Managing Exotic Options Risk 415 12.4.4 Options to Exchange One Asset for Another At the beginning of Chapter 11, we stated that all vanilla options could be viewed as the option to exchange one asset for another. It is equally true, following a result of Margrabe, that every option to exchange one asset for another can be evaluated by the Black‐Scholes option formula used for vanilla options (see Hull 2012, Section 25.13). So why should we try to view these as multiasset options? Because by bringing in a third asset that plays no role in the original contract, we can in some cases increase the liquidity of the option’s valuation. This can most easily be seen by a concrete example. Consider an option to exchange 10,000 ounces of gold for £4.5 million. Clearly, this option will be exercised if and only if an ounce of gold at the expiration of the option is worth more than £450. Equally clearly, this contrast has absolutely no reference or relationship to dollars. However, it can be viewed, as a mathematical equivalence, as a spread option on the difference between the dollar price of 10,000 ounces of gold and the dollar price of £4.5 million. To see this equivalence, consider the following: ■ ■ The option will be exercised if and only if an ounce of gold is worth more than £450. This is equivalent to saying it will be exercised if and only if the dollar price of an ounce of gold is worth more than the dollar price of £450, which is equivalent to saying it will be exercised if and only if the dollar price of an ounce of gold minus the dollar price of £450 is greater than 0. Multiplying by 10,000, this is equivalent to saying it will be exercised if and only if the dollar price of 10,000 ounces of gold minus the dollar price of £4.5 million is greater than 0. If the option is exercised, it can be exercised by buying 10,000 ounces of gold for its then current market price in dollars, exchanging the gold under the options contract for £4.5 million, and selling the £4.5 million for its then current market price in dollars. The (necessarily positive) difference between the dollar sale price and the dollar purchase price represents the payoff of the option. What has been gained by introducing dollars into the picture? If sterling options on gold have no liquid market, but dollar options on gold and dollar‐ sterling options have a liquid market, then the gold‐sterling spread option can be valued and risk managed based on the implied volatilities of dollar‐ gold and dollar‐sterling vanilla option hedges. Some residual uncertainty will still exist due to the assumed correlation level, but this residual uncertainty may be less than the uncertainty of an illiquid gold‐sterling exchange option. As we saw in Table 12.13, this will depend on the gold and sterling‐dollar prices not being too highly correlated with one another. If they are highly 416 FINANCIAL RISK MANAGEMENT correlated, implying a very negative correlation for the long and short positions in the spread basket, then little can be gained from being able to hedge the sensitivity to implied volatilities of dollar‐gold and dollar‐sterling. A particular case of an option to exchange one asset for another that draws considerable attention is the large market in convertible bonds; see Hull (2012, Section 26.4) and Tsiveriotis and Fernandes (1998). Convertible bonds offer the bondholder an option to exchange the bond for a fixed number of shares of the firm issuing the convertible bond. Convertible bonds generally have reasonably liquid markets, so there is rarely a valuation advantage to viewing them as spread options. However, when determining trading strategies and evaluating risk exposures, it is often convenient to assess the dependence of convertible bond valuations on the implied volatility of the equity option (more precisely, the equity‐cash option), the assumed volatility of the option on a straight (nonconvertible) bond issued by the firm, and the assumed correlation between the bond and the stock. As discussed at the end of Section 8.3, one trading strategy often pursued is to try to take advantage of the implied volatility for an equity option on the stock of a particular firm being higher than the equity volatility implied by the price of a convertible bond issued by that firm. A trader may decide that buying a convertible is an inexpensive way of buying volatility on the firm’s equity price. Or a trader might choose to run a basis position long the convertible bond and short the equity option. Risk analysis of such positions should be sensitive to the reasonableness of assumptions about the volatility of the bond option and the correlation between the bond and stock that have been used to conclude that the convertible bond’s equity volatility is cheap. The valuation of a convertible should always be based on observed market prices, not on assumptions about correlation. Another issue that frequently arises in the management of convertible positions is determining the correct delta to use in hedging a convertible position with stock. It has often been observed that when stock prices are so low that the convertible is far from its exercise price, the actual response of the convertible price to changes in the stock price is far larger than would be expected from a delta derived from a model that only accounts for volatility of the stock price. The explanation of this observation can be found in the correlation between the bond and stock. When stock prices are far below its exercise prices, a convertible bond ought to behave very much like a straight bond, but both the bond and stock price will be impacted in similar ways by changes in the outlook for the firm’s earnings (this is discussed in more detail in Section 13.2.4). If a convertible bond behaves more like a straight bond than a stock, then a straight bond would seem like a better hedge. However, there might be reasons for using the stock as a hedge, such as greater liquidity or ease in Managing Exotic Options Risk 417 borrowing the stock relative to the straight bond. In such instances, hedging ratios should certainly reflect the assumed correlation between stock and bond prices. But you must be careful to remember that the correlation assumption drives this delta. For example, if the firm’s risk reports show a sensitivity to credit spread for the convertible, also showing a high sensitivity to stock price for the convertible in the firm’s risk reports would involve a double count of the sensitivity to the bond price—once directly and once though the bond‐stock correlation. 12.4.5 Nonlinear Combinations of Asset Prices When a derivative’s payoff is the function of a nonlinear combination of a set of asset prices, none of the three simplifying characteristics that hold for a linear combination can be assumed to be in force. This can be illustrated by a single concrete example: a quanto forward whose payoff is calculated by the product of an asset price and FX rate. On January 25, 2002, stock in the Sony Corporation was trading at 6,080 yen per share and the yen was trading at 134.79 yen per dollar. So the then current dollar price of a share of Sony stock was 6,080/134.79 = $45.11. The six‐month forward price for Sony stock on that date was also roughly 6,080 yen per share and the six‐month forward exchange rate was 133.51 yen per dollar. Suppose a customer comes to a market maker looking to purchase 1,000,000 shares of Sony stock for six‐month forward delivery at a dollar price. Possible contracts (see Reiner 1992 for a full discussion) could be: ■ ■ ■ Make the purchases at a dollar price fixed in advance. The market maker has a static hedge available (it is an exchange of assets, as discussed in Section 12.4.4). She can purchase 1,000,000 shares for six‐month forward delivery at 1,000,000 × 6,080 = 6,080,000 yen and purchase 6,080,000 yen for six‐month forward delivery at 6,080,000/133.51 = $45,539,660, which is the price, without profit margin, she should charge the customer. Make the purchase at a dollar price based on the exchange rate, which will be in effect in six months. The market maker has a static hedge available. She can purchase 1,000,000 shares for six‐month forward delivery at 1,000,000 × 6,080 = 6,080,000 yen. The dollar price will be determined in six months based on the then prevailing exchange. Agree that the dollar price per share will differ from the current six‐ month forward price of 6,080/133.51 = $45.54 per share by the percentage change in the yen price per share. So if the yen price in six months is 6,080 × 110% = 6,688, the price per share to be paid will be $45.54 × 110% = $50.094. This is a quanto. 418 FINANCIAL RISK MANAGEMENT No static hedge is available for a quanto. The market maker can begin with a purchase of 1,000,000 shares for six‐month forward delivery for 6,080,000 yen and a 6‐month forward exchange of 6,080,000 yen for $45,539,660. However, if the forward share price rises by 10 percent, she now has FX risk on an additional 1,000,000 × 6,080 × 10% = 608,000 yen and must enter into a forward exchange of these yen for dollars. If the forward FX rate rises by 10% to 133.51 × 110% = 146.86 yen per dollar, she now has stock price risk of an additional 10 percent, since her stock price hedge is for a fixed amount of yen and what she needs is a hedge for a fixed amount of dollars. As the yen weakens against the dollar, she needs to increase the amount of hedge denominated in yen to maintain the dollar amount of the hedge. This pattern, a change in one asset price requiring a dynamic change of the hedge amount of the other asset, is typical of derivatives with payoffs based on the product of two asset prices. The formula for valuation of a quantoed forward, under the assumption of a bivariate lognormal distribution, is the price of a standard forward multiplied by exp(ρσSσF), where σS is the volatility of the stock price denominated in yen, σF is the volatility of the FX rate (that is, the yen price denominated in dollars), and ρ is the correlation between the stock price denominated in yen and the FX rate. A brief explanation of this formula can be found in Hull (2012, Section 29.3), and a more detailed derivation can be found in Baxter and Rennie (1996, Section 4.5). Two important consequences follow from this formula. First, the value of the derivative, even though it is not an option, is dependent on the volatilities of the assets and the correlation. Second, if the correlation is zero, then the valuation formula for a quanto is the same as the valuation formula for a standard forward, so the total impact of the dynamic hedging required must balance out to zero (however, this dynamic hedging could still result in transaction costs). A derivative with very similar characteristics to a quanto is a difference swap, in which the payoff is based on the future difference between interest rates in different currencies multiplied by a notional principal denominated in one of the currencies. For example, the difference between a dollar interest rate and a yen interest rate may be multiplied by a dollar notional amount. The future dollar interest rate multiplied by the dollar notional amount represents a quantity that can be statically hedged, but a yen interest rate multiplied by a dollar notional amount is a quantoed combination that requires dynamic hedging of both the yen interest rate and the dollar/ yen FX rate. For more details, see Hull (2012, Section 32.2) and Baxter and Rennie (1996, Section 6.5). Once the bivariate lognormal assumption is dropped, more complex valuation algorithms are required. Both the Monte Carlo and trinomial tree approaches discussed in Section 12.3 have the flexibility to be directly Managing Exotic Options Risk 419 applied to quantos or any other derivative based on a nonlinear combination of asset prices. Both approaches build probability distributions for each asset separately and can incorporate a full volatility surface (and, in the case of Monte Carlo, can incorporate stochastic volatility and price jumps). Both approaches can factor in any desired correlation assumptions between assets. Both approaches can then compute any desired function of the asset prices, no matter how complex, based on the individual asset prices at each node (and Monte Carlo can incorporate full price histories of the assets if they play a role in the function). Nonlinear functions of multiple asset prices can range from the simplicity of the maximum or minimum price of a basket of assets to the complexity of an involved set of rules for successively dropping high and low prices out of a basket on which an average is being calculated. Some assets in the basket may represent quantoed translations from other currencies. As a further step, options can be written on any of these nonlinear functions, and exotic features such as barriers can be introduced. So long as the Monte Carlo or tree is valuing the nonlinear function correctly, it should also value the option correctly. A general designation for derivatives based on nonlinear functions of multiple asset prices and their derived options is a rainbow contract. Hedging considerations for derivatives on nonlinear combinations are exactly parallel to those for derivatives on linear combinations, so the approach in Section 12.4.2 can be applied. The only difference is that the simple approximation formulas used in that section do not apply. Computations of sensitivities to shifts in asset prices, implied volatilities, and assumed correlations generally need to be evaluated by rerunning the Monte Carlo or trinomial tree valuation model with shifted inputs. Another interesting example that is similar in structure to the quanto is counterparty credit exposure on a derivative such as an interest rate swap of a FX forward. As discussed in Section 14.3.5, counterparty credit exposure can grow or diminish through time as a function of the interest rate or FX rate driving the value of the derivative. This credit exposure can be hedged by the purchase of credit derivatives or the short sale of bonds issued by the counterparty. The total value of the credit exposure is then the product of the value of the derivative and the credit spread on the counterparty. Similar to a quanto, a dynamic hedge is required. A change in the value of the derivative requires a change in the size of the credit hedge and a change in the size of the credit spread requires a change in the size of the derivative hedge. In Section 11.3, we examined a case of mean reversion in which there is a narrower dispersion of final underlying price levels than would be implied by a pure random walk and we questioned whether dynamic hedging 420 FINANCIAL RISK MANAGEMENT costs would be a function of the higher short‐term volatility or the lower long‐term dispersion. Our answer, based on both Monte Carlo simulation and theory, was that sufficiently frequent rehedging makes dynamic hedging costs depend entirely on short‐term volatility, but a trader who wanted to take advantage of anticipated lower long‐term dispersion could do so by rehedging less frequently (but with an attendant trade‐off of a higher uncertainty of hedging costs). Let’s ask a parallel question for correlation. Suppose you anticipate that two assets will have a strong correlation in terms of long‐term trend, but very little correlation in terms of short‐term moves. If you are dynamically hedging a position whose valuation depends on correlation, will your dynamic hedging costs be a function of the low short‐term correlation or the high long‐term correlation? You shouldn’t be surprised to find that the answer is the same for correlation as it is for single‐asset volatility. If you rehedge often enough, only the short‐term correlation impacts hedging costs. If you want to take advantage of the anticipated long‐term trend, you must hedge less frequently and accept a higher uncertainty of hedging costs in exchange for expected hedging costs being influenced by the longer‐term correlation. Many people find this conclusion highly nonintuitive. Consider an example. Suppose you are hedging the counterparty credit risk on an FX forward and that over the life of the forward the exposure continues to grow while the credit rating of the counterparty continuously deteriorates, but the individual moves are uncorrelated. As the exposure grows, you are going to have to buy more credit protection, and it may be hard to believe that you will not have to pay for this increased credit protection at the higher price levels brought on by the deteriorating credit rating. To help see how this works mechanically, I have provided the CrossHedge spreadsheet, which enables you to enter a price history of six prices for each asset and which looks at the hedging of an exotic paying the product of the two asset prices. The spreadsheet shows the hedging and its costs under two assumptions: if the price moves between the two assets are completely uncorrelated and if the price moves between the two assets are perfectly correlated. The complete lack of correlation is implemented by having each price move on the first asset precede in time each price move on the second asset, so there is time to change the hedge quantity before the second asset’s price changes. (Remember that for a payoff tied to the product of two asset prices, a change in the price of one asset requires a change in the hedge quantity of the other asset.) Perfect correlation is implemented by simultaneous changes in prices. Table 12.14 shows the case of deteriorating credit on counterparty credit risk. The first asset is the exposure amount and the second asset is the 421 Quanto Asset 1 Asset 2 Total Asset 2 5 10 30 50 80 100 Total Buy/Sell 5 0 5 0 20 0 20 0 30 0 20 –100 P&L –5,950 1,750 4,200 0 1,750 Uncorrelated Asset 1 Price Proceeds Buy/Sell 10 –50 10 20 0 10 20 –100 0 30 0 10 30 –600 0 40 0 10 40 –800 0 50 0 10 50 –1,500 0 60 0 10 60 –1,200 0 60 6,000 –60 4,200 Asset 2 Price Proceeds 5 –50 5 –50 10 0 10 –100 30 0 30 –300 50 0 50 –500 80 0 80 –800 100 0 100 6,000 Hedges Quanto Asset 1 Asset 2 Total 20 –100 30 20 20 5 Buy/Sell 5 P&L –5,950 1,750 3,250 –950 Cross‐Hedge of Deteriorating Credit on a Growing Counterparty Exposure Prices Asset 1 Asset 2 10 5 20 5 20 10 30 10 30 30 40 30 40 50 50 50 50 80 60 80 60 100 Hedges Asset 1 10 20 30 40 50 60 TABLE 12.14 End 4 3 2 Start 1 Start 1 2 3 4 End 60 60 50 40 30 20 1,750 –1,200 6,000 –1,500 –800 –600 –100 10 –60 10 10 10 10 Correlated Asset 1 Price Proceeds Buy/Sell 10 –50 10 100 100 80 50 30 10 3,250 –1,000 6,000 –800 –500 –300 –100 Asset 2 Price Proceeds 5 –50 422 FINANCIAL RISK MANAGEMENT discount on the counterparty’s bonds. As credit deteriorates, the discount goes all the way to 100 percent, corresponding to the worst possible case of default with no recovery. Despite the fact that the exposure is steadily growing while the discount is steadily increasing, the change in the value of the product is completely hedged in the uncorrelated case. Examining the impact of the individual hedges should impart a better sense of how the hedge works—each change in credit quality and exposure has been hedged by having the right size hedge in place at the time of the change. 12.4.6 Correlation between Price and Exercise Standard option pricing assumes a correlation of 100 percent between price and exercise; that is, option buyers will exercise their options when, and only when, the price of the underlying asset makes it profitable to exercise. However, in some instances, it can be argued that a correlation of less than 100 percent should be assumed. These arguments rely on a combination of historical experience, showing that a previous correlation has been less than perfect, and on a behavioral analysis of the option buyers, demonstrating that they have motivations that conflict with optimal option exercise. In terms of game theory, standard option analysis, which assumes a correlation of 100 percent, is equivalent to a zero‐sum game in which a loss by the option seller is exactly offset by a gain for the option buyer. A correlation of less than 100 percent corresponds to a non‐zero‐sum game. Table 12.15 shows the impact of different correlation assumptions, multiplying the payoff based on price and exercise by the probability and summing to get an expected return. For example, it may be argued that a municipality that has the option to require early repayment of a fixed‐rate term deposit without paying any penalty, which is equivalent to a swaption, will only exercise this option in response to a change in its cash needs, which are uncorrelated with interest rate levels. Support for this analysis should certainly include historical studies of how similar municipalities have exercised these options. However, even reasonable explanations of behavior and historical precedent may be questionable evidence. In the absence of any actual legal constraint or internal costs that exercise would entail, it is possible that institutions will become more efficient exercisers of options over time, as they gain financial sophistication or as large economic movements (for example, unusually high interest rates on new deposits) create increased incentives to focus attention. Such arguments may become more plausible when the option must be exercised by a large group of individuals. Correlation now becomes a question of what proportion of a population will exercise options in a timely fashion, and their diversity of circumstances will argue for less than perfect 423 Managing Exotic Options Risk TABLE 12.15 Correlation between Price and Exercise Standard Option Price Exercise No Exercise 110 90 –10 +10 0 0 Probability × 50% 0 0 50% = –5 25% 25% = 0 15% 35% = –2 35% 15% = +2 No Correlation Price Exercise No Exercise 110 90 –10 +10 0 0 Probability × 25% 25% Some Correlation Price Exercise No Exercise 110 90 –10 +10 0 0 Probability × 35% 15% Negative Correlation Price Exercise No Exercise 110 90 –10 +10 0 0 Probability × 15% 35% correlation. An example would be a pension plan that guarantees some minimum return on a particular investment strategy. If option exercise were a zero‐sum game, the individual investors would withdraw from the plan whenever the investment was below the minimum return in order to collect the guarantee. However, financial institutions that provide these guarantees value them based on behavioral assumptions about the individual participants, whose varied circumstances with regard to age, career, and tax status make the cost of exercising the option different for each subgroup. An important example of an option exercised by a large group of individuals is the very sizable market in asset‐backed bonds, where each bond is backed by a pool of mortgages, automobile loans, or other consumer loans. Although these assets often provide consumers the legal right to prepay the loan without penalty, individual circumstances often get in the way of an economically efficient exercise of this right. First, refinancing a loan often involves substantial personal costs (for example, legal fees, title searches, and the time devoted to the transaction). For an institution on a large loan, these would probably be insignificant relative to gains from exercise, but this may not be true for an individual. Second, some consumers may not be able to refinance due to a deteriorating credit rating or decrease in asset value. Others may have strong personal motives that outweigh the costs of financing, such 424 FINANCIAL RISK MANAGEMENT as a required move or a divorce forcing a home sale that causes a desirable rate mortgage to be prepaid or the desire to trade a car for a newer model. Given the enormous size of this asset class and the plausibility of less than perfect correlations, financial firms have invested and continue to invest large amounts of money in research to develop accurate models of this correlation. A good introduction to this asset class is Davidson et al. (2003), with its Chapter 9 introducing the modeling issues involved. For some assets, such as automobile loans, the general conclusion is that correlation tends to be close to zero. For mortgages, correlation is definitely strongly positive; falling mortgage rates trigger massive refinancings, and rising mortgage rates trigger considerably slow refinancings. However, correlation is certainly far from perfect, and the stakes in properly identifying which mortgage bonds represent good investments are sufficiently high to support detailed research trying to predict the relationship between refinancing behavior and prevailing mortgage rates by population subcomponent, such as the geographic region or size of mortgage. The relationships developed are often quite complex. The behavior depends not just on current mortgage rates, but also past mortgage rates and yield curve shape. Consumers are found to be sensitive not only to the current refinancing advantage, but also to beliefs as to whether that advantage will be growing, since the costs of refinancing are high enough to cause consumers to attempt to minimize the number of times they refinance. Another factor, known as burnoutt, indicates that a consumer population that has already experienced a period of low rates will show lower refinancing response (as a proportion of mortgages still outstanding) in a subsequent low rate period. This is presumably due to the proportion of those who did not refinance the first time who cannot afford to refinance. Monte Carlo models of the correlation are used to project consumer behavior under a variety of possible future interest rate movements, and bonds are ranked on the basis of option‐adjusted spread (OAS)—the spread the bond is earning over a comparable‐maturity Treasury after taking into account the cost of the refinancing option based on the assumed correlation. Why does this spread remain? One reason is certainly that these correlation relationships are only estimates based on past data that could prove to be wrong. When unanticipated shifts in consumer behavior on refinancings are observed, such as a prolonged period of very low rates resulting in greater consumer education about refinancings, leading to refinancing levels that substantially exceed those predicted by models based on past data, OAS can show large rapid increases. To some extent, this will later reduce as Monte Carlo models are updated to accommodate the new experience, but some OAS increase may persist, reflecting an increase in uncertainty over the accuracy of such models. Managing Exotic Options Risk 12.5 425 CORRELATION‐DEPENDENT INTEREST RATE OPTIONS Throughout Chapter 11 on vanilla options and in Sections 12.1 and 12.2, we have dealt with options whose underlying can be regarded as a forward to a set future date. As we discussed at the beginning of Chapter 11, all uncertainty about discounting rates for these models can be collapsed into the volatility of the forward. However, some options have payoffs that depend on forwards for several different future dates (but with all forwards on the same spot underlying). The primary example would be an American option that gives the option holder freedom to determine the timing of payoff. More complex dependence on different forwards can be seen in the products we examined in Section 12.3, such as barrier options. Options that depend on forwards for several different future dates can usefully be viewed as options on multiple underlyings with all relationships between these forwards built into the correlation structure assumed between the forwards. Indeed, this is the approach to multifactor interest rate models that has predominated over the past two decades in the form of the Heath‐Jarrow‐Morton (HJM) models (see Hull 2012, Section 31.1) and the LIBOR market models, also known as Brace‐Gatarek‐Musiela (BGM) models (see Hull 2012, Section 31.2). Should we then just view these products as a particular class of options with multiple underlyings and consider their risk management issues as already having been dealt with in Section 12.4? One reason for not availing ourselves of this convenient shortcut is that the large volume of these options that actively trade encourages extra effort to try to find a simpler structure and faster computation time for subsets of this product. Another reason is that this represents the only class of multiasset options where some reasonable liquidity exists in products that require correlation inputs to value, so it is worth studying how much information on correlation can be extracted from observed market prices. Three levels of models are essentially available, of increasing mathematical and computational complexity. The simplest level includes the binomial and trinomial tree models in which the relationship between different forwards is treated as constant. In Section 12.5.1, we examine risk management using these models and the conditions under which more complex models are required. The second level includes the single‐factor interest rate models in which the relationship between different forwards is treated as stochastic. In Section 12.5.2, we examine risk management using these models and the conditions under which the third level of full‐blown multifactor HJM or BGM models are required. Finally, in Section 12.5.3, we look at how much correlation information can be extracted from observed market prices. 426 FINANCIAL RISK MANAGEMENT 12.5.1 Models in Which the Relationship between Forwards Is Treated as Constant We have already encountered binomial and trinomial tree models in which the relationship between forwards is treated as constant—the local volatility models discussed in Section 12.3.2. Recall that this section was devoted to options whose payoff depends on the underlying price of a single asset at several different times. Because values of the asset at several different times are involved, we needed to be concerned with hedging and valuation depending on several different forwards. However, the only way to avoid treating these different forwards as multiple assets is to assume that a constant relation exists between them. This is in effect what is done in the local volatility models of Section 12.3.2, since the only variable changing on the tree is the spot price of the asset and all forward prices are derived based on fixed interest rate relationships between forward and spot prices. In this section, we study the simplest, most widely traded, and best‐ known version of a product that depends on the underlying price of a single asset at several times—the American option. American options differ from European options by a single added feature: the right of the option buyer to exercise the option at any time. A simple variant restricts the right to exercise to several specified times and is variously known as a semi‐European, semi‐American, or (as a geographic middle ground between European and American) Bermudan option. American and Bermudan options have long been valued using binomial trees (the Cox‐Ross‐Rubinstein model) and more recently using trinomial trees to allow for nonflat volatility surfaces. See Hull (2012, Chapter 12) for the mathematics of the binomial tree. See Clewlow and Strickland (1998, Chapter 5) for the use of trinomial trees to incorporate the volatility surface. The key assumption is that the relationship between the forwards remains fixed. Most typically, this is represented by a constant interest rate and forward drift (or constant dividend rate, with driftt defined as the interest rate less the dividend rate). However, any constant set of relationships between forwards can be accommodated with no increase in complexity or cost of computation, as discussed in Hull (2012, Section 20.5). Four factors drive the value of early exercise (all of this discussion is for calls—we are continuing our convention from Chapter 11 of treating all options as calls): 1. Price. When prices rise, it increases the probability of price levels high enough to warrant early exercise, so early exercise value increases. 2. Volatility. The more volatile the price, the greater the incentive not to exercise early in order to take advantage of the time value of the option. Managing Exotic Options Risk 427 However, high volatility means a greater percentage of price moves will be large enough to warrant early exercise. So the two impacts of higher volatility run in opposite directions. In practice, the second effect is usually larger, and higher volatility increases early exercise value. 3. Financing cost. The higher the net cost of funding the delta hedge of the option, the greater the incentive to exercise early. However, if net financing cost is earning the option buyer money on his delta hedge, it discourages early exercise. An equivalent way of viewing this is through the drift of the forward. If drift is positive, this decreases the incentive to exercise the call early since it is likely the call will be worth more after the upward drift. If drift is negative, this increases the incentive to exercise the call early since it is likely the call will be worth less after the downward drift. 4. Discount rate. Early exercise allows earlier receipt of option payoffs. This is more valuable the higher the discount rate, so high discount rates encourage early exercise. The AmericanOption spreadsheet illustrates the computation of American option values using a Cox‐Ross‐Rubinstein binomial tree. It focuses on the computation of the early exercise value, defined as the excess value the American option possesses over the corresponding European option. Table 12.16 shows some sample results. Note from Table 12.16 the relatively small impact of discount rates on early exercise relative to drift. Since exchange‐traded American options are all options on a fixed forward, they all have zero drift, so early exercise value is quite small. This explains the claim made at the start of Chapter 11 that exchange‐traded American options have little valuation difference from European options. Hedges can be established for the impact on the early exercise value of all four of these factors, as illustrated in Table 12.16. For delta and vega, we calculate the ratio of American option delta and vega to the corresponding European option delta and vega, enabling the American to be represented in delta reports and price‐vol matrices for the corresponding European option. For discount and drift, the sensitivity of the early exercise value to a 100 basis point shift is calculated and can be used to establish a hedge. This is a comparable situation to vega hedging an option you are valuing using the Black‐Scholes model—the theory behind the model says volatility is constant, but you are going “outside the model” to hedge against volatility uncertainty. Here we are determining the early exercise value using a model that says that discount rate and drift are constant, but we are establishing a hedge against an uncertain discount rate and drift. The liquid proxy for the American option would be a combination of the corresponding European 428 TABLE 12.16 100.24% 100.18% $163 $844 0.17% 3.46% 0.00% 0.00% 100.02% 100.00% $136 $778 7.89% 7.90% 0.01% 0.39% 100.00% 1 20% 1% 0% 7.97% 7.97% 0.00% 0.40% 100.00% 1 20% 0% 0% 100.96% 99.99% $202 $1,044 1.05% 19.72% 7.44% 7.52% 0.08% 0.39% 100.00% 1 20% 0% –1% Early Exercise Values and Hedges for American Option American / European delta American / European vega Rate sensitivity per $1mm Drift sensitivity per $1mm European price American price Early exercise Vega Early exercise as % of: European price Vega Strike Time Volatility Rate Drift 110.39% 99.84% $214 $1,267 9.23% 136.95% 5.57% 6.09% 0.51% 0.38% 100.00% 1 20% 0% –5% 101.63% 101.15% $201 $971 1.13% 22.74% 7.58% 7.66% 0.09% 0.38% 100.00% 1 20% 5% 0% 103.96% 102.68% $219 $1,037 2.65% 53.11% 7.21% 7.40% 0.19% 0.36% 100.00% 1 20% 10% 0% 100.58% 100.08% $291 $956 0.70% 20.16% 11.37% 11.45% 0.08% 0.39% 100.00% 1 30% 0% –1% Managing Exotic Options Risk 429 option and the extra hedges needed for the exposure to discount rate and drift. This can easily be converted into a Monte Carlo simulation of the differences in final payout between an American option and this liquid proxy, given a simulation of changes in the underlying price, the discount rates, and drift. The critical assumption when calculating these hedges is that discount rate and drift risk can be valued and hedged as variables independent of the spot price risk. Equivalently, the assumption is that the level of forward rates is uncorrelated with the shape of the forward rate curve. This assumption is reasonable for most equities, questionable for FX and commodities (refer back to our discussion of mean reversion in Section 11.3), and certainly false for interest rate options, since high correlation will exist between the rate determining the payoff and the rates determining the discount and drift. Is it possible that the impact of this correlation is small enough to ignore for practical purposes? As shown in Table 12.16, when drift is positive or zero or when it is not too negative, the total size of the early exercise value is not too large so any impact of correlation can probably be ignored. When drift is quite negative, early exercise value becomes significant and it is likely that the impact of correlation between interest rates needs to be taken into account. To do so requires some type of term structure model; the factors influencing choices between these models are discussed in the next section. This is particularly true for options on bonds or on swaps, where the pull to par causes drift to be very negative. Because the duration of a bond or swap gets shorter as time passes, the impact of interest rates on prices is continuously declining. So an option holder faced with an early exercise decision knows that the current price premium is likely to diminish through time—if interest rates don’t move further in her favor, current rate levels will translate into a smaller price advantage in the future. This is true both for options that pay on rising interest rates and those that pay on falling interest rates, since both high bond prices based on low interest rates and low bond prices based on high interest rates move in the direction of par if rates stay the same as time to maturity diminishes. If any substantial reduction in the duration of an underlying bond or swap occurs during the tenor of an option, this negative drift will require a term structure model. If no substantial reduction in duration occurs over the option tenor, then a Cox‐Ross‐Rubinstein model with the duration held constant can be used as a reasonable approximation. A rule of thumb that is often used is that this approximation is suitable as long as the duration of the underlying at the start of the option life is at least 10 times as great as the option tenor. So this rule of thumb would allow the use of a Cox‐Ross‐ Rubinstein model for a six‐month option on a 10‐year bond, but would insist on a term structure model for a one‐year option on a five‐year bond. 430 FINANCIAL RISK MANAGEMENT 12.5.2 Term Structure Models The most liquid options products based on interest rates are caps, floors, and European swaptions. A European swaption is an option to enter into a swap at some fixed future date at a rate fixed at the time of entering into the option. A one‐period swap is a forward rate agreement (FRA) and, by convention, an option on an FRA is called a caplett if it is an option to receive floating and pay fixed (i.e., it pays off when rates are high) and is called a floorlett if it is an option to pay floating and receive fixed (i.e., it pays off when rates are low). Market practice is to sell caplets and floorlets in bundles, called strips, which are called caps and floors, respectively (so a swaption is an option on a bundle of FRAs, a swap, and a cap or floor is a bundle of options on FRAs). For example, a five‐year cap on three‐month LIBOR would consist of a strip of nineteen options on three‐month FRAs that have starting dates beginning at times starting three months from now and ending four years and nine months from now. Market convention will quote a single volatility for a cap or floor, which is then applied to each of the constituent FRA options—but this is just a convention to make it easy to communicate. Actual pricing of a cap or floor evaluates each individual FRA option at the appropriate volatility, sums the resulting prices to arrive at the price of the cap or floor and then solves for a single volatility, which, applied to each individual FRA option, would result in this summed price. A European swaption on a one period swap is identical to a caplet or floorlet. In our discussion of term structure models, the models used to price complex interest rate products, in both this section and in Section 12.5.3, we will for convenience sometimes refer to all of the liquid instruments being used for calibration of the models as swaptions, even though those which are options on individual FRAs are more accurately called caplets or floorlets. Broadly speaking, term structure models come in two varieties: single‐ factor models that assume that the correlation between all forwards is 100 percent and multifactor models that can accommodate less than perfect correlation structures. Both types of model can handle a correlation between the underlying of the option and drift. Multifactor models are obviously more accurate, but add a considerable cost in computation time and complexity. Since American and Bermudan options on swaps and bonds are by far the most utilized exotic in the interest rate options market, there is a strong incentive to try to use single‐factor models for this product as long as accuracy is reasonable. A critical fact about interest rate options, which any term structure model needs to deal with, is that options of the same tenor for bonds (or swaps) of different maturities tend to have lower interest rate volatilities for the long 431 Managing Exotic Options Risk TABLE 12.17 Annualized Volatility of Dollar Par Swap Yields Tenor 1 Year 3 Years 4 Years 5 Years 6 Years 7 Years 8 Years 16.95 16.24 15.95 15.81 15.26 15.18 14.88 Annualized volatility Tenor Annualized volatility 9 Years 10 Years 12 Years 15 Years 20 Years 30 Years 14.72 14.65 14.25 12.50 12.87 11.99 maturity. This can be confirmed both by observations of implied volatilities from market quotes and from historical volatility observations of par bond or swap yields. (For example, Table 12.17 shows annualized volatilities by tenor based on six years of dollar par swap yields between 1996 and 2001— see the DataMetricsRatesData spreadsheet for the underlying data.) Broadly speaking, this fact can be explained by some combination of the following two theses: 1. Forward rates are less than perfectly correlated with one another and the longer the bond maturity, the more its volatility is dependent on the correlation between forwards. 2. Longer‐term forwards have lower volatility than shorter‐term forwards. The latter theory implies that interest rates are mean reverting, since it requires the standard deviation of longer‐term forwards to be lower than that produced by a pure random walk driven by the volatility of shorter‐ term forwards. To see the interaction between the correlation and volatility of longer‐term forwards when explaining swaption volatility, refer to Section 12.5.3. Because it assumes that all correlation between forwards is 100 percent, a single‐factor model must utilize the lower volatility of long‐term forwards to drive the observed volatility structure of swaptions. To what extent does forcing one of these two levers to bear all of the explanatory weight distort valuation and hedging? In principle, to answer this question, build the best multifactor term structure model you can; calibrate both this multifactor model and the single‐factor model that is proposed for production use to the current set of vanilla cap, floor, and European swaption prices; and then compare their output in valuing exotic products. Although this is too daunting a computational task to attempt here, I will give a flavor of what this analysis is like for one very simple case: a three‐year time horizon; three liquid vanilla products—a one‐year caplet on a one‐year LIBOR, a two‐year caplet on a one‐year LIBOR, and a one‐year 432 FINANCIAL RISK MANAGEMENT swaption on a two‐year swap; and a flat implied volatility surface with respect to strike. We will assume the two‐year swap is on a one‐year LIBOR. We will take advantage of the equivalence between swaps and packages of forward rate agreements (FRAs), as noted in Section 10.1.6. The notation we will employ is to label a FRA by the time at which its rate is determined and the time at which it settles. So a 2–3 FRA has a rate determined at the end of two years based on what would then be the one‐year rate. The model will be calibrated to the current one‐year LIBOR, 1–2 FRA and 2–3 FRA, the one‐year volatility of the 1–2 FRA, the first‐year volatility of the 2–3 FRA, the second‐year volatility of the 2–3 FRA, and the one‐year correlation between the 1–2 FRA and the 2–3 FRA. In addition to valuing the liquid vanilla products, we will value four exotics: 1. A two‐year Bermudan swaption that can be exercised either at the end of year 1 based on the then prevailing two‐year LIBOR or at the end of year 2 based on the then prevailing one‐year LIBOR. 2. A two‐year caplet on a one‐year LIBOR that can knock out depending on the level of a one‐year LIBOR in one year. 3. A forward‐start caplet on a one‐year LIBOR that has a one‐year tenor and begins in one year with a strike set to the then one‐year LIBOR. 4. A one‐year tenor option on the spread between a two‐year LIBOR and a one‐year LIBOR. Our full term structure model is in the TermStructure spreadsheet. It is a simple Monte Carlo implementation. It takes advantage of the fact that only two exercise points are available for the Bermudan to value it by the following trick. At the end of two years, exercise is a simple decision. If you are in the money at the end of one year, you have a choice between early exercise, which gives you a two‐year par swap, or waiting a year, which is equivalent to a one‐year caplet on a one‐year LIBOR. So you just choose the maximum value between the two‐year swap and the one‐year caplet on the one‐year LIBOR. Using a flat initial rate curve of one‐year LIBOR = 1–2 FRA = 2–3 FRA = 7 percent, two scenarios can be computed as shown in Table 12.18, which can be verified with the spreadsheet. Notice the following: ■ ■ The inputs have been deliberately chosen to calibrate to the same vanilla option prices in both scenarios. The higher correlation in scenario 2 must be balanced by the lower volatility of the longer‐term 2–3 FRA in the first year in order to match the one‐year swaption price. This must be followed by higher volatility in 433 Managing Exotic Options Risk TABLE 12.18 The Valuation of Interest Rate Volatility Products under Two Scenarios Inputs First‐year volatility of 1–2 FRA First‐year volatility of 2–3 FRA Second‐year volatility of 2–3 FRA First‐year correlation of 1–2 FRA and 2–3 FRA Scenario 1 Scenario 2 20.00% 19.50% 14.83% 50.00% 20.00% 14.00% 20.00% 100.00% Valuations One‐year caplet on one‐year LIBOR One‐year swaption on two‐year swap Two‐year caplet on one‐year LIBOR 0.519 0.810 0.559 0.519 0.810 0.559 Bermudan swaption Knock‐out caplet Forward‐start option Spread option 0.936 0.400 0.645 0.518 0.949 0.447 0.541 0.153 ■ ■ ■ ■ the second year when its time to maturity is shorter so that the combined first‐ and second‐year volatilities fit the price of the two‐year caplet. Despite a very large difference in correlations between the two scenarios, the Bermudan swaption values close to equal in both scenarios. This reflects a trade‐off between lower volatility of the 2–3 FRA in the first year, which decreases the value of early exercise, and higher volatility of the 2–3 FRA in the second year, which increases the value of the option in those cases in which early exercise does not occur. The knock‐out caplet also shows values close to equal in both scenarios. Lower correlation increases the chances that a high 2–3 FRA, which leads to a higher caplet value, will be accompanied by a 1–2 FRA that is low enough that the caplet will not knock out. This leads to a higher caplet value but is offset by the lower second‐year volatility that accompanies the lower correlation. Lower correlation causes the forward‐start option to have a higher value by adding volatility in the relation between the strike and forward to the volatility of the forward. The largest difference between the two scenario valuations is for the spread option, which is the product most directly tied to yield curve shape rather than level. It values much higher when lower correlation permits greater variability in shape. This single case is consistent with the intuition of most practitioners in the interest rate options market. For Bermudan swaptions, a one‐factor 434 FINANCIAL RISK MANAGEMENT model can be calibrated to current vanilla prices and give reasonable results, but as you move toward products that are more dependent on the future shape of the yield curve, multifactor models become more of a necessity. Although this demonstration for a two‐period case is far from conclusive for longer‐term swaptions, see Andersen and Andreasen (2001) for similar conclusions in a more general setting. This spreadsheet can be useful for gaining intuition about the direction and order of magnitude of correlation assumptions on different interest rate exotics. When multifactor models are utilized, traditionally the primary choice has been between models that assume a normal distribution of the short‐term rate, such as Hull‐White, and models that assume a lognormal distribution of the short‐term rate, such as Black‐Derman‐Toy or Black‐Karasinski. See Hull (2012, Section 30.3) and Rebonato (1998, Chapters 12 and 13) for an exposition of these models. The discussion on which of these approaches to use has often centered on whether one believes that normal or lognormal distributions of rates give closer correspondence to historical experience. This line of argument is getting to seem rather dated in light of the almost universal adoption of full volatility surfaces that accommodate mixtures of normal and lognormal assumptions in equity, FX, and commodity options models (see Section 11.6.2). As we discussed with barrier options in Section 12.3.1, not getting the shape of the implied volatility surface correct can result in major errors in the valuation of exotics. Bermudans share a key characteristic of barriers in that the strike level that determines the termination of the option can be different than the strike level that determines the value of the option, making the correct fitting of the relative volatility between these two strike levels an important determinant of valuation. A more modern approach to utilizing the full implied volatility surface when creating a single‐factor interest rate options model can be found in Khuong‐Huu (1999). Other factors that go into the choice and accuracy of a single‐factor model include: ■ ■ The Hull‐White model offers a strong computational advantage in that the forward value of a bond or swap can be computed by analytic formula for any node of the tree (see Hull 2012, Section 30.3). By contrast, lognormal models of the short rate must extend the tree all the way out to the maturity of the bond or swap and solve backwards on the tree to determine a forward value. It is possible for interest rates to become negative in some portion of the tree in normal models of the short rate. If you believe this is economically unrealistic (refer back to the discussion in Section 10.3.2), then you would want to get estimates of the degree of impact this could Managing Exotic Options Risk ■ ■ ■ 435 have on valuations and hedges; see Rebonato (1998, Section 13.9) for a balanced discussion of this issue and other strong and weak points of the Hull‐White model. The limitation of having just a single factor to calibrate with leads to conflicts between the desire to correctly fit observed prices of potential hedging instruments and the desire to avoid unrealistic evolutions of the rate curve; see Rebonato (1998, Sections 12.5 and 13.9) for an extended discussion. Black‐Derman‐Toy is a binomial tree model, in contrast to the trinomial tree models of Hull‐White and Black‐Karasinski, and is far easier to implement and maintain than the trinomial tree models. The price paid for this convenience is that the speed of mean reversion is determined and cannot be set as an input parameter. Overcoming this weakness was the primary motivation for the introduction of Black‐Karasinski (see Hull 2012, Section 30.3). As a result, Black‐Derman‐Toy can only calibrate to a limited subset of vanilla options on any given run. For instance, in our two‐period example, it could only calibrate to the one‐ year swaption on a two‐year swap and the two‐year caplet, but not to the one‐year caplet. This could potentially reduce the number of possible hedging instruments that have been correctly priced by the model; see Rebonato (1998, 12.5) for further discussion. All of the single‐factor models share the issue that shifts in rate levels will cause shifts in the package of vanilla options that form a good hedge for an American or Bermudan option. Table 12.19 shows an illustrative example. This table is based on a 10‐year annually exercisable Bermudan call option on a 10‐year swap with a coupon rate of 7 percent and flat volatility surface at 20 percent. As should be expected, falling rates increase the value of the call, making early exercise more likely and thus increasing the impact of early volatility relative to later volatility. Rising rates decrease the value of the call, making early exercise less likely and thus increasing the impact of late volatility relative to earlier volatility. It is then easy to solve for a set of European options with similar exposure to the forward volatility curve. However, a package of vanilla options that matches the distribution of exposure at one rate level will no longer match the exposure at a different rate level. Rebonato (2002), particularly Chapters 8, 9, and 10, is an excellent source of detailed examples and exposition regarding the subtleties of calibrating term structure models to market prices of caps, floors, and European swaptions. Rebonato’s discussion of term structure models is very much consistent with the conclusions of Gatheral (2006) regarding dynamic hedging models discussed in Section 12.3.2—models that correctly price all of the liquid instruments can still differ substantially in how the volatility 436 TABLE 12.19 FINANCIAL RISK MANAGEMENT Impact of Rate Levels on the Forward Volatility Curve Dependence of a Swaption Year 5% Flat Rate Level 7% 9% 1 2 3 4 5 6 7 8 9 10 1% 12% 13% 13% 12% 11% 11% 10% 9% 8% 0% 4% 8% 11% 12% 13% 13% 13% 13% 12% 0% 0% 3% 6% 8% 13% 15% 17% 19% 20% surface evolves. And differences in volatility surface dynamics can translate into substantial differences in the cost of hedging an exotic instrument with more liquid instruments. Looking back once more to Section 8.4, the risk management approach to this should be Monte Carlo simulation of the P&L resulting from following a hedging strategy implied by a particular model, as recommended by Derman (2001). Once again, the difficulty is the computational burden of needing to compute required rehedging along all the different Monte Carlo paths. In this case, I don’t have a static or quasistatic hedging alternative to offer that I have actually had experience with. A suggested approach would be, to take a Bermudan swaption as an example, to start with a liquid proxy of a package of vanilla swaptions as in Table 12.19, based on current rate levels. The idea would be to hold this package fixed as you go forward on the Monte Carlo path. This approach runs into two problems. The first is that some of the European options will reach expiry and deliver a payoff, leaving the Bermudan option decidedly underhedged. Perhaps a simple rule could be followed, such as every time a European swaption reaches expiry, bring the package of European swaptions back up to 100 percent of the Bermudan swaption by buying new European swaptions in the same proportion as the remaining European swaptions in the original package. The second problem is how to decide when on each path Bermudan options should be exercised without needing repeated reruns of the term structure model. One approach could be to use some rule of thumb to govern exercise. Another approach would be to assume exercise on each path will take place at the time that, looking back at the path from final expiry, would be the least favorable to the trading desk. Managing Exotic Options Risk 437 12.5.3 Relationship between Swaption and Cap Prices Since a European option on a swap or bond can be a reasonably liquid instrument, and since we can view it as equivalent for valuation purposes to an option on the baskets of FRAs, which the swap is equivalent to, we can try to extract information on market‐implied correlations between FRAs from liquid prices. How much correlation information can we extract? Not that much, unless we are willing to make some additional assumptions. To see why, let’s start by considering a simplified market in which only two FRAs trade a 1–2 year and a 2–3 year. The natural options would be a one‐year caplet on the 1–2 year, a two‐year caplet on the 2–3 year, and a one‐year swaption on the combination of 1–2 year and 2–3 year. To price these three options, we need inputs for the following underlying variables: the volatility of the 1–2 year FRA in year 1, the volatility of the 2–3 year FRA in year 1, the correlation between these two FRAs in year 1, and the volatility of the 2–3 year FRA in year 2. Unfortunately, four underlying variables are present and only three options need to be priced. So it will not be possible to extract a correlation from the prices, as we have seen in the example of the previous section, unless we are willing to place some tight restrictions on the possible structure of FRA volatilities. When we move to more realistic market assumptions, the situation does not improve. The Swaptions spreadsheet can take price inputs for one‐year LIBOR caplets from one to 10 years and all possible swaption prices involving an integral number of years less than or equal to 10 (for convenience, the prices are quoted as the equivalent Black‐Scholes implied volatility). Based on an assumption as to correlation structure, the spreadsheet uses the Excel Solver to find a structure of underlying FRA volatilities that explains the prices. From your experimentation with the spreadsheet (see Exercise 12.10), you can confirm that a wide range of different correlation assumptions is consistent with a single set of prices. We have assumed zero volatility skew and smile throughout this discussion, but changing this assumption will not improve the situation. It is possible to come to conclusions about the probability of different underlying FRA volatility structures based on historical observation, and this may result in constraints that would at least give a tight range of possible market‐implied correlations. For example, one proposal that has both intuitive appeal and some empirical support is to assume that the volatility of FRAs is a function of how far they are from maturity. So the volatility of a 2–3 year FRA in its second year, when it is in the final year of its life, should be the same as the first‐year volatility of a 1–2 year FRA and the third‐year volatility of a 3–4 year FRA. The intuition behind this assumption is that new information has its greatest impact on nearby borrowing rates, so we 438 FINANCIAL RISK MANAGEMENT should expect to see greater volatility in nearby rates and lower volatility as you go farther out in maturity (this is equivalent to assuming mean reversion of interest rates, as we saw in Section 12.5.2). So if the caplet volatility in the market for a 1–2 year FRA is 23 percent, but is 22 percent for a 2–3 year FRA, it is reasonable to assume that this 22 percent can be decomposed into a 21 percent volatility in the first year, when the FRA still has over a year to go, and a 23 percent volatility in the second and last year. This assumption is powerful enough to enable all FRA correlations to be derived from swaption prices. To see this, consider that if you have N different FRAs for which you provide volatility assumptions, this can provide N2 N pricing for different swaptions (N in period 1, N – 1 in period 2, 2 N 2 N . The total number of correlations that can be and so on — ∑ i = N ) 2 i =1 2 specified between FRAs is N N since the N correlations of a FRA with 2 itself must be 100 percent and a correlation between FRAi and FRA Aj must equal the correlation between FRA Aj and FRAi. If you specify that FRA volatility is completely determined by time to maturity, it reduces the number of volatilities that can be specified to N. The total of specified volatilities N2 N N2 N N2 N plus specified correlations is then N + = . So if all 2 2 2 swaption prices are specified, a unique set of FRA volatilities and correlations that can explain them must exist. However, it is possible that placing severe constraints on the relationship between different FRA volatilities will not leave enough freedom to find implied correlations that fit market swaption prices. It can also be the case that caplet volatilities decline too steeply with time to be consistent with the assumption of FRA volatility being a function only of time to maturity; compare this with the discussion in Rebonato (1998, Section 4.5). Rebonato (2002, Section 9.1.3) makes a case that swaptions volatilities tend to be persistently higher than caplet volatilities due to supply and demand considerations. This is due to consistently high demand from corporate borrowers for cap protection of borrowing costs, while issuers of puttable bonds and buyers of callable bonds are willing to sell the options they own for a fixed upfront price, creating a supply of swaption protection. In Section 9.1.3, along with Sections 1.2 and 6.1.2, Rebonato warns against trying to fit models of exotic interest rate products to both caplet and swaption volatilities, since the difference in volatility levels due to the imbalance of supply and demand factors may result in unrealistic implications for the evolution of volatilities, which may in turn lead to future trading losses. This point is roughly Managing Exotic Options Risk 439 similar to one made in Section 10.2.1 of this book, the need to account for the trade‐off between basis risk and liquidity risk in considering the degree to which an exact fit to market prices should be attempted in a model designed to infer prices of illiquid instruments from more liquid instrument prices. EXERCISES 12.1 Using the BasketHedge Spreadsheet 1. For a flat volatility assumption (that is, smile = 0 and skew = 0), check the calculation of the square root option in the Main worksheet against another pricing method. The method could be analytic (that is, based on solving a PDE), use Monte Carlo simulation, or use a binomial or trinomial tree. Whatever method you choose, make sure you check its accuracy by pricing ordinary options and comparing the answers to the Black‐Scholes formula. 2. Pick another type of nonlinear payoff. Change Column C in the Main worksheet to calculate a hedge and pricing. Check the results for a flat volatility assumption against another pricing method, as in part 1 of this exercise. 3. Check the impact of smile and skew on the pricing of each of the following: a. The square root option. b. The option you priced in part 2 of this exercise. c. The single‐asset quanto priced in the Quanto worksheet. d. The log contract priced in the Log worksheet. e. The convexity risk hedge priced in the Convexity worksheet. f. The call‐on‐a‐call option priced in the Compound worksheet. 4. Change Column C in the Compound worksheet to price: a. A put‐on‐a‐call compound option. b. A call‐on‐a‐put compound option. c. A chooser option that as of the first expiry time (B1) turns into whichever is more valuable between a call and a put priced at the same strike (B5) to a second expiry time (B4) (see Hull 2012, Section 25.7). 5. For a call‐on‐a‐call option and all three of the options in part 4 of this exercise, use the Compound worksheet to determine how much sensitivity remains to future implied volatility after exposure to the price level has been hedged. 440 FINANCIAL RISK MANAGEMENT 12.2 Using the BinaryMC Spreadsheet Assume you are long one binary option and short a second binary option of the same size. Create a set of examples to show that there is a lower probability of loss: a. The closer the two binary options are in maturity date. b. The closer the two binary options are in strike. c. The greater the correlation in the underlying instruments of the two binary options. Also show that the variability of results can be reduced by narrowing the spread between the call options used as liquid proxies for the binary options. 12.3 Using the CarrBarrier Spreadsheet Using the same price strike, up barrier, down barrier, and original time to expiry as the one used in Table 12.7, perform the following: 1. Test the validity of the claim that unwind P&L is zero whenever drift and skew at unwind are zero. Try different combinations of time to expiry, at‐the‐money volatility, smile, and rate at the time the barrier is hit. Also try different combinations of drift and skew at the time the option is originated. 2. What conclusions can you draw about the pattern of dependence of unwind P&L on different values of drift? 3. What conclusions can you draw about the pattern of dependence of unwind P&L on different values of skew? 12.4 Using the CarrBarrierMC Spreadsheet Create a set of examples to show the sensitivity of loss probability to changes in the standard deviation of skew and the standard deviation of drift. 12.5 Using the OptBarrier Spreadsheet Take a down‐and‐out call case that you have analyzed using CarrBarrier and analyze it using OptBarrier. Use the optimization criterion of 100 percent of the maximum absolute error: 1. First use OptBarrier with four possible times and four possible at‐the‐money volatilities, but only one possible smile, skew, and Managing Exotic Options Risk 2. 3. 4. 5. 441 drift—smile, skew, and drift are all set to zero. Confirm that the values you derive for the option price are close to those that CarrBarrier derived. Change skew to a single value of 10 percent and see what option values result. Change drift to a single value of –3 percent and see what option values result. Change skew to have two values—one 0 and one 10 percent— and see what option values result and what the resulting degree of uncertainty of closeout cost is. Compare this uncertainty of forward cost to that of the CarrBarrier for the same level of skew and drift. Change drift to have two values—one 0 and one –3 percent—see what option values result and what the resulting degree of uncertainty of closeout cost is. Compare this uncertainty of forward cost to that of the CarrBarrier for the same level of skew and drift. 12.6 Using the DermanErgenerKani Spreadsheet 1. Use the spreadsheet to check the results given in Table 12.6. Then examine the impact on unwind P&L of deviations between the assumptions about unwind conditions in C8:C12 and the actual unwind conditions in C17:C21. Create a table to show the impact of changes in rate, drift, smile, and skew. 2. Verify that any changes made in initial conditions in B8:B12 will only change the initial price of setting up the hedge and will not have any impact on unwind P&L. 12.7 Using the BasketOption Spreadsheet 1. Check on the sensitivities shown in Table 12.13. 2. Create some examples to check that the General Case and the 3 Asset Case give the same answers for cases with just two or three assets. 3. Using the General Case, tabulate the rate of change in base case volatility and sensitivity to changes in volatility and correlation as the number of assets increases. How does this differ at base correlation rates of 0, 25, and 50 percent? 12.8 Using the CrossHedge Spreadsheet Try different price paths for the two assets and confirm that they always show zero P&L for the uncorrelated case. What patterns do you 442 FINANCIAL RISK MANAGEMENT observe for the P&L in the correlated case? For example, what distinguishes cases that lead to gains from cases that lead to losses? What influences the size of the gains or losses? 12.9 Using the TermStructure Spreadsheet 1. Reproduce the results in Table 12.18, which will verify that two different combinations of volatility and correlation input can produce the same valuations for vanilla products but different valuations for exotic products. 2. Find other combinations of volatility and correlation inputs that produce the same valuations for the vanilla products and determine the sensitivity of the exotic products to these inputs. 3. Create your own exotic product by specifying a different payout structure in column J and determine its sensitivity to different combinations of input volatility and correlation that leave vanilla product pricing fixed. 12.10 Using the Swaptions Spreadsheet Start with input swaption and FRA rates as follows: ■ ■ All FRA rates at 7.0 percent. Swaption volatilities from Table 12.20. These swaption volatilities display the usual pattern observed in the market of declining as swap tenor increases: 1. Input correlations of 90 percent for all combinations and use the Solver to find a set of FRA volatilities that correspond to this case. 2. Replace all the 90 percent correlations with 80 percent correlations and use the Solver to find a set of FRA volatilities that correspond. 3. You now have two different sets of FRA volatilities that can explain the same set of swaption volatilities—one based on higher correlation levels than the other. What are the patterns of difference you see between these two sets of volatilities, and how would you explain the linkage between these patterns and the difference in correlation levels? 443 1 2 3 4 5 6 7 8 9 10 Option Expiry 1 14.700% 15.100% 15.100% 15.000% 14.200% 13.300% 13.400% 13.700% 12.900% 2 13.700% 13.700% 14.000% 13.800% 13.100% 12.900% 13.200% 13.100% 3 12.800% 13.100% 13.200% 12.900% 12.700% 12.900% 12.900% 4 6 12.000% 12.100% 12.500% 12.700% 12.600% Swap Tenor 5 12.400% 12.600% 12.600% 12.800% 12.800% 12.700% Swaption Volatilities Input for Exercise 12.10 16.000% 17.700% 17.100% 17.200% 16.200% 14.800% 14.400% 14.900% 14.000% 13.000% TABLE 12.20 11.700% 12.000% 12.400% 12.500% 7 11.500% 11.800% 12.200% 8 11.300% 11.600% 9 11.100% 10 CHAPTER 13 Credit Risk he field of credit risk management has undergone major transformations over the past two decades. Traditional commercial bank lenders, whose focus used to be almost exclusively on the analysis of individual borrowers with a small dose of limits to avoid excessive concentration in a region or industry, have increasingly viewed overall portfolio management as a major part of their function. This has opened the door to rapid growth in the use of quantitative risk management techniques. At the same time, the introduction of an array of vehicles for transferring credit risk between creditors—the increased use of loan sales, loan syndication, and short sales of bonds, along with the introduction of many varieties of credit derivatives, asset‐backed securities, and collateralized debt obligations (CDOs)—has served as a tool for portfolio management. Over the same time period, many new players have become active participants in credit risk markets. While there have always been nonbank investors in corporate bonds, such as insurance companies, pension funds, and mutual funds, the variety of new instruments available for investors in credit risk—asset swaps, total return swaps, credit default swaps (CDSs), CDOs— has both introduced new investors, such as hedge funds, and increased the participation of existing investors. In looking at the principles guiding credit risk management, one sees a genuine dichotomy between the views of traditional commercial bank lenders and the views of many nonbank investors. Investors who focus primarily on liquid corporate bonds and CDSs view risk management on these instruments as no different from market risk management of equity or interest rate positions—the general principles of Section 6.1.1 would apply, with emphasis on stop‐loss limits, liquidation of positions, timely marking to market, and use of value at risk (VaR) and stress testing to assess liquidation risk. Traditional commercial bank lenders, with many loans to creditors whose debt has little liquidity and with large positions of illiquid size to creditors whose debt does have liquidity, see little value in such short‐term T 445 446 FINANCIAL RISK MANAGEMENT views of risk and concentrate instead on long‐term (multiyear) analysis of portfolio risk. This dichotomy of views relates back to the discussion in Section 1.2, with the credit risk of commercial bank lenders looking like actuarial risk, requiring an approach more like the one we’ve outlined in Sections 6.1.2 and 8.4. Caught in the middle are investors who have hybrid exposure to liquid and illiquid names—they need to use a mixture of short‐term market risk management techniques for their more liquid risks and long‐term portfolio analysis for their less liquid names. Among the players caught in the middle are market makers in over‐the‐counter derivatives, who almost always have a customer mix of counterparties with both liquid and illiquid debt. The approach in this chapter is to start with the short‐term risk management of liquid positions in Section 13.1, then to turn to long‐term portfolio risk management in Section 13.3. In between, Section 13.2 looks at non‐market‐based methods for the internal analysis of single‐name credit instruments. This topic is important as critical input to the portfolio models of Section 13.3, as a vital supplement to the techniques of Section 13.1 for names with good but limited liquidity, and as a fundamental element in trading models even for the most liquid names. Finally, Section 13.4 looks at the risk management of multiname credit derivatives such as CDS indexes and CDOs, which require a challenging mix of the portfolio management techniques of Section 13.3 and the more market‐based approach of Section 13.1, a challenge that much of the financial industry badly failed in the 2008 crisis. The important topic of the management of credit risk for derivatives counterparties is placed in a separate chapter, Chapter 14, which will draw heavily on the conclusions of this chapter. 13.1 SHORT‐TERM EXPOSURE TO CHANGES IN MARKET PRICES When dealing with sufficiently liquid debt, credit instrument risk management can be designed to look very similar to interest rate risk management, but there are some important differences. As with interest rate risk management, a good part of the challenge is coming up with a unifying principle for combining the risks of many different types of instruments with a wide variety of terms and conditions. As with interest rate risk management, the key tool will be a focus on cash flows as a unifying principle (refer back to the start of Section 10.2). This principle does not work as cleanly for credit instruments as it does for interest rates, but with some modification it will still be able to serve. Credit Risk 447 We will model our discussion in this section closely on our interest rate discussion in Chapter 10. Section 13.1.1 looks at the variety of credit instruments, Section 13.1.2 looks at the mathematical models for valuing credit instruments, and Section 13.1.3 examines the design of risk reports. 13.1.1 Credit Instruments 13.1.1.1 Bonds and Asset Swaps The market for corporate bonds has been around for a long time, and these instruments are generally well quoted for certain firms. It has always been advantageous for companies seeking capital to issue bonds, partly because of resulting tax advantages, and also not to dilute the ownership in the company by issuing too much equity. Most of the time, corporate bonds are fixed‐rate bonds, because this is what most investors in bonds prefer, even though many companies prefer to borrow at a floating rate, generally indexed to the London Interbank Offered Rate (LIBOR) (companies wishing to exchange floating‐rate payments they want to make for the fixed‐rate payments required on their bonds are a major source of demand for interest rate swaps). Most investors in corporate bonds share the following three characteristics: 1. They have cash to invest. 2. They are willing to take on credit risk, because they have a favorable view of the credit prospects of a particular firm or set of firms. 3. They are willing to take on rate risk or have a longer‐term investment horizon and so view locking into long‐term rates desirable. Some investors are interested only in the first two features because they don’t necessarily want to take a position with a view on rate risk. This is why asset swaps were created. An asset swap is a combination of a corporate bond and an interest rate swap contract that swaps the bond’s coupon into a floating payment. So the purchaser of an asset swap will receive a fixed spread as payment so long as the bond does not default. But an even larger market developed for a purer form of credit‐linked instruments, the credit default swap (CDS), that isolates credit risk without either of the other two aspects of corporate bonds. 13.1.1.2 Credit Default Swaps Credit default swaps were created in the 1990s. Their definition is very simple. While there is no default on the underlying, the protection provider receives a fixed spread payment on a regular basis (for example, every six months) from the protection buyer. If there ever is a default during the lifetime of the contract, the protection seller will pay the protection buyer the full par value of the bond. Since the protection seller 448 FINANCIAL RISK MANAGEMENT will then only be able to recover the value of the bond less loss given default (LGD), the seller will have a loss equal to the par value times the loss given default rate. So the protection seller is in exactly the same financial position as the buyer of an asset swap, receiving fixed spread payments if there is no default, losing the par amount times the loss given default rate if there is a default. A CDS is meant to look like an asset swap, but without the need to invest cash. While this feature makes it very attractive to some investors looking to take on credit risk, it is an even more important product for investors with a negative view of a firm’s credit or who are seeking protection against a firm’s default. These investors previously could only achieve the position they desired by selling short a corporate bond. But the market for borrowing corporate bonds is extremely thin and expensive. The advent of the CDS, like any new forward market, provides far greater liquidity to those wishing to take short positions. (You might wonder why an investor seeking protection against a firm’s default could not just sell the asset causing this exposure. But not all assets exposing an investor to losses when a firm defaults are as easy to sell as a corporate bond. Some may be difficult to sell, such as bank loans and extensions of trade credit; others may be impossible to sell, such as counterparty credit exposure on derivatives.) It also provides far greater liquidity for those wishing to express relative value views that one set of credit spreads will widen relative to another set. The growth of the CDS market has been explosive, growing at a rate of about 100 percent per year in many years. The most troublesome issue in the creation of the CDS market has been difficulties in deciding on a settlement mechanism in the event of default. First, since payoff by the protection seller only occurs in the event of a default, exact definition of a default event must be agreed upon. Does default mean any delay in a scheduled payment of the borrower or only one of a particular magnitude? Is a formal declaration of bankruptcy a necessity? What happens if the terms of the borrower’s debt are voluntarily renegotiated with creditors? (And how can you tell how voluntary it has been? The 2011 and 2012 experience with renegotiation of Greek government bonds has been a particularly worrisome example; see the Economistt article “Fingers on the Trigger” of June 2, 2011.) Second, how should the amount owed by the protection seller to the protection buyer in the event of default be determined, and should this determination involve physical settlement or cash settlement? Third, what becomes of a CDS when the reference firm ceases to exist through merger or acquisition? Multiple solutions have been proposed to these issues with many different variants incorporated into documentation of individual deals. This is a particular headache for market makers in CDSs, who must be certain that transactions that seem to offset one another in terms of tenor and reference Credit Risk 449 entity actually do offset one another when contractual details of settlement procedure are considered. Protection sellers would prefer that the debt instruments used for settlement be as narrow as possible, preferably the single most liquid bond issued by the company. But CDSs with such a narrow class of deliverables have led to severe problems in settlement, with protection buyers having to scramble to purchase a deliverable bond, resulting in driving up the price of that bond so high that it is close to par—the resulting profit between the purchase price and sale to the protection seller at par has not been nearly enough to compensate for actual default losses on which protection was sought. (For more details, see the articles from the Economistt: “Is There Money in Misfortune?” July 16, 1998, and “Of Devils, Details and Default,” December 3, 1998.) It has now become much more common to define a broad class of deliverables, even including much less liquid credit instruments such as bank loans and trade credit. This makes it far easier for the protection buyers, since they can often deliver the actual credit instrument on which they were seeking protection and, in any case, have a wide choice of instruments to deliver. But this has made settlement more difficult for the protection seller, both because of lower liquidity of the instrument being delivered and because the protection buyer’s choice of deliverable instrument gives the buyer a cheapest to deliver option, comparable to the cheapest to deliver option into the Treasury bond future, referenced in Section 10.1.4. Both these effects cause protection sellers to demand higher credit spreads than they would otherwise. We discuss issues of relative pricing between bonds and CDSs in Section 13.1.2.3. Some of the impact on market prices of credit protection buyers scrambling to acquire deliverable instruments can be eased by a cash settlement provision defined in terms of quoted prices for a specified bond. But the illiquidity of corporate bond markets, particularly in conditions following the default of the bond issuer, makes quoted prices suspect. This problem has been greatly exacerbated by the growth of multiname credit derivatives that have resulted in the notional value of CDS contracts referenced to a firm exceeding the total value of the firm’s debt. This has led to the establishment of auction procedures for establishing prices at which cash settlement can take place; see Helwege et al. (2009) for details concerning the auction mechanism and its implications for CDS market participants. Another possible solution to this problem is to have a default swap with a fixed payment in the event of default, known as binary credit default swaps. This resolves the issue of how to determine payment, but may not be a good fit to the risk needs of a holder of a bond or loan. Suppose I am holding a $100 million bond issued by ABC. I can buy a standard default swap on $100 million notional. If the loss in the event of default turns out to 450 FINANCIAL RISK MANAGEMENT be $20 million, it should pay roughly $20 million. If it turns out to be $80 million, it should pay roughly $80 million. However, if I buy a default swap with a fixed dollar payout, I must make a guess as to the loss in the event of default and run the risk that I have either purchased too little protection or paid for too much protection. Default swaps with fixed payoffs are also harder to value since this requires an estimate of the probability of default, whereas a standard bond price is based on the product of the probability of default and loss given default. See Section 13.1.2.1 for further discussion of this point. Default swaps, more than any other derivative instrument, have led to the concept of legal basis risk (see Section 3.2.1). A market maker may believe its risk on a default swap is matched exactly by the protection purchased through another default swap, only to find it has to make a payment under the contractual language of the first swap but receives nothing under the slightly different language of the second swap. The International Swaps and Derivatives Association (ISDA), the industry group that sets standards for derivatives contracts, has made several valiant attempts to remedy the situation by standardizing contract wording. The resulting checklist of possible contract terms is a daunting document. Even so, new disputes continue to arise. ISDA has also established determination committees that rule on disputed issues such as the impact of mergers and whether a renegotiation is voluntary or forced. And ISDA has standardized the auction mechanism for determining prices at which cash settlement can take place. Any firm participating in this market needs to be thoroughly aware of all the relevant history of the disputes, of past actions of ISDA determination committees, and of ISDA auction procedures, and needs to be certain it fully understands the terms of the risk it has taken on. A good synopsis of the ISDA standards and the motivation behind them can be found in Gregory (2010, Section 6.3). For further background on the issues, see Henderson (1998), Falloon (1998), Cass (2000), Bennett (2001), Helwege et al. (2009), and the following articles from the Economistt: “Is There Money in Misfortune?” (July 16, 1998); “Of Devils, Details and Default” (December 3, 1998); “Fixing the Holes” (August 12, 1999); “The Swaps Emperor’s New Clothes” (February 8, 2001); “The Tender Age” (April 20, 2006); and “Fingers on the Trigger” (June 2, 2010). The “Legal and Documentation” section of the ISDA website, found under the “Functional Areas” heading, provides many documents relating to contractual disputes and the findings of determination committees (www2.isda.org/functional‐areas/legal‐and‐ documentation). The total return swap, which we encountered in Section 10.1.7, is another derivative instrument that can be structured for investors who want to take on credit risk without putting up cash. Unlike the CDS, which is designed to look like an asset swap, the total return swap is designed to look 451 Credit Risk like a straight investment in a corporate bond. The mechanics are that the investor enters into a swap in which he receives all of the coupon payments from the bond and any change in the bond price (positive or negative, so he may owe payments) and pays an amount equal to LIBOR times the par amount of the bond. So cash flows are very similar to borrowing at LIBOR and investing in the bond, but with the advantage that the counterparty to the total return swap can use it to create as short position in the bond and thereby express a negative view on the credit or protect a credit exposure. Total return swaps have proved to be far less popular instruments than the CDS, perhaps because asset swap positions are more sought after than fixed‐rate corporate bond positions (for those investors not willing to put up cash) and perhaps because the reliance on a single bond raises settlement issues unfavorable to the investor similar to a CDS with a single deliverable. 13.1.2 Models of Short‐Term Credit Exposure In Section 10.2, we were able to base all modeling of interest rate risk on a single principle, that the value of each individual cash flow that is bundled together in an interest rate contract can be determined independently of the value of any other cash flow bundled in that contract. We would like to use a similar principle for credit instruments, but run into three roadblocks, one having to do with the treatment of credit instruments in bankruptcy proceedings, the second due to the large convexity risks of credit instruments, and the third due to basis risk between bonds and CDSs. Before we can address these issues, we first need a fundamental framework within which we can discuss credit risk. Ultimately, the cost of credit risk must be based on expectations and uncertainty concerning loss from default. Without the possibility of default, credit instruments would just be priced based on the risk‐free discount curve. Default loss can be analyzed into three components as follows: D(I) = PD(B) × LD(I) × AD(I) where I B D(II) PD(B) LD(I) = = = = = (13.1) a credit instrument the borrower on the instrument the default loss on the instrument the probability that the borrower will default the percentage loss on the instrument conditioned on default AD(I) = the amount that will be owed on the instrument conditional on default 452 FINANCIAL RISK MANAGEMENT We use PD(B) instead of PD(I) because cross‐default legal provisions come close to guaranteeing that a borrower will default on either all or none of its debt. For liquid instruments like bonds and CDSs, AD(II) is a fixed amount— the amount of currency borrowed—so we need only concern ourselves in this section with the PD(B) × LD(I) term. In Section 13.2, when we look at illiquid instruments, we will encounter cases (lines of credit and counterparty credit risk) for which AD(I) can vary. Market prices of credit instruments cannot distinguish the effects of default probability and loss given default—in other words, you can extract information from market prices on PD(B) × LD(I), but cannot distinguish between PD(B) and LD(II). To get a clear view of the entanglement of default probability and loss given default in market prices for credit instruments, let’s consider the simplest possible case. Suppose company XYZ has a two‐year zero coupon bond that is trading at $85.50 per $100.00 par amount, while a two‐year zero coupon government bond is trading at $90.00 per $100.00 par amount. The $4.50 haircut on the corporate bond implies that the market is pricing the bond as if the expected loss from default over a two‐year period would be 5 percent ($90 × 95% = $85.50). However, this loss could consist of PD(B) = 5%, LD(I) = 100%; PD(B) = 10%, LD(I) = 50%; or any other combination that results in PD(B) × LD(I) = 5%. This inability of splitting probability of default from expected loss given default will need to be kept in mind when we discuss utilizing market data in internal models of credit risk in Section 13.2, and in models of portfolio credit risk in Section 13.3. 13.1.2.1 Impact of Bankruptcy Law The ability to value all cash flows received on the same date using the same discount factor is a vital assumption in the methodology used to maximize liquidity in the forwards markets, as discussed in Section 10.2. The reason this assumption breaks down for credit instruments relates to provisions of bankruptcy law. In almost all jurisdictions, the claim for a two‐year coupon due on a five‐year bond is not the same in bankruptcy as the claim for the same amount of principal on a two‐ year bond. The common rule for bankruptcy is that the holder of a bond or loan can make a claim on the principal, but not on any coupon interest. Offsetting this loss of interest that can be claimed is the ability to call for immediate payment of principal, regardless of maturity. For bonds or loans trading close to par—that is, the coupon on the bond is close to the current par coupon—the advantage and disadvantage almost cancel out. A five‐year bond loses five years’ worth of coupons, but can accelerate principal due by five years, while a two‐year bond loses only two years’ worth of coupons, but can accelerate principal due by only two years. The par coupon can be thought of as the rate of interest that exactly compensates an Credit Risk 453 investor, at current market discount factors, for deferral of receiving principal; therefore, foregoing coupons on the par coupon bond will precisely offset the acceleration of principal. For similar reasons, a floating‐rate bond or loan, whose coupon resets to current market levels, should have the advantage of principal acceleration closely balance out the loss of coupon payment. However, for bonds or loans selling at a premium, either because of a fixed coupon higher than the current par coupon or a floating rate at a positive spread to current market levels, the bankruptcy rules will cause more of a loss on default than that felt by a par bond or loan. Conversely, a bond or loan selling at a discount will experience less of a loss on default than that experienced by a par bond or loan. As a result, the rule that all cash flows on the same date are equivalent, regardless of what package they are part of, breaks down. A coupon payment is worth more in default if it is packaged as part of a discount bond than a coupon payment for the same date that is packaged as part of a premium bond. Exercise 13.1 familiarizes you with the mathematics needed to deal with this situation. The CreditPricer spreadsheet used in the exercise takes as input the current risk‐free zero coupon curve, an assumed set of annual default rates, and an assumed loss given default rate, and computes the resulting par curve for a corporate bond and resulting spreads to the risk‐free par curve. The calculation looks at the value of payments received if no default occurs plus the accelerated principal payments received if default occurs. You will also find this calculation explained and illustrated in Hull (2012, Section 23.4). The exercise demonstrates that spreads to the risk‐free par curve will differ for differing assumptions of loss given default. This shows that it is not just the product PD(B) × LD(I) that matters in this case, but also the individual components, since the value of the principal acceleration depends on the loss given default assumption. The exercise also shows you how to use the same spreadsheet to solve for market‐implied default rates based on an observed par curve and an assumed loss given default rate. It further shows that if prices are available for several coupons with the same maturity, then information about the split between PD(B) and LD(II) can be extracted. One issue in which the difficulty in splitting market quotes into probability of default and loss given default components involves binary credit default swaps. The price of a binary CDS should just be the cost of a standard CDS divided by 1 – LD(I), since the standard CDS will pay 1 – LD(I) dollars for each dollar of principal of the CDS in the event of default, but the binary CDS pays the full principal in the event of default. 13.1.2.2 Convexity of Credit Instruments To illustrate the difficulty that convexity poses for credit risk management based on short‐term exposure to 454 FINANCIAL RISK MANAGEMENT market prices, consider the following simple example. (By contrast, convexity has little impact on interest rate instruments; see Section 10.4.) Consider two obligations of company XYZ: a two‐year zero coupon bond and a 10‐ year zero coupon bond. Assume that a risk‐free two‐year zero is trading at $90 per $100 par value and a risk‐free 10‐year zero is trading at $60 per $100 par value. If the expected loss from default for XYZ is roughly 1 percent a year, we would expect to see a haircut for the two‐year zero of $90.00 × 2% = $1.80 and a haircut for the 10‐year zero of $60.00 × 10% = $6.00. If market confidence in XYZ worsened slightly, expected loss from default might rise from about 1 percent a year to about 1.1 percent a year, resulting in a haircut for the two‐year zero of $90.00 × 2.2% = $1.98 and a haircut for the 10‐year zero of $60.00 × 11% = $6.60. Therefore, the two‐year zero has moved by $1.98 – $1.80 = $0.18 and the 10‐year zero has moved by $6.60 – $6.00 = $0.60, a ratio of $0.60/$0.18 = 3.33, which could also be derived as a ratio of the durations multiplied by the present values: (10 × $60)/(2 × $90). If you want to hedge against small moves in a credit spread, you would sell short $30 million 10‐year bonds against a long position of $100 million two‐year bonds. But what happens if XYZ defaults? You have losses on $100 million balanced by gains on only $30 million. The right ratio for hedging short‐term market movements is an extremely poor ratio for hedging default, due to the severe convexity. The higher the PD(B) component of the in PD(B) × LD(I) product, the greater the probability of default, and the more significant the convexity risk. For large moves that do not go all the way to default, as might be associated with a credit downgrade, a mismatch in correct hedging ratios will still occur, but it will be less severe. This example demonstrates that risk management utilizing short‐term exposures to changes in market price is not sufficient by itself; it needs to be supplemented by an analysis of ultimate default risk. 13.1.2.3 CDS‐Bond Basis Risk To understand the basis risk between CDS and bonds, we must first start with the theoretical arbitrage relationship between them and then see what factors might alter it. We will us a simple illustrative example of the arbitrage relationship, a fuller discussion of which can be found in Duffie and Singleton (2003, Section 8.3). Under ideal circumstances, the spread above LIBOR on a floating‐rate bond issued by a corporation (call this spread S) ought to be equal to the CDS spread for the same maturity (call this spread C). If the purchaser of the bond also purchases a CDS of the same tenor, his return if the issuer does not default is LIBOR + S – C each year plus the return of his principal. If the issuer does default, he can exchange the bond for par under the terms of the CDS. Since the investor always gets back his principal, he has an investment with no credit risk Credit Risk 455 on which the return ought to be LIBOR; hence LIBOR + S – C should equal LIBOR, so S should equal C. In practice, very few companies issue floating‐rate bonds, but an asset swap can be used to turn a fixed‐rate bond into a close approximation of a floating‐rate bond. So the CDS spread ought to equal the spread over LIBOR that the fixed‐rate coupon can be exchanged for in the interest rate swap market, which is the spread between the coupon rate and the swap rate for the bond’s tenor. So why should the actual basis between a CDS spread and a bond’s coupon spread to the swap rate be different than zero? Partly it is because the asset swap is not a perfect substitute for a floating‐rate bond, and partly it is because of features of the CDS that have not been accounted for in the above idealization, such as the cheapest‐to‐deliver option discussed in Section 13.1.1.2. The seminal article on the CDS‐bond basis is Lehman Brothers’ “Explaining the Basis: Cash versus Default Swaps” by O’Kane and McAdie (2001). It analyzes many factors that potentially could make the CDS spread greater than the bond spread (“increase the default swap spread” in the terminology of the paper) or make the bond spread greater than the CDS spread (“decrease the default swap spread”). A summary can be found in O’Kane (2008, Chapter 5). Other sources worth consulting are: ■ ■ ■ ■ DeWit (2006). DeWit’s discussion of factors driving the basis in Section 2 of his paper leans heavily on O’Kane and McAdie, but he adds some analysis and a very comprehensive set of footnotes with references to both empirical and theoretical articles. Table 6 gives a concise comparison of empirical research on the size of the basis, which centers around 5 to 10 basis points with CDS spreads higher than bond spreads. DeWit states: “While we define 14 different economic basis drivers, it is our understanding that four of them (i.e. the CDS cheapest to deliver option, difficulties in shorting cash bonds in a context of structural demand for protection, relative liquidity in segmented markets, and synthetic CDS issuance) are the main determinants of the CDS‐bond basis.” Hull, Predescu, and White (2004) also present empirical evidence that supports the same conclusion as DeWit’s Table 6. Duffie and Singleton (2003, Section 8.3) analyzes the CDS‐bond basis. They are less inclusive than O’Kane and McAdie in considering all possible influences, but are worth looking at for the depth of their analysis of the impact of the difficulty in shorting bonds. The historical relationship of CDS trading about 5 or 10 basis points higher than bond spreads was severely disrupted by the 2007–2008 crisis, with spreads going negative by 250 basis points for 456 FINANCIAL RISK MANAGEMENT investment‐grade firms and by 650 basis points for high‐yield names; see Bai and Collin‐Dufresne (2011) for a detailed discussion of both market behavior and possible causes. The two major drivers of this disruption appear to be: 1. Funding cost. Many holders of cash bonds were now funding at substantially higher rates than LIBOR. While the high bond spread relative to CDS spreads would then seem to offer an arbitrage opportunity to those who could still fund at LIBOR, there may be have been little appetite for such arbitrage in the current environment (a Reuters article “Popular US Credit Trade Turns Sour” of December 13, 2007, stated that “lack of financial balance sheet capacity and a general unwillingness to lend has prolonged the negative basis”). 2. Heightened concern for counterparty risk. If a CDS is not fully collateralized, the buyer of CDS protection may be unwilling to pay the full cost of default risk. 13.1.3 Risk Reporting for Market Credit Exposures A good starting point for risk reporting of market credit risk is to closely parallel the reporting guidelines for forward risk given in Section 10.4. As with forward risk, key questions for market credit risk involve selection of maturity buckets and selection of summary statistics, such as exposure to a parallel shift in the credit spread curve and exposure to linear tilt of the credit spread curve. A measure of credit spread duration is calculated in close parallel to the duration measure for rates and serves as an alternative to the value of a basis point shift in the credit spread curve. Because economic events that have an impact on default probabilities often impact the credit spreads of more vulnerable firms more than those of higher credit quality, many firms utilize a measure of percentage change in credit spread as an alternative to or supplement to a measure of impact of a parallel shift in credit spread. For example, a measure of a 5 percent increase in credit spreads would add together the impact of a 5‐basis‐point increase in a credit that currently has a 100‐basis‐point credit spread with the impact of a 25‐basis‐point increase on a credit that currently has a 500‐basis‐point credit spread. There are two key added factors for the measurement of credit spread exposures relative to rate exposures. One is that credit spreads have far more characteristics to be taken into account when grouping exposures— geographic, industry, and credit quality. The second is the importance of price jumps and convexity for credit spreads, which is of little importance for forwards. Credit Risk 457 Let’s look at grouping characteristics first. As with equity spot risk in Section 9.3, grouping of exposures and limits by geography and industry make sense. For corporate credit, equity exposure and credit spread exposure are two aspects of risk exposure to corporations, so the groupings used should be very similar. All levels of management should see total net credit exposure along with exposure to major geographic regions (e.g., United States, Western Europe, developed Asia, emerging markets) and major industry groups, while lower levels of management should see more detailed net credit exposures by country and specific industries. Reporting and limits for exposure to individual borrowers is also needed. Finally, grouping of exposures and limits are needed for credit quality, with rating agency grades, such as Aa, A, Baa, and Ba, often being used. Given the large impact of convexity on credit spread exposures, as discussed in Section 13.1.2.2, it very important to have measures and limits that capture this risk. Measures and limits that capture default risk will be discussed in Section 13.2. For large credit spread shifts, the most intuitively appealing are measures of and limits on the amount that can be lost in the event of very large shifts in credit spread that might be associated with a major shift in the economic environment. So you might have a measure of exposure to a 1 percent shift in credit spreads to control for ordinary market moves, and a measure of exposure to a 10 percent shift in credit spreads to control for a large move. There is an obvious parallel to the delta and convexity limits on options positions, discussed in Section 11.4. 13.2 MODELING SINGLE‐NAME CREDIT RISK Models of single‐name credit risk are important for several reasons: ■ ■ ■ If you have exposure to a single‐name credit instrument for which you can’t obtain a liquid market price, you will need a model to value it. Even when you can obtain a liquid market price, comparison to a modeled price can be useful in informing trading decisions. Single‐name credit instrument models serve as important inputs to credit portfolio and multiname credit instrument models. Since credit portfolios and multiname credit instruments need to be evaluated over long‐term horizons, just having a liquid price for constituent pieces is not adequate—a model of possible price evolution is also required. The key element in modeling any single‐name credit instrument is modeling expected default loss, since, absent default loss, the instrument is just 458 FINANCIAL RISK MANAGEMENT an interest rate instrument, whose modeling we have already studied in Chapter 10. Referring back to Equation 13.1 in Section 13.1.2, the default loss on a credit instrument can be written as D(I) = PD(B) × LD(I) × AD(I) that is, the product of probability of default, loss given default, and the amount that will be owed conditional on default. The best way of organizing the modeling of default loss is to model these three components separately. Partly, this is just an aid to clear thinking. Partly, it is motivated by probability of default being a function of the borrower, independent of the instrument, while the other two components are instrument dependent. And partly, this is a matter of expertise: Those who are most expert in modeling loss given default may be lending officers with experience in loan work‐outs of borrowers threatened by bankruptcy, while probability of default may best be modeled by those with direct knowledge of a particular firm or industry. Our discussion in this section is accordingly separated into sections on estimating probability of default (13.2.1), estimating loss given default (13.2.2), and estimating amount owed conditional on default (13.2.3). Section 13.2.4 looks at information relative to defaults that can be derived from prices for equity and equity options utilizing an option‐theoretic approach, a topic that cuts across both probability of default and loss given default. 13.2.1 Estimating Probability of Default Default probability is the most critical and most intensely studied of the components of single‐name credit risk. Almost all firms that deal in credit risk instruments will want to form their own assessments of default probability (only if credit instruments are only a small portion of the investment portfolio and are almost all liquid might a firm be satisfied with just basing this assessment on input from an outside service). Firms with heavy investment in credit instruments, such as traditional banks, will devote considerable resources to their own determination of default probability. But all firms should be aware of and make use of independent assessments of default probability, both as input to their own judgments and as reality checks. This is particularly true when the credit valuation is for a business with which the lending firm has a long relationship and detailed, intimate knowledge of the business’s management and operations. Caution needs to be exercised in such cases, since close, longtime relationships can breed complacency and a reluctance to acknowledge unwelcome changes. It is important to have an internal review mechanism in which internal credit ratings that show lower default probabilities than agency default probabilities or Credit Risk 459 those derived from the models or markets are challenged. The review mechanism needs to be run by people with good experience in the credit area but who don’t have the direct client involvement that may lead to complacency. We have divided up the possible sources of independent default probabilities into two broad categories. The first, and most widely used, is direct comparison to rating agency evaluations. The second is the use of statistical modeling that may take as input borrower‐specific information, gauges of the broad economy, and market prices. This last category will lead us into the area of option‐theoretic models, discussed in Section 13.2.4. We discuss these two categories in turn. 13.2.1.1 Rating Agency Evaluations The primary output of rating agency evaluation of individual borrowers is a letter grade. Translating letter grades into default probabilities requires some analysis, but the ratings agencies provide an abundant amount of historical data that can be utilized to make this conversion. While all of the rating agencies provide such historical data, I will, for convenience, make all my references in this section to the Moody’s data, which is updated regularly and appears to be easily available to the public on the web, through the National Association of Insurance Commissioners at www .naic.org. All the data quoted and used in tables comes from Moody’s (2011a). It should be noted that any use of historical rating agency data to translate from current agency ratings to default probabilities does rest on the assumption that the ratings assignment process has been reasonably stable and consistent over time. Arguments for this being a reasonable assumption can be found in the section “The Rating Process” in de Servigny and Renault (2004, Chapter 2), for example, “The criteria according to which any assessment is provided are very strictly defined and constitute the intangible assets of ratings agencies, accumulated over years of experience. Any change in criteria is typically discussed at a worldwide level.” The translation of rating agency grades to default probabilities generally starts with transition matrices that show the probability over a fixed time period that a credit rated in one category at the beginning of the period will default during the period or will transition to another credit rating category at the end of the period. Tables 13.1 and 13.2 show a sample one‐year transition matrix and a cumulative transition matrix that only looks at default. Rating agencies also publish matrices covering many different transition periods (for example, two‐year transitions, three‐year transitions, and so on); matrices with finer credit rating graduations; and matrices based on subsets of this historical data. There are a variety of approaches in using this data to convert agency ratings into default probabilities. Here are some of the major differences. As the one‐year transition matrix in Table 13.1 shows, some borrowers who 460 87.40% 0.97% 0.06% 0.04% 0.01% 0.01% 0.00% 0.00% 8.63% 85.62% 2.69% 0.18% 0.06% 0.03% 0.02% 0.00% Aa Source: Moody’s (2011a, Exhibit 27). Ca–C Aaa Aa A Baa Ba B Caa Aaa 0.60% 7.97% 86.76% 4.52% 0.37% 0.13% 0.02% 0.00% A 0.01% 0.36% 5.27% 84.52% 5.64% 0.34% 0.14% 0.00% Baa One‐Year Transition Matrix, 1970–2010 Initial Rating TABLE 13.1 0.03% 0.05% 0.49% 4.11% 75.76% 4.76% 0.46% 0.32% Ba 0.00% 0.02% 0.11% 0.78% 7.24% 73.52% 8.26% 2.37% B Rating at Year‐End 0.00% 0.01% 0.03% 0.17% 0.53% 5.77% 60.09% 8.88% Caa 0.00% 0.00% 0.00% 0.02% 0.08% 0.67% 4.10% 36.27% Ca–C 0.00% 0.02% 0.05% 0.18% 1.10% 4.23% 14.72% 35.45% Default 3.34% 5.00% 4.53% 5.48% 9.21% 10.54% 12.18% 16.70% Withdrawn 461 Credit Risk TABLE 13.2 Initial Rating Aaa Cumulative Default Rates, 1970–2010 Number of Years 1 0.00% 2 0.01% 3 0.01% 4 0.04% 5 0.10% 7 0.24% 10 0.49% 15 0.92% Aa A Baa Ba B 0.02% 0.06% 0.10% 0.18% 0.27% 0.44% 0.62% 1.26% 0.05% 0.18% 0.36% 0.55% 0.76% 1.24% 2.14% 3.66% 0.18% 0.51% 0.93% 1.43% 1.95% 3.03% 4.90% 8.85% 1.16% 3.19% 5.60% 8.15% 10.45% 14.44% 20.10% 29.70% 4.47% 10.43% 16.33% 21.51% 26.17% 34.72% 44.57% 56.35% Caa–C 18.16% 30.20% 39.71% 47.32% 53.77% 61.18% 72.38% 76.16% Source: Moody’s (2011a, Exhibit 34). receive a rating at the beginning of a period are no longer tracked by the end of the period because they have asked the agency to withdraw its rating. In projecting transition probabilities, a choice must be made between assuming that a request for rating withdrawal indicates anticipation of a downgrade and assuming that a request for rating withdrawal carries no information content or some intermediate assumption (see de Servigny and Renault 2004, Appendix 2A). ■ ■ ■ Trade‐offs exist between using multiyear default data based on the direct observation of cumulative default rates versus generating multiyear cumulative default rates by the matrix multiplication of one‐year transition matrices. The direct use of cumulative default rates suffers from a diminishing data pool for longer tenors and greater potential inaccuracy from withdrawn ratings (firms whose ratings are no longer tracked) (see Gupton, Finger, and Bhatia 1997, Section 6.3.2). Matrix multiplication assumes a Markovian process, where no serial correlation exists between transitions. Alternatively, it could be desirable to derive one‐year transition matrices that are consistent with observed longer‐term cumulative default and transition behavior (see Gupton, Finger, and Bhatia 1997, Section 6.4). However, there is data suggesting that serial correlation between transitions does exist (see Bahar and Nagpal 2000). Default probabilities for tenors that fall in between those for which transition matrices are published can be interpolated (see de Servigny and Renault 2004, Appendix 2A). Rating agencies are very frank about the fact that their ratings represent through‐the‐cycle as opposed to point‐in‐the‐cycle ratings (see de 462 ■ ■ ■ ■ ■ FINANCIAL RISK MANAGEMENT Servigny and Renault 2004, “Time Horizon for External Ratings” in Chapter 2). Ratings are not adjusted just because a movement from an expansionary phase of the economic cycle to a recession increases the likelihood of defaults. Conversion to default probabilities that accurately reflect the current economic environment can be made using data such as that presented in Table 13.3, which shows how five‐year default probabilities differed by starting year. This data could then be correlated with information on the stage of the economic cycle each five‐year period represents. Default rates and transition matrices could be adjusted for the current stage in the economic cycle, based on historical observation of differences during recession and growth periods. Large lending firms may have their own internal data on defaults and transitions that they may want to use to supplement the publicly available data that comes from the ratings agencies. However, even if this data has been well maintained, a trade‐off exists between using data that is more relevant to the particular class of borrowers who are customers of a particular firm and the loss of accuracy that comes from the utilization of a smaller sample. If default and transition data is available broken out by country and industry, this could be used to refine the data available from the ratings agencies. One criticism of ratings agency data is that they are largely based on experience with U.S. firms; see the sections “Quality of Transition Matrices over Time and Region” and “Industry and Geography Homogeneity” in de Servigny and Renault (2004, Chapter 2). However, the same points about small data samples raised in the last bullet may be relevant here. The tables and discussion in this section have referred only to corporate borrowers. The rating agencies publish comparable transition matrices for sovereign government borrowers (see Moody’s 2011b) and other government borrowers, such as municipalities (see Moody’s 2010). Default and transition data from different sources can be blended, such as averaging S&P and Moody’s data, or rating agency and private data. A frequently expressed concern is that agency credit ratings are not updated often enough to fully reflect the probability of default. It reflects the nature of the rating process, which, because of the serious consequences to a firm’s financial health a ratings change can entail, requires that changes be thoroughly deliberated and well documented. This may supply the motivation to supplement this source of independent default probabilities with one of the two other sources we will now discuss. 463 Credit Risk TABLE 13.3 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 Five‐Year Default Rates based on data from Moody’s (2011) Exhibit 42 Aaa Aa A Baa Ba B 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 2.395% 1.449% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.810% 0.797% 0.000% 0.000% 0.000% 0.487% 1.832% 0.789% 1.230% 0.394% 1.001% 0.618% 0.000% 0.316% 0.284% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.231% 0.413% 0.736% 0.336% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.576% 0.832% 0.267% 1.118% 0.262% 1.504% 2.547% 1.865% 1.925% 1.431% 0.758% 0.000% 0.000% 0.000% 0.000% 0.000% 0.000% 0.122% 0.336% 0.540% 0.759% 0.903% 0.729% 0.355% 0.000% 1.355% 1.412% 1.079% 2.061% 1.941% 1.784% 0.819% 0.910% 0.594% 1.406% 2.046% 1.706% 3.365% 2.477% 3.776% 1.729% 2.876% 5.813% 5.006% 3.991% 3.289% 0.629% 0.275% 0.000% 0.547% 0.673% 1.583% 1.327% 2.472% 3.283% 3.100% 2.848% 2.791% 1.949% 0.329% 0.316% 7.058% 3.827% 3.549% 3.206% 4.050% 3.847% 3.731% 2.903% 4.559% 5.980% 8.730% 11.708% 18.565% 14.205% 18.146% 18.452% 20.319% 22.846% 22.673% 23.439% 18.292% 9.527% 2.495% 4.427% 4.989% 7.538% 8.513% 12.393% 13.918% 10.852% 6.281% 6.285% 6.490% 3.155% 3.903% 22.586% 3.846% 7.018% 3.774% 7.143% 9.566% 4.082% 15.249% 22.612% 17.423% 28.033% 27.985% 30.063% 28.702% 27.363% 30.568% 34.910% 40.224% 39.494% 43.744% 38.319% 33.874% 27.086% 19.208% 16.683% 16.651% 18.447% 26.515% 34.987% 37.442% 34.574% 30.406% 18.057% 9.995% 8.633% (continued) 464 TABLE 13.3 FINANCIAL RISK MANAGEMENT (Continued) Aaa Aa A Baa Ba B 2005 2006 0.000% 0.000% 0.210% 0.898% 1.388% 1.496% 1.986% 1.238% 9.231% 9.873% 20.057% 20.838% Mean StdDev Max Max − Mean/SD 0.104% 0.454% 2.395% 5.041 0.267% 0.441% 1.832% 3.549 0.610% 0.677% 2.547% 2.860 1.984% 1.349% 5.813% 2.838 9.729% 6.504% 23.439% 2.108 23.139% 11.394% 43.744% 1.808 13.2.1.2 Statistical Modeling The seminal concept in statistical modeling of default probabilities was Edward Altman’s 1968 Z‐score model that related probability of corporate default to firm‐specific accounting ratios—the ratio to total assets of working capital, retained earnings, earnings before interest and taxes, and sales—and one market price, the market value of equity. Bohn and Stein (2009, Chapter 4) and Saunders and Allen (2010, Chapter 6) give a good exposition of the current state of these models. Market prices can be used in statistical models of default probability in one of three ways. The first is the way Altman used the market value of a firm’s equity in his Z‐score model, as just an independent variable in a regression model or discriminant analysis. The second is to try to bring more theoretical structure to the relationship between equity market prices and default probability, the option‐theoretic models we will examine in Section 13.2.4. The third is to try to find a structural relationship between bond and CDS market prices and default probability. Linking bond and CDS market prices to default probability could be useful in several ways. A bank that is holding too much debt of a particular borrower to be able to consider using the CDS market to liquidate the risk and which therefore must manage the risk using a longer‐term portfolio management approach would still be interested in finding out the default probability that is built into the market price—CDS spreads may reflect new information faster than the bank’s internal review process and would be valuable as input to the internal process. Even when no liquid market exists for the bonds or CDSs of a particular borrower, it might be possible to construct an index of liquid bonds and CDSs for other borrowers related by similar characteristics (e.g., credit rating, nationality, industry), and deriving a default probability for this index could be similarly valuable input to the bank’s internal review process. There are two barriers that must be overcome in deriving default probabilities from market credit spreads. The first is the one discussed in Section 13.1.2, the inability to separate default probability from loss given Credit Risk 465 default (LGD). This would need to be addressed by making a reasonable assumption for the LGD and then deriving the default probability implied by the credit spread. The second barrier is the large difference between actual default probabilities and those implied by market rates, due largely to the systematic risk embedded in credit exposure (this will be discussed further in Section 13.4.4). This difference has been studied extensively over the past few years; good summaries are in Hull (2012, Section 23.5) and Amato and Remolona (2003). Actual default probabilities can be inferred from market‐ implied default probabilities based on observed historical relationships. The downside to this latter approach is that changes in debt prices may reflect many factors other than changes in market sentiment about default probability; technical liquidity factors or changes in the willingness to take on systematic risk can dominate. And even when a borrower does have liquid bonds and CDSs, they may not be very liquid and may not provide an up‐to‐date assessment of market sentiment on the firm’s credit risk. Stock prices are generally far more liquid and less subject to, though not immune to, being impacted by technical liquidity factors (and equity is certainly subject to the same buffering as debt by changes in willingness to take on systematic risk). The greater liquidity of stock prices is a major driving factor behind the use of the option‐theoretic models for credit. 13.2.2 Estimating Loss Given Default De Servigny and Renault (2004, Chapter 4) and Bohn and Stein (2009, Chapter 5) are good introductions to the general topic of estimating loss given default (LGD). Statistical estimates of LGD have been published by the credit rating agencies. A few other published studies are available as well. De Servigny and Renault (2004, Chapter 4); Altman, Resti, and Sironi (2001, Appendix III.1); and Gupton, Finger, and Bhatia (1997, Chapter 7) offer good discussions of the public data available. Table 13.4 provides results from the Moody’s study for bond defaults occurring from 1982 to 2010, as reported in Moody’s (2011a). Distinctions are drawn based on the relative seniority of debt, with bank loans regarded as a separate seniority class from bonds. Published studies usually show recovery rates, which are 100 percent minus the LGD rate, but I have translated into LGD. The measurement of historical LGD can be performed in two different ways. One is to observe the drop in market prices for an instrument about one month after the announcement of default, and is shown in the column labeled “Measured by Postdefault Trading Prices” in the table. The second is to track all cash eventually received in the settlement of claims and to present value these future receipts back to the date of default, utilizing a 466 TABLE 13.4 FINANCIAL RISK MANAGEMENT Comparison of Rates of Loss Given Default Seniority Class First lien bank loans Second lien bank loans Senior unsecured loans Senior secured bonds Senior unsecured bonds Senior subordinated bonds Subordinated bonds Junior subordinated bonds Measured by Ultimate Measured by Postdefault Recoveries (1987–2010) Trading Prices (1982–2010) 19.7% 36.5% 50.8% 70.6% 70.7% 81.6% 34.2% 70.9% 52.2% 49.2% 63.3% 69.3% 68.7% 75.3% Source: Based on Moody’s (2011a, Exhibits 7 and 9). discount rate that suitably reflects the uncertainty of recovery. This measure is shown in Table 13.4 in the column labeled “Measured by Ultimate Recoveries.” Gupton, Finger, and Bhatia (1997, Section 7.1) cite academic studies that conclude that the “bond market efficiently prices future realized liquidation values,” supporting a rough equivalence of these two methods. This conclusion is consistent with the data in Table 13.4. Bohn and Stein (2009, Chapter 6) cite a Moody’s study by Varma and Cantor that “determined that the single B bond spread provided a reasonable proxy for the discount rate that, on average, equated” these two measures. Which measure is more relevant depends on usage. In the context of the management of liquid credit instruments in Section 13.1, postdefault trading prices would be more in line with the exit price approach for liquid instruments. Managers of less liquid credit portfolios would have more flexibility in deciding which method of recovery was more promising for each default event. All losses should be expressed as a percentage of par, given that bankruptcy law uses par amount of the instrument as the basis for a claim (as discussed in Section 13.1.2.1). Volatility of LGD rates is an important issue for the credit portfolio simulations discussed in Section 13.3.2. Tables 4.4 and 4.5 of de Servigny and Renault (2004) display statistics on volatility of LGD rates by seniority class, showing standard deviations in the 25 to 35 percent range. Parallel to our discussion on the estimation of the risk of default, firms may want to supplement published data on LGD with their own internal data. This is particularly an issue with non‐U.S. debt and bank loans. Published data on loss given default is heavily weighted toward the U.S. market, but bankruptcy laws and procedures differ substantially by country and may thus be expected to impact recovery rates. Recovery rate has also been shown Credit Risk 467 to differ significantly by industry; see de Servigny and Renault (2004, Chapter 4) for data in Table 4.5 and discussion; in particular, de Servigny and Renault suggest that “what may appear as an industry effect may actually reflect differences in collateral quality offered by firms in various industries.” The lower loss given default rate on bank loans can be presumed to be due to the attention banks pay to the negotiation of security against default. However, this attention may vary between banks and, even within a bank, by loan type. Firms putting together their own internal data on LGD must be careful in compiling the data on ultimate recoveries. Gupton and Stein (2005, Section 4.3.1) point to a 1999 Moody’s study “Debt Recoveries for Corporate Bankruptcies” by David Hamilton and Lea Carty showing that “15% of the value of recoveries for Senior Secured Loans came in the form of equity of the defaulted form.” Gupton and Stein then comment: Since these payments with equity interests (e.g., common stock, preferred, and warrants) commonly do not trade, their value will be unclear and unrealized for years. When these equity values are eventually realized/known (often well past the write‐off date), it would be atypical for a bank’s accounting system to track flows back to the original charge‐off. When we assist clients in databasing their own institution’s LGD histories, we have always found it necessary to examine archived paper records. The full tracking of default resolution realized values (cash flows) has been far more informative than sourcing simply the accounting write‐offs. Economic modeling of LGD has not received as much attention as economic modeling of probability of default. Jacobs and Karagozoglu (2011), Altman and Kalotay (2010), and Bohn and Stein (2009, Chapter 5) each present economic models for LGD along with discussion of the relevant literature. Moody’s KMV has developed a commercial economic forecasting model for LGD; see Gupton and Stein (2005). Even when forecasts are based on the judgments of experienced credit managers, it is still advisable to be aware of the economic models, at least for sensitivity to the factors that have proved most important. Along with loan structure and ranking of collateral, Bohn and Stein find macroeconomic environment (state of the economy, industry) and firm leverage among the significant factors. Jacobs and Karagozoglu also find firm size to be significant. Gupton and Stein also utilize KMV’s distance‐to‐default measures (discussed in Section 13.1.4) for the firm, the industry average, and the geographic region average. An issue that has drawn significant recent attention is the correlation between the occurrence of default and the rate of loss given default. This is the focus of a report submitted by Altman, Resti, and Sironi (2001) to the 468 FINANCIAL RISK MANAGEMENT International Swaps and Derivatives Association. This study finds significant negative correlation between the occurrence of default and recovery rate, which translates to a strong positive correlation between the occurrence of default and loss given default. This is not surprising on economic grounds, since an economic recession is likely to trigger more defaults while also negatively impacting the ability of a bankrupt firm to realize value on its remaining assets. This correlation has much the same effect as an increase in the level of correlation between defaults, since both result in more clustering of default losses. For example, if we’re projecting the possible default losses for the next year, we might experience a good period for the overall economy that leads to few defaults and small losses on the defaults that do occur, or we might experience a recession that leads to many defaults and a high level of losses on these defaults. To the extent default losses cluster, it implies the need for added capital to guard against large losses, as discussed in Section 13.3.2, and a lower valuation of the senior tranches in CDOs, as discussed in Section 13.4.1. 13.2.3 Estimating the Amount Owed at Default For loans and bonds, amount owed at default is simply the par amount. But for lines of credit and counterparty credit on derivatives, the amount owed at default needs to be modeled. We will consider the modeling of the amount owed at default for counterparty credit on derivatives in Chapter 14. Here we will confine our discussion to lines of credit. Lines of credit enable a borrower to draw funds as needed up to some maximum amount, subject to various terms and conditions. From a completely pessimistic view, AD(I) would be set for a credit line equal to the maximum amount that can be drawn, since just prior to default a borrower will likely try to maximize the use of all available sources of credit. However, this fails to take into account some of the contractual terms that the lender can employ to limit credit line usage when the credit rating of the borrower is declining. It is thus possible that AD(I) will be less than the maximum amount that can be drawn. Two principal forms of credit lines are available—those used for working capital and those used as backstops for commercial paper issuance. Working capital credit lines give a borrower the flexibility of only paying full interest on the amount of funds it needs at a particular point of time without losing the security of knowing that it can draw down a precommitted amount as needed. Commercial paper backup lines act as a safety net for commercial paper issuers. Commercial paper issuance typically occurs for very short time periods, often only a few days, to accommodate the liquidity needs of commercial paper investors. The tenor of the commercial paper is usually shorter Credit Risk 469 than the borrowing need of the commercial paper issuer, leaving the issuer vulnerable to an inability to roll the paper over at maturity, but also leaving the investor vulnerable to not being paid back in the event of rollover failure. The backup line gives assurance to both the borrower and investor in the event of a liquidity squeeze. A backup line is consequently insisted on by rating agencies as a prerequisite for an investment‐grade credit rating on a firm’s commercial paper. Usage on commercial paper backup lines is virtually zero, except in the rare case of rollover difficulty. In measuring the loss given default of credit lines, average usage is obviously of little value, since it fails to deal with the high correlation between line usage and credit deterioration. The key is how much usage will there be if default occurs. As noted previously, backup line usage averages close to zero, but when the lines are used, it is because credit difficulties make rolling commercial paper problematic. If only 1 percent of all commercial paper issuers default, but all of these have their lines drawn by 100 percent just prior to default, and if 0 percent usage appears on the remaining 99 percent of issuers, then the overall line usage will be only 1 percent, but default losses will be just as great as if overall line usage is 100 percent. If credit lines are viewed simply as an option to draw funds exercisable by the borrower, then line usage should be assumed to be 100 percent in the event of default. However, this option is not unconstrained, given that covenants that form part of the contract for the line give lenders the opportunity to reduce line availability in the event of credit deterioration. There will, on one hand, be competitive pressures on the bank not to exercise its full rights under these covenants to avoid damaging the particular relationship and to maintain a reputation with customers as being reliable in a crisis. On the other hand, a bank can pressure a customer to renegotiate loan terms. Araten and Jacobs (2001) aptly describe credit line usage in the event of default as “the outcome of the race between the bank and the borrower with regard to the draw‐down of unused commitments in adverse circumstances.” When a result is the product of complex behavioral assumptions, it is not surprising to see that the dominant method of analysis is historical statistical study. Araten and Jacobs (2001) published the most complete analysis based on a study of 399 defaulted borrowers at Chase Manhattan Bank over a 53∕4‐year period, ending in December 2000. Their main results are shown in Table 13.5. As would be expected, average usage upon default rises with the time elapsed between when a line is committed and when default occurs. This is because the longer the time period elapsed, the more likely that a borrower who started as higher grade and subject to fewer covenants has slipped downward in credit grade. Similar reasoning explains the finding 470 39.7% (22) 26.7% (14) 26.5% (86) 24.5% (100) 32.9% (418) B/B– CCC Total BB–/B – + 46.6% (254) 44.9% (84) 49.8% (100) BBB/BBB– BB 32.0% (81) 39.6% (129) 2 12.1% (1) 75.5% (6) 47.2% (7) 52.1% (20) 1 78.7% (3) 93.9% (1) 54.8% (18) AAA/AA A BBB+/BBB + Facility Risk Grade 62.1% (103) 9.4% (1) 37.3% (5) 62.1% (45) 62.1% (37) 84.0% (1) 41.7% (5) 41.5% (9) 3 68.7% (59) 97.8% (2) 76.0% (17) 62.6% (25) 100% (2) 37.5% (3) 4 Time to Default (in Years) 71.8% (59) 68.3% (4) 100% (4) 100% (2) 5–6 Average Usage Conditional on Default by Facility Risk Grade and Time to Default for Revolving Credits Number of observations in parentheses TABLE 13.5 43.4% (834) 24.6% (115) 30.7% (115) 46.4% (231) 50.1% (295) 12.1% (1) 77.2% (10) 55.5% (15) 52.2% (52) Total Credit Risk 471 that average usage upon default tends to rise with a higher initial credit rating. Of course, it is less likely that a higher‐rated credit will default compared to a lower‐rated credit, but for those who do default, the lower level of covenants results in higher usage. 13.2.4 The Option‐Theoretic Approach Before expounding on the option‐theoretic approach, let us review why it would be very useful to have a model that relates a firm’s equity price to credit spreads, default probability, and loss given default. First, as noted toward the end of Section 13.2.1.2, the generally greater liquidity and more frequently available quotes of equity prices relative to debt prices makes this an attractive potential driver of inputs to portfolio credit models. Second, the greater availability of historical stock price data makes it attractive as a driver of default correlation models, as we will see in Section 13.3.1. Third, credit spreads derived from equity prices can be a useful input to trading decisions about which credit instruments represent good investment values. Fourth, models of correlations between equity prices and credit spreads can be valuable tools in building models of products, such as convertible bonds, that are hybrids of equity and debt. Fifth, models of correlations between equity prices and credit spreads can be useful input to the creation of stress scenarios. And sixth, certain trading strategies, termed capital structure arbitrage, use option‐theoretic analysis to identify mispriced relationships between debt instruments and equity options; Morini (2011, Section 11.2) offers an extensive discussion of these strategies and possible difficulties they may encounter. In the option‐theoretic approach, a firm’s equity is viewed as a call option on the value of the firm’s assets with a strike price equal to the face value of the firm’s debt. This is equivalent to viewing the equity owners of a firm as having a put option to pay off the debt holders with either the face value of the debt or the total value of the firm’s assets, whichever is smaller. So the total economic value of the firm’s debt to the debt holders must be the face value of the debt less the value of this put option. Let us first look at a very simple version of the options model, basically corresponding to the original Merton model, which can be found in Hull (2012, Section 23.6). It is extremely useful as a first approximation, since we will see that it provides a precise relationship between all of the elements we are trying to link with very little computational burden. This model has four key simplifying assumptions: 1. The firm has only a single class of debt outstanding, a zero coupon debt, and the firm will not issue any new debt before this debt matures. 472 FINANCIAL RISK MANAGEMENT 2. If the firm defaults, this will only occur at the time of the maturity of this debt. 3. The firm’s behavior, such as the riskiness of its investments, will not be impacted by how close it is to default. 4. No intermediate payments, such as dividends, will be made to equity holders. At the price of these simplifying assumptions, the model requires only four inputs—the time to maturity of the debt, the market value of the firm’s assets, the present value of the firm’s debts, and the volatility of the firm’s assets. The model can give explicit formulas, in terms of these four inputs, for the probability the firm will default, the loss given default, the required interest rate spread over the risk‐free rate for the firm’s debt, and the market value of the firm’s equity and debt. Using notation close to that in Hull, we’ll denote: V0: The current market value of the firm’s assets D0: The present value of the firm’s debt, which matures at time T, discounted at the risk‐free interest rate σV: The volatility of the firm’s assets PD: The probability of default LD: The loss in the event of default Viewing the equity as a call option on the firm’s value with a strike price of the face amount of the debt, we can write a formula for the current market value of the firm’s equity as: E0 = V0N(d1) – D0N(d2) (13.2) where V0/D0) + σ2VT/2]/σV T d1 = [ln(V d2 = d1 – σV T The current market value of the firm’s debt is just V0 – E0. Following the standard Black‐Scholes analysis, N(d1) is the delta, the partial derivative of E0 with respect to V0, and N(d2) is the probability that the strike price will be exceeded at time T. But this is the probability that the firm will not default so: PD = 1 – N(d2) (13.3) 473 Credit Risk If no default occurs, the debt holders receive the face value of the debt and, if default does occur, the debt holders receive the recovery rate times the face value of the debt, so we can write the market value of the debt as: V0 – E0 = [(1 – PD) + PD(1 – LD)]D0 (13.4) Substituting from Equations 13.2 and 13.3, V0[1 – N(d1)] + D0N(d2) = {(1 – PD) + [1 – N(d2)] (1 – LD)} D0 (13.5) Solving this equation for LD, we get: LD = 1 – (V V0/D0)[1 – N(d1)]/[1 – N(d2)] (13.6) If the debt were truly risk free, its market value would be D0. The credit spread on a zero coupon instrument can be written as s, where the market value (MV V) of the instrument is the face amount (F), discounted by r + s, where r is the risk‐free rate. Thus, MV = Fe−T(r+s) e−Ts = MV V/Fe−Tr s = –ln(MV V/Fe−Tr)/T (13.7) We know the market value of the debt is V0 – E0 and the present value of the debt discounted by the risk‐free rate, Fe−Tr, is D0. Thus, s = –ln [(V V0 – E0)/D0]/T = ln[D0/(V0 – E0)]/T (13.8) Two of the four required inputs, T and D0, are easy to determine, provided all the firm’s debts are reported in some publicly filed statement. To use the model as an approximation when several maturity dates are available for debt and the debt has scheduled coupon payments, T can be calculated as the weighted average duration of the debt. In theory, you could obtain V0 by summing the market prices of all the firm’s equity and debt and estimate σV by looking at the historical volatility of this sum. In practice, most firms have some amount of debt that is not publicly traded and for which a market price would therefore not be available. Inputs that can be obtained easily are the market price of equity, E0, and the volatility of equity price, σE, which can be based on both historical 474 FINANCIAL RISK MANAGEMENT observation and implied volatility from equity options. To obtain V0 and σV from E0 and σE, solve the simultaneous equations: E0 = V0N(d1) – D0N(d2) (13.9) σE E0 = N(d1) σV V0 (13.10) and The latter equation can be derived from Ito’s lemma and the fact that N(d1) is the partial derivative of E0 with respect to V0. The MertonModel spreadsheet takes E0, σE, D0, and T as input and solves for V0, σV PD, LD, MV V, and s. Whenever I have tested this model out on real data, the result has always been the same—reasonable values for PD but unreasonably low values for LD and for s—values produced for LD would be around 10 percent when real experience with loss given default is usually 50 percent or greater, as can be seen in Table 13.5. To explore which of the simplified assumptions of the model considered thus far is leading to this divergence from reality, we could move to a Monte Carlo model that reproduces many possible future paths of the firm’s asset value. The growth rate of the asset value assumed would be the risk‐free rate by the usual risk‐neutral valuation argument. It is easy in the context of a Monte Carlo model to build in payments due to different maturities of debt with coupons, build in rules for when default will occur (such as when the net worth of the firm is below a certain threshold), and build in rules for the distribution of asset value in the event of default to different seniority levels of debt. It is also easy to build in behavioral rules for the firm’s response to different levels of net worth (such as increasing asset volatility as the net worth gets close to the default threshold or issuing new debt as it gets further from the default threshold) and build in rules for dividend policy. By summing over all paths in the Monte Carlo model, it is easy to compute the expected default rates by time period, recovery rates in the event of default by time period and seniority level, and the market value of equity and of each combination of maturity and seniority level of debt. Required spreads over the risk‐free rate for each combination of maturity and seniority level of debt can be computed from the market value. When the assumptions of the simple options model are input to the Monte Carlo model, the same result is obtained as from the simple model. When this model is implemented, we can see what is driving the unrealistic LD and s outputs. If the default threshold is set greater than zero and if asset values are assumed to follow paths without jump processes, then the Credit Risk 475 required spread over the risk‐free rate can be driven as close to zero as desired by increasing the frequency with which observations of the asset value are taken. Increasing the frequency of observation increases the probability of default, but it also causes the loss in the event of default to approach zero by dividing up the assets of the firm among the creditors while they are still sufficient to pay off the creditors in full. This shows that the key issues in determining default loss are behavioral rather than financial; that is, they depend critically on how transparent the operations of the firm are to creditors and how much control the creditors can exercise in forcing bankruptcy in a timely fashion. This may differ significantly by government jurisdiction. The role governments may play in providing help for firms close to default may also differ. There are two ways forward from this impasse. One is to focus on models that do incorporate jump processes. The other is to stick with a simple model but treat it just as a heuristic that can be input to a statistical analysis. We will explore both in turn. 13.2.4.1 Jump Process Models Many such models have been proposed. A good summary with references and discussion for a variety of such models and is Bohn and Stein (2009, Chapter 3). It is important to distinguish between two reasons why a jump process may exist. One is that the asset value of the firm may follow a jump process. The other is that there can be discontinuities in the asset value threshold that will lead to default. CreditGrades (2002) in documenting a model of the second type states, in the introduction to Chapter 2, “In our approach, we model the uncertainty in the default barrier, motivated by the fact that we cannot expect to know the exact leverage of the firm except at the time the firm actually defaults. The uncertainty in the barrier admits the possibility that the firm’s asset value may be closer to the default point than we might otherwise believe.” The advantage of this approach is that it is consistent with the term structure of credit spreads observed in the market; without uncertainty around how close a firm currently is to the default point, one would expect to see much lower short‐term credit spreads than are actually observed. The JumpProcessCredit spreadsheet implements a jump process model closely related to the one documented in CreditGrades (2002) and also documented in Schonbucher (2003, Section 9.5). This model has advantages similar to the Merton model, in requiring very few inputs and being relatively easy to understand. The model assumes the same sort of stochastic evolution of total firm asset value as the Merton model, but assumes that default could occur at any time the asset value falls below a default barrier. Inputs are E0 and σE, as in the Merton model, along with the risk‐free rate 476 FINANCIAL RISK MANAGEMENT and both the mean and standard deviation of the default barrier This input for the default barrier takes the place of the present value of debt that is input to the Merton model. Unlike the Merton model, this model does not attempt to compute a loss given default rate; this is assumed to be estimated by statistical means for each class of debt, as per Section 13.2.2. The model outputs probability of default and credit spread for any desired time period; unlike the Merton model, it is not restricted to a single time period corresponding to the tenor of existing debt. As our quote from CreditGrades (2002) in the paragraph before last indicates, standard deviation of the default barrier is assumed to represent uncertainty about the current level of the default barrier, and hence is independent of time period. Section 2.2 and Figure 2.2 of CreditGrades (2002) show that a standard deviation of 30 percent for the default barrier, derived from historical statistics on actual recovery data, produces a term structure of credit spreads that is consistent with market observations. The CreditGrades model does not have an input for the mean of the default barrier. Instead it is assumed to be the face value of outstanding debt multiplied by the historical average loss given default, averaged over all of the outstanding debt of the firm. While CreditGrades (2002, Section 2.2) presents an argument for the plausibility of this assumption, there is a range of values for the default barrier that would also be plausible given the criteria of the CreditGrades argument. I have chosen in the JumpProcessCredit spreadsheet to leave the mean default barrier as a user input. Users can choose the CreditGrades assumption or experiment with default barrier levels that seem to fit the historical credit spreads of a particular issuer or category of issuers (say a grouping by industry and country). Exercise 13.2 is designed to give you an understanding of the differences between the Merton model and the jump process model in results and in sensitivities to inputs. 13.2.4.2 Statistical Analysis Even simple options models can still play a useful heuristic role in helping to understand the default process. This is the role they play in the models of Moody’s KMV, whose analysis is widely utilized among investors in credit instruments. Crosbie and Bohn (2003) summarize the KMV methodology. De Servigny and Renault (2004, Chapter 3) in the section “KMV Credit Monitor Model and Related Approaches” provide a brief review of the model, along with some reservations. The KMV approach is to utilize a model somewhat like the simple Merton model we first discussed, but the objective is to use it not to try to directly measure default probability, but rather to produce a measure called distance to defaultt, which is then used to project default probabilities based on an empirically fitted statistical model. Technically, the model utilized by Credit Risk 477 KMV treats equity as “a perpetual option with the default point acting as an absorbing barrier for the firm’s asset value” (see Crosbie and Bohn 2003, Section 3). The insight behind this is that, whereas the behavioral nature of default requires the use of statistical observation of past experience, the options model output can be a valuable input to this process when used comparatively to judge which firms are relatively more likely to default than others. In this approach, statistical models, not option‐theoretic ones, are employed in estimating loss in the event of default. KMV presents the following points in favor of this use of the option model: ■ ■ Because the model is based on equity market prices, which are continuously observable, it is more likely to represent the latest available information than the ratings of just a single firm’s credit officers or a rating agency or on statistical models based on accounting information that is only available periodically. It can also be applied to any public company, even one that does not have publicly rated debt, since it is based on equity prices. The model takes into account both the capital structure of a firm and its business and industry risk. Capital structure is represented by the leverage, the ratio of total firm value to equity. Business and industry risk is represented by the volatility of asset values. (For example, you can expect much more volatility from a firm in a high‐tech industry than a utility, or much more volatility from a firm in an emerging market country than one in an established industrial country.) The distance to default is measured by the number of standard deviation movements it would take to put a firm at the point where default is a serious possibility. In terms of the simple model we presented, it would be (V V0 – D0)/(V V0 σV), which is calculated in the MertonModel spreadsheet. The actual model used by KMV to calculate the distance to default is more complex than our simple model in several ways. To highlight a few: ■ Our simple model assumes that default can occur only when firm asset value is insufficient to make a required payment. The KMV model recognizes that firms can be forced to default when their asset values decline sufficiently below the present value of required future payments. Based on empirical studies, KMV has set the default point, which in our model is D0, as the sum of short‐term debt, representing required current payments, and one‐half of long‐term debt, representing payments that will be required in the future. In this way, assets can decline below the required future payments by some amount, but not too far, 478 ■ ■ FINANCIAL RISK MANAGEMENT before default is threatened. De Servigny and Renault (2004, Chapter 3) note that this is a purely empirical rule of thumb that “does not rest on any solid theoretical foundation. Therefore there is no guarantee that the same rule should apply to all countries and jurisdictions and all industries. In addition, little empirical evidence has been shown to provide information about the confidence level associated with this default point.” This critique should be compared to the response to the question “Are default probabilities applicable across countries and industries?” in Crosbie and Bohn (2003, Section 6). The KMV model can handle more liability classes than just straight debt and equity; it can also accommodate hybrid classes—convertible debt and preferred stock. KMV regards Equation 13.10 as too simplistic, since it does not take into account the impact of varying leverage levels through time on the relationship between equity volatility and asset volatility. KMV uses a more complex model to reflect this factor. In particular, the concern is that for a firm whose performance is trending downward, the decline in equity value will result in current leverage being higher than its leverage has been in the past. If asset volatility is estimated from its historical equity volatility and its current leverage, this will tend to understate historical asset volatility, resulting in understating the default probability. The converse of this effect will result in overstating the default probability for a firm whose performance is trending upward. As Crosbie and Bohn (2003, Section 4) state, this “biases the probabilities in precisely the wrong direction.” KMV’s solution is a more granular approach in which a time series of historical daily asset returns is constructed from historical daily equity returns and Equation 13.2, based on an initial guess at σV. These daily asset returns can then be used to compute a new guess at σV, leading to a new series of daily asset returns. The process is repeated until it converges (see Crosbie and Bohn 2003, Section 4). Many aspects of KMV’s methodology are proprietary and undisclosed, but the results they have published have had a major impact on firms that manage credit risk, both as a source of information and as an inspiration for their own research. De Servigny and Renault (2004, Chapter 3) note that “Many banks have developed their own systems to extract early warning information from market variables. Many variants can be found that extract the volatility of the firm from either equity time series, implied volatilities in options markets, or even spreads. . . . Equity‐based models reflect the market’s view about the probability of default of specific issuers and therefore can provide valuable early warning signals. Unfortunately they are no panacea, Credit Risk 479 as they also reflect all the noise and bubbles that affect equity markets. Overall, they can be seen as a useful complement to an analysis of a firm’s fundamentals.” Bohn and Stein (2009, Chapter 3) in the section “Modifying BSM” provide references to empirical research that “cast[s] doubt on the practical viability of structural models” but observe that “numerous financial institutions around the world have successfully implemented and tested credit risk management systems based on the structural framework.” (In this context, “structural” is equivalent to what we have been calling “option‐theoretic”— models that are not just based on statistical linkages but utilize options theory to link default probability to equity prices.) Altman, Fargher, and Kalotay (2010) present results supporting the use of statistical models of default probabilities that combine equity market information with traditional accounting variables (of the type discussed in Section 13.2.1.2). They provide references to other published models utilizing equity market based inputs with a discussion of comparative results. Bohn and Stein (2009, Chapter 3) in the section “Modifying BSM” also observe that a “promising point of departure is that of the hybrid approach, where characteristics of both structural and reduced‐form models . . . or structural and econometric approaches are combined,” and provide many references to published hybrid models. 13.3 PORTFOLIO CREDIT RISK In Section 13.2, we have established the main building blocks that are needed for analyzing portfolio credit risk. The remaining building block is estimation of correlations between defaults, which we will investigate in Section 13.3.1. We will then turn, in Section 13.3.2, to Monte Carlo simulation models that bring all of these building blocks together, and look at computational alternatives to full simulation in Section 13.3.3. Finally, in Section 13.3.4, we will examine how simulation models and the tools of Sections 13.1 and 13.2 can be used in the management and reporting of portfolio credit risk. 13.3.1 Estimating Default Correlations Let’s begin with some points on which almost everyone who has worked on this topic can agree: ■ Strong evidence supports a positive correlation between defaults—that is, that defaults tend to occur in clusters. For example, Table 13.6, from Moody’s (2011a), of default percentages by year for ratings categories 480 TABLE 13.6 FINANCIAL RISK MANAGEMENT Default Percentages by Year Rating Year Baa Ba B 1982 0.32% 2.78% 2.35% 1983 0.00% 0.91% 6.36% 1984 0.36% 0.83% 6.75% 1985 0.00% 1.41% 7.48% 1986 1.01% 2.05% 11.60% 1987 0.00% 2.73% 6.49% 1988 0.00% 1.26% 6.20% 1989 0.60% 3.04% 8.72% 1990 0.00% 3.41% 15.47% 1991 0.28% 4.89% 12.36% 1992 0.00% 0.31% 9.22% 1993 0.00% 0.57% 4.56% 1994 0.00% 0.25% 4.06% 1995 0.00% 0.73% 4.26% 1996 0.00% 0.00% 1.37% 1997 0.00% 0.19% 1.94% 1998 0.12% 1.00% 4.15% 1999 0.11% 1.32% 3.81% 2000 0.39% 0.72% 4.90% 2001 0.20% 1.39% 6.03% 2002 1.10% 1.38% 9.57% 2003 0.00% 1.00% 4.53% 2004 0.00% 0.41% 2.11% 2005 0.18% 0.00% 0.84% 2006 0.00% 0.20% 1.03% 2007 0.00% 0.00% 1.18% 2008 0.47% 1.16% 2.07% 2009 0.86% 2.40% 7.41% 2010 0.00% 0.00% 0.48% Source: Based on Moody’s (2011a, Exhibit 31). Credit Risk ■ ■ 481 Baa, Ba, and B, shows much higher default rates in recession periods, such as 1989 to 1991 and in 2001 to 2003, than in periods of economic growth, such as 1993 to 1997. Estimating this correlation based on the joint default history of firms with the same credit rating is unsatisfactory. Grouping together all firms with the same credit rating ignores factors such as whether firms are in the same industry or whether firms are located in the same geographical region, but these factors are widely believed to influence joint default correlation. See Gupton, Finger, and Bhatia (1997, Section 8.2). The direct estimation of joint default correlation by examining historical defaults categorized by rating, country, and industry is not a feasible approach. Default is a relatively rare event and with this fine a segmentation, there would not be enough observation to allow robust statistical inference. A way around this impasse is to estimate correlation for a variable that can be more frequently observed and can then be utilized to produce default correlations. For KMV, asset returns are a very natural choice for such a variable, since they are directly tied to defaults through the distance‐to‐default measure and its statistical relationship to default probability. KMV utilizes the methodology we discussed in Section 13.2.3 to delever equity returns directly observed in the market and compute asset returns. It is an easy step from creating a time series of asset returns for a large universe of borrowers to computing correlations between asset returns for those borrowers. Monte Carlo simulations of correlated movements in asset returns, which we discuss in the next section, can then be used to calculate the percentage of cases that result in joint default, enabling a default correlation to be computed. The actual methodology employed by KMV does not directly calculate asset return correlations between pairs of borrowers. Instead, a factor analysis is used in which composite asset returns are calculated for sectors—countries and industries as well as groupings of countries and industries. Historical asset return correlations can then be computed between sectors. Asset return correlations between borrowers can then be easily computed based on the statistical relationship between each borrower’s asset return and those of the country and industry sectors. The KMV approach to correlations is described in more detail in the section on “Model of Value Correlation” in Kealhofer and Bohn (2001) and in the section “Calculating Correlations Using Moody’s KMV Portfolio Manager” in Chapter 8 of Saunders and Allen (2010). The CreditMetrics approach to estimating default correlations is very similar to KMV’s, except that correlation between equity returns is used as a proxy for correlation between asset returns. Gupton, Finger, and Bhatia (1997, Section 8.5) provide great detail on this process. 482 FINANCIAL RISK MANAGEMENT The default probability implied by credit spreads is another natural candidate to be used. The drawbacks are that this involves a much smaller universe of borrowers for whom liquid public debt prices are available relative to the number of borrowers for whom liquid equity prices are available and that implied default probabilities from public debt prices significantly overstate actual default probabilities, as we discussed in Section 13.2.1.2. As a result, firms that do decide to use market implied default probabilities as indicators of relative credit quality, but may choose to adjust the overall default probability by a factor that lowers these probabilities to anticipated rates of actual default, following our discussion in Section 13.2.1.2. Whichever variable is used to provide the linkage, the key to transforming shorter‐term correlations into longer‐term default correlations is a simulation of movements through time, which we will come to in the next section. Because of the relative infrequency of default, even high short‐term correlations transform into much smaller default correlations. While much of the early work on building default correlation relationships focused on estimating correlation coefficients, the recent trend has been to also place a lot of attention on the shape of the correlation relationship, the copula. For example, it is frequently the case that large moves in changes in default probability are more closely correlated than smaller moves. The section on “Copulas” in Bohn and Stein (2009, Chapter 8) and Duffie and Singleton (2003, Section 10.4) give an introduction to this topic in the context of estimates from historical data. The assumption that correlation is the same for all sizes of changes in probability is known as the Gaussian copula assumption. Estimates based on market prices of credit correlation products, such as CDOs, will be discussed in Section 13.4.2. One recent trend has been to utilize frailty analysis, a technique borrowed from medical research, to correct for underestimation of correlation due to undetected factors that can have a common impact on many borrowers. A good explanation can be found in Duffie et al. (2009), which provides a frailty model applied to corporate defaults, a statistical analysis of the degree of correlation underestimation that may occur if this correction is not accounted for, and references to related literature. Another recent trend has been to introduce modeling of contagion, the impact that defaults by one or more firms may have in increasing the default probabilities of remaining firms. Rullière, Dorobantu, and Cousin (2010) provide a recent model of this effect with references to related literature. 13.3.2 Monte Carlo Simulation of Portfolio Credit Risk Where we stand, based on the Sections 13.2.1, 13.2.2, 13.2.4, and 13.3.1, is that a variety of methods have been presented for estimating default Credit Risk 483 probability and loss given default for individual borrowers and for estimating the short‐term correlation between borrowers for some variable that can be linked to longer‐term defaults. We now focus on analyzing methods that can provide this linkage and also produce calculations of portfolio risk very similar to the portfolio risk measures that were provided for market risk in Chapter 11. Let’s begin by assuming that all of this analysis will be provided by Monte Carlo simulation. For many of the same reasons stated in our analysis of VaR in Section 11.1, simulation is the most accurate method of generating portfolio risk measures. It has the flexibility to incorporate almost any assumption about statistical distributions we want to make. Later, in Section 13.3.3, we will discuss possible shortcuts to the calculation of portfolio risk under more restrictive distributional assumptions. However, simulation is always the benchmark against which the accuracy of other approximations can be tested. The reason why we only consider Monte Carlo simulation for credit risk, while we considered the alternatives of Monte Carlo simulation and historical simulation for market risk VaR, is that the longer time periods involved in credit risk simulations mean that not enough nonoverlapping historical data points will be available to derive a historical simulation. A Monte Carlo simulation will follow a key variable, whether it is asset value, macroeconomic factors, or default probability, for each borrower to whom credit has been extended. The simulation will be based on assumptions about the volatility of asset returns or transaction matrices for default probabilities and assumptions about short‐term correlations between the borrowers (both correlation coefficients and copulas, as discussed in Section 13.3.1). If asset value is being used as the key variable, it must be converted into default probabilities, using a statistical relationship such as the one developed by KMV between an asset’s distance to default and probability of default. Defaults can then occur at random, based on the probability of default. In the event of default, a random sample is drawn based on the mean and standard deviation of the loss given default for a given seniority class of instrument (if instruments of different seniority are outstanding to the same borrower, a correlation should be enforced between the degree to which the loss given default on each instrument exceeds or is below the average). Correlations between default probability and LGD, as discussed in Section 13.2.2, can be specified as part of the simulation. Detailed descriptions of such simulation models can be found in Duffie and Singleton (2003, Chapter 10), Schonbucher (2003, Chapter 10), and Bohn and Stein (2009, Chapter 8). A Monte Carlo model meeting this description has many possible applications. It can be used with just two borrowers to translate a short‐term correlation of assets values or credit spreads into long‐term default correlations. 484 FINANCIAL RISK MANAGEMENT It can also be used with an entire portfolio of assets to generate statistics on expected credit losses and the full distribution of credit losses, such as losses at the 99th percentile. It can be used for valuing a tranched CDO by tracking losses to each tranche along each of the paths and then calculating the expected losses on each tranche, as we will discuss in Section 13.4.2. Thus far we have been discussing a Monte Carlo simulation that only deals with a single time period in which the relevant outcomes are for each credit to either default or not default. This can be extended in one of two directions. One direction would be the simulation of an end‐of‐period change in credit grade and credit spread in addition to default. This extension requires a tie between the key variable being simulated and credit grade and credit spread. This relationship is straightforward for implied default probabilities and is provided for asset values by KMV’s statistical linkages of the distance to default to credit rating. KMV has also developed a linkage between asset values and credit spreads (see Agrawal, Arora, and Bohn 2004), partially based on the capital asset pricing model and partly on statistical relationships. The other direction would be multiperiod simulation. Multiperiod simulation could be achieved by just computing default loss distributions at different points along the simulation path. However, full accuracy requires some simulation of possible changes in the overall economic climate, factoring in features such as the increased probability of an economic downturn following a period of sustained economic growth (see Wilson 1997). De Servigny and Renault (2004, Chapter 6) give a thorough discussion of four commercially available Monte Carlo simulation models for credit portfolios: the CreditMetrics model documented in Gupton, Finger, and Bhatia (1997); the Moody’s KMV Portfolio Manager model documented in Kealhofer and Bohn (2001); Standard & Poor’s Portfolio Risk Tracker mode documented in de Servigny et al. (2003); and Financial Analytics’s CreditPortfolio View documented in Wilson (1997). De Servigny and Renault’s summary in their Tables 6.1 and 6.3 is particularly useful in showing at a glance the similarities and differences in these four models. De Servigny and Renault’s analysis shows greater similarities than differences. All four models simulate stochastic evolution of default probabilities and use stochastic LGD rates, but only Portfolio Risk Tracker includes correlation between default probabilities and LGD rates. All four models handle correlations between defaults by simulation of common factors that drive default probabilities. They differ on the common factors that drive default probabilities—country and industry factors for CreditMetrics, Portfolio Manager, and Portfolio Risk Tracker, and macroeconomic factors for CreditPortfolio View. Both CreditMetrics and Portfolio Manager derive the relationships that drive these common factors from equity markets (as Credit Risk 485 discussed in more detail in Section 13.3.1), while Portfolio Risk Tracker offers user flexibility to base correlations on equity, credit spread, or empirical data. Their biggest differences involve outputs—Portfolio Manager focuses on the distribution of default losses, CreditMetrics and Portfolio Risk Tracker on the distribution of changes in market value that can be driven by both defaults and ratings changes, while Credit Portfolio View gives a choice between these two distributions. While information on changes in market value can be very useful information for credit portfolio managers, as we will discuss further in Section 13.3.4, I am extremely suspicious of any approach to portfolio credit risk that does not include a focus on distribution of ultimate default losses. Since portfolio credit risk is a long‐term risk not amenable to management with liquid market instruments, as discussed in the introduction to this chapter, the approach of Sections 6.1.2 and 8.4 should govern. Any modeling that relies on future market value is assuming a future liquidity that cannot be relied on. If no statistics on ultimate default losses are available from the model being used, then there needs to be a ready way of translating model output on market value changes into distributions of ultimate default losses. A Monte Carlo simulation of individual loans becomes computationally infeasible for loan portfolios with very large numbers of very small loans. This is certainly true for retail loans such as home mortgages or credit cards. In such cases, the portfolio needs to be analyzed into segments that can be treated as roughly homogeneous. For example, a portfolio of home mortgages could be divided into segments grouped by geography and home value. Each segment now must be treated as a single loan in the Monte Carlo simulation of the entire firm’s loan portfolio. But unlike a true individual loan, which is in either one of two states, default or nondefault, a grouping of small loans must be represented by a percentage of loans that default in a particular time period. An analysis of the history of default patterns can establish statistics to drive the simulation, including correlations with default levels between two segments and between the segment and individual loans being simulated. The best way to derive historical correlations may be through a mutual dependence on macroeconomic factors, such as growth rates in the economy; see Wilson (1997) and Bucay and Rosen (2001). For a general overview of a simulation of defaults on retail credits, see Risk Management Association (2000). One way in which the use of simulation for credit risk differs from its use for market risk VaR is that the expected value of the distribution plays a significant role. Since market risk VaR is computed over very short time periods, expected value can either be ignored or else has only a minor impact. The far longer time horizon of credit risk simulation requires that expected credit loss be accounted for. Expected credit loss should be taken 486 FINANCIAL RISK MANAGEMENT as a charge against earnings, either in the form of a reduction in valuation for a portfolio that is marked to market or in the form of a loan loss reserve for a portfolio that uses accrual accounting. Risk should be measured and capital should be allocated based on the unexpected losses—the distribution of returns around the expected losses. Parallel to the discussion in Section 7.2 of stress testing as a complement to VaR for market risk, it is often desirable to complement the statistical analysis of credit risk for a portfolio with a stress test based on economic insight, for example looking at the impact of an unusually prolonged global recession. This is especially true in evaluating the risk of credit concentration to firms doing business within a particular country. Credit concentration within a country leads to the risk of correlated outcomes since all firms may be impacted by how well the country’s economy performs. This type of correlation risk is very much the same type of risk as the risk of credit concentration within an industry or within a geographical region of a country. All of these correlation risks can be reasonably measured by statistical means. But country risk has an additional dimension. The possibility exists that all firms, individuals, and government bodies within a given country will be prohibited from meeting their contractual obligations. This can arise from the imposition of exchange controls by the government as a defensive measure against adverse currency flows, or from government renunciation of foreign debts, or from disruption of normal contractual relationships due to war or revolution. This form of risk represents a major political discontinuity that statistical analysis of historical economic data will shed little light on. It can best be quantified by looking at the extent of damage in past incidents of political disruption in other countries combined with subjective assessment of the likelihood of occurrence based on economic and political insights into the current conditions within a particular country. See Bouchet, Clark, and Groslambert (2003). Using Monte Carlo methods to design credit portfolio stress tests, parallel to the discussion in Section 7.2.3, is specifically addressed in Breuer and Csiszar (2010, Sections 3.3 and 3.4). 13.3.3 Computational Alternatives to Full Simulation Portfolio credit risk analysis does not have the same need for rapid turnaround that models used for trading liquid instruments do. Changes in the portfolio occur more slowly, you don’t need to respond to the needs of a trading desk requiring an up‐to‐date picture at the start of each trading day, or to contribute to the daily VaR market risk calculations. So there is much greater tolerance for full simulation runs that may require many hours or even a few days to produce statistics. Even so, there will be a desire to see Credit Risk 487 the impact of alternative strategies in building the credit portfolio or in buying protection that may need a speedier approximation to accommodate multiple runs. And definitely the need to produce marginal risk analysis for the impact of a proposed new loan, discussed in Section 13.2.4, will require a fast approximation technique. Fortunately, approximations can easily be tested for accuracy against the full simulation, following the prescriptions of Section 8.2.5. And fortunately a great deal of clever mathematics has been developed to produce good approximations in reasonable time. Much of this work has been in support of credit derivatives, such as CDOs, which are traded in a market environment and have an even greater demand for quick estimation, as we will see in Section 13.4.2. But, since full simulation models for portfolio credit risk and for CDOs are virtually identical, portfolio credit risk can benefit from these quantitative advances. The two most important ideas that have been introduced for approximating full simulations are the large homogeneous portfolio (LHP) approximation and the Vasicek model that utilizes only a single factor to drive correlation relationships. The LHP approximation looks very similar to the approach to simulation modeling of large numbers of very small loans discussed in Section 13.3.2. Loans are grouped by common characteristics (this could include industry, geography, credit rating, estimated probability of default) and each group is simulated as if it were a single loan, but instead of being represented by just two states (default or nondefault), the representing state is the percentage of loans within the group that have defaulted as of a given time step. Within each group, the loans are treated as homogeneous (i.e., all having the same default probabilities, LGD probabilities, and default correlations with loans in other classes). The number of loans within each class is treated as large enough that the class can just be represented by an overall percentage of default without worrying about the actual sizes of individual loans within a category. The less equal the loan sizes are, the less accurate this assumption will be; for example, if there were one very large loan, its default would cause a jump in the percentage of defaults for the category. The fewer categories you use, the faster the simulation will run, but the less accurate the approximation to the full simulation model will be. The Vasicek model utilizes only a single factor, roughly corresponding to the state of the world economy, to drive the simulation. This is obviously a major approximation, since much of the detail of correlation relationships based on industry‐specific and geography‐specific factors will now be lost. But the resulting simplification allows calculations to be performed by much quicker numerical integration methods, as opposed to simulation; see Schonbucher (2003, Section 10.4.3) for details. Even in cases where the 488 FINANCIAL RISK MANAGEMENT decision is to rely on a fuller simulation for risk reporting, this much faster calculation allows quick estimation of sensitivities to input variables that can be valuable for building intuition for portfolio managers. The Vasicek model is a particularly useful approximation for building intuition because of its strong emphasis on the separation of systematic risk and idiosyncratic risk, as we will see in the following pages. Even quicker numerical integration can be achieved by combining the Vasicek model with the LHP approximation. The most frequently encountered version of this combination also assumes a Gaussian copula (see Section 13.3.1) for the correlations. This version, the now infamous Li model (see Section 5.2.5.3), has been well documented, for example Hull (2012, Section 23.9) or Schonbucher (2003, Section 10.4.4), as well as the original Li (2000). The Vasicek model operates by keeping track of all default correlations using a single common factor and calculating losses corresponding to each possible value of this common factor. Tying computations to this common factor, which could be thought of as the state of the economy or, for mortgage portfolios, the level of housing prices, is what makes the Vasicek model so appealing for gaining an intuitive grasp of the impact of systematic risk (we will discuss this further in Section 13.4.4). What makes computation so easy is that all idiosyncratic risk is incorporated through a simple formula applied to each level of the common factor. We will quickly look at how this is done. Since the portfolio is assumed homogeneous, there are only three input variables to describe the underlying credit portfolio of the CDO: the expected default percentage of this portfolio, D; the recovery rate in the event of default, R; and the correlation between changes in asset values, ρ. All the other inputs reflect the structure of the CDO—the attachment and detachment point of each tranche. The model assumes that each individual credit has associated with it a standard normally distributed variable xi that reflects the distance to default of the credit. There is a threshold value such that, if xi goes below the threshold, default will take place. Since we know that the probability of default of each credit is D, this threshold must be N−1(D), where N is the cumulative standard normal distribution and N−1 is its inverse. The variable xi may be written as: xi ρM + 1 − ρZi where M is a common factor affecting all defaults and Zi is a factor affecting only credit i. The variable M and all the Zi variables are assumed to have independent standard normal distributions, so that the relationship assures 489 Credit Risk that all pairs of credits have a correlation of ρ and that all xi variables have a standard normal distribution. The probability of default of any individual credit is xi < N−1(D), which becomes ρM or (N Zi 1 ρZi < N −1 ( D) ( D) − ) M / 1− ρ so that the probability of default is: ( N ⎡⎣ N − ( D) − ) M / 1 − ρ ⎤⎦ The next step is to numerically integrate over many different possible values of M. For each one we can calculate the percentage of total defaults, which multiplied by (1 – R) gives the percentage of total losses. We can easily calculate the losses due to each tranche, utilizing the attachment and detachment points, corresponding to each value of M. We make use of the LHP assumption to treat these loss estimates as exact (given the value of M), rather than just the central point for a probability distribution. We then use the probability distribution of the values of M to infer probability distribution of the tranche losses. The assumption of a Gaussian copula is not at all necessary for the quick numerical integration technique to work; see O’Kane (2008, Sections 21.5 and 21.6) and Schonbucher 2003, Section 10.8.2) for examples. Similar, but more computationally intense, integrations can be used for multifactor approximations (see Schonbucher (2003, Sections 10.4.5 and 10.4.6). The CDO spreadsheet on the website for this book will allow you to experiment with a Vasicek model that uses the LHP approximation but with several choices for the copula. O’Kane (2008, Sections 16.4 and 18.6) analyzes the accuracy of these approximations. The LH+ model is an interesting compromise (see O’Kane 2008, Section 17.3). It’s a Vasicek model that uses the LHP approximation for the entire portfolio except for a single loan that is individually modeled. It can still make its computations using a fast numerical integration, but as O’Kane says, it “allow[s] us to understand the interplay between the characteristics of the single credit and those of the overall portfolio.” This can obviously be very valuable when making decisions about how to place an internal price on a new loan or whether to buy credit protection against an existing one. 490 FINANCIAL RISK MANAGEMENT Alternatives to the LHP assumption can achieve speeded approximation while retaining more detail about the structure of individual loans within the portfolio. O’Kane (2008, Chapter 18) provides a good overview of these approximation techniques. A good starting point for learning about these models would be Hull and White (2004), which is particularly clear in its exposition, presents two models that are relatively easy to implement, and provides references and comparisons to other similar approaches in the literature. Both models work with individual loan data and utilize recurrence relationships in place of Monte Carlo simulation for calculation. The first calculates probabilities of exact loss percentages from the recurrence relationships using Gaussian quadrature. The second approximates the chances of losses falling into user‐specified probability buckets. Two other well‐known models along comparable lines are Andersen, Sidenius, and Basu (2003), which also provides many references to similar approaches, and Laurent and Gregory (2003), which utilizes a fast Fourier transform. Bluhm, Overbeck, and Wagner (2002, Chapter 4) is a good exposition of the commercially available CreditRisk+ model that utilizes recurrence relationships and probability‐generating functions in place of Monte Carlo simulation. Since credit risk calculations are focused on occurrence of default, which is a low‐probability event, improvements in accuracy relative to computation time can be gained utilizing importance sampling, a technique that focuses more of the simulation paths on those probability regions where default is more likely to occur. Glasserman and Li (2005) is a key paper in this area. Glasserman (2004, Section 9.4.3) covers similar material. Giesecke and Shkolnik (2011) is a recent contribution and provides many references to similar approaches. 13.3.4 Risk Management and Reporting for Portfolio Credit Exposures A traditional bank managing a large portfolio of credit risk will need to find the proper balance between the illiquidity of much of its portfolio and the liquid instruments that can allow it to manage some of its risk. On the one hand, the illiquidity of some borrowers and the size of exposures to even liquid borrowers will preclude any chance of using liquid instruments to eliminate all of the exposure. On the other hand, a combination of loan sales, purchases of credit insurance through CDSs, and packaging of some credit in CDOs does permit some choices on the composition of the portfolio. Choices about sales of existing loans or purchases of credit insurance against current positions are not the only tools available to a credit portfolio manager. A key tool is the internal pricing for taking on new credit exposure. When credit managers have a more favorable view of the credit prospects of a particular borrower than the market does, they will convey Credit Risk 491 this to the firm’s relationship managers by quoting internal prices that reflect narrower credit spreads than those quoted in the CDS market. (The same will apply for particular classes of loans where the credit managers’ view of the combination of default probability and LGD is more favorable than is reflected in the market.) As we noted in Section 13.2.1, it is important that the views of the firm’s credit managers be challenged when they are more favorable than ratings agencies or market prices imply, but when the credit managers’ judgment is sustained after such a challenge, it is an appropriate strategy for the firm to encourage such lending—it may very well be that long experience with particular borrowers or industries or particular expertise gives the firm an edge that it should be taking advantage of. Conversely, when the firm’s credit managers have a less favorable view of the credit prospects of a particular borrower than the market does, they will want to discourage relationship managers from extending new credit. But this does not necessarily involve quoting internal prices that reflect wider credit spreads than those quoted in the CDS market. To the extent that the CDS market for the borrower is liquid, the internal quote may just reflect the CDS spread, with the credit managers intending to purchase CDS protection against any new credit extensions to the borrower. But this strategy must be accompanied by some type of limit on the amount of lending that can be offered at this market price, gauged to the liquidity of the market. Internal price quotes more favorable than market quotes are not confined to the situation in which credit managers have a favorable view on a borrower. It may also reflect the composition of the firm’s credit portfolio. If lending to a particular borrower offers diversification of the portfolio, perhaps in terms of industry or geography, this may get reflected in a lower internal price for credit than the market. And conversely, if lending to a particular borrower will add to existing portfolio concentrations, the credit managers may quote an internal price equal to the market price and plan to offset the loan with CDS protection, even when they have a more favorable view of the borrower than the market does. This approach clearly requires reporting from the credit portfolio model that can assess the marginal impact new lending to a particular borrower will have on the firm’s overall credit risk. We will address this information need shortly, when we discuss credit portfolio reporting requirements. Given our discussion of the large differences between actual and market‐implied default probabilities toward the end of Section 13.2.1.2, one might suspect that this would motivate relationship managers to prefer internal loan pricing to market loan pricing. In my experience, this does not turn out to be much of a factor. The reason is that the internal pricing’s charge for capital usage is roughly equal to the difference between default experience and market‐implied probabilities. Looking at Table 13.5, you 492 FINANCIAL RISK MANAGEMENT can see that the difference between average and worst‐case default experience is greater than the average default experience for every credit rating above B. Since allocated capital is roughly comparable to this difference, and since charges against earnings for capital allocations are on the order of 15 percent per year, you can see that internal pricing is unlikely to look more favorable than market pricing just based on this factor. Rather, it is differences between market and internal assessments of credit quality for a given name and portfolio composition considerations just discussed that usually drive preferences between internal and market pricing. Risk reporting for portfolio credit risk is similar to risk reporting for market risk VaR, and many of the recommendations of Sections 7.1.2 and 7.3 can be followed with little modification. The reporting of VaR at different percentiles, the use of shortfall VaR as an alternative to VaR to better capture tail risk and to avoid issues of instability and negative diversification effects, the reporting of exposures by product and business line, and the use of both marginal and stand‐alone measures of risk all carry over quite well to portfolio credit risk. Reporting of exposures by product and business line will help to identify lines of business that should be expanded or whose growth may need to be slowed and to identify priorities for parts of the portfolio that require hedging through loan sales, CDSs, and CDOs. More thoughts along these lines will be found in the section on “Improving Portfolio Performance” in Bohn and Stein (2009, Chapter 8). In making decisions among competing priorities for portfolio hedging, credit managers will need to consider the risks of delay not just in terms of possible defaults, but also in terms of ratings downgrades and widening of market credit spreads. So, while I would still insist on the illiquidity of credit portfolios requiring output based on eventual defaults (see Section 13.3.2), output based on the impact of changes in market value is also needed. This output should also include reports on market value sensitivity to shifts in the economic environment, along the lines of the sensitivity measures for liquid credit discussed in Section 13.1.3. O’Kane (2008, Section 17.2) has an extensive analysis of sensitivity measures for credit portfolio risk in the context of CDO tranches, which will be discussed in Section 13.4.3. The caveats addressed there about the limitations on the usefulness of these sensitivity measures due to illiquidity of credit markets also apply here. One major difference between market risk VaR and credit portfolio risk models is the need to measure the marginal risk contribution of individual loans, in line with our previous discussion. No marginal measure that is this fine‐grained is required for market risk. Since it would be computationally infeasible to generate this information by running the full simulation for each loan, finding computational shortcuts, addressed in Section 13.3.3, is critical for credit portfolio modeling. Credit Risk 13.4 493 RISK MANAGEMENT OF MULTINAME CREDIT DERIVATIVES 13.4.1 Multiname Credit Derivatives Multiname credit derivatives are baskets that bundle together credit exposure to several debt issuers into a single instrument. From the side of credit protection sellers, these instruments offer an opportunity to obtain exposure to a diversified basket of corporate debt—it can be quite difficult for an investor to put together such a basket on his own, owing to the relative illiquidity of corporate bond markets. So a market maker who has the ability to source such a basket of debt can get paid a good spread for selling it in a convenient form. From the side of credit protection buyers, these instruments offer a chance to offset portfolio credit risk and can play a significant role in the management of portfolio credit exposure discussed in Section 13.3.4. There are several forms such credit baskets can take. One that has become particularly popular is to create a derivative tied to a credit market index such as the CDX and iTraxx indexes (see the discussion in Hull (2012, Section 24.3). Since these indexes are calculated and disseminated in a very public manner, they provide good transparency. A spread is set that the credit protection buyer will pay to the credit protection seller. Every time one of the components of the basket defaults, settlement is made on that portion of the basket, following the same rules as settlement of individual CDSs (with a strong bias toward cash settlement). The choice of index components has been designed to balance liquidity of individual names and diversity of credit exposure. Many different strategies for expressing views on relative credit spreads and achieving protection against existing credit risks can be obtained through combinations of long and short exposures to different credit indexes and individual CDSs. Market makers will also construct more tailored indexes for clients with particular needs. As long as all credit protection sellers in a credit basket share equally in all cash flows (both receipt of credit spread and payment of default losses), the credit basket is just a simple summation of the pricing of individual CDSs and so should be risk managed following the principles of Section 12.4.1.1; there are some technical basis risk issues that create differences between the index basket and a portfolio of the individual CDSs comprising the basket— these are well covered in O’Kane (2008, Chapter 10). But a frequent variant is to structure a credit basket into tranches that receive different portions of the cash flows. The motivation for this structuring is that different classes of investors have different tolerance for credit risk and different institutional restrictions on the type of credit risk they can invest in, and the object is to design a structure that will better fit demand. 494 FINANCIAL RISK MANAGEMENT Some CDOs are structured this way, but others are just single tranches (called synthetic tranches) that have been agreed to between a credit protection buyer and a credit protection seller with an agreed reference portfolio. Credit protection buyers may enter into these single tranches either as a way of buying protection against their credit exposure in a piece‐by‐piece fashion or in order to express a particular view on credit risk. Credit protection sellers may want to express a particular view on credit risk or offset previously purchased credit protection they no longer need. The modeling and risk management of tranches is very similar, whether they arose as part of a credit basket or as a synthetic tranche, though there are some differences in payment details that will impact modeling; see O’Kane (2008, Section 12.5). Tranches of CDOs are structured by designating an attachment point and a detachment point. For example, a tranche with an attachment point of 7 percent and a detachment point of 10 percent would not have to make any default payments until default losses in the basket exceed 7 percent of notional principal and then would pay all default losses until default losses in the basket reach 10 percent of notional principal, after which time the tranche no longer exists (it is no longer receiving any credit spread payments and does not owe any obligations on any possible future defaults). In between 7 percent and 10 percent, every time a default takes place, settlement is made on that portion of the tranche, with credit spread payments on that portion ceasing. By market convention, credit spreads paid to the protection seller of a tranche are quoted as percentages of the portion of notional principal that the seller could potentially lose. For example, if an investor sells 7 percent to 10 percent credit protection on a $100 million basket, his largest potential loss would be $100 million * (10% – 7%) = $3 million. If his credit spread is quoted at 3.47 percent, he will receive 3.47% * $3 million = $104,100 per year until such time as default losses exceed 7 percent. Standardized tranches have been created for both the CDX and iTraxx indexes (see Hull, Table 24.6). The tranche that will receive the first losses, the tranche with 0 percent attachment point, is called the equity tranche (since its absorption of losses prior to any losses impacting other tranches is similar to the relation between the equity investors in a corporation relative to the debt holders). The tranche with the highest attachment point is called the super‐senior tranche (since its expected losses are usually even smaller than the highest‐rated AAA corporate debt). Intermediate tranches are called mezzanine tranches for those with lower attachment points and senior tranches for those with higher attachment points. Tranching cash flows from a credit basket introduces a new type of risk that did not previously exist in the basket—exposure to default correlation. This can be illustrated by a simple example. Suppose you have a credit basket on which your expected default losses, net of recovery, over its five‐year Credit Risk 495 life are 3 percent of principal. If you assume a very low level of correlation between defaults, then almost all scenarios will involve some group of companies defaulting and very few will involve a large number defaulting. So a 0 percent to 3 percent equity tranche will almost always lose close to its maximum and a 15 percent to 30 percent super‐senior tranche will experience zero losses. By contrast, at a very high level of default correlation, some scenarios will involve almost no losses while some will incur very heavy losses. So a 0 percent to 3 percent equity tranche will sometimes lose less than the maximum and so have lower losses than under the low correlation assumption while the 15 percent to 30 percent super‐senior tranche will sometimes experience losses and so have higher losses than under the high correlation assumption. This pattern always holds—higher correlation means lower losses for any tranche with a 0 percent attachment point and higher losses for any tranche with a very high attachment point, but you can’t tell in advance of detailed calculations how an intermediate tranche will be impacted. One variant of tranching CDOs is to allocate losses based on number of defaults rather than on the percentage of losses in the portfolio. It is called a default basket. For example, a first‐to‐default tranche absorbs all the losses of the first credit in a basket to default (if any) but loses nothing on any subsequent defaults. Default baskets make sense only when based on a relatively small number of individual credits, anywhere form 2 to 10. O’Kane (2008, Section 12.2) explains the mechanics and basic economics of this product. While its modeling and risk management are closely related to those of standard CDOs, LHP approximations do not make sense, given the small number of credits in the underlying portfolio and the digital nature of loss allocation. Approximation methodology such as the first of the two models in Hull and White (2004), referenced in Section 13.3.3, are more appropriate for default baskets. More complex variants of CDOs, such as CDO‐squareds, constant proportional debt obligations, and options on tranches, will be covered only briefly as part of the next section on modeling of multiname credit derivatives. 13.4.2 Modeling of Multiname Credit Derivatives The modeling of multiname credit derivatives is extremely similar to the modeling of portfolio credit risk covered in Section 13.3. Indeed, many of the techniques discussed there were originally developed in support of analysis of CDOs and CDO tranches. This is not surprising, since multiname credit derivatives represent an attempt to provide a market for the transfer of portfolio credit risk. 496 FINANCIAL RISK MANAGEMENT A first‐cut sketch of the model for a multiname credit derivative instrument would therefore be to start with the portfolio of credits that comprise the underlying basket referenced by the instrument, model the losses in this portfolio using the tools of Section 13.3, and in this modeling keep track of which losses accrue to which tranches. You can see a simple illustration for a single‐factor Vasicek model in the CDO spreadsheet on the website for this book. At each level of the single factor that drives the correlation between defaults of underlying credits, the spreadsheet keeps track of how much loss accrues to each tranche of the CDO. It is then easy to compute a full probability distribution of the losses for each tranche. An exposition of this simple model can be found in O’Kane (2008, Section 16.2) along with an analysis of the sensitivity of model results to input parameters in Section 16.3. In practice, the allocation of losses to tranches may follow complex rules. This is particularly true for tranches of CDOs based on mortgage securities. So modeling of tranche losses requires more sophistication in the simulation of each individual path. Smithson and Pearson (2008) addresses this issue. More complex multiname credit derivatives may require more detailed modeling of the evolution of losses over time. O’Kane (2008) gives a brief introduction to these products (constant proportional debt obligations in Section 22.3, forward‐starting tranches in 22.6, and options on tranches in 22.7) along with an introduction to the more detailed models of evolution of losses in Chapters 23 and 24. A product that became popular in the explosion of subprime mortgage– based CDOs was the multilevel CDO in which tranches of different CDOs are bundled together to form a portfolio that can itself be tranched, called a CDO‐squared (and this process can be repeated to form a CDO‐cubed, and so on). The same fundamental modeling approach can be utilized as for single‐ level CDOs, modeling losses and keeping track of which losses accrue to which tranches and tranches of tranches. However, the computational intensity of keeping proper track of this waterfall may require new techniques for efficient approximation. O’Kane (2008, Section 22.4) is a good introduction to these products and their modeling. As O’Kane illustrates in Figure 22.4, CDO‐squareds have tremendous sensitivity—very large changes in the losses to tranches due to very small variations in losses to the underlying credits. In the wake of the 2007–2008 crisis, when a great many AAA‐rated tranches of CDO‐squareds of subprime mortgages suffered close to 100 percent losses, these products came in for harsh criticism as vehicles for inappropriate levels of leverage. It is doubtful that we will ever see a revival of interest in them (a parallel story to the power options whose unsuitable use in the early 1990s led to the virtual death of the product ever since—see the discussion toward the end of the introductory section in Section 12.1). Credit Risk 497 Credit portfolio managers are typically dealing with just a single firmwide portfolio. By contrast, traders in CDOs and other multiname credit derivatives are typically dealing with a large number of different reference portfolios. This makes computational alternatives to full simulation, which we covered in Section 13.3.3, even more critical to CDO traders than they are to credit portfolio managers. As we already noted in Section 13.3.3, it was in the context of CDO modeling that many of these computational alternatives were first developed. It also means that CDO traders will typically have a far less intimate knowledge of the characteristics of any particular reference portfolio they are dealing with than a credit portfolio manager will have of her portfolio. This should raise a note of caution concerning the accuracy of CDO modeling, to which we will return in the next section on CDO risk management. A critical difference between credit portfolio modeling and CDO modeling is that CDO modelers will frequently be trying to fit to market data on where different CDO instruments are trading. The only market data that was considered in our discussion of credit portfolio modeling in Section 13.3 was market data on individual credits; all input on relationship between defaults of different borrowers came from statistics and subjective probabilities. While CDO traders and risk managers must also be aware of the implications of statistical and subjective estimates, input from markets is vital. Let’s consider a typical situation. A trading desk is asked to price a nonstandard tranche on a particular portfolio (by nonstandard, we mean having different attachment and detachment points from more commonly traded tranches). The desk can obtain market prices for standard tranches on the same portfolio. To price the nonstandard tranche, the traders would like to fit the parameters of a pricing model to correctly price all of the standard tranches and then apply the model to a nonstandard tranche. This very closely fits the interpolation model approach of Sections 8.2.6.1 and 8.3. It would be convenient if a simple model, such as the Li model, could fit the standard tranche prices, but this is virtually never possible. The reasons why sound very much like the reasons we discussed in Section 11.6.2 for why a single implied volatility is unlikely to fit market options prices for several different strikes. For vanilla options, as we saw in Section 11.6.2, it is partly because of different market supply and demand pressures for different strikes, and it is partly because some of the assumptions of the Black‐Scholes model are incorrect. The story is similar for CDOs. Tranches with very low attachment points (equity tranches) and tranches with very high attachment points (super‐senior tranches) are far less popular with buyers of credit risk than are tranches with intermediate attachment points (mezzanine tranches). We discussed some of the reasons for this in Sections 5.2.2 and 5.2.5, in the context of CDOs of mortgages; similar reasons apply to CDOs of corporate 498 FINANCIAL RISK MANAGEMENT loans. As for model assumptions, the assumption of a Gaussian copula is often contradicted by historical data (see Section 13.3.1). There are two basic approaches to fitting market prices for tranches. One focuses on finding a model that more accurately reflects statistical relationships between defaults; the second focuses on pragmatically changing model parameters to achieve a fit. The first has the advantage of trying to build in more economic reality and so is likely to be a more robust model than one that just fits to prices. But the first has the disadvantage that even the most realistic model may not be able to account for supply and demand forces in the market. The more pragmatic approach of just fitting market prices is a very close analogue to utilizing volatility smiles and skews in fitting option prices; whatever supply and demand dictates determines the implied volatility input for each strike and tenor, and options that can’t be directly priced in the market have their volatilities interpolated from those that are directly priced, as in Sections 11.6.1 and 11.6.2. The analogous method for CDOs is called implied correlation skew to parallel the implied volatility skew. O’Kane (2008, Chapter 19) explains this approach in detail. To make the fitting more manageable and arbitrage‐free, it is extremely helpful to do all fitting to base tranches, tranches with an attachment point of 0 (i.e., tranches that absorb all losses up to the detachment point). This approach, known as base correlation, is explained in detail in O’Kane (2008, Chapter 20). Standard base tranche prices can always be constructed directly from standard tranche prices (e.g., a 0 percent to 10 percent base tranche is just the direct summation of a 0 percent to 3 percent tranche, a 3 percent to 7 percent tranche, and a 7 percent to 10 percent tranche). This approach closely parallels one that has been in use for years in the vanilla options market, where all fitting of volatilities by time period is done for time periods starting at the current date; this avoids interpolations that produce negative implied volatilities and makes for smoother fits. The more fundamental approach of finding a model that more accurately reflects statistical relationships has a vast multitude of candidate models—at least as many as the different ideas on alternative copulas discussed in Section 13.3.1. O’Kane (2008, Chapter 21) addresses some of the more popular choices for copulas, and discusses issues of calibration and comparison between models. 13.4.3 Risk Management and Reporting for Multiname Credit Derivatives We will begin with two polar views of risk management and related reporting requirements for multiname credit derivatives and then see how the two Credit Risk 499 views can be blended. At one extreme, we will focus on the fact that holding a CDO position is very similar to holding a credit portfolio position, so the risk management should look very similar to the approach to risk management of portfolio credit given in Section 13.3.4. At the other extreme, we will focus on the greater liquidity of credit derivatives and look to a risk management approach closer to that for other liquid derivatives in Sections 11.4 and 13.1.3. Which approach will have the greater weight in a blended view will depend a lot on just how liquid the market is for multiname credit derivatives. Even if we believed that multiname derivatives were completely illiquid, we would still need to modify the portfolio credit risk approach of Section 13.3.4 to account for the fact that in a derivatives book your positions encompass sales of credit portfolios as well as purchases. But the basic principle would remain the same: simulate default losses all the way to the maturity of positions and look at the full distribution, both expected losses and tail losses. But there are a few additional points to be considered: ■ ■ ■ Some credits may be referenced in multiple tranches. And for multilevel CDO products, such as CDO‐squareds, the same tranche may be referenced in multiple higher‐level tranches. The simulation engine must have the capability of identifying these multiple references and treating them properly; they must show 100 percent correlation on defaults. Traders in CDO tranches will often lack the detailed knowledge about individual underlying credits that a credit portfolio manager would have for credits originated in the firm. The credit portfolio manager might still use a faster simulation method such as an LHP approximation, as addressed in Section 13.3.3, but has the capability of checking the accuracy of the approximation by occasional comparison of results to a full simulation, and then adjusting risk reports to reflect the accuracy of the approximation. A CDO trading desk manager using the same LHP methodology may lack the detailed data on individual underlying credits that would allow the accuracy of the approximation to be checked. Some alternative means of adjusting risk reports for the inaccuracy of approximation must be designed, such as simulating several different possible specifications for the data on underlying credits, calculating the approximation error that each would lead to, and basing a conservative estimate of approximation error on these results. Some adjustment in probability distributions would also be appropriate to reflect the lower certainty regarding estimates of default probabilities and loss given default that is associated with less detailed knowledge of individual credits; compare with Rajan (2010, 128–129). 500 FINANCIAL RISK MANAGEMENT A trader dealing in liquid CDO tranches would want to start with a set of risk measures and limits that looked a lot closer to those of Section 13.1.3, with a focus on exposure to changes in market credit spreads, but supplemented by measures of convexity exposure to large jumps in credit spread and default. But this would need to be modified to take exposure to correlation into account. If the tranches are truly liquid, then it should be possible to manage correlation risk in a manner very close to the management of option risk in Section 11.4, with measures of exposure to changes in correlation levels as well as changes in the shape of the correlation surface (by time bucket and attachment point) and to joint changes in credit spread and correlations. O’Kane (2008, Chapter 17) has a detailed discussion of risk reporting following this approach. The reporting and risk management of a multiname credit derivative portfolio needs a blend of these two approaches, based on actual degree of liquidity. But no matter how liquid the derivatives in the portfolio, some weight should always be given to the approach of Section 13.3.4, since this is the approach best designed to deal with the impact of defaults. This is a parallel point to the one made in Section 13.1.3 on risk management and reporting for single‐name credit instruments; the extreme difference between exposure to credit spread movements and exposure to defaults, illustrated in Section 13.1.2.2, necessitates two different reporting frameworks. Whatever approach is being taken to risk management of CDOs, there needs to be a strong awareness by both traders and risk managers of the extreme sensitivity of some CDO tranches to systematic risk and to changes in assumptions. This will be highlighted in the next section. 13.4.4 CDO Tranches and Systematic Risk Among the general principles for risk management in Section 6.1.1 was the need for risk managers to carefully distinguish between systematic (undiversifiable) and idiosyncratic (diversifiable) risks. Earlier in this chapter, we noted the strong impact of systematic risk on the pricing of credit exposure (Section 13.2.1.2). This becomes a particularly important issue for the most senior tranches of CDOs, because tranching has the effect of concentrating the idiosyncratic risk of the reference portfolio in the more junior tranches and concentrating its systematic risk in the more senior tranches. This effect becomes even more pronounced for the senior tranches of CDO‐ squared products. These issues are very cogently analyzed in Coval, Jurek, and Stafford (2008). One way to understand why this happens is to see that there are likely to be some defaults in the reference portfolio regardless of the state of the economy. So the amount of loss in the tranches that absorb the first losses 501 Credit Risk is likely to be as dependent on the idiosyncratic risk arising from exactly which credits are in the portfolio as it is on the state of the economy. But losses will reach the very senior tranches only in situations where the common economic factor suffers a major negative event. A useful analogy would be a put option purchased as protection against a large decline in a stock index; it will pay off only if there is a severe shock to the economy. But just as we saw in Section 11.6.2 that protection buyers tend to strongly bid up the implied volatility on such put options, we can anticipate that senior tranches of CDOs should be priced at steep premiums to expected losses. Closely related points, highlighted by Coval, Jurek, and Stafford (2008), are that senior tranches have very high volatility of returns and very strong sensitivity to model assumptions. And all of these points apply with even more force to senior tranches of CDO‐squareds. As we noted in Section 13.3.3, a major advantage of the Vasicek model is its ability to build intuition concerning the allocation of systematic risk to tranches. Exercise 13.3, using the CDO spreadsheet on the website for this book, allows you to use the Vasicek model to generate measures of systematic risk and volatility for tranches of a CDO. One point that will be made in this exercise is that the reasonableness of correlation inputs to the Vasicek model can be judged by comparing model results to historical default experience, utilizing data such as that presented in Table 13.5. EXERCISES 13.1 Calculating default rates from bond rates Using the CreditPricer spreadsheet, begin with the following input: 1 2 3 4 5 Risk‐Free Zero‐Coupon Rate Risky Par Rate 7.00% 7.50% 7.75% 8.00% 8.15% 8.00% 8.60% 8.90% 9.20% 9.40% 1. Solve for the default rates and spreads to the risk‐free par curve that corresponds to this case. 502 FINANCIAL RISK MANAGEMENT 2. Change the loss given default to 30 percent and double the default rates. Solve for the risky par bond rates. How does the spread to the risk‐free par curve differ from that in the previous step? This shows that it is not just the product of default rate and loss given default that impacts the valuation of risky cash flows. 3. Assume that the company whose risky par rate curve was shown previously also has a five‐year bond with a 9 percent coupon that is priced in the market at 98.56. Assuming a constant loss given default irrespective of the time at which default occurs, determine a unique loss given default and a set of default rates from this information. What if the 9 percent coupon five‐year bond is selling at 98.46? 13.2 Comparing the jump process credit model to the Merton model 1. Run the MertonModel with a stock price of 40, debt per share of 60, equity volatility of 60 percent, and time to maturity of five years. What are the resulting probability of default and loss given default? 2. Run the JumpProcessCredit model with the same stock price and equity volatility as you used for the MertonModel with a risk‐free rate of 5 percent, a loss given default of 60 percent, and a standard deviation of the default barrier of 50 percent. Try different input values for the default barrier level and see what the impact is on the probability of default and the credit spread for a five‐year maturity. 3. Prepare an analysis comparing the two models in terms of the impact on probability of default for changes in the stock price and changes in the equity volatility. 13.3 Using the Vasicek model for risk measurement of CDO tranches 1. Set the CDO spreadsheet to run the Vasicek model with Gaussian copula (i.e., set all tail factors and correlation factors to 100.00%). In this exercise we will just be experimenting with default rates, so we will not reduce input loss rates for assumed recoveries. 2. Assume that you have a portfolio of Bb loans. Using results from Table 13.5, set the input loss rate to 9.73%. Experiment with Credit Risk 503 different input correlation rates to see the impact on the standard deviation and 2.45th percentile losses for the portfolio. Notice that very low input correlation rates produce standard deviations and 2.45th percentile losses for the portfolio that look unrealistically low relative to historical experience, and that very high input correlation rates produce the opposite effect (for example, a 1% input correlation rate produces a portfolio standard deviation of 1.69% and a 2.45th percentile loss of 13.48%, while a 20% input correlation produces a portfolio standard deviation of 8.31% and a 2.45th percentile loss of 32.38%; Table 13.5 shows the historical standard deviation of Bb loan defaults to be 6.50% and maximum loss over any 5 year period to be 23.44%). 3. Through experimentation find an input correlation rate that produces reasonable results relative to the historical Bb loan default standard deviation and maximum loss. 4. Continuing with this example, experiment with different tranche attachment points to find one that will produce expected losses as a percentage of investment in the most senior tranche (the tranche with a 100% detachment point) roughly equal to the historical 0.27% loss rate for Aa loans from Table 13.5. Then compare the standard deviation of losses and 2.45th percentile loss as a percentage of investment for this senior tranche with the historical standard deviation of 0.44% and maximum loss of 1.83% for Aa loans from Table 13.5. You should see that even though the expected losses of the senior tranche match historical losses of Aa loans, the standard deviation and “worst case” losses are considerably higher for the senior tranche than they are for a portfolio of Aa loans. This illustrates the point made in Section 13.4.4 about the impact of tranching on concentration of systematic risk. 5. Further exercises with the CDO spreadsheet could involve experimenting with attachment and detachment points to try to create tranches that match other credit classes in terms of expected loss as a percentage of investment. You can also experiment with the impact of creating fatter tails than the Vasicek model by using input tail factors and correlation factors higher than 100%. CHAPTER 14 Counterparty Credit Risk ounterparty credit risk management arising from derivative contracts is an extremely important piece of the management of credit risk for reasons discussed in Section 14.1. Since the first edition of this book was published, the first full‐length book treatment of counterparty credit risk, written by one of the leading practitioners in this field, Gregory (2010), has appeared. I will be making frequent reference to this book in what follows and will provide several suggestions for further reading in Gregory that will provide greater detail and examples for points I will raise. C 14.1 OVERVIEW For credit risk, derivatives represent a two‐edged sword. On the one hand they have been valuable tools in reducing credit exposure, but on the other hand the use of derivatives leads to the buildup of credit exposure. The hope is that exposure reduction outweighs exposure buildup, but, without careful management, the full potential for credit exposure reduction by derivatives use will not be achieved. When financial derivatives markets first began to grow in the 1970s, the growth was primarily in currency and interest rate derivatives, and this remains the largest use to the current day (over 85 percent of the notional amount of contracts outstanding, according to figures from Tables 19 and 23A in the Bank for International Settlements’ December 2011 Derivatives Statistics). One use of these derivatives was to take on market exposures that could previously be accomplished only by cash instruments, such as loans, bonds, and deposits. As can be seen from Section 10.1.3 and Table 10.2, derivatives minimize the credit exposure and funding requirements entailed by loans, bonds, and deposits. The management of counterparty credit risk has been accomplished by two very distinct, but related, approaches: the use of derivatives exchanges 505 506 FINANCIAL RISK MANAGEMENT and the credit management of over‐the‐counter (OTC) derivatives that are not traded on exchanges. We will first discuss credit risk management through exchanges in Section 14.2 and the credit risk management for OTC derivatives in Section 14.3. The clear failure of many firms in managing their credit exposure on OTC derivatives (discussed in Section 5.3.1) has led to increasing pressure to move as much counterparty credit risk to exchanges and away from OTC as possible. The potential of and possible problems with this approach were discussed in Section 5.5.7. To see the significance of credit risk generated by derivatives, consider that U.S. commercial banks had $281 billion of credit exposure related to derivatives contracts at the end of 2011, about a quarter the size of their $1,339 billion credit exposure in traditional commercial and industrial loans (figures taken from the Federal Reserve’s H8 report). To see the impact of management of credit risk on derivatives credit exposure, as of June 2011 global credit exposure on all over‐the‐counter derivatives contracts was $19.5 trillion, compared with $707 trillion of notional outstandings, and there were another $83 trillion in notional outstandings on exchange‐traded derivatives contracts, on which there should be virtually no credit exposure, as we will see in Section 14.2 (all figures from Tables 19 and 23A in the Bank for International Settlements’ December 2011 Derivatives Statistics). 14.2 EXCHANGE‐TRADED DERIVATIVES Counterparty credit risk management of exchange‐traded derivatives rests on five key concepts: novation, margining, closeout, netting, and loss mutualization. The most important of these concepts is novation. As soon as two counterparties (let’s call them A and B) agree to a derivative contract traded on an exchange, the contract between the two counterparties is immediately canceled and replaced by two contracts, one between A and the exchange, and the other between the exchange and B; see Gregory (2010, Section 14.1.5) for details. Neither of the two counterparties needs to have any concern with the credit risk of the other—each has a contractual relationship for delivery on the derivatives contract with the exchange, and the exchange always has very low credit risk because it has the backing of all its members (we’ll discuss this further under loss mutualization), because it takes no market risk, and because it carefully controls its credit risk. To keep the discussion simple in what follows, I will write as if exchanges deal directly with all counterparties. Actually, a typical exchange has two classes of counterparties: exchange members who share in loss Counterparty Credit Risk 507 mutualization, and all others. It is only an exchange member who is permitted to be the direct counterparty of the exchange. All other counterparties are actually counterparties of one of the exchange member firms, which places trades with the exchange on behalf of these counterparties. But since exchange members manage their counterparty risk by the exact same method that the exchange handles its counterparty risk, through margining and closeout, a unified description is not too far removed from actual practice. At the end of this section, we discuss the extra detail that is needed to account for the two‐tier reality of exchanges and members. Also, in the interest of simplicity, I will always refer to the contracts as being with the exchange, ignoring the possible distinction between the exchange and its affiliated clearinghouse; see Gregory (2010, Section 14.1.3) for details. The exchange takes no market risk because its only positions arise as a result of novation and hence are always exactly offsetting positions. For example, if A contracts to deliver 100 million dollars to B in exchange for B delivering 70 million euros to A on a certain future date, this contract is replaced by A contracting to deliver 100 million dollars to the exchange for 70 million euros and B contracting to deliver 70 million euros to the exchange for 100 million dollars, both on the same date. So as long as both A and B perform their contractual obligations, the exchange will have no gain or loss, no matter what happens to the dollar/euro exchange rate. Hence, the exchange never bears any market risk. By contrast, the exchange must be very concerned about counterparty credit risk, since each trade leaves it with credit exposure to both parties of the trade. The exchange manages this credit risk through a very well‐ defined system of margining, closeout, and netting. The exchange is continuously monitoring the mark‐to‐market position of every trade, and any mark‐to‐market losses require a counterparty to immediately pay cash to the exchange to cover the loss (the exchange doesn’t keep this cash; it pays it to the counterparty with an offsetting mark‐to‐market gain). Any time a counterparty fails to provide the cash required to cover a mark‐to‐market loss, the exchange will declare the counterparty in default and close out all of the counterparty’s positions with the exchange. In this closeout, all of the counterparty’s positions, whether gains or losses, are netted against one another. The exchange seeks new counterparties to take over these positions. The exchange’s losses on these positions are limited to the change from the mark‐to‐market price the last time the defaulting counterparty posted margin and the price at which a new counterparty is willing to trade. The exchange has three ways in which to cover these losses: 1. First, each counterparty must post with the exchange initial margin at the time it first enters into a trade (this does not need to be in cash; 508 FINANCIAL RISK MANAGEMENT it can be some high‐quality security such as a Treasury bond). Losses in closing out a defaulted position will be charged against this initial margin before any money or securities are returned to the defaulting counterparty. 2. Second, if losses exceed the initial margin, the exchange will sue the defaulting counterparty for the remaining loss. However, recovery may be limited if the defaulting counterparty is actually bankrupt as opposed to just suffering temporary problems in meeting a margin call. 3. Third, any remaining losses are shared among all the members of the exchange. This is the principle of loss mutualization. In evaluating how much initial margin an exchange should demand to protect itself against the possibility of default, a key factor is to estimate probability distribution of price changes between the last mark‐to‐market and the transaction with a new counterparty. This depends crucially on the price volatility of the contract, the liquidity of the contract, and the speed with which the closeout mechanism operates. The more liquid a contract (i.e., the more frequently it trades and the larger the size of trading that occurs), the more confidence the exchange can have that the mark‐to‐market is close to the actual price at which a new trade can be done, and the lower the chance that the forced trading the exchange will do to close out the defaulted position will impact the price. As we already noted in Section 6.1.1, the management of counterparty credit risk through margining can follow very closely the prescription we have detailed for the market risk of trading: the importance of timely and accurate mark‐to‐market (Section 6.1.3), value at risk (VaR) (Section 7.1), and stress test (Section 7.2) calculations. In particular, VaR simulations and stress testing should look almost identical to the discussion in Chapter 7. As with trading positions, VaR will focus on losses that might occur under conditions of normal market liquidity, while stress tests will look at losses that might occur over longer periods between closeout and replacement with a new counterparty that result from unusual conditions of market illiquidity. There are two critical differences between the management of counterparty credit risk on exchanges and the management of the market risk for trading desks that impact VaR calculation methodology. One is that there may be a significant delay between the failure of a counterparty to meet a margin call and the declaration of default (this is called the grace period); it may take time for the exchange to confirm that a counterparty truly cannot or is choosing not to meet a margin call, rather than just a delay caused by an operational error or a communication failure. The time that is necessary to make this determination must be built into the VaR calculation, since it Counterparty Credit Risk 509 is a time period during which prices may fluctuate. Exchanges try to minimize required initial margin, since this is a key factor in the competition for business, and so will try to minimize the grace period. For example, as pointed out in Gregory (2010, Section 14.1.8), large price movements might trigger intraday margin calls, a practice that is becoming increasingly common and is supported by technology advances. But closing out too quickly may also result in loss of business to competitor exchanges, since it will unduly penalize operational errors. The second critical difference is that trading desks are experienced in managing market risk positions, and so can be expected to skillfully manage the required closing out of a position. By contrast, exchanges by their nature are not expected to have model risk positions, so closing out a position is not a task they are well positioned for. Exchanges protect themselves by limiting the number of contracts they will trade to a standardized set (e.g., allowing trading for only four settlement dates each year; see Hull 2012, Section 2.2, “Delivery Months”). By utilizing a limited set of standardized contracts, exchanges cultivate liquidity for each contract traded, making marking to market more robust and closeout easier to perform. Once VaR and stress test computations have been made, an exchange will be in a good position to evaluate the adequacy of initial margin requirements and to estimate the probability that the initial margins will prove insufficient to cover the losses incurred in a closeout. Some of the considerations that will go into evaluating the required size of initial margins are (compare with Gregory 2010, Section 14.1.8): ■ ■ ■ ■ The volatility of prices for the particular contracts involved and the length of the grace period, both of which should be direct inputs to the VaR and stress test computations. The degree of offset likely between netted positions in different contracts. This also should be an integral part of VaR and stress test computations, but with the same concern for the reliability of historical correlation relationships under stressed market conditions discussed in Sections 7.2.2 and 7.2.3. The size of the counterparty’s position relative to the size of trading in the contract. This is a point very similar to that raised in Section 6.1.4 regarding positions that are illiquid due to size. The remedy should be similar to that proposed in Section 6.1.4: simulation of price change between last mark‐to‐market and completed closeout should be over a longer time period to accommodate the larger position. The degree to which a counterparty has financial resources beyond its trading positions. This will impact the likelihood that losses could be recovered through a lawsuit. 510 FINANCIAL RISK MANAGEMENT ■ The degree to which a counterparty’s losses will tend to be correlated with those of a significant number of the exchange’s other counterparties. This might require VaR and stress test calculations that look at the whole universe of counterparties, rather than just one at a time. The methodology that exchanges use to manage counterparty credit risk through margining and closeout offers both drawbacks and advantages to counterparties. On the negative side is the narrow range of allowed contracts, which limits the degree to which derivatives can be tailored to meet specific needs of a customer. Also on the negative side is the operational complexity of meeting continuous margin calls. On the positive side, the heavy reliance of this approach on controlling credit risk through the actual mechanism of trading reduces reliance on credit evaluation of each customer. This can be very attractive to some customers who might not have the track record needed to withstand a credit review but who have confidence in their ability to manage margin calls. Another positive is that since the exchange has no market position, it has no incentive to hide information about prices at which trades have occurred and the depth of the market. Exchanges typically supply a much greater range and quality of price and market size information than do trading desks that are also holding market positions. Not only do exchanges generally provide complete public information on the sizes and prices of all executed trades, but “in typical exchange‐traded markets . . . the best available bid and offer are provided to nearly all market participants nearly instantly” (Duffie, Li, and Lubke 2010). One further negative that must be considered is that exchanges may protect themselves in instances of extreme market volatility by imposing limitations on trading that disadvantage some customers (this point is made forcefully in the section on clearinghouses in Brown (2012, Chapter 10). A very important positive of the exchange counterparty credit methodology is the ease with which a counterparty can offset a position previously entered into. As time and circumstances change, it is very common to wish to reverse a previous transaction. If your contract is with a private firm, as in an over‐the‐counter derivative, you must negotiate with this firm to offset the prior position. If your counterparty still wants to keep the position, you have a choice of either offering price concessions to induce your counterparty to offset the position or entering into an offsetting position with a new counterparty, which would offset the market position but leave you with credit exposure to both your original counterparty and the new counterparty. By contrast, the novation feature of exchange‐traded derivatives makes offset easy. Since your counterparty on any transaction is the exchange, you can find any new counterparty wanting to enter into an offsetting position Counterparty Credit Risk 511 and this will result in the complete cancellation of your original position with the exchange, leaving both you and the exchange with no further credit exposure on the original position or on your new offsetting position. (To make this completely clear, if the original position was between A and B, and later A enters into an offsetting position with C, A will be left with no exposure and the exchange will have offsetting positions with C and B, replacing its original offsetting positions with A and B.) As a final point, let us account for the actual two‐tiered nature of exchanges. As we said toward the beginning of this section, we have been simplifying by writing as if exchanges deal directly with all counterparties. In fact, it is only an exchange member, one who shares in loss mutualization, who is permitted to be the direct counterparty of the exchange. All other counterparties are actually counterparties of one of the exchange member firms, which places trades with the exchange on behalf of these counterparties. When a customer requests a trade through a member, the member is obligated to make that trade on the exchange, so members do not accumulate any market positions with customers. The exchange only needs to manage its credit exposure to its members, while each member needs to manage its credit exposure to its customers. The description we have given thus far, of margining, netting, closeout, VaR, and stress test calculations all apply equally to the exchange’s management of its credit exposure to members and to members’ management of their credit exposure to customers. If a customer’s position requires a margin call by the exchange, it is the member that is obligated to meet the exchange’s margin call, and the member in turn will make a margin call to the customer. From the viewpoint of a customer of a member, there shouldn’t be any difference between placing trades through the exchange and the actual placement of trades through a member—the obligation to pay the customer is the exchange’s obligation. The exchange will make payments due on mark‐to‐ market increases to the member firm, which is in turn obligated to pass these payments on to the customer. The only potential problem would be if the member does not adequately segregate customer funds from its own funds; in this case, if the member goes bankrupt, the customers could lose on initial margin accounts being kept with the member along with any funds the customer kept in excess of required margin, perhaps as an operational convenience to meet future margin calls. This was considered a remote possibility, given exchange rules and legal requirements for member firms. But the 2011 bankruptcy of MF Global and its failure to segregate customer funds left customers will long delays in access to funds and the definite potential for ultimate loss of part of their margin accounts (see Koutoulas and Roe 2012). It remains to be seen what impact this will have on customer views of the safety of exchange‐traded derivatives. 512 14.3 FINANCIAL RISK MANAGEMENT OVER‐THE‐COUNTER DERIVATIVES 14.3.1 Overview Given all the advantages of exchange‐traded derivatives, why do customers enter into OTC derivatives, which require far more credit scrutiny, are much more difficult to offset, and are surrounded by far more secrecy concerning prices and market conditions? The answer has to be largely centered on the two main weaknesses of exchange‐traded derivatives: lack of customization and the operational intensity of managing margin calls. Firms that want to enter into derivative contracts custom‐tailored to a specific need must use OTC derivatives. An additional motivation for using OTC derivatives is that a counterparty may be seeking an extension of credit in connection with its derivatives trading. Initial margin and daily margin calls require cash or securities that the firm may need for other purposes. Unlike an exchange, the provider of an OTC derivative may be willing to extend credit for an amount that is due in the future under a derivative contract. While some OTC derivatives contracts are negotiated directly between two firms looking for opposite sides of a trade, the overwhelming majority of OTC derivative contracts involve a derivatives market maker as one of the counterparties to the trade. This reflects both the willingness of derivatives market makers to structure contracts that fit the particular needs of a customer and the nature of market making in providing continuous liquidity by being willing to take either side of a trade at a reasonable price, as we discussed in Section 2.5. Finding a non‐market‐making firm looking for the opposite side of a trade you want to enter into requires an extensive search. A market maker in derivatives must therefore have both a sophisticated trading operation with regard to market risk and a very high credit rating. In cases where there have been credit concerns regarding a market‐making firm, special arrangements have been made to create a subsidiary that has a higher credit rating than the parent firm that will be the counterparty to all derivatives trades (for details, see Gregory 2010, Section 2.3.1 and Chapter 13). We can therefore see that in many ways the derivatives market maker plays a very similar role to that of the exchange in managing the credit risk of derivatives. Parties taking opposite market positions have credit exposure to a market maker rather than to one another. But the market maker has more freedom than an exchange in deciding how it wants to manage this credit exposure; the loss mutualization rules of the exchange make it answerable to all of its member firms and constrain its options. Counterparty Credit Risk 513 Three primary approaches have been proposed and used for managing counterparty credit risk for OTC derivatives. The earliest approach was to treat the counterparty credit risk on OTC derivatives as much as possible like the traditional credit process for loans. We will discuss this approach in Section 14.3.2. The second approach is to incorporate some of the credit management tools of exchange‐traded derivatives to OTC derivatives— closeout, netting, and margining. This approach will be discussed in Sections 14.3.3 and 14.3.4. In 14.3.5, we will discuss the most recent of the approaches, the use of dynamic hedging to manage counterparty credit risk. 14.3.2 The Loan‐Equivalent Approach The earliest approach to the management of counterparty credit risk on OTC derivatives was to incorporate it into the traditional credit process for loans. Since credit risk managers are used to making decisions on the total amount of credit that it is prudent to extend to a given borrower, it is only necessary to calculate the total loan‐equivalent size of credit extension needed for a given OTC derivative position. The difficulty with this approach is that where a standard loan (other than a line of credit) has a fixed amount that is subject to loss in the event of default, the size of derivative exposure at the time of default depends on the uncertain evolution of market conditions. The standard solution to this problem has been to set some probability threshold (such as the 99th percentile) and then estimate the near‐maximum amount that can be lost in the event of default at this threshold. This near‐ maximum loss amount is treated as a loan equivalent, and credit risk managers are asked to give approval for this added credit extension to the borrower. Before discussing the computational aspects of this approach, let us note two major issues: 1. Credit risk management looks not just at total credit exposure but also at the expected recovery in the event of default. While historical experience has been developed for recovery on different classes of credit exposure (see Table 13.4), the relative rarity of default by OTC derivative counterparties has made comparable data difficult to obtain. Some assumption about this recovery rate needs to be made based on some combination of relevant experience and theoretical considerations. 2. Derivatives marketers and traders feel discriminated against by this traditional approach. They point out, with reason, that the actual amount at risk in the event of default would, on statistical grounds, often be less than the near‐maximum amount used as a loan equivalent, whereas a 514 FINANCIAL RISK MANAGEMENT traditional loan will always have the same fixed exposure. Derivatives marketers and traders want to see notions of expected exposure at default supplement or replace the measure of near‐maximum exposure at default. However, care must be taken to create a comparable measure to traditional loans. If traditional loan exposure is measured by loan amount, the expected exposure on derivatives must be measured by exposure at default and not be based on expected loss, which differs from expected exposure by the amount of expected recovery in the event of default. With respect to the second point, there is near‐universal agreement that expected exposure at default should be measured and that loan officers in making decisions on credit extensions should look at expected exposure along with near‐maximum exposure. There is also near‐universal agreement that pricing credit exposure and allocating capital against credit use, as in Section 13.3.4, should be based on expected exposure. More controversial is the proposal by some derivatives marketers and traders that near‐maximum exposure should not be considered at all and that only expected exposure should be looked at as a measure of credit risk on OTC derivatives. In my experience, this argument has not gained much traction. Certainly for borrowers with very large exposures, the potential impact of default on the lending firm makes it mandatory for credit officers to consider the near‐ maximum impact. Even for smaller borrowers, the discipline of looking at near‐maximum exposure is a healthy incentive to focus loan officers on the soundness of credit extension decisions. Turning to the computational aspects of the loan‐equivalent approach, there are two basic methodologies to consider: simulation and formulas. Consistent with the basic themes of this book, I advocate the use of simulation as the recommended approach (compare with Sections 1.3 and 6.1.1). Simulation is more accurate than formulas in the calculation of credit exposure on a single derivative, is an absolute necessity for looking at credit exposure of the full set of derivatives for a counterparty or for pricing credit exposure for a portfolio of counterparties, is needed for taking into account correlation between market movements and default probability (so‐called wrong‐way risk), and is an absolute necessity for taking into account credit mitigation techniques such as netting and margining. We will postpone the discussion of the details of simulation methodology until after the introduction of credit mitigation in Section 14.3.3, allowing for a unified simulation approach. For all these reasons, calculation of credit exposure through formulas has limited applicability and is relied on only by smaller, less sophisticated firms. However, larger and more sophisticated firms may still utilize Counterparty Credit Risk 515 formulas as a quick first approximation to guide initial discussions between derivative traders and loan officers and as an aid to intuition. These approximations are usually based on the reasonable assumption that uncertainty about market variables will grow with the square root of elapsed time, balanced by the decrease in duration of products such as interest rate swaps. For a swap, increasing uncertainty at first dominates, and credit exposure increases, reaches a peak, and then declines through time as the impact of decreasing duration comes to dominate. Gregory (2010, Section 4.2 and Appendix 4A) contains examples of approximation formulas and graphs illustrating typical cases. For less sophisticated firms attempting to approximate counterparty credit exposure without the use of a full simulation model, portfolio credit risk as calculated in Section 13.3.2 will just have expected loan equivalents representing counterparty exposure as input. Portfolio credit risk computed with this shortcut must be adjusted upward to take into account interactions between credit exposure and market value that would be picked up in a full simulation. This is the so‐called alpha factor explained in detail in Gregory (2010, Sections 10.4 and 10.5). This exposure increase is present even in the absence of any correlation between default probabilities and market values, simply due to the added volatility of market values contributing to higher tail risk of the credit portfolio. 14.3.3 The Collateralization Approach The second approach to managing counterparty credit risk on OTC derivatives has been to combine the first approach just described with tools borrowed from the exchanges’ management of counterparty credit risk. In particular, a combination of netting and closeout is used to combine derivative positions with a single counterparty, and margining is used to obtain collateral that will offset loss in the event of default. Let’s look at these two tools in some more detail. Netting and closeout are discussed at length in Gregory (2010, Sections 3.4 and 3.5). According to Gregory, “Of all risk mitigation methods, netting has had the greatest impact on the structure of the derivatives markets. Without netting, the current size and liquidity in the derivatives markets would be unlikely to exist. . . . The expansion and greater concentration of derivatives has increased the extent of netting from around 50% in the mid‐1990s to close to 100% today.” Netting and closeout require a legal agreement between counterparties, most typically under an ISDA Master Agreement (see Gregory 2010, Section 3.4.6), that permits, in the event of a counterparty default, the nondefaulting counterparty to immediately terminate all outstanding derivative contracts between the two counterparties, 516 FINANCIAL RISK MANAGEMENT determine what is owed on each terminated contract at current market values, and net offsetting amounts owed. It eliminates the possibility of the defaulting counterparty settling contracts on which it owes money at only a recovery fraction of the amount owed, while demanding full payment on contracts on which it is owed money. According to Gregory, “ISDA has obtained legal agreements supporting their Master Agreements in most relevant jurisdictions” (wherever there are doubts about legal enforceability of closeout netting in a jurisdiction, ISDA lobbies for legislative clarity; once clarity has been achieved, ISDA obtains a legal opinion to this effect for the benefit of its members). Another major advantage of the ISDA Master Agreement is that it has standardized procedures for determining what claims can be made in bankruptcy against a defaulting counterparty. The suggested ISDA language defines the amount that can be claimed as the amount that the nondefaulting party “reasonably determines in good faith to be its total losses and costs” as of the closeout date and states that the nondefaulting party “may (but need not) determine its loss by reference to quotations of relevant rates or prices from one or more leading dealers in the relevant markets.” This language makes clear that the nondefaulting party does not have to enter into a replacement transaction in haste in order to establish a price on which to base its claim in bankruptcy proceedings. Instead, it can utilize market quotations, supplemented by industry‐standard models, to establish what the mark‐to‐market of the transaction was at the time of default, base its bankruptcy claim on that, and exercise its best judgment as to when or whether to actually enter into a replacement transaction. Margining is discussed at length in Gregory (2010, Sections 3.6, 3.7, and 5.2.1). It works similarly to margining by exchanges, with a call for posting of margin to cover a mark‐to‐market loss and the failure to post margin constituting a default event that will terminate the trade (and all other trades linked through netting agreements). If OTC derivatives margining worked exactly like exchange margining, it would completely eliminate the advantages of OTC derivatives over exchange‐traded derivatives in operational simplicity and credit extension (though still leaving contract customization as an advantage). To retain these advantages, OTC derivatives market makers usually make their margining requirements less burdensome than exchange margining requirements by one or more of the following conditions: ■ ■ Margin payments may not be required as often as daily, but may have a less frequent period, such as weekly or monthly. Margin payments may be required only once a certain mark‐to‐market loss threshold has been reached. Counterparty Credit Risk ■ ■ ■ 517 Margin may be allowed to be posted as securities of a specified quality rather than necessarily being cash, though this provision has been losing popularity since events of the 2008 crisis (Gregory 2010, Section 3.6.5). Initial margin may not be required. More leniency may be permitted in allowing a grace period during which the counterparty has time in which to post margin. These more lenient margining requirements allow OTC derivatives market makers to accept a greater degree of credit exposure to customers than is normally extended by exchanges. With this background on netting, closeout, and margining, let’s begin to look at the computation of counterparty credit risk exposure by simulation. There are very strong parallels to the use of simulation and stress testing that can be found in Chapter 7, and much of that material is fully applicable to counterparty credit exposure. As in Chapter 7, we are concerned with the value at which a transaction will actually take place—the replacement value at which a derivative contract can be entered into in the event of default for counterparty credit exposure versus the exit value of an existing transaction in the event of forced liquidation for market risk. The primary differences between the market risk simulation of Chapter 7 and the simulation of counterparty credit exposure are length of simulation period and the required statistics. Counterparty credit exposure must be calculated over much longer time periods than VaR, since a firm can exit its market exposures over a period of a few days but has a longer contractual commitment to the credit risk on derivatives. While market risk simulations are concerned only with tail risk, counterparty credit exposure simulations need to calculate expected value as well as the tails, as already explained in Section 14.3.2. For the time being we will assume that the timing of default of the counterparty is independent of the market values of the derivative contracts. We will later drop this assumption in the next section, on wrong‐way risk. Here are some points that must be considered in designing counterparty credit exposure simulations in addition to the points already covered in Chapter 11; for a more detailed description, see Gregory (2010, Chapters 4 and 5), and also compare with Brindle (2000) and Canabarro and Duffie (2003). ■ The longer time period that counterparty credit exposure simulation requires necessitates the use of Monte Carlo simulation. With VaR simulation, we can choose between historical simulation and Monte Carlo simulation only because the short time period being simulated means 518 ■ ■ ■ ■ ■ FINANCIAL RISK MANAGEMENT there are many previous historical periods of the same time length as the period to be simulated. Each path of the Monte Carlo simulation determines credit exposure at each possible default time being considered. Calculations along each path take into account not just the values of the derivative contracts but also account for all netting that would occur in the event of default and any margin calls and collateral postings that would have occurred based on the details of the margining agreement with the counterparty. Since a single counterparty may have entered into many different types of derivative contracts (equity, interest rate, foreign exchange [FX], credit, commodities, etc.), a full range of market variables must be considered, just as in a VaR calculation, with due care exercised on correlation assumptions between variables. As with VaR simulations, full valuation of derivatives along each simulation path may be very resource intensive, and trade‐offs will exist between the accuracy of full valuation and the faster turnaround time and lower cost of approximations (compare with the discussion of valuation approximations in Section 7.1.1.2). This is an even greater issue for counterparty credit simulations than for VaR simulations, since each path also requires valuations at many different time periods; see the section on “Computational Considerations” in Brindle (2000) and Gregory (2010, Section 4.1.3). To speed computation, in addition to the approximation measures discussed in Section 7.1.1.2, the number of default times for which valuation is done may be reduced with interpolation utilized for default times in between the ones evaluated. Gregory (2010, Section 4.1.4) discusses possible issues with interpolation between the discrete time points for which calculations are made and measures for reducing interpolation error. In counterparty credit exposure simulations, the drift (the expected change in a variable through time) plays a more important role than in VaR calculations. Due to the short time frame of VaR calculations, drift can be assumed to be zero, since volatility will totally dominate drift, particularly in tail calculations. But for counterparty credit exposure, over much longer time periods and where expected value is important along with tail values, drift is very important. As Gregory (2010, Section 4.3.2) notes, “in the long run a strong drift will dominate” since volatility varies with the square root of time whereas the drift scales linearly with time. So attention must be paid to forecasting the drifts of market variables in the Monte Carlo simulation. In simulating margining, in addition to all contractual details, assumptions need to be made about delays in the time between a margin call being made and a default for failure to meet the margin call being Counterparty Credit Risk ■ ■ 519 declared. As Gregory (2010, Section 5.2.1) explains in detail, the industry standard incorporated into Basel II is to assume a 10‐business‐day minimum remargin period between margin call and default declaration and closing out of positions. This allows time for both operational issues of processing margin requests and delays in detection of nondelivery, and grace periods allowed to permit a counterparty to cure a failure to post margin. As Gregory notes, longer remargin periods may be appropriate for counterparties that may be granted more leniency to maintain good relations or where the nature of the derivatives may require longer periods to resolve disputes over the mark‐to‐market driving a margin call. Brindle (2000) also notes that in some jurisdictions, statutory stay periods may delay the liquidation of collateral, and contractual agreements may stipulate a minimum delay period. The closeout delays assumed in the simulation should be individually tailored to each counterparty. When a counterparty agreement allows for noncash collateral, the market value of the collateral should also be simulated along each of the simulation paths, with full consideration of correlation between value of the collateral and value of the derivatives. When the value of the derivatives and of the collateral instrument are positively correlated (e.g., a Treasury bond as collateral and a set of swaps on which the counterparty net pays a fixed rate in the same currency), credit exposure will be greater than if collateral was posted in cash. When the value of the derivatives and of the collateral instrument are negatively correlated (e.g., a Treasury bond as collateral and a set of swaps on which the counterparty net receives a fixed rate in the same currency), credit exposure will be less than if collateral was posted in cash. A worked example of the impact of collateral on credit exposure can be found in Gregory (2010, Section 5.2.5). As with VaR, as discussed in Section 7.1.1.2, counterparty credit exposure simulations must account for illiquidity, whether due to infrequent trading or to a large position. Illiquidity must be considered for both the derivative positions and for noncash collateral. Whether due to infrequent trading or to large positions, the basic tool for dealing with illiquidity of derivatives is to lengthen the time assumed between a default event and position closeout. This closely parallels the treatment for illiquidity detailed in Section 6.1.4 and the provision for remargin periods discussed two bullet points previously. Illiquidity will probably have limited impact on counterparty exposure where margining is not used—there is little difference between the price movement in derivative value over, say, two years from now to default and over two years and two weeks, allowing an extra two weeks after default for illiquidity. 520 ■ ■ ■ FINANCIAL RISK MANAGEMENT But illiquidity can have a major impact on counterparty exposure when margining is used. It could now, for example, double the time from default event to closeout from two weeks to four weeks, increasing exposure by 2 . Similarly, illiquidity of collateral can be treated by increasing the time period over which the collateral is assumed to be liquidated and hence the uncertainty of the price realized. When illiquidity of a derivative is due to a position with actuarial risk, a separate treatment is needed. This will be discussed next. For derivatives with actuarial risk, I strongly favor an approach parallel to that recommended in Section 6.1.2: utilize a liquid proxy in the counterparty exposure simulation but make a separate computation for the residual risk. My argument is that a default by the counterparty will result in the nondefaulting party acquiring and now needing to manage the actuarial risk in the same way it would have needed to manage it if it had created it in a trading position. The reserves that would be needed to manage out of the position, as computed in Section 8.4, will now be a potential cost and hence are an addition to near‐maximum credit exposure and need to be accounted for in expected credit cost, multiplied by the proper default probability and loss given default percentage. Another consequence of this argument is that firms should not enter into derivatives positions on transactions for which they lack adequate models and personnel to manage a position that will result from a default. There have been unfortunate examples in which firms have decided to “stand in the middle” between two counterparties on a transaction that they had no experience trading and little understanding of, persuaded that they were “only” taking a counterparty credit risk and not taking any market risk (typically because one of the counterparties was not willing to accept the credit risk of the other and was looking for a counterparty with stronger credit risk). On default of one of the counterparties, these firms found themselves suddenly needing to manage positions they lacked competence to trade. Stress testing as a supplement to simulation of counterparty credit exposure plays a smaller role than it does as a supplement to VaR, for reasons similar to those discussed two bullet points previously concerning the minimal impact of illiquidity of positions on counterparty credit exposure. By parallel reasoning, a stress scenario of a temporary period of market illiquidity in normally liquid positions will have little impact on exposure when no margining is employed but may be quite necessary and of significant impact when margining is employed. The degree to which netting reduces near‐maximum credit exposure is very heavily impacted by correlation assumptions regarding market variables. It is important to make sure that subjective probabilities of Counterparty Credit Risk 521 future periods in which correlations that are either very low or very high by historical standards have been given due consideration. The need for communication of marginal cost of new credit exposures to loan officers discussed in Section 13.3.4 has a parallel requirement for communicating the marginal cost of new counterparty credit exposures to loan officers, traders, and structurers. This is done through the credit value adjustment (CVA), a thorough discussion of which can be found in Gregory (2010, Chapter 7). Gregory’s discussion of measuring marginal exposure contributions in his Section 4.5 is also relevant. I will limit myself to just a few remarks related to cases where the CVA methodology differs in some respect from the methodology of Section 13.3.4: ■ ■ As in the more general case of marginal credit exposures discussed in Section 13.3.4, there is a need for approximations that can be used at the individual credit level. Gregory provides approximation formulas in the appendixes to Chapter 7, but with the important caveat that these work only in the absence of wrong‐way risk (i.e., when there is no dependence between default probability and loss given default). When wrong‐way risk is present, the techniques of the next section, 14.3.4, need to be used; these are very closely related to the computations in Section 13.3.4 and so would need to utilize approximation techniques covered there, though Gregory’s Section 8.3 does provide some approximation formulas specific to CVA for wrong‐way risk. Many firms have employed an accounting procedure that takes into account the impact on derivative contracts of the default probability of the firm itself (this is termed bilateral counterparty risk and is covered by Gregory in Section 7.3). Whatever its virtues as an accounting procedure, it should never be utilized in risk management measures such as CVA. From a risk management standpoint, there are no benefits to a firm from its own default, so utilizing it in risk measures would be completely misleading. Even as an accounting procedure, the benefits of this approach are dubious: an attempt to book profits that will fuel short‐term bonuses at the potential expense of investor confidence in the firm’s reported earnings, as can be seen in the examples in Gregory (2010, 188). 14.3.4 The Collateralization Approach—Wrong‐Way Risk In the previous section on simulation of counterparty credit exposure, we noted that a key assumption in our calculations was independence of counterparty default and market value of the derivatives contracts. For many 522 FINANCIAL RISK MANAGEMENT counterparties, this is a reasonable assumption. When there is a correlation between default and market value, then computations must be different. A positive correlation between probability of default and market value of the derivatives is known as wrong‐way risk and increases exposure and CVA measures from what they would have been in the absence of this correlation. A negative correlation between probability of default and the market value of derivatives is known as right‐way risk and decreases exposures and CVA measures from what they would have been in the absence of this correlation. Gregory (2010, Chapter 8) contains a thorough exposition of wrong‐way and right‐way risk. This section addresses how to modify the simulation methodology of the previous section to accommodate this correlation. The short answer is that the default probability of the counterparty must also be simulated along each path, incorporating correlation with the market variables being simulated. We will provide details and examples shortly, but first let us consider some cases in which wrong‐way risk is so extreme that simulation should be circumvented and a direct analysis should be made. Let’s start with a trade that has, unfortunately, been proposed all too often by trading desks over the past 15 years. With macabre humor, it is sometimes called an “end of the world” trade. It is a proposal to put on a derivative trade that will provide a payoff only if some really extreme event occurs—let’s say 40 percent defaults on a basket of investment‐grade corporate loans. No initial margin is being asked of the counterparty providing the protection, and there is no provision for margin calls. It is easy to see the attractiveness of this trade from the viewpoint of the counterparty providing the protection; it will receive a small annual payment every year, and if the dire circumstances in which it is required to make a payment did occur, it doubts it would still be in business. It is harder to see why the firm purchasing the protection would want to do the trade. In every case I have encountered, when I asked the trading desk proposing the trade whether they thought there was any chance the counterparty would still be in business if it was required to make a payment, the answer was, “No, but even though this has no financial benefit to the firm, it will provide us relief under such‐and‐such regulatory capital calculation.” My response, as a risk manager, was always: (1) we wouldn’t permit trades to be made that cost the firm money with no financial benefit, and (2) even if it appeared to provide regulatory relief, it would be my obligation as someone in a control function to point out to the regulatory authority concerned that it was being gamed. In no way was any modeling required to come to this conclusion. A less obvious case is one in which no margining is required by a counterparty unless the counterparty receives a ratings downgrade below Counterparty Credit Risk 523 a certain level or unless an extremely negative event occurs in the counterparty’s stock price or credit spread, in which case a large margin payment is required. In such circumstances, I have always been opposed to giving any credit in counterparty credit exposure calculations for this margining requirement; I would make the calculations assuming no margin requirement at all. My reasoning is that the type of event that triggers the margin call is just the sort of circumstance in which the counterparty will be strapped for cash and will either be forced to default or will appeal to our firm’s senior management for relief from the margin call to avoid bankruptcy. Indeed, it was just this type of margining provision that pushed Enron into bankruptcy (see McLean and Elkind 2003, 394–395). So this is a case of wrong‐ way risk in which there is a high correlation between a required margin payment and a default that prevents it being made. Another variant on extreme wrong‐way risk is an attempt to avoid reliance on margin calls that have a low probability of being fulfilled by converting the counterparty credit risk into a gap market risk. A detailed and instructive worked example of this mechanism is given in Gregory (2010, Section 8.6.4). I will build on Gregory’s example in the discussion that follows, but with only a brief sketch of Gregory’s details. In the example, the market‐making firm buys or issues a $100 million credit‐linked note (CLN) and enters into a total return swap on the CLN with a hedge fund. The hedge fund posts $10 million in initial margin and benefits from having a highly leveraged position, receiving a return on the $100 million note while only needing to invest $10 million in collateral. The downside for the market maker is that it knows that if the market value of the CLN starts to decline toward $90 million, it is highly unlikely that the hedge fund will be able to post additional margin, since the hedge fund, under the circumstances that credit spreads have risen high enough to create this size market loss on the CLN, will likely be in trouble due to its high leverage and probable losses on similar trades. The market maker’s trading desk knows it is unlikely to get any credit for margin call provisions due to the extreme wrong‐way risk. A possible alternative is to exclude the margin call provision but instead put in a provision that if the value of the CLN gets too close to exhausting the $10 million initial margin, the market maker has the right to close out the position and sell the CLN. In Gregory’s example, a provision is set that if the value of the CLN is at or below $92.2 million, the position can be closed out. This is supposed to leave the evaluation of the trade entirely to market risk managers since there is no credit risk component remaining. The only losses to the market maker can occur if the gap between the $92.2 million trigger point and the price at which the CLN can be sold exceeds the $2.2 million of remaining initial margin. It is the probability of this large market move 524 FINANCIAL RISK MANAGEMENT occurring that is supposed to be evaluated by standard market risk VaR and stress test methodologies. I have always been dubious of this type of attempted end run. I think it just replaces one form of wrong‐way risk with another form of wrong‐ way risk: the high correlation between large drops in price of the CLN and large subsequent gap moves. The fundamental flaw in the appeal to VaR and stress test methodologies in evaluating the gap risk is that VaR and stress testing are designed to evaluate the risk of current positions based on current market conditions. For gap risk, we are being asked to evaluate a future position under future market conditions and one that will be triggered by conditions likely to be unfavorable to us. As such, they fall under one of the criteria proposed for actuarial risk in Section 6.1.1, positions that can be liquidated only under restrictive conditions. Hence, they should be evaluated using the tools of Section 8.4, with very conservative reserves to allow for the illiquidity of the position. Subjective judgment by risk management would be required as to the size of gap moves that could occur following the very negative market events that would cause the trigger to be reached. One more variant of extreme wrong‐way risk is the liquidity puts described in Section 5.2.5.2. Here an investment bank was selling an extremely illiquid asset, a super‐senior tranche of a CDO, but with the provision that if the firm buying this asset encountered funding difficulties it could sell the asset back to the investment bank at par. This type of transaction should be treated for stress testing purposes as if the asset had not been sold at all—the firm buying the asset would probably run into funding difficulties only in a period of widespread financial distress, exactly the circumstances in which the asset is likely to be worth significantly less and be even harder to find another buyer for. Since the assessment of the potential losses on the asset were that it would lose significant value only in a period of unlikely widespread financial distress, allowing it to be placed back to the investment bank in these circumstances reduces the risk reduction for stress testing purposes of selling the asset to a negligible amount. We now turn to the details of simulation incorporating correlation between default probabilities and market variables for those instances of wrong‐way and right‐way risk that do require full calculation. ■ Instead of assuming that default occurs independent of market variables, we now directly simulate default probabilities and allow the Monte Carlo simulation to work from these default probabilities to assign defaults to particular paths and time periods. Expected and near‐ maximum exposure values are computed from only those points at which default has occurred. If those default points are correlated with Counterparty Credit Risk ■ ■ ■ ■ ■ 525 market value of the derivatives positions, this will be reflected in the simulation results. Correlations between default probabilities and market values will need to be established by a combination of subjective judgments based on economic insight and statistical studies of correlations between market variables and credit spreads as a proxy for default probabilities. Much depends on the degree of business diversification of a counterparty. A counterparty with many business lines in different countries and different industries is far less likely to be subject to wrong‐way risk than a counterparty with a highly concentrated business. One of the most obvious examples of wrong‐way risk stems from country risk. A counterparty whose financial health is very dependent on business in a single country is likely to have a high correlation between its default probability and the exchange rate and interest rates of that country. This most frequently impacts long‐term FX forwards or cross‐currency swaps. As pointed out by Gregory (2010, Section 8.2.3), “another way to look at a cross‐currency swap is that it represents a loan collateralized by the opposite currency in the swap. If this currency weakens dramatically, the value of the collateral is strongly diminished.” A business whose viability is likely to be strongly impacted by the price of a particular commodity such as oil should show a strong correlation between default probability and the commodity price. Correlations between default probabilities of firms based on industry and country have already been discussed in Section 13.3.1. This can have a strong impact if a counterparty is highly correlated with a firm on which it is writing credit protection through a credit default swap (CDS). One of the principal sources of wrong‐way risk historically has been the use of CDS counterparties closely related to the firm on which protection is being purchased. The credit portfolio simulations of Section 13.3.2 should be able to capture this. Consider, for example, a loan to Company ABC for which credit protection has been purchased from Company XYZ through a CDS. No loss will occur if ABC defaults and XYZ has not defaulted, since, in this circumstance, XYZ must pay all the costs of the ABC default. If XYZ defaults and ABC has not defaulted, the firm will have a loss (or gain) equal to the replacement cost of the CDS, which is driven by changes in the credit spread for ABC. The simulation calculates this by keeping track of changes in default probabilities and credit spreads for both firms along each simulation path, taking the proper correlation between the default probabilities of the two firms into account, and linking the default probability of ABC to the credit spread of ABC. Gregory (2010, Section 8.4) provides 526 ■ FINANCIAL RISK MANAGEMENT more detail and examples illustrating wrong‐way risk on CDSs, and in Section 8.5 extends this analysis to wrong‐way risk on CDOs. A significant source of wrong‐way risk is counterparties who derive a major portion of their revenues from financial transactions. In such cases, an estimate must be made of how much of the counterparty’s trading positions are similar to those on which your firm holds positions with the counterparty. The more similar overall trading positions are to those with your firm, the more likely that default probability has a high correlation with market variables impacting those positions. While simulation is a requirement for accuracy in measuring wrong‐ way risk, formulas can be utilized for quick approximations that are useful in gaining intuition and to guide initial discussions between derivative traders and loan officers. Examples of useful formulas and illustrated cases can be found in Gregory (2010, Section 8.3) and Winters (1999). 14.3.5 The Active Management Approach The third, and newest, approach to managing counterparty credit risk for OTC derivatives involves the active use of purchased credit protection through CDSs (or, equivalently, by short selling of bonds). As such, it shares many of the characteristics of active management of credit portfolios discussed in Section 13.3.4, involving trade‐off decisions about when to purchase protection versus when to self‐insure by extending credit lines, the communication of internal pricing of new credit extensions based on a combination of the cost to purchase CDS protection and the cost of required capital against self‐insurance risk, the active involvement of marketers and traders in making judgments about whether the extension of new credit is worth paying the internal charge, and the management by a central unit of the cost of loan defaults against the revenue accumulated by internal charges for credit extension. The difference between the active management of counterparty credit risk and of portfolio credit risk is that counterparty credit risk active management involves simultaneous management of the cost of credit exposure and the dynamic changes in size of credit exposure due to changes in the market value of counterparty positions. This requires a very specialized skill set that has led most large derivatives dealers to set up specialized business units (counterparty risk groups [CRGs]) for the dynamic management of counterparty credit risk. Gregory (2010, Chapter 12) gives an extended discussion of how this is done. The centralized unit for managing counterparty exposure will need to create a mechanism for charging trading desks for protection against counterparty risk. This mechanism must follow many of the same criteria as Counterparty Credit Risk 527 outlined in Section 13.3.4 in the context of the more general issue of how to charge marketing areas for the extension of credit risk, but with the added complexities of estimating the credit exposure arising from market movements. These charges should create the incentives for trading desks and derivatives structurers to design contracts that minimize credit use. There will be trade‐offs between customer desire to minimize the use of devices such as margin calls and the reduction in credit charges that result from such devices. It is the task of traders and structurers to find clever designs that bring the greatest reduction in credit charge for the least amount of customer dissatisfaction. To the extent this counterparty risk group (CRG) decides to manage counterparty credit risk with the purchase of CDS protection, it requires the use of dynamic hedging techniques originally developed for multiasset exotic derivatives such as quantos. The size of market exposure at any instant is the product of the credit spread of the counterparty and the size of the credit exposure. As we illustrated in Section 12.4.5, this requires dynamic hedging, with a change in derivative value requiring a change in the size of the credit hedge, and a change in the credit spread requiring a change in the size of the derivative hedge. Essentially, this method amounts to replacing the derivative with another counterparty, not all at once on default, but gradually as the original counterparty’s credit worsens. Correlation assumptions, driven by wrong‐way exposure concerns, will have the intuitively correct effect of increasing the expected cost of the dynamic hedge. The CrossHedge spreadsheet gives a detailed example of the dynamic hedging of a counterparty credit position with results shown in Table 12.13. What the example in Section 12.4.5 illustrates is that, to a good degree of accuracy, the dynamic hedge allows locking in credit protection on the counterparty at the current market credit spread, even though the amount of credit protection will vary over time in a stochastic fashion. This is quite counterintuitive—it would seem that if credit spreads widened at the time that exposure grows you would need to purchase some of the credit protection at higher spreads. But the dynamic hedging approach means that you are always simultaneously hedged against both changes in credit spread and changes in exposure (always with the exception that correlation in price movements between the credit spread and the market exposure caused by wrong‐way exposure leads to extra costs). This allows the CRG to be able to price credit exposure at the time of agreeing to the derivatives contract with reasonable confidence. While the example in Section 12.4.5 is written from the point of view of credit protection on a single derivatives contract, the mechanism actually works for covering an entire portfolio of derivatives— essentially, you just substitute the volatility of the whole portfolio for the volatility of the single contract. 528 FINANCIAL RISK MANAGEMENT In practice, a CRG will choose to use CDS hedging on some exposures and not on others—some counterparties will not have sufficient liquidity in the CDS market to allow the dynamic hedging technique to be used; for other counterparties the credit managers will judge that their view of the credit risk of the name is more favorable than what is priced into the CDS market and they will choose to self‐insure for that name, at least for a time. In other cases, mixed approaches will be taken—names that lack a liquid CDS market but whose exposure is at an uncomfortable level for the credit managers will be proxy hedged with a basket of more liquid CDSs on similar names being used to hedge a basket of less liquid names, with the risk having been transformed from outright default risk to the basis risk on default experience of the basket hedge and default experience of the actual basket. The simultaneous dynamic hedging of credit spread (for the proxy basket) and market exposure works in this case as well. When utilizing dynamic hedging of counterparty credit exposure, a CRG will need to utilize risk measures similar to those we have discussed for dynamic hedging of options in Section 11.4, but with the added complications that exposures to credit and to market variables are being managed simultaneously and that credit risk requires risk measures that include exposure to immediate default. A thorough discussion of the risk measures required can be found in Gregory (2010, Chapter 9). One issue for CRGs that has been much debated and is highlighted by Gregory (2010) in Section 12.4.4 is whether the CRG should engage in dynamic hedging of the market exposure of a derivatives book in a case where it is completely self‐insuring the credit risk for that counterparty. Unlike the dynamic cross‐hedging examples just given, there is no cost of a CDS position that is being offset by the market exposure hedge. All that is being hedged is an accounting entry of the mark‐to‐market of the self‐insurance strategy. The economic value of paying money to hedge accounting entries is regarded with extreme suspicion by many risk management practioners, myself included. But if there is some form of active hedging in the CDS market, even if it is only against a basket of names that provide a liquid proxy, then I would find dynamic hedging of market exposure to be quite reasonable. In taking over management of the counterparty credit risk of derivatives, the CRG must be prepared to manage all aspects of a counterparty default (see Gregory 2010, Section 12.2.6). This includes the settlement process on any CDS protection that has been purchased (which may involve delivery squeezes, as discussed in Section 13.1.1.2), the legal process for recovery of amounts owed, and responsibility for the liquidity costs of replacing defaulted contracts. The CRG must factor all of these possible costs into its pricing of default insurance to the firm’s trading desks. Counterparty Credit Risk 529 There are other strategies that a CRG can pursue in providing protection. It might, for example, contact a counterparty with which the firm has a large outstanding exposure and seek to negotiate a reduction in exposure. This could be especially attractive if deterioration in this counterparty’s credit outlook causes particular concern to the firm’s credit risk managers. Reduction in exposure could come in several different forms: a one‐time posting of margin or renegotiating the terms of existing contracts to provide for tighter terms on posting of margin. Of course, posting margin or tightening margin requirements is costly to the counterparty, so some concession must be offered as inducement—probably as a renegotiation of the financial terms of the derivative contracts to make strikes or spreads more favorable to the counterparty. The CRG would need to compensate the relevant trading desk for any such pricing concessions and must judge whether this up‐front cost is worth the reduction in credit risk. Another strategy that a CRG could pursue in reducing exposure to a counterparty is to offer the counterparty a mutual reduction in exposure— reducing the counterparty’s credit exposure to the firm by changing the financial terms on some derivative contracts on which the firm owes money to the counterparty in exchange for reducing the firm’s credit exposure to the counterparty by changing the financial terms on some derivative contracts on which the counterparty owes money to the firm. These changes in financial terms can be done in such a way as to leave the net amount owed by one party to the other unchanged, but with lower gross amounts owed. While netting and closeout master agreements accomplish much the same thing, actual reduction in gross amounts owed reduces the amounts that will be in contention in litigation that follows a default, and thus offers positive benefits. A greater impact on exposures could be achieved by moving beyond bilateral negotiations for changed financial terms to multilateral negotiations in which a counterparty’s exposure to one firm is reduced in exchange for a reduction in another firm’s exposure to the counterparty. This results in actual reduction in credit exposure, not just the reduction of litigation risk of the bilateral negotiation of changed financial terms discussed previously. Here’s a simple illustration. Suppose Bank A currently is owed $50 million on an interest rate swap by Counterparty B, and Bank C currently owes Counterparty B $50 million on an FX forward. If Counterparty B is willing to renegotiate the financial terms on these two contracts, it would not have to make any payments, since the $50 million it would owe to Bank A for the renegotiation would be offset by the $50 million it is owed by Bank C. Bank C would owe a $50 million payment to Bank A, but Bank A would offer Bank C some discount on this as an inducement to lowering Bank A’s credit exposure to Counterparty B and to compensate Bank C for losing the 530 FINANCIAL RISK MANAGEMENT cushion it had against having a credit exposure to B. In summation, Bank A benefits from reduced credit exposure but may have to pay something for it, Counterparty B is not impacted and in fact might gain slightly by reduced credit exposure to Bank C (though it may ask for some payment from Bank A for its cooperation), and Bank C will benefit to the extent it receives a payment from Bank A. Other creditors of Counterparty B are potentially disadvantaged, since in a default they would no longer have a claim on the amount owed to B by Bank A, but they have no standing in the transaction as long as B is a going concern. Variants of this last transaction have been introduced as a way for derivatives market makers to lower credit usage on derivatives transactions between market makers, and thereby free up credit lines. For example, several market makers get together and engage in trade compression, in which the market makers identify a set of derivative transactions that can be canceled and replaced by another set of derivative transactions, leaving market exposures close to unchanged but with a significant decrease in credit exposures. In addition to canceling trades that offset one another in market exposure, slight differences in contract detail that have little impact on market exposure can be eliminated to increase possibilities for contract cancellation. Some vendors now offer analytical services for developing proposed replacements that optimize the reduction in credit exposure that can be achieved by trade compression. Vause (2010) has a thorough discussion of trade compression and similar counterparty credit reduction techniques with examples. ISDA (2012) provides a detailed exposition of compression in the important case of interest rate swaps and illustrates the trade‐off between a firm’s tolerance for small changes in interest rate exposure and the degree of compression that can be accomplished. Generally speaking, having a derivatives position with a counterparty that is marked to market in your favor gives rise to credit exposure, but there is no offsetting credit benefit from having derivatives positions with a counterparty that is marked to market against you. Many CRGs have been searching for ways to achieve a more symmetrical position. We have just seen (in the next‐to‐previous paragraph) an example in which a firm can benefit from the credit consequences of a mark‐to‐market against it, since Bank C would be paid by Bank A to use its negative exposure to offset A’s positive exposure to Counterparty B. But this captures only part of the value of the exposure. A strategy that has been proposed for capturing the full value of the exposure is to purchase a bond of the counterparty that you net owe money to on derivative contracts with a maturity close to that of your derivative positions. Let’s consider an example to see how this might work. Let’s say you net owe $100 million in derivatives marked to market to a counterparty in a weak credit condition. Say you can purchase $100 million Counterparty Credit Risk 531 face value of its bonds for $90 million owing to its poor credit outlook. If the counterparty does not default, then you gain $10 million from the bond that you purchased at $90 million maturing at $100 million. If it does default, you can use the bond you own as an offset in bankruptcy proceedings to the $900 you owe the counterparty on the derivatives. So you have been able to use the amount you owe on your derivatives contracts to purchase free default protection on the bonds. The CRG would, of course, need to dynamically manage the amount of bond it holds to match changes in the derivatives market exposure in just the same way it dynamically manages the amount of CDS protection it buys when it is owed money on the derivatives position. The risk of this strategy is that a bankruptcy court could possibly object to offsetting the derivatives position and the bond holding. Finally, one option for a CRG would be to just purchase complete protection against the counterparty credit risk on a particular derivatives trade through a contingent credit default swap (CCDS). This is a CDS that in the event of default pays the amount that has been lost on the referenced derivatives trade. So, in effect, the CRG is turning the management of credit risk on this trade over to another firm. The firm selling the CCDS will have all of the issues of managing risk on this trade that we have discussed throughout this chapter and will need to be paid accordingly. There are many negatives arguing against the use of a CCDS, such as the mismatch between the amount of protection purchased and the amount of protection actually needed, since buying protection on a single transaction cannot take reduction in exposure through netting and margining into account. The CCDS is therefore probably a solution for only very large single transactions that are unlikely to have much offset against them. A thorough discussion of CCDSs can be found in Gregory (2010, Section 9.8.2). References A copy of this references list will be maintained on the website for this book, with ongoing updates to the Internet sites referenced. Acharya, Viral, Matthew Richardson, Stijn van Nieuwerburgh, and Lawrence White. 2011. Guaranteed to Fail. Princeton, NJ: Princeton University Press. Agrawal, Deepak, Navneet Arora, and Jeffrey Bohn. 2004. “Parsimony in Practice, and EDF‐Based Model of Credit Spreads.” Moody’s KMV. www.moodysanalytics .com/~/media/Insight/Quantitative‐Research/Credit‐Valuation/04‐29‐04‐Parsimony‐in‐Practice.ashx. Allen, Peter, Stephen Einchcomb, and Nicholas Granger. 2006. “Variance Swaps.” J.P. Morgan Securities European Equity Derivatives Research Investment Strategies no. 28. Allen, Steven, and Otello Padovani. 2002. “Risk Management Using Quasi‐Static Hedging.” Economic Notes 31 (2): 277–336. Almgren, Robert, and Neil Chriss. 2001. “Optimal Execution of Portfolio Transactions.” Journal of Risk 3 (2): 5–39. Altman, Edward, Neil Fargher, and Egon Kalotay. 2010. “A Simple Empirical Model of Equity‐Implied Probabilities of Default.” http://pages.stern.nyu .edu/~ealtman/DefaultModel2010JFI.pdf. Altman, Edward, and Egon Kalotay. 2010. “A Flexible Approach to Modeling Ultimate Recoveries on Defaulted Loans and Bonds.” http://pages.stern.nyu .edu/~ealtman/FlexibleRecovery_v1.1.pdf. Altman, Edward, Andrea Resti, and Andrea Sironi. 2001. “Analyzing and Explaining Default Recovery Rates.” www.defaultrisk.com/pp_recov_28.htm. Amato, Jeffery, and Eli Remolona. 2003. “The Credit Spread Puzzle.” BIS Quarterly Review (December). www.bis.org/publ/qtrpdf/r_qt0312e.pdf. Andersen, Leif, and Jesper Andreasen. 2000. “Static Barriers.” Risk 9:120–122. Andersen, Leif, and Jesper Andreasen. 2001. “Factor Dependence of Bermudan Swaptions: Fact or Fiction?” Journal of Financial Economics 62 (1): 3–37. Anderson, Leif, Jakob Sidenius, and Susanta Basu. 2003. “All Your Hedges in One Basket.” Risk 11:67–72. www.risk.net/data/Pay_per_view/risk/technical/ 2003/1103_credit_der.pdf. Araten, Michel, and Michael Jacobs. 2001. “Loan Equivalents for Revolving Credits and Advised Lines.” RMA Journal 5:34–39. http://michaeljacobsjr.com/Jacobs_ Araten_RMA_LEQ_Pub_April_2001.pdf. 533 534 REFERENCES Artzner, Philippe, Freddy Delbaen, Jean‐Marc Eber, and David Heath. 1997. “Thinking Coherently.” Risk 11:68–71. A longer version is available at www.math .ethz.ch/~delbaen/ftp/preprints/CoherentMF.pdf. Ashcroft, Adam, and Til Schuermann. 2008. “Understanding the Securitization of Subprime Mortgage Credit.” Federal Reserve Bank of New York Staff Reports no. 318. www.newyorkfed.org/research/staff_reports/sr318.pdf. Asiaweek. 1996. “Perils of Profit.” July 5. www‐cgi.cnn.com/ASIANOW/ asiaweek/96/0705/ed2.html. Bahar, Reza, and Krishan Nagpal. 2000. “Modeling the Dynamics of Rating Transition.” Credit 3:57–63. Bai, Jennie, and Pierre Collin‐Dufresne. 2011. “The CDS‐Bond Basis during the Financial Crisis of 2007–2008.” http://papers.ssrn.com/sol3/papers.cfm?abstract_ id=1785756ttp://papers.ssrn.com/sol3/papers.cfm?abstract_id=1785756. Basel Committee on Banking Supervision. 2009a. “Principles for Sound Stress Testing Practices and Supervision.” www.bis.org/publ/bcbs155.pdf. Basel Committee on Banking Supervision. 2009b. “Supervisory Guidance for Assessing Banks’ Financial Instrument Fair Value Practices.” www.bis.org/publ/ bcbs153.pdf. Baxter, Martin, and Andrew Rennie. 1996. Financial Calculus. Cambridge: Cambridge University Press. Bennett, Oliver. 2001. “Splitting Headaches.” Risk 7:36–37. Bluhm, Christian, Ludger Overbeck, and Christoph Wagner. 2002. An Introduction to Credit Risk Modeling. Boca Raton, FL: CRC Press. Bodie, Zvi, Alex Kane, and Alan Marcus. 2009. Investments. 8th ed. New York: McGraw‐Hill. Bohn, Jeffrey, and Roger Stein. 2009. Active Credit Portfolio Management in Practice. Hoboken, NJ: John Wiley & Sons. Bookstaber, Richard. 2007. A Demon of Our Own Design. Hoboken, NJ: John Wiley & Sons. Bouchet, Michel, Ephraim Clark, and Bertrand Groslambert. 2003. Country Risk Assessment. Chichester, England: John Wiley & Sons. Brealey, Richard, Stewart Myers, and Franklin Allen. 2011. Principles of Corporate Finance. 10th ed. New York: McGraw‐Hill. Breuer, Thomas, and Imre Csiszar. 2010. “Stress Tests: From Arts to Science.” https:// homepages.fhv.at/tb/cms/?Working_Papers. Breuer, Thomas, and Grerald Krenn. 2000. “Identifying Stress Test Scenarios.” http:// citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.27.118. Brindle, Andy. 2000. “Calculating with Counterparties.” Risk 1:100–103. www .risk.net/data/Pay_per_view/risk/technical/2000/risk_0100_jpmorgan.pdf. Broadie, Mark, Paul Glasserman, and Gautam Jain. 1997. “Enhanced Monte Carlo Estimates of American Option Prices.” Journal of Derivatives 5 (1): 25–44. A closely related article can be found at www.axelkind.com/teaching/financial_ derivatives/papers/broadie_glasserman_1997_jedc.pdf. Broom, Giles. 2011. “UBS in ‘Disarray.’” Bloomberg News. September 26. www .bloomberg.com/news/2011‐09‐25/ubs‐in‐disarray‐as‐gruebel‐quits‐ermotti‐ named‐interim‐chief.html. References 535 Brown, Aaron. 2012. Red‐Blooded Risk. Hoboken, NJ: John Wiley & Sons. Bruck, Connie. 1988. The Predators’ Ball. New York: Penguin. Brunnermeier, Markus. 2009. “Deciphering the Liquidity and Credit Crunch 2007–2009.” Journal of Economic Perspectives 1:77–100. www.princeton .edu/~markus/research/papers/liquidity_credit_crunch.pdf. Bucay, Nisso, and Dan Rosen. 2001. “Applying Portfolio Credit Risk Models to Retail Portfolios.” Journal of Risk Finance 3:35–61. www.financerisks.com/ filedati/WP/ALGO_PAPER/ch5_retail.pdf. Burghardt, Galen, and Gerald Hanweck. 1993. “Calendar‐Adjusted Volatilities.” In Volatility: New Estimation Techniques for Pricing Derivatives, edited by Robert Jarrow. London: Risk Books. Burghardt, Galen, and Susan Kirshner. 1994. “One Good Turn.” Risk 11:44–54. Canabarro, Eduardo, and Darrell Duffie. 2003. “Measuring and Marking Counterparty Risk.” In Asset/Liability Management of Financial Institutions, edited by Leo Tilman, Chapter 9. London: Euromoney Books. www.darrellduffie.com/ uploads/surveys/DuffieCanabarro2004.pdf. Carr, Peter. 2001. “Closed Form Barrier Option Valuation with Smiles.” www.math .nyu.edu/research/carrp/papers/pdf/closed4.pdf. Carr, Peter, and Andrew Chou. 1996. “Breaking Barriers.” Risk 9:139–145. www .math.nyu.edu/research/carrp/papers/pdf/breakingbarriers.pdf. Carr, Peter, and Andrew Chou . 1997. “Hedging Complex Barrier Options.” www .math.nyu.edu/research/carrp/papers/pdf/multipl3.pdf. Carr, Peter, Katrina Ellis, and Vishal Gupta. 1998. “Static Hedging of Exotic Options.” Journal of Finance 53 (3): 1165–1190. Carr, Peter, and Dilip Madan. 2002. “Towards a Theory of Volatility Trading.” www .math.nyu.edu/research/carrp/papers/pdf/twrdsfig.pdf. Cass, Dwight. 2000. “The Devil’s in the Details.” Risk 8:26–28. Chew, Lillian. 1996. Managing Derivatives Risks. New York: John Wiley & Sons. Chriss, Neil. 1997. Black‐Scholes and Beyond. Chicago: Irwin. Clementi, Gian Luca, Thomas Cooley, Matthew Richardson, and Ingo Walter. 2009. “Rethinking Compensation in Financial Firms.” In Restoring Financial Stability, edited by Viral Acharya and Matthew Richardson. Hoboken, NJ: John Wiley & Sons. Clewlow, Les, and Chris Strickland. 1998. Implementing Derivatives Models. New York: John Wiley & Sons. Cochrane, John. 2001. Asset Pricing. Princeton, NJ: Princeton University Press. Cordell, Larry, Yilin Huang, and Meredith Williams. 2012. “Collateral Damage: Sizing and Assessing the Subprime CDO Crisis.” Working Paper No. 11‐30/R, Research Department, Federal Reserve Bank of Philadelphia. www .philadelphiafed.org/research‐and‐data/publications/working‐papers/2011/ wp11‐30.pdf. Counterparty Risk Management Policy Group. 1999. “Improving Counterparty Risk Management Practices.” www.defaultrisk.com/pp_other_08.htm. Coval, Joshua, Jakub Jurek, and Erik Stafford. 2008. “The Economics of Structured Finance.” Harvard Business School Working Paper no. 09‐060. www.hbs.edu/ research/pdf/09‐060.pdf. 536 REFERENCES CreditGrades. 2002. “CreditGrades Technical Document.” RiskMetrics Group. www.msci.com/resources/research/technical_documentation/CGtechdoc.pdf. Crosbie, Peter, and Jeff Bohn. 2003. “Modeling Default Risk.” Moody’s KMV. www .moodysanalytics.com/~/media/Insight/Quantitative‐Research/Default‐and‐ Recovery/03‐18‐12‐Modeling‐Default‐Risk.ashx. Crouhy, Michel, Dan Galai, and Robert Mark. 2001. Risk Management. New York: McGraw‐Hill. Culp, Christopher. 2001. The Risk Management Process. New York: John Wiley & Sons. Culp, Christopher, and Merton Miller. 1995a. “Hedging in the Theory of Corporate Finance: A Reply to Our Critics.” Journal of Applied Corporate Finance (Spring): 121–127. Culp, Christopher, and Merton Miller. 1995b. “Metallgesellschaft and the Economics of Synthetic Storage.” Journal of Applied Corporate Finance (Winter): 62–75. www.rmcsinc.com/articles/JACF74.pdf. Dash, Eric, and Julie Creswell. 2008. “Citigroup Saw No Red Flags Even as It Made Bolder Bets.” New York Times, November 22. Davidson, Andrew. 2007. “Six Degrees of Separation.” www.securitization.net/pdf/ content/ADC_SixDegrees_1Aug07.pdf. Davidson, Andrew, Anthony Sanders, Lan‐Ling Wolfe, and Anne Ching. 2003. Securitization. Hoboken, NJ: John Wiley & Sons. Davies, Rob. 2008. “Genius or Blunder?” Risk. 3:22–27. www.risk.net/risk‐magazine/ feature/1498128/genius‐blunder. Dembo, Ron. 1994. “Hedging in Markets That Gap.” In Handbook of Derivatives and Synthetics, edited by Robert A. Klein and Jess Lederman. Chicago: Probus. Demeterfli, Kresimir, Emanuel Derman, Michael Kamal, and Joseph Zou. 1999. “Guide to Variance Swaps.” Risk 6:54–59. A longer version is available at www .ederman.com/new/docs/gs‐volatility_swaps.pdf. Derivative Strategies. 1998. “What Really Happened at UBS?” October. www .derivativesstrategy.com/magazine/archive/1998/1098fea1.asp. Derman, Emanuel. 1999. “Regimes of Volatility.” Risk 4:55–59. A longer version is available at www.ederman.com/new/docs/risk‐regimes_of_volatility.pdf. Derman, Emanuel. 2001. “Markets and Models.” Risk 7:48–50. My references are primarily to the longer version available at www.ederman.com/new/docs/ risk‐price_verification.pdf. Derman, Emanuel, Deniz Ergener, and Iraj Kani. 1995. “Static Options Replication.” Journal of Derivatives 2 (4): 78–95. Derman, Emanuel, and Iraj Kani. 1994. “Riding on a Smile.” Risk 2:32–39. Derman, Emanuel, and Iraj Kani . 1998. “Stochastic Implied Trees.” International Journal of Theoretical and Applied Finance 1 (1): 61–110. A longer version is available at www.ederman.com/new/docs/gs‐options_replication.pdf. Derman, Emanuel, and Joseph Zou. 2001. “A Fair Value for the Skew.” Risk 1:111– 114. de Servigny, Arnaud, V. Peretyatkin, W. Perraudin, and Olivier Renault. 2003. “Portfolio Risk Tracker: Description of the Methodology.” Standard & Poor’s Risk Solutions. References 537 de Servigny, Arnaud, and Olivier Renault. 2004. Measuring and Managing Credit Risk. New York: McGraw‐Hill. DeWit, Jan. 2006. “Exploring the CDS‐Bond Basis.” National Bank of Belgium working paper. www.nbb.be/doc/ts/publications/wp/wp104En.pdf. Diebold, Francis, Til Schuermann, and John Stroughair. 2000. “Pitfalls and Opportunities in the Use of Extreme Value Theory in Risk Management.” In Extremes and Integrated Risk Management, edited by Paul Embrechts. London: Risk Books. www.ssc.upenn.edu/~fdiebold/papers/paper21/dss‐f.pdf. Dixit, Avinash, and Robert Pindyck. 1994. Investment under Uncertainty. Princeton, NJ: Princeton University Press. Dowd, Kevin. 2005. Measuring Market Risk. 2nd ed. Hoboken, NJ: John Wiley & Sons. Duc, Francois, and Yann Schorderet. 2008. Market Risk Management for Hedge Funds. Hoboken, NJ: John Wiley & Sons. Duffie, Darrell. 2011. How Big Banks Fail. Princeton, NJ: Princeton University Press. Duffie, Darrell, Andreas Eckner, Guillaume Horel, and Leandro Saita. 2009. “Frailty Correlated Defaults.” Journal of Finance. 5:2089–2123. www.darrellduffie .com/uploads/pubs/DuffieEcknerHorelSaita2009.pdf. Duffie, Darrell, Ada Li, and Theo Lubke. 2010. “Policy Perspectives on OTC Derivatives Market Structure.” Federal Reserve Bank of New York Staff Reports no. 424. www.newyorkfed.org/research/staff_reports/sr424.pdf. Duffie, Darrell, and Kenneth Singleton. 2003. Credit Risk. Princeton, NJ: Princeton University Press. Dunbar, Nicholas, and Elisa Martinuzzi. 2012. “Goldman Secret Greek Loan Shows Two Sinners as Client Unravels.” Bloomberg News, March 5. www .bloomberg.com/news/2012‐03‐06/goldman‐secret‐greece‐loan‐shows‐two‐ sinners‐as‐client‐unravels.html. Dwyer, Paula. 1996. “Sumitomo’s Descent into the Abyss.” Business Week, July 1. www.businessweek.com/1996/27/b348241.htm. Eichenwald, Kurt. 1995. Serpent on the Rock. New York: HarperBusiness. Embrechts, Paul. 2000. “Extreme Value Theory: Potential and Limitations as an Integrated Risk Management Tool.” Derivatives Use, Trading, and Regulation 6:449–456. www.math.ethz.ch/~baltes/ftp/evtpot.pdf. Fabozzi, Frank, Sergio Focardi, and Petter Kolm. 2006. Financial Modeling of the Equity Market. Hoboken, NJ: John Wiley & Sons. Fabozzi, Frank, Sergio Focardi, and Petter Kolm. 2010. Quantitative Equity Investing: Techniques and Strategies. Hoboken, NJ: John Wiley & Sons. Falloon, William. 1998. “The Devil in the Documentation.” Risk 11:32–35. Fay, Stephen. 1996. The Collapse of Barings. London: Richard Cohen Books. Federal Reserve Board. 2009. “The Supervisory Capital Assessment Program: Design and Implementation.” www.federalreserve.gov/bankinforeg/bcreg20090424a1 .pdf. Federal Reserve Board. 2011. “Supervisory Guidance on Model Risk Management.” www.federalreserve.gov/bankinforeg/srletters/sr1107a1.pdf. Financial Crisis Inquiry Report. 2011. New York: PublicAffairs. Downloadable at www.gpo.gov/fdsys/pkg/GPO‐FCIC/pdf/GPO‐FCIC.pdf. 538 REFERENCES Financial Stability Board. 2010. “Implementing OTC Derivatives Markets Reforms.” www.financialstabilityboard.org/publications/r_101025.pdf. Financial Stability Forum. 2008. “Enhancing Market and Institutional Resilience.” www.financialstabilityboard.org/publications/r_0804.pdf. Financial Stability Forum. 2009a. “Addressing Procyclicality in the Financial System.” www.financialstabilityboard.org/publications/r_0904a.pdf. Financial Stability Forum. 2009b. “Principles for Sound Compensation Practices.” www.financialstabilityboard.org/publications/r_0904b.pdf. Garfield, Andrew. 1998. “Peter Young Charged with Morgan Grenfell Fraud.” The Independent. October 20. www.independent.co.uk/news/business/peter‐young‐ charged‐with‐morgan‐grenfell‐fraud‐1179481.html. Gatheral, Jim. 2006. The Volatility Surface. Hoboken, NJ: John Wiley & Sons. Giesecke, Kay, and Alexander Shkolnik. 2011. “Importance Sampling for Event Timing Models.” www.stanford.edu/dept/MSandE/cgi‐bin/people/faculty/giesecke/ pdfs/is.pdf. Gillen, David, Yoolim Lee, and Bill Austin. 1999. “J.P. Morgan’s Korean Debacle.” Bloomberg Magazine 3:24–30. Glasserman, Paul. 2004. Monte Carlo Methods in Financial Engineering. New York: Springer‐Verlag. Glasserman, Paul, and Jingyi Li. 2005. “Importance Sampling for Portfolio Credit Risk.” Management Science 11:1643–1656. www2.gsb.columbia.edu/faculty/ pglasserman/Other/is_credit.pdf. Global Legal Group. 2011. “The International Comparative Legal Guide to Securitization 2011.” www.iclg.co.uk/index.php?area=4&kh_publications_id=194. Greenlaw, David, Jan Hatzius, Anil Kashyap, and Hyun Song Shin. 2008. “Leveraged Losses: Lessons from the Mortgage Market Meltdown.” U.S. Monetary Policy Forum Report no. 2. http://research.chicagobooth.edu/igm/events/docs/ MPFReport‐final.pdf. Gregory, Jon. 2010. Counterparty Credit Risk. Chichester, UK: John Wiley & Sons. Group of Thirty. 1993. “Global Derivatives Study Group: Practices and Principles.” Group of Thirty. 2009. “Financial Reform: A Framework for Financial Stability.” www .group30.org/images/PDF/Financial_Reform‐A_Framework_for_Financial_ Stability.pdf. Gumerlock, Robert. 1999. “The Future of Risk.” Euromoney 6:112–114. Gupta, Ajay. 1997. “On Neutral Ground.” Risk 7:37–41. Gupton, Greg, Christopher Finger, and Mickey Bhatia. 1997.“Creditmetrics-Technical Document.” www.defaultrisk.com/pp_model_20.htm. Gupton, Greg, and Roger Stein. 2005. “Losscalc V2: Dynamic Prediction of LGD.” Moody’s KMV. January. www.defaultrisk.com/_pdf6j4/LCv2_ DynamicPredictionOfLGD_fixed.pdf. Hammond, John, Ralph Keeney, and Howard Raiffa. 1999. Smart Choices. Boston, MA: Harvard Business School Press. Hansell, Saul. 1997. “Joseph Jett: A Scoundrel or a Scapegoat?” New York Times, April 6. Harris, Larry. 2003. Trading and Exchanges: Market Microstructure for Practitioners. New York: Oxford University Press. References 539 Hasanhodzic, Jasmina, and Andrew Lo. 2007. “Can Hedge Fund Returns Be Replicated? The Linear Case.” Journal of Investment Management 2:5–45. http:// jasminah.com/Papers/hasanhodzic_lo07b.pdf. Helwege, Jean, Samuel Maurer, Asani Sarkar, and Yuan Wang. 2009. “Credit Default Swap Auctions.” Federal Reserve Bank of New York Staff Report no. 372. www.newyorkfed.org/research/staff_reports/sr372.pdf. Henderson, Schuyler. 1998. “Credit Derivatives: Selected Documentation Issues.” Journal of International Banking and Financial Law 9. Heston, Steven. 1993. “A Closed‐Form Solution for Options with Stochastic Volatility.” Review of Financial Studies 6 (2): 327–343. www.javaquant.net/papers/ Heston‐original.pdf. Holland, Kelley, and Linda Himelstein. 1995. “The Bankers Trust Tapes.” Business Week, October 16. www.businessweek.com/1995/42/b34461.htm. Holton, Glyn. 2003. Value‐at‐Risk. San Diego, CA: Academic Press. Huertas, Thomas. 2011. Crisis: Cause, Containment and Cure. 2nd ed. Basingstoke, UK: Palgrave Macmillan. Hull, John. 2000. Options, Futures, and Other Derivatives. 4th ed. Upper Saddle River, NJ: Prentice Hall. Hull, John. 2002. Options, Futures, and Other Derivatives. 5th ed. Upper Saddle River, NJ: Prentice Hall. Hull, John. 2009. “The Credit Crunch of 2007.” www.rotman.utoronto.ca/~hull/ downloadablepublications/CreditCrunch.pdfqa Hull, John. 2012. Options, Futures, and Other Derivatives. 8th ed. Upper Saddle River, NJ: Prentice Hall. Hull, John, Mirela Predescu, and Alan White. 2004. “The Relationship between Credit Default Swap Spreads, Bond Yields, and Credit Rating Announcements.” Journal of Banking and Finance (November): 2789–2811. www.rotman .utoronto.ca/~hull/DownloadablePublications/HPWPaperonCDSSpreads.pdf. Hull, John, and Wulin Suo. 2001. “A Methodology for Assessing Model Risk and Its Application to the Implied Volatility Function Model.” www.rotman.utoronto .ca/~hull/DownloadablePublications/HullSuoPaper.pdf. Hull, John, and Alan White. 1994. “Numerical Procedures for Implementing Term Structure Models II: Two‐Factor Models.” Journal of Derivatives 2 (2): 37–48. Hull, John, and Alan White. 2004. “Valuation of a CDO and an nth to Default CDS without Monte Carlo Simulation.” Journal of Derivatives 2:8–23. www.rotman. utoronto.ca/~hull/DownloadablePublications/HullWhiteCDOPaper.pdf. Ineichen, Alexander. 2003. Absolute Returns. Hoboken, NJ: John Wiley & Sons. International Comparative Legal Guide to Securitization. 2011. London: Global Legal Group. International Monetary Fund. 2009. Global Financial Stability Report. April. www .imf.org/External/Pubs/FT/GFSR/2009/01/pdf/text.pdf. ISDA. 2012. “Interest Rate Swaps Compression: A Progress Report.” February 23. www2.isda.org/functional‐areas/research/studies/. Jackel, Peter. 2002. Monte Carlo Methods in Finance. Chichester, UK: John Wiley & Sons. 540 REFERENCES Jackwerth, Jens, and Mark Rubinstein. 1996. “Recovering Probability Distributions from Option Prices.” Journal of Finance 51 (5): 1611–1631. www.er.ethz.ch/ teaching/Risk_neutral_Proba_Nonparametric_1996.pdf. Jacobs, Michael, and Ahmet Karagozoglu. 2011. “Modeling Ultimate Loss Given Default on Corporate Debt.” Journal of Fixed Income 1:6–20. Jameson, Rob. 1998a. “Operational Risk: Getting the Measure of the Beast.” Risk 11:38–41. Jameson, Rob. 1998b. “Operational Risk: Playing the Name Game.” Risk 10:38–42. Jett, Joseph, with Sabra Chartrand. 1999. Black and White on Wall Street. New York: William Morrow. Jewson, Stephen, Anders Brix, and Christine Ziehmann. 2005. Weather Derivatives Valuation. Cambridge: Cambridge University Press. Jorion, Philippe. 2007. Value at Risk. 3rd ed. New York: McGraw‐Hill. Kagel, J. H., and Alvin Roth. 1995. Handbook of Experimental Economics. Princeton, NJ: Princeton University Press. Kealhofer, Stephen, and Jeffrey Bohn. 2001. “Portfolio Management of Default Risk.” Moody’s KMV. www.moodysanalytics.com/~/media/Insight/Quantitative‐ Research/Portfolio‐Modeling/93‐15‐11‐Portfolio‐Management‐of‐Default‐ Risk.ashx. Khuong‐Huu, Philippe. 1999. “Swaptions on a Smile.” Risk 8:107–111. www.risk .net/data/Pay_per_view/risk/technical/1999/risk_0899_jpmorgan.pdf. Kim, Jongwoo, and Christopher Finger. 2000. “A Stress Test to Incorporate Correlation Breakdown.” Journal of Risk. www.risk.net/digital_assets/4867/v2n3a1 .pdf. Kooi, Mari. 1996. “Analyzing Sumitomo.” http://riskinstitute.ch/134800.htm. Kotowitz, Y. 1989. “Moral Hazard.” In The New Palgrave: Allocation, Information, and Markets, edited by John Eatwell, Murray Milgate, and Peter Newman. New York: W.W. Norton. Koutoulas, James, and John Roe. 2012. “White Paper: Background, Impacts & Solutions to MF Global’s Demise.” www.financialsense.com/node/725. Laurent, Jean‐Paul, and Jon Gregory. 2003. “Basket Default Swaps, CDO’s and Factor Copulas.” http://laurent.jeanpaul.free.fr/basket_cdo_factor_copula_2003.pdf. Lee, Roger. 2001. “Local Volatilities under Stochastic Volatility.” International Journal of Theoretical and Applied Finance 4:45–89. Leeson, Nick, with Edward Whitley. 1996. Rogue Trader. Boston: Little, Brown. Leibowitz, Martin, and Alfred Weinberger. 1981. “The Uses of Contingent Immunization.” Journal of Portfolio Management 1:51–55. Leibowitz, Martin, and Alfred Weinberger. 1982. “Contingent Immunization–Part I: Risk Control Procedures.” Financial Analysts Journal 38 (6): 17–31. Leibowitz, Martin, and Alfred Weinberger. 1983. “Contingent Immunization–Part II: Problem Areas.” Financial Analysts Journal 39 (1): 35–50. Lewis, Michael. 2011. The Big Short. New York: W.W. Norton. Li, David. 2000. “On Default Correlation: A Copula Function Approach.” RiskMetrics Working Paper no. 99‐07. http://papers.ssrn.com/sol3/papers.cfm?abstract_ id=187289. References 541 Litterman, Robert. 1997a. “Hot Spots and Hedges, Part 1.” Risk 3:42–45. Litterman, Robert. 1997b. “Hot Spots and Hedges, Part 2.” Risk 5:38–42. http:// home.ku.edu.tr/~cdemiroglu/Teaching/Risk/hotspots.pdf. Litterman, Robert, and Jose Scheinkman. 1988. “Common Factors Affecting Bond Returns.” Journal of Fixed Income 1 (1): 54–61. Lo, Andrew. 2008. “Hedge Funds, Systemic Risk, and the Financial Crisis of 2007– 2008.” Testimony for the U.S. House of Representatives Committee on Oversight and Government Reform November 13 hearing on hedge funds. http:// papers.ssrn.com/sol3/papers.cfm?abstract_id=1301217. Lo, Andrew. 2012. “Reading about the Financial Crisis: A 21‐Book Review.” Journal of Economic Literature 1:151–178. www.argentumlux.org/documents/JEL_6. pdf. Lowenstein, Roger. 2000. When Genius Failed. New York: Random House. Lowenstein, Roger. 2004. Origins of the Crash. New York: Penguin. Lowenstein, Roger. 2010. The End of Wall Street. New York: Penguin. Ludwig, Eugene. 2002. “Report on Allied Irish Bank.” www.aibgroup.com/servlet/ ContentServer?pagename=AIB_Investor_Relations/AIB_Download/aib_d_ download&c=AIB_Download&cid=1015597173380&channel=IRHP. Madan, Dilip, Peter Carr, and Eric Chang. 1998. “The Variance Gamma Process and Option Pricing.” www.math.nyu.edu/research/carrp/papers/pdf/ VGEFRpub.pdf. Malcolm, Fraser, Pawan Sharma, and Joseph Tanega. 1999. Derivatives: Optimal Risk Control. London: Prentice Hall. Matytsin, Andrew. 1999. “Modeling Volatility and Volatility Derivatives.” www .math.columbia.edu/~smirnov/Matytsin.pdf. Mayer, Martin. 1995. “Joe Jett: Did the Computer Make Him Do It?” Institutional Investor 3:7–11. Mayer, Martin . 1997. The Bankers: The Next Generation. New York: Truman Talley Books. McDonald, Robert. 2006. Derivatives Markets. Boston: Pearson Education. McKay, Peter. 1999. “Merrill Lynch to Pay Fines of $25 Million for Scandal.” Dow Jones Newswires, July 1. McLean, Bethany, and Peter Elkind. 2003. The Smartest Guys in the Room. New York: Penguin. McLean, Bethany, and Joe Nocera. 2010. All the Devils Are Here. New York: Penguin. McNeil, Alexander. 2000. “Extreme Value Theory for Risk Managers.” In Extremes and Integrated Risk Management, edited by Paul Embrechts. London: Risk Books. Mello, Antonio, and John Parsons. 1995. “Maturity Structure of a Hedge Matters.” Journal of Applied Corporate Finance (Spring): 106–120. Meucci, Attilio, 2005. Risk and Asset Allocation. Berlin: Springer‐Verlag. Moody’s. 2010. “Municipal Bond Defaults and Recoveries, 1970–2009.” www.naic .org/documents/committees_e_capad_vos_c1_factor_review_sg_related_docs_ moodys_us_municipal_bonds.pdf. 542 REFERENCES Moody’s. 2011a. “Corporate Default and Recovery Rates, 1920–2010.” www.naic .org/documents/committees_e_capad_vos_c1_factor_review_sg_related_docs_ moodys_corporate_default.pdf. Moody’s. 2011b. “Sovereign Default and Recovery Rates, 1983–2010.” www.naic .org/documents/committees_e_capad_vos_c1_factor_review_sg_related_docs_ moodys_sovereign_default.pdf. Moosa, Imad. 2007. Operational Risk Management. New York: Palgrave Macmillan. Moosa, Imad. 2008. Quantification of Operational Risk under Basel II. New York: Palgrave Macmillan. Morini, Massimo. 2011. Understanding and Managing Model Risk. Chichester, UK: John Wiley & Sons. Neuberger, Anthony. 1996. “The Log Contract and Other Power Options.” In The Handbook of Exotic Options, edited by Israel Nelken. Chicago: Irwin. Norris, Floyd. 2009. “A Lack of Rigor Costs MBIA.” New York Times, November 12. O’Brien, Timothy L. 1999. “Taking the Danger out of Risk.” New York Times, January 28. O’Kane, Dominic. 2008. Modelling Single‐Name and Multi‐Name Credit Derivatives. Chichester, UK: John Wiley & Sons. O’Kane, Dominic, and Robert McAdie. 2001. “Trading the Default Swap Basis.” Risk 9, Lehman Brothers–sponsored article. www.scribd.com/ dtayebconsulting/d/26559374‐Explaining‐the‐Basis‐Cash‐Versus‐Default‐Swaps. Oyama, Tsuyoshi. 2010. Post‐Crisis Risk Management. Singapore: John Wiley & Sons. Perold, Andre. 1998. “Capital Allocation in Financial Firms.” Harvard Business School working paper. Perold, Andre. 1999a. Harvard Business School Case Study (9‐200‐007) of Long‐ Term Capital Management (A). Perold, Andre. 1999b. Harvard Business School Case Study (9‐200‐009) of Long‐ Term Capital Management (C). Pirrong, Craig. 2011. “The Economics of Central Clearing: Theory and Practice.” ISDA Discussion Paper Series no. 1. Available at www2.isda.org/functional‐ areas/research/discussion‐papers/. PricewaterhouseCoopers. 2011. A Practitioner’s Guide to Basel III and Beyond. London: Thomson Reuters. Rajan, Raghuram. 2010. Fault Lines. Princeton, NJ: Princeton University Press. Ramberg, John, Edward Dudewicz, Pandu Tadikamalla, and Edward Myktyka. 1979. “A Probability Distribution and Its Use in Fitting Data.” Technometrics 2:201–214. Rawnsley, Judith. 1995. Total Risk. New York: HarperCollins. Rebonato, Riccardo. 1998. Interest‐Rate Options Models. 2nd ed. New York: John Wiley & Sons. Rebonato, Riccardo. 2002. Modern Pricing of Interest Rate Derivatives. Princeton, NJ: Princeton University Press. Rebonato, Riccardo. 2003. “Theory and Practice of Model Risk Management.” In Modern Risk Management: A History, 223–248. London: Risk Books. Available on the web through CiteSeerX. References 543 Rebonato, Riccardo. 2004. Volatility and Correlation. 2nd ed. Hoboken, NJ: John Wiley & Sons. Rebonato, Riccardo. 2007. Plight of the Fortune Tellers. Princeton, NJ: Princeton University Press. Reiner, Eric. 1992. “Quanto Mechanics.” Risk 3:59–63. Richardson, Matthew, and Lawrence White. 2009. “The Ratings Agencies.” In Restoring Financial Stability, edited by Viral Acharya and Matthew Richardson. Hoboken, NJ: John Wiley & Sons. Risk Management Association. 2000. Credit Risk Capital for Retail Products. www .johnmingo.com/pdfdocs/RMARetailCreditCapit614.pdf RiskMetrics Group. 1996. RiskMetrics Technical Document. 4th ed. www.msci .com/resources/research/technical_documentation/td4e.pdf. Roubini, Nouriel, and Stephen Mihm. 2010. Crisis Economics. New York: Penguin. Rullière, Didier, Diana Dorobantu, and Areski Cousin. 2010. “An Extension of Davis and Lo’s Contagion Model.” http://arxiv.org/pdf/0904.1653v2.pdf. Salmon, Felix. 2009. “Recipe for Disaster: The Formula That Killed Wall Street.” Wired, February 23. www.wired.com/techbiz/it/magazine/17‐03/wp_ quant?currentPage=all. Saunders, Anthony, and Linda Allen. 2010. Credit Risk Measurement. 3rd ed. Hoboken, NJ: John Wiley & Sons. Saunders, Anthony, Roy Smith, and Ingo Walter. 2009. “Enhanced Regulation of Large, Complex Financial Institutions.” In Restoring Financial Stability, edited by Viral Acharya and Matthew Richardson. Hoboken, NJ: John Wiley & Sons. Schachter, Barry. 2001. “How Well Can Stress Tests Complement VaR?” www .gloriamundi.org/UploadFile/2010‐2/stressandevt1.pdf. Schonbucher, Philipp. 2003. Credit Derivatives Pricing Models. Chichester, UK: John Wiley & Sons. Schutz, Dirk. 2000. The Fall of UBS. New York: Pyramid Media Group. Senior Supervisors Group. 2008. “Observations on the Risk Management Practices during the Recent Market Turbulence.” March 6. www.newyorkfed.org/ newsevents/news/banking/2008/SSG_Risk_Mgt_doc_final.pdf. Sharpe, William. 1992. “Asset Allocation: Management Style and Performance Measurement.” Journal of Portfolio Management 2:7–19. www.stanford .edu/~wfsharpe/art/sa/sa.htm. Shaw, Julian. 1997. “Beyond VaR and Stress Testing.” In VaR –Understanding and Applying Value‐at‐Risk, 211–224. London: Risk Publications. Shiller, Robert. 2003. The New Financial Order. Princeton, NJ: Princeton University Press. Shiller, Robert. 2008. The Subprime Solution. Princeton, NJ: Princeton University Press. Shirreff, David. 1998. “Another Fine Mess at UBS.” Euromoney 11:41–43. Shirreff, David. 2000. “Lessons from the Collapse of Hedge Fund, Long‐Term Capital Management.” http://elsa.berkeley.edu/users/webfac/craine/e137_ f03/137lessons.pdf. Sifakis, Carl. 1982. Encyclopedia of American Crime. New York: Facts on File. Smithson, Charles. 2000. “What’s New in the Options Markets?” Risk 5:54–55. 544 REFERENCES Smithson, Charles, and Neal Pearson. 2008. “Valuing CDOs of ABSs.” Risk 3:84–87. www.rutterassociates.com/pdf/Smithson&Pearson_ValuingTranchesOfCDOsOnABS_RISK_Mar2008.pdf. Société Générale Mission Green Report. 2008. www.sp.socgen.com/sdp/sdp.nsf/V3 ID/8D7F118212725EC6C1257452005AA90E/$file/report%20part%203.pdf. Société Générale Special Committee of the Board of Directors Report to Shareholders. 2008. www.sp.socgen.com/sdp/sdp.nsf/V3ID/C54C92B634428055C12574 52005A7F61/$file/report%20part%201.pdf. Sorkin, Andrew Ross. 2010. Too Big to Fail. New York: Penguin. Squam Lake Group. 2010. Squam Lake Report: Fixing the Financial System. Princeton, NJ: Princeton University Press. Standard & Poor’s. 1996. “April 1 Credit Week.” Stiglitz, Joseph. 2010. Freefall. New York: W.W. Norton. Stigum, Marcia. 1989. The Repo and Reverse Markets. Homewood, IL: Dow Jones– Irwin. Swensen, David. 2000. Pioneering Portfolio Management. New York: Free Press. Taleb, Nassim. 1997. Dynamic Hedging. New York: John Wiley & Sons. Taleb, Nassim. 2010. The Black Swan. New York: Random House. Tett, Gillian. 2009. Fool’s Gold. New York: Free Press. Thaler, Richard. 1992. The Winner’s Curse. New York: Free Press. Tsiveriotis, Kostas, and Chris Fernandes. 1998. “Valuing Convertible Bonds with Credit Risk.” Journal of Fixed Income 8:95–102. Tuckman, Bruce, and Angel Serrat. 2012. Fixed Income Securities. 3rd ed. Hoboken, NJ: John Wiley & Sons. Turner Review. 2009. “A Regulatory Response to the Global Banking Crisis.” Financial Services Authority. www.fsa.gov.uk/pubs/other/turner_review.pdf. UBS. 2008. “Shareholder Report on UBS’s Write‐Downs.” Available at https://www .ubs.com/global/en/about_ubs/investor_relations/share_information/ shareholderreport.html. Varian, Hal. 1987. “The Arbitrage Principle in Financial Economics.” Journal of Economic Perspectives 1 (2): 55–72. http://ideas.repec.org/a/aea/jecper/ v1y1987i2p55‐72.html. Vause, Nicholas. 2010. “Counterparty Risk and Contract Volumes in the Credit Default Swap Market.” BIS Quarterly Review (December): 59–69. http://finance .eller.arizona.edu/lam/fixi/creditmod/Vause.pdf. Wang, Jin. 2001. “Generating Daily Changes in Market Variables Using a Multivariate Mixture of Normal Distributions.” Proceedings of the 33rd Winter Simulation Conference. Washington, DC: IEEE Computer Society. Weiss, Gary. 1994. “What Lynch Left Out.” BusinessWeek, August 22. www .businessweek.com/stories/1994‐08‐21/what‐lynch‐left‐out. Whaley, Elizabeth, and Paul Wilmott. 1994. “Hedge with an Edge.” Risk 10:82–85. Williams, Jeffrey. 1986. The Economic Function of Futures Markets. Cambridge: Cambridge University Press. Wilson, Charles. 1989. “Adverse Selection.” In The New Palgrave: Allocation, Information, and Markets, edited by John Eatwell, Murray Milgate, and Peter Newman. New York: W.W. Norton. References 545 Wilson, Harry. 2011.“UBS Rogue Trader: Investigations Focus on Fictitious Hedges.” The Telegraph, September 16. www.telegraph.co.uk/finance/financial‐crime/8772540/ Fictitious‐hedges‐see‐UBS‐rogue‐trader‐losses‐climb‐to‐2.3bn.html. Wilson, Thomas. 1997. “Portfolio Credit Risk.” Risk 9:111–117 and 10:56–61. Shorter version available at http://papers.ssrn.com/sol3/papers.cfm?abstract_ id=1028756. Wilson, Thomas. 1998. “Value at Risk.” In Risk Management and Analysis, Vol. 1, edited by Carol Alexander, 61–124. New York: John Wiley & Sons. Winters, Bill. 1999. “Wrong‐Way Exposure.” Risk 7:52–55. www.risk.net/data/Pay_ per_view/risk/technical/1999/risk_0799_jpmorgan.pdf. Wolfe, Eric. 2001. “NatWest Markets: E‐Risk.” www.riskmania.com/pdsdata/ NatWestCaseStudy‐erisk.pdf. Zandi, Mark. 2009. Financial Shock. Updated ed. Upper Saddle River, NJ: FT Press. About the Companion Website his book has an associated website (www.wiley.com/go/frm2e) containing Microsoft Excel spreadsheets that can be used to experiment with many of the concepts covered in the text. Most of the book’s exercises are built around these calculators. Full documentation of the spreadsheets is contained in an accompanying Word document on the website. This appendix briefly describes the spreadsheets that are available. They are listed in the order you will encounter them in the text. I have chosen to build all of these calculators in Excel with minimal use of user‐defined functions for two reasons: T 1. By using Excel rather than a programming language, I am hoping to maximize the number of readers who will be able to follow the calculations. 2. By minimizing user‐defined functions, I am making the machinery of the computations as visible as possible. These calculators have all been built specifically to illustrate the material of this book (and the course I teach on which the book is based). They are not designed to be used to actually manage risk positions. Specifically, they don’t include the sort of detail, such as day count conventions, that is important in a trading environment. This sort of detail can be distracting when trying to learn broad concepts. For similar reasons, I have often chosen simple alternatives over more complex ones to illustrate a point. For example, I have chosen to represent volatility smile and skew through a simple formula that favors the ease of seeing the approximate impact of changes in input variables over the accuracy of a more complex representation. Using such calculators for actual trading would require programs that are easily scalable; that is, they can readily accommodate adding a larger number of positions. I have deliberately sacrificed scalability for the ease of handling a small number of positions. Scalability nearly always requires the use of a programming language as opposed to a primarily spreadsheet‐based approach. For readers who want to pursue building more robust calculators, and for teachers who want to assign exercises involving the building of scalable versions of some of these calculators, these spreadsheets should be able 547 548 ABOUT THE COMPANION WEBSITE to serve as good sources for parallel tests of computations, particularly since Excel gives an immediate display of all the numerical results of the intermediate stages of the calculations. The spreadsheets, in the order of the corresponding material in the text, are as follows. ■ ■ ■ ■ ■ ■ ■ ■ ■ ■ The MixtureOfNormals spreadsheet produces series of random variables displaying fat tails and clustering of large moves by mixing together two normally distributed series. It is utilized for exercises in Sections 1.3 and 7.1.1. The WinnersCurse spreadsheet illustrates the mechanism of the winner’s curse in auction situations, as explained in Section 2.4. The VaR spreadsheet computes VaR using three different methods– historical simulation, Monte Carlo simulation, and variance covariance. It enables the user to compare results obtained through the three methods and explore possible modifications. This is discussed in Section 7.1 and is used in Exercises 7.1 and 7.3. The EVT spreadsheet uses the extreme value theory formulas from the box “Key Results from EVT” in Chapter 7 to calculate VaR and shortfall VaR for selected percentiles. The Rates spreadsheet can be used either to value and compute risk statistics for a portfolio of linear instruments (such as forwards, swaps, and bonds) based on an input set of forward rates or to determine a set of forward rates that achieve an optimum fit with a given set of prices for a portfolio of linear instruments while maximizing the smoothness of the forward rates selected. This is discussed in Sections 10.2.1 and 10.4. The Bootstrap spreadsheet produces a comparison between the bootstrap and optimal fitting methodologies for extracting forward rates from an observed set of swap rates. This spreadsheet was used to produce Figure 10.1 in Section 10.2.1. The RateData spreadsheet contains a historical time series of U.S. interest rate data. It is used in Exercises 10.1 and 10.2. The NastyPath spreadsheet is an illustration of the size of losses that can be incurred when dynamically delta hedging an option. The example follows the dynamic delta hedging of a purchased call option over the 30 days of its life. This is discussed in the example in Section 11.2. The PriceVolMatrix spreadsheet computes the price‐volatility matrix and volatility surface exposure for a small portfolio of vanilla European‐style options. It illustrates the material discussed in Section 11.4. The PriceVolMatrixCycle spreadsheet is a particular run of the PriceVolMatrix spreadsheet that has been used to produce Table 11.5. About the Companion Website ■ ■ ■ 549 The VolCurve spreadsheet fits a forward volatility curve to observed options prices. This spreadsheet is designed for European options other than interest rate caps and floors. This is discussed in Section 11.6.1. The CapFit spreadsheet fits a forward volatility curve to observed options prices for interest rate caps. Since caps are baskets of options, with each option within the basket termed a caplet, the spreadsheet needs to break each cap apart into its constituent caplets and price each one individually. This is discussed in Section 11.6.1. The VolSurfaceStrike spreadsheet interpolates implied option volatilities by strike for a given tenor, utilizing the methods discussed in Section 11.6.2. The interpolation can be performed in two modes: 1. Implied volatilities are input for enough strikes to allow for reasonable interpolation. 2. Implied volatilities are input for only three strikes. ■ ■ ■ ■ ■ ■ The OptionRoll spreadsheet is a variant of the PriceVolMatrix spreadsheet. It differs in the form of the optimization, which is set up to calculate a hedge that will minimize a future roll cost. It illustrates the material discussed in Section 11.6.3. The OptionMC spreadsheet calculates a single path of a Monte Carlo simulation of the delta hedging of a vanilla European‐style call option position. It is designed to help you check your work for the Monte Carlo simulation exercise in Chapter 11 (Exercise 11.2). The OptionMC1000 spreadsheet used in Exercise 11.2 is identical to the OptionMC spreadsheet except that it is set up for 1,000 time steps instead of 20 time steps. The OptionMCHedged spreadsheet used in Exercise 11.2 is a variant on the OptionMC spreadsheet. It calculates a single path of a Monte Carlo simulation of the delta hedging of the European‐style call option hedged by two other call options with the same terms but different strike prices. The OptionMCHedged1000 spreadsheet used in Exercise 11.2 is identical to the OptionMCHedged spreadsheet except that it is set up for 1,000 time steps instead of 20 time steps. The BasketHedge spreadsheet calculates and prices a piecewise‐linear hedge using forwards and plain‐vanilla European options for any exotic derivative whose payoffs are nonlinear functions of the price of a single underlying asset at one particular point in time. The spreadsheet consists of a Main worksheet that can be used for any payoff function and other worksheets that contain illustrations of how the Main worksheet can be used to hedge particular payoff functions. The particular functions 550 ■ ■ ■ ■ ■ ■ ■ ■ ■ ■ ■ ABOUT THE COMPANION WEBSITE illustrated are a single‐asset quanto, a log contract, interest rate convexity, and a compound option. This is discussed in Section 12.1. The BinaryMC spreadsheet provides a Monte Carlo simulation of binary options using the method discussed in Section 12.1.4. The ForwardStartOption spreadsheet is a slight variant on the PriceVolMatrix spreadsheet that can be used for the risk management of forward‐starting options using the method discussed in Section 12.2. The CarrBarrier spreadsheet compares the pricing of barrier options using Carr’s static hedging replication with those computed using standard analytic formulas. The cost of unwinding the static hedge is also calculated. This is discussed in Section 12.3.3. The CarrBarrierMC spreadsheet provides a Monte Carlo simulation of barrier options using Carr’s static hedging replication, as discussed in Section 12.3.3. The OptBarrier spreadsheet illustrates the use of optimization to find a hedge for a down‐and‐out call barrier option, as discussed in Section 12.3.3. The DermanErgenerKani spreadsheet used in Exercise 12.6 calculates the pricing of knock‐out barrier options using the Derman‐Ergener‐ Kani static hedging replication. The cost of unwinding the static hedge is also calculated. It illustrates the material discussed in Section 12.3.3. The DermanErgenerKani20 spreadsheet also calculates the pricing of knock‐out barrier options using the Derman‐Ergener‐Kani static hedging replication. It displays intermediate results more explicitly than the DermanErgenerKani spreadsheet, but is less flexible for expansion to a larger number of time steps. The DermanErgenerKaniDoubleBarrier spreadsheet calculates the pricing of double barrier knock‐out barrier options using the Derman‐ Ergener‐Kani static hedging replication. The cost of unwinding the static hedge is also calculated. This is discussed in Section 12.3.5. The DermanErgenerKaniPartialBarrier spreadsheet calculates the pricing of partial barrier knock‐out barrier options using the Derman‐ Ergener‐Kani static hedging replication. The cost of unwinding the static hedge is also calculated. This is discussed in Section 12.3.5. The BasketOption spreadsheet computes an approximate value for the volatility to be used to price an option on a basket of assets and also computes the sensitivity of this volatility to changes in the volatility of the underlying asset and in the correlation between assets. This is discussed in Section 12.4.1. The CrossHedge spreadsheet simulates the hedging of a quanto that pays the product of two asset prices. The hedge is simulated using two different assumptions: if the asset price moves are completely uncorrelated About the Companion Website ■ ■ ■ ■ ■ ■ ■ 551 and if the asset price moves are completely correlated. This is discussed in Section 12.4.5. The AmericanOption spreadsheet calculates risk statistics for the early exercise value of American call options, as discussed in Section 12.5.1. The TermStructure spreadsheet illustrates the difficulties involved in pricing products that are dependent on yield‐curve shape. It shows that different combinations of input parameters that result in the identical pricing of European caps/floors and swaptions can lead to very different pricings of these products. This is discussed in Section 12.5.2. The Swaptions spreadsheet calculates current swaption volatilities from current forward rate agreement (FRA) levels, forward FRA volatilities, and correlations between FRAs. Using the Solver, it can find forward FRA volatilities that will reproduce observed current swaption volatilities, as discussed in Section 12.5.3. The CreditPricer spreadsheet translates between par yields and default rates for risky bonds and also prices risky bonds based on the derived default rates, as discussed in Section 13.1. The MertonModel spreadsheet calculates default probabilities and the distance to default using the simplified model documented in Section 13.2.4. The JumpProcessCredit spreadsheet calculates default probabilities and credit spreads using the jump process model discussed in Section 13.2.4.1. The CDO spreadsheet calculates default probabilities for tranches of CDOs utilizing a Vasicek model with the large homogeneous portfolio (LHP) assumption, as discussed in Section 13.3.3. Index ACA Insurance, 107 Accounting arbitrage, 107 Accounting risk: defined, 30 as form of reputational risk, 42 Accrual swaps, 378 Acharya, Viral, 84 Ackerlof, George, 20 Active management approach, 526–531 Actuarial risk management: actuarial risk, defined, 2 comparison with financial risk management, 2–4 financial risk in estimating, 2 liquid proxies in, 520 moral hazard in, 13–14 for positions that are born illiquid, 143–144 Adjustable-rate mortgages (ARMs), 86–87. See also Financial crisis of 2007–2008 Adoboli, Kweku, 67 Adverse selection, 19–21 controlling, 22 defined, 20 information asymmetry and, 20–21 legal risk and, 37 Agrawal, Deepak, 484 AIG (American International Group), 86, 93, 97, 102, 107, 108, 114 Allen, Franklin, 141 Allen, Linda, 464, 481 Allen, Peter, 331 Allfirst First Maryland Bancorp. See Allied Irish Bank (AIB) case Allied Irish Bank (AIB) case, 31, 51, 57–59, 64, 65 detection of unauthorized financial positions, 58 development of unauthorized position, 57 failure to detect unauthorized positions, 57–58 further reading, 59 incident, 57 lessons to be learned, 59 result, 57 Almgren, Robert, 255–256 Alpha factor, 515 Altman, Edward, 464–465, 467, 479 Amato, Jeffery, 84, 465 American International Group (AIG), 86, 93, 97, 102, 107, 108, 114 American options: defined, 426 difficulty in valuing, 409 European options versus, 426 hedging, 435 intensity of use, 363 AmericanOption spreadsheet, 427–428, 551 Analysis of overrides, 233–234 Analysis of revenue, 156–157 Andersen, Leif, 404, 434, 490 Andreasen, Jesper, 404, 434 Andrews, Edmund, 87 Araten, Michel, 469 Arbitrage: accounting, 107 Allied Irish Bank (AIB) case, 31, 51, 57–59, 64, 65 arbitrage theory in decomposing risk, 142 553 554 Arbitrage (Continued) Barings Bank case, 31, 51, 55–57, 64, 66 cash-and-carry, 259–260, 300–301, 302 internal, 167 no-arbitrage principle, 239–241 Arbitrage pricing theory (APT), 141 Arbitrageurs, 26–27 Armitstead, Louise, 132 Arora, Navneet, 484 Arthur Andersen, 12–13, 79–80 Artzner, Philippe, 188 Ashcroft, Adam, 85, 90 Asian credit crisis of 1997, 187, 206 Asian options, 413 Askin, David, 66–67 Askin Capital Management, 66–67 Asset-backed securities, 278–279. See also Credit default swaps (CDS) Asset liquidity risk. See also Funding liquidity risk basis risks versus, 3–4, 255–256, 289 defined, 3 funding liquidity risk versus, 42 positions that achieve illiquidity, 143 positions that are born illiquid, 143–144 in risk measurement, 142–147 Asset-or-nothing options, 371 Asset swaps: credit default swaps (CDS) versus, 447, 448 in credit risk management, 447 AT&T, 263 Auctions, winner’s curse and, 22–24 Austin, Bill, 80 Background checks, 33, 87–88 Back office: defined, 10 fraud risk and, 32–35 Back-testing, in value at risk (VaR) analysis, 191, 233 Backwardation, 303 Bahar, Reza, 461 INDEX Bai, Jennie, 456 Bankers Trust (BT) case, 41, 77–79, 367 Bank for International Settlements (BIS), 101, 115, 122, 505–506 Bank of America, 274–275 Bank of England, 126 Bank of New York, 36 Bankruptcy: impact of bankruptcy law, 452–453 legal risk and, 39–40 need for more orderly bankruptcy proceedings, 131 skew pattern in equity markets and, 350 “too big to fail” mentality and, 11, 72, 105–106, 114, 121, 124 Banziger, Hugo, 62 Barings Bank case, 31, 51, 55–57, 64, 66 detection of unauthorized positions, 56 development of unauthorized positions, 55 failure to detect unauthorized positions, 55–56 further reading, 57 incident, 55 lessons to be learned, 56 result, 55 Barrier options, 381–404 barriers, defined, 382 Carr hedge, 382, 388, 391–401 defined, 382 Derman-Ergener-Kani hedge, 387–391, 403–404 double, 382 drift, 382–383 dynamic hedging models, 385–387 knock-in (down and in), 382 knock-out (down and out/up and out), 382 ladder options, 402–403 lookback options, 402 partial-time, 382 put-call symmetry, 391–392 with rebates, 402 standard analytic models, 383–384 Index static hedging models, 387–391 value of barrier based on analytic formula, 384 Base correlation, 498 Basel Committee on Banking Supervision, 29, 46–47, 130, 193, 196, 209, 219 Basis risk: CDS-bond, 454–456 defined, 3, 255–256 liquidity risk versus, 3–4, 255–256, 289 Basis swaps, 298–299 BasketHedge spreadsheet, 364, 368– 370, 381, 439, 549–550 BasketOption spreadsheet, 412–413, 441, 550 Basu, Susanta, 490 Baxter, Martin, 418 Bear Stearns, 95, 101, 106 Bennett, Oliver, 450 Bermudan options, 426 hedging, 435, 436 intensity of use, 363 Bermudan swaptions, 432, 433–434 Bet options. See Binary options Bhatia, Mickey, 461, 465, 466, 481 Bilateral counterparty risk, 521 Binary credit default swaps, 449–450 BinaryMC spreadsheet, 376, 440, 550 Binary options, 371–377 asset-or-nothing option, 371 cash-or-nothing option, 371 Binomial tree model, 425–426 Black-Derman-Toy model, 434, 435 Black-Karasinski model, 434, 435 Black-Scholes option pricing model, 98, 144, 157–158 for exotic options, 359–361, 383, 385, 387, 388, 405, 427 model risk and, 210–211, 220 for vanilla options, 311–324, 331, 335, 344–345, 349–350, 355–356 Black Swan, The (Taleb), 138 Bleed (Taleb), 343–344 Bluhm, Christian, 490 555 Bodie, Zvi, 141 Bohn, Jeffrey, 464–467, 475–479, 481–484, 492 Bonds: CDS-bond basis risk, 454–456 in credit risk management, 447 market for, 447 Book running. See Market makers/ market making Bookstaber, Richard, 95, 132 Bootstrapping, 287–288, 290 Bootstrap spreadsheet, 289, 548 Borrowing costs, 299–304 forward prices and, 303–304 nature of borrowing demand, 299–300 possibility of cash-and-carry arbitrage, 300–301 seasonality, 302 variability of storage costs, 301 Bouchet, Michel, 486 Brace-Gatarek-Museila (BGM) model, 425 Brealey, Richard, 141 Breeden-Litzenberger theorem, 364 Breuer, Thomas, 200, 486 Brindle, Andy, 517–519 British Bankers Association, 268–269, 297 Brix, Anders, 144 Broadie, Mark, 409 Broom, Giles, 67 Brown, Aaron, 27, 132, 265, 510 Bruck, Connie, 237 Brunnermeier, Markus, 85–88, 90, 111 Bucay, Nisso, 485 Burghardt, Galen, 302, 347 Burnout, 424 Business Week, 87 Cabiallavetta, Mathis, 61 Calendar spread, 336, 337 Call spreads, 331–334, 372 Canabarro, Eduardo, 517 Cancel-and-correct activity, 62–63, 65 CapFit spreadsheet, 347, 549 556 Capital asset pricing model (CAPM), 141–142, 484 Capital structure, leverage in, 477 Caps/caplets, 347, 430, 432 Carr, Peter, 245, 364, 382, 386, 388, 391, 394, 404 CarrBarrierMC spreadsheet, 396, 440, 550 CarrBarrier spreadsheet, 394–396, 440, 550 Carr hedge: advantages, 393–394 broader applications, 404 comparison with other static hedging models, 387–390 deriving, 393 development of, 382 key points, 391, 394–396 put-call symmetry and, 391–392 static hedge, 383–384, 397–401 Carty, Lea, 467 Cash-and-carry arbitrage, 259–260, 300–301, 302 Cash-or-nothing options, 371 Cash settle, 274–275 Cass, Dwight, 450 CDO spreadsheet, 496, 551 CDX index, 493 Central counterparty clearinghouse (CCP), 127 Chain letter frauds, 17 Chang, Eric, 386 Chase Manhattan Bank, 105, 123, 141, 202, 205–206, 249, 319, 469 Chase Manhattan Bank/Drysdale Securities case, 45, 51, 52–53 detection of unauthorized positions, 53 development of unauthorized positions, 52–53 failure to detect unauthorized positions, 53 further reading, 53 incident, 52 lessons learned, 53 result, 52 INDEX Chew, Lillian, 51, 57, 77 Chief financial officer, funding liquidity risk control and, 43 Ching, Anne, 424 Cholesky decomposition method, 178 Chou, Andrew, 394, 404 Chriss, Neil, 22, 255–256, 386 Citigroup, 40, 79–80, 94–98, 101, 106 Clementi, Gian Luca, 121 Clewlow, Les, 386, 426 Cliquet options, 378–379 Closeouts: exchange-traded derivatives, 506, 507, 508, 509, 510, 511 over-the-counter derivatives, 513, 515, 516, 517, 519, 520, 529 Cochrane, John, 141 Collateral: Chase Manhattan Bank/Drysdale Securities case, 45, 51, 52–53 continuous collateral calls on futures contracts, 273 nondeliberate incorrect information and, 35 Société Générale case, 31, 61–66, 67 Collateralization approach, 515–526 ISDA Master Agreement, 515–516 wrong-way risk, 521–526 Collateralized debt obligations (CDOs). See also Credit default swaps (CDS); Financial crisis of 2007– 2008 CDO creators in financial crisis of 2007–2008, 88–89, 111, 116–117 computational approximations, 226 credit risk management and, 445–446 default basket, 495 equity tranches, 89, 92–93, 94–95, 494–495 faulty models in financial crisis of 2007–2008, 98–99 illiquidity of, 143–144 mezzanine tranches, 94, 102, 104–105, 494 multiname credit derivatives, 493–501 Index risk management and reporting for portfolio credit exposures, 490–492 senior tranches, 494 super-senior tranches, 94–102, 104–105, 106–107, 113, 494 Collin-Dufresne, Pierre, 456 Commercial paper: in credit contagion of 2007–2008, 109 estimating amount owed at default, 468–469 Commodities: broad definition, 253–254 financial, 254 physical, 254, 259–260, 301–304 Component VaR, 204–205 Compound options, 379–381 Compound worksheet, 381 Comptroller of the Currency, 12 Computer errors, 36 Conduct of customer business: Bankers Trust (BT) case, 41, 77–79, 367 Enron case, 79–80 other cases, 80–81 Constant-maturity Treasury (CMT), 295 Contagion, 482 credit contagion, 108–109, 115, 126–129 market contagion, 109–111, 115, 129–131 Contango, 303 Contingency plans: for disaster risk, 36 for funding liquidity risk, 43 for model risk and evaluation control, 213–214 Contingent credit default swap (CCDS), 531 Contingent immunization strategy, 136, 161 Contingent premium options, 377–378 Continuous review, 232–234 analysis of overrides, 233–234 557 back-testing, 233 daily P&L reconciliation, 232–233 Contracts, risk of unenforceable, 37–40 Control variate technique in modeling, 360 Convergence position, 414 Convexity: convexity adjustments, 297 of credit instruments, 453–454, 457 defined, 330 price-vol-matrix and, 330, 334–335, 343 of single-payout options, 370 Convexity risk, 305–307 Cooley, Thomas, 121 Copula methodology, 98–99, 180, 211, 482, 487, 489–490, 497–498 Cordell, Larry, 83, 93 Correlation between price and exercise, 422–424 Correlation-dependent interest rate options, 425–439. See also Correlation-dependent options Brace-Gatarek-Musiela (BGM) models, 425 described, 362–364 Heath-Jarrow-Morton (HJM) models, 425 intensity of use, 363 relationship between forwards treated as constant, 426–429 relationship between swaption and cap prices, 437–439 term structure models, 430–436 Correlation-dependent options, 404–425 correlation between price and exercise, 422–424 described, 362 index options, 413–414 interest-rate options (see Correlationdependent interest rate options) linear combinations of asset prices, 405–409 nonlinear combinations of asset prices, 417–422 558 Correlation-dependent (Continued) options to exchange one asset for another, 415–417 risk management of options on linear combinations, 409–413 Counterparty credit risk, 505–531 of CDS-bond basis risk, 456 exchange-traded derivatives, 128– 129, 506–511 over-the-counter derivatives, 128– 129, 508–509, 512–531 overview, 505–506 Counterparty risk groups (CRGs), 526–531 Counterparty Risk Management Policy Group, 74 Countrywide, 84 Cousin, Areski, 482 Coval, Joshua, 500–501 Cox-Ross-Rubinstein binomial tree, 426–429 Coy, Peter, 87 Crack spread, 260 Credit concentration, 486 Credit contagion, in financial crisis of 2007–2008, 108–109, 115, 126–129 Credit default swaps (CDS), 447–451 asset swaps versus, 447, 448 binary, 449–450 CDS-bond basis risk, 454–456 counterparty credit exposure through, 525–526 in credit contagion of 2007–2008, 108–109 credit risk management and, 445–446 legal basis risk, 450 loss given default (LGD), 447–448 in market contagion of 2007–2008, 109–111 Monte Carlo simulation and, 178 origins, 447 total return swaps, 276–278, 450–451 Credit exposure mitigation techniques, 291 INDEX Credit Grades, 475–476 Credit instruments, 447–451 asset swaps, 447 bonds, 447, 454–456 collateralized debt obligations (CDOs) (see Collateralized debt obligations [CDOs]) convexity of, 453–454 credit default swaps (see Credit default swaps [CDS]) Credit-linked note (CLN), 523–524 CreditMetrics, 481–484 Creditors: moral hazard and, 14 outside monitors for, 11–12 CreditPricer spreadsheet, 453, 551 Credit rating agencies: criticism of, 12 estimating probability of default, 459–464 in financial crisis of 2007–2008, 89–92, 111, 113, 117–118 information asymmetry and, 11, 12 investment bank reliance on, 99–100 relationship with investment banks, 90 use of forecasts, 90 Credit risk management, 279, 445–503 counterparty (see Counterparty credit risk) credit instruments, 447–451 large money moves and, 195–196 legal risk versus, 39 loan-equivalent approach, 513–515 models of short-term credit exposure, 451–456 multiname (see Multiname credit derivatives) portfolio (see Portfolio credit risk) risk reporting for market credit exposures, 456–457 single-name (see Single-name credit risk) Credit spread curve, 456–457 Credit value adjustment (CVA), 521 Creswell, Julie, 101 Index Crosbie, Peter, 476–477, 478 Cross-currency swaps, 525 CrossHedge spreadsheet, 420, 441–442, 527, 550–551 Crouhy, Michel, 67 Crush spread, 260 Csiszar, Imre, 200, 486 Culp, Christopher, 76, 206 Daily Telegraph, 132 Daiwa Bank, 66 Danish mortgage structure, 116 Dash, Eric, 101 DataMetricsRatesData spreadsheet, 431 Davidson, Andrew, 85, 87, 90, 116, 424 Davies, Rob, 62–64 Ddeltavol (Taleb), 343–344 De Angelis, Anthony (“Salad Oil King”), 52 Default basket, 495 Default risk: comparison of rates of loss given default, 466 correlation with market values, 525 default percentages by year, 480 estimating amount owed at default, 468–471 estimating default correlations, 479–482 estimating loss given default, 465–468 estimating probability of default, 458–465 five-year default rates, 463–464 leverage in measuring, 73 rating agency evaluations, 459–464 statistical modeling, 464–465 Delbaen, Freddy, 188 Dembo, Ron, 396 Demeterfli, Kresimir, 368–369 Derivative Strategies, 61 Derman, Emanuel, 184, 209, 228, 241, 243, 245–247, 342, 361, 385–391, 403–404, 409, 436 DermanErgenerKani20 spreadsheet, 550 DermanErgenerKaniDoubleBarrier spreadsheet, 403–404, 441, 550 559 Derman-Ergener-Kani hedge: broader applications, 403–404 comparison with other static hedging models, 387–390 key points, 390 unwinding, 391 DermanErgenerKaniPartialBarrier spreadsheet, 403–404, 550 DermanErgenerKani spreadsheet, 550 Derman-Kani dynamic hedge, 383 de Servigny, Arnaud, 459, 461–462, 465–467, 476, 478, 484 Deutsche Bank, 62, 104, 106 DeWit, Jan, 455 Diebold, Francis, 190 Digital options. See Binary options Direct borrowing and lending, 264– 267, 270, 280 Direct negotiation, winner’s curse and, 22–24 Disaster risk, 36 Distance to default, 476–477, 488–489 Divergence position, 414 Diversifiable/idiosyncratic risk, 141– 142, 197 Dixit, Avinash, 4 Documentation: of contribution of risk positions, 203–204 legal risk and, 37–40 of model risk and evaluation control, 218–219 of model verification, 220 Dollar gamma, 331 Dorobantu, Diana, 482 Double barrier options, 382 Dowd, Kevin, 170, 173, 175, 176, 178, 180, 181, 183, 185, 186, 188–192, 199, 203–204 Down and in (knock-in), 382 Down and out (knock-out), 382, 383 Drexel Burnham Lambert, 237 Drift, 382–383, 426 Drysdale Government Securities. See Chase Manhattan Bank/Drysdale Securities case 560 Duc, Francois, 161 Dudewicz, Edward, 175 Duffie, Darrell, 128, 454, 455, 482, 483, 510, 517 Dunbar, Nicholas, 80 Dwyer, Paula, 66 Dynamic hedging strategies: dynamic delta hedging, 315–316, 320, 344–345, 375–376 hedge slippage and, 17–18 impact of drift and mean reversion, 328 models for barrier options, 385–387 Monte Carlo simulation versus dynamic delta hedging, 324 nature of, 313–318 path dependence of, 318–323 performance of, 314 simulation of dynamic hedging, 321–329 for vanilla options, 313–329 Eber, Jean-Marc, 188 Economic scenario stress tests, 193–197 Economist magazine, 79, 102–103, 140, 268–269, 448–450 Eichenwald, Kurt, 80 Einchcomb, Stephen, 331 Eisman, Steve, 104 Elkind, Peter, 80, 523 Ellis, Katrina, 391, 404 Embrechts, Paul, 190–191 Enron, 9, 12–13, 40, 79–80, 523 Enterprise risk, 30, 44 Equity spot risk, 258–259 Equity tranches, 89, 92–93, 94–95, 494–495 Ergener, Deniz, 385–391, 403–404 ERisk, 66 European options. See also Vanilla option risk management American options versus, 426 conventions for, 311–312 intensity of use, 363 European swaptions, 430 INDEX EVT spreadsheet, 548 Exchange-traded derivatives, 128–129, 506–511 closeouts, 506, 507, 508, 509, 510, 511 loss mutualization, 506 margining, 506, 507, 508, 510, 511 netting, 506, 507, 511 novation, 506 Exotic option risk management, 359–443 correlation-dependent interest rate options, 362–364, 425–439 correlation-dependent options, 362, 404–425 exotic options, defined, 311, 359 intensity of use of option structures in various markets, 363 path-dependent options, 362, 381–404 single-payout options, 362, 364–378 time-dependent options, 362, 378–381 valuation reserves and, 152 Extrapolation approach: based on time period, 352–355 extreme value theory (EVT) in, 190–191 Extreme value theory (EVT), 190–191, 198, 548 Fabozzi, Frank, 176, 233 Factor-push stress tests, 199–200 Fair value: defined, 159 risk measurement for position taking, 159–161 Falloon, William, 450 Fannie Mae, 84 Fargher, Neil, 479 Fay, Stephen, 57 Federal Deposit Insurance Corporation (FDIC), 319 Federal Reserve Bank of New York, 128 Federal Reserve Bank of Philadelphia, 93 Index Federal Reserve Board (FRB), 123, 209, 213–223, 232, 234, 252 Federal Reserve System, 12, 30 Fernandes, Chris, 416 FICO scores, 87 Finance function, 8, 10 Financial commodities, 254 Financial Crisis Inquiry Commission, 85, 95 Financial Crisis Inquiry Report (FCIR), 84, 85, 91, 92, 94–95, 98, 102, 106–108 Financial crisis of 2007–2008, 1, 83–132 actuarial versus financial risk management and, 3 broader lessons, 132 CDO creators in, 88–89, 111, 116–117, 496–497 credit contagion in, 108–109, 115, 126–129 credit rating agencies in, 89–92, 111, 113, 117–118 crisis in CDOs of subprime mortgages, 85–108 equity tranches in, 89, 92–93, 94–95 FCIR report on, 84, 85, 91, 92, 94–95, 98, 102, 106–108 insurers in, 96–97, 106–108, 114–115, 126 investment banks in, 93–106, 112–114, 118–126 investors in, 92–93, 111, 118 lessons for regulators, 115–131 lessons for risk managers, 111–115 Li’s Gaussian copula formula and, 98–99, 211, 482, 489–490, 497–498 market contagion in, 109–111, 115, 129–131 overview, 83–85 spread of the crisis, 108–111 subprime mortgage originators in, 86–88, 111, 116 561 “too big to fail” mentality, 105–106, 114, 121, 124 Financial disasters, 49–81 conduct of customer business, 77–81 large money moves, 68–77, 195–196, 201–202 misleading reporting, 49–67 Financial risk management: actuarial risk management versus, 2–4 broader applications of, 2 credit risk (see Credit risk management) default risk (see Default risk) essential components, 136–142 financial versus actuarial risk, 3–4 forward risk (see Forward risk management) instruments that lack liquidity, 144–147, 150–151 options risk (see Vanilla option risk management; Exotic option risk management) quantification in, 2 through risk aggregation, 4 risk control, 161–167 through risk decomposition, 4 risk measurement in, 133–161 spot risk (see Spot risk management) Financial Stability Board, 115, 120 Financial Stability Forum, 114, 115, 120–121, 124, 126, 129, 130 Finger, Christopher, 178, 200, 461, 465, 466, 481, 484 Fitch, 89 Floors/floorlets, 347, 430 Flows: indexed, 295–299 in pricing illiquid flows by interpolation, 284–291 representing promised deliveries, 282, 293–295 stack-and-roll hedge and, 291–293 Focardi, Sergio, 176, 233 Fons, Jerome, 91 562 Foreign exchange spot risk, 257–258 Forward contracts, 253, 272, 273, 280. See also Forward risk management models in which relationship between forwards is treated as constant, 426–429 Forward prices, borrowing costs and, 303–304 Forward rate agreements (FRA), 274–275, 280, 430, 432, 437–439 Forward risk, 279 Forward risk management, 263–310 asset-backed securities, 278–279 direct borrowing and lending, 264–267, 270, 280 factors impacting borrowing costs, 265–269, 299–304 firm-level risk management, 307 forward contracts, 253, 272, 273, 280 forward prices for different time periods, 259–260 forward rate agreements (FRAs), 274–275, 280 forward transactions, defined, 253 futures contracts, 272–274, 280 instruments, 269, 270–282 interest rate swaps, 275–276, 281 Kidder Peabody case, 21, 31, 51, 53–55, 57, 64–66, 234–235 models (see Forward risk models) overlap between interest rate risk and credit risk, 266–269 overview, 263–269 repurchase agreements, 69, 271–272, 280, 294–295 risk comparisons, 280–281 risk management reporting system, 269, 304–308 spot versus forward positions, 263–264 total return swaps, 276–278 Forward risk models, 265, 267–269, 282–299 flows representing promised deliveries, 282, 293–295 INDEX indexed flows, 295–299 pricing illiquid flows by interpolation, 284–291 pricing long-dated illiquid flows by stack and roll, 291–293 Forward-start caplets, 432 Forward-start options, 378–379 hedge at rollover, 380 ForwardStartOption spreadsheet, 381, 550 ForwardStart spreadsheet, 379 FRA (forward rate agreements), 274– 275, 280, 430, 432, 437–439 Frailty analysis, 482 Fraud risk, 31–35 deception about earnings, 31 deception about positions, 31–32 reducing, 32–35 Freddie Mac, 84 Friedman, Billings, Ramsey, 87 Front office: components of, 8 defined, 8 fraud risk and, 32–35 healthy skepticism about, 95–96 hedge slippage and, 17–18 information asymmetry and, 8–10 legal risk and, 38–39 modeling choices of, 238 nondeliberate incorrect information and, 35 risks that are difficult to identify, 231 “too big to fail” mentality and, 105–106, 114 Funding cost, of CDS-bond basis risk, 456 Funding liquidity risk, 42–44 asset liquidity risk versus, 42 components of, 42–43 defined, 30 Futures contracts, 272–274, 280 Futures exchanges: credit in, 70 over-the-counter markets versus, 70 Index Galai, Dan, 67 Gamma: defined, 315, 330–331, 343–344 dollar, 331 hedging costs, 327 price-vol-matrix and, 330–331, 334, 343–344 Gap market risk, 523 Garfield, Andrew, 80 Gates, Bill, 138 Gatheral, Jim, 369, 377, 386, 387, 435 Gaussian copula formula, 98–99, 211, 482, 489–490, 497–498 General Electric (GE), 54, 407–408 Generalized autoregressive conditional heteroscedasticity (GARCH), 176 Gibson Greetings, 41, 77–78, 367 Giescke, Henning, 63 Giesecke, Kay, 490 Gilbert, W. S., 14, 303 Gillen, David, 80 Glasserman, Paul, 409, 490 Global Industry Classification Standard (GICS), 259 Global Legal Group, 116–117 Golden parachutes, 106 Goldman Sachs, 22, 80, 106, 203–206 Gone on special, 294–295 Government: conflict of interest, 12 information asymmetry and, 11 lessons from financial crisis of 2007–2008, 115–131 outside monitors for, 12 Grace period, 508 Granger, Nicholas, 331 Granite Capital, 66–67 Granville-Barker, Harley, 18–19 Greece, 80, 448 Greenlaw, David, 110 Greenspan, Alan, 106 Gregory, Jon, 450, 490, 505–507, 509, 512, 515–519, 521–523, 525, 526, 528, 531 Groslambert, Bertrand, 486 563 Gross position regulation, 63–64, 65 Group of Thirty (G-30), 115–119, 125, 129–131 recommendations on trading risk, 136, 137, 147, 156–157, 159, 169 Group of Twenty (G-20), 115, 127–128 Grunkemeyer, Barbara, 87 Gumerlock, Robert, 4 Gupta, Ajay, 327 Gupta, Vishal, 391, 404 Gupton, Greg, 461, 465–467, 481, 484 Guys and Dolls (Runyon), 44 Hamanaka, Yasuo, 66 Hamilton, David, 467 Hammond, John, 6 Hansell, Saul, 55 Hanweck, Gerald, 347 Harris, Larry, 27 Harvard Business School case studies, 75 Hasanhodzic, Jasmina, 243 Hatzius, Jan, 110 Heath, David, 188 Heath-Jarrow-Morton (HJM) model, 425 Heat maps, 331 Hedge funds, need for broader regulatory oversight, 131 Hedge slippage, 17–18, 232, 243 Hedging. See Dynamic hedging strategies; Static hedging strategies Helwege, Jean, 449, 450 Henderson, Schuyler, 450 Heston, Steven, 346, 386 Heston model, 346 High-yield debt, 307 Himelstein, Linda, 78 Historical data: simulation of P&L distribution, 170–171, 173–174, 180–183 stress tests relying on, 192, 197–201 Holland, Kelley, 78 Holton, Glyn, 177 564 Huang, Yilin, 83, 93 Huertas, Thomas, 131 Hull, John, 4, 83–86, 89, 115, 142, 175, 176, 225, 229, 265, 274, 283, 288, 289, 295–297, 311–313, 348–350, 360–361, 370, 378, 379, 383, 392, 408, 409, 413, 415, 416, 418, 425, 426, 434, 435, 453, 455, 465, 471, 488, 490, 493–495, 509 Hull-White model, 434–435 IBM, 407–408 Idiosyncratic risk. See Diversifiable/ idiosyncratic risk Iguchi, Toshihida, 66 Illegal actions, risk of, 40 Illiquid instruments. See also Collateralized debt obligations (CDOs) asset liquidity risk and, 142–147 choice of liquid proxy, 144–146, 243–245, 247–249 choice of model validation approach, 241–243 design of Monte Carlo simulation, 245–247 implications for marking to market, 247–249 implications for risk reporting, 249–250 model validation and, 241–250 risk management, 144–147, 150–151 Illiquid positions, pitfalls in deriving valuations, 150–151 Importance sampling, 185 Incentive asymmetry, information asymmetry and, 9, 11 Independent auditors, criticism of, 12–13, 79–80 Indexed flows, 295–299 described, 295–297 translation into fixed flows, 298–299 Index options, 413–414 IndyMac, 125 Ineichen, Alexander, 243 INDEX Informationally disadvantaged, 21 Information asymmetry: adverse selection and, 20–21 for creditors, 10 government regulation and, 11 incentive asymmetry and, 9, 11 moral hazard in, 7–16, 148–149 nature of, 8–9 outside monitors and, 10–16 potential solutions, 9–10 traders and, 148–149 Initial margin, for exchange-traded derivatives, 508 Insurers: AIG, 86, 93, 97, 102, 107, 108, 114 in financial crisis of 2007–2008, 96–97, 106–108, 114–115, 126 Interest rate swaps, 80, 145, 275–276, 281 Internal arbitrage, 167 International Monetary Fund (IMF), 92, 93, 115 International Swaps and Derivatives Association (ISDA), 38–39, 450, 467–468, 515–516, 530 Interpolation approach: in building a volatility surface, 346–352 in model validation, 227, 230 pricing illiquid flows by interpolation, 284–291 seasonality of borrowing costs, 302 between strikes, 347–352 Intraday margin calls, 509 Investment analysts: conflict of interest, 12–13 information asymmetry and, 11–12 Investment banks: capital requirements reform recommendation, 122–124 CDO creators in financial crisis of 2007–2008, 88–89, 111, 116–117 compensation reform recommendations, 120–122 565 Index conflict of interest, 12–13 failure to account for illiquidity of super-senior tranches, 101–102, 113 faulty CDO models, 98–99, 112 in financial crisis of 2007–2008, 93–106, 112–114, 118–126 inadequate analysis of statistical hedging, 103–105, 113 inadequate derivative protection, 96–97, 112 inadequate stress tests, 102–103, 113 losses in financial crisis of 2007– 2008, 92, 93 off-balance-sheet vehicles, 97–98, 112 overreliance on VaR measures, 100–101, 113 personnel risk and, 36–37 recommendations for, 118–126 reliance on external ratings, 99–100, 113 risk management procedures reform recommendation, 119–120 size and allowable activities reform recommendations, 124–126 “too big to fail” mentality and, 11, 105–106, 114, 121, 124 Investors, in financial crisis of 2007– 2008, 92–93, 111, 118 Irish central bank, 300 iTraxx index, 493, 494 Jackel, Peter, 175 Jackwerth, Jens, 348 Jacobs, Michael, 467, 469 Jain, Gautam, 409 Jameson, Rob, 29, 66 Jett, Joseph, 51, 53–55, 57, 234–235 Jewson, Stephen, 144 Jorion, Philippe, 176, 191, 205, 206 JPMorgan, 80, 105, 231 JPMorgan Chase, 40, 79–80, 84, 106, 205–206, 249 JumpProcessCredit spreadsheet, 475–476, 551 Jump process models, 475–476 Junk bonds, 307 Jurek, Jakub, 500–501 Kalotay, Egon, 467, 479 Kane, Alex, 141 Kani, Iraj, 385–391, 403–404 Karagozoglu, Ahmet, 467 Kashyap, Anil, 110 Kealhofer, Stephen, 481, 484 Keeney, Ralph, 6 Kerviel, Jérôme, 61–66, 67 Khuong-Huu, Philippe, 434 Kidder Peabody case, 21, 31, 51, 53–55, 57, 64, 234–235 detection of unauthorized positions, 55 development of unauthorized positions, 54 failure to detect unauthorized positions, 54–55 further reading, 55 incident, 53–55 lessons to be learned, 55 result, 54 Kim, Jongwoo, 178, 200 King, Mervyn, 126 Kirshner, Susan, 302 KMV approach, 467, 476–484 Knock-in (down and in/up and in), 382 Knock-out (down and out/up and out), 382, 383 Kolm, Petter, 176, 233 Kooi, Mari, 66 Kotowitz, Y., 7–8, 11 Koutoulas, James, 511 Krenn, Gerald, 200 Kurer, Peter, 99–100 Ladder options, 402–403 Large complex financial institutions (LCFIs), 120 Large homogenous portfolio (LHP), 487–490, 495 566 Large money moves, 68–77. See also Long-Term Capital Management (LTCM) case Metallgesellschaft (MG) case, 75–77, 135, 201–202, 273 stress tests and, 195–196, 205 Laurent, Jean-Paul, 490 Law of one price, 239–241 Lee, Roger, 386 Lee, Yoolim, 80 Leeson, Nick, 51, 55–57, 66 Legal-basis risk, 38–39, 450 Legal risk, 37–40 bankruptcy and, 39–40 defined, 30 error in legal interpretation, 50 mitigating, 37–40 risk of illegal actions, 40 of unenforceable contracts, 37–40 Lehman Brothers, 455 Leibowitz, Martin, 161 Lender of last resort facilities, 129 Leonhardt, David, 103 Leverage: in capital structure analysis, 477 as measure of default risk, 73 Lewis, Michael, 102, 104 Li, Ada, 128, 510 Li, David, 98–99, 211, 488 Li, Jingyi, 490 LIBOR (London Interbank Offered Rate), 69, 267–269, 275, 297–298, 447 Limited partnerships, 80 Linear combinations of asset prices, 405–413 approximation of option values, 407–409 derivative characteristics, 405 derivative payoffs as linear functions of, 406–407 risk management of options with, 409–413 rules for dynamic hedging, 407 Lippmann, Greg, 104 Liquid instruments, model risk and, 237–241 INDEX Liquidity risk. See also Asset liquidity risk; Funding liquidity risk costs of liquidation, 139 defined, 3 time required for liquidation, 135–136 Liquidity squeeze, 469 Liquid proxy: control variate technique compared with, 360 for derivatives with actuarial risk, 520 for illiquid instruments, 144–146, 243–245, 247–249 reasons to use, 144–146 Li’s Gaussian copula formula, 98–99, 211, 482, 489–490, 497–498 Litterman, Robert, 203–204, 307 Lo, Andrew, 84, 85, 131, 243 Loan-equivalent approach, 513–515 Loan-to-value ratios, 87 Local volatility models, 385 Log contracts, 367–369 London Interbank Offered Rate (LIBOR), 69, 267–269, 275, 297–298, 447 Long, meanings of, 304 Long-Term Capital Management (LTCM) case: bailout, 71–72 large money moves, 68–77, 109, 122, 160, 193, 195, 201–202, 206, 211, 250, 318 lessons learned, 73–75 management style, 68–69, 73–74 suggestions for improved practices, 74–75 types of positions, 69–70 Union Bank of Switzerland (UBS) and, 59, 61, 318 Lookback options, 402 Loss given default (LGD), 447–448 estimating, 464–468, 484–485, 487, 491 Lowenstein, Roger, 61, 72, 75, 81, 90, 96 Lubke, Theo, 128, 510 Index Ludwig, Eugene, 51, 59 Lynch, Gary, 55 Madan, Dilip, 364, 386 Madoff, Bernie, 17 Malcolm, Fraser, 37, 40 Marcus, Alan, 141 Margin Call (film), 1 Margin calls, 128, 509 Margining: exchange-traded derivatives, 506, 507, 508, 510, 511 over-the-counter derivatives, 513, 514, 516–517, 518–520, 522–523, 531 Mark, Robert, 67 Market contagion: in financial crisis of 2007–2008, 109–111, 115, 129–131 need for broader regulatory oversight, 131 need for more orderly bankruptcy proceedings, 131 need to reduce procyclicality, 129–131 Marketers, in front office, 8 Market makers/market making: gambling analogy, 26 hedging in spot markets, 254–255 impact of customer order flow in spot markets, 255–257 liquidity risk/basis risk trade-off, 255–256 market making, defined, 5 models to perform risk decomposition, 25–26 position taking versus, 24–27 winner’s curse and, 22 Market risk, legal risk versus, 39 Market using. See Position taking Marking to market, 95–96, 110, 130–131, 143, 147–152 analysis of revenue and, 156 caveats concerning, 150–151 dollar versus Japanese yen, 147, 153 567 in establishing exit prices, 148–149, 151–152 for exchange-traded derivatives, 508 by expert panels, 149–150 exposure to market price shifts, 157–159 frequency of, 147 with illiquid positions, 146, 147–151 liquid proxy for illiquid instruments, 144–146, 243–245, 247–248 purpose of, 146 Markowitz, Harry, 141 Martinuzzi, Elisa, 80 Matytsin, Andrew, 350, 386 Maurer, Samuel, 449, 450 Mayer, Martin, 55, 57, 67 MBIA, Inc., 107–108 McAdie, Robert, 455 McDonald, Robert, 260 McKay, Peter, 66 McLean, Bethany, 80, 90, 97, 106, 125, 523 McNeil, Alexander, 190 Mean reversion, 327–328 Mello, Antonio, 76–77 Merck, 407–408 Merger arbitrage, 26–27 Merrill Lynch, 67, 79, 94, 95, 106, 107–108 Merton model, 474–476, 551 MertonModel spreadsheet, 474, 477, 551 Metallgesellschaft (MG) case, 75–77, 135, 201–202, 273 Metropolitan Life, 4 Meucci, Attiolio, 176 Mezzanine tranches, 94, 102, 104–105, 494 MF Global, 511 Middle office, 10 defined, 8 fraud risk and, 32–35 model verification and, 223 Mihm, Stephen, 126 Milken, Michael, 237 Miller, Merton, 76 Miller, William “520 Percent,” 18–19 568 Misleading reporting, 49–67 Allied Irish Bank (AIB) case, 31, 51, 57–59, 64, 65 Barings Bank case, 31, 51, 55–57, 64, 66 Chase Manhattan Bank/Drysdale Securities case, 45, 51, 52–53 deception about earnings, 31 deception about positions, 31–32, 49–51 Kidder Peabody case, 21, 31, 51, 53–55, 57, 64–66, 234–235 other cases, 66–67 risk of nondeliberate, 35–36 Société Générale case, 31, 61–66, 67 Union Bank of Switzerland (UBS) case, 59–61, 67 MixtureofNormals spreadsheet, 5, 6, 175, 178, 207, 548 Model risk, 209–252. See also Model risk evaluation and control defined, 209 illiquid instruments, 241–250 importance of, 210–212 liquid investments, 237–241 as operations risk, 35–36 trading models, 250–252 valuation of illiquid positions, 150 Model risk evaluation and control, 212–237. See also Model risk board of directors role in, 219 business unit accountability for, 215–219 capturing difficult-to-identify risks, 231 components of review, 214 continuous review, 232–234 documentation of, 218–219 model as term, 213–214 model validation, 212, 226–231 model verification, 212, 219–226 periodic review, 234–237 proprietary information and, 217–218, 250–252 roles and responsibilities, 214–219 INDEX scope, 213–214 senior management role in, 219 vender versus in-house models, 213–214 Model validation, 226–231 capturing difficult-to-identify risks, 231 choice of approach for illiquid instruments, 241–243 cost of hedging approach, 227–228 defined, 212, 226–227 illiquid instruments and, 241–250 interpolation approach, 227, 230 liquid instruments and, 237–241 matching to model purpose, 229–231 no-arbitrage principle and, 239–241 by outside reviewers, 238, 245–246, 250–252 prevailing market model approach, 228–229 of specific trading strategies, 230–231 Model verification, 219–226 of approximations, 223–226 components of, 220–221 of deal representation, 222–223 defined, 212 degree of complexity of models, 221–222 independent implementation, 220 model error and, 221–222 nature of, 219–220 rules, 220–221 suggested controls for computational approximation, 224–225 systems implementation, 220 testing on cases with known solutions, 220–221 Money market mutual funds, in credit contagion of 2007–2008, 109 Monopoly rents, 167 Monte Carlo simulation: advantages of, 176–180, 245–246 computational alternatives to full simulation, 486–490 569 Index of counterparty credit exposure, 517–518, 524–525 disadvantage of, 199 dynamic hedging of vanilla options, 321–329 equal probability weights for all simulation runs, 189 for illiquid instruments, 242–243, 245–247 missing/nonsynchronous data in, 176–177, 181–182 model verification using, 225–226 of options hedging, 357–358 of P&L distribution, 170–171, 175–183 of portfolio credit risk, 482–486 with stress tests, 198–199, 200–201 stress tests versus, 192–193 Monte Carlo stress tests, 200–201 Moody’s Investors Service ratings, 11, 89, 91, 108, 268, 459–464, 466, 480 Moody’s KMV, 467, 476–484 Moosa, Imad, 47 Moral hazard, 7–16 in analysis of insurance risks, 13–14 conflict between insiders and outsiders, 8–16 defining, 7–8 information asymmetry, 7–16, 148–149 in risk measurement, 135 taking large risk positions, 50–51 “too big to fail” mentality and, 11, 72, 105–106, 114, 121, 124 in value placed on earnings volatility, 15–16 Morgan Grenfell Asset Management, 80 Morgan Stanley, 4, 95 Morgan Stanley Capital International (MSCI), 259 Morini, Massimo, 209, 211, 212, 215, 220, 221, 226–227, 230–231, 235, 239, 471 Morningstar, 258–259 Mortgage brokers, in financial crisis of 2007–2008, 86–87 Multiname credit derivatives, 493–501 CDO tranches and systematic risk, 500–501 modeling, 495–498 nature of, 493–495 risk modeling and reporting for, 498–500 Myers, Stewart, 141 Myktyka, Edward, 175 Nagpal, Krishan, 461 Narrow banks, 125 NastyPath spreadsheet, 319, 345, 548 National Association of Insurance Commissioners, 459 National Westminster Bank, 67 Netting: exchange-traded derivatives, 506, 507, 511 over-the-counter derivatives, 513, 514, 515, 516, 517, 518, 520–521, 529, 531 Neuberger, Anthony, 368 New York Stock Exchange (NYSE), 277 New York Times, 87, 103 No-arbitrage principle, 239–241 Nobel Prize in economics, 20 Nocera, Joe, 90, 97, 106, 125 Nondiversifiable risk, 141–142, 197 Nonlinear combinations of asset prices, 417–422 Norris, Floyd, 107–108 Northern Rock, 125 Novation, exchange-traded derivatives, 506 Numeraire, 305, 312 O’Brien, Timothy L., 206 Off-balance-sheet vehicles, 97–98 Office of the Comptroller of the Currency, 87 570 “Off-the-run” instruments, 238 O’Kane, Dominic, 161, 455, 489, 490, 492–496, 500 One-way markets, 151 One-year tenor options, 432 “On-the-run” instruments, 238 Operational risk, 29–47 accounting risk, 30, 42 defined, 29 enterprise risk, 30, 44 funding liquidity risk, 30, 42–44 identification of risks, 44–45 legal risk, 30, 37–40 operational risk capital, 45–47 operations risk, 30, 31–37 reputational risk, 30, 41–42 Operational risk capital, 45–47 bottom-up approach, 46 top-down approach, 46 Operations risk, 31–37 defined, 30 disaster risk, 36 personnel risk, 36–37 risk of fraud, 31–35 risk of nondeliberate incorrect information, 35–36 OptBarrier spreadsheet, 396, 440–441, 550 Option-adjusted spread (OAS), 424 OptionMC1000 spreadsheet, 358, 549 OptionMCHedged1000 spreadsheet, 358, 549 OptionMCHedged spreadsheet, 357–358, 549 OptionMC spreadsheet, 357–358, 549 OptionRoll spreadsheet, 352–355, 549 Options risk management. See also Exotic option risk management; Vanilla option risk management options conventions, 311–312, 426–427 options transactions, defined, 253 overview of options risk management, 313–318 Options to exchange one asset for another, 415–417 INDEX Option-theoretic approach, 471–479 jump process models, 475–476 KMV statistical analysis, 476–484 Out-of-the-money calls, 319 Overbeck, Ludger, 490 Override analysis, 233–234 Over-the-counter derivatives, 128–129, 508–509, 512–531 active management report, 526–531 closeout, 513, 515, 516, 517, 519, 520, 529 collateralization approach, 515–526 counterparty risk groups (CRGs), 526–531 loan-equivalent approach, 513–515 margining, 513, 514, 516–517, 518–520, 522–523, 531 netting, 513, 514, 515, 516, 517, 518, 520–521, 529, 531 overview, 512–513 wrong-way risk, 521–526 Oyama, Tsuyoshi, 85 Padovani, Otello, 404 Paine Webber, 54 Pandit, Vikrim, 96–97 Parking, 31 Parsons, John, 76–77 Partial differential equations (PDEs), 383–384 Partial-time barrier options, 382 Path-dependent options, 17–18, 381–404 barrier options with rebates, 402 broader classes, 403–404 deriving the Carr hedge, 393 described, 362 dynamic hedging models for barriers, 385–387 in exotic option risk management, 381–404 intensity of use, 363 ladder options, 402–403 lookback options, 402 put-call symmetry, 391–392 Index standard analytic models for barriers, 383–384 static hedging models for barriers, 387–401 in vanilla option risk management, 318–323 Paulson, John, 104 Pearson, Neal, 496 Pension funds, as investors in financial crisis of 2007–2008, 92–93, 111, 118 Performance attribution, 258–259 Performance measurement, 205–206 Periodic review, 234–237 changes in academic literature, 237 changes in market environment, 236–237 changes in market practices, 237 changes in population of transactions, 234–236 changes in technology, 237 Perold, Andre, 16, 69, 75 Personnel risk, 36–37 Phantom profits, 167 Physical commodities: borrowing costs for, 303–304 defined, 254 financial commodities versus, 254 spot risk, 259–260 storage costs, 301, 302 transportation costs, 259–260 Pindyck, Robert, 4 Pin risk (Taleb), 377 Pirrong, Craig, 128 Plain-vanilla options. See Vanilla option risk management Ponzi, Charles, 18–19 Ponzi schemes, 17–19, 156 broadened meaning, 17–18 hedge slippage and, 17–18, 232 Kidder Peabody case, 21, 31, 51, 53–55, 57, 64–66, 234–235 losses from unauthorized positions and, 51 original meaning, 17, 18–19 571 Portfolio credit risk, 479–493, 515 computational alternatives to full simulation, 486–490 estimating default correlations, 479–482 Monte Carlo simulation of, 482–486 risk management and reporting for portfolio credit exposures, 490–492 Portfolio insurance, 96–97, 315, 320 Portfolio Risk Tracker, 484–485 Portfolio theory, 141 Position managers, in front office, 8 Position taking: defined, 25 gambling analogy, 26 instruments outside area of expertise, 164–165 market making versus, 24–27 models as forecasting tools, 25 risk measurement for, 159–161 Power options, 366–367 Predescu, Mirela, 455 Price taking. See Position taking Price-vol matrix: advantage of, 339 for being a short a call option, 331, 332 for a calendar spread, 336, 337 for a call spread, 331–334 interpolation results based on, 185 for a reduced risk portfolio, 336–341 in vanilla option risk management, 315–317, 323–324, 326, 329–344 PriceVolMatrixCycle spreadsheet, 336, 548 PriceVolMatrix spreadsheet, 329–344, 548 Pricewaterhouse Coopers, 101, 122, 127, 129–131 Prince, Chuck, 94 Private equity funds, need for broader regulatory oversight, 131 Procter & Gamble (P&G), 41, 77–78, 367 572 Program trading, 251 Proprietary trading, 125–126 Prudential-Bache Securities, 80 Pull to par, 429 Pyramid schemes, 17. See also Ponzi schemes Quanto: nonlinear combinations of asset prices, 417–422 single-asset quanto options, 369–370 Quanto worksheet, 370 Rafael, Andrea, 201 Raiffa, Howard, 6 Rainbow contracts, 419 Rajan, Raghuram, 122, 125, 126, 499 Ramberg, John, 175 Random matrix theory/shrinkage estimation, 176 RateData spreadsheet, 308–309, 548 Rates spreadsheet, 283–284, 288, 307, 310, 548 Ratios worksheet, 138 Rawnsley, Judith, 57 Real options, 4 Rebates, barrier options with, 402 Rebonato, Riccardo, 6, 209, 228, 237, 241, 242, 327, 348, 434–435, 438 Rebooking trades, 152–153 Reduced risk portfolio, 336–341 Rehedging, 196, 322, 327–329 Reiner, Eric, 417 Remargin period, 519 Remolona, Eli, 84, 465 Renault, Olivier, 459, 461–462, 465–467, 476, 478, 484 Rennie, Andrew, 418 Repurchase agreements (RPs), 69, 271–272, 280, 294–295 Reputational risk, 41–42, 77–81 accounting risk as form of, 42 Bankers Trust (BT) case, 41, 77–79, 367 defined, 30 INDEX large money moves and, 196 nature of, 41–42 Researchers, in front office, 8 Reserves. See Valuation reserves Resti, Andrea, 465, 467 Revealing positions, problems of, 150–151 Richardson, Matthew, 84, 118, 121 Right-way risk, 522 Risk-adjusted return on capital (RAROC), 44, 206 Risk aggregation, 4 Risk arbitrage. See Merger arbitrage Risk control, 161–167 detailed limits on size of exposure, 162–165 incentive-based approach to, 161–163 internal hedging in, 166–167 risk decomposition and, 166 Risk decomposition: defined, 4 models to perform, 25–26 reporting in, 203–204 risk control and, 166 Risk Identification for Large Exposures (RIFLE), 231 Risk magazine, 99 Risk management. See Financial risk management Risk Management Association, 485 Risk managers, in front office, 8 Risk measurement, 133–161 analysis of revenue, 156–157 exposure to changes in market prices, 157–159 general principles, 133–144 instruments that lack liquidity, 144–147, 150–151 liquidation time and, 135–136 market valuation, 147–152 for position taking, 159–161 principles of risk management in, 136–142 rules for, 133–134 Index stop-loss limit, 133–136 valuation reserves and, 145, 146, 152–156 RiskMetrics Group, 177 Risk of fraud, 31–35 pressures, 32–35 Risk of nondeliberate incorrect information, 35–36 Risk reversals, 334–335, 377 Roe, John, 511 Roseman, Alan, 107 Rosen, Dan, 485 Ross, Stephen, 141 Roubini, Nouriel, 126 Royal Bank of Scotland, 67 Rubinstein, Mark, 348 Rullière, Didier, 482 Runyon, Damon, 44 Rusnak, John, 51, 57–59 Russian debt default of 1998, 71, 206 Salespeople, in front office, 8 Salmon, Felix, 98, 209, 211 Salomon Brothers, 68, 71 Sanders, Anthony, 424 Sarkar, Asani, 449, 450 Saunders, Anthony, 120, 464, 481 Scenario analysis, 211 Schachter, Barry, 190, 191, 194 Scheinkman, Jose, 307 Scheuermann, Til, 190 Schonbucher, Philipp, 99, 475, 483, 487–489 Schorderet, Yann, 161 Schuermann, Til, 85, 90 Schutz, Dirk, 61 Seasonality, of borrowing costs, 302 Securities and Exchange Commission (SEC), 12, 153, 159 Securities Industry Association, 210 Seinfeld (TV program), 1 Sell side. See Market makers/market making Semi-American options, 426 Semi-European options, 426 Senior Supervisors Group report, 100 573 Senior tranches, 494 September 11, 2001 attacks, disaster risk and, 36 Serrat, Angel, 268, 284, 298–299, 304–305, 308 Shakespeare, William, 142–143 Shareholders: information asymmetry and, 11–12 outside monitors for, 11–12 Shareholder value added (SVA), 44, 206 Sharma, Pawan, 37, 40 Sharpe, William, 141, 258–259 Sharpe ratio, 5, 160 Shaw, Julian, 175, 182 Shiller, Robert, 103, 116, 132 Shin, Hyun Song, 110 Shirreff, David, 61, 72 Shkolnik, Alexander, 490 Short, meanings of, 304 Shortfall/expected shortfall VaR, 187–189 Shorting a call option, 331, 332 Short squeeze, 300 Short-term credit exposure, 451–456 CDS-bond basis risk, 454–456 convexity of credit instruments, 453–454 impact of bankruptcy law, 452–453 risk reporting for market credit exposures, 456–457 Sidenius, Jakob, 490 Sifakis, Carl, 18–19 Simulation. See also Monte Carlo simulation advantages of, 139–140 computational alternatives to full simulation, 486–490 historical data in, 5, 170–171, 173–174, 180–183 illiquid positions in, 146 Monte Carlo (see Monte Carlo simulation) nature of, 138–139 of P&L distribution, 170–171, 173–187 574 Simulation (Continued) for risk measurement, 138–140 subjective judgment and, 4–6 Single-asset quanto options, 369–370 Single-name credit risk, 457–479 estimating amount owed at default, 468–471 estimating loss given default, 465–468 estimating probability of default, 458–465 option-theoretic approach, 471–479 Single-payout options, 364–378 accrual swaps, 378 binary options, 371–377 contingent premium options, 377–378 convexity, 370 described, 362 intensity of use, 363 log contract swaps, 367–369 single-asset quanto options, 369–370 variance swaps, 367–369 Singleton, Kenneth, 454, 455, 482, 483 Sironi, Andrea, 465, 467 Skew, 349–350 Smiles, 348–349 Smith, Adam, 11 Smith, Roy, 120 Smithson, Charles, 362, 363, 496 Société Générale case, 31, 61–66, 67 detection of unauthorized positions, 64 development of unauthorized positions, 62 failure to detect unauthorized positions, 62–64 further reading, 66 incident, 61 lessons to be learned, 64–66 result, 62 Sony Corporation, 417 Sorkin, Andrew Ross, 109 Soros, George, 116 South Korea, 80 INDEX SpecComm, 62–63, 63 Speculation. See Position taking Spence, Michael, 20 Split-fee options, 379–381 Spot risk management, 253–261, 279 equity, 258–259 firm-level risk management, 257 foreign exchange, 257–258 overview, 253–257 physical commodities, 259–260 spot trades, defined, 253 Spreadsheets: AmericanOption spreadsheet, 427–428, 551 BasketHedge spreadsheet, 364, 368–370, 381, 412–413, 439, 549–550 BasketOption spreadsheet, 550 BinaryMC spreadsheet, 376, 440, 550 Bootstrap spreadsheet, 289, 548 calculating default rates from bond rates, 501–502 CapFit spreadsheet, 347, 549 CarrBarrierMC spreadsheet, 396, 440, 550 CarrBarrier spreadsheet, 394–396, 440, 550 CDO spreadsheet, 496, 551 comparing the jump process credit model to the Merton model, 502 CreditPricer spreadsheet, 453, 551 CrossHedge spreadsheet, 420, 441–442, 527–531, 550–551 DataMetricsRatesData spreadsheet, 431 DermanErgenerKani20 spreadsheet, 550 DermanErgenerKaniDoubleBarrier spreadsheet, 403–404, 441, 550 DermanErgenerKaniPartialBarrier spreadsheet, 403–404, 550 DermanErgenerKani spreadsheet, 550 EVT spreadsheet, 548 ForwardStartOption spreadsheet, 381, 550 Index ForwardStart spreadsheet, 379 generating fat tails in Monte Carlo simulations, 207 interpolation, 308–309 JumpProcessCredit spreadsheet, 475–476, 551 maximizing diversification, 207 measuring fat tails in historical data, 207 MertonModel spreadsheet, 474, 477, 551 MixtureofNormals spreadsheet, 5, 6, 175, 178, 207, 548 Monte Carlo simulation of options hedging, 357–358 NastyPath spreadsheet, 319, 345, 548 OptBarrier spreadsheet, 396, 440–441, 550 OptionMC1000 spreadsheet, 358, 549 OptionMCHedged1000 spreadsheet, 358, 549 OptionMCHedged spreadsheet, 357–358, 549 OptionMC spreadsheet, 357–358, 549 OptionRoll spreadsheet, 352–355, 549 options portfolio risk measures, 356–357 PriceVolMatrixCycle spreadsheet, 336, 548 PriceVolMatrix spreadsheet, 329– 344, 548 RateData spreadsheet, 308–309, 548 Rates spreadsheet, 283–284, 288, 307, 310, 548 simulation of the impact of trading rules on expected return and risk, 260–261 stack and roll, 309 Swaptions spreadsheet, 437, 442, 551 TermStructure spreadsheet, 432–433, 442, 551 using Vasicek model for risk measurement of CDO tranches, 502–503 575 value-at-risk computations, 206–207 VaR spreadsheet, 138, 181, 189, 206–207 (example), 548 VolCurve spreadsheet, 347, 549 VolSurfaceStrike spreadsheet, 348, 352, 549 WinnersCurse spreadsheet, 24, 548 Squam Lake Group, 118, 119, 121, 127, 131 Stack-and-roll hedge, 291–293 advantages of, 293 described, 291–293 Stafford, Erik, 500–501 Standard & Poor’s (S&P) 500 stock index, 3–4, 141, 157–159, 259, 319, 371–377 Standard & Poor’s (S&P) ratings, 11, 89, 108 Static hedging strategies: for barrier options, 387–401 for exotic options, 361–362 flows representing promised deliveries, 293–295 indexed flows, 295–299 nature of, 313 pricing illiquid flows by interpolation, 284–291 quasistatic representations, 361–362 stack-and-roll hedge, 291–293 Static overhedge, 375 Statistical hedging, inadequate analysis in financial crisis of 2007–2008, 103–105 Stay period, 519 Stein, Roger, 464–467, 475, 479, 482, 483, 492 Sticky delta, 342 Sticky strike, 342 Stiglitz, Joseph, 20, 126 Stigum, Marcia, 51, 53 Stochastic volatility models, 385 Stop-loss limits, 133–136, 162–163 Storage costs, 301, 302 576 Stress tests, 1, 192–201 in assessing credit risk, 75 in capital requirements reform recommendations, 122–124 of counterparty credit exposure, 520 economic scenario stress tests, 192, 193–197 for exchange-traded derivatives, 509–510, 511 factor-push, 199–200 historical data stress tests, 192, 197–201 impact of large money moves and, 205 inadequate, 102–103, 113 large money moves and, 195–196, 205 Monte Carlo simulation versus, 192–193 Monte Carlo simulation with, 198–199, 200–201 overall measures of firm position risk, 201–205 overview, 192–193 performance measurement and, 205–206 for positions that achieve liquidity, 143 Strickland, Chris, 386, 426 Stroughair, John, 190 Structured Finance Litigation blog, 92 Structured investment vehicles (SIVs), 97–98 Structurers, in front office, 8 Subjective judgment. See also Stress tests historical information versus, 5 in Li’s Gaussian copula formula, 98–99 simulation and, 4–6, 140–141 Subprime mortgage originators. See also Financial crisis of 2007–2008 in financial crisis of 2007–2008, 86–88, 111, 116 Sullivan, Arthur, 14, 303 INDEX Sumitomo Corporation of Japan, 66 Suo, Wulin, 229, 455 Super-senior tranches, 94–102, 104– 105, 106–107, 113, 494 Swaps: accrual, 378 basis, 298–299 binary credit default, 449–450 credit default (see Credit default swaps [CDS]) cross-currency, 525 interest rate, 80, 145, 275–276, 281 log contract, 367–369 total return, 276–278, 450–451 variance, 367–369 volatility, 368–369 Swaptions, 413–414 Bermudan, 432, 433–434 European, 430 relationships between cap prices and, 437–439 Swaptions spreadsheet, 437, 442, 551 Swensen, David, 176 Swiss Bank Corporation (SBC), 60 Synthetic tranches, 494 Systematic/nondiversifiable risk, 141–142, 197 Tadikamalla, Pandu, 175 Taleb, Nassim, 14–15, 138, 300, 318–319, 334, 342–344, 347, 350, 355, 377 Tanega, Joseph, 37, 40 Technologists, in front office, 8 Technology stock bubble (2001), 80–81 Telecom, 263 Term structure models, 430–436 TermStructure spreadsheet, 432–433, 442, 551 Tett, Gillian, 88, 91, 95, 97, 99–100, 102, 105 Thaler, Richard, 21–22 Theta: defined, 343–344 price-vol-matrix and, 343–344 Index Tickets in the drawer, 31 Time-dependent options, 378–381 cliquet options, 378–379 compound options, 379–381 described, 362 forward-start options, 378–379 intensity of use, 363 “Too big to fail” mentality, 11, 72, 105–106, 114, 121, 124 Total return swaps, 276–278, 450–451 Totem Market Valuations service, 150 Trade cancellation, 62–63, 65 Trade compression, 530 Traders: adverse selection and, 20–21 collusion and, 63, 65 conservatism versus independence and, 155–156 control personnel versus, 148–149 delta rehedging and, 196 detailed limits on size of exposure, 162–165 fraud risk and, 32–35 in front office, 8 G-30 recommendations on trading risk, 136, 137, 147, 156–157, 159, 169 incentive-based approaches in risk control, 161–163 information asymmetry and, 7–16, 148–149 monopoly rents and, 167 moral hazard and, 14–15, 148–149 positions in instruments outside area of expertise, 164–165 pressure to book immediate profits, 240–241 trading models and, 246–247, 250–252 valuation reserves and, 152–156 Trading and Capital-Markets Activities Manual (Federal Reserve System), 30 Trading models, 250–252 Transportation costs, in physical commodities spot risk, 259–260 577 Treasury function, funding liquidity risk control and, 43–44 Trinomial tree model, 425–426 Tsiveriotis, Kostas, 416 Tuckman, Bruce, 268, 284, 298–299, 304–305, 308 Turner Review, 114–116, 120, 121, 124–125 Twelfth Night (Shakespeare), 142–143 UniCredit Group, 63 Union Bank of Switzerland (UBS): Amplified Mortgage Portfolio (AMPS), 103–104 analysis of financial crisis of 2007– 2008, 95, 99–101, 103–104 VaR methodologies, 99–101, 104 Union Bank of Switzerland (UBS) case, 59–61, 67, 335 development of authorized positions, 60–61 further reading, 61 incident, 59 lessons learned, 61 result, 60 Up and in (knock-in), 382 Up and out (knock-out), 382, 383 Utopia, Limited (Gilbert & Sullivan), 14, 303 Vacation policy, 63, 65 Valuation reserves, 145, 146, 152–156 aging reserve policy, 155 impact of exiting large positions, 154 model verification and, 226 objective standards for reserves, 153–156 to shield earnings from fluctuation, 248–249 Value at risk (VaR) analysis, 1, 104, 136, 169–191 back-testing, 191, 233 based on credit rating agencies, 99–100 based on historical variance/ covariance, 170, 172–173 578 Value at risk (VaR) analysis (Continued) in capital requirements reform recommendations, 122–124 counterparty credit exposure, 517–520 detail recorded on positions and market prices, 185–186 determining all market variables, 183–187 direct measurement of profit and loss, 170, 171–172 earnings volatility and, 205 for exchange-traded derivatives, 508, 509–510, 511 exotic derivative prices and, 186 extreme value theory (EVT) in, 190–191 in financial crisis of 2007–2008, 99–101, 172–173 for forward positions, 184 illiquid positions in, 146 importance sampling in, 185 liquidity considerations in, 146, 186–187 measures of profit and loss distribution, 187–191 nonstatistical measures versus, 169 for option positions, 184–186 overall measures of firm position risk, 201–205, 201–206 overreliance on, 100–101, 113 performance measurement and, 205–206 for positions that are born illiquid, 143 in risk control, 162 shortfall/expected shortfall, 188–189 simulations of P&L, 170–171, 173–187 for spot positions, 183–184 Vanilla option risk management, 311–358 building a volatility surface, 346–355 INDEX conventions, 311–312 delta hedging, 315–316, 320, 344–345 dynamic hedging strategies, 314, 318–329 overview, 313–318 risk reporting and limits, 329–344 tools in, 315–316 vanilla call spread, 372 vanilla options, defined, 311, 426 van Nieuwerburgh, Stijn, 84 VaR analysis. See Value at risk (VaR) analysis Varian, Hal, 406 Variance gamma model, 386 Variance swaps, 367–369 VaR spreadsheet, 138, 181, 189, 206–207, 548 Vasicek model, 98, 102, 487–489, 496, 501 Vause, Nicholas, 530 Vega: defined, 315, 330 price-vol-matrix and, 330, 334–335 Vigorish, 26 Volatility surface, 346–355 extrapolating based on time period, 352–355 interpolating between strikes, 347–352 interpolation between time periods, 346–347 for pricing vanilla options, 346 Volatility swaps, 368–369 Volcker, Paul, 125, 132 Volcker rule, 125 VolCurve spreadsheet, 347, 549 VolSurfaceStrike spreadsheet, 348, 352, 549 Vosey Inheritance, The (GranvilleBarker), 18–19 Wagner, Christoph, 490 Wall Street Journal, 21, 40, 116 Walter, Ingo, 120, 121 579 Index Wang, Jin, 175 Wang, Yuan, 449, 450 Washington Mutual, 84, 125 Wealth of Nations, The (Smith), 11 Weather derivative options, 144 Weinberger, Alfred, 161 Weiss, Gary, 55 Whaley, Elizabeth, 344 White, Alan, 409, 455, 490, 495 White, Lawrence, 84, 118 Williams, Jeffrey, 265 Williams, Meredith, 83, 93 Wilmott, Paul, 344 Wilson, Charles, 20 Wilson, Harry, 67 Wilson, Thomas, 201–202, 205, 484, 485 Winner’s curse, 21–24 application to trading, 22–24 defined, 21 mechanism leading to, 21–24 WinnersCurse spreadsheet, 24, 548 Winters, Bill, 526 Wired magazine, 98 Wolfe, Eric, 67 Wolfe, Lan-Ling, 424 World Bank, 115 Wrong-way risk, 521–526 Y2K crisis, 36, 192, 302 Yield curve, nonstatistical limits on yield curve shape, 307–308 Youngblood, Michael, 87 Zandi, Mark, 94, 130–131 Ziehmann, Christine, 144 Zou, Joseph, 409 Z-score model, 464–465
Source Exif Data:
File Type : PDF File Type Extension : pdf MIME Type : application/pdf PDF Version : 1.6 Linearized : No Author : Allen, Steve L.(Author) Modify Date : 2012:12:21 10:52:40-08:00 Create Date : 2012:11:26 14:45:36+05:30 EBX PUBLISHER : Wiley Page Layout : SinglePage Page Count : 609 Page Mode : UseOutlines Has XFA : No XMP Toolkit : Adobe XMP Core 4.2.1-c041 52.342996, 2008/05/07-21:37:19 Producer : Acrobat Distiller 9.4.0 (Macintosh) Metadata Date : 2012:11:27 17:09:31+05:30 Document ID : uuid:c9eb2f91-37a9-11e2-801d-001451072078 Instance ID : uuid:edc93111-7e09-49a5-a131-290f05545e6b Format : application/pdf Title : Financial Risk Management : A Practitioner’s Guide to Managing Market and Credit Risk Creator : STEVEN ALLENEXIF Metadata provided by EXIF.tools