Guide To Business Analysis Risk Analyst
User Manual:
Open the PDF directly: View PDF .
Page Count: 102
Download | |
Open PDF In Browser | View PDF |
AUGUST 2008 GUIDE TO BUSINESS ANALYSIS RISKANALYST 5.0 © 2008 Moody’s KMV Company. All rights reserved. Moody’s KMV, RiskAnalyst, Credit Monitor, Risk Advisor, CreditEdge, LossCalc, RiskCalc, Decisions, Benchmark, Expected Default Frequency, and EDF are trademarks of MIS Quality Management Corp. used under license. All other trademarks are the property of their respective owners. ACKNOWLEDGEMENTS We would like to thank everyone at Moody’s KMV who contributed to this document. Published by: Moody’s KMV Company 405 Howard Street, Suite 300 San Francisco, CA 94105 USA Phone: +1.415.874.6000 Toll Free: +1.866.321.6568 Fax: +1.415.874.6799 Author: MKMV Email: docs@mkmv.com Website: http://www.moodyskmv.com/ To Learn More Please contact your Moody’s KMV client representative, visit us online at www.moodyskmv.com, contact Moody’s KMV via e-mail at info@mkmv.com, or call us at: NORTH AND SOUTH AMERICA, NEW ZEALAND, AND AUSTRALIA CALL: +1.866.321.MKMV (6568) or +1.415.874.6000 EUROPE, THE MIDDLE EAST, AFRICA, AND INDIA CALL: +44.20.7280.8300 ASIA-PACIFIC CALL: +852.3551.3000 FROM JAPAN CALL: +81.3.5408.4250 TABLE OF CONTENTS PREFACE ..............................................................................................................................7 P.1.1 About This Guide......................................................................................7 P.1.2 Audience ..................................................................................................7 P.1.3 Typographic Conventions ........................................................................7 1 INTRODUCTION TO SINGLE BORROWER CREDIT ANALYSIS WITH RISKANALYST ....9 1.1 Borrower module and Internal Rating Models...................................................9 1.2 Facilities module and Loss Given Default...........................................................9 1.3 Archive module ..................................................................................................10 2 INTERNAL RATING MODEL METHODOLOGY ............................................................11 2.1 Fundamental Analysis methodology.................................................................11 2.2 Scorecard methodology.....................................................................................11 2.3 Ratio assessment methodology ........................................................................11 3 FUNDAMENTAL ANALYSIS METHODOLOGY ............................................................13 3.1 Assessments......................................................................................................13 3.2 Assessments with Categorized Components ...................................................13 3.3 Uncertainty.........................................................................................................16 3.4 Range .................................................................................................................17 3.5 Assessment algorithm ......................................................................................18 3.6 Calculating the Assessment..............................................................................18 3.6.1 Scoring of Inputs....................................................................................19 3.6.2 Aggregation of input distributions ........................................................20 3.6.3 Transformation of distribution..............................................................21 3.7 Calculating the Means and Standard Deviations of an Assessment ...............22 3.7.1 Real Mean: .............................................................................................22 4 SCORECARD METHODOLOGY...................................................................................25 4.1 Scorecard overview ...........................................................................................25 4.2 Meters ................................................................................................................26 4.3 Factors ...............................................................................................................26 4.4 Calculation of Scores for Numerical Factors ...................................................27 4.4.1 Scoring Factors Using Values ...............................................................27 4.4.2 Scoring Factors by Banding ..................................................................27 4.4.3 Transformation of Numerical Inputs ....................................................27 4.4.4 Special Numeric Factors.......................................................................30 4.5 Calculation of Scores for Categorized Factors.................................................30 4.6 Weighting ...........................................................................................................30 5 RATIO ASSESSMENT METHODOLOGY......................................................................33 5.1 Financial Ratios .................................................................................................33 5.2 Standard Ratio Assessment Algorithm.............................................................34 5.2.1 Algorithm for numeric factors ..............................................................35 5.2.2 Algorithm for categorized factors.........................................................35 5.3 5.4 5.5 5.6 5.7 5.8 Absolute assessment ........................................................................................36 Trend assessment .............................................................................................36 5.4.1 Slope ......................................................................................................37 5.4.2 Volatility .................................................................................................37 5.4.3 Peer Trend .............................................................................................38 Peer ....................................................................................................................39 5.5.1 Standard Ranking Algorithm ................................................................39 5.5.2 Alternative Ranking Algorithm .............................................................40 Combining Ratio components ...........................................................................41 Special Conditions .............................................................................................42 5.7.1 Assess Ratio ..........................................................................................42 5.7.2 Ratio Unacceptable ...............................................................................43 5.7.3 Trend Defaults to Good .........................................................................43 Debt Coverage Ratio Assessments ...................................................................43 5.8.1 Earnings and Cash Flow Coverage Ratios............................................43 5.8.2 The Cash Flow Management Assessments..........................................44 6 RATINGS SUMMARY ................................................................................................47 6.1 Structure of the rating .......................................................................................47 6.2 Ratings Summary screen ..................................................................................48 6.2.1 Customer State......................................................................................48 6.2.2 1-Yr EDF.................................................................................................48 6.2.3 Overrides................................................................................................48 6.2.4 Facilities.................................................................................................48 6.2.5 Constraining ..........................................................................................48 7 LOSS GIVEN DEFAULT ANALYSIS............................................................................49 7.1 Determining LGD and EL Values.......................................................................49 7.2 Facility Status ....................................................................................................51 7.3 Calculating Grades ............................................................................................51 8 LOSS GIVEN DEFAULT CALCULATIONS ...................................................................53 8.1 Calculating EADs ...............................................................................................53 8.2 Guarantees.........................................................................................................55 8.2.1 Haircuts Applied to Guarantees............................................................55 8.2.2 Eligibility Criteria for Guarantees .........................................................55 8.3 Collateral ...........................................................................................................56 8.3.1 Mapping RiskAnalyst Collateral Types to Basel II Collateral Types....56 8.3.2 Calculating Haircuts for Financial Collateral.......................................56 8.3.3 Haircuts for Non-Financial Collateral ..................................................57 8.3.4 Prior Liens and Limitations...................................................................58 8.3.5 Eligibility Criteria for Collateral............................................................58 8.3.6 Collateral Minimum LGD% Values .......................................................61 8.4 Allocating CRMs to Facilities.............................................................................63 8.4.1 Automatic Allocation of CRMs using EAD Weighting ...........................63 8.4.2 Manual Allocation of CRMs ...................................................................63 8.5 Derivation of the LGD for the Unsecured portion of the facility.......................64 8.6 Derivation OF the Borrower PD.........................................................................64 9 LOSS GIVEN DEFAULT ALGORITHM.........................................................................65 9.1 Introduction to the LGD Algorithm....................................................................65 9.2 LGD Algorithm Overview ...................................................................................66 9.3 Calculating EAD .................................................................................................67 9.4 Calculate Allocations.........................................................................................67 9.5 Derive CRM Eligibility Per Se ............................................................................68 9.5.1 Deriving CRM Eligibility Per Se .............................................................68 9.6 CRM Sorting .......................................................................................................68 9.6.1 Guarantees First....................................................................................69 9.6.2 Guarantees Second ...............................................................................70 9.7 The LGD Calculation ..........................................................................................71 9.7.1 Guarantee First/Second Processing.....................................................71 9.7.2 Guarantee First Processing ..................................................................71 9.7.3 Guarantees Second Processing ............................................................71 9.8 Perform CRM Allocation....................................................................................72 9.8.1 Mismatch-Adjusted Eligible Amount ....................................................72 9.8.2 Expected Realization .............................................................................73 9.9 Eligibility Per Facility .........................................................................................73 9.9.1 Implementation .....................................................................................73 9.9.2 Determining the Minimum Collateralization Requirement .................75 9.10 Expected Loss ....................................................................................................76 9.11 Calculate EL and LGD Data for the Facility ......................................................76 9.12 Calculate EL and LGD Data Across All Facilities .............................................77 10 FACILITY SUMMARY ................................................................................................79 10.1 Facility Summary ...............................................................................................79 10.1.1 Customer information bar ....................................................................79 10.1.2 Switching between Proposed and Existing Positions ..........................79 10.1.3 Selecting the Evaluation Date ...............................................................79 11 ARCHIVE..................................................................................................................81 11.1 Archive overview ................................................................................................81 A APPENDIX A - INTERNAL RATING TEMPLATES.......................................................83 A.1 The Internal Rating Template Concept .............................................................83 A.2 The Methodology Used to Create Internal Rating Templates..........................84 A.2.1 The Internal Rating Template Creation Process..................................84 A.2.2 Tuning the Internal Rating Template....................................................84 A.2.3 Verification of the Internal Rating Template Performance .................86 A.3 Configuring Internal Rating Templates ............................................................86 A.4 A Typical Internal Rating Template Configuration Life-Cycle .........................87 A.4.1 Model Set-up .........................................................................................89 A.4.2 Pilot Phase.............................................................................................92 A.4.3 Production .............................................................................................94 A.4.4 Some Additional Information on Configuration Tasks .........................95 A.5 Configuration versus Customization.................................................................99 A.6 Making Changes on Your Own...........................................................................99 A.6.1 Configuration Changes..........................................................................99 GUIDE TO BUSINESS ANALYSIS 5 A.7 A.6.2 A Note on Versioning ...........................................................................100 The Facilities Module and The Internal Rating Templates ............................100 PREFACE PREFACE P.1.1 About This Guide The Guide to Business Analysis is intended for users of the Internal Rating Model Author and purchasers of internal rating templates. The guide provides detailed information on performing business analyses using RiskAnalyst. Specifically, it explains the structure and calculations involved in internal rating models and loss given default. After reading this guide, one should have a good understanding of how the system arrives at internal rating model grades and loss given default values. P.1.2 Audience This guide is for Moody’s KMV clients and personnel. It assumes a good understanding of RiskAnalyst. P.1.3 Typographic Conventions This guide uses the following typographic conventions—fonts and other stylistic treatments applied to text—to help you locate and interpret information. TABLE P.1 Convention Typographic Conventions Description Bold Virtual buttons, radio buttons, check boxes, literal key names, menu paths, and information you type into the system appear in bold type — for example, “Click Add” or “Press Enter.” Courier System messages appear in a courier typeface — for example, “The system displays the following message: Added Amazon.com to your portfolio.” Italic Emphasized definitions and words appear in italic type — for example, “Portfolio Tracker is a portfolio tools page.” NOTE Information you should note as you work with the system. WARNING Warning information that prevents you from damaging your system or your work. TIP Additional information you can use to improve the performance of the system. EXAMPLE Information that illustrates how to use the system. GUIDE TO BUSINESS ANALYSIS 7 CHAPTER 1 1 INTRODUCTION TO SINGLE BORROWER CREDIT ANALYSIS WITH RISKANALYST RiskAnalyst™ optionally includes Borrower, Facilities, and Archive modules. The first two modules provide the ability to analyze a business and produce risk ratings at both the borrower and facility level. Once an analysis has been completed, the Archive module is used to create a record of that analysis. 1.1 BORROWER MODULE AND INTERNAL RATING MODELS The Borrower module uses internal rating models to analyze industry specific factors and produce a risk rating grade and PD estimate. Internal rating models can be created or customized using the RiskAnalyst Internal Rating Model Author (component of RiskAnalyst Studio). You can also use a Moody’s KMV designed internal rating template (IRT) as a starting point in producing your own internal rating model. Internal rating models are assigned to a customer in RiskAnalyst in the Analysis Setup dialog box. Access the internal rating model using the internal rating model screen. The methodology behind internal rating models is discussed in detail in Chapters 2-5. 1.2 FACILITIES MODULE AND LOSS GIVEN DEFAULT When assessing the risk associated with providing credit to borrowers, there are two dimensions to the risk assessment: the risk that the borrowers will default on their obligations and the risk associated with any recovery of the obligations from the borrowers if they default. The latter analysis is often referred to as Loss Given Default (LGD) analysis. The Facilities module assesses facility risk by calculating LGD and Expected Loss (EL). As specified by Basel II, LGD and EL are first calculated at the facility level. Each facility's LGD and EL values are aggregated to provide a total LGD and EL value. Finally, the system produces a Facility and EL grade based upon the aggregate values. Access the Facilities module through the Facility Summary screen. GUIDE TO BUSINESS ANALYSIS 9 Chapters 7, 8, and 9 discuss Facilities and the Loss Given Default calculation in detail. 1.3 ARCHIVE MODULE The Archive module stores and retrieves records of analyses. When archiving analyses, the system creates a copy of the customer data contained in the customer database. You can view a history of the customer’s archives in RiskAnalyst. These archives can later be recovered individually to support internal audit functions, as well as in batches for model evaluation purposes. Access the Archive feature through the Archive dialog box. Chapter 11 provides an overview of archiving analyses. 10 CHAPTER 2 2 INTERNAL RATING MODEL METHODOLOGY RiskAnalyst includes two possible approaches to internal rating models: Fundamental Analysis and Scorecard. Both approaches support sophisticated analysis of ratios and financial metrics. It is important to understand that the primary difference between these two approaches is the way the system calculates and scores the inputs of the model. While the methodologies of these approaches differ, they are both based on the same technology platform. Additionally, each internal rating model, no matter the approach used, produces a borrower rating and PD. The Internal Rating Model Author supports the creation and customization of internal rating models using both approaches. This chapter includes a brief description of each of the standard approaches. Subsequent chapters will explore the methodology behind each approach in detail. Please note that many of the calculations in both approaches can be overridden by a highly custom designed model. The following sections and chapters describe the standard calculations and methodologies. 2.1 FUNDAMENTAL ANALYSIS METHODOLOGY The fundamental analysis approach, like the scorecard approach, supports the analysis of key financial factors as well as non-financial, subjective factors. Compared to scorecard internal rating models, however, fundamental analysis internal rating models have a greater degree of complexity in their scoring mechanism. Assessments are the primary output of the fundamental analysis approach and the way in which inputs are scored. Assessments are produced by combining the results of many inputs to produce a number on a scale of 0 to 100. Assessment results display as this number, called the assessment mean, and also as a graphical meter. The final assessment in the model is typically mapped to a borrower rating grade and equivalent PD. Fundamental analysis assessments inherently support uncertainty and can account for missing information in the model. The methodology behind the fundamental analysis approach, that is, the way in which inputs are combined to produce assessments, is described in depth in Chapter 3. 2.2 SCORECARD METHODOLOGY The scorecard approach groups the inputs, or factors, of the internal rating model into sections. Each factor in a section is scored, and the sum of these scores becomes the total score for that section. The section scores are summed to produce an overall score, which is then mapped to the borrower rating grade and equivalent PD. Both the individual factors and the sections can be weighted. The scoring mechanism is simpler than the fundamental analysis approach since typically only sums of scores (with any weights) are considered. Inputs are not assessed on a 0 to 100 scale, but rather use the range of possible scores identified when setting up the input. Because of this, uncertainty and missing information are not supported by scorecard internal rating models. Scorecard methodology is explored more thoroughly in Chapter 4. 2.3 RATIO ASSESSMENT METHODOLOGY Ratio assessments are similar to the assessments used in the fundamental analysis approach, but have unique components and calculations. They can be incorporated in both scorecard and fundamental analysis internal rating models. Ratio assessments combine a peer, trend, and absolute assessment to calculate an overall assessment of a ratio or financial metric. The specific calculations used to produce a ratio assessment can be found in Chapter 5. GUIDE TO BUSINESS ANALYSIS 11 CHAPTER 3 3 FUNDAMENTAL ANALYSIS METHODOLOGY 3.1 ASSESSMENTS Assessments are the primary output of fundamental analysis internal rating models. To create assessments, an internal rating model combines the results of many inputs to produce a number on a scale of 0 to 100. Assessment results display as this number, called the assessment mean, and also as a graphical meter. Meters display the assessment result on a 0 to 100 scale that is divided into the following seven categories. The certainty of an assessment is also visually displayed by shaded and solid bands. See section 3.3 for more information about uncertainty. You can map this scale to your institution's scale using the Configuration Console, Tuning Console, or the Internal Rating Model Author. If you map this scale, the number on the right of the meter corresponds to the institution's scale. The inputs that make up the assessment are called components. For example, an assessment of Management Character could be made up of two components: one that measures Commitment and one that measures Integrity. The types of components that make up an assessment determine how the assessment is calculated: • Assessments with categorized components (such as subjective factors and aggregated assessments). The calculation for these assessments is described in the next section. • Ratio assessments with peer, trend, and absolute components. Ratio assessments are discussed in Chapter 5. 3.2 ASSESSMENTS WITH CATEGORIZED COMPONENTS This section discusses how the system produces assessments from components with categorized inputs. A detailed look at the algorithm behind these calculations can be found in section 3.5. These components have input values that are predefined categories, such as: Non-financial factors. Factors with input values that display as drop-down menus and have categories such as: HIGH, AVERAGE, LOW. Assessments. Individual assessments have categories (Unacceptable through Excellent) and can be combined to form an aggregated assessment. GUIDE TO BUSINESS ANALYSIS 13 To combine categorized inputs into an assessment, the system uses a weight function. The weight function serves two purposes: • It allows components with different categories to be combined in the assessment. • It allows each component category to have more or less importance in the assessment. The following discussion gives a simplified account of how the system uses weight to create an assessment. Uncertainty and soft saturation also contribute to the assessment, but to simplify this discussion they are in a separate section. Weighting Assessment Components The weight function assigns a vote to each component category. If the component category improves the assessment it is assigned a positive vote. If the category worsens the assessment it is assigned a negative vote. The relative importance of the category is defined by the size of the vote: the larger the vote, the more influence the component category has on the assessment. The weight concept in the Tuning Console provides an idea of the importance of individual components in an assessment (i.e., which components impact the assessment to a greater degree). It also allows you to adjust, at a very high level, the votes associated with a component and its categories. It is important to remember that this weight is not explicit in the fundamental analysis approach; weighting is added to an assessment through the use of votes. While you can adjust the weights of components in the Tuning Console, what you are really doing is adjusting the votes. Each component has a list of possible categories and corresponding votes. When a component receives its value, for example, through user selection, the system looks up the category to assign the votes. The sum of the votes for all of the assessment’s components contributes to the assessment value. Votes, which center on zero, must be mapped onto the assessment scale of 0–100. To do this, assessments have an initial distribution which is used as a starting point for the assessment. The initial distribution generally has a mean of 50 (on a 0-100 scale) and a standard deviation of seven. The sum of the component votes is added to the initial mean to produce the assessment mean. This value is then mapped onto the seven assessment meter categories to display the meter. Consider an assessment of Management Character with two components: Commitment and Integrity. The votes of the Commitment and Integrity components are summed and the result is added to the initial mean to produce the assessment mean. The diagram below demonstrates this process. Note that because uncertainty in the assessment is taken into account, as discussed in section 3.3, the assessment has a shaded band to the left of the solid band. 14 The same logic applies when aggregating a group of assessments into an overall assessment. The example below details an assessment with 266 absolute votes to be distributed among five components. The components have different numbers of categories (some 4 and some 5). The 266 votes used in this example are based on a Range value of 59. Category 1 2 3 4 5 Total Absolute Vote Average Absolute Vote Votes for Component 1 -5 20 -10 -20 N/A 55 Votes for Component 2 0 20 -10 -30 N/A 60 Votes for Component 3 12 8 0 -15 -25 60 Votes for Component 4 10 5 0 -8 -13 36 Votes for Component 5 12 8 0 -10 -25 55 (55/4) 13.75 (60/4) 15 (60/5) 12 (36/5) 7.2 (55/5) 11 The total of the Average Absolute votes is: 13.75 + 15 + 12 + 7.2 + 11 = 58.95 Average Absolute Vote as a percentage of 58.95: 23.3% 25.4% 20.4% 12.2% 18.7% Total 100% Adjusting the weight (through the use of votes) of an individual component affects only that component. It is necessary, therefore, to adjust other weights in order to achieve a total of 100% weight across all components. GUIDE TO BUSINESS ANALYSIS 15 Initial Distribution Each assessment has an initial distribution which is the mean and standard deviation of the assessment. Generally, the starting mean is set halfway along the meter at 50, with the standard deviation of seven. Positive voting moves the meter up, and negative voting moves it down. A meter mean point of 50 adds no bias to the assessment. A meter starting mean point value of less than 50 will ensure a built-in conservative bias to the assessment, a value of greater than 50 will ensure an optimistic bias. 3.3 UNCERTAINTY As well as the assessment value, meters display RiskAnalyst’s analysis of the certainty or uncertainty attached to the assessment. Uncertainty can be introduced as a result of: • Missing data • Inherent imprecision of the items under assessment, particularly where these are subjective analyses Meters use solid and shaded bands to indicate both the assessment value and the level of certainty attached to the analysis. A meter displays as either a single, solid green band or a combination of a solid and shaded green bands. The band or bands denote a range with a 90% degree of certainty. The system will color in the smallest number of categories it can to cover 50% of the curve, and then will shade in the smallest number of categories to the left and right to reach 90% of the curve. Where a solid band appears without any shaded bands, the range denoted by the solid band has a 90% degree of certainty. Where a solid band is accompanied by shaded bands, the range denoted by the solid and shaded bands has a 90% degree of certainty, but the range denoted by just the solid band is only 50% certain. Inbuilt Uncertainty The inbuilt uncertainty of a component is one of the determinants of the width of the uncertainty band in the component’s meter. A high uncertainty indicates an inherently uncertain component, and means that, even when all the inputs are given, the meter displays a wide shaded band. Soft Saturation Although meter values theoretically range from 0 to 100, in practice it is usually impossible to achieve either extreme. This is due to a type of damping mechanism—soft saturation—that ensures meters stay within range by applying increasing resistance as the value approaches either extreme. Values lower than about 3 or higher than about 97 would be rare. Calculation of Uncertainty The uncertainty for an assessment is the square root of the sum of: 16 The standard deviation squared The square of the standard deviation associated with each input applied to its weights. The latter is derived from the difference between the input's mean value (as derived using the assessment votes) and the votes for each category in the assessment. The resulting differences are combined by weighting each difference by the probability that the input could take the value associated with the weight. The differences are also squared and summed. Then the system takes the square root of the result to eliminate positive and negative differences canceling out each other. The spread of votes you enter for an assessment will affect the uncertainty if the input itself has some uncertainty. This is due to the use of differences between votes and input mean. Using a greater spread of votes for meters will affect assessments, especially when questions are unanswered. Note that if the input to the assessment is a specific answer to a question then the spread of votes has no impact as its value is certain. 3.4 RANGE The range is the number that controls the boundaries of an assessment and thus its impact – the greater the range the greater the potential movement of the meter (14 votes is equal to approximately one category of the meter). Range sets the maximum number of votes for the components of an assessment, and changing it changes the votes for each component of the assessment, resetting them to a straight line and overriding the existing distribution. Similar to weights in the Tuning Console, changing the range is really just another way to change the votes of the components of an assessment. The range will be changed if tuning (see ‘Range in the Tuning Console’) changes the total number of votes within the components. The highest and lowest possible values for an assessment are determined (indirectly) by a factor using the range. The actual number of votes is determined using the formula explained below. This calculation produces votes on a straight line through a midpoint of 0. It must be carried out for each category in each component using the correct formula (determined by whether a component has an equal or odd number of categories). Weight (as a percentage) = p Number of categories in a component = n Range = r Category number (ID) = c Number of votes = v For each category (c), the vote (v), is given by the following equations: For even values of n: v(c) = 2(npr) * (2c - n - 1) / n2 For odd values of n: v(c) = 2(npr) * (2c - n - 1) / (n2 - 1) For example, consider the Operations Skill component category ABOVE AVERAGE. When Operations Skill has a weight of 24.5%, and the range is 39, the number of votes for the ABOVE AVERAGE category is calculated as follows: p = 0.245, n= 5, c = 4 GUIDE TO BUSINESS ANALYSIS 17 v4 = 2(5 x 0.245 x 39) x (8-5-1)/25 = 95.55 x 2/25 =8 Range in the Tuning Console In the Internal Rating Model Author, you can use the Tuning Console’s Range slider to adjust the total number of votes that contribute to the assessment. Changing the range overwrites the distribution of votes in existing assessments by setting them in a straight line that centers on zero. When range is set to a low value, the system allocates fewer votes between all component categories for the assessment and the resultant meter range is narrower. Conversely, when you set a high value for range the system allocates more votes, the resultant meter range is broader, and the extremes of the meter can be reached in the assessment. Since there is a larger potential movement along the meter between best and worst, the meter has a greater possible impact. The table below displays how the system allocates votes to components using three different range settings: Range Setting 20 40 60 3.5 Votes for Component 1 (50% weight) -15 to +15 -30 to +30 -45 to +45 Votes for Component 2 (30% weight) -9 to +9 -18 to +18 -27 to +27 Votes for Component 3 (20% weight) -6 to +6 -12 to +12 -18 to +18 Assessment Meter (Worst/Best) 23.1 / 76.9 8.3 / 91.7 2.7 / 97.3 Potential Meter Range (/100) 53.8 83.4 94.6 ASSESSMENT ALGORITHM This section describes the main assessment algorithm of the fundamental analysis approach. This algorithm is used to derive all the assessments except for those that assess ratio values. However, even these assessments use the main assessment algorithm as a building block. In order to derive an assessment, each value that an input can take is associated with a vote as described above. A vote denotes the “score” of an input for a given value of that input. If the input takes a numeric value, this value is called a fixed point. If the input takes a category (i.e., Excellent, Good, Fair) as a value, this value is called a category. A set of fixed point – vote pairs or category – vote pairs defines a relative function of “goodness” for the various values of an input and is defined for all inputs. For categorised inputs a vote is typically defined for each possible category. For numeric inputs it would clearly not be possible to specify a vote for each possible value and so a number of votes are specified, each corresponding to one possible value of the input (the fixed point). If the actual value of the input does not match one of the fixed points, linear interpolation is used to determine the corresponding vote. If the input value is greater than the largest fixed point the vote corresponding to this is used. Similarly, if the input value is less than the smallest fixed point the vote corresponding to this is used. 3.6 CALCULATING THE ASSESSMENT Assessments are created using the following steps: 1. 18 Scoring of inputs: For each input the system calculates the mean and standard deviation for a Normal distribution from the set of fixed points or categories associated with this input and the value of this input. The fixed point to vote or category to vote mapping described above is used to determine the score distribution. Fixed points or categories are defined during the design and tuning of a model, and can later be adjusted using the Internal Rating Model Author. 2. Aggregation of input distributions: RiskAnalyst assumes the probability of an assessment is Normally distributed. The total votes of all inputs are aggregated to calculate the mean and standard deviation of this distribution. 3. Transformation of distribution: The final stage is to transform the Normal distribution from the continuous domain on an infinite scale to fit our discrete seven-category assessment on a 0-100 scale. This is done by calculating areas under the standard Normal distribution that correspond the range of each of the categories on the infinite scale. 3.6.1 Scoring of Inputs The methods used to score inputs are slightly different depending on whether they are categorized or numeric. The following sections provide the algorithms for determining the scores for each input. Categorized Inputs Let categorised input x have n categories where 1 ≤ n ≤ 7 . Let: • c • p • v • μ ix ix ix x denote the i-th category value of x . denote the probability associated with the i-th category value of x . denote the votes associated with the i-th category value of x . denote the mean of the Normal distribution associated with x , where: μ = ∑ p .v x • σ ix ix denote the standard deviation of the Normal distribution associated with x , where: x σ • i≤n x ∑ p .(vix − μ x )2 = If the value of x is unknown, i.e. μ σ x ix i≤n x = p ix is undefined, then: max(vix ) + min(vix ) 2 ( = 0.2887. max(vix ) − min(vix ) ) i.e. the mean and standard deviation of the uniform distribution is assumed. GUIDE TO BUSINESS ANALYSIS 19 Numeric inputs Let numeric input x have n fixed-points where 1 ≤ n ≤ 7 . Let: • u • v • w denote the value of x . • μ ix denote the i-th fixed-point value of x . ix denote the votes associated with the i-th fixed point value of x . x x denote the mean of the Normal distribution associated with x , where: ⎧vnx ⎪ ⎪v1x ⎪ μ x = ⎪⎨v jx ⎛ ⎪ ⎜ w x −u jx ⎪v + ⎜ ⎪ jx ⎜ u kx − u jx ⎪⎩ ⎝ ⎞ ⎟ ⎟. vkx − v jx ⎟ ⎠ ( ) if if if w ≥u w ≤u w =u if u x nx x 1x x jx jx for 1 ≤ j ≤ n ≤ wx ≤ u kx for k = j + 1 This is often referred to as linear interpolation. σ Note that typically there is no uncertainty associated with numeric inputs ( x = 0 ). The exception is when the input is the slope where it is possible to measure the standard deviation of the slope. See section 5.4.1 for details. When there is standard deviation associated with the input then the standard deviation, denoted by σ , is calculated by considering the average difference of the vote around the mean for the standard deviation: σ Where = x μ x +σ is calculated as μ x μ x +σ − μ x + μ x −σ − μ x 2 w with x +σ = μ + σ . Similarly for x μ x−σ . If x is unknown the mean and standard deviation are calculated using the same method for categorised inputs described above. In RiskAnalyst, this is not used in practice since the ratio assessment algorithm only assesses defined ratios. 3.6.2 Aggregation of input distributions Let the assessment being constructed be denoted as y . Let pw denote an initial distribution associated with y . Let μ ,σ pw pw denote the mean and standard deviation of this initial distribution. Given m inputs, the mean of the Normal distribution associated with y is calculated as: μ =μ y and the standard deviation as: 20 pw + ∑μ 1≤ x ≤ m x σ 3.6.3 y = σ 2 pw + ∑σ 1≤ x ≤ m 2 x Transformation of distribution c for 1 ≤ i ≤ 7 denote the categories of an assessment, where c = UNACCEPTABLE Lc = EXCELLENT . Let cep denote the class end point of Let i 1 7 i the i-th assessment category on a 0 to 100 scale, i.e., cep = i Let i.100 7 iep denote the interval end point of the i-th assessment category on a − ∞ to ∞ scale. i The interval end point is calculated from the class end point using the following inverse sigmoid function: ⎞ ⎛ cep ⎜ ⎟ i 50 25 . ln = + iepi ⎜ 100 − cep ⎟ i⎠ ⎝ Let Z iy denote the Z-score for the i-th category of an assessment. Z Let F ( Z iy iy = iep − μ i σ Z iy is calculated as: y y ) denote the probability that the value of y is in one of the j categories up to and including i ( 1 ≤ j ≤ i ). Standard lookup tables can be used to determine F ( Z iy ) . F ( Z iy ) corresponds to the area underneath the standard Normal distribution to the left of Z iy . These areas correspond to the probability that the value of the assessment is contained in any of the first i intervals. For efficiency reasons RiskAnalyst uses an approximation method for finding the area under the Standard Normal distribution. It uses linear interpolation with the following values: GUIDE TO BUSINESS ANALYSIS Mean Cumulative Integral -11 0 -10 0 -3.5 0.0002 -3 0.0013 -2.5 0.0062 -2 0.0228 -1.5 0.0668 -1.2 0.1151 -1 0.1587 -0.8 0.2119 -0.5 0.3085 0 0.5 0.5 0.6915 21 0.8 0.7881 1 0.8413 1.2 0.8849 1.5 0.9332 2 0.9772 2.5 0.9938 3 0.9987 3.5 0.9998 10 1 Finally the probability of the i-th category for y , probability F ( Z iy σ y , can be calculated from the cumulative for i = 1 ⎧⎪ F ( Z iy ) =⎨ iy ⎪⎩ F ( Z iy ) − F ( Z ( i −1) y ) for 2 < i ≤ 7 = 0 there is no uncertainty and the mean of the distribution fits entirely into one of the categories. For p 3.7 iy ): p Note that if p σ y =0, p iy is calculated as: ⎧⎪1 for the smallest value of i such that μ < iep y i =⎨ iy ⎪⎩0 Otherwise CALCULATING THE MEANS AND STANDARD DEVIATIONS OF AN ASSESSMENT While the assessment meters provide a powerful way of communicating the evaluation of the assessment, they are limited in that they are complex to describe and are not particularly precise. Often it is desirable to have a single figure that describes the assessment. To satisfy this requirement, we use the assessment mean, or Real Mean. The Real Mean is associated with the mean of the original continuous Normal distribution, and is defined further in section 3.7.1. When combining assessments, the fundamental analysis approach uses the full information from the sub-assessments, and doe not simply combine the real means of the sub-assessments. 3.7.1 Let Real Mean: meanry denote the Real Mean. This is the mean of the continuous Normal distribution transformed to the 0-100 scale using the RiskAnalyst saturation function and is calculated as: ( ) 100 . e mean ry = ( ) 1+ e μ y −50 25 μ y −50 25 22 Let sd yr denote the real standard deviation of an assessment. It is calculated as: ( ) ⎞⎟ ⎟ − 50 ( ) ⎟⎟ σy ⎛ ⎜ 100.e 25 r ⎜ sd y = ⎜ σy ⎜ 1+ e 25 ⎝ GUIDE TO BUSINESS ANALYSIS ⎠ 23 CHAPTER 4 4 SCORECARD METHODOLOGY 4.1 SCORECARD OVERVIEW A scorecard internal rating model consists of a set of factors grouped into sections corresponding to particular areas of analysis. Factors are inputs obtained either from the user or from values entered in other parts of RiskAnalyst, or from external programs. Each factor within a section is scored, and the sum of those scores, with any weighting, is the total score for that section. The sum of each section's total score, with any weighting, is the scorecard's total score. FIGURE 1.1 Section total Each factor within a section can be weighted such that it contributes more or less to the section total than the other factors within that section. Section total scores can also be similarly weighted to have a greater or lesser influence on the total scorecard score. The Internal Rating Model screen displays section scores and ranges with factor weighting, but without section weighting, applied. The Ratings Summary screen shows the section scores and ranges, as well as the total score, with weighting applied. FIGURE 1.2 Total score GUIDE TO BUSINESS ANALYSIS 25 To reiterate, the scoring process is as follows: • Scorecard factors obtain data from user input or from values entered in other parts of RiskAnalyst. • Factors are given a score based on the input value. • Factor scores are weighted and summed to give a section score. • The section scores are weighted and summed to give a scorecard total. • The scorecard total is mapped to a grade and a PD. 4.2 METERS Once the factors are calculated and the section scores totaled, RiskAnalyst displays a meter to give a visual representation of the score. The meter includes both the score and the range of scores possible. The system displays meters for section totals, as well as the total score. FIGURE 1.3 4.3 Scorecard meter FACTORS Scorecard inputs are known as factors and have the following input sources: • User input. Factors that display in scorecards as drop-down menus or input boxes and receive their values from user input. • RiskAnalyst Values. Values entered in other parts of RiskAnalyst. These inputs can be customer information data, ratios, values, or RiskCalc and Public EDF measures. • External Program Values. Values obtained from an external program. A module is necessary to obtain values from an external program and must be configured for use in the Internal Rating Model Author. Input values can be: • Numerical. Values such as ratios or numbers entered by the user. • Categorized. Predetermined categories, such as drop-down menus, where users select from a list of possible values. Each factor is given a score based on its input value. The system calculates numerical factor scores differently than categorized factor scores. The system can also transform numerical input values before they are given a score. An explanation of each calculation, and the RiskAnalyst algorithm for these calculations, follows. Each factor has a name and belongs to a section. To find out more about the individual factors associated with a particular section, click the Clarify button on the Internal Rating Model screen for that section. 26 4.4 CALCULATION OF SCORES FOR NUMERICAL FACTORS Factors that have numerical inputs, such as ratios and financial statement values, are scored in one of two ways. Either RiskAnalyst uses the factor’s value as the score, or it calculates the score using a process called banding. 4.4.1 Scoring Factors Using Values Using this method, the value of the factor feeds directly into the score. In order to display the correct score and range for the section on a meter, there must be a minimum and maximum value for each factor’s input that is scored in this manner. RiskAnalyst uses these values to determine the lower and upper meter values, and displays the section total on this scale. 4.4.2 Scoring Factors by Banding Using this method, a number of bands (or numerical ranges) and a score for each of these bands must be specified: Band upper point Score 20 40 60 80 100 0 5 10 20 30 RiskAnalyst calculates the factor’s score by determining which band the factor’s input value falls into. If the value is larger than the highest upper bound, it is scored as if it were within that band. For example, if the value is 120, a score of 30 is given using the bands shown above. Any value below 20, including a negative number, is given a score of 0. Band upper points are inclusive, such that a value of 40 would score 5. When banding is set up, it assumes that you are using actual units. NOTE If you are entering currency values, they should be entered in the currency you have selected for your scorecard. 4.4.3 Transformation of Numerical Inputs RiskAnalyst can be configured to transform the values of scorecard factors that have numerical inputs. RiskAnalyst transforms the factor values before calculating their scores. RiskAnalyst can apply the following three transformation functions to numerical inputs: • Normalize • Log (base e) • LOGIT Each of these functions is described below. Normalise Use Normalise to standardize raw data. GUIDE TO BUSINESS ANALYSIS 27 To standardize raw data χ, subtract mu μ and divide by sigma σ (the parameters of the normal distribution), where: μ is the mean or typical value of χ. σ is the standard deviation. This is a measure of the spread or volatility of the data, and is equal to the square root of the arithmetic mean of the squares of the deviations from the arithmetic mean of the data. The Normalize function gives the data a bell-shaped distribution that is symmetrical about the mean (or typical value) of the data, as shown below. Approximately 95% of the data will be within ± two standard deviations from the mean, and approximately 99.8% of the data will be within ± three standard deviations from the mean. Example: A company has reported a Gross Margin of 45%. If the mean industry Gross Margin is 25% and the standard deviation is 10, then the standardized value of the company’s Gross Margin figure is: ((45-25)/10) = 2 This tells us that the company’s Gross Margin is two standard deviations above the mean industry figure. Log (base e) The natural logarithm (log in the base e) gears down the importance of higher data values and reduces the data’s spread. In the graph , there is a value a (approximately 2.718282) at which the gradient of the exponential curve at (0,1) is exactly 1. This is an irrational number symbolized by e. In the expression , e is called the base and y is called the exponent, power, or logarithm. Therefore, the expression can also be interpreted as meaning y is the logarithm of the or ln χ. number χ in the base e. This is written as 28 The graph of is shown above. This graph shows that as the value of data χ increases, the gradient of the curve decreases. Therefore, the impact of higher numbers is disproportionately reduced, or geared down. NOTE You cannot enter negative values into logs. Example: Take, for example, an internal rating model that has questions resulting in the following scores: 10 50 90 1000 These results give a data spread of 990, and the highest score is 100 times that of the smallest figure. Transforming the scores with log in the base e results in the following scores: 2.303 3.912 4.499 6.908 Now the spread of data is just 4.6, and the highest score is just 3 times that of the smallest score. LOGIT Use LOGIT to estimate the conditional probability of a positive response or the presence of a characteristic. For example, is a company likely to honor its loan covenants or not? The function is based on an “S” shaped curve, with the result of the transformation formula: This gives the probability of a positive response that falls between zero (no chance of a positive response) and one (100% chance of a positive response). You need to define the LOGIT coefficients A and B. GUIDE TO BUSINESS ANALYSIS 29 4.4.4 Special Numeric Factors The following are two examples of numeric factors that are calculated differently than other numeric factors. Ratios Ratio assessments are a special type of numeric factor, and are calculated by the RiskAnalyst Ratio Assessment algorithm. The calculations behind this algorithm are described in Chapter 5. NOTE It is possible to include a Ratio in a scorecard simply as any other financial value. In this case, the Ratio Assessment algorithm is not used. EDF measures EDF measures are another special instance of a numeric factor. RiskAnalyst gathers EDF values TM ® from Moody's KMV RiskCalc and CreditEdge . Whether or not an EDF value is a factor in the scorecard, the 1-Yr EDF measure will display on the Ratings Summary screen. See Chapter 6 for more information. 4.5 CALCULATION OF SCORES FOR CATEGORIZED FACTORS For categorized factors such as drop-down menus, RiskAnalyst calculates factor scores from the categories. Specify a score for each of the categories in the factor value properties, as shown in the example below: Categories Scores INADEQUATE BELOW AVERAGE AVERAGE ABOVE AVERAGE EXCEPTIONAL -10 0 10 20 30 RiskAnalyst calculates the score when a category is selected, such as when a user makes a menu selection. 4.6 WEIGHTING Weighting can be applied to both factors and sections in scorecards. There are various practical reasons for using weighting in scorecards. The following is an example that uses weighting to replicate an averaging approach. 30 Suppose you want the total score for each section to be an average of the factors in the section. You also want the total score to be the average of the section scores. Section 1 Value Factor 1 3 Factor 2 4 Factor 3 5 Factor 6 6 Section total Average of (3, 4, 5, 6) = 4.5 Section 2 Value Factor 1 5 Factor 2 6 Section total: Average of (5, 6) = 5.5 Scorecard total: Average of (4.5, 5.5) = 5.0 To produce the above result in RiskAnalyst, you would set the following factor and section weights: Factor Weights Section 1 Value * Weight = Weighted Score Factor 1 3 .25 .75 Factor 2 4 .25 1 Factor 3 5 .25 1.25 Factor 4 6 .25 1.5 Section total: 4.5 Section 2 Value * Weight = Weighted Score Factor 1 5 .5 2.5 Factor 2 6 .5 3 Section total: 5.5 Section Weights Scorecard Value * Weight = Weighted Score Section 1 4.5 .5 2.75 Section 2 5.5 .5 Scorecard total: 2.25 5.0 The weights above correspond to 1 / (number of factors) or 1 / (number of sections). This is a simple example of using weighting to produce averaging. Using the above example as a starting point, you could also produce a weighted average, such that a particular section has a greater impact on the total score (i.e., replace Section 2's weight of .5 in the example above with 1.5 to make Section 2 three times as important). GUIDE TO BUSINESS ANALYSIS 31 CHAPTER 5 5 RATIO ASSESSMENT METHODOLOGY 5.1 FINANCIAL RATIOS Both the fundamental analysis and scorecard approaches make use of ratio assessments. Financial ratios or metrics are used to form ratio assessments. RiskAnalyst performs different analyses for each ratio. The different analyses take into account a comparison of each ratio to a peer benchmark, the direction and volatility in the trend of the ratio, and in several ratios, a comparison to a policy-based absolute benchmark. These analyses are then combined to produce an overall ratio assessment. All the information displayed relating to the ratios comes from financial values in RiskAnalyst. This information consists of the calculated ratios and the peer benchmarks, and is dependant on the financial periods and the peer group. Ratio Assessment Analyses A ratio assessment is the weighted sum of a calculated peer/trend assessment and an absolute assessment. The trend assessment also includes a volatility analysis component. The absolute analysis is performed for certain ratios only. The diagram below gives an overview of the various analysis components that may be involved in the ratio assessment and how the components are combined together to give the final ratio assessment. FIGURE 1.4 Structure of ratio assessment The Peer Trend assessment is included in the Internal Rating Model Author as a potential analysis component of ratio assessments. However, the ratio assessments included with the RiskAnalyst internal rating templates do not include the Peer Trend assessment. NOTE Each of the sub-components of the analysis is described in more detail below. Component Description Peer Assessment This is an assessment of the ratio’s performance compared to industry values. The calculated value of the ratio for the current period is compared to the industry values to derive the company’s ranking. Trend Assessment The overall trend assessment is itself the weighted combination of the trend analyses and the volatility assessment of this trend. Absolute Assessment For some of the ratios, there are certain standards common in the lending community above or below which it can safely be said that the company is performing well or poorly without regard to the rest of the industry. These standards of ratio performance are used in the ratio assessment to counter the effects of making ratio comparisons in a very strong or weak industry. The absolute values used for this analysis may vary depending on the financial template used to analyze the company. Raw Trend Assessment This assessment considers the trend of the ratio over the periods analyzed. Remember that trend analysis will only be performed if sufficient historical statements are available. Typically, a minimum of three historical values are GUIDE TO BUSINESS ANALYSIS 33 Component Description required in order to provide a trend analysis. Comparative analyses, like % Turnover Growth, require a minimum of four historical statements Volatility Analysis The volatility component is an extension of the trend analysis. Generally, a ratio trend that is either stable or increasing/decreasing at a consistent rate can be analyzed with greater certainty than ratios that fluctuate from period to period. If the historical trend of the ratio shows a pattern of wide fluctuation, the volatility assessment will be unfavorable. Conversely, if the trend of the ratio shows stability or a relatively even pattern of change, the volatility assessment will be favorable. However, even an extremely favorable volatility assessment will have a relatively small impact on the overall trend assessment. The main function of the volatility analysis is to worsen the overall trend assessment. Peer Trend Assessment This assessment compares the average rate of change (slope) of the ratio value for the company over the past three or four years to the average rate of change (slope) of the median ratio value for the company's peers. To use this assessment, the average rate of change (slope) for each peer group needs to be calculated, and this information must be populated in the PEERINFO reference table. 5.2 STANDARD RATIO ASSESSMENT ALGORITHM This section describes the standard ratio assessment algorithm. It is used for all historical ratio assessments. The algorithm has three main components: 1. Absolute: Compares the most recent ratio values to a set of benchmarks. 2. Trend: Looks at how the ratio values for this borrower have changed over time. Derived from two to three sub-components (see below). 3. Peer: Ranks the ratio values against the borrower’s peer group. Trend is itself determined from between two and three further sub-components: • Slope: An assessment of the slope of recent ratio values. • Volatility: An assessment of the volatility of recent ratio values. • Peer Trend: Compares the slope of the borrower’s ratio values with those of its peer group. It is rare for this information to be available and this part of the ratio assessment is not often used. FIGURE 1.5 Components of the Standard Ratio Assessment The three top level ratio assessment components (absolute, trend, and peer) are combined using a simple weighted sum. This section details each of these components, how they are combined, and documents a number of special conditions used in the ratio assessment algorithm. 34 By default, RiskAnalyst bases its ratio calculations on statements listed in the Analysis Setup screen in RiskAnalyst (generally the latest annual, historical statements). Other annual statements can be selected by hiding statements in the RiskAnalyst grid. Projected statements can be used in the analysis by selecting the projection to use in the Analysis Setup dialog box. See the RiskAnalyst Help for more information. For slope and volatility calculations, RiskAnalyst uses values from prior statements. The two algorithms below, discussed in 5.2.1 and 5.2.2, are used in the main ratio assessment algorithm, and will be referenced in the following sections. 5.2.1 Algorithm for numeric factors Let numeric input x have n fixed points where 1 ≤ n ≤ 7 . Let: • u • v • w denote the value of x . • μ ix denote the i-th fixed point value of x . ix denote the votes associated with the i-th fixed point value of x . x x denote the mean of the Normal distribution associated with x , where: ⎧vnx ⎪ ⎪v1x ⎪ μ x = ⎪⎨v jx ⎛ ⎪ ⎜ w x −u jx ⎪v + ⎜ ⎪ jx ⎜ u kx − u jx ⎪⎩ ⎝ ⎞ ⎟ ⎟. vkx − v jx ⎟ ⎠ ( ) if if if w ≥u w ≤u w =u if u x nx x 1x x jx jx for 1 ≤ j ≤ n ≤ wx ≤ u kx for k = j + 1 This is often referred to as linear interpolation. σ Note that typically there is no uncertainty associated with numeric inputs ( x = 0 ). The exception is when the input is the slope where it is possible to measure the standard deviation of the slope. See section 5.4.1 for details. When there is standard deviation associated with the input, then the standard deviation, denoted by σ , is calculated by considering the average difference of the vote around the mean for the standard deviation: μ x +σ − μ x + μ x −σ − μ x σx= μ x +σ is calculated as μ x 5.2.2 with 2 w x +σ = μ + σ . A similar calculation is used for x μ x−σ . Algorithm for categorized factors Let categorized input x have n categories where 1 ≤ n ≤ 7 . Let: • c ix denote the i-th category value of x . GUIDE TO BUSINESS ANALYSIS 35 • p • v • μ ix ix x denote the probability associated with the i-th category value of x . denote the votes associated with the i-th category value of x . denote the mean of the Normal distribution associated with x , where: μ = ∑ p .v x • σ ix ix denote the standard deviation of the Normal distribution associated with x , where: x σ • i≤n x ∑ p .(vix − μ x )2 = If the value of x is unknown, i.e. μ σ x ix i≤n x = p ix is undefined, then: max(vix ) + min(vix ) 2 ( = 0.2887. max(vix ) − min(vix ) ) i.e., the mean and standard deviation of the uniform distribution is assumed. 5.3 ABSOLUTE ASSESSMENT RiskAnalyst assesses a ratio value against a set of benchmark fixed points. These fixed points are defined during modelling and are entered in the Internal Rating Model Author. The absolute assessment is created using the algorithm for numeric factors defined in section 5.2.1. The intial distribution for the absolute assessment is a mean of 50 and standard deviation of 0. 5.4 TREND ASSESSMENT RiskAnalyst assesses the trend of the ratio by considering three components. These are combined using the algorithm for categorized factors described in section 5.2.2. The intial distribution for the trend assessment is a mean of 50 and standard deviation of 3. We need to deal with a ratio over a number of data points. For the data points, let i denote an integer associated with the date of the ratio, where i =1 is the earliest data point. Let t denote a number specifying the point in time to which the statement date refers. If the statements are all for twelve month periods, then i =t. For example: i t Statement date 1 1 12/2000 2 2 12/2001 3 3 12/2002 4 4 12/2003 If statements are for periods other than twelve months, t will be a non-integer for some periods. For example: 36 i t Statement date 1 1 12/2000 2 2 12/2001 3 2.75 9/2002 4 3.75 9/2003 Let n denote the number of ratio values and xi denotes the ratio value at the i-th data point. For each ratio, it is possible to choose the maximum and minimum values that n can take. RiskAnalyst will only consider a maximum of five data points. If the number of data points is greater than the maximum, only the latest n ratio values are used. If the number of data points is less than the minimum, the trend component is not considered in the assessment of the ratio. Except where otherwise stated, the summations below all sum over (n − max_nos+ 1) < i ≤ n , where max_nos is the specified maximum number of periods to be used for the ratio. 5.4.1 Slope Let sloperatio denote the fitted slope of the ratio values, where: slope ratio = n∑ t i xi − ∑ xi ∑ t i ( ) n∑ t i − ∑ t i 2 2 As mentioned in the algorithm for numeric factors in section 5.2.1, RiskAnalyst also calculates the standard deviation of the slope as an input into the slope assessment. This is one-twelfth of slope denote the standard deviation of the the Residual Standard Deviation calculation. Let σ ratio slope, where: ( n∑ t x − ∑ x ∑ t ) i ⎛⎜ n ⎞ − (∑ x ) ⎟ − ∑ x i ⎠ ⎝ n∑ t − (∑ t ) i (n − 2)⎛⎜ n∑ t − (∑ t ) ⎞⎟ i ⎠ ⎝ 2 2 2 i slope σ ratio = 1 12 i i i 2 2 i 2 2 i RiskAnalyst assesses sloperatio against a set of benchmark fixed points. These fixed points are defined during modelling and are entered with the tuning tool. The slope assessment is created using the algorithm for numeric factors defined in section 5.2.1, using both sloperatio and slope σ ratio . The intial distribution for the slope assessment is a mean of 50 and standard deviation of 3. 5.4.2 Volatility Let slopei denote the slope between the i-th and i-1 ratio: slopei = x −x t −t i i i −1 ,2 ≤ i ≤ n i −1 Let ad ratio denote the average difference between consecutive pairs of ratio values, where: GUIDE TO BUSINESS ANALYSIS 37 ad ratio = Let vi ratio vi ∑ (slope ) i 2≤i ≤ n (n − 1) denote the volatility index for a given ratio, where: ratio = 2 ⎛ ⎞ ⎛ ⎞ ⎜⎜ ∑ ⎜ slope − ad ⎟ ⎟⎟ ratio i ⎠ ⎠ ⎝ 2 ≤i ≤ n ⎝ ( n − 1) ∑x 1≤i ≤ n i n vi against a set of benchmark fixed points. There are three fixed points RiskAnalyst assesses ratio defined in the system for HIGH, MEDIUM, and LOW volatility groups. The group associated with a particular ratio is defined during modelling and entered with the Internal Rating Model Author. The volatility assessment is created using the algorithm for numeric factors defined in section 5.2.1. The intial distribution for the volatility assessment is a mean of 50 and standard deviation of 3. 5.4.3 Peer Trend Let ptratio denote the difference between the fitted slope of the ratio values and the peer trend, where: peer ptratio = sloperatio − trend ratio peer trend ratio is the measured trend of a ratio given a peer group. This value is entered directly into the RiskAnalyst reference tables. Due to the difficulty in maintaining this data, the peer trend component has no impact on the trend assessment in the standard models shipped with RiskAnalyst. Since there is a standard deviation associated with the slope, this is passed to the peer trend pt , and it is the same as the standard deviation used in the slope assessment. It is denoted as σ ratio calculation, i.e. pt slope σ ratio = σ ratio RiskAnalyst assesses ptratio against a set of benchmark fixed points. These fixed points are defined during modelling and are entered with the tuning tool. The slope assessment is created using the algorithm for numeric factors defined in section 5.2.1. It uses both ptratio and pt σ ratio as inputs. The intial distribution for the peer trend assessment is a mean of 50 and standard deviation of 3. 38 5.5 PEER RiskAnalyst estimates the peer rank from a ratio value and three quartile cut-off points. There are two different peer rank algorithms. The standard algorithm is used to assess the peer rank under most circumstances. An alternative algorithm is used for gearing/leverage ratios if the third quartile is negative. RiskAnalyst can also be configured so that the alternative ranking algorithm can be used by other ratios (possibly new ones). Let Q1 , Q 2 , and Q3 denote the first, second, and third quartile cut-off points for a given ratio th th th and peer respectively. Note that these correspond to the 75 , 50 , and 25 percentiles. Depending on the ratio, the Q1 ratio value may be smaller or larger than the Q3 value. This 1 depends on whether the ratio is deemed to be ascending or descending . Let rank denote the estimated rank of the ratio against its peer group, where: ⎧ altRank if Q3<0 and ratio is a gearing ratio rank = ⎨ otherwise ⎩stdRank stdrank and altrank are defined below. RiskAnalyst assesses rank against a set of benchmark fixed points. These fixed points are defined during modelling and are entered with the tuning tool. The peer assessment is created using the assessment algorithm for numeric factors defined in section 5.2.1. The initial distribution for the peer assessment is a mean of 50 and a variable standard deviation depending on the sample size of the peer: Sample size greater than or equal to: Initial Distribution Standard Deviation 100 3 10 5 Default 7 Note that the sample size information must be configured manually in the database tables as this is not done automatically by the benchmark tool. In RiskAnalyst, this information is added to the fpRatioPeerSample table. 5.5.1 Standard Ranking Algorithm Let stdrank denote the standard rank, where: 1 if (Q3 > Q1 and x > Q 2) • ⎧2( r −1) ⎪ stdrank = ⎨2( r −1) ⎪1 − 2( r −1) ⎩ • x is the ratio value • (Q 2 − x ) ⎧ ⎪⎪ max(Q3, Q1) − Q 2 if x > Q 2 r=⎨ (Q 2 − x ) ⎪ Otherwise ⎪⎩ min(Q3, Q1) − Q 2 if (Q3 < Q1 and x < Q 2) Otherwise For an ascending ratio, we deem higher values to be generally better. The inverse applies to descending ratios. GUIDE TO BUSINESS ANALYSIS 39 This algorithm creates a smooth, S-like mapping from ratio values to peer ranks. For example, for an ascending ratio that takes values in (0,1) with quartiles at 0.25, 0.5, and 0.75, the mapping can be described by the following diagram: 100.00% Peer Rank 80.00% 60.00% 40.00% 20.00% 0.00% 0 0.2 0.4 0.6 0.8 1 1.2 Ratio Value FIGURE 1.6 5.5.2 The Standard Ranking Algorithm (an Ascending Ratio) Alternative Ranking Algorithm The alternate ranking fits a log curve to the peer quartiles. The figure below illustrates the curve where the quartiles are 1, 10, and -5. Note that for gearing and leverage ratios in RiskAnalyst, if the ratio is negative, we default the rank to 0. Therefore, the lower section of the curve is flat at 0 for these ratio values. 120.00% Peer Rank 100.00% 80.00% 60.00% 40.00% 20.00% 0.00% -150 -100 -50 0 50 100 150 Ratio Value FIGURE 1.7 The Alternative Ranking Algorithm The alternate ranking algorithm assumes that the map between ratio and rank can be approximated by a pair of log linear functions: ln( ratio ) = m.Rank + c One function is approximated for each side of the discontinuity in the mapping. Let altrank denote the alternative rank, where: 40 ⎧ ⎪ ⎪max(crossOver, HighRank ) 100 ⎪ ⎪0 altRank = ⎨ ⎪ ⎪ ⎪ ⎪min(crossOver, LowRank ) 100 ⎩ • • if x ≥ 0 if x < 0 and if gearing or leverage ratio Otherwise (Note : This option is not used with the shipped configuration of RiskAnalyst) gm denotes the gradient multiplier constant of a ratio. This is defined in the PEER_RANK_MULT table. • lhCutoff and rhCutoff denote the left- and right-hand cut-off constants respectively. These are defined in the NEG_RANK_CUTOFF table. • mh = (ln (Q1) − ln (Q 2 )) 2 5 • c h = ln (Q1) − 75.mh • ml = mh .gm • cl = ln( Q3 ) − 25.ml • HighRank = min(100, max(25, (ln( x ) − c h ) / mh )) • LowRank = min(50, max(0, (ln( x ) − cl ) / ml )) • rhCutoffRank = min(100, max(25, (ln( rhCutoff ) − c h ) / mh )) • lhCutoffRank = min(50, max(0, (ln( lhCutoff ) − c l ) / ml )) • 5.6 crossOver = (lhCutoffRank + rhCutoffRank ) 2 COMBINING RATIO COMPONENTS The absolute, trend, and peer components are combined to give an assessment of the ratio. A weight is defined for each ratio component. These weights must add up to one, i.e. w abs + wtrnd + w peer = 1 The mean of the ratio assessment is calculated as: μ ratio =μ abs . wabs + μ trnd . wtrnd + μ trnd . wtrnd ) + peer . w peer The standard deviation is calculated as: σ GUIDE TO BUSINESS ANALYSIS ratio = (σ . w ) + (σ 2 abs abs 2 (σ . w ) 2 peer peer 41 5.7 SPECIAL CONDITIONS RiskAnalyst has a number of special conditions that determine the behaviour of ratio assessments. Many conditions are configurable using the Internal Rating Model Author. 5.7.1 Assess Ratio A ratio is only assessed if all of the following conditions are true. • The peer contribution is entered for the ratio. It could be zero. • At least one of the components of the ratio assessment with a non-zero weight is defined. • The ratio has a value for the current statement date. • The ratio values are not all zero. • If the ratio is “Gross Margin” AND the industry division is Services or Transportation (defined in the SPECIAL_INDS table) THEN all of the following must be TRUE if the ratio is to be assessed: • At least one of the “Gross Margin” ratio values considered for analysis is not 100%. • The quartile values are NOT all 100% OR all 0. • The ratio value is NOT 100% for both the current and penultimate periods. • If the ratio is “Creditor Days” OR “Debtor Days” OR “Stock Days” OR “A/P Days” OR “A/R Days” OR “Inventory Days” AND the industry division is Services or Transportation (defined in the SPECIAL_INDS table) THEN at least one quartile value must be non-zero if the ratio is to be assessed. • If the ratio is “% NPBT/TNW” THEN Tangible Net Worth must be positive. 42 5.7.2 Ratio Unacceptable The ratio value is set to "Unacceptable" if either of the following is true: • The ratio is “Sales/WC” or “Turnover/WC” AND the ratio value is negative in the current period AND the industry is not retail. (The industries are defined in the SPECIAL_INDS table using the “SWCUnac” key). • The ratio is one of “OFF BS LEVERAGE”, “OFF BS GEARING”, “ADJ GEARING”, “ADJ LEVERAGE”, “DEBT/TNW” AND the ratio is negative for the current period. 5.7.3 Trend Defaults to Good The ratio value is set to "Good" if the following condition applies: • This is a gearing ratio (one of “OFF BS LEVERAGE”, “OFF BS GEARING”, “ADJ GEARING”, “ADJ LEVERAGE”, “DEBT/TNW”) AND the ratio is either undefined or negative for at least one of the periods which qualify for trend assessment (except the latest). 5.8 DEBT COVERAGE RATIO ASSESSMENTS There are a few unique ratios that behave differently from the calculations described above. These debt coverage ratio assessments are used only in the fundamental analysis approach and are assessed using the calculations described in the following sections. 5.8.1 Earnings and Cash Flow Coverage Ratios The Earnings Coverage and Cash Flow Coverage ratios are assessed by determining a weighted average ratio value and comparing these against a set of benchmarks. Let x denote the value of the ratio for period i i 2 ds and ⎛ ∑ ⎜⎜ x 1≤i ≤ n , ≠0 i 0 x dsratio = ⎝ ⎛ ∑ ⎜⎜ 1≤i ≤ n , ≠0 i 0 x The ratios ( i ⎝ 1 ratio ( n −i ) ⎞ ⎟ ⎟ ⎠ ( n −i ) ⎞ ⎟ ⎟ ⎠ 2 2 denote the weighted average ratio. x ) are calculated as either earnings or cash flow over total debt service. If total debt i service is zero for any period then the system substitutes its value ( ± ∞ ) with a proxy. This proxy is set as 10 if the earnings or cash flow are positive and –10 if negative. If both the ratio numerator and denominator are zero, the ratio is not included in the calculation. Note that in 2 The indices are specified in the same way as in section 5.4. GUIDE TO BUSINESS ANALYSIS 43 this circumstance the index of the other ratios are not altered and the weighting factor for that ratio is not included in the denominator of . ratio ds RiskAnalyst assesses dsratio against a set of benchmark fixed points. These fixed points are defined during modelling. The earnings coverage and cash coverage assessments are created using the algorithm for creating assessments using numeric inputs defined in section 5.2.1. 5.8.2 The Cash Flow Management Assessments An internal rating model using the fundamental analysis approach determines a cash impact of management variables assessment for each financial period and then aggregates these assessments to determine an overall cash flow management assessment. The individual cash impact of management variables assessments are computed by assessing the cash impact ratio against a set of benchmark fixed points. The aggregation of the individual assessments is performed using a weighted sum. Up to four periods of data are considered in the analysis. The individual cash impact of management variables assessments are determined using the methodology for assessing numeric assessments described in section 5.2.1. The sole input is the cash impact ratio. The fixed points are defined during modelling. The mean and standard deviation are 50 and 3 respectively. Let μ ir and σ ir be the real mean and standard deviation for the cash impact of management 3 r r and σ cfm be the real mean and standard variables assessment for period i, and let μ cfm deviation of the aggregate cash impact of management variables assessment. Let n be the number of periods under consideration. A maximum of four periods’ data can be used. The real mean of the combined cash impact of management variables assessment is calculated as: r μ cfm ⎞ ⎛μr ∑ ⎜⎜ i ( n −i ) ⎟⎟ = ⎝ 2 ⎠ ⎞ ⎛ ∑ ⎜⎜ 1 ( n −i ) ⎟⎟ ⎝ 2 ⎠ The standard deviation as: r = σ cfm 44 ⎛σ r ⎞ ∑ ⎜⎜ i ( n−i ) ⎟⎟ ⎝ 2 ⎠ ⎛ ⎞ ∑ ⎜⎜ 1 ( n−i ) ⎟⎟ ⎝ 2 ⎠ 2 2 We need to map this distribution onto a meter. This is done in the same way as for ratio assessments (see section 5.6) using the method described in section 3.6.3. Again, since the input means are already on the 0-100 scale the Z function uses the class end points rather than the interval end points in this case, i.e. for this step: Z iy = except for i=7 where GUIDE TO BUSINESS ANALYSIS Z 7y cep − μ i σ y y = ∞. 45 CHAPTER 6 6 RATINGS SUMMARY Each internal rating model produces a borrower rating by combining the inputs of the model using the methodologies described in earlier chapters. RiskAnalyst displays a summary of the internal rating model with the overall assessment or score as well as the borrower rating and probability of default (PD). 6.1 STRUCTURE OF THE RATING The calculations used on the areas of analysis of an internal rating model vary depending on the approach used. The figure below displays a typical fundamental analysis internal rating model, in which the borrower rating is based on a financial and business assessment. The overall financial assessment is based on several financial inputs assessed in specific analysis areas (i.e., Operations, Liquidity, etc.). Similarly the overall business assessment is based on several subjective inputs assessed in the areas of Company, Management, and Industry. Assessments of the specific areas of analysis are aggregated according to the methodology described in Chapter 3. The grade and equivalent PD displayed in the Ratings Summary screen are derived from a mapping of the resulting mean of the overall assessment. The diagram below shows how the factors in a scorecard internal rating model are summed to form the section scores, which are then summed to form the scorecard total score. The factor, section, and overall scores are calculated according to the methodology described in Chapter 4. The grade and equivalent PD displayed in the Ratings Summary screen are derived from a mapping of the total score. GUIDE TO BUSINESS ANALYSIS 47 6.2 RATINGS SUMMARY SCREEN The Ratings Summary screen displays the overall assessments or scores for the different areas of analysis in the internal rating model. The areas of analysis are aggregated to form an overall rating, with a borrower grade and equivalent PD. If constraining is activated for the internal rating model, the Ratings Summary screen will display the result of constraining on the borrower grade. 6.2.1 Customer State The Customer State field lists different key points in the ratings process and is used when archiving a customer. The items in this drop-down list can be edited by using the Configuration Console in RiskAnalyst Studio. See Chapter 11 for more information about the role of archiving in assessing borrower risk. 6.2.2 1-Yr EDF If you subscribe to RiskCalc and/or Public EDF (CreditEdge) integration for RiskAnalyst, the Ratings Summary screen will display an EDF value. If using Public EDF, the Ratings Summary screen will also include the Moody's Rating for the customer. The screen displays the 1-Yr Bond Default Rate Mapping if using a compatible RiskCalc model. These values can be used as a gauge for the grade and equivalent PD determined by the borrower rating. There are some instances when an EDF value will not be displayed, including incorrect RiskCalc or Public EDF settings, missing Source/Target Currency, and missing Exchange Rate settings. Ensure your RiskCalc or Public EDF connection and settings are working properly by running EDF Measures reports. 6.2.3 Overrides The Override feature allows users with a certain level of security to override the grade and PD determined by the scorecard total score. A higher security level is required to authorize the override. See the RiskAnalyst Help and the Configuration Console Help for more information about overrides. 6.2.4 Facilities The Facilities Total section provides total EAD, LGD, and EL% values and a total facility grade and EL grade. Totals are given for both Proposed and Existing Positions. 6.2.5 Constraining Internal rating models can be designed to include a constraining mechanism. The constraining mechanism can be any assessment or factor which can realistically be mapped to the internal grading system configured for the internal rating model. This assessment or factor is then used to prevent the borrower rating grade from varying too high or too low from the grade attached to the assessment or factor. A common example of constraining is to use a combined assessment of the 1-Yr and 5-Yr EDF values to constrain the borrower grade. If constraining is being used, the Ratings Summary screen displays the initial borrower grade, the grade associated with the assessment/factor being used to constrain the borrower grade, and the constrained borrower grade. 48 CHAPTER 7 7 LOSS GIVEN DEFAULT ANALYSIS As specified by Basel II, RiskAnalyst’s LGD calculation engine focuses at the level of the individual facility. To assist management and validation at a customer level, it also aggregates values up from each facility to create global customer values. NOTE This is based on Moody's KMV understanding of the provisions of the June 2004 document, International Convergence of Capital Measurement and Capital Standards as applied to the IRB Foundation Approach For each facility, RiskAnalyst calculates an Exposure at Default (EAD) value. This is an estimate of the amount that would be outstanding on the facility if the borrower defaulted. 4 RiskAnalyst then considers recoveries that could be obtained if the borrower were to default . The expected recovery can be improved by using Credit Risk Mitigants (CRMs) to enhance the quality of the obligation. These commonly take the form of guarantees and collateral. Collateral typically consists of a charge placed on a borrower's asset by the creditor. RiskAnalyst also calculates Expected Loss (EL), which is the expected value of losses due to default. This measure brings together the risk assessments of the borrower and of the borrower’s facility or facilities. 7.1 DETERMINING LGD AND EL VALUES RiskAnalyst determines borrower-level LGD and EL simply by summing the LGD and EL amounts determined for each facility. To derive the LGD of each facility, RiskAnalyst breaks up the exposure (EAD) into three possible parts: guaranteed, collateralized, and unsecured. For each of these parts, RiskAnalyst determines an LGD separately and then sums the LGD amounts to derive the aggregate LGD. RiskAnalyst calculates EL in a similar way. 4 Recovery can be considered the opposite of LGD. If both are expressed as percentage factors, Recovery = 1 – LGD%. GUIDE TO BUSINESS ANALYSIS 49 The following diagram and description summarizes the process used to determine the LGD value for each component part of the overall LGD calculation. Please note that Chapter 8 provides further details of the calculations. A given credit risk mitigant (CRM) may be applied to more than one facility (crosscollateralization). RiskAnalyst can represent most cross-collateralization scenarios by using a consistent approach when allocating CRMs to facilities and sharing the risk mitigating effects of the CRMs between the facilities. By default, RiskAnalyst considers guarantees first (before collateral), as specified in Basel II. The system sorts each facility’s guarantees at two levels: first, according to their probability of default (PD) and second, according to their LGD. There are complex rules governing which guarantees are eligible to act as credit risk mitigants. Accordingly, RiskAnalyst next considers 5 eligibility criteria for each guarantee, applying the appropriate haircut and, if applicable, limiting the guarantee’s value according to any limit specified by the user. RiskAnalyst then determines what proportion of the guarantee to allocate to the facility by considering the size of the exposure and the available guarantee value. Two factors impact this calculation: 1. Whether the guarantee is shared across other facilities, and how much is already allocated to these. RiskAnalyst uses a weighting system based on EAD and user input. This is described in more detail below. 2. How much exposure remains uncovered on this facility. If the remaining exposure is less than the available guarantee value, the allocation is capped at this value. Having determined the allocation for an individual guarantee, RiskAnalyst multiplies it by the guarantor’s LGD to calculate the LGD amount. The EL is simply the product of the LGD amount and the guarantor’s PD. The guarantee allocations to the facility can be summed across all the guarantees on the facility to determine a total eligible guarantee portion. Having considered the guarantees, RiskAnalyst then considers the collateral allocations. The process used is very similar to guarantees. One difference is that where guarantee values may 6 have user-defined limits, collateral items may have prior liens and limitations . When the system comes to consider the amount of uncovered exposure, it naturally removes the amount already covered by guarantees. It is also important to note that when calculating EL for 5 The haircuts for guarantees are all zero in the version of RiskAnalyst delivered out of the box. Almost all collateral items have haircuts. In the cases where Basel II specifies ‘overcollateralization amounts’, these values have been converted to haircuts. Note that an overcollateralization amount is equal to 1/(1- haircut). 6 50 7 each collateral item , it is the borrower’s PD or expected default frequency (EDF) measure that is used. If any of the exposure is still left uncovered after applying all available guarantees and collateral, RiskAnalyst considers this unsecured and uses the unsecured LGD% to determine an unsecured LGD amount. 7.2 FACILITY STATUS Facilities and Credit Risk Mitigants can be assigned to one of three different states. • Completed (CRMs only) • Committed (Facilities only) • Proposed These different states allow you to add or remove facilities and CRMs from the LGD and EL calculations without deleting facilities and CRMs from the system. 7.3 CALCULATING GRADES RiskAnalyst calculates a facility grade for the LGD dimension by mapping the LGD grade from the LGD% value calculated for the facility. The system displays this as the Facility Grade on the Facilities Summary screen. Additionally, the system displays an EL grade, which is either a direct mapping from the EL value or a combined mapping of the borrower grade and facility grade. 7 As for guarantees, EL is the product of the PD and the LGD amount. GUIDE TO BUSINESS ANALYSIS 51 CHAPTER 8 8 LOSS GIVEN DEFAULT CALCULATIONS In the previous section, we presented the structure that RiskAnalyst uses to analyze LGD. In this section we fill in the detail, specifying how RiskAnalyst performs each of the steps already outlined. We first consider EAD calculations and then move on to the CRMs. For both guarantees and collateral, we look at the eligibility criteria and parameters used to determine the 8 CRM values. In addition, we specify minimum LGD values for collateral. Then we move on to describe the methods used to allocate these CRMs to facilities. The calculation of the unsecured portion’s LGD is considered next. Finally, we examine the way in which PDs are brought into the LGD module. 8.1 CALCULATING EADS Facility-level exposure at default (EAD) is determined using the following process: • Determine the used and unused parts of the facility. • The used part is simply the utilization entered by the user. • The unused part is the commitment (also entered by the user) less the used part. • Determine the Credit Conversion Factor (CCF) to be utilized for the used and unused parts of the facility. These will be based on the Facility Type and up to three sub-types. CCFs are simply numbers used to convert off-balance-sheet items into credit equivalents by recognizing the inherent risk in various types of items. • Multiply the used part of the facility by the CCF for the used part and add this to the product of the unused part of the facility and the unused CCF for the facility. If the Facility has both utilization and a commitment, then the EAD is the sum of: • The utilization and CCF for the used part of the facility • (Commitment minus utilization) multiplied by the CCF for the unused portion If the Facility only has utilization, then the EAD is the value entered multiplied by the used portion CCF. If the Facility only has a commitment, then the EAD is the CCF for the unused portion multiplied by the value entered. If there are no CCFs specified for the facility (or any of the components of the MOF), then EAD is left blank. In some circumstances the system does not calculate an EAD, for example, derivative transactions. In this case, the user inputs an EAD value. 8 The LGD% values for guarantees are user input. GUIDE TO BUSINESS ANALYSIS 53 The following tables document the CCFs that RiskAnalyst uses: Type CCF – Utilized Portion CCF – Unutilized Portion Not Immediately Cancelable Immediately Cancelable Not Immediately Cancelable Immediately Cancelable Short Term Long Term Short Term Short Term Short Term Revolving Line of Credit 100% 100% 100% 100% 75% 75% 0% 0% Term Loan 100% 100% 100% 100% 75% 75% 0% 0% Short Term Loan 100% 100% 100% 100% 75% 75% 0% 0% Letter of Credit, Trade Transaction Related 20% 50% 20% 50% 20% 50% 0% 0% Letter of Credit, Financial Guarantee 100% 100% 100% 100% 100% 100% 0% 0% Bonds 50% 50% 50% 50% 50% 50% 0% 0% Banker’s Acceptance 20% 50% 20% 50% 20% 50% 0% 0% Indemnities 50% 50% 50% 50% 50% 50% 0% 0% Bank Guarantee 100% 100% 100% 100% 100% 100% 0% 0% Forward Currency Contract N/A N/A N/A N/A N/A N/A N/A N/A Interest Rate Swap N/A N/A N/A N/A N/A N/A N/A N/A Currency Swap N/A N/A N/A N/A N/A N/A N/A N/A Other Derivative N/A N/A N/A N/A N/A N/A N/A N/A Floor Plan Line of Credit 100% 100% 100% 100% 75% 75% 0% 0% Credit Cards 100% 100% 100% 100% 75% 75% 0% 0% Negotiations 20% 50% 20% 50% 20% 50% 0% 0% Overdraft 100% 100% 100% 100% 75% 75% 0% 0% Long Term Long Term Long Term If using UK English for your language settings, the system uses the following facility types: Type Overdraft 54 CCF – Utilised Portion CCF – Unutilised Portion Not Immediately Cancellable Immediately Cancellable Not Immediately Cancellable Immediately Cancellable Short Term Long Term Short Term Long Term Short Term Long Term Short Term Long Term 100% 100% 100% 100% 75% 75% 0% 0% Short Term Loan 100% 100% 100% 100% 75% 75% 0% 0% Term Loan 100% 100% 100% 100% 75% 75% 0% 0% Letter of Credit, Trade Transaction Related 20% 50% 20% 50% 20% 50% 0% 0% Letter of Credit, Financial Guarantee 100% 100% 100% 100% 100% 100% 0% 0% Revolving Credit 100% 100% 100% 100% 75% 75% 0% 0% Bonds 50% 50% 50% 50% 50% 50% 0% 0% Banker’s Acceptance 20% 50% 20% 50% 20% 50% 0% 0% Indemnities 50% 50% 50% 50% 50% 50% 0% 0% Bank Guarantee 100% 100% 100% 100% 100% 100% 0% 0% Forward Currency Contract N/A N/A N/A N/A N/A N/A N/A N/A Interest Rate Swap N/A N/A N/A N/A N/A N/A N/A N/A Currency Swap N/A N/A N/A N/A N/A N/A N/A N/A Other Derivative N/A N/A N/A N/A N/A N/A N/A N/A Stocking Loan 100% 100% 100% 100% 75% 75% 0% 0% Credit Cards 100% 100% 100% 100% 75% 75% 0% 0% Negotiations 20% 50% 20% 50% 20% 50% 0% 0% 8.2 8.2.1 GUARANTEES Haircuts Applied to Guarantees A haircut is normally applied to a CRM in order to account for the risk of a fall in the value of the CRM held prior to default or during the close-out period. RiskAnalyst is shipped with haircuts for guarantees with values of zero. The system user is expected to enter the true expected realization of the guarantee into the ‘Valuation’ field when adding or editing guarantees in RiskAnalyst. Type Eligible Personal Min PD Max Eligible PD Haircut Corporate Yes No 0.03% 0.09% 0 Bank Yes No 0.03% N/A 0 Securities Firm Yes No 0.03% N/A 0 Public Sector Entity Yes No 0 N/A 0 Sovereign Yes No 0 N/A 0 Personal No Yes N/A N/A 0 8.2.2 Eligibility Criteria for Guarantees By default, RiskAnalyst uses the following criteria to determine whether guarantees are eligible: • The guarantee must be issued by an organization. Personal guarantees are not eligible. • The guarantee PD must be less than or equal to the borrower’s. • The guarantee EL (PD × LGD%) must be less than the product of the borrower’s PD and the LGD% value used for the unsecured portion of the facility. • If the guarantor is a corporate, the guarantor’s PD must be less than or equal to nine basis points (0.09%). GUIDE TO BUSINESS ANALYSIS 55 8.3 COLLATERAL 8.3.1 Mapping RiskAnalyst Collateral Types to Basel II Collateral Types The following table provides a mapping from the RiskAnalyst Collateral Types to their Basel II equivalents: RiskAnalyst Collateral Type Debtors/ Accounts Receivable (Less Than 1 Year) Basel II Collateral Type Receivables Stock/ Inventory Other Plant/Equipment/Furniture/Fixtures Agricultural Charge Land/Property/Real Estate Debt Securities Cash Gold UCITS/Mutual Funds Life Policies Other Other Other CRE/RRE Financial Collateral Financial Collateral Financial Collateral Financial Collateral Other Other 8.3.2 Calculating Haircuts for Financial Collateral RiskAnalyst uses the following table for specifying the base haircuts for financial collateral. These are taken from Basel II. Issue rating for debt securities Residual Maturity Sovereigns Other Issuers AAA to AA-/A-1 ≤ 1 year 0.5 1 >1 year, ≤ 5 years 2 4 > 5 years 4 8 ≤ 1 year 0.5 1 1 2 >1 year, ≤ 5 years 3 6 > 5 years 6 12 All 15 N/A A+ to BBB-/ A-2/A-3 and unrated bank securities per para 116(d) BB+ to BB- 56 Main index equities and Gold 15 Other equities listed on a recognized exchange UCITS/Mut Fund, Cash only 25 0 UCITS/Mut Fund, Sovrgn AA-/A-1 Bonds or better 4% UCITS/Mut Fund, Other AA-/A-1 Bonds or better 8% UCITS/Mut Fund, Sovrgn BBB-/C-1 Bonds or better 6% UCITS/Mut Fund, Other BBB-/C-1 Bonds or better 12% UCITS/Mut Fund, Sovrgn BB- Bonds or better 15% UCITS/Mut Fund, Main Index Equities or better 15% UCITS/Mut Fund, Oth Eqties Lstd on Recgnzd Exchnge or better 25% UCITS/Mut Fund, Other N/A Cash in the same currency 0 These haircuts are based on the assumption of ‘daily mark to market, daily remargining and a ten business day holding period’. Basel II also states that, for secured lending, a minimum holding period of 20 days is appropriate and provides two formulae for transforming the haircuts. One formula allows transformation of the holding period under which the asset’s volatility was determined into the minimum holding period: HM=HN√(TM/TN) • HM is the haircut under the minimum holding period. • HN is the haircut based on a holding period TN • TM is the minimum holding period for the transaction type (in this case, secured lending). • TN is the holding period used by the bank to determine HN A subsequent formula allows the determination of the appropriate haircut when the frequency of remargining or revaluation is longer than the minimum. H = HM √{[NR + (TM – 1)]/TM} • H is the required haircut. • NR is the actual number of days between revaluations or remargining. RiskAnalyst uses the above formulae to transform the supervisory haircuts. It assumes secured lending and daily revaluation. Therefore, TM is 20, TN is 10, and NR is 1. 8.3.3 Haircuts for Non-Financial Collateral As mentioned above, the overcollateralization levels specified by Basel II can be transformed into haircuts using the formula: haircut = 1/(1 – overcollateralization level). RiskAnalyst uses the following haircuts for eligible non-financial collateral: Type Receivables (≤1yr) – Not Affiliate Haircut 20% Receivables (≤1yr) – Affiliate N/A Inventory – Physical Inspection Performed 28.57143% Inventory – No Physical Inspection Performed N/A Plant/Equipment/Furniture/Fixtures (excl Leasehold Imprvmnts) 28.57143% Leasehold Improvements N/A CRE/RRE – No Dependence btwn Borrower & Property 28.57143% CRE/RRE – Dependence btwn Borrower & Property N/A Agricultural Charge 28.57143% Life Policies – With Surrender Value 28.57143% Life Policies - Other N/A Other 28.57143% GUIDE TO BUSINESS ANALYSIS 57 If using UK English for your language setting, the system uses the following collateral types: RiskAnalyst Collateral Type Debtors (≤1yr) – Not Affiliate Haircut 20% Debtors (≤1yr) – Affiliate N/A Stock – Physical Inspection Performed 28.57143% Stock – No Physical Inspection Performed N/A Plant/Equipment/Furniture/Fixtures 28.57143% CRE/RRE – No Dependence btwn Borrower & Property 28.57143% CRE/RRE – Dependence btwn Borrower & Property N/A Agricultural Charge 28.57143% Life Policies – With Surrender Value 28.57143% Life Policies - Other N/A Other 28.57143% 8.3.4 Prior Liens and Limitations Users can enter the values of any prior liens and limitations representing the situation where other creditors have charges issued against a piece of collateral. For each item of collateral, RiskAnalyst deducts any prior liens and then caps the collateral valuation to the value of the limitation. This is done after any haircut has been applied to the collateral valuation. 8.3.5 Eligibility Criteria for Collateral There are a number of eligibility criteria that must be met in order for an item of collateral to be considered eligible within RiskAnalyst. These are based around the Basel II requirements. However, it is envisaged that a bank will need to implement its own policy to handle any additional aspects not addressed as standard. To support this, where appropriate, the help system describes some of the Basel II requirements. The following sections describe the criteria encoded within RiskAnalyst. First Charge Required Some non-financial collateral types are not eligible if a second or subsequent charge has been taken. The following table documents this: Type 58 Receivables (≤1yr) – Not Affiliate First Charge Required No Receivables (≤1yr) – Affiliate N/A Inventory – Physical Inspection Performed Yes Inventory – No Physical Inspection Performed N/A Plant/Equipment/Furniture/Fixtures (excl Leasehold Imprvmnts) Yes Leasehold Improvements N/A CRE/RRE – No Dependence btwn Borrower & Property No CRE/RRE – Dependence btwn Borrower & N/A Property Agricultural Charge Yes Life Policies – With Surrender Value Yes Life Policies - Other N/A Other Yes If using UK English for your language setting, the system uses the following collateral types: Debtors (≤1yr) – Not Affiliate RiskAnalyst Collateral Type First Charge Required No Debtors (≤1yr) – Affiliate N/A Stock – Physical Inspection Performed Yes Stock – No Physical Inspection Performed N/A Plant/Equipment/Furniture/Fixtures Yes CRE/RRE – No Dependence btwn Borrower & Property No CRE/RRE – Dependence btwn Borrower & Property N/A Agricultural Charge Yes Life Policies – With Surrender Value Yes Life Policies – Other N/A Other Yes Eligibility Criteria based on Seniority If the facility ‘Seniority’ is set to ‘Subordinated,’ only financial collateral is eligible. The following table describes RiskAnalyst’s behavior: (The only differences between MMAS, IFT, and ESM are minor spelling changes). Type Cash Eligible for Subordinated Facilities Yes Gold Yes Equity, Listed on Main Index Yes Equity, Listed on Recognized Exchange Yes Equity, Unquoted N/A Debt Secrty, Sovrgn, AAA/AA/A-1, ≤1yr Yes Debt Secrty, Other, AAA/AA/A-1, ≤1yr Yes Debt Secrty, Sovrgn, AAA/AA/A-1, >1, ≤5yrs Yes Debt Secrty, Other, AAA/AA/A-1, >1, ≤5yrs Yes Debt Secrty, Sovrgn, AAA/AA/A-1, >5yrs Yes Debt Secrty, Other, AAA/AA/A-1, >5yrs Yes Debt Secrty, Sovrgn, A/BBB/A-2/A-3/P-3, ≤1yr Yes Debt Secrty, Other, A/BBB/A-2/A-3/P-3, ≤1yr Yes Debt Secrty, Other Qualifying Bank, •1yr, see Help Yes Debt Secrty, Sovrgn, A/BBB/A-2/A-3/P-3, >1, ≤5yrs Yes GUIDE TO BUSINESS ANALYSIS 59 Debt Secrty, Other, A/BBB/A-2/A-3/P-3, >1, ≤5yrs Yes Debt Secrty, Other Qualifying Bank, >1, ≤5yrs, see Help Yes Debt Secrty, Sovrgn, A/BBB/A-2/A-3/P-3, >5yrs Yes Debt Secrty, Other, A/BBB/A-2/A-3/P-3, >5yrs Yes Debt Secrty, Other Qualifying Bank, >5yrs, see Help Yes Debt Secrty, Sovrgn, BB Yes Debt Secrty, Other N/A UCITS/Mut Fund, Cash only Yes UCITS/Mut Fund, Sovrgn AA-/A-1 Bonds or better Yes UCITS/Mut Fund, Other AA-/A-1 Bonds or better Yes UCITS/Mut Fund, Sovrgn BBB-/C-1 Bonds or better Yes UCITS/Mut Fund, Other BBB-/C-1 Bonds or better Yes UCITS/Mut Fund, Sovrgn BB- Bonds or better Yes UCITS/Mut Fund, Main Index Equities or better Yes UCITS/Mut Fund, Oth Eqties Lstd on Recgnzd Exchnge or better Yes UCITS/Mut Fund, Other N/A Minimum Collateralization Level Restrictions 9 Under Basel II, a minimum collateralization level is specified for certain collateral types . This can be summarized as a requirement that, if the collateral is to be eligible, the collateral’s value must provide significant coverage of the exposure. 10 The following table provides details of minimum collateralization levels by collateral type . Note that all financial collateral has a minimum collateralization level of 0. Type 9 Minimum Collateralization Level Receivables (≤1yr) – Not Affiliate 0 Receivables (≤1yr) –Affiliate N/A Inventory – Physical Inspection Performed 30% Inventory – No Physical Inspection Performed N/A Plant/Equipment/Furniture/Fixtures (excl Leasehold Imprvmnts) 30% Leasehold Improvements N/A CRE/RRE – No Dependence btwn Borrower & Property 30% CRE/RRE – Dependence btwn Borrower & Property N/A Agricultural Charge 30% Life Policies – With Surrender Value 30% Life Policies - Other N/A Other 30% These types are CRE/RRE and Other Collateral. A minimum collateralization level of zero effectively means that there is no minimum collateralization level applicable to the collateral type. 10 60 If using UK English for your language setting, the system uses the following collateral types: RiskAnalyst Collateral Type Debtors (≤1yr) – Not Affiliate Minimum Collateralisation Level 0 Debtors (≤1yr) –Affiliate N/A Stock – Physical Inspection Performed 30% Stock – No Physical Inspection Performed N/A Plant/Equipment/Furniture/Fixtures 30% CRE/RRE – No Dependence btwn Borrower & Property 30% CRE/RRE – Dependence btwn Borrower & Property N/A Agricultural Charge 30% Life Policies – With Surrender Value 30% Life Policies - Other N/A Other 30% The purpose of the Minimum Collateralization Requirement is to ensure that any piece of nonfinancial collateral allocated to a facility it is of sufficient value compared to the facility to make the allocation worthwhile. RiskAnalyst uses the following to determine whether collateral has met the required Minimum Collateralization level. • • • • • • • • • Determine the discounted value for each piece of eligible collateral for which the minimum collateralization amount is zero. o The collateral details MIN_COLLAT_AMOUNT specifies the minimum collateralization amount. o The discounted value is the valuation times the haircut. Determine the amount apportioned to this facility the Allocation Percent. Sum the discounted values for zero min-collateralization collateral. Deduct this value from the EAD. If guarantees are taken first, deduct guarantee value from the EAD as well. Call the value of the EAD remaining after deducting items with zero min-collateralization and guarantees EADForNonZeroMinCollat. Sum the base values of the remaining eligible collateral items (i.e., non-zero mincollateralization) o If a limitation is specified, the base value is the limitation plus any prior liens. Otherwise it is the valuation value. Call the sum of these base values NonZeroMinCollat Determine M as follows: NonZeroMinCollat divided by EADForNonZeroMinCollat. If M is less than the maximum MIN_COLLAT_AMNT for the piece of collateral, then the piece of collateral has not passed the Minimum Collateralization test. Otherwise, it has. Other Eligibility Criteria In addition to the above criteria, RiskAnalyst ignores collateral where the collateral LGD% is greater than the unsecured LGD% for the facility. 8.3.6 Collateral Minimum LGD% Values The following table specifies the minimum LGD% values for each collateral type. Note that all financial collateral has a minimum LGD of 0%. GUIDE TO BUSINESS ANALYSIS 61 RiskAnalyst Collateral Type Receivables (≤1yr) – Not Affiliate Minimum LGD Value 35% Receivables (≤1yr) –Affiliate N/A Inventory – Physical Inspection Performed 40% Inventory – No Physical Inspection Performed N/A Plant/Equipment/Furniture/Fixtures (excl Leasehold Imprvmnts) 40% Leasehold Improvements N/A CRE/RRE – No Dependence btwn Borrower & Property 35% CRE/RRE – Dependence btwn Borrower & Property N/A Agricultural Charge 40% Life Policies – With Surrender Value 40% Life Policies - Other 40% Other 40% If using UK English for your language setting, the system uses the following collateral types: Debtors (≤1yr) – Not Affiliate RiskAnalyst Collateral Type Minimum LGD Value 35% Debtors (≤1yr) –Affiliate N/A Stock – Physical Inspection Performed 40% Stock – No Physical Inspection Performed N/A Plant/Equipment/Furniture/Fixtures 40% CRE/RRE – No Dependence btwn Borrower & Property 35% CRE/RRE – Dependence btwn Borrower & Property N/A Agricultural Charge 40% Life Policies – With Surrender Value 40% Life Policies - Other 40% Other 40% The user can override these values in RiskAnalyst. The system has a configurable restriction on these overrides and, as shipped, these LGD% values can only be increased by the user. 62 8.4 ALLOCATING CRMS TO FACILITIES Allocating CRMs to facilities is straightforward in cases where a single CRM is allocated to a single facility, but in practice it is common for more complex relationships to exist between facilities and CRMs. To handle this, RiskAnalyst uses both algorithmic and user-specified approaches. This allows the user to capture any legal nuances while RiskAnalyst does the number crunching. Users allocate CRMs to facilities either by auto-allocating or by manually defining a percentage of the CRM to each facility. Using this method, users model the many-to-many type relationships that often occur between facilities and CRMs. In addition, fields for prior liens and limitations allow users to capture details of charges made to other creditors. If a CRM is shared between two or more facilities, its value needs to be split between these facilities. RiskAnalyst uses an algorithm based on EAD weighting to apportion the CRMs. The user is also given the option to control the allocation manually if required. 8.4.1 Automatic Allocation of CRMs using EAD Weighting 11 RiskAnalyst first allocates all of the guarantees to facilities. The following process is used: • For each guarantee, determine all the facilities to which the guarantee is allocated. • Find the total EAD of all these facilities. • Determine a weighting to allocate the guarantee to each individual facility; this is the facility EAD divided by the total EAD calculated in the previous step. • For each facility, the guarantee allocation is the guarantee value12 times the weighting derived in the previous step. Having allocated guarantees, RiskAnalyst then allocates collateral. The process is similar: • For each collateral item, determine all the facilities to which the collateral is allocated and eligible for consideration. • Determine the residual EAD after deducting the guarantee allocations. • Find the total EAD of all these facilities. • Determine a weighting to allocate the collateral to each individual facility; this is the facility EAD divided by the total EAD calculated in the previous step. • For each facility, the collateral allocation is the eligible collateral value13 times the weighting derived in the previous step. 8.4.2 Manual Allocation of CRMs RiskAnalyst also allows manual allocation of CRMs. This allocation method would be used in situations where the EAD weighting method is not optimal. Examples include situations where a particular CRM is tied to a specific facility, while all other CRMs are available to all facilities. 11 The text describes the configuration with which RiskAnalyst is shipped. If RiskAnalyst is configured to consider collateral before guarantees, the ordering in this section will be reversed. 12 The expected realization after discounting with a haircut, if applied, and limiting with a limit, if present. 13 The valuation after discounting with a haircut and applying any prior liens and limitations. GUIDE TO BUSINESS ANALYSIS 63 In this situation, it is possible that CRMs could be over-allocated to the first facility while leaving the other facilities only partially covered. To use manual allocation, you can specify an allocation portion as a percentage value for each facility when adding or editing a CRM. RiskAnalyst multiplies the CRM value by the facility's allocation to determine the portion allocated to each part of the CRM. Manual allocation and EAD weighting can be used with the same CRM. 8.5 DERIVATION OF THE LGD FOR THE UNSECURED PORTION OF THE FACILITY RiskAnalyst uses the facility ‘seniority’ to determine the LGD% for the unsecured portion of the facility. This is 45% for senior facilities and 75% for subordinated ones. 8.6 DERIVATION OF THE BORROWER PD The PD used in the LGD module can come from one of two sources: • The 1-Yr EDF credit measure supplied by either RiskCalc or CreditEdge. • The equivalent PD derived from the scorecard total and grade. The source is configurable within the Facility Configuration application in RiskAnalyst Studio. 64 CHAPTER 9 9 LOSS GIVEN DEFAULT ALGORITHM 9.1 INTRODUCTION TO THE LGD ALGORITHM The previous chapter provided a detailed overview of the LGD calculations in RiskAnalyst. This chapter describes the details of the LGD algorithm used in RiskAnalyst. It is designed to enable you to understand better the LGD functionality of RiskAnalyst. While it is more technical and programmatic than the previous chapter, it is not a full and complete technical specification, in that is does not describe functionality at the code level. The flowchart below shows a simplified version of the LGD algorithm. The following sections then describe each of the functional blocks referred to in the flowchart. The block operations are described in two sections: • A description section, which provides a high level English description of the purpose of the operation, and • An implementation section, which provides a detailed technical description of how the operation is implemented. GUIDE TO BUSINESS ANALYSIS 65 Begin Calculate EAD for all Facilities Calculate Allocations for all Facilities and CRMs Sort CRMs Derive Eligibility Per Se for all CRMs LGD Calculation Loop through all Facilities Loop through all CRMs CRM eligible for Facility? Yes Yes No Allocate CRM to Facility Yes Facilities Remaining CRMs Remaining? Calculate EL and LGD for Facility No No Calculate Summary Level Data End 9.2 LGD ALGORITHM OVERVIEW Description The LGD Component operates in three distinct phases: 1. A setup phase, during which data concerning Facilities, CRMs, and allocations is entered into the engine. Facilities data is passed into the LGD engine from the LGD wrapper 2. A calculation phase, during which the LGD algorithm is run. 3. A results phase, during which results are gathered from the engine and sent to the user. Implementation When the calculation phase begins, the order of operation is as follows: 66 1. Calculate EAD for each facility 2. Run the allocation algorithm for each CRM and each facility 3. Identify ineligible CRMs 4. Sort CRMs according to user-specified order 5. Calculate LGD for each facility 6. Calculate Summary Level Data 9.3 CALCULATING EAD Description EAD (Exposure at Default) is an estimate of the exposure on the facility at the time of default. Implementation EAD is either calculated or user-supplied. If a user-supplied value is entered, and it is larger than the calculated EAD, the supplied value is the EAD value that is used. Otherwise, EAD is calculated as follows: EAD = (Utilized Portion * CCF for utilized Portion) + (Unutilized Portion * CCF For Unutilized Portion) where • Utilized Portion = utilization • Unutilized portion = commitment – utilization. • CCF (Credit Conversion Factor) values are read from the LGDRefFacility table for each facility type. CCF values can differ depending on whether or not the facility is immediately cancelable, so this is used as another key in the LGDRefFacility table. 9.4 CALCULATE ALLOCATIONS Description Determine the percentage of each CRM to allocate to a facility based on the facility’s EAD relative to the total of all EADs. Implementation • It is important the allocation occurs after EAD calculation, as the calculated EAD values are required in the allocation algorithm. • The system allocation is stored internally as a double, so 50% is stored as 0.5. When the value is displayed to the user it is multiplied by 100 to turn it into a percentage. • Users can supply their own allocation, in which case the user allocation will override the system allocation. GUIDE TO BUSINESS ANALYSIS 67 For each CRM Total EAD = sum (EAD of all facilities linked to this CRM) For each facility This EAD = EAD for this facility System Allocation = This EAD / Total EAD DERIVE CRM ELIGIBILITY PER SE 9.5 Description Not all CRMs can be used to support a given facility. Whether or not a CRM can be used to support a facility is determined by its eligibility. There are two levels of CRM eligibility: 1. Per Se eligibility, whereby a CRM can be eligible or ineligible of itself, irrespective of any relation to any facility, and 2. Per Facility eligibility, in which a facility may be eligible per se, but is nonetheless not eligible for a particular facility. This is derived during the Allocation process. Deriving CRM Eligibility Per Se 9.5.1 Personal Guarantees • In RiskAnalyst, Personal Guarantees are never eligible. Organizational Guaranatees An organization guarantee is eligible per se if all the conditions below are true: • The BIS2_Eligible column in Guarantee Table (Reference Database) is true. • The PD is less than the value in the MAX_PD column of the Guarantee Table, for this Guarantee type. • PD is greater than the minimum PD given in the Guarantee Table. Collateral A piece of collateral is eligible per se if: • For this Credit Risk Mitigant type, the BIS2_ELIGIBLE column in the Collateral_Details table in the reference database is set to true. • There are Prior Liens and the REQ_FIRST_CHG column for this collateral type is set to false. 9.6 CRM SORTING The order in which CRMs are considered when they are being allocated to facilities makes a difference to the results that are obtained. Two CRM ordering algorithms are supported. 68 9.6.1 Guarantees First Description This algorithm allocates Guarantees, then Financial Collateral, then Non-Financial Collateral. Implementation This algorithm lists the guarantees towards the top of the Credit Risk Mitigant list. This is the Basel preferred option, and the default in RiskAnalyst. The order is: 1. Eligible Guarantees a. Organizational Guarantees First b. Lowest EL first c. Highest Expected Realization First d. Most time to maturity first e. Lowest ID first 2. Financial Collateral a. Low LGD First b. Where LGDs are the same (quite likely in many cases), then ordered by Minimum Collateralization Amount (taken from the MIN_COLLAT column of the LGDRefCollateral table). c. Where LGD and Min. Collat. Amt. are equal, order by Potential Eligible Amount for Facility (calculated value), largest first. d. Most time to maturity first 3. Non-Financial Collateral a. Low LGD First b. Where LGDs are the same (quite likely in many cases), then ordered by Minimum Collateralization Amount (taken from the MIN_COLLAT_AMOUNT column of the SECURITY_DETAILS table). c. Where LGD and Min. Collat. Amt. are equal, order by Potential Eligible Amount for Facility (largest first). d. Most time to maturity first 4. Ineligible Guarantees (by ID) 5. Ineligible Financial Collateral (by ID) 6. Ineligible Non-Financial Collateral (by ID) GUIDE TO BUSINESS ANALYSIS 69 9.6.2 Guarantees Second Description This algorithm allocates Financial Collateral, then Non-Financial Collateral, and then Guarantees. Implementation This option was requested from customers. The order is: 1. Financial Collateral a. Low LGD First b. Where LGDs are the same (quite likely in many cases), then ordered by Minimum Collateralization Amount (taken from the MIN_COLLAT column of the LGDRefCollateral table). c. Where LGD and Min. Collat. Amt. are equal, order by Potential Eligible Amount for Facility (largest first). d. Most time to maturity first 2. Non-Financial Collateral a. Low LGD First b. Where LGDs are the same (quite likely in many cases), then ordered by Minimum Collateralization Amount (taken from the MIN_COLLAT_AMOUNT column of the SECURITY_DETAILS table). c. Where LGD and Min. Collat. Amt. are equal, order by Potential Eligible Amount for Facility (largest first). d. Most time to maturity first 3. Eligible Guarantees a. Organizational Guarantees First b. Lowest EL first c. Highest Expected Realization First d. Lowest ID first e. Most time to maturity first 70 4. Ineligible Financial Collateral (by ID) 5. Ineligible Non-Financial Collateral (by ID) 6. Ineligible Guarantees (by ID) 9.7 THE LGD CALCULATION Description Once all the data concerning CRMs and Facilities has been entered into the LGD Component, the LGD Calculations are started by running the LGD Algorithm. Implementation The operation of the LGD algorithm is different depending on whether the CRM ordering method is Guarantees First or Guarantees Second. If Ordering is Guarantee First Perform Guarantee First Processing and return EAD remaining after allocation Otherwise Perform Guarantee Second Processing and return EAD remaining after allocation 9.7.1 Guarantee First/Second Processing Description There are two differences between Guarantee First processing and Guarantee Second processing: 1. The CRM sorting algorithms are different, and 2. The order in which guarantees, Zero Minimum Collateralization, and Non-Zero Minimum Collateralization collateral are allocated to facilities is different. 9.7.2 Guarantee First Processing Description Guarantee First allocation allocates guarantees before collateral. Implementation 1. Sort Guarantees 2. Allocate Guarantees 3. Sort Collateral items with zero Minimum Collateralization amounts (ZMC Collateral) 4. Allocate ZMC Collateral 5. Sort Collateral with non-zero Minimum Collateralization amounts (NZMC Collateral) 6. Allocate NZMC Collateral 9.7.3 Guarantees Second Processing Description Guarantee Second allocation allocates collateral before guarantees. GUIDE TO BUSINESS ANALYSIS 71 Implementation 1. Sort ZMC collateral 2. Allocate eligible ZMC collateral 3. Derive eligibility for NZMC collateral, taking minimum collateralization into account 4. Allocate eligible NZMC collateral 5. Sort guarantees 6. Allocate guarantees 9.8 PERFORM CRM ALLOCATION Description Allocation is the process of allocating portions of a CRM to a facility. In general, the amount of a CRM allocated to a facility (called the ‘Eligible Amount’) will be the Mismatch-Discounted Expected Realization multiplied by the allocation percentage. The Mismatch-Discounted Expected Realisztion is the Expected Realiation of the CRM, discounted according to the CRM mismatch formula specified by Basel II. The order that the CRMs are taken in is very important, as different results will be obtained if a different ordering algorithm is used. Implementation The allocation process is almost the same whether we are allocating ZMC collateral, NZMC collateral, or guarantees. The only slight difference is that when allocating NZMC collateral the eligibility check has to take Minimum Collateralization into account. For each CRM If the CRM is eligible for the facility If the mismatch-adjusted eligible amount > EAD remaining Allocate the EAD remaining to the facility Otherwise Allocate the mismatch-adjusted eligible amount to the facility Subtract the allocated amount from the EAD remaining The LGD Amount for the CRM is (CRM LGD% x allocated amount) 9.8.1 Mismatch-Adjusted Eligible Amount Description This is the eligible amount, discounted to take the CRM mismatch into account. It is a property of the CRM and a particular Facility. Implementation The mismatch-adjusted eligible amount is calculated from the eligible amount as follows: 72 The Capped CRM Residual Maturity is Min (Residual Maturity of CRM in years, 5) The Capped Facility Residual Maturity is Min (Residual Maturity Facility in years, Capped CRM Residual Maturity) The Adjustment Factor is (Capped CRM Residual Amount – 0.25) divided by (Capped Facility Residual Amount – 0.25) The Adjusted Eligible Amount is (Expected Realisation x Allocation % x Adjustment Factor) 9.8.2 Expected Realization Description The Expected Realization is a property of the CRM. It is the valuation of the CRM taking the haircut, valuation, and any prior liens into account. Implementation If a limitation is defined Min(Max((Valuation x (1 – Haircut)) – Prior Liens),0),Limitation) Otherwise Max((Valuation x (1 – Haircut)) – Prior Liens),0) ELIGIBILITY PER FACILITY 9.9 Description This concerns the case in which a facility may be eligible per se, but is nonetheless not eligible for a particular facility. 9.9.1 Implementation CRM Mismatches CRM Mismatches feature heavily when determining the mismatch Eligibility Conditions, as defined below. This section defines precisely what constitutes a mismatch as far as RiskAnalyst is concerned. Data Requirements When deciding whether there is a CRM mismatch, we need to consider the following data items. Original maturities are entered in months and reflect the overall lifespan of the CRM or Facility. Maturity Dates reflect that date when the CRM or Facility Matures. The original maturity of a Facility is also required to calculate the Facility EAD, and as such is a compulsory data requirement. • The original maturity of the CRM • The original maturity of the Facility • The maturity date of the CRM GUIDE TO BUSINESS ANALYSIS 73 • The maturity date of the Facility Data Conditions a) If there is no Facility Maturity Date but there is a CRM Maturity Date, the CRM is not eligible. b) Otherwise, if any of the mismatch data requirements are missing, there is no mismatch so the CRM is eligible as far as CRM maturity is concerned. c) If neither (a) nor (b) applies, then a CRM mismatch is defined as the case where the residual maturity of the CRM is less than that of the Facility. Mismatch Eligibility Conditions These conditions apply to all CRMs. Mismatches are defined as above. • Any allocation where there is a mismatch and where the residual maturity is less than 3 months is not eligible. The residual maturity is the difference between the maturity date of the CRM and the evaluation date. • Any allocation where there is a mismatch and the original maturity of the CRM is less than 1 year is not eligible. • Any allocation where the adjusted expected realization is zero is not eligible. The adjusted expected realization is calculated in the same way as the adjusted eligible amount, but using an allocation of 100%. Organizational Guarantees An organizational guarantee is eligible for a given facility if it is eligible per se, and, for this facility: • Either The Guarantor PD (PD of the guarantee) is less than the Borrower’s PD (PD of the customer), or (Guarantor PD x Guarantor LGD%) is less than (Borrower PD x Unsecured LGD%) Collateral A piece of collateral is eligible for a given facility if it is eligible per se and, for this facility: 74 • If the facility is set as senior or it is set as subordinated and the INC_SUBORDINATED column of the LGDRefCollateral table is set to true. • The LGD% of the Credit Risk Mitigant is less than the LGD% of the unsecured portion. • Minimum Collateralization has been met (Non-Zero Minimum Collateralization Collateral only. See below). 9.9.2 Determining the Minimum Collateralization Requirement Description The purpose of the Minimum Collateralization Requirement is to ensure that any piece of nonfinancial collateral allocated to a facility is of sufficient value compared to the facility to make the allocation worthwhile. Implementation • Note that the minimum collateralization test can only be carried out after: Guarantees and ZMC items have been allocated (if guarantees taken first), or ZMC items have been allocated (if guarantees are taken second). • In RiskAnalyst, the results of minimum collateralization do not affect the facility/CRM auto-allocations. • Determine the discounted value for each piece of eligible collateral for which the minimum collateralization amount is zero. The collateral details MIN_COLLAT_AMOUNT specifies the minimum collateralization amount. The discounted value is the valuation times the haircut. • Determine the amount of the discounted value apportioned to this facility using both the Allocation Percent and the EAD weighting. • Sum the discounted values for zero min-collateralization collateral. Deduct this value from the EAD. • If guarantees are taken first, deduct guarantee value from the EAD as well. • Call the value of the EAD remaining after deducting items with zero mincollateralization and guarantees EADForNonZeroMinCollat. • Sum the base values of the remaining eligible collateral items (i.e., non-zero mincollateralization) If a limitation is specified, the base value is the limitation plus any prior liens. Otherwise, it is the valuation value. • Call the sum of these base values NonZeroMinCollat • Determine M as follows: NonZeroMinCollat divided by EADForNonZeroMinCollat. • If M is less than the maximum MIN_COLLAT_AMNT for the piece of collateral, then the piece of collateral has not passed the Minimum Collateralization test. Otherwise, it has passed the Minimum Collateralization test. GUIDE TO BUSINESS ANALYSIS 75 9.10 EXPECTED LOSS Description If no guarantees are present this is the borrower's PD multiplied by the total LGD amount. If guarantees are present it is the borrower's PD multiplied by the sum of LGD Amounts for the Collateralized and Unsecured portions of the facility plus, for each guarantee, the PD for the guarantee multiplied by its LGD. Implementation If (Eligible Amount for Guarantees) is not zero For each Guarantee associated with the facility Determine the EL amount (Guarantor PD x Guarantee LGD Amount) Determine the Total EL Amount for Guarantees EL Amount for Collateralized Portion = LGD Amount for Collateral x Borrower PD EL Amount for Unsecured Portion = LGD Amount for Unsecured Portion x Borrower PD EL Amount = EL Amount for Guarantees + EL Amount for Collateralized Portion + EL Amount for Unsecured Portion Otherwise EL Amount = Borrower PD x Total LGD Amount 9.11 CALCULATE EL AND LGD DATA FOR THE FACILITY After calling the main LGD Calculation algorithm, the following values are calculated: • Unsecured Amount = EAD remaining after allocation • Unsecured LGD Amount = Unsecured Amount x LGD% for unsecured portion • UnGuaranteed LGD Amount = Unsecured LGD Amount + LGD Amount for Collarateral • UnGuaranteed LGD Percent = UnGuaranteed LGD Amount / Portion of EAD not covered by guarantees • UnGuaranteed EL Amount = Borrower PD x UnGuaranteed LG Amount • EL Percent = EL Amount / EAD 76 9.12 CALCULATE EL AND LGD DATA ACROSS ALL FACILITIES Description Once EL and LGD data have been calculated for each facility, we can then produce summary level data across all the facilities in the system. The values calculated are as follows: • Total EAD = Sum of all EAD of all Facilities • Total LGD Amount = Sum of LGD Amounts for all Facilities • Aggregate LGD Percent = Total LGD Amount / Total EAD • Aggregate EL Amount = Sum of EL Amounts for all Facilities • Aggregate EL Percent = Aggregate EL Amount / Total EAD GUIDE TO BUSINESS ANALYSIS 77 CHAPTER 10 10 FACILITY SUMMARY 10.1 FACILITY SUMMARY The Facility Summary screen displays the list of facilities, with their status and the calculated LGD and EL values. It also shows the aggregated values, providing a picture of the overall facility risk for the customer. The LGD value is mapped to a Facility Grade, and the EL Grade represents a combination of the borrower grade and facility grade or a direct mapping from the EL value (This is configurable in Facility Configuration). 10.1.1 Customer information bar The customer information bar provides useful information about the customer, including the Borrower Grade and Equivalent PD. These values are taken from the Ratings Summary screen, and if an override has been authorized, the override values will be the ones displayed. The system also displays the 1 Yr EDF value as well as the Expected Loss PD. The latter value is the Equivalent PD or the 1 Yr EDF value divided by 100, depending on how the LGD model you are using has been configured. Again, if an override has been authorized, and the LGD model is configured to use the Equivalent PD, the system will use the Override PD in this calculation. 10.1.2 Switching between Proposed and Existing Positions The Facility Summary, Facility Detail, and CRM Summary screens include radio buttons to select Proposed or Existing Positions. This allows you to view the impact of proposed facilities and proposed Credit Risk Mitigants on your overall LGD and EL calculations. In effect, the toggle lets you choose which facilities and CRMs to include or exclude from the calculations. If you select Proposed Position, all committed and proposed facilities and CRMs will be taken into account. Alternatively, selecting Existing Position only includes those facilities and CRMs that are committed or completed. The radio buttons may appear grayed-out if you must be in a certain position to perform the current function. For example, if you try to edit a proposed facility from the Facility Summary dialog box, the system will display the Facility Detail dialog box locked into the Proposed Position. Since you must be in Proposed Position to edit a proposed facility, the system will not allow you to change to the Existing Position. 10.1.3 Selecting the Evaluation Date RiskAnalyst uses the evaluation date selected on the Facility Summary screen to determine maturity mismatches. The system uses the date selected in this field as the reference point against which to compare the maturity dates entered for facilities and CRMs. For example, if you select April of the current year as your evaluation date, CRMs and facilities with a maturity date of March of the current year will be considered by the system to have reached maturity. You can select a distinct evaluation date for the proposed position and for the existing position. Simply select the position, and then select your evaluation date. GUIDE TO BUSINESS ANALYSIS 79 CHAPTER 11 11 ARCHIVE 11.1 ARCHIVE OVERVIEW The Archive module captures all data in the customer database, both financial and rating, and stores it in an archive database. Archive allows you to capture data at key points in the case’s lifecycle so that you can record a ratings history of your borrowers. You can archive an individual customer manually, configure the system to automatically archive a customer based on its current state, or perform a batch archive on several customers at once. In addition to the data in the customer database, Archive also stores special archive fields, such as the reason for the Archive, which can later be queried in the database. While the system comes with some of these fields predefined, you can customize them and add new ones to allow for more successful queries. Archive also stores RiskAnalyst, financial template, and internal rating model version information to ensure the data is restored within the proper analysis context. For more information on archiving data, retrieving archived data, and configuring the Archive module, consult the RiskAnalyst Help, RiskAnalyt Batch Help, and Configuration Console Help. GUIDE TO BUSINESS ANALYSIS 81 APPENDIX A A APPENDIX A - INTERNAL RATING TEMPLATES A.1 The Internal Rating Template Concept Moody’s KMV is well known as a leading provider of benchmark rating models. We offer proprietary models developed to extremely high standards to assess both borrower and portfolio risk. The borrower models are developed using large data sets specific to regions of the world and provide very powerful external default predictions. According to recent proposals under the auspices of Basel II, lending institutions wishing to follow the ‘Internal Ratings Based Approach’ need to create and own their rating models and ensure that they meet certain criteria. The models should be tailored to their portfolio of customers which dovetail into their credit culture. Typically, the creation of an internal rating system is a major undertaking for an institution. It requires considerable resources to design and build both model and delivery platform. We believe that in many cases a vendor solution can provide a more expedient solution than the undertaking to create such a system from scratch. We supply tools and services to address this need. The RiskAnalyst platform is designed to assist institutions in deploying their models. Internal rating templates (IRTs) are designed to form the basis for the rating models delivered within the rating system. The aim of the IRTs is to provide pre-constructed borrower rating models that can be used by institutions as a starting point for their own rating models. In order to achieve this goal, IRTs (as shipped) have been designed with the following requirements in mind: • They should have the capability to rank-order borrowers effectively. • They should provide an internal rating model that can be used by lenders throughout the 14 world with limited configuration . • The analysis is to be based on a rating methodology that is intuitive to credit analysts. • The set of factors included should be broad, as it is easier to remove factors than add new ones. • The set of factors should span the key attributes that an experienced credit analyst would consider important to an analysis. • The factors should be tailored to be as objective as possible, supporting consistent evaluation by credit analysts. In order to attain these objectives, Moody’s KMV works with selected partners to create the IRTs. These partners typically comprise members of the RiskAnalyst Advisory Group (MRAAG). The MRAAG members are senior risk managers from credit institutions selected from the key users of Moody’s KMV internal rating solutions. MRAAG provides Moody’s KMV with strategic direction and expertise that is incorporated within the RiskAnalyst framework. 14 It is our view that it is not possible to ship an internal rating template that involves a judgmental analysis fully configured to predict default. GUIDE TO BUSINESS ANALYSIS 83 The partners’ role involves providing input into the IRT design and its parameters, reviewing the IRT, and providing data in the form of anonymous customer data together with the grades assigned to those customers. Approximately 100 customers are provided by each participant. A.2 The Methodology Used to Create Internal Rating Templates A.2.1 The Internal Rating Template Creation Process The following table outlines the approach that Moody’s KMV uses to create the IRTs. More detail about the participants follows the table. Phase Purpose/Task Participants Identify scope Determine which business areas to address (industry, size, regions) and the implementation framework required (scorecard, multi-layer). MRAAG, Moody’s KMV Identify partners Identify the partners who will participate and determine the roles that they will play. MRAAG, Moody’s KMV, Other interested parties Design Determine main components of IRTs, including preliminary (paperbased) tuning. Moody’s KMV Partner review Review specifications. Partners, Domain Experts, Moody’s KMV Build prototype Build a working prototype of the design. Moody’s KMV Acquire data Partners provide a pre-specified number of cases to support IRT verification. Partners, Moody’s KMV Tune Moody’s KMV and Domain Experts adjust IRT parameter values using expertise and a sample of the available data. Moody’s KMV, Domain Experts Test performance The IRT is tested against the remaining data and other quality standards. Partners, Moody’s KMV, Domain Experts Implement The model is implemented as a production quality piece of software into the RiskAnalyst framework. Moody’s KMV Quality assurance Testing is performed to ensure that the implemented IRT and its documentation meet Moody’s KMV standards. Tests are performed to ensure the IRT rates companies identically to the completed prototype Regression test scripts are created. Moody’s KMV Create documentation Creation of documentation to support the model. Moody’s KMV Domain Experts are experts in the particular area of lending that the IRT addresses. They may come from within Moody’s KMV or one of the organizations supporting the IRT development. Domain experts help with the design and tuning of the IRT. Partners are the organizations that have volunteered to support the verification of the development. Organizations may be Partners and provide Domain Experts. A.2.2 Tuning the Internal Rating Template Tuning is the process of specifying and adjusting the parameters used within the IRT. The approach used to determine the parameters of the IRTs uses a combination of expert judgment and data-driven optimization using sample data provided by the development partners. This approach supports the creation of an IRT with many parameters using limited data sets and 15 differs considerably to the approach we use for developing our quantitative models . This is due to several differences: 15 We provide detailed information on model development methodologies for our quantitative models in the form of white papers. These can be found at http://www.moodyskmv.com/research/defaultrisk.html. 84 1. Size of sample data sets upon which the IRTs are built. The IRTs evaluate judgmental inputs in addition to financial ones. It is not normally possible to obtain large numbers of cases with judgmental data within the timeframe for building the IRTs. This is due to the cost of acquiring the data; for each test case an experienced and knowledgeable credit analyst must determine the values for each judgmental attribute. 2. Optimization criteria. The IRT is optimized against the internal ratings assigned to businesses by experienced lenders. It is generally not possible to obtain large enough sample data sets that include judgmental factors to optimize against default information. In addition, it is problematic to determine the values of judgmental inputs retrospectively; for example, if a credit analyst knows that a customer has defaulted, this is likely to impact the analyst’s evaluations of the judgmental inputs. 3. Expected use of the IRT. It is envisaged that each purchaser of an IRT will make changes to ensure that it fits better with the local requirements associated with their credit culture and portfolio. Therefore, although the performance of the standard IRT should be good, it does not need to be as optimized as an external model that will be used without any subsequent changes. The hybrid IRT tuning approach is outlined below: Tuning Phase Activities Participants Initial Tuning Preliminary parameters specified by business experts. These parameters are estimated independently of the data. Partners, Moody’s KMV Configuration of Ratio Assessments The parameters for the sub-components of the ratio assessments (absolute, peer, slope, volatility) are largely driven from the data with expert review to determine the final parameters. The weightings to determine the ratio assessments from the subcomponents are derived in part by expert opinion and in part by looking at how well each sub-component correlates to the required ranking. To determine the parameters for each sub-component, a comparison is made between the component value for individual cases and the associated ranking of the case in the sub-portfolio of test-cases. For example, for the absolute component, the raw ratio value for each case is plotted against the case’s ranking within the sub-portfolio. This provides a curve from which parameter values can be inferred. A similar process is performed for the peer rank, the slope, and the volatility index. The sub-components are weighted together to determine an overall assessment. This weighting is determined by a combination of expert review and consideration of the correlation of each sub-component and the ranking of case within the sub-portfolio. For more information on the algorithm used to calculate ratio assessments, see the Guide to Business Analysis. Moody’s KMV Refinement of scores and weights associated with factors and sections The IRT is run against a sub-portfolio of the data provided by the 16 development partners and a genetic algorithm is used to identify potential improved models. The algorithm is constrained to prevent it suggesting models that diverge significantly from the original model determined using the expert’s views. Candidate models identified by the genetic algorithm are reviewed to determine potential improvements to the initial model. These improvements are evaluated based on their impact on model performance, and whether they make business sense. Improvements satisfying these criteria are incorporated into the final model. Moody’s KMV 16 A genetic algorithm is an iterative optimization technique based on evolutionary principles. It uses the following approach: Many possible models are generated randomly to form an initial generation. Each model within a generation is assessed against a function chosen to provide a measure of the desirability of that model. A probabilistic function is used to determine whether to replicate models or kill them off based on the desirability of the model. Remaining models are then combined and mutated to form the next generation using probabilistic criteria. The process continues iteratively. The function that we use to determine desirability considers two factors: the ability of the model to rank order the sample data in the way that the partners had, and the ‘closeness’ of the model to that specified by the experts. GUIDE TO BUSINESS ANALYSIS 85 A.2.3 Verification of the Internal Rating Template Performance The applicability and performance of the IRTs are evaluated using a combination of expert review and testing against customer data rated by experienced credit professionals. To distinguish this process from the testing performed on Moody’s KMV quantitative models, the term verification (as opposed to validation) is used. Verification is designed to achieve the following: • Determine that the IRT can rank order companies. This rank ordering should be reasonably robust across portfolios. This ranking should be evaluated against ratings determined by credit professionals for cases with which they are familiar. • Ensure that the IRT provides a sufficiently broad starting point for credit providers who wish to use it as a basis for their internal rating system. • Ensure that all the factors can be assessed as objectively as possible and that the data required to enter them is available. • Ensure that all interim assessments are meaningful. Verification is conceptually different to validation and calibration. The data sets available for the verification will be much smaller than those typically used within a validation exercise and the outputs would not be mapped to a probability of default (PD) estimate. Although validation and calibration can, and should, be performed on internal rating models, we believe this can best be performed in the context of the specific institution in which it will be used. Verification is performed during two of the phases within the model development process outlined above. Verification Phase Activities Participants Partner Review IRT design is reviewed against the following criteria: • Parameters are meaningful and widely applicable • Attributes are sufficiently broad to provide a good starting point for a credit provider’s internal rating system • Factors can be assessed as objectively as possible and the information required to enter them is available • Factor labels and answers are meaningful and intuitive • Help text is adequate to fully explain how to answer each question Partners, Domain Experts, Moody’s KMV Performance Testing The IRT is tested against data provided by the development partners. This is different data to that used for Tuning. The revised IRT and the revised parameters are reviewed by domain experts. Partners, Domain Experts, Moody’s KMV A.3 Configuring Internal Rating Templates The IRTs, when released, are designed to provide a robust starting point for a credit provider’s internal rating model. Experienced risk personnel have contributed to their design, and data sets from several portfolios have been used to construct and test the IRTs. However, Moody’s KMV expects that the performance of IRTs can be improved if adapted to the local factors of each credit provider’s portfolio and the credit culture they operate with. Indeed, one of the aims of these products is to provide lenders with IRTs that they can quickly and easily adapt to their specific requirements without having to build an internal rating model from scratch. We call the process of adapting IRTs to the local environment localization. Localization is important for a number of reasons. These include: 86 • The importance of factors differs between portfolios. Those included in the IRT are designed to be general and therefore are not optimized for your portfolio. Some factors may be less relevant to your specific portfolio than is generally the case. • There may be factors that are not included in the model that your organization deems important or useful to the analysis of its borrowers. • Factors included in the IRT may prove problematic for your credit analysts and/or relationship managers. For example, the information required to answer specific factors may be difficult or costly to accurately obtain. • Terminology issues: nuances of specific questions or answers may differ in different regions. • Business norms differ in different regions of the world and this may reflect in the benchmarks against which the financial factors are assessed. Therefore, it is important that a review of the IRT is performed. Based on our experience, this review should occur over a period of time as you learn more about the internal rating model and the way it behaves in your environment. It is important that you test your internal rating model using a data set that is representative of the borrowers that you rate. In addition, as grades assigned to borrowers normally differ between lending institutions and as 17 Probabilities of Default (PDs) vary between portfolios , a calibration of grades and PDs will need to be performed. Normally it is not possible to properly calibrate the internal rating model before a significant amount of data has been fed through it and experience has been gained of the borrowers rated through the system. Often institutions perform an initial calibration of internal rating models to their grading system but avoid calibrating to PDs until this can be done with a greater level of confidence. Another important step is internal rating model validation. This task involves determining how accurately the internal rating model predicts default and the confidence with which that measure of accuracy can be stated. This section provides an overview of the process required to configure an IRT. While reading this section, you should bear in mind that it is our experience that there is no one correct or normal way of configuring and implementing an IRT. Instead, we find that institutions have their own unique circumstances and requirements and that the configuration life-cycle varies depending on these circumstances. A.4 A Typical Internal Rating Template Configuration Life-Cycle The following diagram represents a typical set of tasks involved when building an internal rating model from scratch and implementing it within a credit risk rating system: 17 The IRTs are not shipped with mappings to default probabilities. GUIDE TO BUSINESS ANALYSIS 87 Model Set-up Pilot Production Factor Selection Model Design Build Model Tuning Pilot Production Model Review Model Review Val/Cal Tailor Tailor Tailor A typical internal rating model development life-cycle With the IRTs, a significant part of this work has already been performed. This is reflected in the following diagram. Model Set-up Pilot Tuning Pilot Production Model Review Model Review Val/Cal Tailor Tailor Tailor Production A typical configuration life-cycle for an IRT As is illustrated in the diagram, the development can be structured into three broad phases. Model Set-up involves work performed on the IRT prior to use. This may involve some data gathering, and is likely to require input from your credit officers. The pilot phase normally involves a limited rollout of the IRT operated in parallel to an existing rating process. This allows you to test it in a real environment and to capture data necessary for further model optimization and testing; note that while financial data can often be obtained retrospectively, it is problematic to do this with judgmental inputs. 88 In production, the IRT is implemented as an internal rating model. The production phase, or phases, involve(s) using the internal rating model in a live environment, typically without the back-up of running another rating system in parallel. The internal rating model will be run with regular reviews of its performance and re-configuration as necessary. Often the reviews are performed annually. The length and content of each phase will vary depending on requirements and experience. One of the key design aims for IRTs was to minimize the effort required within the model set-up phase. However, there is naturally a trade-off between the effort that you put into the model development, the timescales, and the quality of the internal rating model released into production. This guide is intended to assist in understanding and making these decisions. The following sections provide more detail on each of the phases and tasks. Details of phases are presented separately to the tasks that may be performed within them. A.4.1 Model Set-up The Model Set-up phase is geared towards ensuring that the IRT is ready for use as part of a pilot. The objectives of this phase are: • Identify those responsible for agreeing that the model is ready for pilot • Ensure they understand RiskAnalyst and the IRT sufficiently • Determine quality criteria for release of model into pilot • Determine data reporting requirements for pilot • Verify that the IRT adheres to the quality criteria and amend if necessary The scope of this phase will vary. Many organizations spend very little time on this phase, 18 possibly just holding a workshop to iron out issues and determine a path forward . It is important that you understand the IRT and understand any compromises made for the sake of expediency. The following paragraphs outline considerations and tasks relevant to this phase. Identify staff It is recommended that staff from both the relationship and the review functions are involved at this stage. The staff selected should be experienced; they should understand how the lending processes currently works, and should work. In addition, it is also recommended that IT staff is involved as they will need to support the implementation and manage the data captured for analysis. Training We provide documentation on the structure and use of our tools and IRTs. In addition, we offer a number of training courses on the structure, use and configuration of our products. Review IRT factors Review the qualitative factors in the IRT. Ensure that these are applicable to your portfolio or customers. Also review the possible answers to these questions and again determine whether they are meaningful for your portfolio and can be answered by your users. 18 Moody’s KMV can lead such workshops if desired. GUIDE TO BUSINESS ANALYSIS 89 Consider whether there are factors not contained in the IRT that would normally be considered important for your portfolio of customers. If you are interested in data on qualitative criteria, or criteria that cannot be determined from the financial statements or other sources, it may be advisable to add these at this stage. This will allow analysis of these factors using the data captured during the Pilot. Review the ratios used within the IRT. Are there any ratios that your users would normally consider important that should be added? If a similar ratio exists within the IRT, you may choose not to add the new ratio at this stage. Note that the financial spreads entered by the users in the pilot will be retained and these can be used to retrospectively determine the values for ratios not initially included in the IRT. If new factors are added, it will be necessary to assign scores to these. Review IRT parameters (scores and weights) At this stage in the internal rating model development life-cycle, it is not normally 19 recommended that significant changes are made to the IRT’s parameters . However, if you wish to analyze a portfolio that is significantly different to that upon which the IRT was based, 20 then this may be necessary . Clearly, if new factors are added to the IRT, care needs to be taken to ensure that the parameters used to assess these produce good results. Adding and deleting factors requires use of the Internal Rating Model Author. See section A.6 Making Changes on Your Own for more information. In addition, one area that often merits some review is the set of scores assigned to each industry. These are deigned to be widely applicable and your knowledge of your local markets may indicate that some changes should be made. Mapping the score to your Grading System The IRT can be used within the Pilot without any grades being visible, or alternatively you can create a mapping from the IRT scores/assessment means to a grading system. You may choose a temporary grading system or use your existing one. Neither solution is without problems. Using your established grading system has the merit that the meaning of the internal rating model’s output will be readily understood by pilot users. However, some care should be exercised; if the system is mis-calibrated, this familiarity may be misleading. In addition, when corrections are made (if significant), this can further impact confidence in the system. Therefore, if you use this method, it may be worth undertaking the effort to run a number of cases through the system first to form a basis for calibration. Alternatively, if you choose to use an alternate grading system, you will avoid the risk of confusion if the grades are mis-calibrated. However, you will place an additional learning overhead on the users of the system as they will need to relate to a different grading system. 19 You may choose to perform a thorough overhaul of the parameters, but this would normally require analyzing a number of customers within the system to capture data and determine levels for parameters. Normally a review of this level would be performed after the Pilot. 20 To illustrate this, consider using the Middle Market IRT to assess quoted entities. Short term funding for quoted companies is very different and therefore the whole area of liquidity is very different. In addition, intangible assets are often more important and quantifiable for quoted companies (consider the brand of a large supermarket vs. a small convenience store). Therefore, assessments of ratios based on tangible net worth will not adequately capture the business fundamentals. 90 Mapping Grades to PDs Normally, creating a mapping of the internal rating model output to PDs requires significant data. This data should be representative of the portfolio under consideration and include sufficient defaults. Ideally, this data would be gathered from the portfolio under consideration, but this is not always practical. We can provide support in this area. Our Modeling Services team has experience performing such work and can use our Credit Research Database (CRD) to augment your data (financial values only), if applicable. Pooling data with other institutions 21 may provide an alternate solution. Overrides If you have set up a grading system for the pilot, then you will almost certainly wish to allow the users to perform overrides. By capturing override information, you will see where user views and assigned ratings differ to those calculated by the IRT. In addition, override reasons allow capture of a reason for the override. This can be used to support the analysis of the system’s performance after the pilot. The set of override reasons you choose for your pilot may be more extensive than those that you would use in a live environment. Consider whether multiple variants of the IRT will be used If you wish to use the IRT over a number of business units, or over different portfolios or regions, you may wish to ultimately have multiple variants of the IRT. If the IRT will be used for multiple segments, is your strategy to have one internal rating model, or multiple variants? If the latter, consider what differences may be required between the internal rating models and whether this has implications for the IRT(s) used in the pilot. Bear in mind that each time you split your portfolio into separate internal rating models, you reduce the data set available for future validation and calibration. Consider Peer Data Determine whether to use the peer database that is shipped with RiskAnalyst, or create a custom one. As is noted below in section A.4.4, the Middle Market IRT was built around the RMA (Risk Management Association) peer database. The results proved robust even though some of the portfolios used in optimization and testing did not come from North America. Therefore, it may be more expedient to use the RMA database during the pilot, particularly in view of the time and data constraints involved in the creation of custom peers. However, it may be beneficial during the set-up or pilot phases to review the peer quartile values against local benchmarks and adjust the ratio assessments where necessary. Implementation Issues This guide is not intended as an implementation guide for RiskAnalyst. However, this section contains some brief notes on this topic that are particularly relevant to implementing the IRT. • Implementation Architecture – consideration needs to be given to factors such as how the internal rating model will be distributed to its users, how updates are distributed, and where databases storing parameters and customer data will reside. In addition, standard procedures need to be followed (e.g., around security, data back-ups, etc). • Data considerations – the data captured during the pilot will be used to optimize and test the IRT before releasing it into a live environment. This data needs to be available for review at the end of the pilot. Consider whether this data should be taken directly from the customer database, or from the archive database. The customer database will contain the 21 There are other methods that are used to derive PDs. One method involves using a reference rating provided by a rating agency (or other source) where the PD is known. GUIDE TO BUSINESS ANALYSIS 91 current state of each case, and therefore will contain the customer data correct as at the end of the pilot. Customers can be archived to an archive database at set points in their lifecycle (e.g., when a borrower grade is agreed or facilities are approved) and snapshots of the case recorded for subsequent analysis. • Extra pressure on your users - the pilot will need to be run in parallel with your existing rating system. The users will be performing more work. Can anything be done to assist them? • Linking allocated grades to the grades/scores assigned by the IRT - when reviewing the pilot, it will be necessary to compare the scores calculated by the IRT to those actually allocated to the borrowers and also, if different, those determined by your existing rating system. • Business units/groups involved in pilot - if the IRT will eventually be used by multiple segments of your business (different divisions, portfolios, regions, etc.), you may wish to select more than one segment for the pilot and carefully select the segment(s) to maximize the utility of the pilot. • Integration within your enterprise - consider any integration requirements with your existing systems. Normally we advise customers to limit integration work at this stage and use manual processes instead. This reduces the upfront project scope and risk. A.4.2 Pilot Phase The Pilot Phase is designed to allow you to test out the IRT with real customers using a process similar to that intended for your production rating system. We advise that the pilot is performed in parallel to your existing rating system. This serves two purposes. The first is that it reduces the risk that you make erroneous lending decisions, and the second is that it makes it easier to identify differences between old and new systems and to determine whether these are desirable or not. The objectives of the Pilot Phase are: • to assess the IRT’s performance both quantitatively and qualitatively, • to identify and mitigate any model deficiencies, • to gather data to enable you to determine how the IRT can be best optimized for your portfolio, • to assist you to determine whether to remove factors from the IRT (and possibly add new ones), • to gather sufficient data to perform a mapping of the IRT scores to your grading system (it is unlikely that the pilot will have sufficient duration or scope to support a mapping to PDs without supplementary data), • to gather data that could be used to create a peer database (if required). The following paragraphs outline items handled at the beginning or during the Pilot Phase: Select Pilot Users Determine which business area(s) will be involved in the pilot. If multiple segments of the business will be using the IRT, you may wish to include more than one within the Pilot. 92 Train Pilot Users Provide training for the users of the Pilot system. We can deliver this training ourselves or train your staff to deliver it. The latter approach is often used by our clients. Implementation Issues As mentioned above, this document is not an implementation guide and the following provides a summary of the tasks required: • Set up security profiles for users • Roll-out system to pilot users • Ensure that there are mechanisms to update the IRT or RiskAnalyst if issues arise that you choose to address during the pilot phase • Ensure that the data captured during the pilot is available for analysis Data Capture and Quality Issues During the pilot, the data captured should be monitored. Useful information may be obtained about completeness and quality that can be addressed prior to the completion of the pilot. The following tasks are performed at the end of the Pilot Phase to enhance the IRT and ensure that it is ready for production. Assess and Review IRT Performance This task is typically addressed in several ways: • The users of the system will be able to provide valuable insights into problems that exist with the system. For example, certain questions may be difficult to answer, or the ratings for some sub-assessments or borrowers may not fit with the user’s view. • The actual ranking performance of the system should be compared to a benchmark, possibly ratings assigned by your credit personnel, default probability measures from other 22 sources , if available, or a default study if sufficient data is available. • Review the usefulness of the factors. Do they contribute to the quality of the rating process and do they improve the power of the IRT? Peer Database Considerations Review whether it is desirable to continue using the RMA peer database or whether you want to create a peer database based on the data collected in the pilot or from data available from a third party supplier. Alternatively, differences may often be addressed by tuning a ratio’s peer assessment. Optimize IRT Make amendments to factors, if necessary, based on the IRT performance review. Then optimize the IRT parameters. A number of techniques are available to perform this task. These include the process used to create the IRT and the use of statistical regression techniques. Our Modeling Services team will often use the latter approach. 22 For example, Moody’s KMV RiskCalc product. GUIDE TO BUSINESS ANALYSIS 93 Test Optimized IRT Test the improved IRT to determine how it performs against the data collected within the Pilot. You may choose to segment the data collected during the Pilot to ensure that the data used to test the IRT is not used for optimization. Calibrate the IRT to your grading system Review the IRT scores produced using the data gathered during the Pilot. Determine cut-offs for each grade to allow a mapping to be derived from the IRT’s score to your grading system. If data is available, determine PD values for grades If sufficient data is available, determine the PDs for the grades, or specific score ranges. Review Overrides Based on the experience of the pilot, review and amend the override reasons. A.4.3 Production Production involves running the IRT in a live environment. At this point, the IRT becomes an internal rating model and should be robust and effective. However, there will still be a need to review and improve the internal rating model at regular intervals. The objectives of this phase, from an internal rating model maintenance/development perspective are: • Continue to gather data to support optimization, validation, and calibration of the internal rating model • Monitor performance • Identify any deficiencies and amend The following paragraphs provide further information on tasks and considerations. Data Capture Unlike the Pilot Phase, where there is an ‘end,’ at any point in time in the Production Phase, it is likely that cases will be at various stages within their life-cycle and at various stages of completeness. Some cases will contain work in progress, with partial data, while others will reflect the state at a particular stage (e.g., as at the last loan review). It is recommended that the data used for internal rating model development contains only cases at fixed points in their lifecycle that are in a ‘completed’ state. One way to address this is to use the Archive functionality and ensure that cases are archived at key decision points. This will allow you to pick off cases that are in a ‘completed’ state and also determine the status of the case at that point in time. In addition, the history of the case is maintained using the Archive, enabling you to review each case at its key decision points. Data Quality and Monitoring It is important to monitor the use of the system to ensure that it is being used consistently and properly. Considerations for data quality include: • 94 Is the system being used by all users and regions? • Are cases being archived when they should? • Are there differences in usage between different divisions or regions? If so, why? • Are there patterns in the use of the system that indicate some sort of gaming? For example, you could measure the ratio of judgmental scores to financial scores produced by the model and see if this varies across user groups or across time. Assess and Review Internal Rating Model Performance At regular periods, the performance should be reviewed against loss experience to ensure that the internal rating model performance is not deteriorating. Feedback from users should be obtained, as this might indicate changes in the environment that have a negative impact on performance. Re-optimize It may be considered important to add, amend, or remove factors within the internal rating model based on the review. In addition, it may prove beneficial to amend the internal rating model parameters periodically as new data becomes available. This task is likely to be performed less often than the review process or the process of calibration to PDs. It is generally considered important to ensure that the internal rating model remains relatively stable. Validate Periodically, it will be necessary to formally determine the internal rating model’s performance at predicting default. Calibrate The internal rating model will need to be recalibrated at regular intervals to ensure that the PDs it produces remain accurate. Considerations for multiple versions If the internal rating model addresses multiple segments of borrowers, the performance for each segment should be considered. It should be determined whether a single internal rating model can be used across segments and, if not, how different the variants should be (i.e., just different parameters, or different parameters and factors). A.4.4 Some Additional Information on Configuration Tasks This section supplements the previous three, providing additional information on key topics. Factor/ Internal Rating Model Review In the text above, the topic of reviewing the internal rating model was covered. Often this task will need to be performed without significant data to support decision making. The following notes are intended to provide a set of steps to address this topic: • Perform an overall review of the internal rating model: o Does the structure (its division into sections) fit with the way your organization looks at borrowers? GUIDE TO BUSINESS ANALYSIS 95 o • • Are you happy with the weightings assigned to the different sections (for example, you may wish to increase the impact of the Historical Financials at the expense of the judgmental sections that are more open to manipulation)? Perform a review of the judgmental factors. Consider the following: o Are the judgmental factors appropriate to your area of operations and its culture? o Can data be readily accessed to enable users to answer them? o Are the answers appropriate? o Is the Clarify text appropriate? Does it provide sufficient guidance for the users? o Do the scores and weights assigned appear in line with your experience? o If data is available, does the data support or contradict the weights and scores? Perform a review of the ratios and their assessments. Ideally this would be performed after gathering data, but value can be gained by performing this task judgmentally: o Do you wish to retain all the ratios? Note that some had limited power against the portfolios used to develop the IRT. They were nevertheless retained because they are commonly used in our clients’ internal rating systems. o Is financial data available and sufficiently robust to allow the ratios to be calculated reliably? o If peer data is available, are the Peer Assessment fixed points/votes appropriate? Note that for some of the ratios in the IRT, RMA peer data is not available. o Are the Absolute fixed points/votes appropriate? o Are the slope fixed points/votes appropriate? Adding Factors to the IRT The technical details of adding factors to an internal rating model are described in the Internal Rating Model Author Help sytem. The following notes are intended to provide supplementary advice: • When adding a section: o • 96 Do any of the existing factors fit better on this new section? When adding judgmental factors: o Is the question answerable? Is the information readily available to the user? o Are the answers as clear and objective as possible? Note that there is often a trade-off between utility and objectivity when designing questions. Sometimes it takes several objective questions to address the issue that a single subjective question (“How do you rate the quality of the management?”) could address. However, the more general and subjective a question, the less consistency there will be with the answers provided by the users and therefore, the less it can be relied upon. o There is merit in keeping the number of answers for each question to a minimum. If you create a question with many possible answers, and wish to use a data-driven approach to determine/refine the scores assigned to each answer, you will need larger data samples to ensure that you have sufficient number of cases with each answer selected. o Creating good Clarify text can be challenging. Consideration should be given to determine what the user needs to know, and what should be considered to correctly answer each question. We put considerable effort into creating Clarify for the questions within the IRT. Care was taken to make things as clear as possible to the user, often explaining each answer in turn. This took time, but did lead to us redefining ambiguous questions and answers. • • • When adding ratios: o Can peer data be obtained? If not, is this important? Can the ratio be assessed effectively using absolute and trend assessments alone? o Are the accounts required to determine its value normally available? Are they defined in the chart of accounts that you are using? o Is the proposed ratio very similar to an existing ratio? If so, should the existing ratio be removed? When adding values where currency matters: o If all of the borrowers in your portfolio report their accounts in a single currency, then this topic is unlikely to be of concern. o Otherwise, the following issues need consideration. The value of financial items may change if entered in, or targeted to different currencies. For example, non-ratio values (e.g., total assets) will clearly be different when specified in different currencies. In addition, ratio values that use accounts from more than one statement (e.g., growth ratios, average ratios, and cash flow ratios) will change if converted into a different currency if there are movements between the currencies. The latter issue will have a relatively minor impact unless there are significant movements between the currencies. In addition, there may be some debate about which currency is ‘correct,’ as this may 23 depend on perspective . Therefore, in the interests of simplicity of internal rating model construction and use, you may choose not to address the issue directly within 24 the internal rating model . The currency assigned to a model can only be changed using the Internal Rating Model Author. Currently our IRTs use ratio values, so there is minimal concern for currency differences. To include an EDF measure as a component to the IRT: o Consider whether the EDF measure should complement or replace the Historical Ratio Assessments. o You will need to determine if the EDF will feed into the internal rating model, and if so, how its contribution will be utilized. One method might be to allow a certain contribution to be driven directly from the EDF produced. o You can also use an EDF to constrain the borrower grade, such that the final grade cannot vary more than a certain amount from the EDF grade. Amending/Specifying Parameters for Factors There are many ways in which the model parameters can be derived. The best method, if there is such a thing, will vary depending on the expertise available (business, statistical/data mining), and available data. Consequently, instead of prescribing an approach, the following notes provide some insights into the approach that we used to create IRTs. • The IRTs are designed to follow a methodology similar to one that a human credit expert would use. Typically, this means that a larger number of factors are considered than would be used with a model that was based on a statistical design where parsimony is desirable. As a result, the IRTs typically have many parameters. To fit this large number of parameters using purely statistical techniques is problematic, especially when the data sets used are small. Therefore, when building the IRT, human expertise forms the basis for most of the 23 For example, if a borrower reports in Argentine Pesos but must make repayments in US Dollars, are you interested in US Dollar Sales Growth and Cash Flow Coverage or Argentine Peso figures? 24 It may be wise to monitor the effect of such issues if your portfolio consists of borrowers reporting in many currencies. GUIDE TO BUSINESS ANALYSIS 97 parameters, with data being used to adjust the parameters and spot misconceptions made by the experts. The only place where the data is used to drive the base parameters is for the sub-components of the ratio assessments. To gain more detailed information on our approach, see Section A.2.2, Tuning the Internal Rating Template. • When designing the model, we aim to score each attribute within a range of (0, 100), with 50 being a normal or average value. For some attributes, a skewing is applied. For example, in most Middle Market companies intangibles have very little intrinsic value. Consequently for ratio assessments we use Tangible Net Worth (TNW) in places where an equity value may be used and we penalize negative TNW harshly. However, some entities will have intangibles that have real value and we wish to compensate for this and improve the overall assessment under these circumstances. The question ‘Intrinsic Value of Intangibles’ is designed to capture this information and to improve the score when appropriate. Hence we skewed the score. • Each section has an overall score. When designing the IRTs, we aim to ensure that this is also scored in the range (0, 100). As some of the answers to particular questions are skewed, it is necessary to adjust the scores assigned to questions within each section to satisfy this 25 requirement . If data is available, statistical techniques are often used. Our Modeling Services team has used such techniques to optimize many customers’ internal rating models with considerable success. Peer Database The peer database is used by the IRT to determine an assessment for each ratio. The database consists of quartile values for each ratio, for different industry groups. The quartile values th th comprise the median, upper quartile (75 percentile), and lower quartile (25 percentile) values for the peer group and ratios. They allow ratio values calculated for the borrower to be compared against the borrower’s peers, and for a ranking to be determined. RiskAnalyst is shipped with a peer database provided by RMA (Risk Management Association). This database is built from North American data and so may not be directly equivalent to the peers for the borrowers that you will rate. However, this database is used by Moody’s KMV to determine parameters for the ratio assessments. It is also used to test the IRTs, and so it is anticipated that the assessments produced will be robust for many portfolios in different parts of the world. A custom peer database can be created if you have sufficient financial data or can use data supplied by a third party. We provide a tool, Moody’s KMV Benchmark, to create a peer database from financial statement data. A custom peer database, as well as providing better peer ranking within your portfolio, will allow you to create peers for new ratios that you may choose to create, and for ratios within the IRT where RMA does not calculate peer data. Financial Conditions It is possible to specify conditions that prevent a borrower rating from being displayed in RiskAnalyst. Conditions could include: 25 • the balance sheet balances, • retained earnings reconcile across statements, and • a peer group has been selected. This can have a minor detrimental impact on model performance. 98 You may wish to add extra conditions. Internal Rating Model Validation and Calibration This guide is not intended to address the detail of model validation and calibration. These topics require a more thorough coverage than is possible to include here. Our Modeling Services Team specializes in performing such work for customers and we have published several papers 26 on the topic. These are available on our web site . A.5 Configuration versus Customization The sections above are all geared to configuring the IRTs. We have used the word configuration to mean a specific set of changes that can be performed without changing the platform. As is implied above, these changes broadly consist of: • Adding, amending, or deleting sections • Adding, amending, or deleting judgmental and other user input factors that appear within sections • Adding, amending, or deleting ratio assessments • Changes related to the inclusion of, or not, factors currently captured in the RiskAnalyst Financial Package (e.g., audit method, financial statement values, ratios not assessed using ratio assessments) • Including EDF measures (where these are available to RiskAnalyst) • Specifying a mapping of the IRT’s score to your grading system • Specifying a mapping of the IRT’s score to a PD • Specifying override reasons You may wish to make additional changes. Usually these will involve a change to the platform or the creation of an additional program that links into RiskAnalyst. Such changes are often termed customization and can be performed by Moody’s KMV Professional Services Team. A.6 Making Changes on Your Own A.6.1 Configuration Changes Moody’s KMV provides RiskAnalyst Studio, a suite of tools which support making configuration and tuning changes. This tool enables non-technical staff to perform many of the configuration changes to IRTs with little or no technical support. The components of RiskAnalyst Studio used to configure IRTs, and included with RiskAnalyst, are: Configuration Console. The configuration console allows you to modify grade and PD mappings, constraining options, enter override information, as well as configure presettable inputs and conditions. Tuning Console. The Tuning Console allows you to modify the weights and votes, or scores, associated with individual factors and sections of an internal rating template. Additionally, clients can license the Internal Rating Model Author as part of RiskAnalyst Studio. With this application, you can add or remove factors, configure almost any facet of a model, and create your own internal rating models. This application requires a greater degree of 26 See the section on Model Validation and Testing: http://www.moodyskmv.com/research/riskmodeling.html. GUIDE TO BUSINESS ANALYSIS 99 knowledge and technical skill to use properly. If you are interested in purchasing the Internal Rating Model Author, contact your Moody’s KMV representative. Training is included, and required, as part of licensing the Internal Rating Model Author. Also in RiskAnalyst Studio, the Facility Configuration application allows you to configure many of the parameters of the facility model included in RiskAnalyst. A.6.2 A Note on Versioning In order to assist with version control, versioning information is stored with the various components of the product. If you are configuring the product, it is important that you are aware of these. Versioning and the Internal Rating Model Author/Tuning Console The Internal Rating Model Author or Tuning Console maintains the version number of internal rating models. The version is identified by a four-point version number. The first two digits represent the earliest RiskAnalyst release version (major, minor) with which the internal rating model is compatible. The third digit represents the Moody’s KMV internal build number. The fourth number contains the configuration build number from the Internal Rating Model Author or Tuning Console. The version is stored with the internal rating model when a case is archived. Financial Templates are versioned in the same way; this number is held in a table called ModelVersion. If, as part of a configuration project, any of the financial template tables are updated, then the fourth digit of the template version should also be incremented. A.7 The Facilities Module and The Internal Rating Templates RiskAnalyst is shipped with a facility model that calculates LGD (Loss Given Default). This is based around the methodology prescribed under the Basel II Foundation Approach. The model, as shipped, is configured for the types of facilities and credit risk mitigants common in middle market lending. While this guide does not describe how to configure this module, it is important to know that a 27 PD must be supplied for this module to work . This PD can be derived from an EDF model (if available to RiskAnalyst), or the PD calculated by the IRT can be used. If you have access to an EDF model within RiskAnalyst and wish to use the facilities module, then we recommend that you use the EDF model to drive the PD until you have sufficient data to calibrate the IRT to a PD. Subsequently, we anticipate that you will choose to use the PD derived by the IRT. Further details on how to perform this change can be found in the Facility Configuration Help in RiskAnalyst Studio. 27 Note that, although in its simplest form LGD is independent of PD, a PD is required to determine the eligibility of guarantors (their PDs must be less than the borrower’s), and this can impact the overall LGD. 100 GUIDE TO BUSINESS ANALYSIS 101
Source Exif Data:
File Type : PDF File Type Extension : pdf MIME Type : application/pdf PDF Version : 1.4 Linearized : Yes XMP Toolkit : 3.1-701 Producer : Acrobat Distiller 7.0.5 (Windows) Create Date : 2008:06:18 14:19:50-04:00 Creator Tool : Acrobat PDFMaker 7.0.7 for Word Modify Date : 2008:06:18 14:20:16-04:00 Metadata Date : 2008:06:18 14:20:16-04:00 Format : application/pdf Title : Guide to Business Analysis Creator : Moody's KMV Document ID : uuid:245992fd-3f49-4b07-9fbc-d116dd8bdb53 Instance ID : uuid:7c0ba29b-a6ac-4437-9209-0522f592e5d4 Company : Moody's KMV Page Count : 102 Page Layout : OneColumn Author : Moody's KMVEXIF Metadata provided by EXIF.tools