1 / 49

Validation of Rating System

Validation of Rating System. 1. Outline. What is validation of rating systems? Two components of validation of rating systems discrimination, calibration. Discrimination methods Calibration methods. 2. What is validation of rating systems?. 3.

jabir
Download Presentation

Validation of Rating System

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Validation of Rating System 1

  2. Outline • What is validation of rating systems? • Two components of validation of rating systems • discrimination, • calibration. • Discrimination methods • Calibration methods 2

  3. What is validation of rating systems? 3

  4. What is validation of rating systems? Having set up a rating system, it is obvious that we wants to assess its quality. Validation of rating system rely on using a lot of distinguish methods to assess its quality. There are two dimensions along which rating systems are commonly assesed: discrimination and calibration 4

  5. Components of validation of rating systems 5

  6. Components of validation of rating systems Discrimination: • In checking discrimination, we ask: How well does a rating system rank borrowers according to their probability of default (PD)? Calibration: • When examining calibration, we ask: How well do estimated PDs match true PDs? 6

  7. Discrimination methods 7

  8. Discrimination methods Accrording to Basel Committeeon Banking Supervision we can mention the following statistical methodologies for the assessment of discriminatory: • Cumulative Accuracy Profile (CAP), • Accuracy Ratio (AR), • Receiver Operating Characteristic (ROC), • ROC measure (AUC) approximated by Mann-Whitney statistic, • Pietra Index – approximated by Kolmogorov-Smirnov statistic, • Conditional entropy, Kullback-Leibler distance, • Conditional Information Entropy Ratio (CIER), • Information value (divergence, stability index), 8

  9. Discrimination methods Accrording to Basel Committeeon Banking Supervision we can mention the following statistical methodologies for the assessment of discriminatory (continue): • Bayesian error rate, • Kendall’s τ and Somers’ D (for shadow ratings), • Brier Score. 9

  10. Receiver Operating Characteristic (ROC) as a function of cutt-off 10

  11. Receiver Operating Characteristic (ROC) 11

  12. Receiver Operating Characteristic (ROC) 12

  13. Receiver Operating Characteristic (ROC) (the first approach) The area A is 0.5 for a random model without discriminative power and it is 1.0 for a perfect model. It is between 0.5 and 1.0 for any reasonable rating model in practice. Matlab file: ROC.m Call: [xy,AUC]=ROC(ratings, defaults) input: ratings – column vector of rarting value (higher value denote more risky) defaults – 1 denote default, 0 denote non-default Output: xy- points collection to plot ROC, AUC- Area Under Curve -------------------------------------------------------------------------------------------------------------- Matlab file data example: capexample.mat 13

  14. Receiver Operating Characteristic (ROC) (the second approach) Matlab file: groc.m Call: [AUC,AUC_m]=groc(pd,is_default,spos,is_pic) input: pd – column vector of rarting value is_default– 1 denote default, 0 denote non-default spos – three aproaches is_pic – show curve Output: AUC- Area Under Curve using empirical data AUC_m - Area Under Curve using aproximation 14

  15. ROC measure (AUC) approximated by Mann-Whitney statistic Assumption:The scoresof the defaulter and the non-defaulter can be interpreted asrealisations of the two independent continuous random variables SD and SND The area under theROC curve (AUC) is equal to the probability that SD produces a smaller rating score than SND. AUC=P(SD<SND) This interpretation relates to the U-test of Mann-Whitney. We split default vector on two vector: defaulters and non-defaulters and calculatethe test statistic Û of Mann-Whitney which is defined as where Û is defined as 15

  16. ROC measure (AUC) approximated by Mann-Whitney statistic Matlab file: AUC_Whitney.m Call: [U]=AUC_Whitney(ratings, defaults) input: ratings – column vector of rarting value (higher value denote more risky) defaults – 1 denote default, 0 denote non-default Output: U - the test statistic Û of Mann-Whitney -------------------------------------------------------------------------------------------------------------- Matlab file data example: capexample.mat 16

  17. Pietra Index – approximated by Kolmogorov-Smirnov statistic It is possible to interpret the Pietra Index as the maximumdifference between the cumulative frequency distributions of good andbad cases. Interpreting the Pietra Index as the maximum difference between the cumulativefrequency distributions for the score values of good and bad cases makes itpossible to perform a statistical test for the differences between these distributions.This is the Kolmogorov-Smirnov Test (KS Test) for two independent samples. The null hypothesis tested is: The score distributions of good and bad cases are identical. It mean that when the null hypothesis wil be rejected so significant differencesexist between the rating values of good and bad cases (discrimination). 17

  18. ROC curbe model Y=normcdf( a + b * norminv(x) ) 18

  19. Pietra Index – approximated by Kolmogorov-Smirnov statistic Matlab file: pietra.m Call: [pietra_index,H,confidence_level]=pietra(rating,is_default,q) input: ratings – column vector of rarting value (higher value denote more risky) is_default – 1 denote default, 0 denote non-default q - significance levels (default q=0.05) Output: pietra_index - asymptotic P-value H : 1-The null hypothesis was rejected (discrimination) 0- The null hypothesis wasn’t rejected confidence_level – 1-q (default 95%) 19

  20. Cumulative Accuracy Profile (CAP) and Accuracy Ratio (AR) We present two ways of calculation of Cumullative Accuracy Profile and Accuracy Ratio: • (1) the first approach: calculation based on sorted rating e.g. (AA,B...), • (2) the second approach: calculation based on rating which is number without classifying to concrete well-known rating e.g. (AA,B...). Characteristic of the first approach (discrete approach): • calculation based on every rating so we have the same number of points on the curve as number of group of rating, • advantage: this way is less risky because we avoid sorting problem • disadvantage: the curve is less accurate (number of points=number of distinct ratings • Characteristic of the second approach (continuous approach): • calculation based on every observations (so we have the same number of points on the curve as number of observation, • advantage: the curve is more accurate than first approach • disadvantage: this way is more risky because in the case the same value of rating can occur sorting problem 20

  21. Cumulative Accuracy Profile (CAP) (the first approach) 21

  22. Cumulative Accuracy Profile (CAP) (the first approach) The cumulative accuracy profile (CAP) provides a way of visualizing discriminatory power. The key idea is the following: if a rating system discriminates well, defaults should occurmainly among borrowers with a bad rating. To graph a CAP, we need historical data of ratings and default behavior. The example was presented on the below picture 22

  23. Accuracy Ratio (AR) (the first approach) An accuracy ratio (AR) condenses the information contained in CAP curves into a singlenumber. It can be obtained by relating the area under the CAP but above the diagonal tothe maximum area the CAP can enclose above the diagonal. Thus, the maximum accuracyratio is 1. We compute the accuracy ratioas A/B, where A is the area pertaining to the rating system under analysis, and B is theone pertaining to the ‘perfect’ rating system. 23

  24. Cumulative Accuracy Profile (CAP) (the first approach) Matlab file: CAP.m Call: [xy,AR]=CAP(ratings, defaults) input: ratings – column vector of rarting value (higher value denote more risky) defaults – 1 denote default, 0 denote non-default Output: xy- points collection to plot CAP, AR- Accuracy Ratio -------------------------------------------------------------------------------------------------------------- Matlab file data example: capexample.mat 24

  25. Cumulative Accuracy Profile (CAP) (the second approach) Matlab file: cap_continuous.m Call: [AR]=cap_continuous(pd,is_def) input: pd – column vector of rarting value is_def– 1 denote default, 0 denote non-default Output: AR- Accuracy Ratio 25

  26. Conditional entropy, Kullback-Leibler distance Consider as rating system that, applied to an obligor, produces a random score S. If Ddenotes the event “obligor defaults” and D denotes the complementary event “obligor doesnot default”, we can apply the information entropy H to the P(D |S), the conditionalprobability of default given the rating score S. The result of this operation can be considereda conditional information entropy of the default event, and as such is a random variable whose expectation can be calculated. This expectation iscalled Conditional Entropy of the default event (with respect to the rating score S), and canformally be written as The Conditional Entropy of the default event is at most as large as theunconditional Information Entropy of the default event, i.e. For the empirical default rate : 26

  27. Receiver Operating Characteristic (ROC) An analytic tool that is closely related to the Cumulative Accuracy Profile is the ReceiverOperating Characteristic (ROC). The ROC can be obtained by plotting the fraction ofdefaulters against the fraction of non-defaulters. Thetwo graphs thus differ in the definition of the x-axis. Asummary statistic of a ROCanalysis is the area under the ROC curve (AUC). Reflecting the fact that the CAP is verysimilar to the ROC, there is an exact linear relationship between the accuracy ratio and thearea under the curve: (AR)=2× AUC−1 27

  28. Conditional entropy, Kullback-Leibler distance H can be interpreted as a measure of chaotic character.The difference of H(p)and HS should be as large as possible because in this case the gain of information byapplication of the rating scores would be a maximum: For the normalization case: we get Conditional Information Entropy Ratio(CIER): The value of CIER will be the closer to one the more information about the default event iscontained in the rating scores S. 28

  29. Conditional Information Entropy Ratio (CIER) Matlab file: CIER.m Call: [wsk, HPS]=cier(pd,is_default,spos) input: pd– column vector of rarting value (higher value denote more risky) is_default – 1 denote default, 0 denote non-default spos – various approaches Output: wsk-CIER normalised form of conditional entropy HPS – unnormalised form of conditional entropy 29

  30. Conditional Information Entropy Ratio (CIER) The Kullback Leibler divergence can be coputed as follows: This expression can be interpreted as a information divergence (information gain, relative entropy ) between a scoer density for the default and nondefault population. Becouse of its unsymetrically character its more comfortably to use another form as a System Stability Index (SSI) of course both of these vaues should be as much as it is possible

  31. Conditional entropy, Kullback-Leibler distance Matlab file: kullback_leibler.m Call: [D_DN, D_ND, SSI]=kullback_leibler(pd,is_default,spos) input: pd– column vector of rarting value (higher value denote more risky) is_default – 1 denote default, 0 denote non-default spos – various approaches Output: D_DN – distance from f(D) to f(ND) D_ND- distance from f(ND) to f(D) SSI - System Stability Index 31

  32. Bayesian error rate Denote with pD the rate of defaulters in the portfolio and define the hit rate HR and the falsealarm rate FAR as above. In case of a concave ROC curve the Bayesian error rate then canbe calculated via As a consequence, the error rate is then equivalent to the Pietra Index and theKolmogorov-Smirnov statistic. 32

  33. Bayesian error rate Matlab file: bayesian_error_rate.m Call: [ber]=bayesian_error_rate(rating,is_default,spos) input: rating– column vector of rarting value (higher value denote more risky) is_default – 1 denote default, 0 denote non-default spos – various approaches Output: ber - Bayesian error rate 33

  34. Kendall’s τ and Somers’ D Kendall’s τ and Somers’ D are so-called rank order statistics, and as such measure thedegree of comonotonic dependence of two random variables. The notion of comonotonicdependence generalises linear dependence that is expressed via (linear) correlation. Inparticular, any pair of random variables with correlation 1 (i.e. any linearly dependent pair ofrandom variables) is comonotonically dependent. But in addition, as soon as one of thevariables can be expressed as any kind of increasing transformation of the other, the twovariables are comonotonic. In the actuarial literature, comonotonic dependence is consideredthe strongest form of dependence of random variables. Kendall noted that the number of concordances minus the number of discordances is compared to the total number of pairs, n(n-1)/2, this statistic is the Kendall's Tau a: 34

  35. Kendall’s τ and Somers’ D Kendall’s τ and Somers’ D are so-called rank order statistics, and as such measure thedegree of comonotonic dependence of two random variables. The notion of comonotonicdependence generalises linear dependence that is expressed via (linear) correlation. Inparticular, any pair of random variables with correlation 1 (i.e. any linearly dependent pair ofrandom variables) is comonotonically dependent. But in addition, as soon as one of thevariables can be expressed as any kind of increasing transformation of the other, the twovariables are comonotonic. In the actuarial literature, comonotonic dependence is consideredthe strongest form of dependence of random variables. Kendall noted that the number of concordances minus the number of discordances is compared to the total number of pairs, n(n-1)/2, this statistic is the Kendall's Tau a: and Sommers’D: 35

  36. Kendall’s τ and Somers’ D Matlab file: tau_somersd.m Call: [Tau,SomersD,Tau_a,Gamma]=tau_somersd(pd, is_default) input: pd– vector of estimated default from model is_default - vector of the real pd Output: Tau –Tau calculated using Matlab function corr.m SomersD –value of SomersD Tau_a – Tau calculate is traditional way Gamma – the same value as SomersD in case there aren’t the same value of pd (T) 36

  37. Brier Score The Brier score is a method for the evaluation of the quality of the forecast of a probability. Ithas its origins in the field of weather forecasts. But it isstraightforward to apply this conceptto rating models. The Brier Score is denifed as RMSE (root mean squared error) • where i indexes the N observations, di is an indicator variable that takes the value 1 if borrower i defaulted (0 otherwise), and PDi is the estimated probability of default of borrower i. • the Brier score lies between 0 and 1, • better default probability forecasts are associated with lower score values. 37

  38. Brier Score Matlab file: Brier.m Call: [out]=Brier(PDs, defaults) input: PDs – vector of estimated default from model default - vector of the real pd Output: out –Brier Score 38

  39. Calibration methods 39

  40. Calibration methods Accrording to Basel Committeeon Banking Supervision we can mention the following methodologies assessing the quality of the PDestimates: • Binomial test, • Normal test, • Normal test with asset correlation, • Traffic lights approach, • Chi-square test (Hosmer-Lemeshow). 40

  41. Binomial test In many rating systems used by financial institutions, obligors are grouped into rating categories. The default probability of a rating category can then be estimated in different ways. Regardless of the way in which a default probability for a rating grade was estimated, wemay want to test whether it is in line with observed default rates. From the perspective of riskmanagement and supervisors, it is often crucial to detect whether default probability estimatesare too low. On the start we can assume that defaults are independent (so defaultcorrelation is zero). The number of defaults Dktin a given year t and grade k then follows a binomial distribution. The number of trials is Nkt, the number of obligors in grade k atthe start of the year t; the success probability is PDkt , the default probability estimated at the start of year t. 41

  42. Binomial test At a significance level of (e.g. α =1%), we can reject the hypothesis that the default probability is not underestimated if: where BINOM(x, N, q) denotes the binomial probability of observing x successes out of N trials with success probability q. If above condition is true, we need to assume an unlikely scenario to explain the actual default count Dkt (or a higher one). This would lead us to conclude that the PD has underestimated the true default probability. 42

  43. Normal test For large N, the binomial distribution converges to the normal, so we can also use a normalapproximation to equation from the previous slide. If defaults follow a binomial distribution with default probabilityPDkt, the default count Dkt has a standard deviation: The default count’s mean is: Instead of equation using bonomial distribution we can now examine: Where Φ denotes the cumulative standard normal distribution. If above condition is true, we need to assume an unlikely scenario to explain the actual default count Dkt (or a higher one). This would lead us to conclude that the PD has underestimated the true default probability. 43

  44. Normal test with asset correlation When we can assume that defaults are not independent (so defaultcorrelation is not zero) we have to introduce a asset correlation ρ. Now we examine the following equation using asset correlation ρ: Where - is inverse of the normal cumulative distribution function. If above equation is true, we conclude that the PD estimate was too low with the asset correlation ρ. 44

  45. Normal test with asset correlation Matlab file: Binomial.m Call: [ALLTest]=Binomial(PD_est,PD,N,alpha,p) input: PD_est - Historically default rates e.g. for year 1990-2004, for every rating categories (grade) -The default probability of a rating category PD - probability of default of every grade rating N - amount of all trials in every grade rating alpha - significance level e.g. alpha=1% p - asset correlation e.g. p=0.07 Output: ALLTest – matrix of results first column- binomial second column – normal third column – normal with correlation p ------------------------------------------------------------------------------------------------------------- Matlab file data example: binDataExample.mat 45

  46. Traffic lights approach Decisions on significance levels are somewhat arbitrary. In a traffic lights approach, we choose two rather or more than one significance level. If the p-value of a test is below red, we assign an observation to the red zone, meaning that an underestimation of the default probability is very likely. If the p-value is above red but below orange, we interpret the result as a very important warning that the PD might be an underestimate (orange zone). If the p-value is above orange but below yellow, we interpret the result as a warning that the PD might be an underestimate (yellow zone).Otherwise, we assign it to the green zone. For example we assume the following significance level for trafficlight approach: red <=0.01 , orange (0.01,0.05>, yellow (0.05,0.07> , green >0.7 46

  47. Normal test with asset correlation Matlab file: TrafficLight.m Call: [ALLTraficResult]=TrafficLight(PD_est,PD,N,p) input: PD_est - Historically default rates e.g. for year 1990-2004, for every rating categories (grade) -The default probability of a rating category PD - probability of default of every grade rating N - amount of all trials in every grade rating p - asset correlation e.g. p=0.07 In the file exists significance level for trafficlight: red <=0.01 , yellow (0.01,0.05>, orange (0.05,0.07> , green >0.7 Output: ALLTraficResult– matrix of results Number „4” denotes- red light, Number „3” denotes- orange light, Number „2” denotes- yellow light, Number „1” denotes - green light first column - binomial, second column – normal, third column – normal with correlation ---------------------------------------------------------------------------------------------------- Matlab file data example: binDataExample.mat 47

  48. Chi-square test (Hosmer-Lemeshow) Let 0, , p … pK denote the forecasted default probabilities of debtors in the rating categories0,1,…,k. Define the statistic with ni = number of debtors with rating i and θi = number of defaulted debtors with rating i.By the central limit theorem, when ni → ∞ simultaneously for all i, the distribution of Tk willconverge in distribution towards a χ2k +1-distribution if all the pi are the true defaultprobabilities. The p-value of a χ2 k +1-test could serve as a measure of the accuracy of the estimated defaultprobabilities: the closer the p-value is to zero, the worse the estimation is. 48

  49. Chi-square test (Hosmer-Lemeshow) Matlab file: hosmer_lemeshow.m Call: p=hosmer_lemeshow(pd,is_default) input: pd – vector of estimated default from model is_default- vector of the real pd Output: The p-value of a χ2k +1 49

More Related