SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI Abstract. We propose an approach to sovereign market implied ratings based on informations coming both from Credit Default Swap spreads and bond spreads in a unified way. Operationally speaking, we implement a Support Vector Machine type of selection in the plane CDS-bond. Our numerical results seem to confirm that introducing the bond dimension accounts for implied ratings more accurate and with greater predictive power with respect to the 1-dimensional CDS implied ratings. Contents 1. Introduction 2. Market implied ratings: various existing approaches 3. sovereign market implied ratings: a CDS penalty-function analysis 4. CDS+bond based analysis 5. statistical analysis: SVM vs. KV implied ratings 5.1. accuracy 5.2. precision & recall 5.3. gap-conditioned analysis 6. Conclusions Appendix A. data details Credit Default Swaps Bonds Agency ratings References 1 2 3 5 10 11 15 17 19 24 24 24 25 25 1. Introduction Ratings provided by credit rating agencies (CRAs) are meant as indicators/opinions about the issuer’s ability to pay back their debt in due time [1, 2, 3]. Their role in the modern financial system could be hardly overestimated. They are of key importance in at least three respects: for investors, who need information on the risks in which they are going to incur if investing in certain products; for issuers of financial products, who need ratings in order to attract investors and inform them on their creditworthiness; finally, credit ratings have become increasingly important in regulatory policies: indeed, financial institutions can use credit ratings from approved agencies in order to assess their capital requirements. Given that their importance is out of doubt, nevertheless there is plenty of criticism about CRAs and ratings themselves, in particular after their crucial role during the 2008 mortgage crisis and the 2011 euro-zone crisis. Leaving aside the structural critics, one of the main concerns about agency ratings is their (physiological) delay in capturing the actual health or creditworthiness of the issuer. This is the reason why, aside and complementing the information of agency ratings, the idea of capturing some kind of early warnings Date: November 26, 2014. JEL Classification. G24, C44, C55. Key words and phrases. credit ratings, market implied ratings, support vector machine, credit default swaps, default risk. 1 2 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI extracted by the market has become increasingly popular. These warnings are known as market implied ratings or “point in time” ratings. It is worth to stress that agency ratings and market implied ratings should be considered as complementary to each other: the stability in time of agency ratings and their independence of short term market movements are important features, that investors and emitters look for. At the same time, in particular for the investor side and for portfolio management, the market is a source of high frequency information that should not be neglected and can very well suit the role of giving sort of temperature to have a prompt idea of the ’health’ of debt emitters. It is in this spirit that market implied ratings have grown in importance and popularity, and that are now concocted and sold by CRAs themselves besides the ’standard’ ratings (see, e.g. [4, 5, 6, 7, 8, 9, 10]). In this paper we focus on market implied ratings for sovereign debt emitters and we propose the idea of using both the market of Credit Default Swaps and that of bonds. Indeed, in [11] qualitative arguments are presented, by which, for sovereign debt emitters and in particular in situations in which various countries share a common currency as in the Euro area, CDS alone may not capture the exact information about the real creditworthiness of a state. As far as we know, ratings derived by market data have up to now been calculated with one single source of information as input, namely ratings implied by CDS or bonds or equity. The original proposal of this work is precisely to present a way of calculating market implied ratings taking into account in one single procedure two different sources of market information. The procedure proposed is nonetheless sufficiently flexible to be generalized to different sectors and different input variables. The paper is structured as follows: section 2 briefly reviews the literature on market implied ratings; in section 3 we derive CDS-implied ratings for our panel of sovereign countries, using the technique of [4] and [5]; we describe our proposal in section 4, and then calculate the implied ratings for our dataset; finally, section 5 is devoted to the statistical analysis of the performances of the two sets of implied ratings. Concluding remarks follow. 2. Market implied ratings: various existing approaches Most of the literature on market implied ratings refers to corporate agents and typically deals with Credit Default Swap spreads, bonds and equity data. The first attempt in this direction dates back to 1974, with the so-called Merton model [12] whose idea is to see equity as a call option on assets with debt as strike, use equity and equity volatility market data to invert the relation, thus finding the asset value and asset volatility of the firm: from assets and debts a “distance to default” is calculated, which then can be mapped to a probability of default and to a rating score. This methodology, which is at the core of what are now called the “structural” models for default probabilities, has become very popular during the years, and it is actually implemented by both Moody’s and Fitch as an early warning tool over their long-term ratings (Moody’s KMV and Fitch EIR [9, 8]). Other, non-structural, attempts make a direct use of CDS or bond spreads, which are indeed a proxy of the credit risk of the issuer (see, for instance [13]). One of the first proposals (2003) can be found in [4] and is based on the assumption that “agency ratings are, on average, informative”. In this study, the authors create a penalty function minimizing which results in obtaining spread regions of implied ratings. This approach was recovered by Jainming Kou and Simone Varotto in 2008 [5] with some modifications and with statistical analysis to measure the goodness of the implied ratings in anticipating the agencies down/upgrades. The same penalty-function approach is implemented by Fitch [7], with the difference (among others) that the dataset is now composed by CDS spreads rather than bond spreads. Moody’s MIR - Market Implied Ratings [9] provides three types of market implied ratings: bond-implied, CDSimplied and equity-implied. While we have already mentioned the latter, bond-implied and CDS-implied MIR ratings are derived by filtering the data excluding extreme values, then calculating the median spread for each agency rating class and taking boundaries between two classes by geometric means of two contiguous median spreads. Standard & Poor’s MDS - Market Derived Signals [6] uses a linear regression approach, with regressors: agency rating class, the industrial sector (for the corporate case), the Document Type (the definition of what is a credit default) and others. SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 3 In a recent paper [14] a proposal that uses unsupervised learning methods1 was presented, namely a clustering algorithm based on an Hidden Markov Model, taking as inputs surviving probability curves bootstrapped by CDS spreads. As far as we are aware, all the published studies are mainly dedicated to the corporate sector with some minor hints to sovereigns. Here we deal explicitly with the sovereign sector, firstly calculating implied ratings with a standard penalty-function methodology over CDS (which we call, for simplicity, Kou-Varotto (KV) approach), and then proposing and implementing a new methodology for calculating market implied sovereign ratings (but portable to corporates as well). We then use some statistical metrics to compare the performances of the two implied ratings sets. 3. sovereign market implied ratings: a CDS penalty-function analysis In this section, we are going to briefly review the ideas of using a penalty-function to calculate implied ratings from CDS spreads and then we present the results of such a calculation on a pool of 36 sovereign countries for a 10-year time interval starting from 2004. The basic assumption for this kind of non-structural approaches, mentioned in the previous section, is that the agency ratings are considered on average informative, but with “errors”, due e.g. to delay of the agencies in officially updating their opinion on the country’s ability to pay back its debt. Namely, in this kind of approaches, the agency ratings are inputs to the calculation. Thus, the problem one faces is the following: we have, for each trading day, a certain number of CDS spreads, let us say one for each country we are considering2; attached to each country-spread there is a rating label (given by a chosen agency, or by an average rating among all agencies); if agency ratings exactly captured the credit risk inherent in the CDS spreads on that particular day, then the spreads of each rating class would be perfectly separated along the CDS dimension: typically, this is not the case. We refer to figure 1 for an example: you can see that AAA and AA rated countries (according to S&P’s) do not perfectly separate along the CDS dimension, even if, on average, they do. The penalty-function approach aims to find a clever way to draw a boundary in the spreads in order to give a new label (the implied ratings) on the basis of that boundary, as clarified in figure 1. The penalty function is derived from [4] and [5], and has the following expression (for AAA/AA boundary) N∑ NAA AAA 1 1 ∑ + + penalty(b) = (sAAA − b) + (b − sAA (3.1) j ) , NAAA j=1 j NAA j=1 where: • NAAA (NAA ) is the number of AAA (AA) rated countries, • sAAA (sAA j j ) is the spread of the j-th country among the AAA (AA) rated ones, • b is the spread of the boundary that we are looking for, • (·)+ denotes the positive part. The penalty function increases its value whenever a AA rated country has a spread lower than b or a AAA rated country trades at spreads higher than b. Minimizing 3.1 with respect to b gives the desired spread of the boundary between AA implied and AAA implied rated spreads. What we are actually doing is to choose the boundary that minimizes the “errors” of the agency ratings; the measure of the error is simply the distance from the boundary itself. The dataset we shall use comprehends a set of 36 countries3 covering all the rating classes; 5-year maturity CDS spreads (being the more liquid ones) as given by Markit; long-term ratings according to Standard & Poor’s; a time interval going from January, 1 st, 2004 to January, 14-th, 2014. In order for statistical analysis to be more reliable, we map the S&P’s rating classes4 into 5 larger classes, according to the following scheme: AAA → AAA 1 See, e.g., [15]. Actually, for each country there are CDS for various tenors, but let us focus on one single maturity. 3See the appendix A for more details on the dataset. 4See, e.g., the S&P’s website www.standardandpoors.com for reference to S&P’s ratings classes and their meaning. 2 4 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 0.006 ● ● ● implied AA ● CDS5Y RTG ● 0.004 ● ● ● ● ● ● ● ● ● ● 0.002 ● ● ● ● ● ● ● boundary ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● implied AAA ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● Oct 06 Oct 07 Oct 08 Oct 09 Oct 10 time ● AA ● ● ● ● ● ● ● ● ● ● AAA ● Figure 1. snapshot of the 5-year maturity CDS spreads of AAA, AA rated countries (according to S&P’s) in early October 2008, with a sketch of CDS implied rating construction for one day. AA+,AA,AA- → AA A+,A,A- → A BBB+,BBB,BBB- → BBB BB+,BB,BB-,B+,B,B-,CCC+,CCC,CCC-,CC,SD → BB the last of which is necessary since we have few data for the single categories of lowest ratings. The methodology is the following: for each day collect the spreads of the past 30 days (around 22 trading days); divide all the spreads according to the 5 rating classes; minimize the penalty function 3.1 for each couple of nearby ratings (AAA/AA, AA/A, …) thus finding 4 boundaries per day. The reason for using data of the entire previous month is to prevent the implied ratings to be too volatile and too connected to the daily movements of the market: Kou and Varotto in [5] collect data of two months before the desired date. We believe that an accurate analysis of the best temporal window from which collecting the data to calculate the daily boundary is worth doing, and we shall possibly deal with that in future work. Figure 2 shows the result of the calculation: you see the 4 barriers during the 10-year period. Of course, the barriers move in time, reflecting the fact that there are some dynamical features common to all the ratings, depending on common underlying scenarios. Indeed, the values of spreads are not influenced by credit risk alone: market-wide liquidity situation, average risk aversion of investors, are just some of the factors impacting on CDS premiums. From the plot, it is easy to see that before the 2008 crisis all the spreads were extremely low and rather stable, then you have a huge increase in all the spreads, with very high correlation among the different classes; there is a peak in early 2009, then a decrease in late 2009 and 2010, and then a new bump, nearly as high as the previous one, during the so-called euro-zone sovereign crisis; then again a decrease in 2013. Notice that, after the triggering of the 2008 crisis, the spread barriers never revert to something comparable to the pre-crisis situation. SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 5 AAA−AA AA−A A−BBB BBB−BB 0.04 0.03 0.00 0.01 0.02 CDS5Y 0.05 0.06 KV barriers 2004 2006 2008 2010 2012 2014 time Figure 2. time evolution of KV-implied rating boundaries The core idea is that a crossing of one of these barriers by the CDS spread related to one country detects precisely a change in the genuine credit risk part of the information encoded in spreads. Once you have the boundaries, calculating the implied ratings for the countries is straightforward. As an example, we plot the dynamics of the S&P’s and implied ratings for three countries, Italy, Spain and Germany in figures 3, 4, 5 respectively5. Some comments on the results shown in figures 3, 4, 5: firstly, it is clear that the KV implied ratings anticipate all the downgrades, both for Italy and Spain; secondly, the anticipation horizon is quite long: strong signals of downgrade starts between 1 and 2 years before the actual rating adjustment, sometimes even more; KV implied ratings seem to be a little bit on the pessimistic side, particularly regarding the recent years. These stylized features are common to the majority of the 36 countries analyzed. Moreover, focusing on the German situation, we see that the CDS market was giving signals of downgrade for many days in a row, both for the 2008 crisis and for the more recent euro-zone sovereign distress period, while the CRAs didn’t change their excellent opinion on the creditworthiness of Germany. 4. CDS+bond based analysis As we mentioned at the end of the previous section 3, KV analysis gives signals of downgrade for Germany during the two main crisis of the last few years. Actually, even if German CDS spreads were raising, resulting in downgrade market signals, we know that precisely in that periods, Germany, in particular German bund, was the “safe haven” of the entire euro-zone. Indeed, as figure 6 shows, in 2011-2013, while the CDS spread was very high, the bond was actually very low. The very same pattern was in place during the 2008 crisis as well. As was pointed out in [11], for sovereign countries the CDS alone may not be able to capture enough information to assess credit risk. In [11] it is also argued that taking into account both CDS and bonds should improve the ability of deducing from the market the rating for countries. And that is precisely what we are going to investigate in what follows. 5The dataset for Italy starts in 2006 and not in 2004 because of a lack of data for the previous period. 6 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI kv implied ratings for ITALY SnP kv AAA AA A BBB BB 2004 2006 2008 2010 2012 2014 time Figure 3. Kou-Varotto implied rating together with historical S&P’s rating for Italy kv implied ratings for SPAIN SnP kv AAA AA A BBB BB 2004 2006 2008 2010 2012 time Figure 4. KV implied rating together with historical S&P’s rating for Spain 2014 SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 7 kv implied ratings for Germany AAA SnP kv AA 2004 2006 2008 2010 2012 2014 time Figure 5. KV implied rating together with historical S&P’s rating for Germany BND5Y −0.25 −0.50 −0.75 ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ●● ●●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●●● ● ● ● ● ● ● ● ●● ● ●● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ●● ● ● ● ● ●●●● ● ● ●● ● ● ●● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ●●●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ●●● ● ● ●● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ●● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ●● ● ● ●● ● ●● ● ● ● ●● ● ● ● ● ●● ● ●● ●● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ●● ●● ● ●●● ● ● ●● ● ● ●●● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ●● ●● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ●● ● ● ● ● ●● ● ●●● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ●● ● ● ● ● ●● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● 2004 2006 2008 2010 2012 RTG ● AAA 2014 time 0.012 ● ● ● ● ●● ● ● ● ● ● ● ● ●● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ●● ●●● ● ● ● ● ● ● ● ●● ●●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●● ● ● ●●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●●● ● ●● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ●● ● ● ● ● ● ●● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●● ● ●● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ●●● ●● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ●●● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ●● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● SPREAD5Y 0.009 0.006 0.003 0.000 ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 2004 2006 2008 2010 2012 2014 time Figure 6. plot of bond spread (upper) and CDS spread (lower) for Germany RTG ● AAA 8 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI Adding 5-year maturity bonds means adding one dimension to the input space: a single sovereign state is represented no more by a single number (the CDS spread), but by a couple bond spread + CDS spread. Thus, as time goes on, each country describes a trajectory in the bond-CDS plane. Now the problem amounts to divide, on a daily basis, such a plane in regions, each corresponding to a different implied rating. As in the one dimensional case we needed a prescription (the Kou-Varotto minimization approach) by which separate the points into implied-rating classes, we need a prescription here as well. One natural choice to classify points in multi dimensional spaces is to use the so called Support Vector Machine technique [16, 17, 18, 19, 15, 20]. While we refer to the bibliography for a thorough description of this well known approach of computer science, here we merely try and summarize the key points, for sake of clarity and reader’s comprehension. Suppose you have points on a plane, with coordinates xi , i = 1 . . . , N where N is the total number of points, and suppose that each point comes together with a label yi ∈ {−1, 1}. We want to separate these points with a line, in such a way that we divide the +1’s from the −1’s. The configuration can be as easy as in the left panel of figure 7, which is called linearly separable: the +1’s and −1’s can be exactly separated with a line. Actually, with an infinite number of lines. Among them, the Support Vector Machine approach chooses the one that maximizes the margin as depicted in the same figure. In this situation, the SVM is just a minimization procedure to find a unique line separating points on a plane. Referring again to the left panel of figure 7 for the notation: • (β0 , β) are the parameters of the separating hyperplane, with ∥β∥ = 1 (namely, β is the unit vector normal to the hyperplane), • x1 and x2 denote the closest points to the plane, the so called support vectors • M is the half-margin considering that f (xi ) = xTi β + β0 is precisely the (signed) distance of the point xi from the hyperplane x : f (x) = 0, then, without loss of generality, we can take f (namely β, β0 ) such that yi f (xi ) ≥ 0 is the distance of xi from the plane. Then, the problem of maximizing the margin looks like: max M β,β0 ,∥β∥=1 subject to yi (β T xi + β0 ) ≥ M , ∀i (4.1) which is pretty self explanatory. Actually, there a scale invariance in the problem: to see this, notice that one can replace the constraints ∥β∥ = 1, yi (β T xi + β0 ) ≥ M with yi (β T xi + β0 ) ≥ M ∥β∥ (where β0 has been redefined appropriately); now notice that this constraint is invariant to rescaling of the vector (β0 , β). Then, we can arbitrarily fix the norm ∥β∥. A smart choice is 1 . M Then, the problem (4.1) is equivalent to the following ∥β∥ = min β,β0 ∥β∥2 2 subject to: yi (β T xi + β0 ) ≥ 1 , ∀i (4.2) In most cases, and for sure in our situation, data are not linearly separable. SVM prescribes a procedure to deal with non-linearly separable data, as in the example of the right panel of figure 7: we still try to maximize the margin, but we allow some points to be inside the margin or even on the “wrong” side of the boundary line. These “errors” are weighted with the parameter ξi that is the distance by which they enter inside the margin. The minimization problem to uniquely fix the linear boundary reads: ) ( n 2 ∑ ∥β∥ +C ξi ; C → cost parameter min β,β0 2 (4.3) optimization i=1 subject to: y (β T x + β ) ≥ 1 − ξ , ξ ≥ 0 , i = 1 . . . n i i 0 i i SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA x2 yi = −1 9 yi = −1 M M ξ>0 x1 ξ>1 margin ξ>1 yi = 1 xT β + β 0 = 0 yi = 1 xT β + β 0 = 0 Figure 7. Support Vector Machine separating plane: linearly separable (left) and nonlinearly separable (right) cases. where C is an exogenous parameter that measures how much we are willing to make “errors” at all. The minimization problem is then a trade off between the maximization of the margin and the minimization of the cost we pay when making “errors”. Actually, the real utilization of SVM is to use that boundary line to classify new data of which we know x but we don’t know y. That’s precisely the philosophy behind all the Machine Learning algorithms: to train the machine with some known data and to use it to classify unknown ones. We have instead to face a slightly different problem: we want to separate points on a plane of which we know the classification (agency ratings) but we also know that this classification contains “errors”, and we need a good way to perform that separation. Of course, SVM is not confined to work on a plane: SVM can deal with an arbitrary (finite, but in some cases also infinite) number of dimensions. Moreover, SVM can deal with non-linear boundaries as well, via so called kernel methods. In the present work we confine ourselves to linear boundaries and to 2-dimensional analysis, but these are two features that we are going to investigate in future research: adding to CDS spreads and bonds some extra information, and analyzing the possibility of implementing some curved decision boundary. The dataset we work with is the following: same 36 countries, ratings and CDS spreads as for the previous section 3; 5-year maturity bond interest rate spreads over zero-interest curves, as calculated by UniCredit risk management staff. The time interval is again January, 1st, 2004 → January, 14th, 2014. The methodology for calculation is as follows: • for each trading day, we collect CDS spreads and bond spreads for the 36 countries for 30 days backward in time, resulting in approximately 750 input data per day, • with these data we calculate the boundary lines for that day, with the SVM procedure, • having the plane for that day divided into implied rating regions, it is easy to infer the implied rating for each CDS-bond couple for that day. All the calculations were done with R open software [21] and in particular with package e1071 [22] and caret [23] for SVM analysis and cross-validation tuning respectively. Actually, there are a couple of technicalities we must dwell on a while: one regarding the cost parameter C, and the other referring to the multi-class nature of our problem. The cost parameter C must be fixed a priori of the actual boundary calculation. Remember that fixing C means fixing the price we are willing to pay for “misplaced” points on the plane. In the literature on SVM, one can find several approaches regarding the parameter tuning of the support vector approach, among which we choose the cross-validation. Namely, the routine does the following: she divides in 10 parts all the training data for that day (∼ 750 points), then uses 9/10 of the data as training set on which she calculates the boundaries, and the other 1/10 as a test set. This procedure is repeated for each 1/10 of the data set, and the overall procedure is repeated for 10 times independently of one another. Then, the accuracy (i.e. how many 10 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI points of the test set are classified correctly by the trained boundaries), averaged over all these trials, is calculated. This is done for every value of the cost parameter along the sequence (2−3 , 2−2 . . . 0, 2, 4). Finally, the cost parameter that maximizes the accuracy is chosen. This procedure is run through every ten training days, keeping the latest optimal cost for the other 9 days. The overall scheme results in a further (this time statistical) optimization of the separation boundary. Now with the multi-class issue: the SVM procedure we described above, refers to binary classification problems, namely, problems where data belongs to one among 2 classes. Here, instead, we are dealing with a 5-class classification. The routine we used adopts the so-called one-versus-one approach, which runs like this: she calculates 10 linear boundaries, 1 for each couple of rating classes; then, each point on the plane in assigned an implied rating by a voting scheme: the point is classified according to all the 10 couples, and the rating class that obtains more “votes” is the one assigned to that point. This procedure determines, daily, the implied rating regions of the CDS-bond plane6. Another minor technical remark is about the asymmetry of data: namely, for each day we have (sometimes big) differences in the numbers of points for each class (namely, we have a lot of AAA countries but few BB, see appendix A). Given that accuracy is the metric for cross-validation, unbalanced data may result in unbalanced model selection. The solution to this problem is to assign a cost parameter C different for each class, weighted on the basis of the size of the training dataset of each class. See, e.g., [16] for some details. Unfortunately, showing the 2-dimensional boundaries evolving in time is neither illustrative nor easy as in the one-dimensional case. We are going to show just a couple of daily snapshots, in order to give an idea of how the SVM works. Some general comments regarding the two figures 8, 9: • interestingly, you see that while the high ratings (AAA,AA) are separated from the others along the SE-NW direction, the low ones tend to be separated along the orthogonal direction: this means that while for high rated countries CDS and bond spreads tend to be negatively correlated, the opposite holds for low rated countries, • a comparison of the two plots clarifies the very idea of implied ratings and of SVM as well: in fig. 9 you see that Romania and Turkey, despite being rated BB by Standard & Poor’s, have actually quite low bond spreads, and not very high CDS spreads as well (comparable to BBB values); SVM decides however to rate them BB, minimizing the errors; in fig. 8 on the other hand, while Romania is still reclassified BB, Turkey is too mixed with AA guys, and classifying it BB would have cost way too much: thus the calculation puts Turkey in the AA region. Indeed, the market on those days whose trading Turkey debt issues and CDS at the level of most AA countries. As for the KV case, let us show the time evolution of implied ratings for Italy, Spain and Germany. Figures 10, 11, 12 shows the results: historical S&P’s ratings, SVM implied ratings and, for comparison, KV implied ratings. Regarding the issues we addressed for KV implied ratings, we see here that SVM classification has at least two differences/improvements: • it is more timely, meaning that the downgrading signals usually start nearer to the actual rating change, • it is more “optimistic”, meaning that tends to give higher implied ratings, in particular in the recent period. Regarding the German situation, we clearly see that adding the bund has the (expected) result of getting rid of the downgrading signals both for the 2008 and for the euro-zone crisis. 5. statistical analysis: SVM vs. KV implied ratings In the present section, we are going to present the results of some statistical tests done to measure the goodness of our implied ratings, in comparison with the KV CDS-based approach. 6Another popular scheme in dealing with k-class classification is the one-versus-all, in which only k SVM boundaries are calculated, and the classes are assigned according to the distance of a point from the k boundaries, on the basis of the fact that the more you are far from a boundary the more likely your classification is right SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 11 SVM classification plot 0.06 0.05 BB romania italy BBB turkey spain 0.03 germany A SPREAD5Y 0.04 AA 0.02 AAA 0.01 malaysia −1 0 1 norway 2 3 4 5 6 BND5Y Figure 8. (part of the) bond-CDS plane on 2012-09-15: different colored regions denote different implied ratings as calculated by the SVM procedure; the colored circles are the data used to train the machine for this day. Colors of circles correspond to agency (S&P’s) ratings according to: black-AAA, red-AA, green-A, blue-BBB, cyan-BB. Some countries are indicated explicitly. The usual tests to check the goodness of a classification are the ROC (Receiver Operator Characteristic) curve and the Accuracy Ratio, with varying time horizon to test the predictive power as well. Both these tools measure the ability of rating assignments essentially by considering the percentage of defaulters that were assigned low ratings. Unfortunately, these are completely useless in our settings (and, in general, for sovereigns), since the number of defaults of sovereign states, even if we considered longer period of time and larger pool of states, is ridiculous in the perspective of having a statistical test. Thus, we are obliged to abandon the idea of having a test of the absolute goodness of our implied ratings, and we check their performance with respect to S&P’s. 5.1. accuracy. In order to check the accuracy of SVM implied ratings with respect to S&P’s, and to compare it to that of KV implied ratings, we do the following: we take the implied ratings on each day in the interval considered (2004, January 1st to 2014, January 14th) and we calculate the accuracy of the SVM (and KV) implied ratings versus the agency rating of 1 year later. Then we plot the results of accuracy day by day in figure 13. Recall that accuracy in a classification problem is simply given by accuracy = true positives + true negatives ; entire population namely, the fraction of right guesses7 made by the implied ratings. In figures 14-16 other accuracy graphs are displayed, namely the daily accuracy of prediction for 3month, 6-month and 1 month horizon. It is pretty clear that in all the situations considered the SVM implied ratings behave better than KV. Just to be clear, we can calculate the cumulate accuracy for the 7“Right” with respect to the (time-shifted) agency ratings. 12 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI SVM classification plot romania turkey italy A 0.03 BBB spain 0.02 0.01 AA slovenia germany 0 1 2 AAA SPREAD5Y 0.04 BB 0.05 3 4 5 BND5Y Figure 9. (part of the) bond-CDS plane on 2013-09-12: different colored regions denote different implied ratings as calculated by the SVM procedure; the colored circles are the data used to train the machine for this day. Colors of circles correspond to agency (S&P’s) ratings according to: black-AAA, red-AA, green-A, blue-BBB, cyan-BB. Some countries are indicated explicitly. SVM AAA’ AA’ A’ BBB’ BB’ KV AAA’ AA’ A’ BBB’ BB’ AAA 27562 3634 223 4 0 AAA 24604 6051 768 0 0 AA AA 3828 9423 1531 496 53 3598 8270 2883 398 182 A A 365 2640 8805 1506 175 136 3260 8009 1297 789 8 184 1871 8411 961 0 125 2219 6265 2826 BBB BBB BB 61 331 47 1252 5283 BB 0 13 808 1346 4807 Table I. cumulate contingency table for SVM (left) and KV (right) implied ratings for 1-month horizon (primed ratings stand for implied). SVM AAA’ AA’ A’ BBB’ BB’ KV AAA’ AAA 25141 2490 61 0 0 AAA 22634 AA 3434 8550 1280 218 9 AA 2941 365 2641 7729 1151 2 136 A A BBB 8 301 1845 7182 1249 BBB 0 61 351 266 1787 3654 0 BB BB Table II. cumulate contingency table for SVM (left) and for 1-year horizon (primed ratings stand for implied). AA’ A’ BBB’ BB’ 4586 472 0 0 7625 2698 171 56 3359 7473 841 79 119 1947 5952 2567 14 905 1267 3933 KV (right) implied ratings entire period for the different horizons: svm 1 month: ∼ 0.756 svm 3-month: ∼ 0.768 svm 6-month: ∼ 0.763 svm 1yr: ∼ 0.757 SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 13 svm vs kv implied ratings for ITALY ● AAA svm SnP kv AA A BBB BB 2008 2010 2012 2014 time Figure 10. SVM implied ratings for Italy, in comparison with KV implied ratings and together with the historical S&P’s. svm vs kv implied ratings for SPAIN ● AAA svm SnP kv AA A BBB BB 2004 2006 2008 2010 2012 2014 time Figure 11. SVM implied ratings for Spain, in comparison with KV implied ratings and together with the historical S&P’s. 14 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI svm vs kv implied ratings for Germany AAA ● AA 2004 2006 2008 2010 2012 svm SnP kv 2014 time Figure 12. SVM implied ratings for Germany, in comparison with KV implied ratings and together with the historical S&P’s. daily accuracy of prediction for 1yr horizon 0.8 accuracy irs SVM KV 0.6 0.4 2004 2006 2008 2010 2012 2014 time Figure 13. plot of daily accuracy of prediction for 1 year horizon, both for SVM and KV implied ratings (irs). SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 15 daily accuracy of prediction for 3−month horizon 1.0 0.8 accuracy irs SVM KV 0.6 0.4 2004 2006 2008 2010 2012 2014 time Figure 14. plot of daily accuracy of prediction for a 3-month horizon for SVM and KV implied ratings. kv 1 month: ∼ 0.665 kv 3-month: ∼ 0.670 kv 6-month: ∼ 0.677 kv 1yr: ∼ 0.686 , The best performance is given by SVM implied ratings over a 1 month predicting horizon. You can also see the contingency tables for 1-year and 1-month horizon in tables I and II. This analysis reflects two facts: • SVM implied ratings are, on average, fairly better in matching future agency ratings with respect to KV ones, • SVM implied ratings (for the very same way in which they are concocted) tend to be pretty stable around the agency ratings, resulting in an increasing accuracy the shorter the horizon. As we shall see in section 5.3, this second point does not actually mean that the shorter term is the better in predicting upgrades and downgrades of agency ratings. Namely, accuracy captures the information of prediction in every situation, also when there is actually no upgrade or downgrade in the future. Some kind of analysis which focuses mainly on predicting the agency ratings movements is necessary for a thorough description of the goodness of implied ratings. 5.2. precision & recall. Another standard test we are going to show, is the precision & recall for each class. For sake of clarity, we just give the cumulative precision and average recall over the 10-year interval for each class, and we limit ourselves to the 1-month and 1-year horizons of prediction. The results are summarized in tables III and IV Let us remind what precision and recall are: precision of class k = recall of class k = # true positives = probability of being in class k given that the test was positive # positives # true positives = probability of positive test given that the individual is in class k # members of class k 16 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI daily accuracy of prediction for 6−month horizon 0.8 accuracy irs SVM KV 0.6 0.4 2004 2006 2008 2010 2012 2014 time Figure 15. plot of daily accuracy of prediction for a 6-month horizon for SVM and KV implied ratings. daily accuracy for 1 month horizon 1.0 0.8 accuracy irs SVM KV 0.6 0.4 2004 2006 2008 2010 2012 2014 time Figure 16. plot of daily accuracy of prediction for a 1-month horizon for SVM and KV implied ratings SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 17 horizon precision (cumulate over 2004→2014) recall (cumulate over 2004→2014) 1 month AAA AA A BBB BB AAA AA A BBB BB svm 0.866 0.581 0.706 0.721 0.816 0.877 0.615 0.654 0.736 0.758 kv 0.868 0.467 0.545 0.673 0.559 0.783 0.539 0.594 0.548 0.689 Table III. precision & recall for the 5 classes, cumulate over the entire time interval and with 1-month predicting horizon. horizon precision (cumulate over 2004→2014) recall (cumulate over 2004→2014) 1 year AAA AA A BBB BB AAA AA A BBB BB svm 0.867 0.597 0.6910 0.698 0.744 0.908 0.634 0.650 0.679 0.597 kv 0.880 0.486 0.554 0.723 0.593 0.817 0.565 0.629 0.562 0.643 Table IV. precision & recall for the 5 classes, cumulate over the entire time interval and with 1-year predicting horizon. horizon F 1 (cumulate over 2004→2014) 1 month AAA AA A BBB BB svm 0.872 0.597 0.678 0.728 0.785 kv 0.823 0.500 0.568 0.604 0.617 Table V. F 1-score (as by formula (5.1)) for the 5 classes, cumulate over the entire time interval and with 1-month predicting horizon. horizon F 1 (cumulate over 1 year AAA AA A svm 0.887 0.615 0.670 kv 0.848 0.522 0.589 Table VI. F 1-score (as by formula (5.1)) for the time interval and with 1-year predicting horizon. 2004→2014) BBB BB 0.687 0.662 0.633 0.617 5 classes, cumulate over the entire The higher precision the fewer type I errors, the higher recall the fewer type II errors. Namely, with high class k precision you can trust the test in signals of being in class k, with high class k recall you can be quite sure that if a guy is in class k the test is going to get it right. We can add as a final indicator the popular F 1-score, which is simply the geometric mean of precision and recall, namely precision × recall F1 = 2 × , (5.1) precision + recall and is a standard way of summarizing together precision and recall. The cumulate results for 1-month horizon are in table V, while in figure 17 we plot the daily values of F 1 for all the 5 classes. In table VI and figure 18 you can see the same indicator but for a 1-year predicting horizon. As it is apparent from table III, both precision and recall (and thus F 1 as well) are quite high for all the rating classes, and the SVM implied classification performs systematically better, with the (very slight) exception of the precision of AAA. 5.3. gap-conditioned analysis. In this section we want to address the capacity of sovereign implied ratings to predict agency rating adjustments. Namely, as was mentioned at the end of section 5.1, the metrics seen so far do not discriminate predictions of stability from predictions of down/upgrades. While predicting stable ratings is of course important, a focus on rating adjustments is needed to understand the goodness of implied ratings. For the statistical test analysis of the present section, we take inspiration by Moody’s report [9], where the authors do a similar thing for corporates (with their bond-implied ratings). 18 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI F1 − AAA F1 − AA 1.0 1.00 irs 0.8 0.6 irs 0.75 SVM KV SVM 0.50 KV 0.25 0.4 2004 2006 2008 2010 2012 2014 2004 2006 F1 − A 2008 2010 2012 2014 F1 − BBB 1.00 1.00 irs 0.75 irs 0.75 SVM 0.50 KV 0.25 SVM 0.50 KV 0.25 2004 2006 2008 2010 2012 2014 2004 2006 2008 2010 2012 2014 F1 − BB 1.0 irs 0.8 0.6 SVM 0.4 KV 0.2 2004 2006 2008 2010 2012 2014 Figure 17. F 1-score for 1-month horizon for the 5 rating classes. F1 − AAA F1 − AA 1.0 1.00 irs 0.8 0.6 irs 0.75 SVM KV SVM 0.50 KV 0.25 0.4 2004 2006 2008 2010 2012 2014 2004 2006 F1 − A 2008 2010 2012 2014 F1 − BBB 1.00 1.00 irs 0.75 irs 0.75 SVM 0.50 KV 0.25 SVM 0.50 KV 0.25 2004 2006 2008 2010 2012 2014 2004 2006 2008 2010 2012 F1 − BB 1.0 irs 0.8 SVM 0.6 KV 0.4 2004 2006 2008 2010 2012 2014 Figure 18. F 1-score for 1-year horizon for the 5 rating classes. 2014 SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 19 In figure 19 you see a graph displaying predictive power of SVM implied ratings. The methodology is the following: we take the 36 countries at the beginning of each month from 2004, January 1st to 6 months before our end date, which is 2014, January 14; we calculate, for each country in each of these days, the rating gap, namely: rating gap = iRTG − RTG were we have mapped rating to integers according to AAA → 1, AA → 2, . . . , BB → 5 ; thus, a positive rating gap means that, on that particular day, the SVM procedure assigns a rating worse than the agency does. Then we go and look to the situation 6 months later and we count how many countries have been downgraded/upgraded/untouched by S&P’s, conditioned to their rating gap (of 6 months before). We aggregate the data for all the countries and the entire time interval and plot them in a bar graph. Each bar corresponds to a rating gap and counts all the countries that, being in that rating gap class, are subject to an (agency) upgrade (green), a downgrade (blue), or are stable (red). Notice that the bars show only fraction values, namely the count is normalized with respect to the number of events for each class. If the implied ratings are to lead the agency ratings (i.e. to predict them) than we would expect the following: a very big red bar in the 0-gap position; as we move rightward, an increasing blue region (a worse implied rating ’implies’ an increasing fraction of future downgrades); as we move leftward, an increasing green region (a better implied ratings ’implies’ an increasing fraction of future upgrades). As we can clearly see, while for gap ≥ 0 we can indeed conclude that implied ratings lead the agency ratings, we cannot say the same for negative rating gaps. The same calculation is done for a 1 year horizon: the SVM performance is here even better, which means that, when focusing on rating adjustments, the short term horizon is not the optimal choice for prediction. Table VII and plot 20 show the same results, but with absolute numbers. You can see that the vast majority of events is in the region: gap = 0 – change = stable. Actually, notice that the bars corresponding to gap -4, -3, 3 have far too few events to be statistically meaningful: it is better to consider only the bars with gaps form -2 to 2. Moreover, it is apparent that the upgrade events are actually just a tiny fraction of the overall number, both in the 6-month and 1 year scenario, thus resulting in a statistically poor ensemble for making any prevision regarding the upgrades. down stable up down stable up -4 0 3 0 -4 0 3 0 -3 -3 0 15 0 0 15 0 -2 -2 0 26 0 0 26 0 -1 0 382 19 -1 6 331 33 0 63 2500 22 0 115 2276 38 1 44 291 0 1 84 227 0 2 15 21 0 2 23 13 0 3 0 0 3 0 0 3 3 Table VII. 6-month and 1 year SVM gap-conditioned contingency tables In figure 21 it is displayed the same bar graph, obtained with the very same methodology described above, with rating gap calculated with Kou-Varotto implied ratings, while VIII and 22 are the corresponding tables and absolute counting bar graphs, respectively. A comparison shows clearly that, as far as downgrades are concerned, SVM implied ratings perform slightly better than KV. 6. Conclusions Credit ratings have a crucial role in many aspects of modern financial markets, even more so after the 2008 mortgage crisis (when CRAs have been strongly criticized for their ‘opinions’ on mortgagebacked securities) and in particular euro-zone crisis. Given that rating agencies are quite slow in updating 20 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI 6−month rating change conditioned to SVM gap 1.00 count (%) 0.75 down stable up 0.50 0.25 0.00 −3 −2 −1 0 rating gap 1 2 3 1yr rating change conditioned to SVM gap 1.00 count (%) 0.75 down stable up 0.50 0.25 0.00 −4 −3 −2 −1 0 1 2 3 rating gap Figure 19. 6-month and 1 year predictive power: on x axis the gap between SVM implied rating and the corresponding historical one (positive gap means implied rating is worse than agency rating); the bars show the fraction of stable/upward/downward rating change after 6 months-1 year. SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 21 counting of SVM gap−conditioned 6−month horizon rating change 2000 count 1500 down stable up 1000 500 0 −3 −2 −1 0 rating gap 1 2 3 counting of SVM gap−conditioned 1yr horizon rating change 2500 2000 1500 count down stable up 1000 500 0 −4 −3 −2 −1 0 1 2 3 rating gap Figure 20. 6-month and 1 year absolute counting for rating change conditioned to SVM rating gap on the entire time interval. 22 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI 6−month rating change conditioned to KV gap 1.00 count (%) 0.75 down stable up 0.50 0.25 0.00 −3 −2 −1 0 rating gap 1 2 3 1yr rating change conditioned to KV gap 1.00 count (%) 0.75 down stable up 0.50 0.25 0.00 −3 −2 −1 0 rating gap 1 2 3 Figure 21. 6-month and 1 year predictive power: on x axis the gap between KV implied rating and the corresponding historical one (positive gap means implied rating is worse than agency rating); the bars show the fraction of stable/upward/downward rating change after 6 months-1 year. SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 23 counting of KV gap−conditioned 6−month horizon rating change 2000 count 1500 down stable up 1000 500 0 −3 −2 −1 0 rating gap 1 2 3 counting of KV gap−conditioned 1yr horizon rating change 2000 count 1500 down stable up 1000 500 0 −3 −2 −1 0 rating gap 1 2 3 Figure 22. 6-month and 1 year absolute counting for rating change conditioned to KV rating gap on the entire time interval. 24 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI down stable up down stable up -3 0 1 0 -3 0 1 0 -2 0 45 0 -2 0 42 0 -1 1 420 30 -1 2 361 53 23 2200 11 69 2021 18 0 0 1 64 513 0 1 104 432 0 2 32 57 0 2 51 32 0 3 5 2 0 3 5 2 0 Table VIII. 6-month and 1 year KV gap-conditioned contingency tables their credit assessments – particularly for sovereign countries – market implied ratings have become increasingly necessary in providing timely early warnings for institutions/investors interested in higher frequency information about creditworthiness of counterparts. In this paper, we focused our study on sovereign market implied ratings, whilst the literature is mainly devoted to the corporate sector. We proposed a methodology for deriving implied ratings by market data, exploiting both CDS and bond market informations. As far as we know, this methodology, even if inspired by the works in [4] and [5], is entirely new in the respect of using two different market sources together and also in the implementation of Machine Learning techniques for this purpose. We performed the calculation of the implied ratings for a panel of 36 countries during a 10-year period starting in 2004; we calculated both “standard” implied ratings (namely, using only CDS spreads as input information and implementing [4]-[5] technique) and the new ones. The results show that both calculations provide quite good early warnings, both for anticipating upgrades/downgrades and for predicting stable ratings. However, the implied ratings derived with the new methodology perform systematically better in all respects. This is intended as preliminary work on this subject, further analysis must be put in place in order to improve and generalize the idea here proposed and to test the results here obtained with larger samples and possibly other statistical metrics. Appendix A. data details In the following, some detailed information about the dataset used for statistical analysis is presented. We acknowledge UniCredit for providing all the data used in our calculations, in particular we thank Dr. Andrea Basile and the entire Group of Risk Methodologies and Architecture. The 36 countries considered are Australia, Austria, Belgium, Bulgaria, Canada, Croatia, Czech Republic, Cyprus, Denmark, Finland, France, Germany, Greece, Hong Kong, Hungary, Ireland, Italy, Japan, Malaysia, Netherlands, New Zealand, Norway, Poland, Portugal, Romania, Russia, Singapore, Slovenia, Slovak Republic, Spain, Sweden, Swiss, Turkey, Venezuela, UK, US. Credit Default Swaps. CDS data are from Markit Corporation. They are updated to 14-01-2014, with quoted maturities in {0.5, 1, 2, 3, 4, 5, 7, 10, 15, 20, 30} years, among which we used the most liquid one, namely 5-year. The Debt Tier considered is Foreign Currency Debt (Domestic Debt CDS quotations are much more rare), and currencies for considered debt are dollars for all countries except for US who has euros denominated contracts. Bonds. Bond data are constructed starting from bond prices by Bloomberg Corporation. The spread derivation goes through the construction of two curves: the zero curve, or risk-free curve, one for each currency area (namely, we have a zero curve for each currency by which the bond contracts are denominated), and the zero-coupon curve based on treasuries (of course, one curve for each country considered). The actual data used are the spreads between the treasury (zero-coupon) yield and the risk-free yield (of the corresponding currency). The zero curves are built essentially starting from quotations of basic derivatives on currency exchange rates. On theoretical grounds put forward in [24], the zero-coupon curves for treasuries are bootstrapped via the the standard risk-free procedures, even if treasuries are in fact defaultable. As for CDS, the most liquid maturity (5-year) only has been used in calculation. SOVEREIGN RATINGS IMPLIED BY COUPLED CDS-BOND MARKET DATA 25 Agency ratings. Agency ratings are by Standard & Poor’s. We use the Long Term Foreign Currency rating assessment, remapped into 5 classes as described in section 3. S&P’s is the agency typically more timely in updating/reviewing its rating assessments. On average along the period analyzed, roughly 39% of the countries considered are rated AAA by S&P’s8, 20% AA, 17% A, 14% BBB and 10% BB. As a last remark, notice that we have not complete data for all the countries on the entire period considered (01-01-2004 → 14-01-2014). This refers almost completely to the bond dataset, where countries like Swiss, Italy, Singapore, New Zealand, Bulgaria, Portugal, Malaysia, Venezuela and Romania, have yield curves starting after 2004 (typically around 2007). This, of course, does not influence at all the analysis of this paper, it only amounts to have a shorter time period of analysis for some of the countries. References [1] Herwig Langohr and Patricia Langohr. The Rating Agencies and Their Credit Ratings: What They Are, How They Work, and Why They are Relevant . Wiley Finance, January 2009. [2] Norbert Gaillard. A Century of sovereign ratings . Springer, October 2012. [3] Jacob De Haan and Fabian Amtenbrink. Credit Rating Agencies. DNB Working Paper, January 2011. SSRN: http: //ssrn.com/abstract=1760951 or http://dx.doi.org/10.2139/ssrn.1760951. [4] Ludovic L. Breger, Lisa R. Goldberg, and Oren Cheyette. Market-implied Ratings. Risk Magazine, July 2003. SSRN: http://ssrn.com/abstract=402800 or http://dx.doi.org/10.2139/ssrn.402800. [5] Jainming Kou and Simone Varotto. Timeliness of spread implied ratings. European Financial Management, 14:503–527, June 2008. http://dx.doi.org/10.1111/j.1468-036X.2007.00362.x. [6] Sten Bergman, Martin Hampel, Jayson Rome, Iris Shi, Lauren Taralli, and Xi Yang. How Standard and Poor’s arrives at Market Derived Signals. Standard and Poor’s, May 2009. http://dx.doi.org/10.1111/j.1468-036X.2007.00362.x. [7] Greg Gupton, Ahmet E. Kocagil, and Alexander Reyngold. Fitch CDS Implied Ratings (CDS-IR) model. Fitch, Inc, June 13 2007. https://www.fitchratings.com/web_content/product/methodology/cdsir_methodology.pdf. [8] Greg Gupton, Ahmet E. Kocagil, and Bo Liu. Fitch Equity Implied Rating and Probability of Default model. Fitch, Inc, June 13 2007. https://www.fitchratings.com/web_content/product/methodology/cdsir_methodology.pdf. [9] John et al. Gibbon. Moody’s Market Implied Ratings: Description, methodology, and analytical applications. Moody’s Investors Service, Report 106083, December 2007. http://efinance.org.cn/cn/FEshuo/Moody%27s%20Market% 20Implied%20Ratings.pdf. [10] Richard et al. Cantor. An explanation of Market Implied Ratings. Moody’s Investors Service, November 2005. [11] Niccolò Cottini, Stefano Marmi, and Aldo Nassigh. Why was USA Downgraded and Germany not? A market dynamics analysis. http://ssrn.com/abstract=2071091. [12] Robert C. Merton. On the pricing of corporate debt: the risk structure of interest rates. Journal of Finance, pages 449–470, May 1974. [13] Dominic O’Kane. Modelling Single-name and Multi-name Credit Derivatives. Wiley Finance, July 2008. [14] Drew D. Creal, Robert B. Gramacy, and Ruey S. Tsay. Market-based credit ratings. Journal of Business & Economic Statistics, 32(3):430–444, 2014. [15] Trevor Hastie, Robert Tibshirani, and Jerome Friedman. Element of Statistical Learning. Springer, February 2009. http://statweb.stanford.edu/~tibs/ElemStatLearn/. [16] Asa Ben-Hur and Jason Weston. A user’s guide to Support Vector Machines. In Oliviero Carugo and Frank Eisenhaber, editors, Data Mining Techniques for the Life Sciences, volume 609 of Methods in Molecular Biology, pages 223–239. Humana Press, 2010. http://dx.doi.org/10.1007/978-1-60327-241-4_13 or http://pyml.sourceforge.net/doc/howto. pdf. [17] J.C. Burges, Christopher. A tutorial on Support Vector Machines for pattern recognition. Data Mining and Knowledge Discovery, pages 121–167, March 1998. available at http://research.microsoft.com/pubs/67119/svmtutorial.pdf. [18] Chih-Wei Hsu, Chih-Chung Chang, and Chih-Jen Lin. A practical guide to support vector classification. LIBSVM project (http://www.csie.ntu.edu.tw/~cjlin/libsvm). http://www.csie.ntu.edu.tw/~cjlin/papers/guide/guide. pdf. [19] Chih-Chung Chang and Chih-Jen Lin. LIBSVM: a library for Support Vector Machines. LIBSVM project (http: //www.csie.ntu.edu.tw/~cjlin/libsvm), March 2013. http://www.csie.ntu.edu.tw/~cjlin/papers/libsvm.pdf. [20] Trevor Hastie, Gareth James, Robert Tibshirani, and Daniela Witten. Introduction to Statistical Learning. Springer, 2013. http://www.springer.com/statistics/statistical+theory+and+methods/book/978-1-4614-7137-0. [21] R Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria, 2014. [22] David Meyer, Evgenia Dimitriadou, Kurt Hornik, Andreas Weingessel, and Friedrich Leisch. e1071: Misc Functions of the Department of Statistics (e1071), TU Wien, 2014. R package version 1.6-4. 8Taking into account the remapping. 26 STEFANO MARMI, ALDO NASSIGH, AND DANIELE REGOLI [23] Max Kuhn. Contributions from Jed Wing, Steve Weston, Andre Williams, Chris Keefer, Allan Engelhardt, Tony Cooper, Zachary Mayer, and the R Core Team. caret: Classification and Regression Training, 2014. R package version 6.0-35. [24] Kenneth J. Singleton Darrell Duffie. Modeling term structures of defaultable bonds. Rev. Financ. Stud., pages 687–720, 1999. http://dx.doi.org/10.1093/rfs/12.4.687. Scuola Normale Superiore and C.N.R.S. UMI 3483 Laboratorio Fibonacci, Pisa, Italy E-mail address: [email protected] UniCredit Group, Milano and Università Commerciale Luigi Bocconi – Milan, Italy E-mail address: [email protected] Scuola Normale Superiore, Pisa, Italy E-mail address: [email protected]
© Copyright 2024 Paperzz