Risk Analysis of Collateralized Debt Obligations: Kay Giesecke and Baeho Kim February 13, 2009 This Draft March 3, 2010
Risk Analysis of Collateralized Debt Obligations: Kay Giesecke and Baeho Kim February 13, 2009 This Draft March 3, 2010
Risk Analysis of Collateralized Debt Obligations: Kay Giesecke and Baeho Kim February 13, 2009 This Draft March 3, 2010
Abstract
Collateralized debt obligations, which are securities with payoffs that are tied to
the cash flows in a portfolio of defaultable assets such as corporate bonds, play a
significant role in the financial crisis that has spread throughout the world. Insuf-
ficient capital provisioning due to flawed and overly optimistic risk assessments is
at the center of the problem. This paper develops stochastic methods to measure
the risk of positions in collateralized debt obligations and related instruments tied
to an underlying portfolio of defaultable assets. It proposes an adaptive point pro-
cess model of portfolio default timing, a maximum likelihood method for estimating
point process models that is based on an acceptance/rejection re-sampling scheme,
and statistical tests for model validation. To illustrate these tools, they are used
to estimate the distribution of the profit or loss generated by positions in multi-
ple tranches of a collateralized debt obligation that references the CDX High Yield
portfolio, and the risk capital required to support these positions.
∗
Department of Management Science & Engineering, Stanford University, Stanford, CA 94305-
4026, USA, Phone (650) 723 9265, Fax (650) 723 1614, email: [email protected], web:
www.stanford.edu/∼giesecke.
†
Korea University Business School, Anam-dong, Seongbuk-gu, Seoul 136-701 Korea, Phone +82 2
3290 2626, Fax +82 2 922 7220, email: [email protected].
‡
We are grateful for default data from Moody’s and market data from Morgan Stanley. We thank
Eymen Errais, Igor Halperin and Jack Kim for helpful discussions. We are also grateful to three anonymous
referees, an associate editor, Shahriar Azizpour, Lisa Goldberg, Steve Kou, and especially Jeremy Staum
for insightful comments.
1
1 Introduction
The financial crisis highlights the need for a holistic, objective and transparent approach
to accurately measuring the risk of investment positions in portfolio credit derivatives
such as collateralized debt obligations (CDOs). Portfolio credit derivatives are securities
whose payoffs are tied, often through complex schemes, to the cash flows in a portfolio
of credit instruments such as corporate bonds, loans, or mortgages. They facilitate the
trading of insurance against the default losses in the portfolio. An investor providing the
insurance is exposed to the default risk in the portfolio.
There is an extensive literature devoted to the valuation and hedging of portfolio
derivatives.1 The basic valuation problem is to estimate the price of default insurance,
i.e., the arbitrage-free value of the portfolio derivative at contract inception. This value
is given by the expected discounted derivative cash flows relative to a risk-neutral pricing
measure. After inception, the derivative position must be marked to market; that is, the
value of the derivative under current market conditions must be determined. The basic
hedging problem is to estimate the sensitivities of the derivative value to changes of
the default risk of the portfolio constituents. These sensitivities determine the amount
of constituent default insurance to be bought or sold to neutralize the derivative price
fluctuations due to small changes of the constituent risks.
The valuation and hedging problems are distinct from the risk analysis problem,
which is to measure the exposure of the derivative investor, who provides default in-
surance, to potential payments due to defaults in the portfolio. More precisely, the goal
is to estimate the distribution of the investor’s cumulative cash flows over the life of the
contract. The distribution is taken under the actual measure describing the empirical like-
lihood of events, rather than a risk-neutral pricing measure. The distribution describes the
risk/reward profile of a portfolio derivative position, and is the key to risk management
applications. For example, it allows the investor or regulator to determine the amount of
risk capital required to support a position. The financial crisis indicates the significance of
these applications and the problems associated with the traditional rating based analysis
of portfolio credit derivative positions; see SEC (2008).
The risk analysis problem has been largely ignored in the academic literature. This
paper provides stochastic methods to address this problem. It makes several contributions.
First, it develops a maximum likelihood approach to estimating point process models of
portfolio default timing from historical default experience. Second, it devises statistical
tests to validate a fitted model. Third, it formulates, fits and tests an adaptive point
process model of portfolio default timing, and demonstrates the utility of the estimation
and validation methods on this model. Fourth, it addresses important risk management
1
See Arnsdorf & Halperin (2008), Brigo, Pallavicini & Torresetti (2007), Chen & Glasserman (2008),
Cont & Minca (2008), Ding, Giesecke & Tomecek (2009), Duffie & Garleanu (2001), Eckner (2009), Errais,
Giesecke & Goldberg (2009), Kou & Peng (2009), Longstaff & Rajan (2008), Lopatin & Misirpashaev
(2008), Mortensen (2006), Papageorgiou & Sircar (2007) and many others.
2
applications, including the estimation of profit and loss distributions for positions in mul-
tiple tranches of a CDO. These distributions quantify and differentiate the risk exposure
of alternative investment positions, and the impact of complex contract features. They
are preferable to agency ratings, which are often based on the first moment only.
Estimating a stochastic point process model of portfolio default timing under the
actual probability measure presents unique challenges. Most importantly, inference must
be based on historical default timing data, rather than market derivative pricing data.
However, the default history of the reference portfolio underlying the credit derivative is
often unavailable, so direct inference is usually not feasible. We confront this difficulty
by developing an acceptance/rejection re-sampling scheme that allows us to generate
alternative portfolio default histories from the available economy-wide default timing data.
These histories are then used to construct maximum likelihood estimators for a portfolio
point process model that is specified in terms of an intensity process. A time-scaling
argument leads to testable hypotheses for the fit of a model.
The re-sampling approach is predicated on a top-down formulation of the portfolio
point process model. This formulation has become popular in the credit derivatives pricing
literature. Here, the point process intensity is specified without reference to the portfolio
constituents. In our risk analysis setting, the combination of top-down formulation and re-
sampling based inference leads to low-dimensional estimation, validation and prediction
problems that are highly tractable and fast to address even for the large portfolios that
are common in practice. Alternative bottom-up formulations in Das, Duffie, Kapadia
& Saita (2007), Delloye, Fermanian & Sbai (2006) and Duffie, Eckner, Horel & Saita
(2009) require the specification and estimation of default timing models for the individual
portfolio constituent securities. They allow one to incorporate firm-specific data into the
estimation, but lead to high-dimensional computational problems.
To demonstrate the effectiveness of the re-sampling approach and the appropriateness
of the top-down formulation, we develop and fit an adaptive intensity model. This model
extends the classical Hawkes (1971) model by including a state-dependent drift coefficient
in the intensity dynamics. The state-dependent drift involves a reversion level and speed
that are proportional to the intensity at the previous event. While this specification is
as tractable as the Hawkes model, it avoids the constraints imposed by the constant
Hawkes reversion level and speed. This helps to better fit the regime-dependent behavior
of empirical default rates. In- and out-of-sample tests show that our adaptive model
indeed captures the clustering in the default arrival data. Because it can capture the
default clustering better than other models that have been used in practice, our model
leads to more realistic estimates of the probabilities of losses to senior tranches.
This rest of this paper is organized as follows. Section 2 develops the re-sampling
approach to point process model estimation and validation. Section 3 formulates and
analyzes the adaptive point process model, and uses the re-sampling approach to fit it.
Section 4 applies the fitted model to the risk analysis of synthetic CDOs, while Section 5
analyzes the risk of cash CDOs. Section 6 concludes. There are several appendices.
3
2 Re-sampling based inference
This section develops a re-sampling approach to estimating a stochastic point process
model of default timing. It also provides a method to validate the estimators.
Proposition 2.1. Let Z ∗ be a predictable process with values in [0, 1]. Select an economy-
wide event time Tn∗ with probability ZT∗n∗ . If the economy-wide default process N ∗ has
intensity λ∗ , then the counting process of the selected times has intensity Z ∗ λ∗ .
Proposition 2.1 states that the counting process obtained by thinning N ∗ according
to Z ∗ has intensity given by the product of the thinning process Z ∗ and the intensity of N ∗ .
4
Algorithm 1 (Acceptance/Rejection Re-Sampling) Generating a sample path of
the portfolio default process N from the realization of the economy-wide default process
N ∗ over the sample period [0, τ ].
1: Initialize m ← 0.
2: for n = 1 to Nτ∗ do
3: Draw u ∼ U (0, 1).
4: if u ≤ ZT∗n∗ then
5: Assign Tm+1 ← Tn∗ and update m ← m + 1.
6: end if
7: end for
E(Nt+ − Nt | Ft )
Zt∗ (ω) = lim ∗
(ω) (1)
→0 E(Nt+ − Nt∗ | Ft )
in all those points (ω, t) ∈ Ω × (0, ∞] where the limit exists.2 The quotient on the right
side of equation (1), which is taken to be zero when the denominator vanishes, represents
the conditional probability at time t that the next defaulter is a reference name, given
that a default occurs in the economy by time t + .
2
The limit in (1) exists and is equal to Zt∗ (ω) almost surely with respect to a certain measure on the
product space Ω × (0, ∞]. See Giesecke et al. (2009) for more details.
5
We specify Z ∗ nonparametrically, guided by formula (1). Intuitively, Z ∗ must reflect
the relation between the issuer composition of the economy and the issuer composition
of the reference portfolio. We propose to describe the issuer composition in terms of the
credit ratings of the respective constituent issuers. In this case, N ∗ is identified with the
default process in the universe of rated issuers.
The use of ratings does not limit the applicability of our specification, because the
reference portfolios of most derivatives consist of rated names only. Moreover, the rating
agencies maintain extensive and accessible data bases that record credit events including
defaults in the universe of rated names, and further attributes associated with these
events, such as recovery rates. The agencies have maintained a high firm coverage ratio
throughout the sectors of the economy, and therefore the universe of rated names is a
reasonable representation of the firms in the economy.
Let [0, τ ] be the sample period, with τ denoting the (current) analysis time. Let R
be the set of rating categories, Xτ (ρ) be the number at time τ of reference firms with
rating ρ ∈ R, and Xt∗ (ρ) be the number at time t ∈ [0, τ ] of ρ-rated firms in the universe
of rated names. The number of firms X ∗ (ρ) is an adapted process. It varies through time
because issuers enter and exit the universe of rated names. Exits can be due to mergers
or privatizations, for example. We assume that the thinning process Z ∗ is equal to the
predictable projection of the process defined for times 0 < t ≤ τ by3
Xτ (ρ∗Nt−
∗ +1 )
∗
1{Xt−
∗ (ρ∗ )>0} (2)
Xt− (ρ∗Nt−
∗ +1 )
N∗
t− +1
where ρ∗n ∈ FTn∗ is the rating at the time of default of the nth defaulter in the economy.4
We require that Xτ (ρ) ≤ Xt∗ (ρ) for all t ≤ τ and ρ ∈ R. Since Xτ (ρ) is a fixed integer
∗
prescribed by the portfolio composition and Xt− (ρ) is predictable for fixed ρ,
X Xτ (ρ)
Zt∗ = ∗
P (ρ∗Nt−
∗ +1 = ρ | Ft− ) (3)
ρ∈R∗
X t− (ρ)
t−
almost surely, where Rt∗ is the set of rating categories ρ ∈ R for which Xt∗ (ρ) > 0. Formula
(3) suggests to interpret the value Zt∗ as the conditional “empirical” probability that the
next defaulter is a reference name. This conditional probability respects the ratings of
the reference names, as P (ρ∗Nt−∗ +1 = ρ | Ft− ) is the conditional probability that the next
defaulter has rating ρ. Our estimator νt∗ (ρ) of this latter conditional probability is based
on the ratings of the defaulters in [0, t). For ρ ∈ R, it is given by
∗
PNt−
1{ρ∗ =ρ} + α
νt∗ (ρ) = PN ∗ n=1 n 1{ρ∈Rt−
∗ } (4)
t− ∗
n=1 1{ρ ∗ ∈R∗ } + α|R
n t− t− |
3
Here and below, if Y is a right-continuous process with left limits, Yt− = lims↑t Ys .
4
Taking Z ∗ to be the predictable projection of the process given by formula (2) guarantees that Z ∗ is
defined up to indistinguishability; see Dellacherie & Meyer (1982).
6
where α ∈ (0, 1] is an additive smoothing parameter guaranteeing that νt∗ (ρ) is well-defined
for t < T1∗ . For α = 0, equation (4) defines the empirical rating distribution, which treats
the observations ρ∗1 , . . . , ρ∗Nt−
∗ as independent samples from a common distribution and
ignores all other information contained in Ft− . Our implementation assumes α = 0.5, a
value that can be justified on Bayesian grounds, see Box & Tiao (1992, pages 34-36).5
Proposition 2.2. Suppose that Zt∗ > 0, almost surely. Then the economy-wide default
process N ∗ is a standard Poisson process under a change of time defined by
Z t
∗ 1
At = λ ds.
∗ s
(6)
0 Zs
7
Algorithm 2 (Replacement Thinning) Generating a sample path of the portfolio
default process N 0 without replacement from a path of the portfolio default process N
with replacement over [τ, H], for a horizon H > τ .
1: Initialize m ← 0 as we define T00 = τ , and set Y (ρ) ← Xτ (ρ) for all ρ ∈ R.
2: for n = Nτ + 1 to NH do
3: Draw u ∼ U (0, 1).
4: if u ≤ ZTn then
5: Assign Tm0 ← Tn and update m ← m + 1.
6: Draw ρm ∼ ν 0 , where
Y (ρ)
ν 0 (ρ) = ν(ρ)/ZTn , ρ ∈ R. (9)
Xτ (ρ)
Consider event times Tn of N over some interval [τ, H], where H > τ . The event
times Tn0 of the portfolio default process without replacement, N 0 , can be obtained from
the Tn by removing event times due to replaced defaulters. This is done by thinning. To
formalize this, let Xt (ρ) be the number of ρ-rated reference names at time t ≥ τ , assuming
defaulters are not replaced. Thus, for fixed ρ, Xt (ρ) ≤ Xτ (ρ) almost surely for every t ≥ τ .
For fixed ρ, the process X(ρ) decreases and vanishes when all ρ-rated reference names are
in default. It suffices to specify Z, the thinning process for N , at the event times Tn ≥ τ
of N . Motivated by formula (3), we suppose that
X XT − (ρ)
ZTn = n
P (ρn = ρ | FTn− ) (7)
ρ∈R
X τ (ρ)
τ
where Rτ is the set of rating categories ρ ∈ R for which Xτ (ρ) > 0, and ρn is the rating of
the firm defaulting at Tn . Note that the thinning probability (7) vanishes when all firms
in the portfolio are in default. We estimate the conditional distribution P (ρn = · | FTn− ) by
the smoothed empirical distribution ν of the ratings of the defaulters in the re-sampling
scenarios {N (ωi ) : i ≤ I}, where
PI PNτ (ωi )
i=1 n=1 1{ρn (ωi )=ρ} + α
ν(ρ) = PI PNτ (ωi ) 1{ρ∈Rτ } (8)
i=1 n=1 1{ρn (ωi )∈Rτ } + α|Rτ |
for an additive smoothing parameter α ∈ [0, 1]; see formula (4). This estimator treats the
observations ρn (ωi ) of all paths ωi as independent samples from a common distribution.
Algorithm 2 summarizes the steps required to generate N 0 from N .
8
3 An adaptive intensity model
This section demonstrates the effectiveness of the re-sampling approach. It formulates,
fits and evaluates an adaptive intensity model for the CDX High Yield portfolio, which is
a standard portfolio that is referenced by a range of credit derivatives. The fitted intensity
model will be used in Sections 4 and 5 to analyze the risk of such derivatives.
9
4 12
3.5
10
8
2.5
Percent
Percent
2 6
1.5
4
2
0.5
0 0
1970 1975 1980 1985 1990 1995 2000 2005 1970 1975 1980 1985 1990 1995 2000 2005
Figure 1: Left panel : Annual default rate, relative to the number of rated names at the
beginning of a year, in the universe of Moody’s rated corporate issuers in any year between
1970 and 2008, as of 11/7/2008. Source: Moody’s Default Risk Service. Right panel : Mean
annual default rate for the CDX.HY6 portfolio for I = 10K re-sampling scenarios.
Using in- and out-of-sample tests, Azizpour & Giesecke (2008b) found that prediction of
economy-wide default activity based on past default timing outperforms prediction based
on exogenous economic covariates. Intuitively, the timing of past defaults provides infor-
mation about the timing of future defaults that is statistically superior to the information
contained in exogenous covariates. Further, if the past default history is the conditioning
information set, then the inclusion of additional economic covariates does not improve
economy-wide default forecast performance. These empirical findings motivate the formu-
lation of a parsimonious portfolio intensity model whose conditioning information set is
given by the past default history.
We assume that λt is a function of the path of N over [0, t]. More specifically, we
propose that λ evolves through time according to the equation
where λ0 > 0 is the initial intensity value, κt = κλTNt is the decay rate, ct = cλTNt is the
reversion level, and J is a response jump process given by
X
Jt = max(γ, δλTn− )1{Tn ≤t} . (11)
n≥1
The quantities κ > 0, c ∈ (0, 1), δ > 0 and γ ≥ 0 are parameters. We denote by θ
the vector (κ, c, δ, γ, λ0 ). We give a sufficient condition guaranteeing that Nt < ∞ almost
surely, for all t. This condition relates the reversion level parameter c to the parameter δ,
which controls the magnitude of a jump of the intensity at an event.
Proposition 3.1. If c(1 + δ) < 1, then the counting process N is non-explosive.
10
The intensity (10) follows a piece-wise deterministic process with right-continuous
sample paths. It jumps at an event time Tn . The jump magnitude is random. It is equal
to max(γ, δλTn− ), and depends on the intensity just before the event, which itself is a
function of the event times T1 , . . . , Tn−1 . The minimum jump size is γ. From Tn onwards
the intensity reverts exponentially to the level cλTn , at rate κλTn . Since the reversion rate
and level are proportional to the value of the intensity at the previous event, they depend
on the times T1 , . . . , Tn and change adaptively at each default. For Tn ≤ t < Tn+1 , the
behavior of the intensity is described by the FTn -measurable function
The dependence of the reversion level ct , reversion speed κt and jump magnitude
max(γ, δλTn− ) on the path of the counting process N distinguishes our specification (10)
from the classical Hawkes (1971) model. The Hawkes intensity follows a piece-wise de-
terministic process dλt = κ(c − λt )dt + δdUt , where U = u1 + · · · + uN and the jump
magnitudes un are drawn from a fixed distribution on R+ . This model is more rigid than
our adaptive model (10): it imposes a global, state-independent reversion level c and speed
κ, and the magnitude of a jump in the Hawkes intensity is drawn independently of past
event times. Figure 2 contrasts the sample paths of (λ, N ) for the Hawkes model with
those for our model (10). The paths exhibit different clustering behavior, with the Hawkes
model generating a more regular clustering pattern. While Azizpour & Giesecke (2008a)
found that a variant of the Hawkes model performs well on the economy-wide default
data, we had difficulty fitting this and several other variants of the Hawkes model to the
portfolio default times generated by the re-sampling mechanism. We found the constant
reversion level and speed to be too restrictive. Our adaptive specification (10) relaxes
these constraints while preserving parsimony and computational tractability.
The jumps of the intensity process (10) are statistically important. They generate
event correlation: an event increases the likelihood of further events in the near future.
This feature facilitates the replication of the event clusters seen in Figure 1, a fact that
we establish more formally below. The intensity jumps can also be motivated in economic
terms. They represent the impact of a default on the other firms, which is channeled
through the complex web of contractual relationships in the economy. The existence of
these feedback phenomena is indicated by the ripple effects associated with the default
of Lehman Brothers on September 15, 2008, and is further empirically documented in
Azizpour & Giesecke (2008a), Jorion & Zhang (2007) and others. Our jump size specifi-
cation guarantees that the impact of an event increases with the default rate prevailing at
the event: the weaker the firms the stronger the impact. An alternative motivation of the
intensity jumps is Bayesian learning: an event reveals information about the values of un-
observed event covariates, and this leads to an update of the intensity, see Collin-Dufresne,
Goldstein & Helwege (2009), Duffie et al. (2009) and others.
11
35 40 10 40
9
30 35 35
8
30 30
25 7
Number of Defaults
Number of Defaults
25 25
6
20
Intensity
Intensity
20 5 20
15
4
15 15
10 3
10 10
2
5 5 5
1
0 0 0 0
0 2 4 6 8 10 0 2 4 6 8 10
Figure 2: Left panel : Sample path of the intensity (left scale, solid) and default process
(right scale, dotted) for the adaptive model (10) with θ = (0.25, 0.05, 0.4, 0.8, 2.5), values
that are motivated by our estimation results in Section 3.4. The reversion level and speed
change at each event. Algorithm 3 is used to generate the paths. Right panel : Sample
path of the intensity and default process for the Hawkes model dλt = κ(c − λt )dt + δdUt
where the jump magnitudes un = 1 so U = N , λ0 = 2.5, and κ = 2.5 and c = δ = 1.5 are
chosen so that the expected number of events over 10 years matches that of the model
(10), roughly 37. Note that the Hawkes intensity cannot fall below the global reversion
level c. The algorithm in Ogata (1981) is used to generate the Hawkes model paths.
12
Algorithm 3 (Default Time Simulation) Generating a sample path of the portfolio
default process N over [τ, H] for a horizon H > τ and the model (10).
1: Draw (Nτ , TNτ , λTNτ ) uniformly from {(Nτ (ωi ), TNτ (ωi ), λTNτ (ωi )) : i ≤ I}.
2: Initialize (n, S) ← (Nτ , τ ) and λS ← cλTn + (1 − c)λTn exp(−κλTn (S − Tn )).
3: loop
4: Draw E ∼ Exp(λS ) and set T ← S + E.
5: if T > H then
6: Exit loop.
7: end if
8: λT ← cλTn + (λS − cλTn ) exp(−κλTn (T − S))
9: Draw u ∼ U (0, 1).
10: if u ≤ λT /λS then
11: λT ← λT + max(γ, δλT )
12: Assign Tn+1 ← T and update n ← n + 1.
13: end if
14: Set S ← T and λS ← λT .
15: end loop
13
Parameter κ c δ γ λ0
MLE 0.254 0.004 0.419 0.810 8.709
95% CI (A) (0.252,0.255) (0.004,0.004) (0.417,0.422) (0.806,0.813) (8.566,8.851)
95% CI (B) (0.250,0.258) (0.003,0.004) (0.406,0.433) (0.791,0.829) (8.538,8.882)
Median 0.263 0.004 0.430 0.779 9.405
Table 1: Maximum likelihood estimates of the parameters of the intensity λ for the
CDX.HY6 portfolio, along with estimates of asymptotic (A) and bootstrapping (B) 95%
confidence intervals (10K bootstrap samples were used). The “Median” row indicates the
median of the empirical distribution of the per-path MLEs over all re-sampling paths. The
estimates are based on I = 10K re-sampling scenarios, generated by Algorithm 1 from the
observed defaults in the universe of Moody’s rated names from 1/1/1970 to 11/7/2008.
also provide asymptotic and bootstrapping confidence intervals. The left panel of Figure
3 shows the path of the fitted mean intensity λ b = I −1 PI λθb(ωi ).
i=1
To provide some perspective on the parameter estimates, we employ an alternative
inference procedure. Instead of maximizing the total likelihood (5) associated with all
paths of N , we maximize the path log-likelihood
Z τ
θ(ω )
sup (log λs− i (ωi )dNs (ωi ) − λθ(ω
s
i)
(ωi )ds)
θ(ωi )∈Θ 0
for each i = 1, 2, . . . , I. The last row in Table 1 shows the median of the empirical distri-
bution of the per-path MLE θ(ωi ) over all paths ωi . These values are in good agreement
with the MLEs, supporting our total likelihood estimation strategy.
14
20 110
Fitted Mean Portfolio Intensity Semi−annual Defaults
18 100 Fitted Economy−wide Intensity
16 90
80
14
70
12
60
10
50
8
40
6
30
4 20
2 10
0 0
1970 1975 1980 1985 1990 1995 2000 2005 1970 1975 1980 1985 1990 1995 2000 2005
large deviations of the (Wn∗ ) from their theoretical mean 1. Prahl shows that if the (Wn∗ )
are independent samples from a standard exponential distribution, then
1 X W∗
M= 1− n (13)
m n:W ∗ <µ µ
n
15
9
1
8
5 0.6
0.4
3
2
0.2
1
0 0
0 1 2 3 4 5 6 7 8 9 0 0.5 1 1.5 2 2.5 3 3.5 4
Empirical Quantiles
A benefit of the time-scaling tests is that they can be applied to alternative model for-
mulations, facilitating a direct comparison of fitting performance. Consider the bottom-up
specification of Das et al. (2007), which appears to be much richer than ours: they specify
firm-level intensity models with conditioning information given by a set of firm-specific
and macro-economic covariates. Das et al. (2007) find that the time-scaled, economy-wide
times generated by their model deviate significantly from those of a Poisson process. In
particular, they find that their specification does not completely capture the event clusters
in the data. Thus, they reject this model formulation at standard confidence levels. This
may indicate that, relative to our portfolio-level formulation with conditioning informa-
tion given by past default timing, the additional modeling and estimation effort involved
in a firm-level intensity model formulation with a large set of exogenous covariates may
not translate into better fits and forecast performance.
Finally we note that the time-scaling tests can also be used to evaluate the specifica-
tion of λ directly on the re-sampling scenarios {N (ωi ) : i ≤ I}. For each path ωi , we time-
R·
scale the portfolio default process N (ωi ) with its fitted compensator A(ωi ) = 0 λs (ωi )ds,
and calculate the statistics of the KS test and Prahl’s test for the time-scaled inter-arrival
R Tn (ωi ) θb
times Wn = Tn−1 λ (ωi )ds, which can be calculated exactly thanks to formula (12).
(ωi ) s
The 95% confidence interval for the p-value of the KS test is 0.24 ± 0.01. The 95% confi-
dence interval for Prahl’s test statistic, defined for (Wn ) in analogy to (13), is 0.81 ± 0.07.
Also these tests indicate that the adaptive model (10) is well-specified.
16
10 8
Realized Loss Realized Loss
9
7
8
6
7
5
6
1 Year Loss
1 Year Loss
5 4
4
3
3
2
2
1
1
0 0
1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007
Figure 5: Out-of-sample test of loss forecasts for a test portfolio with the same rating
composition as the CDX.HY6. We show the [25%, 75%] percentiles (box), the [15%, 85%]
percentiles (whiskers) and the median (horizontal line) of the fitted conditional distribu-
tion of the incremental portfolio loss L0τ +1 − L0τ given Fτ for τ varying annually between
1/1/1996 and 1/1/2007, estimated from 100K paths generated by Algorithms 3 and 2,
and the realized portfolio loss during [τ, τ + 1]. Left panel : The test portfolio is selected
at the beginning of each period. Right panel : The test portfolio is selected in 1996.
17
survived to the beginning of the period, and compare it with the realized portfolio loss
during the forecast period. This setting is motivated by the situation of a buy and hold
investor. The graphs indicate that the portfolio loss forecasts are very accurate.
The left panel of Figure 6 shows the conditional distribution of the normalized 5 year cu-
mulative tranche loss UH (K, K)/C(K − K) for the CDX.HY6 on 3/27/2006, the Series 6
contract inception date,14 for each of several standard attachment point pairs. The matu-
rity date H for Series 6 contracts is 6/27/2011.15 To estimate the tranche loss distribution,
we generate default scenarios during 1/1/1970 – 3/27/2006 for the CDX portfolio from
Moody’s default history by the re-sampling Algorithm 1. Next we estimate the intensity
model (10) from these scenarios, as described in Section 3.4, with parameter estimates re-
ported in Table 2. Then we generate event times during the prediction interval 3/27/2006
14
Every 6 months, the CDX High Yield index portfolio is “rolled.” That is, a new portfolio with a new
serial number is formed by replacing names that have defaulted since the last roll, and possibly other
names. The index and tranche swaps we consider are tied to a fixed series.
15
Although the actual time to maturity is 5 years and 3 months at contract inception, here and below,
we follow the market convention of referring to the contract as a “5 year contract.” At an index roll, new
“5 year contracts” are issued, and these mature in 5 years and 3 months from the roll date.
18
Index 0−10% 10−15% 15−25% Index 0−10% 10−15% 15−25%
1 1
0.9 0.9
0.8 0.8
Cumulative Loss Distribution
0.6 0.6
0.5 0.5
0.4 0.4
0.3 0.3
0.2 0.2
0.1 0.1
0 0
0 20 40 60 80 100 0 20 40 60 80 100
Index/Tranche Loss (%) Index/Tranche Loss (%)
– 6/27/2011 from the fitted intensity using Algorithm 3, which we thin using Algorithm 2
to obtain paths of portfolio default times and losses without replacement. The trajectories
for U (K, K) thus obtained lead to an unbiased estimate of the desired distribution.
The loss distributions indicate the distinctions between the risk profiles of the different
tranches. The equity tranche, which has attachment points 0 and 10%, carries the highest
exposure among all tranches. For the equity protection seller, the probability of trivial
losses is roughly 15%, and the probability of losing the full tranche notional is 20%.
For the (10%, 15%)-mezzanine protection seller, the probabilities are roughly 80% and
10%, respectively. The mezzanine tranche is less risky, since the equity tranche absorbs
the first 10% of the total portfolio loss. For the (15%, 25%)-senior protection seller, the
probabilities are roughly 90% and 5%, respectively.
To provide some perspective on these numbers, we also estimate the risk-neutral
tranche loss distributions implied by the market prices paid for 5 year tranche protection
on 3/27/2006. The estimation procedure, explained in Appendix B, is based on the in-
tensity model (10) relative to a risk-neutral pricing measure. The risk-neutral intensity
parameters are chosen such that the model prices match the market index and tranche
prices as closely as possible. The fitting errors are small, which indicates that our adaptive
intensity model (10) performs also well under a risk-neutral pricing measure.
The right panel of Figure 6 shows the risk-neutral distribution of UH (K, K)/C(K−K)
19
Parameter κ c δ γ λ0
MLE 0.260 0.003 0.439 0.856 8.494
95% CI (A) (0.259,0.261) (0.003,0.003) (0.437,0.441) (0.853,0.859) (8.368,8.621)
95% CI (B) (0.256,0.265) (0.003,0.004) (0.424,0.454) (0.836,0.876) (8.327,8.650)
for the CDX.HY6, for each of several standard attachment point pairs. For the equity
protection seller, the risk-neutral probability of trivial losses is zero, and the risk-neutral
probability of losing the full tranche notional is 70%. Compare with the actual probabilities
indicated in the left panel of Figure 6. For any tranche, the risk-neutral probability of
losing more than any given fraction of the tranche notional is much higher than the
corresponding actual probability. The distinction between the probabilities reflects the risk
premium the protection seller requires for bearing exposure to the correlated corporate
default risk in the reference portfolio. Figure 7 shows the distributions of the normalized
cumulative portfolio loss L0H /C for multiple horizons H. For all horizons, the risk-neutral
distribution has a much fatter tail than the corresponding actual distribution. In other
words, when pricing index and tranche contracts, the market overestimates the probability
of extreme default scenarios relative to historical default experience.
Our tools can also be used to analyze more complex investment positions. Investors
often trade the CDO capital structure, i.e. they simultaneously sell and buy protection in
different tranches. A popular trade used to be the “equity-mezzanine” trade, in which the
investor sells equity protection and hedges the position by buying mezzanine protection
with matching notional and maturity. Figure 8 contrasts the conditional distribution on
3/27/2006 of the normalized cumulative loss (UH (0, 0.1) − UH (0.1, 0.15))/0.1C generated
by this trade with the conditional distribution of the normalized cumulative equity loss
UH (0, 0.1)/0.1C, both for the CDX.HY6. While the mezzanine hedge does not alter the
probability of trivial losses in the equity-only position, it does reduce the probability of a
total loss of notional from 20% to virtually zero. This is because the mezzanine protection
position generates cash flows when equity is wiped out. The magnitude of these cash flows
is however capped at 50% of the total equity notional, and this property generates the
point mass at 5% in the loss distribution of the equity-mezzanine position.
We can also measure the risk of positions in tranches referenced on different portfolios.
For example, an investor may sell and buy protection on several of the reference portfolios
in the CDX family, including the High Yield, Investment Grade and Crossover portfolios.
In this case, we estimate default and loss processes for each of the portfolios based on the
20
0.1 0.1
0.08 0.08
0.06 0.06
0.04 0.04
0.02 0.02
0 0
10 10
9 9
8 8
7 7
6 40 6 40
5 30 Time to 5 30
Time to 20 20
Maturity (Years) 4 Maturity (Years) 4
10 10
3 0 3 0
Loss (%) Loss (%)
realization of the economy-wide default process. We generate events for each portfolio,
and then aggregate the corresponding position losses as in the equity-mezzanine case.
21
1
0.9
0.8
0.6
0.5
0.4
0.3
0.2
0.1 Equity−Mezzanine
Equity Only
0
0 2 4 6 8 10
Loss (%)
per-period coupon rate v. The total interest income to the SPV in period m is therefore
When a reference bond defaults, the SPV collects the recovery at the coupon date following
the event. The total recovery cash flow in period m is
The total cash flows from coupons and recoveries are invested in a reserve account that
earns interest at the risk-free rate r. For period m, they are given by
W (m) + K(m) if 1 ≤ m < M
B(m) = 0
W (m) + K(m) + C − Ntm if m = M.
The SPV issues three tranches, of which two are debt tranches that promise to pay a
specified coupon at times (tm ). There is one senior debt tranche, represented by a sinking-
fund bond with initial principal p1 and per-period coupon rate c1 , and a junior debt tranche
that is represented by a sinking-fund bond with initial principal p2 and per-period coupon
rate c2 . The initial principal of the residual equity tranche is p3 = C −p1 −p2 . Each tranche
has maturity date H.
For the debt tranches j = 1, 2 and coupon period m, we denote by Fj (m) the re-
maining principal. The scheduled interest payment is Fj (m)cj , and the accrued unpaid
22
Figure 9: Cash flow “waterfall” of a sample cash CDO.
interest is Aj (m − 1). If Qj (m), the actual interest paid in period m, is less than Fj (m)cj ,
then the difference is accrued at cj to generate an accrued unpaid interest of
The actual payments to the debt tranches are prioritized. We consider two prioritization
schemes, the uniform and fast schemes, which were introduced by Duffie & Garleanu
(2001) and which are reviewed in Appendix C for completeness. A prioritization scheme
specifies the actual interest payment Qj (m), the pre-payment of principal Pj (m), and the
contractual unpaid reduction in principal Jj (m). The total cash payment in period m is
Qj (m) + Pj (m). The remaining principal after interest payments is
The par coupon rate on tranche j is the scheduled coupon rate cj with the property that
the initial market value of the bond is equal to its initial face value Fj (0) = pj .
The residual equity tranche does not make scheduled coupon or principal payments
so c3 = 0. Instead, at maturity H, after the debt tranches have been paid as scheduled,
any remaining funds in the reserve account are allocated to the equity tranche.
We analyze the exposure of a tranche investor for a 5 year cash CDO referenced on
the CDX.HY6 of C = 100 names. This choice of reference portfolio allows us to compare
the results with those for the synthetic CDO. Before we can start this analysis, we need
to price the reference bonds and the debt tranches.16 The first step is to estimate the
par coupon rate of the reference bonds. This is the scheduled coupon rate v ∗ with the
16
In practice, this is done by the SPV, which then offers the tranche terms to potential investors. We
have to perform this task here since we do not have access to actual data for the structure we analyze.
23
Reference Uniform Scheme Fast Scheme
Bond Senior Junior Senior Junior
Annualized Par Coupon Rate (bp) 811.23 476.58 1299.42 483.38 749.75
Par Coupon Spread (bp) 339.20 23.68 846.52 11.29 277.77
Table 3: Fitted annualized par coupon rates and spreads on 3/27/2006 for the constituent
bonds and debt tranches of a 5 year cash CDO referenced on the CDX.HY6, for each
of two standard prioritization schemes. The principal values are (p1 , p2 , p3 ) = (85, 10, 5).
The fitting procedure is described in Appendix D.
property that the initial market value of a reference bond is equal to its initial face value
1. Given v ∗ , the second step is to estimate the par coupon rates c∗j of the debt tranches.
Appendix D gives details on these steps, and Table 3 reports the annualized par coupon
rates and spreads. The par coupon spread is the difference between the par coupon rate
and the (hypothetical) par coupon rate that would be obtained if the reference bonds
were not subject to default risk.
Next we estimate the conditional distribution of the discounted cumulative tranche
loss, which is the loss the tranche investor faces during the life of the contract, discounted
at the risk-free rate r. At a post-inception time t ≤ H, the discounted cumulative tranche
loss is given by
Ujt (H, pj , v ∗ , c∗1 , c∗2 ) = V jt (H, pj , v ∗ , c∗1 , c∗2 ) − Vjt (H, pj , v ∗ , c∗1 , c∗2 ) (16)
where Vjt (H, pj , v ∗ , c∗1 , c∗2 ) is the present value at time t of the coupon and principal cash
flows actually paid to tranche j over the life of the tranche, and V jt (H, pj , v ∗ , c∗1 , c∗2 ) is the
present value of these cash flows assuming no defaults occur during the remaining life. For
a debt tranche, V jt (H, pj , v ∗ , c∗1 , c∗2 ) represents the present value of all scheduled coupon
and principal payments. For the equity tranche, this quantity represents the present value
of the reserve account value after the debt tranches have been paid as scheduled.
Note that due to the complex structure of the cash flow prioritization, the tranche loss
Ujt (H, pj , v ∗ , c∗1 , c∗2 ) does not admit a simple analytic expression in terms of the portfolio
default count and loss of the reference portfolio. This leaves simulation as the only feasible
tool for analyzing the cash CDO tranche loss, regardless of whether or not the models of
the portfolio default count and loss processes are analytically tractable.
Figure 10 shows the fitted conditional distribution on 3/27/2006 of the normalized
loss Ujτ (H, pj , v ∗ , c∗1 , c∗2 )/V jτ (H, pj , v ∗ , c∗1 , c∗2 ) for a 5 year structure whose maturity date
H is 6/27/2006, for each of several tranches. To estimate that distribution, we proceed as
described in Section 4 for the synthetic CDO, and generate paths of the fitted portfolio
default and loss processes. These are then fed into the cash flow calculator, which computes
Vjt (H, pj , v ∗ , c∗1 , c∗2 ) for each path. We assume that coupons are paid quarterly. The risk-
free interest rate r is deterministic, and is estimated from Treasury yields for multiple
maturities on 3/27/2006, obtained from the website of the Department of Treasury.
24
1 1
0.9 0.9
0.8 0.8
Cumulative Loss Distribution
0.6 0.6
0.5 0.5
0.4 0.4
0.3 0.3
0.2 0.2
Senior Senior
0.1 Junior 0.1 Junior
Residual Residual
0 0
0 20 40 60 80 100 0 20 40 60 80 100
Tranche Loss (%) Tranche Loss (%)
Figure 10: Kernel smoothed conditional distribution on 3/27/2006 of the normalized dis-
counted loss Ujτ (H, pj , v ∗ , c∗1 , c∗2 )/V jτ (H, pj , v ∗ , c∗1 , c∗2 ) for a 5 year cash CDO referenced on
the CDX.HY6, whose maturity date H is 6/27/2011. The initial tranche principals are
p1 = 85 for the senior tranche, p2 = 10 for the junior tranche, and p3 = 5 for the resid-
ual equity tranche. We apply the model and fitting methodology developed in Sections 2
and 3, based on I = 10K re-sampling scenarios for N , and 100K replications. Left panel :
Uniform prioritization scheme. Right panel : Fast prioritization scheme.
As in the synthetic CDO case, the risk profile of a cash CDO tranche depends on the
degree of over-collateralization, i.e. the location of the attachment points. It also depends
on the prioritization scheme specified in the CDO terms. With the uniform scheme, the
cash flows generated by the reference bonds are allocated sequentially to the debt tranches
according to priority order, and default losses are applied to all tranches in reverse priority
order. With the fast scheme, the cash flows are used to retire the senior tranche as soon as
possible. After the senior tranche is retired, the cash flows are used to service the junior
tranche. After the junior tranche is retired, any residual cash flows are distributed to
the equity tranche. Therefore, the senior tranche investor is less exposed under the fast
scheme: the probability of zero losses is increased from roughly 95% for the uniform scheme
to roughly 98%. This reduction in risk is reflected by the par coupon spreads reported in
Table 3. For the junior investor, the probability of zero losses increases from roughly 82%
for the uniform scheme to roughly 93% for the fast scheme, but the probability of a loss
equal to the present value of all scheduled coupon and principal payments increases from
zero to about 3%. Nevertheless, the junior tranche commands a much smaller spread
under the fast scheme. The risk reduction for the debt tranche investors comes at the
expense of the equity investor: while the probability of zero losses is about 15% for both
schemes, the probability of a loss equal to the present value of the reserve account value
after the debt tranches have been paid as scheduled increases from zero to roughly 8% for
the fast scheme. On the other hand, the equity investor has a much higher upside under
25
1
Uniform Scheme
0.9 Fast Scheme
0.8
0.7
Cumulative Distribution
0.6
0.5
0.4
0.3
0.2
0.1
0
−100 −50 0 50 100 150 200 250 300
Tranche P&L (%)
Figure 11: Kernel smoothed conditional distribution on 3/27/2006 of the normalized dis-
counted profit and loss (V3τ (H, 5, v ∗ , c∗1 , c∗2 ) − 5)/5 for the residual equity tranche of a 5
year cash CDO referenced on the CDX.HY6, whose maturity date H is 6/27/2011. We
apply the model and fitting methodology developed in Sections 2 and 3, based on I = 10K
re-sampling scenarios for N , and 100K replications.
the fast scheme. Figure 11 shows the distribution of the normalized discounted profit and
loss (V3τ (H, 5, v ∗ , c∗1 , c∗2 ) − 5)/5 for an equity tranche position. For the equity investor, the
probability of more than doubling the principal p3 after discounting is roughly 75% under
the fast scheme, while it is only 60% under the uniform scheme. Figure 12 shows that
for the debt tranches, the upside is much more limited, for any scheme. For example, the
normalized discounted profit on the senior tranche can be at most 0.5% under the fast
scheme and about 1% under the uniform scheme.
A standard measure to quantify the risk of a position is value at risk, a quantile of
the position’s loss distribution. We estimate the value at risk at time t ≤ H of a position
in a cash CDO tranche maturing at H, given by
for some level of confidence α ∈ (0, 1). Figure 13 shows the 99.5% value at risk for the 5
year senior tranche with maturity date H given by 6/27/2011, for analysis times varying
weekly between 3/27/2006 and 11/7/2008, for each of the two prioritization schemes.
Assuming risk capital is allocated according to the value at risk, a value in the time series
represents the amount of risk capital that is needed at that time to support the tranche
position over its remaining life. For both prioritization schemes, the time series behavior
reflects the rising trend in corporate defaults that started in 2008 and that is evidenced
in Figure 1. The fast scheme requires less risk capital than the uniform scheme but leads
26
1 1
Uniform Scheme Uniform Scheme
0.9 Fast Scheme 0.9 Fast Scheme
0.8 0.8
0.7 0.7
Cumulative Distribution
Cumulative Distribution
0.6 0.6
0.5 0.5
0.4 0.4
0.3 0.3
0.2 0.2
0.1 0.1
0 0
−12 −10 −8 −6 −4 −2 0 2 −100 −50 0 50
Tranche P&L (%) Tranche P&L (%)
Figure 12: Kernel smoothed conditional distribution on 3/27/2006 of the normalized dis-
counted profit and loss (Vjτ (H, pj , v ∗ , c∗1 , c∗2 ) − pj )/pj for the debt tranches of a 5 year
cash CDO referenced on the CDX.HY6, whose maturity date H is 6/27/2011. We apply
the model and fitting methodology developed in Sections 2 and 3, based on I = 10K re-
sampling scenarios for N , and 100K replications. Left panel : Senior tranche with principal
p1 = 85. Right panel : Junior tranche with principal p2 = 10.
6 Conclusion
This paper develops, implements and validates stochastic methods to measure the risk of
investment positions in collateralized debt obligations and related credit derivatives tied
to an underlying portfolio of defaultable assets. The ongoing financial crisis highlights the
need for sophisticated yet practical tools allowing potential and existing investors as well
as regulators to quantify the exposure associated with such positions, and to accurately
estimate the amount of risk capital required to support a position.
The key to address the risk analysis problem is a model of default timing that cap-
tures the default clustering in the underlying portfolio, and a method to estimate the
model parameters based on historical default experience. This paper contributes to each
of these two sub-problems. It formulates an adaptive, intensity-based point process model
of default timing that performs well according to in- and out-of-sample tests. Moreover,
it develops a maximum likelihood approach to estimating point process models. This ap-
proach is based on an acceptance/rejection re-sampling scheme that generates alternative
portfolio default histories from the available economy-wide default process.
The point process model and inference method have potential applications in other
areas dealing with correlated event arrivals, within and beyond financial engineering.
27
Uniform (Senior) Fast (Senior)
23 16
22 15
21 14
20
99.5% Value−at−Risk (%)
14 8
13 7
12 6
Mar−06 Aug−06 Jan−07 Jun−07 Nov−07 Apr−08 Sep−08 Mar−06 Aug−06 Jan−07 Jun−07 Nov−07 Apr−08 Sep−08
Figure 13: 99.5% value at risk VaR1τ (0.995, H, 85, v ∗ , c∗1 , c∗2 ) for the 5 year senior tranche
of a cash CDO referenced on the CDX.HY6 with maturity date 6/27/2011, for τ varying
weekly between 3/27/2006 and 11/7/2008, based on I = 10K re-sampling scenarios for
N , and 100K replications. Left panel : Uniform prioritization scheme. Right panel : Fast
prioritization scheme.
Financial engineering examples include the pricing and hedging of securities exposed to
correlated default risk, and order book modeling. Other example areas are insurance,
healthcare, queuing, and reliability.
A Proofs
Proof of Proposition 2.1. Let (Un∗ ) be a sequence of standard uniform random variables
that are independent of one another and independent of the (Tn∗ ). Let π(du, dt) be the
random counting measure with mark space [0, 1] associated with the marked point process
RtR1
(Tn∗ , Un∗ ). Note that Nt∗ = 0 0 π(du, ds), and that π(du, dt) has intensity λπ (u, t) = λ∗t−
for u ∈ [0, 1], which we choose to be predictable. This means that the process defined by
RtR1
0 0
(π(du, ds) − λπ (u, s)duds) is a local martingale. Now for u ∈ [0, 1] define the process
I(u, ·) by I(u, t) = 1{u≤Zt∗ } . Then the counting process N generated by the selected event
times can be written as
X Z tZ 1
∗ ∗
Nt = I(Un , Tn )1{Tn∗ ≤t} = I(u, s)π(du, ds).
n≥1 0 0
Since Z ∗ is predictable, I(u, ·) is predictable for u ∈ [0, 1]. The process I(u, ·) is also
bounded for u ∈ [0, 1]. Therefore, a local martingale M is defined by
Z tZ 1
Mt = Nt − I(u, s)λπ (u, s)du ds.
0 0
28
But the definitions of I(u, s) and λπ (u, s) yield
Z t Z Zs∗ Z t
∗
Mt = Nt − λs− du ds = Nt − Zs∗ λ∗s ds,
0 0 0
The time change theorem of Meyer (1971) implies the result, since A∗ is continuous and
increases to ∞, almost surely.
Rt
Proof of Proposition 3.1. It suffices to show that 0 λs ds < ∞ almost surely for each
t > 0. We assume that γ = 0, without loss of generality. For s ≥ 0 and h(0) > 0, let
The function h describes the behavior of the intensity (10) between events. For Tn ≤ t <
Tn+1 and h(0) = λTn , we have that λt = h(t − Tn ). The inverse h−1 to h is given by
−1 1 h(0)(1 − c)
h (u) = log
κh(0) u − ch(0)
29
where M is strictly less than 1 and independent of n = 0, 1, 2, . . . It follows that the
unconditional probability P (λTn+1 > λTn ) = M , for any n. Further, an induction argument
can be used to show that for any n and k = 1, 2, . . .
P (λTn+k > λTn+k−1 > · · · > λTn ) = M k .
Now letting
Ctk = {ω : Nt (ω) ≥ k and λTn+1 (ω) > λTn (ω) for n = Nt (ω) − 1, . . . , Nt (ω) − k}
for t > 0 and k = 1, 2, . . ., we have that
P (Ctk ) = P (Nt ≥ k and λTNt > λTNt −1 > · · · > λTNt −k )
= Mk
independently of t. We then conclude that
Z t
P λs (ω)ds < ∞ ≥ P sup λs (ω) < ∞
0 s∈[0,t]
∞
\ c
≥P Ctk (17)
k=1
∞
\
=1−P Ctk
k=1
≥ 1 − lim M k
k→∞
= 1.
Equation (18) is due to the fact that Ctk ∈ Ft for all k = 1, 2, . . . and Ct1 ⊇ Ct2 ⊇ · · · . To
justify the inequality (17), we argue as follows. For given t > 0 and ω ∈ Ω, define
At (ω) = k ∈ {0, 1, . . . , Nt (ω) − 1} : λTk+1 (ω) ≥ λTk (ω) .
Let ω̂ ∈ { ∞ k c k
T
k=1 Ct } . Then, by definition of Ct , either (i) Nt (ω̂) < ∞, or (ii) there exists
some n < ∞ such that λTn+1 (ω̂) < λTn (ω̂) and |At (ω̂) \ ATn (ω̂)| < ∞.
In case (i), thanks to the condition |At (ω̂)| ≤ Nt (ω̂) < ∞, each λTk+1 (ω̂)−λTk (ω̂) < ∞
for all k ∈ At (ω̂). Thus, λs (ω̂) < ∞ for all s ∈ [0, t]. Hence, we conclude that ω̂ ∈ {ω :
sups∈[0,t] λs (ω) < ∞} in this case.
Now consider case (ii). Note that for any s ∈ [0, t], the condition cλTNs (ω̂) < λs (ω̂)
implies that λs (ω̂) remains at infinity once λTNs (ω̂) achieves infinity. Hence, λTn+1 (ω̂) <
λTn (ω̂) implies that λTn (ω̂) < ∞, and we conclude that sups∈[0,Tn ] λs (ω̂) < ∞. Moreover,
due to the condition |At (ω̂) \ ATn (ω̂)| < ∞, we conclude that sups∈[Tn ,t] λs (ω̂) < ∞.
Therefore, { ∞ k c
T
k=1 Ct } ⊆ {ω : sups∈[0,t] λs (ω) < ∞} and (17) is justified.
30
B Risk-neutral tranche loss distributions
To describe the estimation of the risk-neutral tranche loss distributions discussed in Sec-
tion 4, it is necessary to explain the arbitrage-free valuation of index and tranche swaps.
These contracts are based on a portfolio of C credit swaps with common notional 1,
common maturity date H and common quarterly premium payment dates (tm ).
In an index swap, the protection seller covers portfolio losses as they occur (default
leg), and the protection buyer pays Sam (C − Nt0m ) at each premium date tm , where S is
the swap rate and am is the day count fraction for coupon period m (premium leg). The
value at time t ≤ H of the default leg is given by Dt (H, 0, 1), where
Z H Z s
Dt (H, K, K) = EQ exp − ru du dUs (K, K) Ft . (19)
t t
The index rate at t is the solution S = St (H) to the equation Dt (H, 0, 1) = Pt (S).
In a tranche swap with upfront rate G and running rate S, the protection seller covers
tranche losses (14) as they occur (default leg) and, for K < 1, the protection buyer pays
GKC at inception and Sam (KC − Utm ) at each premium date tm , where K = K − K is
the tranche width (premium leg). The value at τ ≤ H of the default leg is given by (19).
The value at time t ≤ H of the premium leg is given by
X Z tm
Pt (K, K, G, S) = GKC + S am EQ exp − ru du (KC − Utm ) Ft . (21)
tm ≥t t
For a fixed upfront rate G, the running rate S is the solution S = St (H, K, K, G) to the
equation Dt (H, K, K) = Pt (K, K, G, S). For a fixed rate S, the upfront rate G is the
solution G = Gt (H, K, K, S) to the equation Dt (H, K, K) = Pt (H, K, K, G, S).
The valuation relations are used to estimate the risk-neutral portfolio and tranche
loss distributions from market rates of index and tranche swaps. First we formulate a
parametric model for the risk-neutral dynamics of N and L, the portfolio default and loss
processes with replacement. Our risk-neutral model parallels the model under P . Suppose
that N has risk-neutral intensity λQ with Q-dynamics
dλQ Q Q Q
t = κt (ct − λt )dt + dJt (22)
where λQ Q Q Q Q Q Q
0 > 0, κt = κ λTNt is the decay rate, ct = c λTNt is the reversion level, and J is
a response jump process given by
X
Jt = max(γ Q , δ Q λQ )1
T − {Tn ≤t}
. (23)
n
n≥1
31
The quantities κQ > 0, cQ ∈ (0, 1), δ Q > 0 and γ Q ≥ 0 are parameters such that
cQ (1 + δ Q ) < 1. We let θQ = (κQ , cQ , δ Q , γ Q , λQ
0 ).
Q
Since the Q-dynamics of λ mirror the P -dynamics of the P -intensity λ, we can apply
the algorithms developed above to estimate the expectations (19)–(21) and to calculate
the model index and tranche rates for the model (22). We first generate event times Tn of
N by Algorithm 3, with λ and its parameters replaced by their risk-neutral counterparts,
and with initial condition (Nτ , TNτ , λQ Q
TNτ ) = (0, 0, λτ ). Then we apply the replacement
Algorithm 2 as stated to generate event times Tn0 without replacement, and the corre-
sponding paths of N 0 and L0 required to estimate the expectations (19)–(21). In this last
step, we implicitly assume that the thinning probability (7), the rating distributions (8)
and (9), and the distribution µ` of the loss at an event are not adjusted when the measure
is changed from P to Q. The risk-free interest rate r is assumed to be deterministic, and
is estimated from Treasury yields for multiple maturities on 3/27/2006, obtained from
the website of the Department of Treasury.
The risk-neutral parameter vector θQ is estimated from a set of market index and
tranche rates by solving the nonlinear optimization problem
X MarketMid(i) − Model(i, θQ ) 2
min (24)
θQ ∈ΘQ
i
MarketAsk(i) − MarketBid(i)
where ΘQ = (0, 2)×(0, 1)×(0, 2)2 ×(0, 20) and the sum ranges over the data points. Here,
MarketMid is the arithmetic average of the observed MarketAsk and MarketBid quotes.
We address the problem (24) by adapted simulated annealing. The algorithm is initialized
at a set of random parameter values, which are drawn from a uniform distribution on the
parameter space ΘQ . For each of 100 randomly chosen initial parameter sets, the algorithm
converges to the optimal parameter values given in Table 5. The market data and fitting
results for 5 year index and tranche swaps referenced on the CDX.HY6 on 3/27/2006 are
reported in Table 4. The model fits the data, with an average absolute percentage error
of 2.9%.
32
Contract MarketBid MarketAsk MarketMid Model
Index 332.88 333.13 333.01 333.08
0-10% 84.50% 85.00% 84.75% 86.75%
10-15% 52.75% 53.75% 53.25% 48.19%
15-25% 393.00 403.00 398.00 398.88
25-35% 70.00 85.00 77.50 79.43
MinObj 33.77
AAPE 2.90%
Table 4: Market data from Morgan Stanley and fitting results for 5 year index and tranche
swaps referenced on the CDX.HY6 on 3/27/2006. The index, (15 − 25%) and (25 − 35%)
contracts are quoted in terms of a running rate S stated in basis points (10−4 ). For these
contracts the upfront rate G is zero. The (0, 10%) and (10, 15%) tranches are quoted in
terms of an upfront rate G. For these contracts the running rate S is zero. The values in
the column Model are fitted rates based on model (22) and 100K replications. We report
the minimum value of the objective function MinObj and the average absolute percentage
error AAPE relative to market mid quotes.
Unpaid reductions in principal from default losses, Jj (m), occur in reverse priority order,
so that the residual equity tranche suffers the reduction
where
ξ(m) = max{0, L0tm − L0tm−1 − [W (m) − Q1 (m) − Q2 (m)]}
is the cumulative loss since the previous coupon date minus the collected and undistributed
interest income. Then, the debt tranches are reduced in principal by
With uniform prioritization there are no early payments of principal, so P1 (m) = P2 (m) =
0 for m < M . At maturity, principal and accrued interest are treated identically, while
the remaining reserve is paid in priority order. The payments of principal at maturity are
where R(M ) is the value of the reserve account at M . The residual equity tranche receives
D3 (M ) = R(M ) − Q1 (M ) − P1 (M ) − Q2 (M ) − P2 (M ).
33
Parameter κQ cQ δQ γQ λQτ
Estimate 0.522 0.333 0.366 1.646 8.586
As long as the senior tranche receives payments the junior tranche accrues coupons. After
the senior tranche has been retired, the junior tranche is allocated interest and principal
until maturity or until its principal is written down, whichever is first:
The equity tranche receives any residual cash flows. There are no contractual reductions
in principal.
where the notation B(m) = B(m, v) indicates the dependence of the cash flow B(m) on
the coupon rate v of a reference bond. The par coupon rate of a reference bond is the
number v = v ∗ such that Ot0 (v) = C at the CDO inception date t0 .
The par coupon rates for the debt tranches are determined similarly. Let Ojt (v ∗ , c1 , c2 )
be the value at time t ≤ H of the bond representing tranche j = 1, 2 when the reference
bonds accrue interest at their par coupon rate v ∗ , and when the coupon rates on the debt
tranches are c1 and c2 , respectively. Because of the complexity of the cash flow “waterfall,”
34
this quantity does not admit a simple analytic expression. The par coupon rates are given
by the pair (c1 , c2 ) = (c∗1 , c∗2 ) such that Ojt0 (v ∗ , c1 , c2 ) = pj for j = 1, 2.
For the CDX.HY6 of reference bonds, assuming that the cash CDO is established on
3/27/06 and has a 5 year maturity, the par coupon rates are estimated as follows. We first
generate a collection of 100K paths of N 0 and L0 under the risk neutral measure, based on
the risk-neutral intensity model (22) with calibrated parameters in Table 5, as described
in Appendix B. Based on these paths, we can estimate Ot0 (v) for fixed v, and then solve
numerically for the par coupon rate v ∗ . The risk-free interest rate r is deterministic, and
is estimated from Treasury yields for multiple maturities on 3/27/2006. Given v ∗ , c1 , c2 ,
we can calculate the tranche cash flows for a given path of (N 0 , L0 ) according to the spec-
ified prioritization scheme, and then estimate Ojt0 (v ∗ , c1 , c2 ). We then solve numerically
a system of two equations for (c∗1 , c∗2 ).
References
Arnsdorf, Matthias & Igor Halperin (2008), ‘BSLP: markovian bivariate spread-loss model
for portfolio credit derivatives’, Journal of Computational Finance 12, 77–100.
Azizpour, Shahriar & Kay Giesecke (2008a), Premia for correlated default risk. Working
Paper, Stanford University.
Azizpour, Shahriar & Kay Giesecke (2008b), Self-exciting corporate defaults: Contagion
vs. frailty. Working Paper, Stanford University.
Box, George & George Tiao (1992), Bayesian Inference in Statistical Analysis, Wiley, New
York.
Brigo, Damiano, Andrea Pallavicini & Roberto Torresetti (2007), ‘Calibration of CDO
tranches with the dynamical generalized-Poisson loss model’, Risk 20(5), 70–75.
Chen, Zhiyong & Paul Glasserman (2008), ‘Fast pricing of basket default swaps’, Opera-
tions Research 56(2), 286–303.
Collin-Dufresne, Pierre, Robert Goldstein & Jean Helwege (2009), How large can jump-
to-default risk premia be? Modeling contagion via the updating of beliefs. Working
Paper, Carnegie Mellon University.
Cont, Rama & Andreea Minca (2008), Recovering portfolio default intensities implied by
CDO quotes. Working Paper, Columbia University.
Das, Sanjiv, Darrell Duffie, Nikunj Kapadia & Leandro Saita (2007), ‘Common failings:
How corporate defaults are correlated’, Journal of Finance 62, 93–117.
35
Dellacherie, Claude & Paul-André Meyer (1982), Probabilities and Potential, North Hol-
land, Amsterdam.
Delloye, Martin, Jean-David Fermanian & Mohammed Sbai (2006), ‘Dynamic frailties and
credit portfolio modeling’, Risk 19(1), 101–109.
Ding, Xiaowei, Kay Giesecke & Pascal Tomecek (2009), ‘Time-changed birth processes
and multi-name credit derivatives’, Operations Research 57(4), 990–1005.
Duffie, Darrell, Andreas Eckner, Guillaume Horel & Leandro Saita (2009), ‘Frailty corre-
lated default’, Journal of Finance 64, 2089–2123.
Duffie, Darrell & Nicolae Garleanu (2001), ‘Risk and valuation of collateralized debt
obligations’, Financial Analysts Journal 57(1), 41–59.
Eckner, Andreas (2009), ‘Computational techniques for basic affine models of portfolio
credit risk’, Journal of Computational Finance 15, 63–97.
Errais, Eymen, Kay Giesecke & Lisa Goldberg (2009), Affine point processes and portfolio
credit risk. Working Paper, Stanford University.
Giesecke, Kay, Lisa Goldberg & Xiaowei Ding (2009), A top-down approach to multi-name
credit. Operations Research, forthcoming.
Glasserman, Paul & Nicolas Merener (2003), ‘Numerical solution of jump-diffusion LIBOR
market models’, Finance and Stochastics 7, 1–27.
Hamilton, David (2005), Moodys senior ratings algorithm and estimated senior ratings.
Moodys Investors Service.
Hawkes, Alan G. (1971), ‘Spectra of some self-exciting and mutually exciting point pro-
cesses’, Biometrika 58(1), 83–90.
Jorion, Philippe & Gaiyan Zhang (2007), ‘Good and bad credit contagion: Evidence from
credit default swaps’, Journal of Financial Economics 84(3), 860–883.
Kou, Steven & Xianhua Peng (2009), Default clustering and valuation of collateralized
debt obligations. Working Paper, Columbia University.
Lando, David & Mads Stenbo Nielsen (2009), Correlation in corporate defaults: Contagion
or conditional independence? Working Paper, Copenhagen Business School.
Longstaff, Francis & Arvind Rajan (2008), ‘An empirical analysis of collateralized debt
obligations’, Journal of Finance 63(2), 529–563.
36
Lopatin, Andrei & Timur Misirpashaev (2008), ‘Two-dimensional Markovian model for
dynamics of aggregate credit loss’, Advances in Econometrics 22, 243–274.
Ogata, Yosihiko (1981), ‘On Lewis’ simulation method for point processes’, IEEE Trans-
actions on Information Theory 27, 23–31.
Papageorgiou, Evan & Ronnie Sircar (2007), ‘Multiscale intensity models and name group-
ing for valuation of multi-name credit derivatives’, Applied Mathematical Finance
15(1), 73–105.
Prahl, Jürgen (1999), A fast unbinned test of event clustering in poisson processes. Work-
ing Paper, Universität Hamburg.
SEC (2008), Summary report of issues identified in the commission staff’s examinations
of select credit rating agencies. Report, US Securities and Exchange Commission.
37