Você está na página 1de 140

Stochastic volatility models and hybrid

derivatives
Claudio Albanese
Department of Mathematics / Imperial College London

Presented at Bloomberg and at the Courant Institute, New
York University
New York, September 22nd, 2005
Co-authors.
Oliver Chen
(National University of Singapore)
Antonio Dalessandro
(Imperial College London)
Manlio Trovato
(Merrill Lynch London)
available at:
www.imperial.ac.uk/mathn
1
Contents.
Part I. A stochastic volatility term structure model.
Part II. Credit barrier models with functional lattices.
Part III. Estimations under P and under Q.
Part IV. Pricing credit-equity hybrids.
PART V. Credit correlation modeling and synthetic CDOs.
Part VI. Pricing credit-interest rate hybrids.
2
PART I. A stochastic volatility term structure model
It is widely recognized that xed income exotics should be priced by
means of a stochastic volatility model. Callable constant maturity
swaps (CMS) are a particularly interesting case due to the sensitivity
of swap rates to implied swaption volatilities for very deep out of the
money strikes. In this paper we introduce a stochastic volatility term
structure model based on a continuous time lattice which allows for
a numerically stable and quite ecient methodology to price xed
income exotics and evaluate hedge ratios.
3
Introduction
The history of interest rate models is characterized by a long series of
turns. The Black formula for caplets and swaptions was designed to
take as underlying a single forward rate under the appropriate forward
measure, see (Joshi & Rebonato 2003). This has the advantage to
lead to a simple pricing formula for European options but also the
limitation of not being extendable to callable contracts. To have a
more consistent model, short rate models where introduced in (Cox,
Ingersoll & Ross 1985), (Vasicek 1977), (Black & Karasinski 1991)
and (Hull & White 1993). These models are distinguished by the exact
specication of the spot rate dynamics through time, in particular the
form of the diusion process, and hence the underlying distribution of
the spot rate.
4
LMM models
Next came LIBOR market models, also known as correlation models.
First introduced in (Brace, Gatarek & Musiela 1996) and (Jamshidian
1997), forward LIBOR models armed themselves as a mainstream
methodology and are now discussed in textbooks such as for instance
(Brigo & Mercurio 2001). Various extensions of forward LIBOR mod-
els that attempt to incorporate volatility smiles of interest rates have
been proposed. Local volatility type extensions were pioneered in
(Andersen & Andreasen 2000). A stochastic volatility extension is
proposed in (Andersen & Brotherton-Ratclie 2001), and is further
extended in (Andersen & Andreasen 2002). A dierent approach to
stochastic volatility forward LIBOR models is described in (Joshi &
Rebonato 2003). Jump-diusion forward LIBOR models are treated
in (Glasserman & Merener 2001), (Glasserman & Kou 1999). A cali-
bration framework is proposed in (Piterbarg 2003).
5
Stochastic volatility models
Modeling stochastic volatility within LIBOR market models is a chal-
lenging task from an implementation viewpoint. In fact, Monte Carlo
methods tend to be slow and inecient in the presence of a large num-
ber of factors. In a strive to achieve a more reliable pricing framework,
in recent years, we witnessed a move away from correlation models
and the emergence and recognition as market standard of the SABR
model by (S.Hagan, Kumar, S.Lesniewski & E.Woodward 2002) in the
xed income domain.
6
SABR
SABR however is unlikely to be the denitive solution and is perhaps
rather just yet another stepping stone in a long chain. In fact, like-
wise to the Black formula approach, SABR takes up a single forward
rate under the corresponding forward measure as underlier. As a con-
sequence, within this framework it is not possible to price callable
swaps and Bermuda swaptions. There are also calibration inconsis-
tencies. Implied swaption volatilities with very large strikes are probed
by constant maturity swaps (CMS), structures which receive xed, or
LIBOR plus a spread, and pay the equilibrium swap rate of a given
maturity. The asymptotic behavior of implied volatilities for very large
strikes turns out to atten out to a constant, as opposed to diverging
rapidly as SABR would predict. Finally, some pricing inconsistencies
may emerge with SABR due to the fact that the model is solved by
means of asymptotic expansions with a nite, irreducible error.
7
Stochastic volatility term structure models
In this article we attempt to go beyond SABR by introducing a stochas-
tic volatility short rate model which has the correct asymptotic behav-
ior for implied swaption volatilities and can be used for callable swaps
and Bermuda swaptions. Our model is solved eciently by means of
numerical linear algebra routines and is based on continuous time lat-
tices of a new type. No calculation requires the use of Montecarlo or
asymptotic methods and prices and hedge ratios are very stable, even
for extreme values of moneyness.
8
Nearly stationary calibration
Our model is made consistent with the term structure of interest rates
and the term structure of implied at-the-money volatilities by means
of a calibration procedure that greatly reduces the degree of time
dependency of coecients. As a consequence, the model is nearly
stationary over time horizons in excess of 30 years.
9
Applications
In this presentation, I discuss the model by reviewing in detail an
implementation example. While leaving it up to the interested reader
to pursue the many conceivable variations and extensions, we describe
the salient features of our modeling by focusing in detail to the problem
of pricing and nding hedge ratios for Bermuda swaptions and callable
CMS swaps.
10
The model
Our model is built upon a specication of a short rate process r
t
which combines local volatility, stochastic volatility and jumps. We
make our best eort to calibrate the model to a stationary process
and introduce the least possible degree of explicit time dependence in
such a way to rene ts of the term structure of interest rates and of
selected at-the-money swaption volatilites. The model is specied in
a largely non-parametric fashion within a functional analysis formalism
and expressed in terms of continuous time lattice models.
A sequence of several steps is required to specify the short rate process
r
t
.
11
The conditional local volatility processes
We introduce M states of volatility. The process conditioned to stay
in one of such states {1, ...M} is related to the solution r
t
of the
following equation:
dr
t
=

r
t
)dt +

t
dW. (1)
12
Short rate volatilities for each of the four volatility states
13
The functional analysis formalism
In the functional analysis formalism we use, these SDEs are associated
to the Markov generators
L
r

r
t
)

r
+

2

r
2

t
2

2
r
2
. (2)
14
Functional lattices
To build a continuous time lattice (also called functional lattice), we
discretize the short rate variable and constrain it to belong to a nite
lattice containing N +1 points r(x) 0, where x = 0, ...N, r(0) = 0
and the following ones are in increasing order. The discretized Markov
generator L
r

is dened as the operator represented by a tridiagonal


matrix such that

y
L
r

(x, y) = 0

y
L
r

(x, y)(y x) =

r(x))

y
L
r

(x, y)(y x)
2
=
2

r(x)
2

.
15
Model parameters
In our example, we select an inhomogeneous grid of N = 70 points
spanning short rates from 0% to 50%. We also choose to work with
M = 4 states for volatility and make the following parameter choices:

0 31% 30% 2.10% .17


1 46% 40% 5.50% .18
2 75% 50% 8.50% .23
3 100% 60% 12.00% .24
16
How to solve in the special case of a local volatility model (M=1)
and without jumps
Although our model is more complex than a simple local volatility
process, it is convenient to describe our resolution method in the
specic case of the operator L
r

with constant . This method can


then be generalized and is at the basis of other extensions such as the
introduction of jumps (see below).
17
Spectral analysis
We start by considering the following pair of eigenvalue problems:
L
r

u
n
=
n
u
n
L
rT

v
n
=
n
v
n
where the superscript
T
denotes matrix transposition, u
n
and v
n
are
the right and left eigenvectors of L
r

, respectively, whereas
n
are the
corresponding eigenvalues. Except for the simplest cases, the Markov
generator L
r

is not a symmetric matrix, hence u


n
and v
n
are dierent.
18
Spectral analysis
Also, in general the eigenvalues are not real. We are only guaranteed
that their real part is non-positive Re
n
0 and that complex eigen-
values occur in complex conjugate pairs, in the sense that if
n
is an
eigenvalue then also

n
is an eigenvalue. We set boundary conditions
in such a way that there is absorption at the endpoints, and in par-
ticular at the point corresponding to zero rates r(0) = 0. With this
choice, we are also guaranteed that there exists a zero eigenvalue.
19
Spectral analysis
There is no guarantee, in the most general case, that there exists
a complete set of eigenfunctions. However, the chance that such a
complete set does not exist for a Markov generator specied non-
parametrically is zero, so we can safely assume that this is the case.
In the unlikely case that this assumption is not correct, the numerical
linear algebra routines needed to solve our lattice model will identify
the problem and a small perturbation of the given operator will suce
to rectify the situation. Assuming completeness, the diagonalization
problem can be rewritten in the following matrix form:
L
r

= UU
1
(3)
where U is the matrix having as columns the right eigenvectors and
is the diagonal matrix having the eigenvalues
i
as elements.
20
Functional calculus
Key to our constructions is the remark that, if the Markov generator
is diagonalisable, we can apply an arbitrary function F to it by means
of the following formula:
F(L
r

) = UF(
r

)U
1
(4)
This formula is at the basis of the so-called functional calculus.
As Itos formula regarding functions of stochastic processes is central
in the mathematical Finance for diusion processes, functional cal-
culus for Markov generators plays a pivotal role in our framework for
stochastic volatility models.
21
Functional calculus
This formula has several applications. An immediate one allows us to
express the pricing kernel u(r(x), t; r(y), T) of the process as follows:
u(r(x), t; r(y), T) = (e
(Tt)L
r

)(x, y) =

n
e

n
(Tt)
u
n
(x)v
n
(y). (5)
22
Introducing jumps
At this stage of the construction one has the option to also add jumps.
Although in the example discussed in this paper we are mostly focused
on long dated callable swaps and swaptions for which we nd that the
impact of jumps can be safely ignored, adding jumps involves negligible
additional complexities and is thus worth considering and implement-
ing in other situations. To add jumps, one can follow the following
procedure which accounts for the need to assign dierent intensities
to up-jumps and down-jumps. Jump processes are associated to a
special class of stochastic time changes given by monotonously non-
decreasing processes T
t
with independent increments.
23
Stochastic time changes
The time changes characterizing Levy processes with symmetric jumps
are known as Bochner subordinators and are characterized by a Bochner
function () such that
E
0
_
e
T
t
_
= e
()t
(6)
24
The variance-gamma model
For example, the case of the variance gamma process which received
much attention in Finance corresponds to the function
() =

2

log
_
1 +

_
(7)
where is the mean rate and is the variance rate of the variance
gamma process.
25
The generator of the jump process
The generator of the jump process can be expressed using functional
calculus as the operator (L
r

). To produce asymmetric jumps,


we specify the two parameters dierently for the up and down jumps
and compute separately two Markov generators
L

(L
r

) = U

()V

(8)
where:

() =

2

log(1 +

) (9)
26
The generator of the process with asymmetric jumps
The new generator for our process with asymmetric jumps is obtained
by combining the two generators above
L
r
j
=
_
_
_
_
_
_
_
_
0 0
L

(2, 1) d(2, 2) L
+
(2, 3) L
+
(2, n)
.
.
.
.
.
.
.
.
.
.
.
.
L

(n 1, 1) L

(n 1, 2) d(n 1, n 1) L
+
(n 1, n)
0 0 0
_
_
_
_
_
_
_
_
27
Probability conservation and boundary conditions
Here the element of the diagonal are chosen in such a way to satisfy
probability conservation:
d(x, x) =

y=x
L
r
j
(x, y) (10)
Also notice that we have zeroed out the elements in the matrix at the
upper and lower boundary: this ensures that there is no probability
leakage in the process.
28
Drift condition
In our setting, we choose a short rate as a modeling primitive and
we thus do not need to impose a martingale condition. Otherwise,
were we working with a forward rate instead, the appropriate method
of restoring the martingale condition would be to modify the matrix
elements of the resulting generator on the rst sub-diagonal and the
rst super-diagonal.
29
The Local Levy generator
At this stage of the construction, we have therefore obtained a gener-
ator L
r
j
for the short rate process, whose dynamics is characterized
by a combination of state dependent local volatility and asymmetric
jumps. We note that the addition of jumps has not increased the di-
mensionality of the problem and is therefore computationally ecient.
30
Modeling the dynamics of stochastic volatility
As a third step, we dene a dynamics for stochastic volatility by as-
signing a Markov generator to the volatility state variable which
depends on the rate coordinate x. Namely, conditioned to the rate
variable being x, the generator has the following form
L
sv
x
= (x)L
sv
+
+L
sv

(11)
where
L
sv
+
=
_
_
_
_
_
0.7 0.7 0 0
0 1.1 0.8 0.3
0 0 1.5 1.5
0 0 0 0
_
_
_
_
_
, L
sv

=
_
_
_
_
_
0 0 0 0
1.4 1.4 0 0
0 3 3 0
0 0 5 5
_
_
_
_
_
.
(12)
31
Excitability function (x)
32
The total generator
Out of the two generators we just dened, we form a Markov generator
L acting on functions of both the rate variable x and the volatility
variable . This generator has matrix elements given as follows:
L(x, ; y, ) = L
r
j
(x, y)
,
+L
sv
x
(, )
xy
. (13)
33
Numerical analysis
In our working example, the matrix L has total dimension MN = 280.
For matrices of this size, diagonalization routines such as dgeev in
LAPACK are very ecient. Since our underlier is a short rate though,
we are not interested in the pricing kernel but rather in the discounted
transition probabilities given by
p(x, t; y, T) = E
_
e

_
T
t
r
s
ds
, |r
t
= r(x), r
T
= r(y)
_
. (14)
34
Numerical analysis
This kernel satises the following backward equation

t
p(x, t; y, T) +(Lp)(x, t; y, T) = r(x)p(x, t; y, T). (15)
In functional calculus notations, the solution is given by
p(x, t; y, T) = e
G(Tt)
(x, y) where G(x, y) L(x, y) r(x)
xy
.
(16)
35
Numerical analysis with stochastic volatility
The same diagonalization method illustrated above for the local volatil-
ity case applies also in this situation. By representing the matrix G in
the form
G = UU
1
(17)
where is diagonal, and writing the matrix of discounted transition
probabilities as follows
e
G(Tt)
= Ue
(Tt)
U
1
. (18)
36
Calibration and Pricing
In our example, to calibrate our model we aim at matching forward
swap rates and at-the-money swaption volatilities, both referring to
swaps of 5 year tenor. We start from the following data
1y 2y 3y 4y 5y 7y 10y 15y 20y 30y
forward 2.999% 3.311% 3.587% 3.800% 3.984% 4.226% 4.393% 4.477% 4.301% 4.114%
ATM vol 21.506% 19.443% 17.962% 16.967% 16.189% 14.897% 13.801% 12.460% 12.665% 11.728%
37
Nearly stationary calibration
The calibration procedure has two steps. In a rst step we search for
a best t using the model above without introducing any explicit time
dependency. In a second step, we then introduce time dependency to
achieve a perfect t. As a consequence of this procedure, the degree
of time variability of model parameters is kept to a bare minimum.
To introduce time dependence we combine two operations: a shift of
the short rate by a time varying, deterministic function of time and a
deterministic time change, i.e.
r
t
r
t
=

b(t)r
b(t)
+a(t). (19)
38
Nearly stationary calibration
Here b(t) is monotonously increasing and

b(t) denotes its time deriva-
tive. Using the new process, discounted transition probabilities can be
computed as follows:
E
_
e

_
T
t
r
s
ds
, | r
t
= r(x), r
T
= r(y)
_
= e

_
T
t
a(s)ds
G(x, b(t); y, b(T)) (20)
where G is the kernel for the stationary process dened above.
Our choice in the working example is b(t) = 1.095t + 0.008t
2
. The
function a(t) is then dened in such a way to match the term structure
of forward swap rates. This adjustment is given in the next slide.
39
Deterministic yield adjustment (EUR)
40
Degree of time dependence
As one can see from this picture, the yield adjustment is less than 20
basis points in absolute value. This ensures that the probability of the
modied short rate process r
t
to attain negative values is small. In a
typical implementation of the Hull-White model along similar lines, the
short rate adjustment is typically of a few percent. The discrepancy
is linked to the fact that the richer stochastic volatility model we
construct is capable of explaining most of the salient features of the
zero curve even with the constraint that the process be stationary.
41
Advantages of nearly stationary model calibration
The advantage of having a nearly stationary model is that the shapes
of yield curves that one obtains depend on the short rate and the
volatility state but are largely independent of time. The gure in
the next slide shows the yield curves corresponding to dierent initial
volatility states and dierent starting values for the short rate. As
the graphs indicate, yield curves are sensitive to the initial volatility
state as they raise if initial volatilities raise. Moreover graphs show
that curves invert for high values of the short rate. In our model, this
behavior is consistent over all time frames except for corrections of
the order of 10 basis points.
42
Yield curves for dierent values of the initial volatility state and
of the short rate
43
Pricing swaptions and callable constant maturity swaps
Implied volatilities for European swaptions are given in the next slide.
Here we graph extreme out of the money strikes of up to 15% for
swaptions of varying maturities where the deliverable is a 5Y swap.
One can notice that implied volatilities naturally atten out at long
maturities, a behavior consistent with what observed in the CMS mar-
ket where such extreme strike levels are probed.
44
Implied volatility for European swaptions (EUR)
45
Implied volatility for European swaptions (JPY)
46
Term structure of implied 5Y swaption volatilities (JPY)
47
Bermuda swaptions
Exercise boundaries for 10Y Bermuda swaptions are given in the next
slides. The rst graph refers to payer swaptions and the second to
receiver swaptions.
48
Exercise boundaries for payer Bermuda options
49
Exercise boundaries for receiver Bermuda options
The corresponding graphs for callable CMSs are given below. Notice
that the exercise boundaries depend on the volatility state.
50
Exercise boundaries for callable payer CMSs
51
Exercise boundaries for callable receiver CMSs
52
Sensitivities for Bermuda swaptions
Sensitivities for Bermuda swaptions are given in the next slides. These
sensitivities are computed by holding the volatility state variable xed
and are dened as the derivative of the price for a 10Y payer Bermuda
swaption with respect to the rate of the 10Y swap.
53
Delta of a 10Y Bermuda swaption, with semi-annual exercise
schedule, with respect to the 10Y swap rate. This is computed
while holding xed the volatility state variable.
54
Gamma of a 10Y Bermuda swaption, with semi-annual exercise
schedule, with respect to the 10Y swap rate. This is computed
while holding xed the volatility state variable.
55
Sensitivities for Constant Maturity Swaps
Sensitivities of callable constant maturity swaps are given in the next
slides. The delta and gamma are computed similarly to what done for
Bermuda swaptions, while the vega is calculated instead with respect
to the 10Y into 5Y European swaption.
56
Delta of a 10Y callable CMS swap, paying the 5Y swap rate
with semi-annual exercise schedule, with respect to the 15Y
swap rate . This is computed while holding xed the volatility
state variable.
57
Gamma of a 10Y callable CMS swap, paying the 5Y swap rate
with semi-annual exercise schedule, with respect to the 15Y
swap rate. This is computed while holding xed the volatility
state variable.
58
Vega of a 10Y callable CMS swap, paying the 5Y swap rate with
semi-annual exercise schedule, with respect to the 10Y into 5Y
European swaption price. This is computed while holding xed
the short rate.
59
Conclusions
We present a stochastic volatility term structure model, providing a
consistent framework for pricing European and Bermuda options, as
well as callable CMS swaps. The model is built upon a specication
of a short rate process, which combines local volatility, stochastic
volatility and jumps. The richness of the model allows to keep the
degree of time variability of model parameters to a bare minimum,
and obtain a nearly stationary behaviour. The solution methodology
is based on a new type of continuous time lattices, which allow for a
numerically stable and quite ecient technique to price xed income
exotics and evaluate hedge ratios.
60
PART II. Credit Barrier Models
Statistical data that we would like a credit model to t includes:
historical default probabilities given an initial credit rating
historical transition probabilities between credit ratings
interest rate spreads due to credit quality
61
Credit-rating based models
The early models of this class considered the credit-rating migration
and default process as a discrete, time-homogenous Markov chain
and took the historical transition probability matrix as the Markov
transition matrix.
Deciencies:
dicult to correlate
risk-neutralization leads to unintuitive results
62
Analytic closed form solutions versus numerical linear algebra
methods
The former framework for credit barrier models leveraged on solvable
models.
In the newer version recently developed we have a exible non-parametric
framework, whereby tractability comes from the use of numerical lin-
ear algebra as opposed to coming from the analytical tractability of
special functions.
63
The underling diusion process
The rst building block of our construction is a Markov chain process
x
t
on the lattice = {0, h, ..., hN} [0, 1] where N is a positive integer
and h = 1/N. In the case of a discretized diusion with state depen-
dent drift and volatility, the innitesimal generator L, of the process
x
t
is a tridiagonal matrix and can be expressed as follows in terms of
nite dierence operators:
L
x
= a(x)+[b(x) a(x)]
+
where x and
(f)(x) = f(x+1)+f(x1)2f(x), and (
+
f)(x) = f(x+1)f(x).
(21)
64
Continuous space limit
To ensure the existence of a continuous space limit, we derive the
functions a(x), b(x) from a drift function () and a volatility function
(), where [0, 1], which is identiable as the credit quality process
and can be expressed in terms of its innitesimal by imposing the
following conditions:

y
L(x, y)(y x) = (hx)

y
L(x, y)(y x)
2
= (hx)
2

y
L(x, y) = 0
65
The P and the Q measure
In our model, we actually use two drift functions:
P
() and
Q
(),
one dening the P or statistical measure and the latter modeling the Q
or pricing measure. We postulate that the only dierence between the
P and the Q measure lies in the specication of these two drift func-
tions. Correspondingly, we use the subscripts P and Q to identify the
Markov generator and transition probabilities under the corresponding
measure. Whenever the subscripts are omitted as here below, formulas
apply to both the P and the Q measure.
66
Eigenvalue problems and functional calculus
To manipulate the Markov generator by means of functional calculus,
the rst step is to diagonalize it. Let
n
be the eigenvalues of the
operator L and let u
n
(x) and v
n
(x) be the right eigenvectors, so that
Lu
n
=
n
u
n
.
67
Numerical methods for eigenvalue problems
In most cases, Markov generators admit a complete set of eigenvec-
tors. Although there are exceptions where diagonalization is not pos-
sible and one can reduce the operator at most to a non trivial Jordan
form with non-zero o-diagonal elements, these exceptional situations
occur very rarely both in a measure theoretic sense, as exceptions span
a set of zero measure, and in a topological sense as their complement
is dense in the space of all generators. In practical terms, this implies
that non-diagonalizable operators arise very rarely if at all in practice
and whenever they do, a professional numerical diagonalization algo-
rithm would detect the problem and a small perturbation of the model
parameters would rectify. To carry out numerical diagonalization, we
nd that the function dgeev in the public domain package LAPACK is
quite suitable.
68
Diagonalizing the Markov generator
We just assume that the operator L admits a complete set of eigen-
vectors. In this case, we can form the matrix U whose columns are
given by the eigenvectors u
n
(x) and write
L = UU
1
. (22)
We denote with V the operator U
1
and with v
n
(x) its row vectors.
69
Functional calculus
Key to our constructions is the remark that if the matrix operator L
is diagonalizable we can apply an arbitrary function F to it by means
of the following formula:
F(L) = UF()U
1
(23)
This formula is at the basis of the so-called functional calculus. As
Itos formula regarding functions of stochastic processes is central in
the stochastic analysis for diusion processes, functional calculus for
Markov generators plays a pivotal role in our framework for stochastic
volatility models. This formula has several applications. An immediate
one allows us to express the pricing kernel u(x, t; y, t

) of the process
as follows:
u(x, t; y, t

) = (e
(t

t)L
)(x, y) =

n
e

n
(t

t)
u
n
(x)v
n
(y). (24)
70
Introducing jumps
At this stage of the construction we add jumps. Jumps are ubiquitous
in credit model and we nd that a jump component is necessary in
order to reconcile observed default probabilities with credit migration
probabilities. Within our framework, adding jumps involves marginal
additional complexities from the numerical viewpoint.
71
Asymmetric jumps
To reect asymmetries in the jump intensities, we model separately
up and down jumps. A particularly interesting class of jump processes
is associated to stochastic time changes given by monotonously non-
decreasing processes T
t
with independent increments. These time
changes are known as Bochner subordinators and are characterized by
a Bochner function () such that
E
0
_
e
T
t
_
= e
()t
(25)
For example, the case of the variance gamma process which received
much attention in Finance corresponds to the function
() =

2

log
_
1 +

_
(26)
where is the mean rate and is the variance rate of the variance
gamma process.
72
Functional calculus with subordinated generators
The generator of the jump process corresponding to the subordination
of a process of generator L can be expressed using functional calculus
as the operator (L). To produce asymmetric jumps, we specify
the two parameters dierently for the up and down jumps and compute
separately two Markov generators
L

(L) = U

)V

(27)
where:

() =

2

log
_
1 +

_
(28)
73
Generators with asymmetric jumps
The new generator for our process with asymmetric jumps is obtained
by combining the two generators above
L =
_
_
_
_
_
_
_
_
0 0
L

(2, 1) d(2, 2) L
+
(2, 3) L
+
(2, n)
.
.
.
.
.
.
.
.
.
.
.
.
L

(n 1, n) L

(n 1, 2) d(n 1, n 1) L
+
(n 1, n)
0 0 0
_
_
_
_
_
_
_
_
Here the element of the diagonal are chosen in such a way to satisfy
the condition of probability conservation
d(i, i) =

j=i
L(i, j) (29)
Also notice that we have zeroed out the elements in the matrix at the
upper and lower boundary: this ensures that there is no probability
leakage in the process.
74
Adding jumps
At this stage of the construction, we have therefore obtained a gen-
erator L
j
for the process of distance to default, whose dynamics is
characterized by a combination of state dependent local volatility and
asymmetric jumps. We note that the addition of jumps has not in-
creased the dimensionality of the problem and is therefore computa-
tionally ecient.
75
PART III: Estimation and calibration: P measure
We rst estimate the process for distance to default x
t
with respect to
the statistical measure P by matching transition probabilities over one
year and default probabilities over time horizons of 1, 3 and 5 years.
A credit rating system consists of a number K of dierent classes. In
the case of the extended system by Moodys, K = 18 and the ratings
are:
{0, 1, . . . , 17} {Default, Caa, B3, Ba3, Ba2, . . . , Aa3, Aa2, Aa1, Aaa}
76
Introducing barriers
We subdivide the nodes of the lattice into K subintervals of adjacent
nodes:
I
i
= [x
i1
, ...x
i
] (30)
where 0 = x
0
< x
1
< ... < x
K
= N and #(x
i
x
i1
) =
N
K
, for i =
1, ..., K. The interval I
i
corresponds to the i-th rating class. If a
process is in I
i
at time t, then is said to have a credit rating of i. i,
x
i
I
i
denotes the initial node. The conditional transition probability
p
ij
(t) that an obligor with a given initial rating i at time 0 will have
a rating j at a later time t > 0 can be estimated by matching it with
historical averages provided by credit assessment institutions.
77
Introducing barriers
For our purposes, we model this quantity as follows:
p
ij
(t) =
a
j
1

y=a
j1
u
P
(0, x
i
; t, y).
where x
i
is a point in the interval I
i
which represents the barycenter
of the population density in that credit class and is part of the model
specication. For simplicitys sake, we take x
i
to be the midpoint of
the interval I
i
.
78
The state of default
Absorption into the state x = 0 is interpreted as the occurrence of
default. The probability that starting from the initial rating i and
reaching a state of default by time t is given by
p
D
i
(t) = u
P
(0, x
i
; t, 0).
The model under P is characterized by a drift function
P
(), a
volatility function () and jump intensities. The rst two func-
tions are graphed below, while the variance rates we estimated are

+
= 7.5,

= 4.
79
Local volatility () vs. distance to default
80
Local drifts
P
() and
Q
() vs distance to default under the P
and the Q measure, respectively.
81
Comparison of discrete model (lines) and historical (dots) one
year transition probabilities.
82
Comparison of discrete model (lines) and historical (dots) de-
fault probabilities.
83
Estimation and calibration: Q measure
Risk neutralization is dened by changing the drift function
P
() into

Q
(), while leaving everything else unaltered.
The new drift is chosen in such a way to t spread curves. Term
structures of probability of default for each rating class are given by
q
D
i
(t) = u
Q
(0, x
i
; t, 0). (31)
84
CDX index spreads
In our example, we use CDS spreads for 125 names in the Dow Jones
CDX index. We looked at 5 datasets by Mark-it Partners correspond-
ing to the last business days of the months of January, February,
March, April and May 2005. The datasets provide CDS spreads at
maturities: 6m, 1y, 2y, 3y, 5y, 7y, 10y and tentative recovery rates for
each name. We insist that the CDS spreads be matched by our model
and take the liberty of adjusting the term structure of recovery rate
for each name. Besides having to estimate the drift under Q we also
estimate the current distance to default for each name. The objective
is to ensure that the term structure of recovery rates be as at as
possible and as close as possible to the tentative input value.
85
Comparison of discrete model (lines) and market (dots) for
spread curves of investment grade bonds (Data taken 02/10/2003)
86
Comparison of discrete model (lines) and market (dots) for
spread curves of speculative grade bonds (Data taken 02/10/2003)
87
Liquidity spreads
From these pictures one can notice a systematic bias in spreads. Our
model appears to systematically underestimate BB spreads and over-
estimate BBB spreads.
This can be interpreted in terms of the dierential liquidity in the two
market sectors.
88
CDS Spreads: Investment Grades (Data from March 2005)
89
CDS Spreads: Non-Investment Grades (Data from March 2005)
90
Implied term structure of recovery rates
We observe that the implied term structures of recovery rates are
highly correlated across names and to the general spread level. This is
not surprising as recovery levels are known to be linked to the economic
cycle. Hence implied recoveries reect the market perception of the
future economic cycle. As we compare the implied recovery cycles
on the last business day of January, March and May 2005, we notice
that the implied recovery cycle appears equally pronounced in the three
months. However, the ones in January and March show a much greater
degree of coherence across names, perhaps a signature of the fact that
in January and March markets were rather tranquil and ecient, while
in May 2005 dislocations occurred.
91
Implied recovery cycles for the CDX components on January
31st, 2005.
92
Implied recovery cycles for the CDX components on March 31st,
2005.
93
Implied recovery cycles for the CDX components on May 31st,
2005.
94
Risk-neutral transition probabilities
In the risk-neutral setting we can also calculate risk-neutral transition
probabilities. These are necessary to price credit-rating dependent
options.
How do we expect risk-neutral transition probabilities to behave? In-
dependent of the model, since risk-neutral default probabilities are
greater than historical default probabilities, one would expect down-
grades in credit-rating to be more probable in the risk-neutral setting
than historically.
95
96
97
98
PART IV. Equity default swaps (EDS)
Equity default swaps are a new class of instruments that several dealers
started marketing this year. They are dened as out of the money
American digital puts struck at 30% of the spot price. Typical maturity
is 5 years and the premium is paid in installments by means of a semi-
annual coupon stream.
In this example, I will compare CEV prices with the prices one obtains
from credit barrier models. The latter, are models estimated to ag-
gregate data, namely the credit transition matrix, default probabilities
and credit spread curves. The credit equity mapping is obtained by
tting at-the-money implied volatilities as a function of the ratings.
99
Main Finding
It appears that the market is currently pricing EDSs by means of
diusion local volatility models and that this is not entirely consistent
with credit derivative data. The marked dierences in prices are due
to the fact that the credit barrier model accounts for the phenomenon
of fallen angels by introducing and calibrating a jump component in
the process.
100
Credit-Equity mapping
101
Mapping to equity
The credit quality is mapped to equity prices via a deterministic, mono-
tonic function at some horizon date T:
S
T
() = e
rT
()
For t
i
< T, we take the discounted expectation of :
S
t
i
() = e
rt
i
E[()|
t
i
]
102
A snapshot of market EDS spreads
103
Credit quality versus stock price (the credit equity mapping)
104
Local volatility
105
At-the-money implied vols as a function of credit quality
106
EDS spreads as a function of rating
107
CDS to EDS spread ratio as a function of rating
108
CDS to EDS spread ratio based on the CEV model
109
PART V. Credit correlation modeling and lattice models for
synthetic CDOs.
Having characterized the process for credit quality x
t
and identied
starting points for each individual procsess, the next step is to in-
troduce correlations by conditioning to economic cycle scenarios, thus
introducing a correlation structure among the credit quality processes.
The economic cycle is modeled by means of a non-recombining lat-
tice of the structure sketched below. The underlying index variable is
allowed to take up two values on each period t. An upturn corre-
sponds to a good period while a downturn to a bad period for
the economy. In our example, we chose the time step to be t = 1y
and nd that this choice is sucient to provide great exibility in the
tuning of the correlation structure.
110
Conditioning the Lattice to a Market Index
To explain our methodology to introduce correlations, we consider
rst a simple case whereby the model is characterized by a pair of
complementary transition probabilities w, (1w) [0, 1] at each node,
which we assume constant. In order to condition the continuous time
lattice corresponding to a given credit quality process to the economic
index variable we introduce the notion of local beta given by function
() which provides the corresponding sensitivity. The limiting cases of
() = 0 and () = 1 correspond to zero and full correlation between
a name with a given credit quality hx [0, 1] and the cycle variable.
Along the path of each given scenario on the tree, the unconditional
kernel of the credit quality process is replaced by conditional transition
probabilities dened as follows:
u

w,
(t, x; t +t, y) = (1 (hx))u
0
(t, x; t +t, y) +(hx)u

1
(t, x; t +t, y)
111
Here u
0
= u is the unconditional kernel and corresponds to a zero
(hx). In the opposite case of (hx) = 1, conditional kernel u

1
(x, y)
has the following form:
u
+
1
(x, y) =
1
1 w
_

_
u(x, y) if y > m(w, x)
_
w

y>m(w,x)
u(x, a(w, x))
_
if y = m(w, x)
u(x, y) = 0 if y < m(w, x)
(32)
and
u

1
(x, y) =
1
w
_

_
0 if y > m(w, x)
u(x, m(w, x)) u
+
1
(x, m(w, x)) if y = m(w, x)
u(x, y) if y < a(w, x).
(33)
where
m(w, x) = inf
_
m = 0, ...N|

y<m
u(x, y) w
_
. (34)
Notice that, for all specications of () and w [0, 1], we have that
u(x, y) = wu

(x, y) +(1 w)u


+

(x, y). (35)


Conditioning is achieved by forming a weighted sum over all paths in
the event tree. On a given path, we use u

for a bad period scenario


and u
+

for a good one. The weight of a path is the product of


a number of factors w equal to the number of bad periods and a
number of factors (1w) for each one of the good periods. With this
method, marginal probabilities are kept unchanged while correlations
are induced on the single name processes. More specically, one can
price all credit sensitive instruments specied with the given names one
can rst evaluate the conditional prices P

by means of the following


multiperiod kernel:
e
(t
i
t
i1
)L

i
. . . e
(t
n
t
n1
)L

i
.
where = {
1
, . . . ,
n
} runs over the sets of conditional paths due to
the scenario of the index. The (unconditional) price is then given by:
P =

w
n

()
(1 w)
n
+
()
P

.
This construction can be generalized. Consider a number M > 1 of
percentile levels 0 < w
1
< ... < w
M
< 1 and let q
i
[0, 1], i = 1...M be
a corresponding set of probabilities summing up to one, i.e.

i
q
i
= 1.
Then we can set
u

w,
(t, x; t +t, y) =
M

i=1
q
i
u

w
i
,
(t, x; t +t, y). (36)
The formulas above extend also to this case as long as one replaces
the weight w with the average weight

i
q
i
w
i
.
The choices we make for the March and June datasets are graphed
below. Here one can observe that the levels we were led to choose in
June are lower and the probabilities more uneven than in March. This
can be interpreted as saying that the model is detecting a higher level
of implied correlation between jumps in the June data than in March.
Specications for the weights q
i
and percentile levels w
i
.
112
The local beta function
Modeling correlation is key to pricing basket credit derivatives. Buyers
and sellers of basket credit derivatives have a wide range of arbitrage-
free prices to choose from, and it is the market that determines, both in
principle and in practice, a denite price. In our framework, tranches
of varying seniority are priced by calibrating the local beta function
().
113
Specications for the function (x) in March and June 2005.
114
Decoupling of correlation
Notice that as an eect of GM and Ford being downgraded, the local
beta function responded by lowering on the side of low quality grades
while rising on the high qualities. This resulted in a simultaneous fall
of equity prices and widening of senior spreads.
115
Contagion skew
A useful graph to assess the impact of the specication of the local
() function on our correlation model is the contagion skew in the
next slide. This graph is constructed as follows: we rst compute the
unconditional default probabilities as a function of credit quality. Next,
for each value of credit quality, assuming that a name of that quality
defaults within a time horizon of 5 years, we compute the conditional
probability of defaults for all other name. Finally, we take an average
over all values of credit quality of the ratio between the conditional
and unconditional probabilities. As the graph shows, the higher is the
credit quality of a defaulted name, the larger is the impact on all other
names. The steepness of this curve controls precisely the discrepancy
between prices for senior tranches as compared to junior ones.
116
Contagion Skew: ratio between conditional and unconditional
probability of defaults, where conditioning is with respect to the
default of a name whose credit quality is on the X axis.
117
Pricing CDOs
Although CDX index tranches are written on 125 underlying names, we
observe that our lattice model performs quite eciently. We separate
the numerical analysis in two dierent steps. In the rst we go through
all names and generate conditional lattices. We choose t = 1y and a
time horizon of 5y, so that we obtain a total of 32 scenarios. This is a
pre-processing step which is independent of the CDO structure. This
step typically takes a few minutes for a hundred names and could
be carried out periodically and oine for the universe of all traded
names. The pricing step instead takes only a few seconds and requires
generating the probability distribution function for CDO portfolios over
the given time horizon.
118
Expected Loss Distribution for CDX index tranches in March
2005
119
Expected Loss Distribution for CDX index tranches in June 2005
120
Pricing CDOs
The model can be calibrated by adjusting the function (), [0, 1],
the thresholds w
i
, i = 1, ...M and the corresponding probabilities q
i
.
121
Tranche prices for March 20th 2005
attachment detachment spread mktspread
0% 3% 499.6 bp (+32% u) 500 bp(+32% u)
3% 6% 187.4 bp 189bp
6% 9% 108.6 bp 64bp
9% 12% 56.5 bp 22bp
12% 22% 6.7 bp 8bp
where u stands for upfront fee.
122
Calibration
Notice that a good agreement can be reached with the equity, junior
mezzanine and senior tranche. On the other hand, the model appears
to over-estimate the price of the two senior mezzanine tranches 6-9
and 9-12 by a factor 2-3. This might be in relation to the high degree
of liquidity of these tranches and appetite for this risk prole.
123
Tranche prices for June 20th 2005
attachment detachment spread mktspread
0% 3% 499.7 bp (+49% u) 500 bp(+49% u)
3% 6% 170.1 bp 177bp
6% 9% 30.4 bp 54bp
9% 12% 27.5 bp 24bp
12% 22% 10.0 bp 12bp
124
Hedge ratios of the various CDO tranches for March 2005 plot-
ted against 5Y CDS spreads.
125
Hedge ratios.
One can notice that the hedge curves for the equity and the junior
mezzanine are fairly dierent and as a consequence it does not appear
as appropriate to use the mezzanine as a proxy to hedge credit expo-
sure at the equity tranche level. The dierentiation among the two
proles is a direct consequence of the steep aspect of the local beta
function.
126
Conclusions
We propose a novel approach to dynamic credit correlation modeling
that is based on continuous time lattice models correlated by con-
ditioning to a non-recombining tree. The model describes not only
default events but also rating transitions and spread dynamics, while
single name marginal processes are preserved.
127
PART VI. Credit-interest rate hybrids
Functional lattices for the dynamic CDO model and for the term struc-
ture model covered above can be combined and correlated while pre-
serving the specication of marginal processes. This opens the possi-
bility of pricing credit - interest rate hybrid instruments.
As an example of these applications, in the following, we consider
cancellable interest rate swaps which are linked to the default of either
one name in the CDX index, or to the rst default event of a name
in a given basket, or to the default of the CDX equity tranche. We
also consider interest oors that cancel upon the default of the equity
tranche. In all cases, we evaluate also hedge ratios.
128
Single name, credit linked cancelable swaps
129
First to default cancelable interest rate swap
130
Basis of a CDO subordinated interest rate swap
131
Hedge ratios for a CDO subordinated interest rate swap
132
Price of a CDO subordinated interest rate oor
133
Hedge ratios for a CDO subordinated interest rate oor
134
Conclusions
We nd that our model is well suited for interest rate hybrids. It is
numerical ecient and since it does not involve a Montecarlo step,
hedge ratios have no simulation noise.
We nd that, within a local beta model for credit correlations, hedge
proles tend to be relatively higher for the better quality ratings which
are more correlated to the business cycle.
135

Você também pode gostar