April 18, 2014

Program in Probability and its Applications

Workshop on Probability in Finance
Tuesday January 26, 1999 -- Saturday January 30, 1999

Titles and Abstracts

Tomasz Bielecki, Northeastern Illinois University, Chicago
"Recent Results in Credit Risk Modeling: A Multiple Credit Ratings Case"

A new approach to modeling and analysis of defaultable term structure is presented. The approach is based on the HJM-type model of the term structure of credit spreads. We assume that the 'pre-default' values of defaultable bonds are given a priori, and we search for an arbitrage free setup that would support these values. This is achieved by formulating conditions that relate credit spreads, equivalent martingale measures and intensities of transitions of bond processes between different credit classes. In addition, by taking into account credit class dependent recovery rates we arrive at arbitrage free model that involves both default-free bonds and defaultable term structure with multiple ratings.

Joint work with Marek Rutkowski, Technical University of Warsaw, Warszawa, Poland.

Michel Crouhy, CIBC
"A Comparative Analysis of Credit Risk Models and Thoughts for Future Research"

Mark Davis, Tokyo-Mitsubishi International, London
"Valuation of Convertible Bonds"

A convertible bond (CB) confers on the holder the right -- exercisable at specified time periods -- to exchange it for ordinary shares of the issuer. Valuation has to some extent been controversial because the CB "looks like a bond" when out of the money and "looks like a share" when in the money, leaving open the question of what discount factor should be used. In this talk we introduce an arbitrage-free model which explicitly includes the possibility of default. Correct calibration of this model to the market, and the dependence of the value on credit spread volatility, will be discussed.

Freddy Delbaen, ETH Zurich
"Bessel Processes and Passport Options"

Passport options are financial instruments that allow the owner to freely develop bounded investment strategies in an underlying asset. She can keep the positive gains but she is not liable for the losses. There are different kinds of passport options, each of them requiring special tools from stochastic analysis. The talk will be devoted to two kinds of options. In both cases Skorohod's lemma is central to reduce the problem to a Russian type option. After that, the representation of the Brownian-exponential as a time transformed Bessel process reduces the pricing problem to a barrier problem concerning Bessel processes. In contrast to earlier papers, where PDE's were used, the present analysis only uses stochastic analysis. We will also give a new martingale equality for general continuous martingales.

This is joint work with Marc Yor.

Ron Dembo, Algorithmics Inc.
"Liquidity and The True Simulation of Dynamic Portfolio Risk"

Liquidity risk is a killer; it killed the Hunt Brothers, Metalgesellshaft, Barings, Long Term Capital Management and more. There's nothing like stumbling on a scenario where the cost of a trade is prohibitive, where a hedge is no longer a hedge because the one side cannot trade.

Today's risk standards do not allow for the measurement of liquidity risk. They are based on measures that assume a static portfolio. The reality is that liquidity risk is manifested when a portfolio must be rebalanced or a market event requires added margin payments. To measure liquidity true dynamic portfolio risk measurement is essential. A liquidity crisis often occurs over a period and not at a single point in time. It's the cumulative erosion of capital that can bring a trader to her knees. Liquidity risk must be measured over time and not at a single horizon. The dynamics of the portfolio must be accounted for or else the simulation will be misleading. Proper accounting for portfolio aging and scenario dependent valuation are a must.

This talk describes the framework we have implemented to accomplish this on large-scale industrial portfolios.

Darrell Duffie, Stanford University
"Correlated Default Timing and Valuation"

This talk will suggest simple models and illustrative calculations for the valuation and simulation of contingent claims that depend on the times and identity of correlated credit events, such as defaults. Examples include credit derivatives with a first-to-default feature, credit derivatives signed with a defaultable counterparty, credit-enhancement or guarantees, and collateralized debt obligations.

Daniel Dufresne, University of Melbourne
"Laguerre Series for Asian Options"

We consider the problem of pricing continuously averaged Asian options, when the risky asset is modelled as Geometric Brownian motion. This problem is equivalent to finding the distribution of the integral of geometric Brownian motion over a finite interval (denoted A in the sequel). (1) Some new results are derived: -- the law of 1/A is determined by its moments; -- expressions for the moments of 1/A. (2) It is shown how Laguerre series apply to density functions and option values; the formulas are simpler if expressed in terms of the ladder height dsitribtution. (3) Series expressions are obtained for the density function of A(t) and also for Asian options. Numerical illustrations show perfect fit with simulation results.

Robert J. Elliott, University of Alberta
"Affine Bond Prices and Stochastic Flows"

Stochastic flows and their Jacobians are used to show why, when the short rate is Gaussian (as in the Vasicek or Hull-White models), or square root (as in the Cox-Ingersoll-Ross, or Duffie-Kan models), the bond price is an exponential affine function

Paul Embrechts, ETH Zurich
"What Financial Risk Managers can learn from Actuaries"

The flux of methodological input in the financial industry has very much been one from banking towards insurance. Recently, various fields of finance have witnessed a reversed flow, so much so that the banking industry is well-advised to take notice of so-called insurance-analytics (a term coined by Till Guldimann, Infinity). Examples of this flow are to be found in:

risk management (going beyond VaR)
credit risk management (actuarial reserving)
credit derivatives (using survival analytic methods).
Some examples of the above will be discussed, both from a theoretical as well as applied point of view.

This is joint work with C.Klueppelberg and T.Mikosch.

Helyette Geman, University Paris Dauphine and ESSEC
"Stochastic Time Changes and Asset Price Modeling"

Despite its pivotal role in the theoretical financial literature (Capital Asset Pricing Model, Black-Scholes-Merton formula), the normality of asset returns has been consistently refuted in empirical research. The first part of the talk establishes on a high-frequency database of S & P500 returns that a remarkable quasi-perfect normality can be recovered using a stochastic clock driven by the number of trades (where no a priori distribution is assumed for the transaction time). This result is consistent with the beautiful theorem established by Monroe (1978) on "Processes that can be embedded in Brownian motion".

The second part of the talk argues, in full agreement with recent market moves observed around the world, that price processes should be represented as pure jump processes. Continuity and normality may be obtained after a time change related to the order flow. Different types of Levy processes for the modeling of asset prices are analyzed, as well as the relationship between the corresponding Levy measure and the intensity of the economic activity.

The talk is based on two articles by Ane and Geman (1997) and Geman-Madan-Yor (1998).

David C. Heath, Cornell University / CMU
"Futures-based Term Structure Models"

There are currently two paradigms for term structure modelling: modelling the spot rate, and modelling the term structure of forward rates. Each has advantages and disadvantages: For spot rate modelling the question of model choice is unclear, while for most HJM models computations are difficult. We present a new class of term structure models essentially as general as either of the above and for which differences between models are easy to understand and, for a class of interesting models, computations are easy.

Bjarne Hojgaard, Aalborg University
"Optimal Risk Controland Dividend Distribution Policies for Insurance Corporations"

We consider a model of a financial corporation which has to find an optimal policy balancing its risk and expected profits. The example treated is related to an insurance company with the risk control method being reinsurance. Under this scheme the insurance company divert part of its premium stream to another company in exchange of an obligation to pick up that amount of each claim which exceeds a certain level 'a' (rentention level). This reduces the risk but it also reduces the potential profit. The objective is to make a dynamic choice of the retention level and find the dividend distribution policy, which maximizes the cumulative expected discounted dividend pay-outs.

Consider the classical Cramer-Lundberg model, that is (when no dividends are distributed) the reserve R(t) of the company is assumed to be given by R(t)=x+p(a,l)t-(U(a,1)+...+U(a,N(t)) where N(t) is a Poisson process with intensity b>0, U(a,i) is the i.i.d. claim sizes, when the retention level is a and p(a,l)=(1+l)bE(U(a,1)), where l>0 is the relative safety loading. We then have that the process lR(t/l^2) converges in distribution to a BM(m(a),s(a)),where m(a)=bE(U(a,1)) and s^2(a)=bE([U(a,1)]^2).

Hence we consider the following problem: A policy is a pair (a(t),L(t)), where a(t) denotes the retention level at time t and L(t) denotes the total amount of dividend distributed until time t. The reserve r(t) is assumed to be governed by dr(t)=m(a(t))dt+s(a(t))dW(t)-dL(t) and the objective is to maximize present value of dividend pay-out until eventual ruin.

We consider two different reinsurance strategies: 1. Proportional reinsurance, where the retention level a is between 0 and 1 and U(a,i)=aU(i). 2. Excess-of-loss reinsurance, where the retention level a is non-negative U(a,i)=min(U(i),a).

Mathematically this becomes a mixed singular-regular control problem for diffusion processes. Its analytical part is related to a free boundary (Stephan) problem for a linear second order differential equation and closed form solutions are found in both cases.

Ioannis Karatzas, Columbia University
"Dynamic Measures of Market-Risk"

Suppose that we operate in the framework of a standard financial market over a finite time-horizon [0,T], at the end of which we face a certain liability C -- a random quantity representing a payment that has to be made at time t=T . Suppose also that (due to market incompleteness, or insufficient initial funds, or both) we find it impossible to hedge at t=T this liability perfectly, that is, with probability one.

How can we then quantify, at the outset t=0, the risk associated with the hedging of the liability C at time t=T? One way is to try and maximize the probability of perfect hedge (cf. Karatzas (1997), or Foellmer and Leukert (1998)). This is, in a sense, equivalent to a dynamic version of the familiar "value at risk" concept.

Another approach is to try to minimize the "expected shortfall" E[max(C-X(T; x, p(.))), 0)] over admissible portfolios p(.), and then to maximize the resulting quantity over all risk-neutral probability measures P. Here x is the initial capital X(T)=X(T; x, p(.)) the terminal wealth corresponding to x and to the portfolio p(.), and E denotes expectation with respect to the probability measure P. The resulting max-min quantity can then be used as a measure of risk associated with the liability C; as such, it satisfies a number of desirable "coherence" properties postulated by Artzner, Eber, Delbaen and Heath (1996).

In the case of a complete market, when there is only one risk-neutral probability measure P, we present a fully-developed theory for this problem -- along with specific examples of contingent claims C for which explicit computations of risk are possible. The classes of admissible portfolios p(.) and "scenarios" (probability measures) P are rich enough to accommodate margin requirements and uncertainty about the actual values of stock appreciation rates, respectively. We also survey recent work on this problem in the context of incomplete markets, and point out to connections with the generalized Neyman-Person lemma when testing a simple hypothesis against a composite alternative.

This is joint work with J. Cvitanic.

Alexander Levin and Alexander Tchernitser, Bank of Montreal
"Multifactor Stochastic Variance Value-at-Risk Model"

A standard Value-at-Risk (VaR) model corresponds to stable market conditions and assumes a multivariate normal distribution for risk factors with known constant volatilities and correlations. However, the actual risk factor distributions exhibit significant deviations from normality. Excess kurtosis, skewness, and volatility fluctuations are typical for many market variables. Fat-tailed and skewed distributions result in the underestimation of actual VaR by the standard model.

The Stochastic Variance VaR model developed by the Bank of Montreal accounts for uncertainty and instability of the risk factor volatilities. The model naturally describes the dynamics of underlying asset returns for short holding periods typical for VaR calculations. The SV-VaR model fits the actual historical distributions of the risk factors better than the traditional VaR model. Higher moments (skewness, kurtosis) are more accurately captured with the SV-VaR model, which also incorporates correlations between risk factors, as well as correlations between risk factors and their volatilities.

The one-period exponential distribution for the stochastic variance is derived from the Maximum Entropy Principle. This model is extended to the Gamma SV Model that gives the Bessel distribution for the probability density of the risk factor. Corresponding stochastic processes with closed form solutions for the stochastic variance and risk factor dynamics are obtained. Derived simple volatility term structure differs from the term structure for well-known diffusion SV models in the case of short holding periods and better describes an empirical term structure of the risk factor kurtosis.

A general calibration procedure for the class of multifactor SV-VaR models is developed. A closed form solution for the VaR of one-factor linear portfolios is obtained. For the multifactor nonlinear portfolios, a simple two-step Monte Carlo simulation procedure is proposed. Numerical results for equity, commodity, interest rate, and foreign exchange rate risk are presented.

Andrew W. Lo, MIT
"When is Time Continuous?"

Continuous-time stochastic processes have become central to many disciplines, yet the fact that they are approximations to physically realizable phenomena is often overlooked. We quantify one aspect of the approximation errors of continuous-time models by investigating the replication errors that arise from delta-hedging derivative securities in discrete time. We characterize the asymptotic distribution of these replication errors and its joint distribution with other assets as the number of discrete time periods increases. We introduce the notion of temporal granularity of a continuous-time stochastic process, which allows us to characterize the degree to which discrete-time approximations of continuous-time models can track the payoff of a derivative security. We derive closed form expressions for the temporal granularity of geometric Brownian motion and an Ornstein-Uhlenbeck process using call options. We also introduce alternative measures of the replication error and analyze their properties.

This is joint work with D. Bertsimas and L. Kogan.

Ludger Overbeck, Deutsche Bank AG
"Credit Portfolio Risk Management Based on Coherent Risk Measures"

A financial institution uses the economic capital for credit risk as a protection against severe losses in the entire credit portfolio. Mathematically, it is usually defined as a quantile of the distribution of future losses, or even simpler as a multiplier of the standard deviation of this distribution. The classical portfolio theory explains then how to distribute the capital across the whole portfolio.

Since the fundamental work of Artzner et. al about coherent risk measures, other risk measures, like the conditional expectation of the losses given that the loss already exceeded a given threshold, are analyzed in research as well as in applications. Some features of these risk measures are exploit. In particular we present a capital allocation process in the spirit of the exceedance over threshold measures. These measures are compared with classical portfolio theory, i.e. with the risk contributions based on a variance/covariance approach.

L.C.G. Rogers, University of Bath, England
"Designing and Estimating Models of High-Frequency Data"

Most financial houses have access to high-frequency data, which typically gives the time, price and amount of every trade (or quote) in a particular asset. Such detailed information should be more revealing than a single price per day, but it will be hard to extract the additional value if one tries to use a model which supposes that the observed prices are a diffusion process! In this talk, we present a class of models for such data which treat the data as intrinsically discrete, and we show how easily-updated estimation procedures can recover parameter values from a range of simulated examples.

Stephen Ross, MIT
"Topics in Finance"

In finance, as in pathology, we can learn more from failure than from success. This paper exhumes three famous financial failures, the Hunt Brothers silver ventures, Metallgesellschaft's oil futures losses, and the recent LTCM and related hedge fund failures. We do a post mortem on each and see what we can learn. Not surprisingly, the cause of death was similar in each case, or, to put it more familiarly, those who pay no heed to history are doomed to repeat it.

Hiroshi Shirakawa, Tokyo Institute of Technology
"Evaluation of Yield Spread for Credit Risk"

We study the rational evaluation of yield spread for defaultable credit with fixed maturity. The default occurs when the asset value hits a given fraction of the nominal credit value. The yield spread is continuously accumulated to the initial credit as an insurance fee for future default. By the rational credit pricing, we prove the unique existence of equilibrium yield spread which satisfies the arbitrage free property. Furthermore we show that this spread yield is independent of the choice of interest rate process. For the quantitative study of rational yield spread, we derive an explicit analytic formula for the equilibrium and show numerical example for various parameters.

Steven E. Shreve, CMU
"Pricing and Hedging Dangerous Exotic Options"

The Black-Scholes "delta-hedging" approach cannot be implemented for exotic options which exhibit large "gamma" values (e.g., options which knock out in the money), because this would require frequent large changes in the hedge position. For such options, one can build a "margin of safety" into the price, and use this margin to avoid large changes of position in the underlying. A general methodology for this, based on the idea of pricing and hedging under portfolio constraints, will be presented.

Stuart Turnbull, CIBC
"The Intersection of Market and Credit Risk"

Economic theory tells that market risk and credit risks are intrinsically related to each other and are not separable. We start by describing the two main approaches to pricing credit risky instruments: the structural approach and the reduced form approach. It is argued that the standard approaches to credit risk management - Credit Metrics, Credit Risk Plus and KMV - are of limited value, if applied to portfolios of interest rate sensitive instruments.

Empirically it is observed that returns on high yield bonds have a higher correlation with the return on an equity index and a lower correlation with the return on a Treasury bond index than do low yield bonds. The KMV and Credit Metrics methodologies cannot reproduce these empirical observations given their assumptions of constant interest rates. Altman (1983) and Wilson (1997) have shown that macro economic variables appear to influence the aggregate rate of business failures. We show how to incorporate empirical observations into the reduced form Jarrow-Turnbull (1995) model. The volatility of the credit spread can be used to determine the sensitivities of the credit spread to the different factors. Correlation plays an important role in existing methodologies. Here default probabilities are correlated due to their common dependence on the same economic factors. We discuss the implications for pricing, given different assumptions about a bond holder's claim in the event of default. We compare the Duffie-Singleton ( 1997) assumption to the legal claim approach, where a bond holder's claim is assumed to be accrued interest plus principal. Default risk and the uncertainty associated with the recovery rate may not be the sole determinants of the credit spread. We show how to incorporate a convenience yield as one of the determinants of the credit spread.

Incorporating market and credit risk implies that it is necessary to use the martingale distribution for pricing and the natural distribution to describe the value of the portfolio in order to calculate the value-at-risk. We show how to generalize the Credit Metrics methodology to incorporate stochastic interest rates.


Hans Foellmer, Humboldt Universitaet - Berlin

"Probabilistic Problems arising from Finance"

We review some recent developments in Probability which are motivated by problems of hedging derivatives in incomplete financial markets. This will include new variants of decomposition theorems for semimartingales, the construction of efficient hedges which minimize the shortfall risk under some cost constraint, and some results on Brownian motion related to the heterogeneity of information among financial agents.

Return to Finance Workshop Homepage

Return to Workshop Schedule