Backtesting Value-at-Risk: A Duration-Based Approach 1

Size: px
Start display at page:

Download "Backtesting Value-at-Risk: A Duration-Based Approach 1"

Transcription

1 Backtesting Value-at-Risk: A Duration-Based Approach 1 Peter Christoffersen 2 McGill University, CIRANO and CIREQ Denis Pelletier 3 Université de Montreal, CIRANO and CIREQ January 31, The first author acknowledges financial support from IFM2, FCAR, and SSHRC, and the second author from FCAR and SSHRC. We are grateful for helpful comments from Frank Diebold, Jean-Marie Dufour, Rob Engle, Eric Ghysels, James MacKinnon, Nour Meddahi, and Matt Pritsker. The usual disclaimer applies. 2 Corresponding author. Faculty of Management, 1001 Sherbrooke Street West, Montreal, Quebec, Canana H3A 1G5. Phone: (514) Fax: (514) Peter.Christoffersen@McGill.ca 3 Department de Sciences Economiques, CP 6128, succursale Centre-Ville, Montreal, Quebec, H3C 3J7, Canada. Denis.Pelletier@UMontreal.ca.

2 Abstract Financial risk model evaluation or backtesting is a key part of the internal model s approach to market risk management as laid out by the Basle Commitee on Banking Supervision (1996). However, existing backtesting methods such as those developed in Christoffersen (1998), have relatively small power in realistic small sample settings. Methods suggested in Berkowitz (2001) fare better, but rely on information such as the shape of the left tail of the portfolio return distribution, which is often not available. By far the most common risk measure is Value-at-Risk (VaR), which is defined as a conditional quantile of the return distribution, and it says nothing about the shape of the tail to the left of the quantile. Our contribution is the exploration of a new tool for backtesting based on the duration of days between the violations of the VaR.Thechief insightisthatifthevar model is correctly specified for coverage rate, p, then the conditional expected duration between violations should be a constant 1/p days. We suggest various ways of testing this null hypothesis and we conduct a Monte Carlo analysis which compares the new tests to those currently available. Our results show that in realistic situations, the duration based tests have better power properties than the previously suggested tests. The size of the tests is easily controlled using the Monte Carlo technique of Dufour (2000).

3 1 Motivation Financial risk model evaluation or backtesting is a key part of the internal model s approach to market risk management as laid out by the Basle Committee on Banking Supervision (1996). However, existing backtesting methods such as those developed in Christoffersen (1998), has relatively small power in realistic small sample settings. Methods suggested in Berkowitz (2001) fare better, but rely on information such as the shape of the left tail of the portfolio return distribution, which is often not available. By far the most common risk measure is Value-at-Risk (VaR), which is defined as a conditional quantile of the return distribution, and it says nothing about the shape of the tail to the left of the quantile. We will refer to an event where the ex-post portfolio loss exceeds the ex-ante VaRmeasureasa violation. Of particular importance in backtesting is the clustering of violations. An institution s internal risk management team as well as external supervisors explicitly want to be able to detect clustering in violations. Large losses which occur in rapid succession are more likely to lead to disastrous events such as bankruptcy. In the previous literature, due to the lack of real portfolio data, the evaluation of VaR techniques were largely based on artificial portfolios. Examples in this tradition include Beder (1995), Christoffersen, Hahn and Inoue (2001), Hendricks (1996), Kupiec (1995), Marshall and Siegel (1997), and Pritsker (1997). But recently, Berkowitz and O Brien (2002) have reported on the performance of actual VaR forecasts from six large (and anonymous) U.S. commercial banks. 1 Figure 1 reproduces a picture from their paper which shows the VaRexceedences from the six banks reported in standard deviations of the portfolio returns. Even though the banks tend to be conservative they have fewer than expected violations the exceedences are large and appear to be clustered in time and across banks. From the perspective of a regulator worried about systemic risk, rejecting a particular bank s risk model due to the clustering of violations is particularly important if the violations also happen to be correlated across banks. The detection of violation clustering is particularly important because of the widespread reliance on VaRs calculated from the so-called Historical Simulation (HS) technique. In the HS methodology, a sample of historical portfolio returns using current portfolio weights is first constructed. The VaRis then simply calculated as the unconditional quantile from the historical sample. The HS method thus largely ignores the last 20 years of academic research on conditional asset return models. Time variability is only captured through the rolling historical sample. In spite of forceful warnings, such as Pritsker (2001), the model-free nature of the HS technique is viewed as a great benefit by many practitioners. The widespread use of HS the technique motivates us to focus attention on backtesting VaRs calculated using this method. 1 Barone-Adesi, Giannopoulos and Vosper (2000) provides another example using real-life portfolio returns. 1

4 While alternative methods for calculating portfolio measures such as the VaRhave been investigated in for example Jorion (2000), and Christoffersen (2002), available methods for backtesting are still relatively few. Our contribution is thus the exploration of a new tool for backtesting based on the duration of days between the violations of the risk metric. The chief insight is that if the VaR model is correctly specified for coverage rate, p, then the conditional expected duration between violations should be a constant 1/p days. We suggest various ways of testing this null hypothesis and we conduct a Monte Carlo analysis which compares the new tests to those currently available. Our results show that in many realistic situations, the duration based tests have better power properties than the previously suggested tests. The size of the tests is easily controlled using the Monte Carlo testing approach of Dufour (2000). This procedure is described in detail below. We hasten to add that the sort of omnibus backtesting procedures suggested here are meant as complements to and not substitutes for the statistical diagnostic tests carried out on various aspects of the risk model in the model estimation stage. The tests suggested in this paper can be viewed either as a final diagnostic for an internal model builder or alternatively as a feasible diagnostic for an external model evaluator for whom only limited, aggregate portfolio information is available. Our paper is structured as follows: Section 2 outlines the previous first-order Markov tests, Section 3 suggests the new duration-based tests, and Section 4 discusses details related to the implementation of the tests. Section 5 contains Monte Carlo evidence on the performance of the tests. Section 6 suggests various extensions to the analysis, and Section 7 concludes. 2 Extant Procedures for Backtesting Value-at-Risk Consider a time series of daily ex-post portfolio returns, R t, and a corresponding time series of ex-ante Value-at-Risk forecasts, VaR t (p) with promised coverage rate p, such that ideally Pr t 1 (R t < VaR t (p)) = p. ThenegativesignarisesfromtheconventionofreportingtheVaR as a positive number. Define the hit sequence of VaR t violations as ( 1, if R t < VaR t (p) I t = 0, else Notice that the hit sequence appears to discard a large amount of information regarding the size of violations etc. Recall, however, that the VaR forecast does not promise violations of a certain magnitude, but rather only their conditional frequency, i.e. p. Thisisamajordrawback of the VaRrisk measure which we will discuss below. 2

5 Christoffersen (1998) tests the null hypothesis that I t i.i.d. Bernoulli(p) against the alternative that I t i.i.d. Bernoulli(π) and refers to this as the test of correct unconditional coverage (uc) H 0,uc : π = p which is a test that on average the coverage is correct. The above test implicitly assumes that the hits are independent an assumption which we now test explicitly. In order to test this hypothesis an alternative is defined where the hit sequence follows a first order Markov sequence with switching probability matrix " # 1 π01 π 01 Π = 1 π 11 π 11 where π ij is the probability of an i on day t 1 being followed by a j on day t. The test of independence (ind) is then H 0,ind : π 01 = π 11 Finally one can combine the two tests in a test of conditional coverage (cc) H 0,cc : π 01 = π 11 = p The idea behind the Markov alternative is that clustered violations represent a signal of risk model misspecification. Violation clustering is important as it implies repeated severe capital losses to the institution which together could result in bankruptcy. Notice however, that the Markov first-order alternative may have limited power against general forms of clustering. The first point of this paper is to establish more general tests for clustering which nevertheless only rely on information in the hit sequence. 3 Duration-Based Tests of Independence The above tests are reasonably good at catching misspecified risk models when the temporal dependence in the hit-sequence is of a simple first-order Markov structure. However we are interested in developing tests which have power against more general forms of dependence but which still rely only on estimating a few parameters. 3

6 The intuition behind the duration-based tests suggested below is that the clustering of no-hit durations will result in an excessive number of relatively short and relatively long durations, corresponding to market turbulence and market calm respectively. Motivated by this intuition we consider the duration of time (in days) between two VaRviolations (i.e. the no-hit duration) as D i = t i t i 1 where t i denotes the day of violation number i. 2 Under the null hypothesis that the risk model is correctly specified, the no-hit duration should have no memory and a mean duration of 1/p days. To verify the no memory property note that under the null hypothesis we have the discrete probability distribution Pr (D =1) = p Pr (D =2) = (1 p) p Pr (D =2) = (1 p) 2 p... Pr (D = d) = (1 p) d 1 p. A duration distribution is often best understood by its hazard function, which has the intuitive definition of the probability of a getting a violation after D daysgiven thatwehavegoned days without a violation. The above probability distribution implies a flat discrete hazard function as the following derivation shows λ (d) = Pr (D = d) 1 P j<d Pr (D = d) (1 p) d 1 p = 1 P d 1 i=1 (1 p)i p (1 p) d 1 p = 1 P d 2 j=0 (1 p)j p = p. The only memory free (continuous) 3 random distribution is the exponential, thus we have that under the null the distribution of the no-hit durations should be f exp (D; p) =p exp ( pd). 2 For a general introduction to duration modeling, see Kiefer (1988) and Gourieroux (2000). 3 Notice that we use a continuous distribution even though we are counting time in days. This discreteness bias will we acounted for in the Monte Carlo tests. The exponential distribution can also be viewed as the continuous time limit of the above discrete time process. See Poirier (1995). 4

7 In order to establish a statistical test for independence we must specify a (parsimonious) alternative which allows for duration dependence. As a very simple case, consider the Weibull distribution where f W (D; a, b) =a b bd b 1 exp (ad) b. The Weibull distribution has the advantage that the hazard function has a closed form representation, namely λ W (D) f W (D) 1 F W (D) = ab bd b 1 where the exponential distribution appears as a special case with a flat hazard, when b =1. The Weibull will have a decreasing hazard function when b<1, which corresponds to an excessive number of very short durations (very volatile periods) and an excessive number of very long durations (very tranquil periods). This could be evidence of misspecified volatility dynamics in the risk model. Due to the bankruptcy threat from VaR violation clustering the null hypothesis of independence is of particular interest. We therefore want to explicitly test the null hypothesis H 0,ind : b =1. We could also use the Gamma distribution under the alternative hypothesis. The p.d.f. in this case is f Γ (D; a, b) = ab D b 1 exp ( ad) Γ (b) which also nests the exponential when b =1. In this case we therefore also have the independence test null hypothesis as H 0,ind : b =1. The Gamma distribution does not have a closed-form solution for the hazard function, but the first two moments are b and b respectively, so the notion of excess dispersion which is a a 2 defined as the variance over the squared expected value is simply 1. b Note that the average duration in the exponential distribution is 1/p, and the variance of durations is 1/p 2, thus the notion of excess dispersion is 1 in the exponential distribution. 3.1 A Conditional Duration Test The above duration tests can potentially capture higher order dependence in the hit sequence by simply testing the unconditional distribution of the durations. Dependence in the hit sequence may show up as an excess of relatively long no-hit durations (quiet periods) and an excess of relatively short no-hit durations, corresponding to violation clustering. However, in the above 5

8 tests, any information in the ordering of the durations is completely lost. The information in the temporal ordering of no-hit durations could be captured using the framework of Engle and Russel s (1998) Exponential Autoregressive Conditional Duration (EACD) model. In the EACD(1,0) model, the conditional expected duration takes the following form E i 1 [D i ] ψ i = ω + αd i 1 with α [0, 1). Assuming an underlying exponential density with mean equal to one, the conditional distribution of the duration is f EACD (D i ψ i )= 1 µ exp D i ψ i ψ i The null of independent no-hit durations would then correspond to H 0,ind : α =0. Excess dispersion in the EACD(1,0) model is defined as V [D i ]/E[D i ] 2 = 1 1 2α 2 so that the ratio of the standard deviation to the mean duration is above one if α>0. 4 Test Implementation We will first discuss the specific implementation of the hit sequence tests suggested above. Later, we will simulate observations from a realistic portfolio return process and calculate risk measures from the popular Historical Simulation risk model, which in turn provide us with hit sequences for testing. 4.1 Implementing the Markov Tests The log-likelihood function for a sample of T i.i.d. observations from a Bernoulli variable, I t, with known probability p is written as ln L (I,p)=p T 1 (1 p) T T 1 where T 1 is the number of ones in the sample. The log-likelihood function for an i.i.d. Bernoulli with unknown probability parameter, π 1, to be estimated is ln L (I,π 1 )=π T 1 1 (1 π 1 ) T T 1. 6

9 TheMLestimateofπ 1 is ˆπ 1 = T 1 /T and we can thus write a likelihood ratio test of unconditional coverage as LR uc = 2(lnL (I, ˆπ 1 ) ln L (I,p)). For the independence test, the likelihood under the alternative hypothesis is ln L (I,π 01,π 11 )=(1 π 01 ) T 0 T 01 π T (1 π 11 ) T 1 T 11 π T where T ij denotes the number of observations with a j following an i. The ML estimates are and the independence test statistic is ˆπ 01 = T 01 /T 0 ˆπ 11 = T 11 /T 1 LR ind = 2(lnL (I, ˆπ 01, ˆπ 11 ) ln L (I, ˆπ 1 )). Finally the test of conditional coverage is written as LR cc = 2(lnL (I, ˆπ 01, ˆπ 11 ) ln L (I,p)). We note that all the tests are carried out conditioning on the first observation. The tests are asymptotically distributed as χ 2 with degree of freedom one for the uc and ind tests and two for the cc test. But we will instead rely on finite sample p-values below. Finally, as a practical matter, if the sample at hand has T 11 =0, which can easily happen in small samples and with small coverage rates, then we calculate the first-order Markov likelihood as ln L (I,π 01,π 11 )=(1 π 01 ) T 0 T 01 π T and carry out the tests as above. 4.2 Implementing the Weibull and EACD Tests In order to implement our tests based on the duration between violations we first need to transform the hit sequence into a duration series D i. While doing this transformation we also create the series C i to indicate if a duration is censored (C i =1)ornot(C i =0). Except for the first and last duration the procedure is straightforward, we just count the number of days between each violation and set C i =0. For the first observation if the hit sequence starts with 0 then 7

10 D 1 is the number of days until we get the first hit. Accordingly C 1 =1because the observed duration is left-censored. If instead the hit sequence starts with a 1 then D 1 is simply the number of days until the second hit and C 1 =0. Theprocedureissimilarforthelastduration. Ifthelastobservationofthehitsequenceis 0 then the last duration, D N(T ), isthenumberofdaysafterthelast1 in the hit sequence and C N(T ) =1because the spell is right-censored. In the same manner if the last observation of the hit sequence is a 1 then D N(T ) = t N(T ) t N(T ) 1 and C N(T ) =0. The contribution to the likelihood of an uncensored observation is its corresponding p.d.f. For a censored observation, we merely know that the process lasted at least D 1 or D N(T ) so the contribution to the likelihood is not the p.d.f. but its survival function S(D i )=1 F(D i ). Combining the censored and uncensored observations, the log-likelihood is L(D; Θ) = C 1 ln S(D 1 )+(1 C 1 )lnf(d 1 )+ N(T ) 1 X i=2 ln(f(d i )) +C N(T ) ln S(D N(T ) )+(1 C N(T ) )lnf(d N(T ) ). Once the durations are computed and the truncations taken care of, then the likelihood ratio tests can be calculated in a straightforward fashion. The only added complication is that the ML estimates are no longer available in closed form, they must be found using numerical optimization. 4.3 Finite Sample Inference While the large-sample distributions of the likelihood ratio tests we have suggested above are well-known, 4 they may not lead to reliable inference in realistic risk management settings. The nominal sample sizes can be reasonably large, say two to four years of daily data, but the scarcity of violations of for example the 1% VaRrenders the effective sample size small. In this section, we therefore present the technique of Monte Carlo tests [see Dufour (2000)]. For the case of a continuous test statistic, the procedure is the following. We first generate N independent realizations of the test statistic, LR i,i=1,...,n. WedenotebyLR 0 the test computed with the original sample. Under the hypothesis that the risk model is correct we know that the hit sequence is i.i.d. Bernoulli with the mean equal to the coverage rate in our application. We thus benefit from the advantage of not having nuisance parameters under the null hypothesis. 4 Testing α =0in the EACD(1,0) model presents a potential difficulty asymptotically in that it is on the boundary of the parameter space. 8

11 We next rank LR i,i=0,...,n in non-decreasing order and obtain the Monte Carlo p-value ˆp N (LR 0 ) where ˆp N (LR 0 )= NĜN(LR 0 )+1 N +1 with Ĝ N (LR 0 )= 1 NX 1 (LR i >LR 0 ) N i=1 where 1 ( ) takes on the value 1 if is true and the value 0 otherwise. When working with binary sequences the test values can only take a countable number of distinct values. Therefore, we need a rule to break ties between the test value obtained from the sample and those obtained from Monte Carlo simulation under the null hypothesis. The tie-breaking procedure is as follows: For each test statistic, LR i,i=0,...,n,wedrawan independent realization of a Uniform distribution on the [0; 1] interval. Denote these draws by U i,i=0,...,n. The Monte-Carlo p-value is now given by with G N (LR 0 )=1 1 N p N (LR 0 )= N G N (LR 0 )+1 N +1 NX 1 (LR i <LR 0 )+ 1 N i=1 NX 1 (LR i = LR 0 ) 1 (U i U 0 ). i=1 5 Backtesting VaRs from Historical Simulation We now assess the power of the proposed duration tests in the context of a Monte Carlo study. Consider a portfolio where the returns are drawn from a GARCH(1,1)-t(d) model with leverage, that is p R t+1 = σ t+1 ((ν 2) /ν)zt+1, with ³p 2 σ 2 t+1 = ω + ασ 2 t ((ν 2) /ν)zt θ + βσ 2 t where the innovation z t+1 s are drawn independently from a Student s t (ν) distribution. Notice that the innovations have been rescaled to ensure that the conditional variance of return will be σ 2 t+1. 9

12 In the simulations below we choose the following parameterization α = 0.1 θ = 0.5 β = 0.85 ω = e 6 ν = 8 where ω is set to target an annual standard deviation of The parameters imply a daily volatility persistence of 0.975, a mean of zero, a conditional skewness of zero, and a conditional (excess) kurtosis of 1.5. This particular DGP is constructed to form a realistic representation of an equity portfolio return distribution. The risk measurement method under study is the popular Historical Simulation (HS) technique. It takes the Value at Risk on a certain day to be simply the unconditional quantile of the past T e daily observations. Specifically VaR p t+1 = Percentile({R τ } t τ=t T e +1, 100p). From the return sample and the above VaR, we are implicitly assuming that $1 is invested each day. Equivalently, the VaRcan be interpreted as being calculated in percent of the portfolio value. In practice, the sample size is often determined by practical considerations such as the amount of effort involved in valuing the current portfolio holdings using past prices on the underlying securities. For the purposes of this Monte Carlo experiment, we set T e =250or T e =500 corresponding to roughly one or two years of trading days. The VaR coverage rate, p, is typically chosen in practice to be either 1% or 5%, and below we assess the power to reject the HS model using either of those rates. Figure 2 shows a return sample path from the above GARCH-t(d) process along with the 1% and 5% VaRsfromtheHS model (with T e = 500). Notice the peculiar step-shaped VaRs resulting from the HS method. Notice also the infrequent changes in the 1% VaR. The VaR exceedences from the return sample path and the 1% VaR are shown in Figure 3 reported in daily standard deviations of returns. The simulated data in Figure 3 can thus be compared with the real-life data in Figure 1, which was taken from Berkowitz and O Brien (2002). The clustering and the magnitude of the exceedences are quite similar across the two plots. Note that we have simulated 1,000 observations in Figure 3, while Figure 1 contains between 550 and 750 observations per bank. Figure 3 contains more violations than Figure 1 because of these differences in the sample size and because the banks in Figure 1 tend to report VaRswhichon average lead to fewer than p violations. 10

13 Before assessing the finite sample power results we simulate one very long realization (5 million observations) of the GARCH return process and calculate 1% and 5% VaRsfromHis- torical Simulation with a rolling set of 500 in-sample returns. The zero-one hit sequence is then calculated from the ex-post daily returns and the ex-ante VaRs. Figure 4 plots the hazard functions of the duration between violations in the long simulation of GARCH data and Historical Simulation VaRs. The hazard from the 1% VaR is shown in the top panel, and the 5% VaR in the bottom panel. The hazard functions are estimated nonparametrically via the Kaplan-Meier product-limit estimator of the survival function, which is described in Kiefer (1988). These hazards are estimated over intervals of 15 days so if there is a probability p of getting a hit at each day then the probability that a given duration will last less than 15 days is 15X Pr(D = i) = 15X i=1 i=1 (1 p) i 1 p = 1 (1 p) 15. For p equal to 1% and 5% we get a constant hazard of 0.14 and 0.54 respectively over a 15-day interval. We see in Figure 4 that the estimated hazard is at first bigger and then lower than what we would get with a constant probability of getting a hit. Notice the distinctly downward sloping hazard functions, which correspond to positive duration dependence. Finally, Figure 5 shows the simple histograms of durations between the violations. The top panel again shows the 1% VaRand the bottom panel shows the 5% VaR. Data and other resource constraints often force risk managers to backtest their models on a relatively limited backtesting samples. We therefore conduct our power experiment with samples sizes from 500 to 1500 days in increments of 250 days. Thus our backtesting samples correspond to approximately two through six years. BelowwesimulateGARCHreturns,calculate HSVaRs and the various test statistics over 1,000 Monte Carlo replications. The power of the tests are then simply calculated as the number of simulations, divided by 1000, in which the Monte Carlo p-value is smaller than the chosen level. The rejection frequencies are calculated at the 1%, 5% and 10% significance levels. In order to compute p-values we simulate N =9999hit sequence samples under the null hypothesis that the sequences are distributed i.i.d. Bernoulli(p). In order to make sure that we can calculate the test statistics, we do not use Monte Carlo samples with zero or one VaRviolations. 5 This of course constitutes a nontrivial sample selection rule for the smallest sample sizes and the 1% VaR coverage rate. As it is done for all the tests 5 The likelihood of the Weibull distribution can be unbounded when we have only one uncensored observation. When it happens we discard the sample. We get an unbounded likelihood for less than 3% of the draws when the 11

14 considered, the results are still comparable across tests. It also appears to be realistic that a risk management team would not start backtesting unless at least a couple of violations had occurred. The rejection frequencies below reflect this sample selection which is particularly important for the low (e.g. 1%) VaRcoverage rates and in the smallest samples (500 observations). 5.1 Results The results of the Monte Carlo simulations are presented in Tables 1 and 2. We report the empirical rejection frequencies (power) for the Markov, Weibull and EACD independence tests for various significance test levels, VaR coverage rates, and backtesting sample sizes. Table 1 reports power for a Historical Simulation Risk model with T e = 500 rolling estimation sample observations and Table 2 for T e =250rolling estimation sample observations. 6 The results are quite striking. The main result is that the Weibull test is virtually almost more powerful than the Markov and EACD tests in rejecting the HS risk models. This result holds across inference sample sizes, VaR coverage rates and significance levels chosen. The only twoexceptionsoccurintable1forasignificance level of 1%, a coverage rate of 5% and a sample of 500 where the EACD is better and in Table 1 for a significance level of 1%, a coverage rate of 1% and a sample of 500 where the Markov test is slightly better. The differences in power are sometimes very large. For example in Table 1 using a 1% significance level, the 5% VaR in a sample of 1,250 observations has a Weibull rejection frequency of 65.2% and a Markov rejection frequency of only 29.8%. The Weibull test clearly appears to pick up dependence in the hit violations which is ignored by the Markov test. The performance of the EACD test on the other hand is quite sporadic. It appears to do quite well at smaller sample sizes but relatively poorly at larger sample sizes. We suspect that the nonlinear estimate of the α parameter is poorly behaved. Note that the rejection frequencies for a given test are not always increasing in inference sample size. This is due to the sample selection procedure in which we discard samples with less than two violations. This sample selection is going to increase power ceteris paribus, and it is going to have the biggest effect in cells corresponding to the fewest average number of violations. These are of course the smallest sample sizes and the smallest coverage rate. Comparing rejection frequencies across coverage rates in Table 1 we also note that in small samples the power is sometimes higher for the 1% VaR coverage rate than for the corresponding coverage rate is 1% and the sample size is 500, and the probability is smaller than 0.5% for higher coverage rates and sample sizes. 6 We focus solely on the independence tests here because the historical simulation risk models under study are correctly specified unconditionally. 12

15 5% VaR. This may appear to be surprising as the hit sequences from the 5% VaRs contain many more violations which is the source of power. But the sample selection procedure will again have the largest effect for the 1% coverage rate and for the smallest samples. The selected samples for 1% coverage will thus tend to display more dependence on average than those selected for 5% coverage rate. Comparing numbers across Tables 1 and 2, we note that for a coverage rate of 1% the HS VaR with T e =500rolling sample observations always has a higher rejection frequency than the HS VaR with T e =250rolling sample observations. This result is interesting because practitioners often work very hard to expand their data bases enabling them to increase their rolling estimation sample period. Our results indicate that such efforts may be futile. When the return volatility process is very persistent, it is better to use a relatively short rolling estimation sample period. 6 Possible Extensions 6.1 The Monte Carlo Study Encouraged by the results in Tables 1-2, we now briefly outline some possible extensions to the Monte Carlo study: We only investigated one particular parameterization of the GARCH process above. It may be interesting to calculate the power of the test for processes with different volatility persistence, different degrees of conditional kurtosis and different leverage effects. One could also consider more elaborate data generating processes. Engle and Lee (1999) consider a component GARCH model which delivers long-memory like patterns in volatility. Hansen (1994) considers GARCH-t(v t )modelswherethedegreesoffreedom,ν t, varies over time in an autoregressive fashion. Structural breaks in the underlying return models, such as those investigated by Andreou and Ghysels (2002), may be of interest as well. Hamilton and Jorda (2002) have recently introduced a class of dynamic hazard models. Exploring these for the purpose of backtesting could be interesting. Finally, before even venturing into the backtesting of actual risk models it may be useful to conduct a more basic Monte Carlo analysis drawing violation sequences and duration data directly. Specifically, if the violation sequence is generated by a first-order Markov process, what is then the power of the different tests? Conversely, if the violation sequence is constructed from simulated duration data with dependence, then what would the power of the different tests be? 13

16 6.2 Backtesting Tail Density Forecasts The choice of Value-at-Risk as a portfolio risk measure can be criticized on several fronts. Most importantly,the quantilenatureofthevar implies that the shape of the return distribution to the left of the left is ignored. Particularly in portfolio s with highly nonlinear distributions, such as those including options, this shortcoming can be crucial. Theoreticians have criticized the VaR measure both from a utility-theoretic perspective (Artzner et al, 1999) and from a dynamic trading perspective (Basak and Shapiro, 2000). Although some of these criticisms have recently been challenged (Cuoco, He, and Issaenko, 2001), it is safe to say that risk managers ought to be interested in knowing the entire distribution of returns, and in particular the left tail. Backtesting distributions rather than VaRs then becomes important. Consider the standard density forecast evaluation approach 7 of calculating the uniform transform variable U t = F t (R t ) where F t ( ) is the a priori density forecast for time t. forecast is optimal corresponds to The null hypothesis that the density U t i.i.d. Uniform(0, 1) Berkowitz (2001) argues that the bounded support of the uniform variable renders standard inference difficult. One is forced to rely on nonparametric tests which have notoriously poor small sample properties. He suggests a simple transformation using the inverse normal c.d.f. Z t = Φ 1 (U t ) after which the hypothesis Z t i.i.d. Normal(0, 1) can easily be tested. Berkowitz further argues that confining attention to the left tail of the distribution has particular merit in the backtesting of risk models where the left tail contains the largest losses, which are most likely to impose bankruptcy risk. He defines the censored variable Z t = ( Zt,ifR t <VaR t Φ 1 (VaR t ),else and tests the null that Z t Censored Normal(0, 1,VaR t ) 7 See for example Diebold, Gunther and Tay (1998). 14

17 We note first that Berkowitz (2001) only tests the unconditional distribution of Zt. The information in the potential clustering of the VaRexceedences is ignored. Second, note that the censored variable complication is not needed. If we want to test that thetransformsofthep100 largest losses are themselves uniform, then we can simply multiply the subset of the uniform by 1/p, apply the transformation and test for standard normality again. 8 That is We then have that Z i ( Ui Ut /p, ifr t <VaR t = Else not defined = Φ 1 (U i ) i.i.d. Normal(0, 1) Note that due to the censoring there is no notion of time in the sequence Zi. We might want to make a joint analysis of both Zi and the duration between violations D i. Todothis we would like to write a joint density for these two processes under the alternative. We know that under the null hypothesis that the risk model is correctly specified the Zi should be i.i.d. N(0, 1), D i should be i.i.d. exponential with mean 1/p, and the processes should be independent. The question is how to write a joint density for these two processes as the alternative hypothesis knowing that, for example, the marginal p.d.f. of D i is a Weibull and some other p.d.f. for Zi? Copulas provide a useful tool for doing so. A (bivariate) copula is a function C from [0; 1] [0; 1] to [0; 1] with the following properties: 1. For every u, v in [0; 1], and C(u, 0) = 0 = C(0,v) C(u, 1) = u and C(1,v) = v; 2. For every u 1,u 2,v 1,v 2 in [0; 1] such that u 1 u 2 and v 1 v 2, C(u 2,v 2 ) C(u 2,v 1 ) C(u 1,v 2 )+C(u 1,v 1 ) 0 In order to explain how copulas can be used we apply Sklar s theorem (Nelsen, 1998), which states: Let H be a joint distribution function with margins F and G. Then there exists a copula C such that for all x, y in R, H(x, y) =C(F (x),g(y)). If F and G are continuous then C is unique. Conversely, if C is a copula and F and G are distribution functions then H is a joint distribution function with marginal densities F and G. 8 We are grateful to Nour Meddahi for pointing this out. 15

18 So if we have two densities under the alternative (e.g. f(d i ) and g(zi )) thenwecaneasily construct a joint density by applying a copula. Suppose the considered bivariate copula C(u, v; θ) isafunctionofauniqueparameterθ andthatwehavec(u, v; θ 0 )=uv and C(u, v; θ) 6= uv for θ 6= θ 0. This gives us a basis for a test because C(F (x),g(y); θ 0 )=F(x)G(y) means that x and y are independent. An example of such a copula is the Ali-Mikhail-Haq family of copulas where C(u, v; θ) = uv ; θ [ 1, 1] 1 θ(1 u)(1 v) and we have C(u, v; θ) =uv if θ =0. A possible alternative hypothesis could be that D i is i.i.d. Weibull(a, b), Zi is i.i.d. N(µ, σ 2 ) and C(u, v; θ) is from the Ali-Mikhail-Haq family of copulas. We could then test H 0 : a = p, b =1,µ=0,σ=1,θ=0 H 1 : atleastoneoftheseequalitiesdoesnothold in a likelihood ratio framework similar to the one considered for the VaRtests above. We plan to the pursue the implementation of such tests in future work. 7 Summary We have presented a new set of procedures for backtesting risk models. The chief insight is that if the VaR model is correctly specified for coverage rate, p, then the conditional expected duration between violations should be a constant 1/p days. We suggest various ways of testing this null hypothesis and we conduct a Monte Carlo analysis which compares the new tests to those currently available. Our results show that in many of the situations we consider, the duration based tests have much better power properties than the previously suggested tests. The size of the tests is easily controlled through finite sample p-values, which we calculate using Monte Carlo simulation. The immediate potential extensions to our Monte Carlo results are many. We could consider alternative data generating processes for returns and alternative risk models. Allowing for realistic nonstationarities such as structural breaks in the return process could be interesting as well. The majority of financial institutions use VaR as a risk measure, and many calculate VaR using the so-called Historical Simulation approach. While the main focus of our paper has thus been backtesting VaRs from Historical Simulation, we also suggest extensions to density and density tail backtesting. 16

19 References [1] Andreou and Ghysels (2002), Quality Control for Value at Risk: Monitoring Disruptions in the Distribution of Risk Exposure, Manuscript, University of North Carolina. [2] Artzner, P., F. Delbaen, J.-M. Eber and D. Heath (1999), Coherent Measures of Risk, Mathematical Finance, 9, [3] Barone-Adesi, G., K. Giannopoulos and L. Vosper (2000), Backtesting Derivative Portfolios with FHS, Manuscript, USI and City Business School. [4] Basak, S. and A. Shapiro (2000), Value at Risk Based Risk Management: Optimal Policies and Asset Prices, Review of Financial Studies, 14, [5] Basle Committee on Banking Supervision (1996), Amendment to the Capital Accord to Incorporate Market Risks. Basle. [6] Beder, T. (1995), VaR: Seductive but Dangerous, Financial Analysts Journal, September- October, [7] Berkowitz, J. (2001), Testing Density Forecasts, Applications to Risk Management Journal of Business and Economic Statistics, 19, [8] Berkowitz, J. and J. O Brien (2002), How Accurate are the Value-at-Risk Models at Commercial Banks? Journal of Finance, 57, [9] Christoffersen, P. (1998), Evaluating Interval Forecasts, International Economic Review, 39, [10] Christoffersen, P. (2002), Elements of Financial Risk Management, Academic Press, Forthcoming. [11] Christoffersen, P., J. Hahn and A. Inoue (2001), Testing and Comparing Value-at-Risk Measures Journal of Empirical Finance, 8, [12] Cuoco, D., H. He, and S. Issaenko (2001), Optimal Dynamic Trading Strategies with Risk Limits, Manuscript, Yale University. [13] Diebold, F.X., T. Gunther, and A. Tay (1998), Evaluating Density Forecasts, with Applications to Financial Risk Management, International Economic Review, 39,

20 [14] Dufour, J.-M. (2000), Monte Carlo Tests with Nuisance Parameters : A General Approach to Finite-Sample Inference and Nonstandard Asymptotics in Econometrics, Manuscript, Université de Montréal. [15] Engle, R. and G.J. Lee (1999), A Permanent and Transitory Component Model of Stock Return Volatility, in ed. R. Engle and H. White Cointegration, Causality, and Forecasting: A Festschrift in Honor of Clive W.J. Granger, Oxford University Press, [16] Engle, R. and J. Russel (1998), Autoregressive Conditional Duration: A New Model for Irregularly Spaced Transaction Data, Econometrica, 66, [17] Gourieroux, C. (2000) Econometrics of Qualitative Dependent Variables. Translated by Paul B. Klassen. Cambridge University Press. [18] Hamilton, J. and O. Jorda (2002), A Model for the Federal Funds Rate Target, Journal of Political Economy. Forthcoming. [19] Hansen, B. (1994), Autoregressive Conditional Density Estimation, International Economic Review, 35, [20] Hendricks, D. (1996), Evaluation of Value-at-Risk Models Using Historical Data, Economic Policy Review, Federal Reserve Bank of New York, April, [21] Kiefer, N. (1988), Economic Duration Data and Hazard Functions, Journal of Economic Literature, 26, [22] Kupiec, P. (1995), Techniques for Verifying the Accuracy of Risk Measurement Models, Journal of Derivatives, 3, [23] Jorion, P. (2000), Value-at-Risk: The New Benchmark for Controlling Financial Risk. Chicago: McGraw-Hlill. [24] Nelsen, R.(1998), An Introduction to Copulas, Lectures Notes in Statistics, 139, Springer Verlag. [25] Poirier, D. (1995), Intermediate Statistics and Econometrics: A Comparative Approach. Cambridge, MA: MIT Press. [26] Pritsker, M. (1997), Evaluating Value at Risk Methodologies: Accuracy versus Computational Time, Journal of Financial Services Research, [27] Pritsker, M. (2001), The Hidden Dangers of Historical Simulation, Manuscript, Federal Reserve Board. 18

21 Figure 1 Value-at-Risk Exceedences From Six Major Commercial Banks Berkowitz and O Brien (2002) 19

22 Figure 2 GARCH-t(d) Simulated Portfolio Returns with 1% and 5% Value-at-Risk from Historical Simulation with T e =

23 Figure 3 GARCH-t(d) Simulated Portfolio Returns with Exeedences of 1% VaRs from Historical Simulation with T e =500 Reported in Standard Deviations of Returns

24 Figure 4 Hazard Functions of Duration between VaRViolations GARCH-t(d) Portfolio Returns Historical Simulation Risk Model with T e = % Value-at-Risk % Value-at-Risk

25 Figure 5 Histograms of Duration between VaRViolations GARCH-t(d) Portfolio Returns Historical Simulation Risk Model with T e = % Value-at-Risk % Value-at-Risk

26 Table 1: Empirical Power in Independence Tests: Historical Simulation with T e = 500 Significance Level: 1% Significance Level: 5% Significance Level: 10% Coverage Rate: 1% Coverage Rate: 1% Coverage Rate: 1% Test: Markov Weibull EACD Test: Markov Weibull EACD Test: Markov Weibull EACD Sample size Sample size Sample size Coverage Rate: 5% Coverage Rate: 5% Coverage Rate: 5% Test: Markov Weibull EACD Test: Markov Weibull EACD Test: Markov Weibull EACD Sample size Sample size Sample size

27 Table 2: Empirical Power in Independence Tests: Historical Simulation with T e = 250 Significance Level: 1% Significance Level: 5% Significance Level: 10% Coverage Rate: 1% Coverage Rate: 1% Coverage Rate: 1% Test: Markov Weibull EACD Test: Markov Weibull EACD Test: Markov Weibull EACD Sample size Sample size Sample size Coverage Rate: 5% Coverage Rate: 5% Coverage Rate: 5% Test: Markov Weibull EACD Test: Markov Weibull EACD Test: Markov Weibull EACD Sample size Sample size Sample size

Backtesting Value-at-Risk: A Duration-Based Approach 1

Backtesting Value-at-Risk: A Duration-Based Approach 1 Backtesting Value-at-Risk: A Duration-Based Approach 1 Peter Christoffersen 2 McGill University, CIRANO and CIREQ Denis Pelletier 3 North Carolina State University October 24, 2003 1 The first author acknowledges

More information

Backtesting value-at-risk: a comparison between filtered bootstrap and historical simulation

Backtesting value-at-risk: a comparison between filtered bootstrap and historical simulation Journal of Risk Model Validation Volume /Number, Winter 1/13 (3 1) Backtesting value-at-risk: a comparison between filtered bootstrap and historical simulation Dario Brandolini Symphonia SGR, Via Gramsci

More information

How Accurate are Value-at-Risk Models at Commercial Banks?

How Accurate are Value-at-Risk Models at Commercial Banks? How Accurate are Value-at-Risk Models at Commercial Banks? Jeremy Berkowitz* Graduate School of Management University of California, Irvine James O Brien Division of Research and Statistics Federal Reserve

More information

Experience with the Weighted Bootstrap in Testing for Unobserved Heterogeneity in Exponential and Weibull Duration Models

Experience with the Weighted Bootstrap in Testing for Unobserved Heterogeneity in Exponential and Weibull Duration Models Experience with the Weighted Bootstrap in Testing for Unobserved Heterogeneity in Exponential and Weibull Duration Models Jin Seo Cho, Ta Ul Cheong, Halbert White Abstract We study the properties of the

More information

GMM for Discrete Choice Models: A Capital Accumulation Application

GMM for Discrete Choice Models: A Capital Accumulation Application GMM for Discrete Choice Models: A Capital Accumulation Application Russell Cooper, John Haltiwanger and Jonathan Willis January 2005 Abstract This paper studies capital adjustment costs. Our goal here

More information

Value at risk might underestimate risk when risk bites. Just bootstrap it!

Value at risk might underestimate risk when risk bites. Just bootstrap it! 23 September 215 by Zhili Cao Research & Investment Strategy at risk might underestimate risk when risk bites. Just bootstrap it! Key points at Risk (VaR) is one of the most widely used statistical tools

More information

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is:

**BEGINNING OF EXAMINATION** A random sample of five observations from a population is: **BEGINNING OF EXAMINATION** 1. You are given: (i) A random sample of five observations from a population is: 0.2 0.7 0.9 1.1 1.3 (ii) You use the Kolmogorov-Smirnov test for testing the null hypothesis,

More information

Analyzing Oil Futures with a Dynamic Nelson-Siegel Model

Analyzing Oil Futures with a Dynamic Nelson-Siegel Model Analyzing Oil Futures with a Dynamic Nelson-Siegel Model NIELS STRANGE HANSEN & ASGER LUNDE DEPARTMENT OF ECONOMICS AND BUSINESS, BUSINESS AND SOCIAL SCIENCES, AARHUS UNIVERSITY AND CENTER FOR RESEARCH

More information

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS Answer any FOUR of the SIX questions.

More information

Bayesian Estimation of the Markov-Switching GARCH(1,1) Model with Student-t Innovations

Bayesian Estimation of the Markov-Switching GARCH(1,1) Model with Student-t Innovations Bayesian Estimation of the Markov-Switching GARCH(1,1) Model with Student-t Innovations Department of Quantitative Economics, Switzerland david.ardia@unifr.ch R/Rmetrics User and Developer Workshop, Meielisalp,

More information

Financial Econometrics

Financial Econometrics Financial Econometrics Volatility Gerald P. Dwyer Trinity College, Dublin January 2013 GPD (TCD) Volatility 01/13 1 / 37 Squared log returns for CRSP daily GPD (TCD) Volatility 01/13 2 / 37 Absolute value

More information

Exam 2 Spring 2015 Statistics for Applications 4/9/2015

Exam 2 Spring 2015 Statistics for Applications 4/9/2015 18.443 Exam 2 Spring 2015 Statistics for Applications 4/9/2015 1. True or False (and state why). (a). The significance level of a statistical test is not equal to the probability that the null hypothesis

More information

Course information FN3142 Quantitative finance

Course information FN3142 Quantitative finance Course information 015 16 FN314 Quantitative finance This course is aimed at students interested in obtaining a thorough grounding in market finance and related empirical methods. Prerequisite If taken

More information

Equity, Vacancy, and Time to Sale in Real Estate.

Equity, Vacancy, and Time to Sale in Real Estate. Title: Author: Address: E-Mail: Equity, Vacancy, and Time to Sale in Real Estate. Thomas W. Zuehlke Department of Economics Florida State University Tallahassee, Florida 32306 U.S.A. tzuehlke@mailer.fsu.edu

More information

Margin Backtesting. August 31st, Abstract

Margin Backtesting. August 31st, Abstract Margin Backtesting Christophe Hurlin Christophe Pérignon August 31st, 2011 Abstract This paper presents a validation framework for collateral requirements or margins on a derivatives exchange. It can be

More information

Backtesting Trading Book Models

Backtesting Trading Book Models Backtesting Trading Book Models Using Estimates of VaR Expected Shortfall and Realized p-values Alexander J. McNeil 1 1 Heriot-Watt University Edinburgh ETH Risk Day 11 September 2015 AJM (HWU) Backtesting

More information

IEOR E4602: Quantitative Risk Management

IEOR E4602: Quantitative Risk Management IEOR E4602: Quantitative Risk Management Basic Concepts and Techniques of Risk Management Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Backtesting value-at-risk: Case study on the Romanian capital market

Backtesting value-at-risk: Case study on the Romanian capital market Available online at www.sciencedirect.com Procedia - Social and Behavioral Sciences 62 ( 2012 ) 796 800 WC-BEM 2012 Backtesting value-at-risk: Case study on the Romanian capital market Filip Iorgulescu

More information

Experience with the Weighted Bootstrap in Testing for Unobserved Heterogeneity in Exponential and Weibull Duration Models

Experience with the Weighted Bootstrap in Testing for Unobserved Heterogeneity in Exponential and Weibull Duration Models Experience with the Weighted Bootstrap in Testing for Unobserved Heterogeneity in Exponential and Weibull Duration Models Jin Seo Cho, Ta Ul Cheong, Halbert White Abstract We study the properties of the

More information

Amath 546/Econ 589 Univariate GARCH Models: Advanced Topics

Amath 546/Econ 589 Univariate GARCH Models: Advanced Topics Amath 546/Econ 589 Univariate GARCH Models: Advanced Topics Eric Zivot April 29, 2013 Lecture Outline The Leverage Effect Asymmetric GARCH Models Forecasts from Asymmetric GARCH Models GARCH Models with

More information

Dependence Structure and Extreme Comovements in International Equity and Bond Markets

Dependence Structure and Extreme Comovements in International Equity and Bond Markets Dependence Structure and Extreme Comovements in International Equity and Bond Markets René Garcia Edhec Business School, Université de Montréal, CIRANO and CIREQ Georges Tsafack Suffolk University Measuring

More information

Violation duration as a better way of VaR model evaluation : evidence from Turkish market portfolio

Violation duration as a better way of VaR model evaluation : evidence from Turkish market portfolio MPRA Munich Personal RePEc Archive Violation duration as a better way of VaR model evaluation : evidence from Turkish market portfolio Ekrem Kilic Finecus Financial Software and Consultancy 1. May 2006

More information

Chapter 7: Point Estimation and Sampling Distributions

Chapter 7: Point Estimation and Sampling Distributions Chapter 7: Point Estimation and Sampling Distributions Seungchul Baek Department of Statistics, University of South Carolina STAT 509: Statistics for Engineers 1 / 20 Motivation In chapter 3, we learned

More information

CAN LOGNORMAL, WEIBULL OR GAMMA DISTRIBUTIONS IMPROVE THE EWS-GARCH VALUE-AT-RISK FORECASTS?

CAN LOGNORMAL, WEIBULL OR GAMMA DISTRIBUTIONS IMPROVE THE EWS-GARCH VALUE-AT-RISK FORECASTS? PRZEGL D STATYSTYCZNY R. LXIII ZESZYT 3 2016 MARCIN CHLEBUS 1 CAN LOGNORMAL, WEIBULL OR GAMMA DISTRIBUTIONS IMPROVE THE EWS-GARCH VALUE-AT-RISK FORECASTS? 1. INTRODUCTION International regulations established

More information

Much of what appears here comes from ideas presented in the book:

Much of what appears here comes from ideas presented in the book: Chapter 11 Robust statistical methods Much of what appears here comes from ideas presented in the book: Huber, Peter J. (1981), Robust statistics, John Wiley & Sons (New York; Chichester). There are many

More information

Scaling conditional tail probability and quantile estimators

Scaling conditional tail probability and quantile estimators Scaling conditional tail probability and quantile estimators JOHN COTTER a a Centre for Financial Markets, Smurfit School of Business, University College Dublin, Carysfort Avenue, Blackrock, Co. Dublin,

More information

ROM SIMULATION Exact Moment Simulation using Random Orthogonal Matrices

ROM SIMULATION Exact Moment Simulation using Random Orthogonal Matrices ROM SIMULATION Exact Moment Simulation using Random Orthogonal Matrices Bachelier Finance Society Meeting Toronto 2010 Henley Business School at Reading Contact Author : d.ledermann@icmacentre.ac.uk Alexander

More information

Business Statistics 41000: Probability 3

Business Statistics 41000: Probability 3 Business Statistics 41000: Probability 3 Drew D. Creal University of Chicago, Booth School of Business February 7 and 8, 2014 1 Class information Drew D. Creal Email: dcreal@chicagobooth.edu Office: 404

More information

Chapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29

Chapter 5 Univariate time-series analysis. () Chapter 5 Univariate time-series analysis 1 / 29 Chapter 5 Univariate time-series analysis () Chapter 5 Univariate time-series analysis 1 / 29 Time-Series Time-series is a sequence fx 1, x 2,..., x T g or fx t g, t = 1,..., T, where t is an index denoting

More information

Lecture 5: Fundamentals of Statistical Analysis and Distributions Derived from Normal Distributions

Lecture 5: Fundamentals of Statistical Analysis and Distributions Derived from Normal Distributions Lecture 5: Fundamentals of Statistical Analysis and Distributions Derived from Normal Distributions ELE 525: Random Processes in Information Systems Hisashi Kobayashi Department of Electrical Engineering

More information

MEASURING PORTFOLIO RISKS USING CONDITIONAL COPULA-AR-GARCH MODEL

MEASURING PORTFOLIO RISKS USING CONDITIONAL COPULA-AR-GARCH MODEL MEASURING PORTFOLIO RISKS USING CONDITIONAL COPULA-AR-GARCH MODEL Isariya Suttakulpiboon MSc in Risk Management and Insurance Georgia State University, 30303 Atlanta, Georgia Email: suttakul.i@gmail.com,

More information

Introduction to Algorithmic Trading Strategies Lecture 8

Introduction to Algorithmic Trading Strategies Lecture 8 Introduction to Algorithmic Trading Strategies Lecture 8 Risk Management Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com Outline Value at Risk (VaR) Extreme Value Theory (EVT) References

More information

Week 7 Quantitative Analysis of Financial Markets Simulation Methods

Week 7 Quantitative Analysis of Financial Markets Simulation Methods Week 7 Quantitative Analysis of Financial Markets Simulation Methods Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 November

More information

Financial Econometrics Notes. Kevin Sheppard University of Oxford

Financial Econometrics Notes. Kevin Sheppard University of Oxford Financial Econometrics Notes Kevin Sheppard University of Oxford Monday 15 th January, 2018 2 This version: 22:52, Monday 15 th January, 2018 2018 Kevin Sheppard ii Contents 1 Probability, Random Variables

More information

Homework Problems Stat 479

Homework Problems Stat 479 Chapter 10 91. * A random sample, X1, X2,, Xn, is drawn from a distribution with a mean of 2/3 and a variance of 1/18. ˆ = (X1 + X2 + + Xn)/(n-1) is the estimator of the distribution mean θ. Find MSE(

More information

Assicurazioni Generali: An Option Pricing Case with NAGARCH

Assicurazioni Generali: An Option Pricing Case with NAGARCH Assicurazioni Generali: An Option Pricing Case with NAGARCH Assicurazioni Generali: Business Snapshot Find our latest analyses and trade ideas on bsic.it Assicurazioni Generali SpA is an Italy-based insurance

More information

Statistical Methods in Financial Risk Management

Statistical Methods in Financial Risk Management Statistical Methods in Financial Risk Management Lecture 1: Mapping Risks to Risk Factors Alexander J. McNeil Maxwell Institute of Mathematical Sciences Heriot-Watt University Edinburgh 2nd Workshop on

More information

Evaluating the Accuracy of Value at Risk Approaches

Evaluating the Accuracy of Value at Risk Approaches Evaluating the Accuracy of Value at Risk Approaches Kyle McAndrews April 25, 2015 1 Introduction Risk management is crucial to the financial industry, and it is particularly relevant today after the turmoil

More information

Introduction Dickey-Fuller Test Option Pricing Bootstrapping. Simulation Methods. Chapter 13 of Chris Brook s Book.

Introduction Dickey-Fuller Test Option Pricing Bootstrapping. Simulation Methods. Chapter 13 of Chris Brook s Book. Simulation Methods Chapter 13 of Chris Brook s Book Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 April 26, 2017 Christopher

More information

Financial Risk Forecasting Chapter 9 Extreme Value Theory

Financial Risk Forecasting Chapter 9 Extreme Value Theory Financial Risk Forecasting Chapter 9 Extreme Value Theory Jon Danielsson 2017 London School of Economics To accompany Financial Risk Forecasting www.financialriskforecasting.com Published by Wiley 2011

More information

A Test of the Normality Assumption in the Ordered Probit Model *

A Test of the Normality Assumption in the Ordered Probit Model * A Test of the Normality Assumption in the Ordered Probit Model * Paul A. Johnson Working Paper No. 34 March 1996 * Assistant Professor, Vassar College. I thank Jahyeong Koo, Jim Ziliak and an anonymous

More information

Box-Cox Transforms for Realized Volatility

Box-Cox Transforms for Realized Volatility Box-Cox Transforms for Realized Volatility Sílvia Gonçalves and Nour Meddahi Université de Montréal and Imperial College London January 1, 8 Abstract The log transformation of realized volatility is often

More information

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics Unit 5: Sampling Distributions of Statistics Statistics 571: Statistical Methods Ramón V. León 6/12/2004 Unit 5 - Stat 571 - Ramon V. Leon 1 Definitions and Key Concepts A sample statistic used to estimate

More information

The University of Chicago, Booth School of Business Business 41202, Spring Quarter 2009, Mr. Ruey S. Tsay. Solutions to Final Exam

The University of Chicago, Booth School of Business Business 41202, Spring Quarter 2009, Mr. Ruey S. Tsay. Solutions to Final Exam The University of Chicago, Booth School of Business Business 41202, Spring Quarter 2009, Mr. Ruey S. Tsay Solutions to Final Exam Problem A: (42 pts) Answer briefly the following questions. 1. Questions

More information

Unit 5: Sampling Distributions of Statistics

Unit 5: Sampling Distributions of Statistics Unit 5: Sampling Distributions of Statistics Statistics 571: Statistical Methods Ramón V. León 6/12/2004 Unit 5 - Stat 571 - Ramon V. Leon 1 Definitions and Key Concepts A sample statistic used to estimate

More information

ADVANCED OPERATIONAL RISK MODELLING IN BANKS AND INSURANCE COMPANIES

ADVANCED OPERATIONAL RISK MODELLING IN BANKS AND INSURANCE COMPANIES Small business banking and financing: a global perspective Cagliari, 25-26 May 2007 ADVANCED OPERATIONAL RISK MODELLING IN BANKS AND INSURANCE COMPANIES C. Angela, R. Bisignani, G. Masala, M. Micocci 1

More information

PERFORMANCE AND CONSERVATISM OF MONTHLY FHS VAR: AN INTERNATIONAL INVESTIGATION

PERFORMANCE AND CONSERVATISM OF MONTHLY FHS VAR: AN INTERNATIONAL INVESTIGATION PERFORMANCE AND CONSERVATISM OF MONTHLY FHS VAR: AN INTERNATIONAL INVESTIGATION Stéphane Chrétien Frank Coggins * This draft: September 008 Abstract This study examines sixteen models of monthly Value-at-Risk

More information

Internet Appendix for Asymmetry in Stock Comovements: An Entropy Approach

Internet Appendix for Asymmetry in Stock Comovements: An Entropy Approach Internet Appendix for Asymmetry in Stock Comovements: An Entropy Approach Lei Jiang Tsinghua University Ke Wu Renmin University of China Guofu Zhou Washington University in St. Louis August 2017 Jiang,

More information

Absolute Return Volatility. JOHN COTTER* University College Dublin

Absolute Return Volatility. JOHN COTTER* University College Dublin Absolute Return Volatility JOHN COTTER* University College Dublin Address for Correspondence: Dr. John Cotter, Director of the Centre for Financial Markets, Department of Banking and Finance, University

More information

Which GARCH Model for Option Valuation? By Peter Christoffersen and Kris Jacobs

Which GARCH Model for Option Valuation? By Peter Christoffersen and Kris Jacobs Online Appendix Sample Index Returns Which GARCH Model for Option Valuation? By Peter Christoffersen and Kris Jacobs In order to give an idea of the differences in returns over the sample, Figure A.1 plots

More information

Lecture 5: Univariate Volatility

Lecture 5: Univariate Volatility Lecture 5: Univariate Volatility Modellig, ARCH and GARCH Prof. Massimo Guidolin 20192 Financial Econometrics Spring 2015 Overview Stepwise Distribution Modeling Approach Three Key Facts to Remember Volatility

More information

Estimation Risk in Financial Risk Management

Estimation Risk in Financial Risk Management Estimation Risk in Financial Risk Management Peter Christoffersen McGill University and CIRANO Sílvia Gonçalves Université de Montréal and CIRANO December 6, 24 Abstract Value-at-Risk (VaR) is increasingly

More information

THE INFORMATION CONTENT OF IMPLIED VOLATILITY IN AGRICULTURAL COMMODITY MARKETS. Pierre Giot 1

THE INFORMATION CONTENT OF IMPLIED VOLATILITY IN AGRICULTURAL COMMODITY MARKETS. Pierre Giot 1 THE INFORMATION CONTENT OF IMPLIED VOLATILITY IN AGRICULTURAL COMMODITY MARKETS Pierre Giot 1 May 2002 Abstract In this paper we compare the incremental information content of lagged implied volatility

More information

High-Frequency Data Analysis and Market Microstructure [Tsay (2005), chapter 5]

High-Frequency Data Analysis and Market Microstructure [Tsay (2005), chapter 5] 1 High-Frequency Data Analysis and Market Microstructure [Tsay (2005), chapter 5] High-frequency data have some unique characteristics that do not appear in lower frequencies. At this class we have: Nonsynchronous

More information

Assessing Regime Switching Equity Return Models

Assessing Regime Switching Equity Return Models Assessing Regime Switching Equity Return Models R. Keith Freeland, ASA, Ph.D. Mary R. Hardy, FSA, FIA, CERA, Ph.D. Matthew Till Copyright 2009 by the Society of Actuaries. All rights reserved by the Society

More information

MEASURING TRADED MARKET RISK: VALUE-AT-RISK AND BACKTESTING TECHNIQUES

MEASURING TRADED MARKET RISK: VALUE-AT-RISK AND BACKTESTING TECHNIQUES MEASURING TRADED MARKET RISK: VALUE-AT-RISK AND BACKTESTING TECHNIQUES Colleen Cassidy and Marianne Gizycki Research Discussion Paper 9708 November 1997 Bank Supervision Department Reserve Bank of Australia

More information

ABILITY OF VALUE AT RISK TO ESTIMATE THE RISK: HISTORICAL SIMULATION APPROACH

ABILITY OF VALUE AT RISK TO ESTIMATE THE RISK: HISTORICAL SIMULATION APPROACH ABILITY OF VALUE AT RISK TO ESTIMATE THE RISK: HISTORICAL SIMULATION APPROACH Dumitru Cristian Oanea, PhD Candidate, Bucharest University of Economic Studies Abstract: Each time an investor is investing

More information

OPTIMAL PORTFOLIO OF THE GOVERNMENT PENSION INVESTMENT FUND BASED ON THE SYSTEMIC RISK EVALUATED BY A NEW ASYMMETRIC COPULA

OPTIMAL PORTFOLIO OF THE GOVERNMENT PENSION INVESTMENT FUND BASED ON THE SYSTEMIC RISK EVALUATED BY A NEW ASYMMETRIC COPULA Advances in Science, Technology and Environmentology Special Issue on the Financial & Pension Mathematical Science Vol. B13 (2016.3), 21 38 OPTIMAL PORTFOLIO OF THE GOVERNMENT PENSION INVESTMENT FUND BASED

More information

Cross-Sectional Distribution of GARCH Coefficients across S&P 500 Constituents : Time-Variation over the Period

Cross-Sectional Distribution of GARCH Coefficients across S&P 500 Constituents : Time-Variation over the Period Cahier de recherche/working Paper 13-13 Cross-Sectional Distribution of GARCH Coefficients across S&P 500 Constituents : Time-Variation over the Period 2000-2012 David Ardia Lennart F. Hoogerheide Mai/May

More information

Assessing Value-at-Risk

Assessing Value-at-Risk Lecture notes on risk management, public policy, and the financial system Allan M. Malz Columbia University 2018 Allan M. Malz Last updated: April 1, 2018 2 / 18 Outline 3/18 Overview Unconditional coverage

More information

Non-informative Priors Multiparameter Models

Non-informative Priors Multiparameter Models Non-informative Priors Multiparameter Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Prior Types Informative vs Non-informative There has been a desire for a prior distributions that

More information

12. Conditional heteroscedastic models (ARCH) MA6622, Ernesto Mordecki, CityU, HK, 2006.

12. Conditional heteroscedastic models (ARCH) MA6622, Ernesto Mordecki, CityU, HK, 2006. 12. Conditional heteroscedastic models (ARCH) MA6622, Ernesto Mordecki, CityU, HK, 2006. References for this Lecture: Robert F. Engle. Autoregressive Conditional Heteroscedasticity with Estimates of Variance

More information

The Hidden Dangers of Historical Simulation

The Hidden Dangers of Historical Simulation The Hidden Dangers of Historical Simulation Matthew Pritsker April 16, 2001 Abstract Many large financial institutions compute the Value-at-Risk (VaR) of their trading portfolios using historical simulation

More information

Regime Switching in the Presence of Endogeneity

Regime Switching in the Presence of Endogeneity ISSN 1440-771X Department of Econometrics and Business Statistics http://business.monash.edu/econometrics-and-businessstatistics/research/publications Regime Switching in the Presence of Endogeneity Tingting

More information

Sample Size for Assessing Agreement between Two Methods of Measurement by Bland Altman Method

Sample Size for Assessing Agreement between Two Methods of Measurement by Bland Altman Method Meng-Jie Lu 1 / Wei-Hua Zhong 1 / Yu-Xiu Liu 1 / Hua-Zhang Miao 1 / Yong-Chang Li 1 / Mu-Huo Ji 2 Sample Size for Assessing Agreement between Two Methods of Measurement by Bland Altman Method Abstract:

More information

Window Width Selection for L 2 Adjusted Quantile Regression

Window Width Selection for L 2 Adjusted Quantile Regression Window Width Selection for L 2 Adjusted Quantile Regression Yoonsuh Jung, The Ohio State University Steven N. MacEachern, The Ohio State University Yoonkyung Lee, The Ohio State University Technical Report

More information

Gamma Distribution Fitting

Gamma Distribution Fitting Chapter 552 Gamma Distribution Fitting Introduction This module fits the gamma probability distributions to a complete or censored set of individual or grouped data values. It outputs various statistics

More information

Operational Risk Aggregation

Operational Risk Aggregation Operational Risk Aggregation Professor Carol Alexander Chair of Risk Management and Director of Research, ISMA Centre, University of Reading, UK. Loss model approaches are currently a focus of operational

More information

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage

Point Estimation. Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage 6 Point Estimation Stat 4570/5570 Material from Devore s book (Ed 8), and Cengage Point Estimation Statistical inference: directed toward conclusions about one or more parameters. We will use the generic

More information

1. You are given the following information about a stationary AR(2) model:

1. You are given the following information about a stationary AR(2) model: Fall 2003 Society of Actuaries **BEGINNING OF EXAMINATION** 1. You are given the following information about a stationary AR(2) model: (i) ρ 1 = 05. (ii) ρ 2 = 01. Determine φ 2. (A) 0.2 (B) 0.1 (C) 0.4

More information

Assessing Regime Switching Equity Return Models

Assessing Regime Switching Equity Return Models Assessing Regime Switching Equity Return Models R. Keith Freeland Mary R Hardy Matthew Till January 28, 2009 In this paper we examine time series model selection and assessment based on residuals, with

More information

Section 3 describes the data for portfolio construction and alternative PD and correlation inputs.

Section 3 describes the data for portfolio construction and alternative PD and correlation inputs. Evaluating economic capital models for credit risk is important for both financial institutions and regulators. However, a major impediment to model validation remains limited data in the time series due

More information

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi

Chapter 4: Commonly Used Distributions. Statistics for Engineers and Scientists Fourth Edition William Navidi Chapter 4: Commonly Used Distributions Statistics for Engineers and Scientists Fourth Edition William Navidi 2014 by Education. This is proprietary material solely for authorized instructor use. Not authorized

More information

An Improved Skewness Measure

An Improved Skewness Measure An Improved Skewness Measure Richard A. Groeneveld Professor Emeritus, Department of Statistics Iowa State University ragroeneveld@valley.net Glen Meeden School of Statistics University of Minnesota Minneapolis,

More information

Modeling the Market Risk in the Context of the Basel III Acord

Modeling the Market Risk in the Context of the Basel III Acord Theoretical and Applied Economics Volume XVIII (2), No. (564), pp. 5-2 Modeling the Market Risk in the Context of the Basel III Acord Nicolae DARDAC Bucharest Academy of Economic Studies nicolae.dardac@fin.ase.ro

More information

Lecture 9: Markov and Regime

Lecture 9: Markov and Regime Lecture 9: Markov and Regime Switching Models Prof. Massimo Guidolin 20192 Financial Econometrics Spring 2017 Overview Motivation Deterministic vs. Endogeneous, Stochastic Switching Dummy Regressiom Switching

More information

Backtesting Trading Book Models

Backtesting Trading Book Models Backtesting Trading Book Models Using VaR Expected Shortfall and Realized p-values Alexander J. McNeil 1 1 Heriot-Watt University Edinburgh Vienna 10 June 2015 AJM (HWU) Backtesting and Elicitability QRM

More information

ECON Introductory Econometrics. Lecture 1: Introduction and Review of Statistics

ECON Introductory Econometrics. Lecture 1: Introduction and Review of Statistics ECON4150 - Introductory Econometrics Lecture 1: Introduction and Review of Statistics Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 1-2 Lecture outline 2 What is econometrics? Course

More information

An implicit backtest for ES via a simple multinomial approach

An implicit backtest for ES via a simple multinomial approach An implicit backtest for ES via a simple multinomial approach Marie KRATZ ESSEC Business School Paris Singapore Joint work with Yen H. LOK & Alexander McNEIL (Heriot Watt Univ., Edinburgh) Vth IBERIAN

More information

discussion Papers Some Flexible Parametric Models for Partially Adaptive Estimators of Econometric Models

discussion Papers Some Flexible Parametric Models for Partially Adaptive Estimators of Econometric Models discussion Papers Discussion Paper 2007-13 March 26, 2007 Some Flexible Parametric Models for Partially Adaptive Estimators of Econometric Models Christian B. Hansen Graduate School of Business at the

More information

ARCH and GARCH models

ARCH and GARCH models ARCH and GARCH models Fulvio Corsi SNS Pisa 5 Dic 2011 Fulvio Corsi ARCH and () GARCH models SNS Pisa 5 Dic 2011 1 / 21 Asset prices S&P 500 index from 1982 to 2009 1600 1400 1200 1000 800 600 400 200

More information

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions.

UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. UQ, STAT2201, 2017, Lectures 3 and 4 Unit 3 Probability Distributions. Random Variables 2 A random variable X is a numerical (integer, real, complex, vector etc.) summary of the outcome of the random experiment.

More information

A Regime Switching model

A Regime Switching model Master Degree Project in Finance A Regime Switching model Applied to the OMXS30 and Nikkei 225 indices Ludvig Hjalmarsson Supervisor: Mattias Sundén Master Degree Project No. 2014:92 Graduate School Masters

More information

Copula-Based Pairs Trading Strategy

Copula-Based Pairs Trading Strategy Copula-Based Pairs Trading Strategy Wenjun Xie and Yuan Wu Division of Banking and Finance, Nanyang Business School, Nanyang Technological University, Singapore ABSTRACT Pairs trading is a technique that

More information

Lecture 8: Markov and Regime

Lecture 8: Markov and Regime Lecture 8: Markov and Regime Switching Models Prof. Massimo Guidolin 20192 Financial Econometrics Spring 2016 Overview Motivation Deterministic vs. Endogeneous, Stochastic Switching Dummy Regressiom Switching

More information

Discussion Paper No. DP 07/05

Discussion Paper No. DP 07/05 SCHOOL OF ACCOUNTING, FINANCE AND MANAGEMENT Essex Finance Centre A Stochastic Variance Factor Model for Large Datasets and an Application to S&P data A. Cipollini University of Essex G. Kapetanios Queen

More information

FINITE SAMPLE DISTRIBUTIONS OF RISK-RETURN RATIOS

FINITE SAMPLE DISTRIBUTIONS OF RISK-RETURN RATIOS Available Online at ESci Journals Journal of Business and Finance ISSN: 305-185 (Online), 308-7714 (Print) http://www.escijournals.net/jbf FINITE SAMPLE DISTRIBUTIONS OF RISK-RETURN RATIOS Reza Habibi*

More information

Amath 546/Econ 589 Univariate GARCH Models

Amath 546/Econ 589 Univariate GARCH Models Amath 546/Econ 589 Univariate GARCH Models Eric Zivot April 24, 2013 Lecture Outline Conditional vs. Unconditional Risk Measures Empirical regularities of asset returns Engle s ARCH model Testing for ARCH

More information

Strategies for Improving the Efficiency of Monte-Carlo Methods

Strategies for Improving the Efficiency of Monte-Carlo Methods Strategies for Improving the Efficiency of Monte-Carlo Methods Paul J. Atzberger General comments or corrections should be sent to: paulatz@cims.nyu.edu Introduction The Monte-Carlo method is a useful

More information

Chapter 7: Estimation Sections

Chapter 7: Estimation Sections 1 / 40 Chapter 7: Estimation Sections 7.1 Statistical Inference Bayesian Methods: Chapter 7 7.2 Prior and Posterior Distributions 7.3 Conjugate Prior Distributions 7.4 Bayes Estimators Frequentist Methods:

More information

A potentially useful approach to model nonlinearities in time series is to assume different behavior (structural break) in different subsamples

A potentially useful approach to model nonlinearities in time series is to assume different behavior (structural break) in different subsamples 1.3 Regime switching models A potentially useful approach to model nonlinearities in time series is to assume different behavior (structural break) in different subsamples (or regimes). If the dates, the

More information

Market Risk Prediction under Long Memory: When VaR is Higher than Expected

Market Risk Prediction under Long Memory: When VaR is Higher than Expected Market Risk Prediction under Long Memory: When VaR is Higher than Expected Harald Kinateder Niklas Wagner DekaBank Chair in Finance and Financial Control Passau University 19th International AFIR Colloquium

More information

Idiosyncratic risk, insurance, and aggregate consumption dynamics: a likelihood perspective

Idiosyncratic risk, insurance, and aggregate consumption dynamics: a likelihood perspective Idiosyncratic risk, insurance, and aggregate consumption dynamics: a likelihood perspective Alisdair McKay Boston University June 2013 Microeconomic evidence on insurance - Consumption responds to idiosyncratic

More information

Dealing with Downside Risk in Energy Markets: Futures versus Exchange-Traded Funds. Panit Arunanondchai

Dealing with Downside Risk in Energy Markets: Futures versus Exchange-Traded Funds. Panit Arunanondchai Dealing with Downside Risk in Energy Markets: Futures versus Exchange-Traded Funds Panit Arunanondchai Ph.D. Candidate in Agribusiness and Managerial Economics Department of Agricultural Economics, Texas

More information

Forecasting Volatility movements using Markov Switching Regimes. This paper uses Markov switching models to capture volatility dynamics in exchange

Forecasting Volatility movements using Markov Switching Regimes. This paper uses Markov switching models to capture volatility dynamics in exchange Forecasting Volatility movements using Markov Switching Regimes George S. Parikakis a1, Theodore Syriopoulos b a Piraeus Bank, Corporate Division, 4 Amerikis Street, 10564 Athens Greece bdepartment of

More information

Market Risk Analysis Volume II. Practical Financial Econometrics

Market Risk Analysis Volume II. Practical Financial Econometrics Market Risk Analysis Volume II Practical Financial Econometrics Carol Alexander John Wiley & Sons, Ltd List of Figures List of Tables List of Examples Foreword Preface to Volume II xiii xvii xx xxii xxvi

More information

2. Copula Methods Background

2. Copula Methods Background 1. Introduction Stock futures markets provide a channel for stock holders potentially transfer risks. Effectiveness of such a hedging strategy relies heavily on the accuracy of hedge ratio estimation.

More information

FORECASTING PERFORMANCE OF MARKOV-SWITCHING GARCH MODELS: A LARGE-SCALE EMPIRICAL STUDY

FORECASTING PERFORMANCE OF MARKOV-SWITCHING GARCH MODELS: A LARGE-SCALE EMPIRICAL STUDY FORECASTING PERFORMANCE OF MARKOV-SWITCHING GARCH MODELS: A LARGE-SCALE EMPIRICAL STUDY Latest version available on SSRN https://ssrn.com/abstract=2918413 Keven Bluteau Kris Boudt Leopoldo Catania R/Finance

More information

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی

درس هفتم یادگیري ماشین. (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی یادگیري ماشین توزیع هاي نمونه و تخمین نقطه اي پارامترها Sampling Distributions and Point Estimation of Parameter (Machine Learning) دانشگاه فردوسی مشهد دانشکده مهندسی رضا منصفی درس هفتم 1 Outline Introduction

More information

Operational Risk Aggregation

Operational Risk Aggregation Operational Risk Aggregation Professor Carol Alexander Chair of Risk Management and Director of Research, ISMA Centre, University of Reading, UK. Loss model approaches are currently a focus of operational

More information