File Name: Multiple files (see chapter for details of example files used)
Location: Various places in Modeling Toolkit
Brief Description: Illustrates the use of several banking models to develop an integrated risk management paradigm for the Basel II/III Accords
Requirements: Modeling Toolkit, Risk Simulator
Modeling Toolkit Functions Used: MTProbabilityDefaultMertonImputedAssetValue, MTProbabilityDefaultMertonImputedAssetVolatility, MTProbabilityDefaultMertonII, MTProbabilityDefaultMertonDefaultDistance, MTProbabilityDefaultMertonRecoveryRate, MTProbabilityDefaultMertonMVDebt
With the Basel II/III Accords, internationally active banks are now allowed to compute their own risk capital requirements using the internal ratings-based (IRB) approach. Not only is adequate risk capital analysis important as a compliance obligation, but it also provides banks the ability to optimize their capital by computing and allocating risks, performing performance measurements, executing strategic decisions, increasing competitiveness, and enhancing profitability. This chapter discusses the various approaches required to implement an IRB method, and the step-by-step models and methodologies in implementing and valuing economic capital, value at risk, probability of default, and loss given default, the key ingredients required in an IRB approach, through the use of advanced analytics such as Monte Carlo and historical risk simulation, portfolio optimization, stochastic forecasting, and options analysis. The use of Risk Simulator and the Modeling Toolkit software in computing and calibrating these critical input parameters is illustrated. Instead of dwelling on theory or revamping what has already been written many times, this chapter focuses solely on the practical modeling applications of the key ingredients to the Basel II/III Accords. Specifically, these topics are addressed:
Please note that several white papers on related topics such as the following are available by request (send an e-mail request to admin@realoptionsvaluation.com):
To follow along with the analyses in this chapter, we assume that the reader already has Risk Simulator, Real Options SLS, and Modeling Toolkit installed and is somewhat familiar with the basic functions of each software. If not, please refer to www.realoptionsvaluation.com (click on the Download link) and watch the getting started videos, read some of the getting started case studies, or install the latest trial versions of these software programs. Alternatively, refer to the website to obtain a primer on using these software programs. Each topic discussed will start with some basic introduction to the methodologies that are appropriate, followed by some practical hands-on modeling approaches and examples.
The probability of default measures the degree of likelihood that the borrower of a loan or debt (the obligor) will be unable to make the necessary scheduled repayments on the debt, thereby defaulting on the debt. Should the obligor be unable to pay, the debt is in default, and the lenders of the debt have legal avenues to attempt a recovery of the debt, or at least partial repayment of the entire debt. The higher the default probability a lender estimates a borrower to have, the higher the interest rate the lender will charge the borrower as compensation for bearing the higher default risk.
The probability of default models are categorized as structural or empirical. Structural models look at a borrower’s ability to pay based on market data such as equity prices and market and book values of assets and liabilities, as well as the volatility of these variables. Hence, these structural models are used predominantly to estimate the probability of default of companies and countries, most applicable within the areas of commercial and industrial banking. In contrast, empirical models or credit scoring models are used to quantitatively determine the probability that a loan or loan holder will default, where the loan holder is an individual, by looking at historical portfolios of loans held and assessing individual characteristics (e.g., age, educational level, debt to income ratio, and so forth). This second approach is more applicable to the retail banking sector.
The probability of default models are models that assess the likelihood of default by an obligor. They differ from regular credit scoring models in several ways. First of all, credit scoring models usually are applied to smaller credits—individuals or small businesses—whereas default models are applied to larger credits—corporations or countries. Credit scoring models are largely statistical, regressing instances of default against various risk indicators, such as an obligor’s income, home renter or owner status, years at a job, educational level, debt to income ratio, and so forth, discussed later in this chapter. Structural default models, in contrast, directly model the default process and typically are calibrated to market variables, such as the obligor’s stock price, asset value, debt book value, or the credit spread on its bonds. Default models find many applications within financial institutions. They are used to support credit analysis and determine the probability that a firm will default, value counterparty credit risk limits, or apply financial engineering techniques in developing credit derivatives or other credit instruments.
The model illustrated in this chapter is used to solve the probability of default of a publicly-traded company with equity and debt holdings, and accounting for its volatilities in the market (Figure 92.1). This model is currently used by KMV and Moody’s to perform credit risk analysis. This approach assumes that the book value of asset and asset volatility are unknown and solved in the model; that the company is relatively stable; and that the growth rate of the company’s assets is stable over time (e.g., not in start-up mode). The model uses several simultaneous equations in options valuation coupled with optimization to obtain the implied underlying asset’s market value and volatility of the asset in order to compute the probability of default and distance to default for the firm.
It is assumed that the reader is well versed in running simulations and optimizations in Risk Simulator. The example model used is the Probability of Default – External Options Model and can be accessed through Modeling Toolkit | Prob of Default | External Options Model (Public Company).
To run this model (Figure 92.1), enter in the required inputs:
The comparable option parameters are shown in cells G18 to G23. All these comparable inputs are computed except for Asset Value (the market value of assets) and the Volatility of Asset. You will need to input some rough estimates as a starting point so that the analysis can be run. The rule of thumb is to set the volatility of the asset in G22 to be one-fifth to half of the volatility of equity computed in G10, and the market value of assets (G19) to be approximately the sum of the market value of equity and book value of liabilities and debt (G9 and G11).
Then, optimization needs to be run in Risk Simulator in order to obtain the desired outputs. To do this, set Asset Value and Volatility of Asset as the decision variables (make them continuous variables with a lower limit of 1% for volatility and $1 for assets, as both these inputs can only take on positive values). Set cell G29 as the objective to minimize as this is the absolute error value. Finally, the constraint is such that cell H33, the implied volatility in the default model, is set to exactly equal the numerical value of the equity volatility in cell G10. Run a static optimization using Risk Simulator.
If the model has a solution, the absolute error value in cell G29 will revert to zero (Figure 92.2). From here, the probability of default (measured in percent) and the distance to default (measured in standard deviations) are computed in cells G39 and G41. Then the relevant credit spread required can be determined using the Credit Analysis – Credit Premium model or some other credit spread tables (such as using the Internal Credit Risk Rating model).
The results indicate that the company has a probability of default at 0.56% with 2.54 standard deviations to default, indicating good creditworthiness (Figure 92.2).
A simpler approach is to use the Modeling Toolkit functions instead of manually running the optimization. These functions have internal intelligent optimization routines embedded in them. For instance, the following two functions perform multiple internal optimization routines of simultaneous stochastic equations to obtain their respective results, which are then used as an input into the MTProbabilityDefaultMertonII function to compute the probability of default:
MTProbabilityDefaultMertonImputedAssetValue
MTProbabilityDefaultMertonImputedAssetVolatility
See the model for more specific details.
Several other structural models exist for computing the probability of default of a firm. Specific models are used depending on the need and availability of data. In the previous example, the firm is a publicly-traded firm, with stock prices and equity volatility that can be readily obtained from the market. In the present example, we assume that the firm is privately held, meaning that there would be no market equity data available. This example essentially computes the probability of default or the point of default for the company when its liabilities exceed its assets, given the asset’s growth rates and volatility over time (Figure 92.3). Before using this model, first, review the preceding example using the Probability of Default – External Options Model. Similar methodological parallels exist between these two models, and this example builds on the knowledge and expertise of the previous example.
In Figure 92.3, the example firm with an asset value of $12M and a debt book value of $10M with significant growth rates of its internal assets and low volatility returns a 1.15% probability of default. Instead of relying on the valuation of the firm, external market benchmarks can be used, if such data are available. In Figure 92.4, we see that additional input assumptions are required, such as the market fluctuation (market returns and volatility) and the relationship (correlation between the market benchmark and the company’s assets). The model used is the Probability of Default – Merton Market Options Model accessible from Modeling Toolkit | Prob of Default | Merton Market Options Model (Industry Comparable).
Figure 92.1: Default probability model setup
Figure 92.2: Default probability of a publicly traded entity
Figure 92.3: Default probability of a privately held entity
Figure 92.4: Default probability of a privately held entity calibrated to market fluctuations
As mentioned, empirical models of the probability of default are used to compute an individual’s default probability, applicable within the retail banking arena, where empirical or actual historical or comparable data exist on past credit defaults. The dataset in Figure 92.5 represents a sample of several thousand previous loans, credit, or debt issues. The data show whether each loan had defaulted or not (0 for no default, and 1 for default) as well as the specifics of each loan applicant’s age, education level (1 to 3 indicating high school, university, or graduate professional education), years with current employer, and so forth. The idea is to model these empirical data to see which variables affect the default behavior of individuals, using Risk Simulator’s Maximum Likelihood Model. The resulting model will help the bank or credit issuer compute the expected probability of default of an individual credit holder having specific characteristics.
The example file is Probability of Default – Empirical and can be accessed through Modeling Toolkit | Prob of Default | Empirical (Individuals). To run the analysis, select the data on the left or any other dataset (include the headers) and make sure that the data have the same length for all variables, without any missing or invalid data. Then, using Risk Simulator, click on Risk Simulator | Forecasting | Maximum Likelihood Models. A sample set of results is provided in the MLE worksheet, complete with detailed instructions on how to compute the expected probability of default of an individual.
The Maximum Likelihood Estimates (MLE) approach on a binary multivariate logistic analysis is used to model dependent variables to determine the expected probability of success of belonging to a certain group. For instance, given a set of independent variables (e.g., age, income, education level of credit card or mortgage loan holders), we can model the probability of default using MLE. A typical regression model is invalid because the errors are heteroskedastic and non-normal, and the resulting estimated probability estimates will sometimes be above 1 or below 0. MLE analysis handles these problems using an iterative optimization routine. The computed results show the coefficients of the estimated MLE intercept and slopes.¹
The coefficients estimated are actually the logarithmic odds ratios and cannot be interpreted directly as probabilities. A quick but simple computation is first required. The approach is straightforward. To estimate the probability of success of belonging to a certain group (e.g., predicting if a debt holder will default given the amount of debt he holds), simply compute the estimated Y value using the MLE coefficients. Figure 92.6 illustrates an individual with 8 years at a current employer and current address, a low 3% debt to income ratio, and $2,000 in credit card debt has a log odds ratio of –3.1549. The inverse antilog of the odds ratio is obtained by computing:
¹For instance, the coefficients are estimates of the true population β values in the following equation: Y = β0 + β1X1 + β2X2 + … + βnXn. The standard error measures how accurate the predicted coefficients are, and the Z-statistics are the ratios of each predicted coefficient to its standard error. The Z-statistic is used in hypothesis testing, where we set the null hypothesis (Ho) such that the real mean of the coefficient is equal to zero, and the alternate hypothesis (Ha) such that the real mean of the coefficient is not equal to zero. The Z-test is very important as it calculates if each of the coefficients is statistically significant in the presence of the other regressors. This means that the Z-test statistically verifies whether a regressor or independent variable should remain in the model or it should be dropped. That is, the smaller the p-value, the more significant the coefficient. The usual significant levels for the p-value are 0.01, 0.05, and 0.10, corresponding to the 99%, 95%, and 99% confidence levels.
So, such a person has a 4.09% chance of defaulting on the new debt. Using this probability of default, you can then use the Credit Analysis – Credit Premium model to determine the additional credit spread to charge this person given this default level and the customized cash flows anticipated from this debt holder.
Figure 92.5: Empirical analysis of probability of default
the p-value are 0.01, 0.05, and 0.10, corresponding to the 99%, 95%, and 99% confidence levels.
Figure 92.6: MLE results
As shown previously, the probability of default is a key parameter for computing the credit risk of a portfolio. In fact, the Basel II/III Accord requires that the probability of default as well as other key parameters, such as the loss given default (LGD) and exposure at default (EAD), be reported as well. The reason is that a bank’s expected loss is equivalent to:
Expected Losses = (Probability of Default) × (Probability of Default) × (Exposure at Default)
or simply EL = PD × LGD × EAD.
PD and LGD are both percentages, whereas EAD is a value. As we have shown how to compute PD earlier, we will now revert to some estimations of LGD. There are several methods used to estimate LGD. The first is through a simple empirical approach where we set LGD = 1 – Recovery Rate. That is, whatever is not recovered at default is the loss at default, computed as the charge-off (net of recovery) divided by the outstanding balance:
Therefore, if market data or historical information are available, LGD can be segmented by various market conditions, types of obligor, and other pertinent segmentations. LGD can then be readily read off a chart.
A second approach to estimate LGD is more attractive in that if the bank has available information, it can attempt to run some econometric models to create the best-fitting model under an ordinary least squares approach. By using this approach, a single model can be determined and calibrated, and this same model can be applied under various conditions, with no data mining required. However, in most econometric models, a normal transformation will have to be performed first. Supposing the bank has some historical LGD data (Figure 92.7), the best-fitting distribution can be found using Risk Simulator (select the historical data, click on Risk Simulator | Analytical Tools | Distributional Fitting (Single Variable) to perform the fitting routine). The result is a beta distribution for the thousands of LGD values.
Figure 92.7: Fitting historical LGD data
Then, using the Distribution Analysis tool in Risk Simulator, obtain the theoretical mean and standard deviation of the fitted distribution (Figure 92.8). Then transform the LGD variable using the MTNormalTransform function in the Modeling Toolkit software. For instance, the value of 49.69% will be transformed and normalized to 28.54%. Using this newly transformed dataset, you can run some nonlinear econometric models to determine LGD.
For instance, a partial list of independent variables that might be significant for a bank, in terms of determining and forecast the LGD value, might include:
Figure 92.8: Distributional analysis tool
Economic capital is critical to a bank as it links a bank’s earnings and returns to risks that are specific to business lines or business opportunities. In addition, these economic capital measurements can be aggregated into a portfolio of holdings. Value at Risk (VaR) is used in trying to understand how the entire organization is affected by the various risks of each holding as aggregated into a portfolio, after accounting for cross-correlations among various holdings. VaR measures the maximum possible loss given some predefined probability level (e.g., 99.90%) over some holding period or time horizon (e.g., 10 days). Senior management at the bank usually selects the probability or confidence interval, which is typically a decision made by senior management at the bank and reflects the board’s risk appetite. Stated another way, we can define the probability level as the bank’s desired probability of surviving per year. In addition, the holding period is usually chosen such that it coincides with the time period it takes to liquidate a loss position.
VaR can be computed in several ways. Two main families of approaches exist: structural closed-form models and Monte Carlo risk simulation approaches. We showcase both methods in this chapter, starting with the structural models.
The second and much more powerful approach is Monte Carlo risk simulation. Instead of simply correlating individual business lines or assets, entire probability distributions can be correlated using mathematical copulas and simulation algorithms, by using Risk Simulator. In addition, tens to hundreds of thousands of scenarios can be generated using simulation, providing a very powerful stress-testing mechanism for valuing VaR. Distributional fitting methods are applied to reduce the thousands of data points into their appropriate probability distributions, allowing their modeling to be handled with greater ease.
This first VaR example model used is Value at Risk – Static Covariance Method, accessible through Modeling Toolkit | Value at Risk | Static Covariance Method. This model is used to compute the portfolio’s VaR at a given percentile for a specific holding period, after accounting for the cross-correlation effects between the assets (Figure 92.9). The daily volatility is the annualized volatility divided by the square root of trading days per year. Typically, positive correlations tend to carry a higher VaR compared to zero correlation asset mixes, whereas negative correlations reduce the total risk of the portfolio through the diversification effect (Figures 92.9 and 92.10). The approach used is a portfolio VaR with correlated inputs, where the portfolio has multiple asset holdings with different amounts and volatilities. Each asset is also correlated to each other. The covariance or correlation structural model is used to compute the VaR given a holding period or horizon and percentile value (typically 10 days at 99% confidence). Of course, the example illustrates only a few assets or business lines or credit lines for simplicity’s sake. Nonetheless, using the VaR function (MTVaRCorrelationMethod) in the Modeling Toolkit, many more lines, assets, or businesses can be modeled.
Figure 92.9: Computing Value at Risk using the structural covariance method
Figure 92.10: Different correlation levels
The model used is Value at Risk – Portfolio Operational and Credit Risk VaR Capital Adequacy and is accessible through Modeling Toolkit | Value at Risk | Portfolio Operational and Credit Risk VaR Capital Adequacy. This model shows how operational risk and credit risk parameters are fitted to statistical distributions and their resulting distributions are modeled in a portfolio of liabilities to determine the VaR (99.50th percentile certainty) for the capital requirement under Basel II/III requirements. It is assumed that the historical data of the operational risk impacts (Historical Data worksheet) are obtained through econometric modeling of the Key Risk Indicators.
The Distributional Fitting Report worksheet is a result of running a distributional fitting routine in Risk Simulator to obtain the appropriate distribution for the operational risk parameter. Using the resulting distributional parameter, we model each liability’s capital requirements within an entire portfolio. Correlations can also be inputted if required, between pairs of liabilities or business units. The resulting Monte Carlo simulation results show the VaR capital requirements.
Note that an appropriate empirically based historical VaR cannot be obtained if distributional fitting and risk-based simulations were not first run. The VaR will be obtained only by running simulations. To perform distributional fitting, follow these steps:
Figure 92.11: Sample historical bank loans
Figure 92.12: Data fitting results
Figure 92.13: Simulated forecast results and the 99.50% Value at Risk value
Another example on VaR computation is shown next, where the model Value at Risk – Right Tail Capital Requirements is used, available through Modeling Toolkit | Value at Risk | Right Tail Capital Requirements.
This model shows the capital requirements per Basel II/III requirements (99.95th percentile capital adequacy based on a specific holding period’s VaR). Without running risk-based historical and Monte Carlo simulation using Risk Simulator, the required capital is $37.01M (Figure 92.14) as compared to only $14.00M that is required using a correlated simulation (Figure 92.15). This is due to the cross-correlations between assets and business lines and can only be modeled only using Risk Simulator. This lower VaR is preferred as banks can be required to hold less required capital and can reinvest the remaining capital in various profitable ventures, thereby generating higher profits.
Figure 92.14: Right-tail VaR model
To run the model, follow these steps:
Figure 92.15: Simulated results of the portfolio VaR
Figure 92.16: Setting correlations one at a time
Figure 92.17: Setting correlations using the correlation matrix routine
If risk simulation was not run, the VaR or economic capital required would have been $37M, as opposed to only $14M. All cross-correlations between business lines have been modeled, as are stress and scenario tests, and thousands and thousands of possible iterations are run. Individual risks are now aggregated into a cumulative portfolio level VaR.
As a side note, by performing portfolio optimization, a portfolio’s VaR actually can be reduced. We start by first introducing the concept of stochastic portfolio optimization through an illustrative hands-on example. Then, using this portfolio optimization technique, we apply it to four business lines or assets to compute the VaR or an un-optimized versus an optimized portfolio of assets, and see the difference in computed VaR. You will note that in the end, the optimized portfolio bears less risk and has a lower required economic capital.
The optimization model used to illustrate the concepts of stochastic portfolio optimization is Optimization – Stochastic Portfolio Allocation and can be accessed via Modeling Toolkit | Optimization | Stochastic Portfolio Allocation. This model shows four asset classes with different risk and return characteristics. The idea here is to find the best portfolio allocation such that the portfolio’s bang for the buck or returns to risk ratio, is maximized. That is, in order to allocate 100% of an individual’s investment among several different asset classes (e.g., different types of mutual funds or investment styles: growth, value, aggressive growth, income, global, index, contrarian, momentum, etc.), an optimization is used. This model is different from others in that there exist several simulation assumptions (risk and return values for each asset), as seen in Figure 92.18.
In other words, a simulation is first run, then optimization is executed, and the entire process is repeated multiple times to obtain distributions of each decision variable. The entire analysis can be automated using Stochastic Optimization.
In order to run an optimization, several key specifications on the model have to be identified first:
Objective: Maximize Return to Risk Ratio (C12)
Decision Variables: Allocation Weights (E6:E9)
Restrictions on Decision Variables: Minimum and Maximum Required (F6:G9)
Constraints: Portfolio Total Allocation Weights 100% (E11 is set to 100%)
Simulation Assumptions: Return and Risk Values (C6:D9)
The model shows the various asset classes. Each asset class has its own set of annualized returns and annualized volatilities. These return and risk measures are annualized values such that they can be compared consistently across different asset classes. Returns are computed using the geometric average of the relative returns, while the risks are computed using the logarithmic relative stock returns approach.
Column E, Allocation Weights, holds the decision variables, which are the variables that need to be tweaked and tested such that the total weight is constrained at 100% (cell E11). Typically, to start the optimization, we will set these cells to a uniform value, where in this case, cells E6 to E9 are set at 25% each. In addition, each decision variable may have specific restrictions in its allowed range. In this example, the lower and upper allocations allowed are 10% and 40%, as seen in columns F and G. This setting means that each asset class can have its own allocation boundaries.
Figure 92.18: Asset allocation model ready for stochastic optimization
Column H shows the return to risk ratio, which is simply the return percentage divided by the risk percentage, where the higher this value, the higher the bang for the buck. The remaining sections of the model show the individual asset class rankings by returns, risk, return to risk ratio, and allocation. In other words, these rankings show at a glance which asset class has the lowest risk, or the highest return, and so forth.
To run this model, simply click on Risk Simulator | Optimization | Run Optimization. Alternatively, and for practice, you can set up the model using the following approach:
You may also try other optimization routines where:
Static Optimization is an optimization that is run on a static model, where no simulations are run. This optimization type is applicable when the model is assumed to be known and no uncertainties exist. Also, a static optimization can be run first to determine the optimal portfolio and its corresponding optimal allocation of decision variables before more advanced optimization procedures are applied. For instance, before running a stochastic optimization problem, a static optimization is run first to determine if there exist solutions to the optimization problem before a more protracted analysis is performed.
Dynamic Optimization is applied when Monte Carlo simulation is used together with optimization. Another name for such a procedure is Simulation-Optimization. In other words, a simulation is run for N trials, and then an optimization process is run for M iterations until the optimal results are obtained, or an infeasible set is found. That is, using Risk Simulator’s Optimization module, you can choose which forecast and assumption statistics to use and replace in the model after the simulation is run. Then, these forecast statistics can be applied in the optimization process. This approach is useful when you have a large model with many interacting assumptions and forecasts, and when some of the forecast statistics are required in the optimization.
Stochastic Optimization is similar to the dynamic optimization procedure except that the entire dynamic optimization process is repeated T times. The results will be a forecast chart of each decision variable with T values. In other words, a simulation is run and the forecast or assumption statistics are used in the optimization model to find the optimal allocation of decision variables. Then another simulation is run, generating different forecast statistics, and these new updated values are then optimized, and so forth. Hence, each of the final decision variables will have its own forecast chart, indicating the range of the optimal decision variables. For instance, instead of obtaining single-point estimates in the dynamic optimization procedure, you can now obtain a distribution of the decision variables, and, hence, a range of optimal values for each decision variable, also known as stochastic optimization.
Figure 92.19: Setting up the stochastic optimization problem
Stochastic optimization is performed when a simulation is first run and then the optimization is run. Then the whole analysis is repeated multiple times. The result is a distribution of each decision variable, rather than a single-point estimate (Figure 92.20). This means that instead of saying you should invest 30.57% in Asset 1, the optimal decision is to invest between 30.10% and 30.99% as long as the total portfolio sums to 100%. This way the optimization results provide management or decision makers a range of flexibility in the optimal decisions. Refer to Chapter 11 of Modeling Risk: Applying Monte Carlo Simulation, Real Options Analysis, Forecasting, and Optimization by Dr. Johnathan Mun for more detailed explanations about this model, the different optimization techniques, and an interpretation of the results. Chapter 11’s appendix also details how the risk and return values are computed.
Figure 92.20: Simulated results from the stochastic optimization approach
Now that we understand the concepts of optimized portfolios, let us see what the effects are on computed economic capital through the use of a correlated portfolio VaR. This model uses Monte Carlo simulation and optimization routines in Risk Simulator to minimize the Value at Risk of a portfolio of assets (Figure 92.21). The file used is Value at Risk – Optimized and Simulated Portfolio VaR, which is accessible via Modeling Toolkit | Value at Risk | Optimized and Simulated Portfolio VaR. In this example, we intentionally used only four asset classes to illustrate the effects of an optimized portfolio. In real life, we can extend this to cover a multitude of asset classes and business lines. Here we illustrate the use of a left-tail VaR, as opposed to a right-tail VaR, but the concepts are similar.
First, simulation is used to determine the 90% left-tail VaR. The 90% left-tail probability means that there is a 10% chance that losses will exceed this VaR for a specified holding period. With an equal allocation of 25% across the four asset classes, the VaR is determined using simulation (Figure 92.21). The annualized returns are uncertain and hence simulated. The VaR is then read off the forecast chart. Then optimization is run to find the best portfolio subject to the 100% allocation across the four projects that will maximize the portfolio’s bang for the buck (returns to risk ratio). The resulting optimized portfolio is then simulated once again, and the new VaR is obtained (Figure 92.22). The VaR of this optimized portfolio is a lot less than the not-optimized portfolio.
Figure 92.21: Computing Value at Risk (VaR) with simulation
Figure 92.22: Non-optimized Value at Risk
Figure 92.23: Optimal portfolio’s Value at Risk through optimization and simulation
Another related item in the discussion of risk in the context of Basel II/III Accords is the issue of hurdle rates, or the required rate of return on investment that is sufficient to justify the amount of risk undertaken in the portfolio. There is a nice theoretical connection between uncertainty and volatility whereby the discount rate of a specific risk portfolio can be obtained. In a financial model, the old axiom of high risk, high return is seen through the use of a discount rate. That is, the higher the risk of a project, the higher the discount rate should be to risk-adjust this riskier project so that all projects are comparable.
There are two methods for computing the hurdle rate. The first is an internal model, where the VaR of the portfolio is computed first. This economic capital is then compared to the market risk premium. That is, we have
That is, assuming that a similar set of comparable investments are obtained in the market, based on tradable assets, the market return is obtained. Using the bank’s internal cash flow models, all future cash flows can be discounted at the risk-free rate in order to determine the risk-free return. Finally, the difference is then divided into the VaR risk capital to determine the required hurdle rate. This concept is very similar to the capital asset pricing model (CAPM), which often is used to compute the appropriate discount rate for a discounted cash flow model (weighted average cost of capital, hurdle rates, multiple asset pricing models, and arbitrage pricing models are the other alternatives but are based on similar principles). The second approach is the use of the CAPM to determine the hurdle rate.