Methods and Apparatus for Improving Factor Risk Model Responsiveness
Construction of factor risk models that better predict the future volatility of returns of a portfolio of securities such as stocks, bonds, or the like is addressed. More specifically, improved factor-factor covariance estimation is made even when the covariances change rapidly over time. Methods and techniques for achieving better accuracy, responsiveness, and stability of factor risk models are addressed.
Latest Axioma, Inc. Patents:
- Performance attribution for portfolios with composite investments
- Methods and apparatus employing hierarchical conditional value at risk to minimize downside risk of a multi-asset class portfolio and improved graphical user interface
- Identifying and compensating for model mis-specification in factor risk models
- Methods and Apparatus for Implementing Improved Notional-free Asset Liquidity Rules
- Performance Attribution for Portfolios with Composite Investments
The present application is a divisional of U.S. patent application Ser. No. 13/503,698 filed on Apr. 24, 2012 which claims the benefit of PCT/US11/37231 filed on May 19, 2011 which claims the benefit of U.S. Provisional Application Ser. No. 61/435,439, filed Jan. 24, 2011, which are incorporated by reference herein in their entirety.
FIELD OF INVENTIONThe present invention relates generally to the estimation of the risk, or active risk, of an investment portfolio using factor risk models. More particularly, it relates to improved computer based systems, methods and software for more accurate estimation of the risk or active risk of an investment portfolio. The invention addresses techniques allowing a factor risk model's risk estimates to be more accurate, more stable, and more responsive.
BACKGROUND OF THE INVENTIONFinancial time series analysis often assumes that the statistical properties of equity returns do not vary over time. However, the statistical properties of actual returns data from financial markets do vary over time. In particular, empirical evidence suggests that volatility or risk, the square root of the variance of returns, changes with time.
The challenge for commercial risk model vendors is to produce risk models that predict future volatility, or, in other words, accurately predicting the realized risk 200 shown in
-
- 1. Prediction Accuracy. The difference between the realized and predicted volatilities.
- 2. Stability. The risk model predictions should not exhibit the smaller, transient changes observed in realized risk. In other words, the risk predictions should be smoother than the realized risk. Such smoothness ensures that portfolio rebalancing and risk management decisions are not driven by market transients of shorter duration than the investment holding horizon.
- 3. Responsiveness. When the overall level of market volatility rises or falls, the predicted risk should respond similarly with as little time lag in the response as possible.
Stability and responsiveness both bear on how changes in realized risk are tracked by risk model predictions. On the one hand, stability requires that smaller, temporary changes in realized volatility should not appear in the risk model predictions. On the other hand, responsiveness requires that larger, sustained changes in realized volatility should appear. Thus smaller changes are interpreted as noise that should not affect investment decisions while the larger changes are interpreted as meaningful changes that can and should affect investment decisions. The difference between smaller and larger changes or temporary and sustained changes depends, of course, on the manner in which the risk model is used. A portfolio manager who trades every day may consider a weeklong change in realized volatility a sustained change that should be captured by a high quality risk model, while a portfolio manager who invests over a time horizon of months may consider a weeklong change in realized volatility a temporary effect that should be filtered out of a high quality risk model. In both cases, the portfolio manager wants to react to meaningful changes in market volatility that cause material changes to his or her investment decisions while simultaneously avoiding any overreaction to temporary, noisy market conditions that may lead to unnecessary trading. Stability seeks to ensure that the risk model predictions are smooth over a sufficiently long period of time, while responsiveness seeks to ensure that the risk model predictions change and respond to market changes in volatility over a sufficiently short period of time.
In
In
There are several well known mathematical modeling techniques for estimating the risk of a portfolio of financial assets such as securities and for deciding how to strategically invest a fixed amount of wealth given a large number of financial assets in which to potentially invest.
For example, mutual funds often estimate the active risk associated with a managed portfolio of securities, where the active risk is the risk associated with portfolio allocations that differ from a benchmark portfolio. Often, a mutual fund manager is given a “risk budget”, which defines the maximum allowable active risk that he or she can accept when constructing a managed portfolio. Active risk is also sometimes called portfolio tracking error. Portfolio managers may also use numerical estimates of risk as a component of performance contribution, performance attribution, or return attribution, as well as, other ex-ante and ex-post portfolio analyses. See for example, R. Litterman, Modern Investment Management: An Equilibrium Approach, John Wiley and Sons, Inc., Hoboken, N.J., 2003 (Litterman), which gives detailed descriptions of how these analyses make use of numerical estimates of risk and which is incorporated by reference herein in its entirety.
Another use of numerically estimated risk is for optimal portfolio construction. One example of this is mean-variance portfolio optimization as described by H. Markowitz, “Portfolio Selection”, Journal of Finance 7(1), pp. 77-91, 1952 which is incorporated by reference herein in its entirety. In mean-variance optimization, a portfolio is constructed that minimizes the risk of the portfolio while achieving a minimum acceptable level of return. Alternatively, the level of return is maximized subject to a maximum allowable portfolio risk. The family of portfolio solutions solving these optimization problems for different values of either minimum acceptable return or maximum allowable risk is said to form an “efficient frontier”, which is often depicted graphically on a plot of risk versus return. There are numerous, well known, variations of mean-variance portfolio optimization that are used for portfolio construction. These variations include methods based on utility functions, Sharpe ratio, and value at risk.
Suppose that there are N assets in an investment portfolio and the weight or fraction of the available wealth invested in each asset is given by the N-dimensional column vector w. These weights may be the actual fraction of wealth invested or, in the case of active risk, they may represent the difference in weights between a managed portfolio and a benchmark portfolio as described by Litterman. The risk of this portfolio is calculated, using standard matrix notation, as
V=wTQw (1)
where V is the portfolio variance, a scalar quantity, and Q is an N×N positive semi-definite matrix whose elements are the variance or covariance of the asset returns. Risk or volatility is given by the square root of V.
The individual elements of Q are the expected covariances of security returns and are difficult to estimate. For N assets, there are N(N+1)/2 separate variances and covariances to be estimated. The number of securities that may be part of a portfolio, N, is often over one thousand, which implies that over 500,000 values must be estimated. Risk models typically cover all the assets in the asset universe, not just the assets with holdings in the portfolio, so N can be considerably larger than the number of assets in a managed or benchmark portfolio.
To obtain reliable variance or covariance estimates based on historical return data, the number of historical time periods used for estimation should be of the same order of magnitude as the number of assets, N. Often, there may be insufficient historical time periods. For example, new companies and bankrupt companies have abbreviated historical price data and companies that undergo mergers or acquisitions have non-unique historical price data. As a result, the covariances estimated from historical data can lead to matrices that are numerically ill conditioned. Such covariance estimates are of limited value.
Factor risk models were developed, in part, to overcome these short comings. See for example, R. C. Grinold, and R. N. Kahn, Active Portfolio Management: A Quantitative Approach for Providing Superior Returns and Controlling Risk, Second Edition, McGraw-Hill, New York, 2000, which is incorporated by reference herein it its entirety, and Litterman.
Factor risk models represent the expected variances and covariances of security returns using a set of M factors, where M is much less than N, that are derived using statistical, fundamental, or macro-economic information or a combination of any of such types of information. Given exposures of the securities to the factors and the covariances of factor returns, the covariances of security returns can be expressed as a function of the factor exposures, the covariances of factor returns, and a remainder, called the specific risk of each security. Factor risk models typically have between 20 and 80 factors. Even with 80 factors and 1000 securities, the total number of values that must be estimated is just over 85,000, as opposed to over 500,000.
In a factor risk model, the covariance matrix Q is modelled as
Q=BTΣB+Δ2 (2)
where B is an N×M matrix of factor exposures, Σ is an M×M matrix of factor-factor covariances, and Δ2 is a matrix of specific variances. Normally, Δ2 is assumed to be diagonal.
The factor-factor covariance matrix Σ is typically estimated from a time series of historical factor returns, ft, for each of the M factors, while the specific variances are estimated from a time series of historical specific returns.
Risk models used in quantitative portfolio management partly address the issues of stability and responsiveness when predicting time varying volatility by relying on an exponentially weighted covariance estimator since this estimator places greater emphasis on current observations, implicitly assuming that the most recent subset of return values often vary around a constant value. The returns can be asset returns, or they may be factor returns or specific returns used for estimating a risk model covariance. Given a time series of T returns {rt, rt−1, rt−2, . . . , rt−T+1}, we form the weighted returns series {{tilde over (r)}t}
{{tilde over (r)}t}={(wt,rt),(wt−1,rt−1), . . . ,(wt−T+1,rt−T+1)} (3)
wt−k=2−k/H, k=0, . . . T−1 (4)
where H is the half-life parameter. The exponentially weighted covariance estimator gives
E[var(rt+1)|t]≡{circumflex over (σ)}t+12=var({tilde over (r)}t) (5)
This is frequently seen expressed in the RiskMetrics™ specification in which the half-life is reformulated as a decay factor λ. See, for example, J. Longerstaey and M. Spencer, RiskMetrics™—Technical Document, Morgan Guaranty Trust Company, New York, 4th ed., 1996, which is incorporated by reference herein it its entirety. Equation (5) can be rewritten as:
{circumflex over (σ)}t+12=λσt2+(1−λ)rt2 (6)
Ease and speed of computation, robustness, and parsimony have largely been responsible for the widespread adoption of exponentially weighted covariance estimates in commercial risk models. Exponential weighting generally improves the accuracy of the risk model.
However, when realized risk changes rapidly, the risk predictions of risk models using exponentially weighted covariance estimates often lag realized risk changes over considerable periods of time. In other words, exponential weighting does not always lead to the desired level of responsiveness in a risk model. This lag is shown in
One problem with exponentially weighted covariance estimates recognized and addressed by the present invention is that large returns have a disproportionate effect on the covariance estimate even with exponential weighting. These large returns can inflate risk estimates, and they impact risk estimates for very long times, resulting in lagged risk predictions, especially when volatility falls from a high level, such as shown by gap 201 in 2003 and gap 203 in 2009.
In order to produce stable risk predictions, risk models typically require a long history of data for the covariance estimate. The longer the data history, however, the more likely it is that the return history will span a time period over which the volatility of the older returns is at a substantially different level than the volatility of the recent returns. Although the exponentially weighted covariance estimate will give the older return data less weight than the most recent returns, the resulting volatility forecast may noticeably lag, in other words, not be as responsive to the realized volatility results as desired, if the volatility of the older return data is substantially different than the volatility of the more recent return data.
One approach to the problem of lagging risk model predictions is to use shorter data histories and/or aggressive decay factors in order to reduce the influence of the older data on the forecasts. However, if the data history is too short or the decay factors too aggressive, the stability of the risk model predictions may be jeopardized.
Other methods besides more aggressive half-lives have been proposed to address the issue of non-stationarity of asset returns, factor returns, and specific returns. For example, generalized autoregressive conditional heteroskedasticity (GARCH,) models have been proposed. See, for example, Tim Bollerslev, “Generalized Autoregressive Conditional Heteroskedasticity”, Journal of Econometrics, 31:307-327, 1986, which is incorporated by reference herein it its entirety. However, GARCH models normally produce risk models that are too unstable for use in commercial risk models.
SUMMARY OF THE INVENTIONThe present invention recognizes that the lag in responding to rapidly changing market volatility in 2003 and 2009 can be improved upon when compared with the responsiveness to existing risk models. One aspect of the present invention is to provide a methodology for improving risk model responsiveness with minimal negative impact on both the risk model accuracy and stability. In some cases, as addressed further below, the accuracy and stability may also be improved.
One goal of risk model prediction in accordance with the present invention, then, is to obtain a smooth curve of predicted risks that closely tracks the realized risk but that does not exhibit lags when the overall level of volatility changes substantially. Exponential weighting alone does not solve the problem, nor does the use of shorter data histories or more aggressive decay factors.
Another goal of the present invention is to improve responsiveness of the predicted risk without substantially increasing the change in forecast risk from one period to another.
Another aspect of the present invention is to improve risk model responsiveness over long periods of time.
Among its several aspects, the present invention addresses three things: (1), improving the responsiveness of the risk model; (2), maintaining the same level of stability found in traditional risk models, where stability is measured using the change in month-to-month predicted risk; and (3), maintaining accurate risk predictions over long periods of time during which the overall level of market volatility may or may not change substantially.
The present invention may be suitably implemented as a computer based system, in computer software which is stored in a non-transitory manner and which may suitably reside on computer readable media, such as solid state storage devices, such as RAM, ROM, or the like, magnetic storage devices such as a hard disk or floppy disk media, optical storage devices, such as CD-ROM or the like, or as methods implemented by such systems and software.
One embodiment of the invention has been designed for use on a stand-alone personal computer running in Windows (Microsoft XP, Vista, Windows 7). Another embodiment of the invention has been designed to run on a Linux-based server system.
According to one aspect of the invention, it is contemplated that the computer 12 will be operated by a user in an office, business, trading floor, classroom, or home setting.
As illustrated in
As further illustrated in
The output information may appear on a display screen of the monitor 22 or may also be printed out at the printer 24. The output information may also be electronically sent to an intermediary for interpretation. For example, risk predictions for many portfolios can be aggregated for multiple portfolio or cross-portfolio risk management. Or, alternatively, trades based, in part, on the factor risk model predictions, may be sent to an electronic trading platform. Other devices and techniques may be used to provide outputs, as desired.
With this background in mind, we turn to a detailed discussion of the invention and its context. The invention is herein referred to as Dynamic Volatility Adjustment (DVA). DVA seeks to find a weighting scheme for historical returns that transforms them so that they more closely resemble a weakly stationary time-series. In the discussion that follows, algorithms may be suitably implemented as software stored in memory and executed by a processor or processors in computer 12. Data may be input by a user or retrieved from a database or other storage. Data entered by a user may be entered using a keyboard, mouse, touchscreen display or other data entry device or means. Output data may be printed by a printer, displayed by a display, transmitted over the network to another user or users, or otherwise output utilizing an output device or means. Equity returns rt are weakly stationary when the first two moments of their distributions are stationary:
E[rt]=μ (7)
cov(rt,rt−τ)=γτ (8)
for any τ where we assume that rt and rt−τ have finite and time invariant first and second moments, and that these values only depend on τ. When τ=0, equation (8) becomes variance, which is often used as a measure of market volatility. Weak stationarity is a handy condition because it allows inferences and predictions to be made about future returns. See, for example, Ruey Tsay, Analysis of Financial Time Series, John Wiley & Sons Inc., 2005, which is incorporated by reference herein it its entirety.
DVA seeks a weighting scheme for a set of historical factor returns data that transforms its second moment into a weakly stationary statistic. Specific returns are not modified by the DVA algorithm as they are too unstable. Let ft be the observed time history of factor returns, and gt be a weighting function to be determined. Weak stationarity of the volatility of (gt, ft) requires
cov[(gt,ft),(gt−τ,ft−τ)]=γτ (9)
For a finite set of T observed returns, {f1, f2, f3, f4, . . . , fT}, where f1 is the oldest return and fT is the most recent return, we want the series of weighted returns, {(g1,f1), (g2,f2), (g3,f3), (g4,f4), . . . , (gT,fT)}, to fluctuate with a relatively constant level of variance, computed as {circumflex over (σ)}2=var {(g1, f1), (g2, f2), (g3,f3), (g4,f4), . . . , (gT,fT)}.
There are many weighting functions gt that will satisfy weak stationarity of the covariance. As originally formulated, DVA includes the following steps.
First, assume that the T data points can be grouped into N overlapping segments of length K, where T=(N+1)K/2, and one final data segment of length K/2. Each segment except the earliest shares half of the points of the segment immediately before it, and each segment except the latest shares half the points of the segment immediately after it. The latest segment is only half the length of the others, and is the ‘reference’ segment, containing the most recent data. For example, with T=12, N=5, and K=4, we obtain the segments:
{S1}={f1,f2,f3,f4}
{S2}={f3,f4,f5,f6}
{S3}={f5,f6,f7,f8}
{S4}={f7,f8,f9,f10}
{S5}={f9,f10,f11,f12}
{S6}={f11,f12} (10)
Each segment of historical data is denoted by {Sn}, n=1, . . . N+1, and is used to define a distinct volatility regime. The last segment, {SN+1}, is referred to as the reference chuck or reference segment. In practice, the values for T, N, and K would be larger than for this simple example. For example, T is often the entire factor return history. For Axioma's US Equity model, there are daily returns going back to Jan. 3, 1995, which is more than 4000 factor returns. For a fundamental factor risk model, T normally corresponds to four years of data, making T=1000. For a statistical factor risk model, T normally corresponds to one year of data, making T=250. K normally corresponds to about 6 months or 125. Hence, N=7 for a fundamental factor risk model and N=3 for a statistical factor risk model. Compute the N mean absolute deviations for each {Sn}:
Second, compute the N scaling factors for each νn
The N+1 scaling factors are clipped to lie within 0.8≦δn≦1.25. This prevents the scaling values having too large an impact on the returns, which improves stability. This potentially adversely affects the stationarity of the resulting time series, but is imposed for the sake of model stability and robustness to noisy data.
Third, assume a piecewise constant approximation of the N+1 scaling factors δn to compute the T weighting values gt. That is
Since δN+1≡1, we have gT=1 so that the most recently observed return is unchanged when weighted.
The advantages of DVA can be seen in the results shown in
In
The present invention recognizes, for example, that in
Although both the DVA and shorter half life models have comparable responsiveness, the stability of the DVA model is superior to that of the shorter half-life model.
The present invention addresses a formulation of DVA that improves upon several aspects of the original formulation.
In the original DVA formulation, the weighting values gt are constant over finite time intervals. The jump that occurs when the weighting values change from one scaling factor, δn, to another can lead to undesirable changes in the risk model prediction. In other words, the fact that the weighting values approximate a non-differentiable function can negatively impact the stability of the risk predictions.
In the original formulation of DVA, the length of the reference chunk {SN+2 } is half that of the other chunks. This makes it difficult to define the reference time horizon of the resulting risk model as the scaling values are defined over data segments of different length.
In the original DVA formulation, scaling values lie within [0.8, 1.25]. Although this ensures a degree of stability, it also means that an older period of excessively high volatility will never be scaled down by a factor of greater than 0.8, no matter how distantly in the past it lies.
In the original DVA formulation, the long term accuracy of the DVA enable risk model may be worse than that of a risk model with a shorter half life, as shown in
To address these issues, an improved version of DVA incorporates the following improvements. This improved DVA is the preferred embodiment of the invention.
First, rather than segmenting the history of factor returns into N segments of length K and a final segment of length K/2, the history is segmented into only N segments of length K. Hence, with T=12, N=5, and K=4, we obtain the segments:
{S1}={f1,f2,f3,f4}
{S2}={f3,f4,f5,f6}
{S3}={f5,f6,f7,f8}
{S4}={f7,f8,f9,f10}
{S5}={f9,f10,f11,f12} (14)
With this change, the N scaling factors are redefined as
where now δN=1.
Secondly, rather than use a piecewise constant approximation to estimate the weighting values, use cubic spline interpolation on the N scaling factors δn to compute the T weighting values gt assuming
gT−(n−1)K=δN−(n−1), n=1, . . . ,N. (16)
Since δN=1, we have gT=1 so that the most recently observed return is unchanged when weighted. Unlike the piecewise constant approximation of the original DVA formulation, this approximation varies smoothly and continuously.
Thirdly, rather than clip the scaling values to be within [0.8, 1.25], the requirement that the ratio between any two consecutive scaling factors is no more than 10% is imposed. That is,
This ensures stability while simultaneously allowing older time periods of excessively high volatility to be appropriately scaled. It also improves the cubic spline interpolation. If a series of scaling factors is clipped at the same level, then the cubic-spline interpolation oscillates around the clipped value. The 10% figure was chosen empirically to balance responsiveness and stability, and to minimize prediction differences between improved DVA and original DVA during periods of relatively stable volatility. These changes substantially improve the performance of DVA.
In all cases, the original version of DVA is more responsive in January 2009, but its accuracy erodes over the rest of 2009 in comparison to improved DVA and the shorter half-life model.
However, there are significant differences in the stability of the three model variants.
For each of the 45 portfolios, three bars are shown: a light bar on the left 270 representing the shorter half life risk model predictions; a medium bar on the right 272 representing the original DVA predictions; and a dark bar in the center 274 representing the improved DVA predictions.
The improved DVA model predictions 274 show a clear reduction in forecast change in comparison with both the original DVA model 272 and the shorter half-life model 270. In other words, its forecasts are much smoother, day on day, without losing accuracy. These results show that the improved DVA gives better responsiveness without sacrificing smoothness of forecast.
While the present invention has been disclosed in the context of various aspects of presently preferred embodiments, it will be recognized that the invention may be suitably applied to other environments consistent with the claims which follow.
Claims
1. A computer-based method of estimating the variance of a factor in a factor risk model comprising the steps of:
- storing data for the factor in a memory;
- determining a time series history of factor returns for the factor over a set of historical times by a programmed processor cooperating with the memory and with software;
- calculating a set of exponentially decaying weights with a fixed half life corresponding to the time series history of factor returns by the programmed processor cooperating with the memory and with software;
- computing a metric of volatility for each historical time by the programmed processor cooperating with memory and with software;
- calculating a set of volatility adjustment multipliers by the programmed processor cooperating with the memory and with software as the ratios of most recent volatility metric to the computed volatility metric;
- determining when at least one volatility adjustment multiplier is outside a predetermined range;
- adjusting the at least one volatility adjustment multiplier to a value in the predetermined range;
- computing the factor-factor covariance for the time series of factor returns using the volatility adjustment multipliers within the range and any adjusted volatility adjustment multipliers for any volatility adjustment multipliers determined to be outside the range by the programmed processor cooperating with the memory and with software; and
- outputting the factor variance as part of a factor risk model as an electronic output by an output device.
2. The method of claim 1 where the output factor variance is used in the computation of the volatility of a portfolio of assets.
3. The method of claim 1 where the output factor variance is used to rebalance an investment portfolio.
4. The method of claim 1 where the output factor variance is used in a performance attribution analysis.
5. A computer-based apparatus for estimating the variance of a factor in a factor risk model comprising:
- a programmed processor cooperating with memory and with software to: determine a time series history of factor returns over a set of historical times selected utilizing an input device; calculate a set of exponentially decaying weights with a fixed half life corresponding to the time series history of factor returns; compute a metric of volatility for each historical time; calculate a set of volatility adjustment multipliers that is the ratio of most recent volatility metric to the measured volatility metric; determine at least one volatility adjustment multiplier is outside a predetermined range; adjust the at least one volatility adjustment multiplier to a value in the predetermined range; compute the factor variance for the time series of factor returns using the set of exponentially decaying weights, and volatility adjustment multipliers within the range and any adjusted volatility adjustment multiplier for any volatility multiplier determined to be outside the range; and
- an output means for outputting the factor variance as part of a factor risk model as an electronic output.
6. The apparatus of claim 5 where the output factor variance is used in the computation of the volatility of a portfolio of assets stored in a database.
7. The apparatus of claim 5 where the output factor variance is used by the programmed processor to rebalance an investment portfolio.
8. The apparatus of claim 5 where the output factor variance is used by the programmed processor to perform a performance attribution analysis.
Type: Application
Filed: Mar 11, 2014
Publication Date: Jul 17, 2014
Applicant: Axioma, Inc. (New York, NY)
Inventors: Simon Wannasin Bell (Hove), Frank Pak-Ho Siu (Hong Kong)
Application Number: 14/203,807
International Classification: G06Q 40/06 (20120101);