Falling Water

The current 15-year drought in the South West is the most severe since recordkeeping for the Colorado River began in 1906.  Lake Mead, which supplies much of the water to Colorado Basin communities, is now more than half empty.

bathrub

A 120 foot high band of rock, bleached white by the water, and known as the “bathtub ring” encircles the lake, a stark reminder of the water crisis that has enveloped the surrounding region.  The Colorado River takes a 1,400 mile journey from the Rockies to Mexico, irrigating over 5 million acres of farmland in the Basin states of Wyoming, Utah, Colorado, New Mexico, Nevada, Arizona, and California.

 

map

The Colorado River Compact signed in 1922 enshrined the States’ water rights in law and Mexico was added to the roster in 1994, taking the total allocation to over 16.5 million acre-feet per year.  But the average freshwater input to the lake over the century from 1906 to 2005 reached only 15 million acre-feet.  The river can’t come close to meeting current demand and the problem is only likely to get worse. A 2009 study found that rainfall in the Colorado Basin could fall as much as 15% over the next 50 years and the shortfall in deliveries could reach 60% to 90% of the time.

 

 

Impact on Las Vegas

With an average of only 4 inches of rain a year, and a daily high temperatures of 103 o F during the summer, Las Vegas is perhaps the most hard pressed to meet the demand of its 2 million residents and 40 million visitors annually.   bellagio

Despite its conspicuous consumption, from the tumbling fountains of the Bellagio to the Venetian’s canals, since 2002, Las Vegas has been obliged to cut its water use by a third, from 314 gallons per capita a day to 212. The region recycles around half of its wastewater which is piped back into Lake Mead, after cleaning and treatment.  Residents are allowed to water their gardens no more than one day a week in peak season, and there are stiff fines for noncompliance.

SSALGOTRADING AD

The Third Straw

WATER_Last straw_sidebar 02_1

Historically, two intake pipes carried water from Lake Mead to Las Vegas, about 25 miles to the west. In 2012, realizing that the highest of these, at 1050 feet, would soon be sucking air, the Southern Nevada Water Authority began construction of a new pipeline. Known as the Third Straw, Intake No. 3 reaches 200 feet deeper into the lake—to keep water flowing for as long as there’s water to pump.  The new pipeline, which commenced operations in 2015, doesn’t draw more water from the lake than before, or make the surface level drop any faster. But it will keep taps flowing in Las Vegas homes and casinos even if drought-stricken Lake Mead drops to its lowest levels.

WATER_three images

 

 

 

 

 

 

Modeling Water Levels in Lake Mead

The monthly reported water levels in Lake Mead from Feb 1935 to June 2016 are shown in the chart below. The reference line is the drought level, historically defined as 1,125 feet.

chart1

 

One statistical technique widely applied in hydrology involves fitting a Kumaraswamy distribution to the relative water level.  According to the Arizona Game and Fish Department, the maximum lake level is 1229 feet.  We model the water level relative to the maximum level, as follows.

code1

 

chart2

The fit of the distribution appears quite good, even in the tails:

ProbabilityPlot[relativeLevels, edist]

chart3

 

Since water levels have been below the drought level for some time, let’s instead consider the “emergency” level, 1,075 feet.  According to this model, there is just over a 6% chance of Lake Mead hitting the emergency level and, consequently, a high probability of breaching the emergency threshold some time over before the end of 2017.

 

code2 code4

chart4

 

One problem with this approach is that it assumes that each observation is drawn independently from a random variable with the estimated distribution.  In reality, there are high levels of autocorrelation in the series, as might be expected:  lower levels last month typically increase the likelihood of lower levels this month.  The chart of the autocorrelation coefficients makes this pattern clear, with statistically significant coefficients at lags of up to 36 months.

ts[“ACFPlot”]

chart5

 

 

An alternative methodology that enables us to take account of the autocorrelation in the process is time series analysis.  We proceed to fit an autoregressive moving average (ARMA) model as follows:

tsm = TimeSeriesModelFit[ts, “ARMA”]

arma

The best fitting model in an ARMA(1,1) model, according to the AIC criterion:

armatable

Applying the fitted ARMA model, we forecast the water level in Lake Mead over the next ten years as shown in the chart below.  Given the mean-reverting moving average component of the model, it is not surprising to see the model forecasting a return to normal levels.

chart6

 

There is some evidence of lack of fit in the ARMA model, as shown in the autocorrelations of the model residuals:

chart7

A formal test reveals that residual autocorrelations at lags 4 and higher are jointly statistically significant:

chart8

The slowly decaying pattern of autocorrelations in the water level series suggests a possible “long memory” effect, which can be better modelled as a fractionally integrated process.  The forecasts from such a model, like the ARMA model forecasts, display a tendency to revert to a long term mean; but the reversion process is dampened by the reinforcing, long-memory effect captured in the FARIMA model.

chart9

The Prospects for the Next Decade

Taking the view that the water level in Lake Mead forms a stationary statistical process, the likelihood is that water levels will rise to 1,125 feet or more over the next ten years, easing the current water shortage in the region.

On the other hand, there are good reasons to believe that there are exogenous (deterministic) factors in play, specifically the over-consumption of water at a rate greater than the replenishment rate from average rainfall levels.  Added to this, plausible studies suggest that average rainfall in the Colorado Basin is expected to decline over the next fifty years.  Under this scenario, the water level in Lake Mead will likely continue to deteriorate, unless more stringent measures are introduced to regulate consumption.

 Economic Impact

The drought in the South West affects far more than just the water levels in Lake Mead, of course.  One study found that California’s agriculture sector alone had lost $2.2Bn and some 17,00 season and part time jobs in 2014, due to drought.  Agriculture uses more than 80% of the State’s water, according to Fortune magazine, which goes on to identify the key industries most affected, including agriculture, food processing, semiconductors, energy, utilities and tourism.

Dry fields and bare trees at Panoche Road, looking west, on Wednesday February 5, 2014, near San Joaquin, CA. California drought has hit the Central Valley hard.
Dry fields and bare trees at Panoche Road, looking west, on Wednesday February 5, 2014, near San Joaquin, CA. California drought has hit the Central Valley hard.

In the energy sector, for example, the loss of hydroelectric power cost CA around $1.4Bn in 2014, according to non-profit research group Pacific Institute.  Although Intel pulled its last fabrication plant from California in 2009, semiconductor manufacturing is still a going concern in the state. Maxim Integrated, TowerJazz, and TSI Semiconductors all still have fabrication plants in the state. And they need a lot of water. A single semiconductor fabrication plant can use as much water as a small city. That means the current plants could represent three cities worth of consumption.

The drought is also bad news for water utilities, of course. The need to conserve water raises the priority on repair and maintenance, and that means higher costs and lower profit. Complicating the problem, California lacks any kind of management system for its water supply and can’t measure the inflows and outflows to ground water levels at any particular time.

The Bureau of Reclamation has studied more than two dozen options for conserving and increasing water supply, including importation, desalination and reuse. While some were disregarded for being too costly or difficult, the bureau found that the remaining options, if instituted, could yield 3.7 million acre feet per year in savings and new supplies, increasing to 7 million acre feet per year by 2060.  Agriculture is the biggest user by far and has to be part of any solution. In the near term, the agriculture industry could reduce its use by 10 to 15 percent without changing the types of crops it grows by using new technology, such as using drip irrigation instead of flood irrigation and monitoring soil moisture to prevent overwatering, the Pacific Institute found.

Conclusion

We can anticipate that a series of short term fixes, like the “Third Straw”, will be employed to kick the can down the road as far as possible, but research now appears almost unanimous in finding that drought is having a deleterious, long term affect on the economics of the South Western states.  Agriculture is likely to have to bear the brunt of the impact, but so too will adverse consequences be felt in industries as disparate as food processing, semiconductors and utilities.  California, with the largest agricultural industry, by far, is likely to be hardest hit.  The Las Vegas region may be far less vulnerable, having already taken aggressive steps to conserve and reuse water supply and charge economic rents for water usage.

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

Modeling Water Levels in Lake Mead

 

The monthly reported water levels in Lake Mead from Feb 1935 to June 2016 are shown in the chart below. The reference line is the drought level, historically defined as 1,125 feet.

 

 

One statistical technique widely applied in hydrology involves fitting a Kumaraswamy distribution to the relative water level.  According to the Arizona Game and Fish Department, the maximum lake level is 1229 feet.  We model the water level relative to the maximum level, as follows.

 

 

 

The fit of the distribution appears quite good, even in the tails:

 

ProbabilityPlot[relativeLevels, edist]

 

 

Since water levels have been below the drought level for some time, let’s instead consider the “emergency” level, 1,075 feet.  According to this model, there is just over a 6% chance of Lake Mead hitting the emergency level and, consequently, a high probability of breaching the emergency threshold some time over before the end of 2017.

 

 

 

 

 

 

One problem with this approach is that it assumes that each observation is drawn independently from a random variable with the estimated distribution.  In reality, there are high levels of autocorrelation in the series, as might be expected:  lower levels last month typically increase the likelihood of lower levels this month.  The chart of the autocorrelation coefficients makes this pattern clear, with statistically significant coefficients at lags of up to 36 months.

 

 

ts[“ACFPlot”]

 

 

 

An alternative methodology that enables us to take account of the autocorrelation in the process is time series analysis.  We proceed to fit an autoregressive moving average (ARMA) model as follows:

 

tsm = TimeSeriesModelFit[ts, “ARMA”]

 

The best fitting model in an ARMA(1,1) model, according to the AIC criterion:

 

 

Applying the fitted ARMA model, we forecast the water level in Lake Mead over the next ten years as shown in the chart below.  Given the mean-reverting moving average component of the model, it is not surprising to see the model forecasting a return to normal levels.

 

 

There is some evidence of lack of fit in the ARMA model, as shown in the autocorrelations of the model residuals:

 

 

 

 

A formal test reveals that residual autocorrelations at lags 4 and higher are jointly statistically significant:

 

 

 

The slowly decaying pattern of autocorrelations in the water level series suggests a possible “long memory” effect, which can be better modelled as a fractionally integrated process.  The forecasts from such a model, like the ARMA model forecasts, display a tendency to revert to a long term mean; but the reversion process is dampened by the reinforcing, long-memory effect captured in the FARIMA model.

 

 

The Prospects for the Next Decade

Taking the view that the water level in Lake Mead forms a stationary statistical process, the likelihood is that water levels will rise to 1,125 feet or more over the next ten years, easing the current water shortage in the region.

 

On the other hand, there are good reasons to believe that there are exogenous (deterministic) factors in play, specifically the over-consumption of water at a rate greater than the replenishment rate from average rainfall levels.  Added to this, plausible studies suggest that average rainfall in the Colorado Basin is expected to decline over the next fifty years.  Under this scenario, the water level in Lake Mead will likely continue to deteriorate, unless more stringent measures are introduced to regulate consumption.

 

 

Economic Impact

 

The drought in the South West affects far more than just the water levels in Lake Mead, of course.  One study found that California’s agriculture sector alone had lost $2.2Bn and some 17,00 season and part time jobs in 2014, due to drought.  Agriculture uses more than 80% of the State’s water, according to Fortune magazine, which goes on to identify the key industries most affected, including agriculture, food processing, semiconductors, energy, utilities and tourism.

In the energy sector, for example, the loss of hydroelectric power cost CA around $1.4Bn in 2014, according to non-profit research group Pacific Institute.

 

Although Intel pulled its last fabrication plant from California in 2009, semiconductor manufacturing is still a going concern in the state. Maxim Integrated, TowerJazz, and TSI Semiconductors all still have fabrication plants in the state. And they need a lot of water. A single semiconductor fabrication plant can use as much water as a small city. That means the current plants could represent three cities worth of consumption.

The drought is also bad news for water utilities, of course. The need to conserve water raises the priority on repair and maintenance, and that means higher costs and lower profit. Complicating the problem, California lacks any kind of management system for its water supply and can’t measure the inflows and outflows to ground water levels at any particular time.

 

The Bureau of Reclamation has studied more than two dozen options for conserving and increasing water supply, including importation, desalination and reuse. While some were disregarded for being too costly or difficult, the bureau found that the remaining options, if instituted, could yield 3.7 million acre feet per year in savings and new supplies, increasing to 7 million acre feet per year by 2060.  Agriculture is the biggest user by far and has to be part of any solution. In the near term, the agriculture industry could reduce its use by 10 to 15 percent without changing the types of crops it grows by using new technology, such as using drip irrigation instead of flood irrigation and monitoring soil moisture to prevent overwatering, the Pacific Institute found.

 

Conclusion

We can anticipate that a series of short term fixes, like the “Third Straw”, will be employed to kick the can down the road as far as possible, but research now appears almost unanimous in finding that drought is having a deleterious, long term affect on the economics of the South Western states.  Agriculture is likely to have to bear the brunt of the impact, but so too will adverse consequences be felt in industries as disparate as food processing, semiconductors and utilities.  California, with the largest agricultural industry, by far, is likely to be hardest hit.  The Las Vegas region may be far less vulnerable, having already taken aggressive steps to conserve and reuse water supply and charge economic rents for water usage.

 

 

 

 

 

 

 

 

 

The Amazon Killer

Amazon (NASDAQ:AMZN) has been on a tear over the last decade, especially since the financial crisis of 2008.  If you had been smart (or lucky) enough to buy the stock at the beginning of 2010, each $1,000 you invested would now be worth over $5,700, giving a CAGR of over 31%.

fig 1

Source:  Yahoo! Finance

It’s hard to argue with success, but could you have done better? The answer, surprisingly, is yes – and by a wide margin.

 SSALGOTRADING AD

Introducing AMZN+

I am going to reveal this mystery stock in due course and, I promise you, the investment recommendation is fully actionable.  For now, let’s just refer to it as AMZN+.

A comparison between investments made in AMZN and AMZN+ over the period from January 2010 to June 2016 is shown the chart following.

 

fig 2

 

 

Visually, there doesn’t appear to be much of a difference in overall performance.  However, it is apparent that AMZN+ is a great deal less volatile than its counterpart.

The table below give a more telling account of the relative outperformance by AMZN+.

table 1

 

The two investments are very highly correlated, with AMZN+ producing an extra 1% per annum in CAGR over the 6 ½ year period.

The telling distinction, however, lies on the risk side of the equation.  Here AMZN+ outperforms AMZN by a wide margin, with an annual standard deviation of only 21% compared to 29%.  What this means is that AMZN+ produces almost a 50% higher return than AMZN per unit of risk (information ratio 1.51 vs. 1.07).

The more conservative risk profile of AMZN+ is also reflected in lower maximum drawdown (-13.43% vs -23.74%), semi-deviation and higher Sortino Ratio (see this article for an explanation of these terms).

Bottom line:  you can produce around the same rate of return with substantially less risk by investing in AMZN+, rather than AMZN.

 

Comparing Equally Risky Investments

There is another way to make the comparison, which some investors might find more appealing.  Risk is, after all, an altogether more esoteric subject than return, which every investor understands.

So let’s say the investor adopts the same risk budget as for his original investment in AMZN, i.e. an annual volatility of just over 29%.  We can produce the same overall level of risk in AMZN+, equalizing the riskiness of the two investments, simply by leveraging the investment in AMZN+ by a factor of 1.36, using margin money.  i.e. we borrow $360 and invest a total of $1,360 in AMZN+, for each $1,000 we would have invested in AMZN.  Look at the difference in performance:

fig 3

 

The investor’s total return in AMZN+ would have been 963%, almost double the return in AMZN over the same period and with a CAGR of over 44.5%, more than 13% per annum higher than AMZN.

 

table 2

 

Note that, despite having an almost identical annual standard deviation, AMZN+ still enjoys a lower maximum drawdown and downside risk than AMZN.

 

The Big Reveal

Ok, so what is this mystery stock, AMZN+?  Actually it isn’t a stock:  it’s a simple portfolio, rebalanced monthly, with 66% of the investment being made in AMZN and 34% in the Direxion Daily 20+ Yr Trsy Bull 3X ETF (NYSEArca: TMF).

Well, that’s a little bit of a cheat, although not much of one:  it isn’t too much of a challenge to put $667 of every $1,000 in AMZN and the remaining $333 in TMF, rebalancing the portfolio at the end of every month.

The next question an investor might want to ask is: what other stocks could I apply this approach to?  The answer is: a great many of them.  And where you end up, ultimately, is with the discovery that you can eliminate a great deal of unnecessary risk with a portfolio of around 20-30 well-chosen assets.

 

The Fundamental Lesson from Portfolio Theory

We saw that AMZN incurred a risk of 29% in annual standard deviation, compared to only 21% for the AMZN+ portfolio.  What does the investor gain by taking that extra 8% in annual risk? Nothing at all – in fact he would have achieved a slightly worse return.

The key take-away from this simple example is the fundamental law of modern portfolio theory:

 

The market will not compensate an investor for taking diversifiable risk

 

As they say, diversification is the only free lunch on Wall Street.  So make the most of it.

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

Crash-Proof Investing

As markets continue to make new highs against a backdrop of ever diminishing participation and trading volume, investors have legitimate reasons for being concerned about prospects for the remainder of 2016 and beyond, even without consideration to the myriad of economic and geopolitical risks that now confront the US and global economies. Against that backdrop, remaining fully invested is a test of nerves for those whose instinct is that they may be picking up pennies in front an oncoming steamroller.  On the other hand, there is a sense of frustration in cashing out, only to watch markets surge another several hundred points to new highs.

In this article I am going to outline some steps investors can take to match their investment portfolios to suit current market conditions in a way that allows them to remain fully invested, while safeguarding against downside risk.  In what follows I will be using our own Strategic Volatility Strategy, which invests in volatility ETFs such as the iPath S&P 500 VIX ST Futures ETN (NYSEArca:VXX) and the VelocityShares Daily Inverse VIX ST ETN (NYSEArca:XIV), as an illustrative example, although the principles are no less valid for portfolios comprising other ETFs or equities.

SSALGOTRADING AD

Risk and Volatility

Risk may be defined as the uncertainty of outcome and the most common way of assessing it in the context of investment theory is by means of the standard deviation of returns.  One difficulty here is that one may never ascertain the true rate of volatility – the second moment – of a returns process; one can only estimate it.  Hence, while one can be certain what the closing price of a stock was at yesterday’s market close, one cannot say what the volatility of the stock was over the preceding week – it cannot be observed the way that a stock price can, only estimated.  The most common estimator of asset volatility is, of course, the sample standard deviation.  But there are many others that are arguably superior:  Log-Range, Parkinson, Garman-Klass to name but a few (a starting point for those interested in such theoretical matters is a research paper entitled Estimating Historical Volatility, Brandt & Kinlay, 2005).

Leaving questions of estimation to one side, one issue with using standard deviation as a measure of risk is that it treats upside and downside risk equally – the “risk” that you might double your money in an investment is regarded no differently than the risk that you might see your investment capital cut in half.  This is not, of course, how investors tend to look at things: they typically allocate a far higher cost to downside risk, compared to upside risk.

One way to address the issue is by using a measure of risk known as the semi-deviation.  This is estimated in exactly the same way as the standard deviation, except that it is applied only to negative returns.  In other words, it seeks to isolate the downside risk alone.

This leads directly to a measure of performance known as the Sortino Ratio.  Like the more traditional Sharpe Ratio, the Sortino Ratio is a measure of risk-adjusted performance – the average return produced by an investment per unit of risk.  But, whereas the Sharpe Ratio uses the standard deviation as the measure of risk, for the Sortino Ratio we use the semi-deviation. In other words, we are measuring the expected return per unit of downside risk.

There may be a great deal of variation in the upside returns of a strategy that would penalize the risk-adjusted returns, as measured by its Sharpe Ratio. But using the Sortino Ratio, we ignore the upside volatility entirely and focus exclusively on the volatility of negative returns (technically, the returns falling below a given threshold, such as the risk-free rate.  Here we are using zero as our benchmark).  This is, arguably, closer to the way most investors tend to think about their investment risk and return preferences.

In a scenario where, as an investor, you are particularly concerned about downside risk, it makes sense to focus on downside risk.  It follows that, rather than aiming to maximize the Sharpe Ratio of your investment portfolio, you might do better to focus on the Sortino Ratio.

 

Factor Risk and Correlation Risk

Another type of market risk that is often present in an investment portfolio is correlation risk.  This is the risk that your investment portfolio correlates to some other asset or investment index.  Such risks are often occluded – hidden from view – only to emerge when least wanted.  For example, it might be supposed that a “dollar-neutral” portfolio, i.e. a portfolio comprising equity long and short positions of equal dollar value, might be uncorrelated with the broad equity market indices.  It might well be.  On the other hand, the portfolio might become correlated with such indices during times of market turbulence; or it might correlate positively with some sector indices and negatively with others; or with market volatility, as measured by the CBOE VIX index, for instance.

Where such dependencies are included by design, they are not a problem;  but when they are unintended and latent in the investment portfolio, they often create difficulties.  The key here is to test for such dependencies against a variety of risk factors that are likely to be of concern.  These might include currency and interest rate risk factors, for example;  sector indices; or commodity risk factors such as oil or gold (in a situation where, for example, you are investing a a portfolio of mining stocks).  Once an unwanted correlation is identified, the next step is to adjust the portfolio holdings to try to eliminate it.  Typically, this can often only be done in the average, meaning that, while there is no correlation bias over the long term, there may be periods of positive, negative, or alternating correlation over shorter time horizons.  Either way, it’s important to know.

Using the Strategic Volatility Strategy as an example, we target to maximize the Sortino Ratio, subject also to maintaining very lows levels of correlation to the principal risk factors of concern to us, the S&P 500 and VIX indices. Our aim is to create a portfolio that is broadly impervious to changes in the level of the overall market, or in the level of market volatility.

 

One method of quantifying such dependencies is with linear regression analysis.  By way of illustration, in the table below are shown the results of regressing the daily returns from the Strategic Volatility Strategy against the returns in the VIX and S&P 500 indices.  Both factor coefficients are statistically indistinguishable from zero, i.e. there is significant no (linear) dependency.  However, the constant coefficient, referred to as the strategy alpha, is both positive and statistically significant.  In simple terms, the strategy produces a return that is consistently positive, on average, and which is not dependent on changes in the level of the broad market, or its volatility.  By contrast, for example, a commonplace volatility strategy that entails capturing the VIX futures roll would show a negative correlation to the VIX index and a positive dependency on the S&P500 index.

Regression

 

Tail Risk

Ever since the publication of Nassim Taleb’s “The Black Swan”, investors have taken a much greater interest in the risk of extreme events.  If the bursting of the tech bubble in 2000 was not painful enough, investors surely appear to have learned the lesson thoroughly after the financial crisis of 2008.  But even if investors understand the concept, the question remains: what can one do about it?

The place to start is by looking at the fundamental characteristics of the portfolio returns.  Here we are not such much concerned with risk, as measured by the second moment, the standard deviation. Instead, we now want to consider the third and forth moments of the distribution, the skewness and kurtosis.

Comparing the two distributions below, we can see that the distribution on the left, with negative skew, has nonzero probability associated with events in the extreme left of the distribution, which in this context, we would associate with negative returns.  The distribution on the right, with positive skew, is likewise “heavy-tailed”; but in this case the tail “risk” is associated with large, positive returns.  That’s the kind of risk most investors can live with.

 

skewness

 

Source: Wikipedia

 

 

A more direct measure of tail risk is kurtosis, literally, “heavy tailed-ness”, indicating a propensity for extreme events to occur.  Again, the shape of the distribution matters:  a heavy tail in the right hand portion of the distribution is fine;  a heavy tail on the left (indicating the likelihood of large, negative returns) is a no-no.

Let’s take a look at the distribution of returns for the Strategic Volatility Strategy.  As you can see, the distribution is very positively skewed, with a very heavy right hand tail.  In other words, the strategy has a tendency to produce extremely positive returns. That’s the kind of tail risk investors prefer.

SVS

 

Another way to evaluate tail risk is to examine directly the performance of the strategy during extreme market conditions, when the market makes a major move up or down. Since we are using a volatility strategy as an example, let’s take a look at how it performs on days when the VIX index moves up or down by more than 5%.  As you can see from the chart below, by and large the strategy returns on such days tend to be positive and, furthermore, occasionally the strategy produces exceptionally high returns.

 

Convexity

 

The property of producing higher returns to the upside and lower losses to the downside (or, in this case, a tendency to produce positive returns in major market moves in either direction) is known as positive convexity.

 

Positive convexity, more typically found in fixed income portfolios, is a highly desirable feature, of course.  How can it be achieved?    Those familiar with options will recognize the convexity feature as being similar to the concept of option Gamma and indeed, one way to produce such a payoff is buy adding options to the investment mix:  put options to give positive convexity to the downside, call options to provide positive convexity to the upside (or using a combination of both, i.e. a straddle).

 

In this case we achieve positive convexity, not by incorporating options, but through a judicious choice of leveraged ETFs, both equity and volatility, for example, the ProShares UltraPro S&P500 ETF (NYSEArca:UPRO) and the ProShares Ultra VIX Short-Term Futures ETN (NYSEArca:UVXY).

 

Putting It All Together

While we have talked through the various concepts in creating a risk-protected portfolio one-at-a-time, in practice we use nonlinear optimization techniques to construct a portfolio that incorporates all of the desired characteristics simultaneously. This can be a lengthy and tedious procedure, involving lots of trial and error.  And it cannot be emphasized enough how important the choice of the investment universe is from the outset.  In this case, for instance, it would likely be pointless to target an overall positively convex portfolio without including one or more leveraged ETFs in the investment mix.

Let’s see how it turned out in the case of the Strategic Volatility Strategy.

 

SVS Perf

 

 

Note that, while the portfolio Information Ratio is moderate (just above 3), the Sortino Ratio is consistently very high, averaging in excess of 7.  In large part that is due to the exceptionally low downside risk, which at 1.36% is less than half the standard deviation (which is itself quite low at 3.3%).  It is no surprise that the maximum drawdown over the period from 2012 amounts to less than 1%.

A critic might argue that a CAGR of only 10% is rather modest, especially since market conditions have generally been so benign.  I would answer that criticism in two ways.  Firstly, this is an investment that has the risk characteristics of a low-duration government bond; and yet it produces a yield many times that of a typical bond in the current low interest rate environment.

Secondly, I would point out that these results are based on use of standard 2:1 Reg-T leverage. In practice it is entirely feasible to increase the leverage up to 4:1, which would produce a CAGR of around 20%.  Investors can choose where on the spectrum of risk-return they wish to locate the portfolio and the strategy leverage can be adjusted accordingly.

 

Conclusion

The current investment environment, characterized by low yields and growing downside risk, poses difficult challenges for investors.  A way to address these concerns is to focus on metrics of downside risk in the construction of the investment portfolio, aiming for high Sortino Ratios, low correlation with market risk factors, and positive skewness and convexity in the portfolio returns process.

Such desirable characteristics can be achieved with modern portfolio construction techniques providing the investment universe is chosen carefully and need not include anything more exotic than a collection of commonplace ETF products.

Developing A Volatility Carry Strategy

By way of introduction we begin by reviewing a well known characteristic of the  iPath S&P 500 VIX ST Futures ETN (NYSEArca:VXX).  In common with other long-volatility ETF /ETNs, VXX has a tendency to decline in value due to the upward sloping shape of the forward volatility curve.  The chart below which illustrates the fall in value of the VXX, together with the front-month VIX futures contract, over the period from 2009.


VXXvsVX

 

 

This phenomenon gives rise to opportunities for “carry” strategies, wherein a long volatility product such as VXX is sold in expectation that it will decline in value over time.  Such strategies work well during periods when volatility futures are in contango, i.e. when the longer dated futures contracts have higher prices than shorter dated futures contracts and the spot VIX Index, which is typically the case around 70% of the time.  An analogous strategy in the fixed income world is known as “riding down the yield curve”.  When yield curves are upward sloping, a fixed income investor can buy a higher-yielding bill or bond in the expectation that the yield will decline, and the price rise, as the security approaches maturity.  Quantitative easing put paid to that widely utilized technique, but analogous strategies in currency and volatility markets continue to perform well.

The challenge for any carry strategy is what happens when the curve inverts, as futures move into backwardation, often giving rise to precipitous losses.  A variety of hedging schemes have been devised that are designed to mitigate the risk.  For example, one well-known carry strategy in VIX futures entails selling the front month contract and hedging with a short position in an appropriate number of E-Mini S&P 500 futures contracts. In this case the hedge is imperfect, leaving the investor the task of managing a significant basis risk.

SSALGOTRADING AD

The chart of the compounded value of the VXX and VIX futures contract suggests another approach.  While both securities decline in value over time, the fall in the value of the VXX ETN is substantially greater than that of the front month futures contract.  The basic idea, therefore, is a relative value trade, in which we purchase VIX futures, the better performing of the pair, while selling the underperforming VXX.  Since the value of the VXX is determined by the value of the front two months VIX futures contracts, the hedge, while imperfect, is likely to entail less basis risk than is the case for the VIX-ES futures strategy.

Another way to think about the trade is this:  by combining a short position in VXX with a long position in the front-month futures, we are in effect creating a residual exposure in the value of the second month VIX futures contract relative to the first. So this is a strategy in which we are looking to capture volatility carry, not at the front of the curve, but between the first and second month futures maturities.  We are, in effect, riding down the belly of volatility curve.

 

The Relationship between VXX and VIX Futures

Let’s take a look at the relationship between the VXX and front month futures contract, which I will hereafter refer to simply as VX.  A simple linear regression analysis of VXX against VX is summarized in the tables below, and confirms two features of their relationship.

Firstly there is a strong, statistically significant relationship between the two (with an R-square of 75% ) – indeed, given that the value of the VXX is in part determined by VX, how could there not be?

Secondly, the intercept of the regression is negative and statistically significant.  We can therefore conclude that the underperformance of the VXX relative to the VX is not just a matter of optics, but is a statistically reliable phenomenon.  So the basic idea of selling the VXX against VX is sound, at least in the statistical sense.

Regression

 

 

Constructing the Initial Portfolio

In constructing our theoretical portfolio, I am going to gloss over some important technical issues about how to construct the optimal hedge and simply assert that the best one can do is apply a beta of around 1.2, to produce the following outcome:

Table1

VXX-VX Strategy

 

While broadly positive, with an information ratio of 1.32, the strategy performance is a little discouraging, on several levels.  Firstly, the annual volatility, at over 48%, is uncomfortably high. Secondly, the strategy experiences very substantial drawdowns at times when the volatility curve inverts, such as in August 2015 and January 2016.  Finally, the strategy is very highly correlated with the S&P500 index, which may be an important consideration for investors looking for ways to diversity their stock portfolio risk.

 

Exploiting Calendar Effects

We will address these issues in short order.  Firstly, however, I want to draw attention to an interesting calendar effect in the strategy (using a simple pivot table analysis).

Calendar

As you can see from the table above, the strategy returns in the last few days of the calendar month tend to be significantly below zero.

The cause of the phenomenon has to do with the way the VXX is constructed, but the important point here is that, in principle, we can utilize this effect to our advantage, by reversing the portfolio holdings around the end of the month.  This simple technique produces a significant improvement in strategy returns, while lowering the correlation:

Table2

 

Reducing Portfolio Risk and Correlation

We can now address the issue of the residual high level of strategy volatility, while simultaneously reducing the strategy correlation to a much lower level.  We can do this in a straightforward way by adding a third asset, the SPDR S&P 500 ETF Trust (NYSEArca:SPY), in which we will hold a short position, to exploit the negative correlation of the original portfolio.

We then adjust the portfolio weights to maximize the risk-adjusted returns, subject to limits on the maximum portfolio volatility and correlation.  For example, setting a limit of 10% for both volatility and correlation, we achieve the following result (with weights -0.37 0.27 -0.65 for VXX, VX and SPY respectively):

 

Table3

 

 

VXX-VX-SPY

 

Compared to the original portfolio, the new portfolio’s performance is much more benign during the critical period from Q2-2015 to Q1-2016 and while there remain several significant drawdown periods, notably in 2011, overall the strategy is now approaching an investable proposition, with an information ratio of 1.6 and annual volatility of 9.96% and correlation of 0.1.

Other configurations are possible, of course, and the risk-adjusted performance can be improved, depending on the investor’s risk preferences.

 

Portfolio Rebalancing

There is an element of curve-fitting in the research process as described so far, in as much as we are using all of the available data to July 2016 to construct a portfolio with the desired characteristics. In practice, of course, we will be required to rebalance the portfolio on a periodic basis, re-estimating the optimal portfolio weights as new data comes in.  By way of illustration, the portfolio was re-estimated using in-sample data to the end of Feb, 2016, producing out-of-sample results during the period from March to July 2016, as follows:

Table4

 

A detailed examination of the generic problem of how frequently to rebalance the portfolio is beyond the scope of this article and I leave it to interested analysts to perform the research for themselves.

 

Practical Considerations

In order to implement the theoretical strategy described above there are several important practical steps that need to be considered.

 

  • It is not immediately apparent how the weights should be applied to a portfolio comprising both ETNs and futures. In practice the best approach is to re-estimate the portfolio using a regression relationship expressed in $-value terms, rather than in percentages, in order to establish the quantity of VXX and SPY stock to be sold per single VX futures contract.
  • Reversing the portfolio holdings in the last few days of the month will add significantly to transaction costs, especially for the position in VX futures, for which the minimum tick size is $50. It is important to factor realistic estimates of transaction costs into the assessment of the strategy performance overall and specifically with respect to month-end reversals.
  • The strategy assumed  the availability of VXX and SPY to short, which occasionally can be a problem. It’s not such a big deal if you are maintaining a long-term short position, but flipping the position around over a few ays at the end of the month might be problematic, from time to time.
  • Also, we should take account of stock loan financing costs, which run to around 2.9% and 0.42% annually for VXX and SPY, respectively. These rates can vary with market conditions and stock availability, of course.
  • It is highly likely that other ETFs/ETNs could profitably be added to the mix in order to further reduce strategy volatility and improve risk-adjusted returns. Likely candidates could include, for example, the Direxion Daily 20+ Yr Trsy Bull 3X ETF (NYSEArca:TMF).
  • We have already mentioned the important issue of portfolio rebalancing. There is an argument for rebalancing more frequently to take advantage of the latest market data; on the other hand, too-frequent changes in the portfolio composition can undermine portfolio robustness, increase volatility and incur higher transaction costs. The question of how frequently to rebalance the portfolio is an important one that requires further testing to determine the optimal rebalancing frequency.

 

Conclusion

We have described the process of constructing a volatility carry strategy based on the relative value of the VXX ETN vs the front-month contract in VIX futures.  By combining a portfolio comprising short positions in VXX and SPY with a long position in VIX futures, the investor can, in principle achieve risk-adjusted returns corresponding to an information ratio of around 1.6, or more. It is thought likely that further improvements in portfolio performance can be achieved by adding other ETFs to the portfolio mix.

 

Is Internal Bar Strength A Random Walk? The Case of Exxon-Mobil

For those who prefer a little more rigor in their quantitative research, I can offer more a somewhat more substantive statistical argument in favor of the IBS indicator discussed in my previous post.

Specifically, we can show quite convincingly that the IBS process is stationary, a highly desirable property much sought-after in, for example, the construction of statistical arbitrage strategies.  Of course, by construction, the IBS is constrained to lie between the values of 0 and 1, so non-stationarity in the mean is highly unlikely.  But, conceivably, there could be some time dependency in the process or in its variance, for instance.  Then there is the further question as to whether the IBS indicator is mean-reverting, which would indicate that the underlying price process likewise has a tendency to mean revert.

Let’s take the IBS series for Exxon-Mobil (XOM) as an example to work with. I have computed the series from the beginning of 1990, and the first 100 values are shown in the plot below.


XOMIBS

 

 

Autocorrelation and Unit Root Tests

There appears to be little patterning in the process autocorrelations, and this is confirmed by formal statistical tests which fail to reject the null hypothesis that the first 20 autocorrelations are not, collectively, statistically significant.

XOMIBS Autocorrelations

 

XOMIBS acf test

 

SSALGOTRADING AD

Next we test for the presence of a unit root in the IBS process (highly unlikely, given its construction) and indeed, unsurprisingly, the null hypothesis of a unit root is roundly rejected by the Dickey-Fuller and Phillips-Perron tests.

 

XOMIBS Unit Root

 

Variance Ratio Tests

We next conduct a formal test to determine whether the IBS series follows a random walk.

The variance ratio test assesses the null hypothesis that a univariate time series y is a random walk. The null model is

y(t) = c + y(t–1) + e(t),

where c is a drift constant (assumed zero for the IBS series) and e(t) are uncorrelated innovations with zero mean.

  • When IID is false, the alternative is that the e(t) are correlated.
  • When IID is true, the alternative is that the e(t) are either dependent or not identically distributed (for example, heteroscedastic).

 

We test whether the XOM IBS series is a random walk using various step sizes and perform the test with and without the assumption that the innovations are independent and identically distributed.

Switching to Matlab, we proceed as follows:

q = [2 4 8 2 4 8];
flag = logical([1 1 1 0 0 0]);
[h,pValue,stat,cValue,ratio] = vratiotest(XOMIBS,’period’,q,’IID’,flag)

Here h is a vector of Boolean decisions for the tests, with length equal to the number of tests. Values of h equal to 1 indicate rejection of the random-walk null in favor of the alternative. Values of h equal to 0 indicate a failure to reject the random-walk null.

The variable ratio is a vector of variance ratios, with length equal to the number of tests. Each ratio is the ratio of:

  • The variance of the q-fold overlapping return horizon
  • q times the variance of the return series

For a random walk, these ratios are asymptotically equal to one. For a mean-reverting series, the ratios are less than one. For a mean-averting series, the ratios are greater than one.

For the XOM IBS process we obtain the following results:

h =  1   1   1   1   1   1
pValue = 1.0e-51 * [0.0000 0.0000 0.0000 0.0000 0.0000 0.1027]
stat = -27.9267 -21.7401 -15.9374 -25.1412 -20.2611 -15.2808
cValue = 1.9600 1.9600 1.9600 1.9600 1.9600 1.9600
ratio = 0.4787 0.2405 0.1191 0.4787 0.2405 0.1191

The random walk hypothesis is convincingly rejected for both IID and non-IID error terms.  The very low ratio values  indicate that the IBS process is strongly mean reverting.

 

Conclusion

While standard statistical tests fail to find evidence of any non-stationarity in the Internal Bar Strength signal for Exxon-Mobil, the hypothesis that the series follows a random walk (with zero drift) is roundly rejected by variance ratio tests.  These tests also confirm that the IBS series is strongly mean reverting, as we previously discovered empirically.

This represents an ideal scenario for trading purposes: a signal with the highly desirable properties that is both stationary and mean reverting.  In the case of Exxon-Mobil, there appears to be clear evidence from both statistical tests and empirical trading strategies using the Internal Bar Strength indicator that the tendency of the price series to mean-revert is economically as well as statistically significant.

The Internal Bar Strength Indicator

Internal Bar Strength (IBS) is an idea that has been around for some time.  IBS is based on the position of the day’s close in relation to the day’s range: it takes a value of 0 if the closing price is the lowest price of the day, and 1 if the closing price is the highest price of the day.

More formally:

IBS  =  (Close – Low) / (High – Low)

The IBS effect may be related to intraday over-reaction to news or market movements, which are then ”corrected” the next day.  It serves as a measure of the tendency of a price series to mean-revert over daily horizons.  I use the term “daily” advisedly: so far as I am aware, there has been no research (including my own) demonstrating the existence of an IBS effect at time horizons shorter, or longer, than one day.  Indeed, there has been very little in the way of academic research into the concept of any kind, which is strange considering how compelling are the results it is capable of producing.  Practitioners have been happy enough with that state of affairs, content to deploy this neglected indicator in their trading strategies, where it has often proved to be extremely useful (we use IBS in one of our volatility strategies). Since 2013, however, the cat has been let out of the bag, thanks to an excellent research paper by Alexander Pagonidis, who writes an interesting quantitative finance blog.

The essence of the idea is that stocks that close in the lowest part of the daily range, with an IBS of below, say, 0.2, will tend to rally the next day, while stocks that close in the highest quintile will often decline in value in the following session.  In his paper “The IBS Effect: Mean Reversion in Equity ETFs” (2013), Pagonidis researches the IBS effect in equity index ETFs in the US and several international markets.  He confirms that low IBS values in these assets are associated with high returns in the following day session, while high IBS values are associated with low returns. Average returns when IBS is below 0.20 are .35% ,while average returns when IBS is above 0.80 are -0.13%. According to his research, this effect has been present in equity ETFs since the early 90s and has been highly consistent through time.

SSALGOTRADING AD

IBS Strategy Performance

To give the reader some idea of the potential of the IBS effect, I have reproduced below equity curves for the IBS strategy for the SPDR S&P 500 ETF Trust (SPY) and iShares MSCI Singapore ETF (EWS) index ETFs over the period from 1999 to 2016.  The strategy buys at the close when IBS is below 0.2, and sells at the close when IBS exceeds 0.8, liquidating the position at the following market close. Strategy CAGR over the period has been of the order of 13% for SPY and as high as 40% for EWS, ignoring transaction costs.

IBS Strategy Chart SPY EWS

 

Note that in both cases strategy returns for SPY and EWS have diminished in recent years, turning negative in 2015 and 2016 YTD and this is true for ETFs in general.  It remains to be seen whether this deterioration in strategy performance is temporary or permanent.  There are some indications that the latter holds true, but the evidence is not quite definitive.  For example, the chart below shows daily equity curve for the SPY IBS strategy, with 95% confidence intervals for the latest 100 trades (up to the end of May 2016), constructed using Monte-Carlo bootstrap.  The equity curve appears to have penetrated the lower bound, indicating a statistically significant deterioration in the performance of the IBS strategy for SPY over the last year or so (EWS is similar).  That said, the equity curve does fall inside the boundaries of the 99% confidence interval, so those looking for greater certainty about the possible breakdown of the effect will need to wait a little longer for confirmation.

 

SPY IBS MSA

 

Whatever the outcome may be for SPY and other ETFs going forward, it is certainly true that IBS effects persist strongly for some individual equities, Exxon-Mobil Corp. (XOM) being a case in point (see below).  It’s worth taking note of the exceptional performance of the XOM IBS strategy during the latter quarter of 2008.  I will have much more to say on the application of the IBS indicator for individual equities in a future blog post.

 

XOM IBS Strategy

 

The Role of Range, Volume, Bull/Bear Markets, Volatility and Seasonality

Pagonidis goes on to detail several further important findings in relation to IBS.  It is clear from his research that high volatility is related to increased predictability of returns and a more powerful IBS effect, in particular the high IBS-negative return aspect.  As might be expected, the effect is also larger after days with high range, both for high and low IBS extremes.

Volume turns out to be especially important for  U.S. index ETFs:  in fact, the IBS effect only appears to work on high-volume days.

Pagonidis also separates the data into bull and bear market environments, based on whether 200-day returns are positive or not.  The size of the effect is roughly similar in each environment (slightly larger in bear markets), but it is greater in the direction of the overall trend: high IBS readings are followed by larger negative returns during bear markets, and vice versa.

Day of Week Effect

The IBS effect is also strongly seasonal, having the greatest impact on returns from Monday’s close to Tuesday’s close, as illustrated for the SPY ETF in the chart below.  This accounts for the phenomenon known popularly as “Turnaround Tuesday”, i.e. the tendency for the market to recover strongly from losses on a Monday.  The day-of-week effect is weakest for Fridays.

 

SPY DOW

 

The mean of the returns distribution is not the only aspect that IBS can predict. Skewness also varies significantly between IBS buckets, with low IBS readings being followed by highly skewed returns, and vice versa. Close-to-close returns after a bottom-bucket IBS day have average skewness of 0.65 across Equity Index ETF products, while top-bucket IBS days are followed by returns with skewness of 0.03. This finding has very useful risk management applications for investors concerned with tail risk.

IBS as a Filter for a Swing Trading Strategy in QQQ

The returns to an IBS-only strategy are both statistically and economically significant. However, commissions will greatly decrease the returns and increase the maximum drawdowns, however, making such an approach challenging in the real world. One alternative is to combine the IBS effect with mean reversion on longer timescales and only take trades when they align.

Pagonidis offers a simple demonstration using the Cutler’s RSI indicator that shows how the IBS effect can be used to boost returns of a swing trading strategy while significantly decreasing the number of trades needed.

Cutler’s RSI at time t is calculated as follows:

 

RSI

 

Pagonidis tests a simple, long-only strategy that trades the PowerShares QQQ Trust, Series 1 (QQQ) ETF using the Cutler’s RSI(3) indicator:

• Go long at the close if RSI(3) < 10

• Maintain the position while RSI(3) ≤ 40

 filter these returns by adding an additional rule based on the value of IBS:

• Enter or maintain long position only if IBS ≤ 0.5

Pangonis claims that the strategy produces rather promising results that “easily beats commissions”;  however, my own rendition of the strategy, assuming commissions of $0.005 per share and slippage of a further $0.02 per share produces results that are distinctly less encouraging:

EC0

 

Pef0

Strategy Code

For those interested, the code is as follows:

Inputs:
RSILen(3),
RSI_Entry(10),
RSI_Exit(40),
IBS_Threshold(0.5),
Initial_Capital(100000);
Vars:
nShares(100),
RSIval(0),
IBS(0);
RSIval=RSI(C,RSILen);
IBS = (C-L)/(H-L);

nShares = Round(Initial_Capital / Close,0);

If Marketposition = 0 and RSIval > RSI_Entry and IBS < IBS_Threshold then begin
Buy nShares contracts next bar at market;
end;
If Marketposition > 0 and ((RSIval > RSI_Exit) or (IBS_Threshold > IBS_Threshold)) then begin
Sell next bar at market;
end;

Strategy Optimization and Robustness Testing

One can further improve performance by optimizing the trading system parameters, using Tradestation’s excellent Walk Forward Optimization (WFO) module.  This allows us to examine the effect of re-calibrating the strategy parameters are regular intervals, testing the optimized model on out-of-sample data sets of various sizes.  WFO can be used, not only optimize a strategy, but also to examine the sensitivity of its performance to changes in the levels of key parameters.  For example, in the case of the QQQ swing trading strategy, we find that profitability increases monotonically with the length of the RSI indicator, and this effect is especially marked when an IBS threshold level of 0.2 is used:

Sensitivity

 

Likewise we can test the consistency of the day-of-the-week effect over several OS data sets of  varying size and these tests are consistent with the pattern seen earlier for the IBS indicator, confirming its role as a filter rule in enhancing system profitability:

Distribution Analysis

 

A model that is regularly re-calibrated using WFO is subjected to a series of tests designed to ensure its robustness and consistency in live trading.   The tests include the following:

 

WFO

 

In order to achieve an overall pass rating, the system is required to pass all five tests of its out-of-sample performance, from which Tradestation deems it likely that the system will continue to perform well in live trading.  The results from this procedure appear much more promising than the strategy in its original form, as can be seen from the performance table and equity curve chart shown below.

EC1

Perf1

 

However, these results include both in-sample and out-of-sample periods.  An examination of the results from the WFO indicate that the overall efficiency of the strategy is around 55%, meaning that the P&L produced by the system in out-of-sample periods amounts to a little over one half of the rate of profit produced during in-sample periods.  Going forward, therefore, we might expect the performance of the system in live trading to be only around half as good as shown here.  While this is still superior to the original system, it may not be considered good enough.  Nonetheless, for the purpose of illustrating the benefits of the IBS indicator as a trade filter, it makes the point.

Another interesting example of an IBS-based trading strategy in the QQQ and SPY ETFs can be found in the following blog post.

Conclusion

Internal Bar Strength is a powerful mean-reversion indicator for equity products traded at daily frequencies, with a consistent effect that has continued from the 1990s through to the current decade. IBS can be used on its own in mean-reversion strategies that have worked well for both US equities and US and International equity index ETFs, or used as a trade filter when combined with other alpha signals.

While there is evidence of a weakening of the IBS effect since around 2013 this is not yet confirmed statistically (at the 99% confidence level) and may simply be the result of normal statistical variation in its efficacy.

 

 

Seasonal Effects in Equity Markets

There are a plethora of seasonal anomalies documented in academic research.  For equities these include the Halloween effect (“Sell in May”), January effect, turn-of-the-month effect, weekend effect and holiday effect. For example, Bouman and Jacobsen (2002) and Jacobsen and Visaltanachoti (2009) provide empirical evidence on the Halloween effect, Haug and Hirschey (2006) on the January effect, Lakonishok and Smidt (1988) on the turn-of-the-month (TOM) effect, Cross (1973) on the weekend effect, and Ariel (1990) on the holiday effect.

An excellent paper entitled An Anatomy of Calendar Effects in the Journal of Asset Management (13(4), 2012, pp. 271-286) by Laurens Swinkels of Erasmus University Rotterdam and Pim van Vliet of Robeco Asset Management gives a very good account of the various phenomena and their relative importance.  Using daily returns data on the US value-weighted equity market over the period from July 1963 to December 2008, the researchers find that Halloween and turn-of-the-month (TOM) are the strongest effects, fully diminishing the other three effects to zero. The equity premium over the sample 1963-2008 is 7.2% if there is a Halloween or TOM effect, and -2.8% in all other cases. These findings are robust with respect to transactions costs, across different samples, market segments, and international stock markets. Their empirical research narrows down the number of calendar effects from five to two, leading to a more powerful and puzzling summary of seasonal effects.

SSALGOTRADING AD

The two principal effects are illustrated here with reference to daily returns in the S&P 500 Index, using data from 1980-2016.

Halloween Effect

The Halloween effect refers to the tendency of markets to perform better in the six month period from November to April, compared to the half year from May to October.  In fact, for the S&P 500 index itself, performance during the months of May and October has historically been above the monthly average, as you can see in the chart below.  According to this analysis, the period to avoid spans the four months from June to September, with September being the “cruelest month”, by far.  Note that, between them, the months of November, December and April account for over 50% of the average annual return in the index since 1980.

Halloween Effect S&P500 Index

 

Turn-of-the-Month Effect

The TOM effect refers to the finding that above average returns tend to occur on the last trading day of the month and (up to) the first four trading days of the new calendar month. For the S&P 500 index the TOM effect spans a shorter period comprising the last trading day of the month and the first two trading days of the new month.  It is worth noting also the anomalous positive returns arising on 16th – 18th of the month and negative returns around the 19th and 20th of the month.    My speculative guess is that these mid-month effects arise from futures/option expiration.

TOM Effect S&P500 Index

Seasonal Tactical  Allocation

Let’s assume we allocate to equities (in the form of the S&P 500 Index) only during the period from October to May, or on the last or first two trading days of each month. How do the returns from that seasonal portfolio compare to the benchmark buy and hold portfolio?  If we ignore transaction costs (and income from riskless Treasury investments when we are out of the market), the seasonal portfolio outperforms the buy and hold benchmark over the 36 year period since 1980 by around 88bp per annum (continuously compounded), and with an annual volatility that is 258bp lower.  The outperformance of the seasonal portfolio becomes particularly noticeable after the 2000/2001 crash.

 

perfSeasonal

 

Seasonal vsB&H

 

A much more rigorous analysis of the performance characteristics of the seasonal portfolio is given in the research paper, taking account of transaction costs, with summary results as follows:

 

table 5

 

fig3

Conclusion

There is a sizable body of credible academic research demonstrating the importance of calendar effects and this paper suggests that investors’ focus should be on the Halloween and TOM effects in particular.  A tactical allocation program that increases the allocation to equities towards the end of the month and first few trading days of the new month, and during the November to April calendar months is likely to significantly outperform a buy-and-hold portfolio, according to these findings.

There remain unaccounted-for seasonal effects in the mid-section of the month that may arise from the expiration of futures and option contracts. These are worthy of further investigation.

Trading With Indices

In this post I want to discuss ways to make use of signals from relevant market indices in your trading.  These signals can add value regardless of whether you trade algorithmically or manually.  The techniques described here are one of the most widely applicable in the quantitative analyst’s arsenal.

Let’s motivate the discussion by looking an example of a simple trading system trading the VIX on weekly bars.  Performance results for the system are summarized in the chart and table below.  The system outperforms the buy and hold return by a substantial margin, with a profit factor of over 3 and a win rate exceeding 82%.  What’s not to like?

VIX EC

VIX Performance

Well, for one thing, this isn’t really a trading system – because the VIX Index itself isn’t tradable. So the performance results are purely notional (and, if you didn’t already notice, no slippage or commission is included).

It is very easy to build high-performing trading system in indices – because they are not traded products,  index prices are often stale and tend to “follow” the price action in the equivalent traded market.

This particular system for the VIX Index took me less than ten minutes to develop and comprises only a few lines of code.  The system makes use of a simple RSI indicator to decide when to buy or sell the index.  I optimized the indicator parameters (separately for long and short) over the period to 2012, and tested it out-of-sample on the data from 2013-2016.

inputs:
Price( Close ) ,
Length( 14 ) ,
OverSold( 30 ) ;

variables:
RSIValue( 0 );

RSIValue = RSI( Price, Length );
if CurrentBar > 1 and RSIValue crosses over OverSold then
Buy ( !( “RsiLE” ) ) next bar at market;

.

The daily system I built for the S&P 500 Index is a little more sophisticated than the VIX model, and produces the following results.

SP500 EC

SP500 Perf

 

Using Index Trading Systems

We have seen that its trivially easy to build profitable trading systems for index products.  But since they can’t be traded, what’s the point?

The analyst might be tempted by the idea of using the signals generated by an index trading system to trade a corresponding market, such as VIX or eMini futures.  However, this approach is certain to fail.  Index prices lag the prices of equivalent futures products, where traders first monetize their view on the market.  So using an index strategy directly to trade a cash or futures market would be like trying to trade using prices delayed by a few seconds, or minutes – a recipe for losing money.

SSALGOTRADING AD

Nor is it likely that a trading system developed for an index product will generalize to a traded market.  What I mean by this is that if you were to take an index strategy, such as the VIX RSI strategy, transfer it to VIX futures and tweak the parameters in the hope of producing a profitable system, you are likely to be disappointed. As I have shown, you can produce a profitable index trading system using the simplest and most antiquated trading concepts (such as the RSI index) that long ago ceased to offer any predictive value in actual traded markets.  Index markets are actually inefficient – the prices of index products often fail to fully reflect all relevant, available information in a timely way. Such simple inefficiencies are easily revealed by indicators such as moving averages.  Traded markets, by contrast, are highly efficient and, with the exception of HFT, it is going to take a great deal more than a simple moving average to provide insight into the few inefficiencies that do arise.

bullbear

Strategies in index products are best thought of, not as trading strategies, but rather as a means of providing broad guidance as to the general condition of the market and its likely direction over the longer term.  To take the VIX index strategy as an example, you can see that each “trade” spans several weeks.  So one might regard a “buy” signal from the VIX index system as an indication that volatility is expected to rise over the next month or two.  A trader might use that information to lean on the side of being long volatility, perhaps even avoiding any short volatility positions altogether for the next several weeks.  Following the model’s guidance in that way would would certainly have helped many equity and volatility traders during the market sell off during August 2015, for example:

 

Vix Example

The S&P 500 Index model is one I use to provide guidance as to market conditions for the current trading day.  It is a useful input to my thinking as to how aggressive I want my trading models to be during the upcoming session. If the index model suggests a positive tone to the market, with muted volatility, I might be inclined to take a more aggressive stance.  If the model starts trading to the short side, however, I am likely to want to be much more cautious.    Yesterday (May 16, 2016), for example, the index model took an early long trade, providing confirmation of the positive tenor to the market and encouraging me to trade volatility to the short side more aggressively.

 

SP500 Example

 

 

In general, I would tend to classify index trading systems as “decision support” tools that provide a means of shading opinion on the market, or perhaps providing a means of calibrating trading models to the anticipated market conditions. However, they can be used in a more direct way, short of actual trading.  For example, one of our volatility trading systems uses the trading signals from a trading system designed for the VVIX volatility-of-volatility index.  Another approach is to use the signals from an index trading system as an indicator of the market regime in a regime switching model.

Designing Index Trading Models

Whereas it is profitability that is typically the primary design criterion for an actual trading system, given the purpose of an index trading system there are other criteria that are at least as important.

It should be obvious from these few illustrations that you want to design your index model to trade less frequently than the system you are intending to trade live: if you are swing-trading the eminis on daily bars, it doesn’t help to see 50 trades a day from your index system.  What you want is an indication as to whether the market action over the next several days is likely to be positive or negative.  This means that, typically, you will design your index system using bar frequencies at least as long as for your live system.

Another way to slow down the signals coming from your index trading system is to design it for very high accuracy – a win rate of  70%, or higher.  It is actually quite easy to do this:  I have systems that trade the eminis on daily bars that have win rates of over 90%.  The trick is simply that you have to be prepared to wait a long time for the trade to come good.  For a live system that can often be a problem – no-one like to nurse an underwater position for days or weeks on end.  But for an index trading system it matters far less and, in fact, it helps:  because you want trading signals over longer horizons than the time intervals you are using in your live trading system.

Since the index system doesn’t have to trade live, it means of course that the usual trading costs and frictions do not apply.  The advantage here is that you can come up with concepts for trading systems that would be uneconomic in the real world, but which work perfectly well in the frictionless world of index trading.  The downside, however, is that this might lead you to develop index systems that trade far too frequently.  So, even though they should not apply, you might seek to introduce trading costs in order to penalize higher frequency trading systems and benefit systems that trade less frequently.

Designing index trading systems in an area in which genetic programming algorithms excel.  There are two main reasons for this.  Firstly, as I have previously discussed, simple technical indicators of the kind employed by GP modeling systems work well in index markets.  Secondly, and more importantly, you can use the GP system to tailor an index trading system to meet the precise criteria you have in mind, such as the % win rate, trading frequency, etc.

An outstanding product that I can highly recommend in this context is Mike Bryant’s Adaptrade Builder.  Builder is a superb piece of software whose power and ease of use reflects Mike’s engineering background and systems development expertise.


Adaptrade

 

 

Some Further Notes on Market Timing

Almost at the very moment I published a post featuring some interesting research by Glabadanidis (“Market Timing With Moving Averages”  (2015), International Review of Finance, Volume 15, Number 13, Pages 387-425 – see Yes, You Can Time the Market. How it Works, And Why), several readers wrote to point out a recently published paper by Valeriy Zakamulin, (dubbed the “Moving Average Research King” by Alpha Architect, the source for our fetching cover shot) debunking Glabadanidis’s findings in no uncertain terms:

We demonstrate that “too good to be true” reported performance of the moving average strategy is due to simulating the trading with look-ahead bias. We perform the simulations without look-ahead bias and report the true performance of the moving average strategy. We find that at best the performance of the moving average strategy is only marginally better than that of the corresponding buy-and-hold strategy.

So far, no response from Glabadanidis – from which one is tempted to conclude that Zakamulin is correct.

I can’t recall the last time a paper published in a leading academic journal turned out to be so fundamentally flawed.  That’s why papers are supposed to be peer reviewed.   But, I guess, it can happen. Still, it’s rather alarming to think that a respected journal could accept a piece of research as shoddy as Zakamulin claims it to be.

What Glabadanidis had done, according to Zakamulin, was to use the current month closing price to compute the moving average that was used to decide whether to exit the market (or remain invested) at the start of the same month.  An elementary error that introduces look-ahead bias that profoundly impacts the results.

Following this revelation I hastily checked my calculations for the SPY marketing timing  strategy illustrated in my blog post and, to my relief, confirmed that I had avoided the look-ahead trap that Glabadanidis has fallen into.  As the reader can see from the following extract from the Excel spreadsheet I used for the calculations, the decision to assume the returns for the SPY ETF or T-Bills for the current month rests on the value of the 24 month MA computed using prices up to the end of the prior month.  In other words, my own findings are sound, even if Glabadanidis’s are not, as the reader can easily check for himself.


Excel Workbook

 

Nonetheless, despite my relief at having avoided Glabadanidis’s  blunder, the apparent refutation of his findings comes as a disappointment.  And my own research on the SPY market timing strategy, while sound as far as it goes, cannot by itself rehabilitate the concept of market timing using moving averages.  The reason is given in the earlier post.  There is a hidden penalty involved in using the market timing strategy to synthetically replicate an Asian put option, namely the costs incurred in exiting and rebuilding the portfolio as the market declines below the moving average, or later overtakes it.  In a single instance, such as the case of SPY, it might easily transpire simply by random chance that the cost of replication are far lower than the fair value of the put.  But the whole point of Glabadanidis’s research was that the same was true, not only for a single ETF or stock, but for many thousands of them.  Absent that critical finding, the SPY case is no more than an interesting anomaly.

Finally, one reader pointed out that the effect of combining a put option with a stock (or ETF) long position was to create synthetically a call option in the stock (ETF).  He is quite correct.  The key point, however, is that when the stock trades down below its moving average, the value of the long synthetic call position and the market timing portfolio are equivalent.

 

 

The Information Content of the Pre- and Post-Market Trading Sessions

I apologize in advance for this rather “wonkish” post, which is aimed chiefly at the high frequency fraternity, or those at least who trade intra-day, in the equity markets.  Such minutiae are the lot of those engaged in high frequency trading.  I promise that my next post will be of more general interest.

Pre- and Post Market Sessions

The pre-market session in US equities runs from 8:00 AM ET, while the post-market session runs until 8:00 PM ET.  The question arises whether these sessions are worth trading, or at the very least, offer a source of data (quotes, trades) that might be relevant to trading the regular session, which of course runs from 9:30 AM to 4:00 PM ET.  Even if liquidity is thin and trades infrequent, and opportunities in the pre- and post-market very limited, it might be that we can improve our trading models by taking into account such information as these sessions do provide, even if we only ever plan to trade during regular trading hours.

It is somewhat challenging to discuss this in great detail, because HFT equity trading is very much in the core competencies of my firm, Systematic Strategies.  However, I hope to offer some ideas, at least, that some readers may find useful.

SSALGOTRADING AD

 

A Tale of Two Pharmaceutical Stocks

In what follows I am going to make use of two examples from the pharmaceutical industry: Alexion Pharmaceuticals, Inc. (ALXN), which has a market cap of $35Bn and trades around 800,000 shares daily, and Pfizer Inc. (PFE), which has a market cap of over $200Bn and trades close to 50M shares a day.

Let’s start by looking at a system trading ALXN during regular market hours.  The system isn’t high frequency, but trades around 1-2 times a day, on average.  The strategy equity curve from 2015 to April 2016 is not at all impressive.

 

ALXN Regular

ALXN – Regular Session Only

 

But look at the equity curve for the same strategy when we allow it to run on the pre- and post-market sessions, in addition to regular trading hours.  Clearly the change in the trading hours utilized by the strategy has made a huge improvement in the total gain and risk-adjusted returns.

 

ALEXN with pre-market

ALXN – with Pre- and Post-Market Sessions

 

The PFE system trades much more frequently, around 4 times a day, but the story is somewhat similar in terms of how including the pre- and post-market sessions appears to improve its performance.

PFE Regular

PFE – Regular Session Only

PFE with premarket

PFE – with Pre- and Post-Market Sessions

 

Improving Trading Performance

In both cases, clearly, the trading performance of the strategies has improved significantly with the inclusion of the out-of-hours sessions.  In the case of ALXN, we see a modest increase of around 10% in the total number of trades, but in the case of PFE the increase in trading activity is much more marked – around 30%, or more.

The first important question to ask is when these additional trades are occurring.  Assuming that most of them take place during the pre- or post-market, our concern might be whether there is likely to be sufficient liquidity to facilitate trades of the frequency and size we wish to execute.  Of various possible hypotheses, some negative, other positive, we might consider the following:

(a) Bad ticks in the market data feed during out-of-hours sessions give rise to apparently highly profitable “phantom” trades

(b) The market data is valid, but the trades are done in such low volume as to be insignificant for practical purposes (i.e. trades were done for a few hundred lots and additional liquidity is unlikely to be available)

(c) Out-of-hours sessions enable the system to improve profitability by entering or exiting positions in a more timely manner than by trading the regular session alone

(d) Out-of-hours market data improves the accuracy of model forecasts, facilitating a larger number of trades, and/or more profitable trades, during regular market hours

An analysis of the trading activity for the two systems provides important insight as to which of the possible explanations might be correct.


ALXN Analysis

(click to enlarge)

Dealing first with ALXN, we that, indeed, an additional 11% of trades are entered or exited out-of-hours.  However, these additional trades account for somewhere between 17% (on exit) and 20% (on entry) of the total profits.  Furthermore, the size of the average entry trade during the post-market session and of the average exit trade in the pre-market session is more than double that of the average trade entered or exited during regular market hours. That gives concerns that some of the apparent increase in profits may be due to bad ticks at prices away from the market, allowing the system enter or exit trades at unrealistically low or high prices.  Even if many of the trades are good, we will have concerns about the scalability of the strategy in out-of-hours trading, given the relatively poor liquidity in the stock. On the other hand, at least some of the uplift in profits arises from new trades occurring during the regular session. This suggests that, even if we are unable to execute many of the trading opportunities seen during pre- or post-market, the trades from those sessions provides useful additional data points for our model, enabling it to increase the number and/or profitability of trades in the regular session.

Next we turn to PFE.  We can see straight away that, while the proportion of trades occurring during out-of-hours sessions is around 23%, those trades now account for over 50% of the total profits.  Furthermore, the average PL for trades executed on entry post-market, and on exit pre-market, is more than 4x the average for trades entered or exited during normal market hours.  Despite the much better liquidity in PFE compared to ALXN, this is a huge concern – we might expect to see significant discrepancies occurring between theoretical and actual performance of the strategy, due to the very high dependency on out-of-hours trading.

PFE Analysis

(click to enlarge)

As we dig further into the analysis, we do indeed find evidence that bad data ticks play a disproportionate role.  For example, this trade in PFE which apparently occurred at around 16:10 on 4/6 was almost certainly a phantom trade resulting from a bad data point. It turns out that, for whatever reason, such bad ticks are a common occurrence in the stock and account for a large proportion of the apparent profitability of out-of-hours trading in PFE.

 

PFE trade

 

CONCLUSION

We are, of course, only skimming the surface of the analysis that is typically carried out.  One would want to dig more deeply into ways in which the market data feed could be cleaned up and bad data ticks filtered out so as to generate fewer phantom trades.  One would also want to look at liquidity across the various venues where the stocks trade, including dark pools, in order to appraise the scalability of the strategies.

For now, the main message that I am seeking to communicate is that it is often well worthwhile considering trading in the pre- and post-market sessions, not only with a view to generating additional, profitable trading opportunities, but also to gather additional data points that can enhance trading profitability during regular market hours.