This post is available as a PDF download here.
Summary
- Value investing continues to experience a trough of sorrow. In particular, the traditional price-to-book factor has failed to establish new highs since December 2006 and sits in a 25% drawdown.
- While price-to-book has been the academic measure of choice for 25+ years, many practitioners have begun to question its value (pun intended).
- We have also witnessed the turning of the tides against the size premium, with many practitioners no longer considering it to be a valid stand-alone anomaly. This comes 35+ years after being first published.
- With this in mind, we explore the evidence that would be required for us to dismiss other, already established anomalies. Using past returns to establish prior beliefs, we simulate out forward environments and use Bayesian inference to adjust our beliefs over time, recording how long it would take for us to finally dismiss a factor.
- We find that for most factors, we would have to live through several careers to finally witness enough evidence to dismiss them outright.
- Thus, while factors may be established upon a foundation of evidence, their forward use requires a bit of faith.
In Norse mythology, Fimbulvetr (commonly referred to in English as “Fimbulwinter”) is a great and seemingly never-ending winter. It continues for three seasons – long, horribly cold years that stretch on longer than normal – with no intervening summers. It is a time of bitterly cold, sunless days where hope is abandoned and discord reigns.
This winter-to-end-all-winters is eventually punctuated by Ragnarok, a series of events leading up to a great battle that results in the ultimate death of the major gods, destruction of the cosmos, and subsequent rebirth of the world.
Investment mythology is littered with Ragnarok-styled blow-ups and we often assume the failure of a strategy will manifest as sudden catastrophe. In most cases, however, failure may more likely resemble Fimbulwinter: a seemingly never-ending winter in performance with returns blown to-and-fro by the harsh winds of randomness.
Value investors can attest to this. In particular, the disciples of price-to-book have suffered greatly as of late, with “expensive” stocks having outperformed “cheap” stocks for over a decade. The academic interpretation of the factor sits nearly 25% belowits prior high-water mark seen in December 2006.
Expectedly, a large number of articles have been written about the death of the value factor. Some question the factor itself, while others simply argue that price-to-book is a broken implementation.
But are these simply retrospective narratives, driven by a desire to have an explanation for a result that has defied our expectations? Consider: if price-to-book had exhibited positive returns over the last decade, would we be hearing from nearly as large a number of investors explaining why it is no longer a relevant metric?
To be clear, we believe that many of the arguments proposed for why price-to-book is no longer a relevant metric are quite sound. The team at O’Shaughnessy Asset Management, for example, wrote a particularly compelling piece that explores how changes to accounting rules have led book value to become a less relevant metric in recent decades.1
Nevertheless, we think it is worth taking a step back, considering an alternate course of history, and asking ourselves how it would impact our current thinking. Often, we look back on history as if it were the obvious course. “If only we had better prior information,” we say to ourselves, “we would have predicted the path!”2 Rather, we find it more useful to look at the past as just one realized path of many that’s that could have happened, none of which were preordained. Randomness happens.
With this line of thinking, the poor performance of price-to-book can just as easily be explained by a poor roll of the dice as it can be by a fundamental break in applicability. In fact, we see several potential truths based upon performance over the last decade:
- This is all normal course performance variance for the factor.
- The value factor works, but the price-to-book measure itself is broken.
- The price-to-book measure is over-crowded in use, and thus the “troughs of sorrow” will need to be deeper than ever to get weak hands to fold and pass the alpha to those with the fortitude to hold.
- The value factor never existed in the first place; it was an unfortunate false positive that saturated the investing literature and broad narrative.
The problem at hand is two-fold: (1) the statistical evidence supporting most factors is considerable and (2) the decade-to-decade variance in factor performance is substantial. Taken together, you run into a situation where a mere decade of underperformance likely cannot undue the previously established significance. Just as frustrating is the opposite scenario. Consider that these two statements are not mutually exclusive: (1) price-to-book is broken, and (2) price-to-book generates positive excess return over the next decade.
In investing, factor return variance is large enough that the proof is not in the eating of the short-term return pudding.
The small-cap premium is an excellent example of the difficulty in discerning, in real time, the integrity of an established factor. The anomaly has failed to establish a meaningful new high since it was originally published in 1981. Only in the last decade – nearly 30 years later – have the tides of the industry finally seemed to turn against it as an established anomaly and potential source of excess return.
Thirty years.
The remaining broadly accepted factors – e.g. value, momentum, carry, defensive, and trend – have all been demonstrated to generate excess risk-adjusted returns across a variety of economic regimes, geographies, and asset classes, creating a great depth of evidence supporting their existence. What evidence, then, would make us abandon faith from the Church of Factors?
To explore this question, we ran a simple experiment for each factor. Our goal was to estimate how long it would take to determine that a factor was no longer statistically significant.
Our assumption is that the salient features of each factor’s return pattern will remain the same (i.e. autocorrelation, conditional heteroskedasticity, skewness, kurtosis, et cetera), but the forward average annualized return will be zero since the factor no longer “works.”
Towards this end, we ran the following experiment:
- Take the full history for the factor and calculate prior estimates for mean annualized return and standard error of the mean.
- De-mean the time-series.
- Randomly select a 12-month chunk of returns from the time series and use the data to perform a Bayesian update to our mean annualized return.
- Repeat step 3 until the annualized return is no longer statistically non-zero at a 99% confidence threshold.
For each factor, we ran this test 10,000 times, creating a distribution that tells us how many years into the future we would have to wait until we were certain, from a statistical perspective, that the factor is no longer significant.
Sixty-seven years.
Based upon this experience, sixty-seven years is median number of years we will have to wait until we officially declare price-to-book (“HML,” as it is known in the literature) to be dead.3 At the risk of being morbid, we’re far more likely to die before the industry finally sticks a fork in price-to-book.
We perform this experiment for a number of other factors – including size (“SMB” – “small-minus-big”), quality (“QMJ” – “quality-minus-junk”), low-volatility (“BAB” – “betting-against-beta”), and momentum (“UMD” – “up-minus-down”) – and see much the same result. It will take decades before sufficient evidence mounts to dethrone these factors.
HML | SMB4 | QMJ | BAB | UMD | |
Median Years-until-Failure | 67 | 43 | 132 | 284 | 339 |
Now, it is worth pointing out that these figures for a factor like momentum (“UMD”) might be a bit skewed due to the design of the test. If we examine the long-run returns, we see a fairly docile return profile punctuated by sudden and significant drawdowns (often called “momentum crashes”).
Since a large proportion of the cumulative losses are contained in these short but pronounced drawdown periods, demeaning the time-series ultimately means that the majority of 12-month periods actually exhibit positive returns. In other words, by selecting random 12-month samples, we actually expect a high frequency of those samples to have a positive return.
For example, using this process, 49.1%, 47.6%, 46.7%, 48.8% of rolling 12-month periods are positive for HML, SMB, QMJ, and BAB factors respectively. For UMD, that number is 54.7%. Furthermore, if you drop the worst 5% of rolling 12-month periods for UMD, the average positive period is 1.4x larger than the average negative period. Taken together, not only are you more likely to select a positive 12-month period, but those positive periods are, on average, 1.4x larger than the negative periods you will pick, except for the rare (<5%) cases.
The process of the test was selected to incorporate the salient features of each factor. However, in the case of momentum, it may lead to somewhat outlandish results.
Conclusion
While an evidence-based investor should be swayed by the weight of the data, the simple fact is that most factors are so well established that the majority of current practitioners will likely go our entire careers without experiencing evidence substantial enough to dismiss any of the anomalies.
Therefore, in many ways, there is a certain faith required to use them going forward. Yes, these are ideas and concepts derived from the data. Yes, we have done our best to test their robustness out-of-sample across time, geographies, and asset classes. Yet we must also admit that there is a non-zero probability, however small it is, that these are false positives: a fact we may not have sufficient evidence to address until several decades hence.
And so a bit of humility is warranted. Factors will not suddenly stand up and declare themselves broken. And those that are broken will still appear to work from time-to-time.
Indeed, the death of a factor will be more Fimulwinter than Ragnarok: not so violent to be the end of days, but enough to cause pain and frustration among investors.
Addendum
We have received a large number of inbound notes about this commentary, which fall upon two primary lines of questions. We want to address these points.
How were the tests impacted by the Bayesian inference process?
The results of the tests within this commentary are rather astounding. We did seek to address some of the potential flaws of the methodology we employed, but by-in-large we feel the overarching conclusion remains on a solid foundation.
While we only presented the results of the Bayesian inference approach in this commentary, as a check we actually tested two other approaches:
- A Bayesian inference approach assuming that forward returns would be a random walk with constant variance (based upon historical variance) and zero mean.
- Forward returns were simulated using the same bootstrap approach, but the factor was being discovered for the first time and the entire history was being evaluated for its significance.
The two tests were in effort to isolate the effects of the different components of our test.
What we found was that while the reported figures changed, the overall magnitude did not. In other words, the median death-date of HML may not have been 67 years, but the order of magnitude remained much the same: decades.
Stepping back, these results were somewhat a foregone conclusion. We would not expect an effect that has been determined to be statistically significant over a hundred year period to unravel in a few years. Furthermore, we would expect a number of scenarios that continue to bolster the statistical strength just due to randomness alone.
Why are we defending price-to-book?
The point of this commentary was not to defend price-to-book as a measure. Rather, it was to bring up a larger point.
As a community, quantitative investors often leverage statistical significance as a defense for the way we invest.
We think that is a good thing. We should look at the weight of the evidence. We should be data driven. We should try to find ideas that have proven to be robust over decades of time and when applied in different markets or with different asset classes. We should want to find strategies that are robust to small changes in parameterization.
Many quants would argue (including us among them), however, that there also needs to be a why. Why does this factor work? Without the why, we run the risk of glorified data mining. With the why, we can choose for ourselves whether we believe the effect will continue going forward.
Of course, there is nothing that prevents the why from being pure narrative fallacy. Perhaps we have simply weaved a story into a pattern of facts.
With price-to-book, one might argue we have done the exact opposite. The effect, technically, remains statistically significant and yet plenty of ink has been spilled as to why it shouldn’t work in the future.
The question we must answer, then, is, “when does statistically significant apply and when does it not?” How can we use it as a justification in one place and completely ignore it in others?
Furthermore, if we are going to rely on hundreds of years of data to establish significance, how can we determine when something is “broken” if the statistical evidence does not support it?
Price-to-book may very well be broken. But that is not the point of this commentary. The point is simply that the same tools we use to establish and defend factors may prevent us from tearing them down.
Tactical Portable Beta
By Corey Hoffstein
On May 6, 2019
In Carry, Portfolio Construction, Risk & Style Premia, Term, Trend, Value, Weekly Commentary
This post is available as a PDF download here.
Summary
In November 2017, I was invited to participate in a Bloomberg roundtable discussion with Barry Ritholtz, Dave Nadig, and Ben Fulton about the future of ETFs. I was quoted as saying,
Shortly after the publication of the discussion, we penned a research commentary titled Portable Beta which extolled the potential virtues of employing prudent leverage to better exploit diversification opportunities. For investors seeking to enhance returns, increasing beta exposure may be a more reliable approach than the pursuit of alpha.
In August 2018, WisdomTree introduced the 90/60 U.S. Balanced Fund (ticker: NTSX), which blends core equity exposure with a U.S. Treasury futures ladder to create the equivalent of a 1.5x levered 60/40 portfolio. On March 27, 2019, NTSX was awarded ETF.com’s Most Innovative New ETF of 2018.
The idea of portable beta was not even remotely uniquely ours. Two anonymous Twitter users – “Jake” (@EconomPic) and “Unrelated Nonsense” (@Nonrelatedsense) – had discussed the idea several times prior to my round-table in 2017. They argued that such a product could be useful to free up space in a portfolio for alpha-generating ideas. For example, an investor could hold 66.6% of their wealth in a 90/60 portfolio and use the other 33.3% of their portfolio for alpha ideas. While the leverage is technically applied to the 60/40, the net effect would be a 60/40 portfolio with a set of alpha ideas overlaid on the portfolio. Portable beta becomes portable alpha.
Even then, the idea was not new. After NTSX launched, Cliff Asness, co-founder and principal of AQR Capital Management, commented on Twitter that even though he had a “22-year head start,” WisdomTree had beat him to launching a fund. In the tweet, he linked to an article he wrote in 1996, titled Why Not 100% Equities, wherein Cliff demonstrated that from 1926 to 1993 a 60/40 portfolio levered to the same volatility as equities achieved an excess return of 0.8% annualized above U.S. equities. Interestingly, the appropriate amount of leverage utilized to match equities was 155%, almost perfectly matching the 90/60 concept.
Source: Asness, Cliff. Why Not 100% Equities. Journal of Portfolio Management, Winter 1996, Volume 22 Number 2.
Following up on Cliff’s Tweet, Jeremy Schwartz from WisdomTree extended the research out-of-sample, covering the quarter century that followed Cliff’s initial publishing date. Over the subsequent 25 years, Jeremy found that a levered 60/40 outperformed U.S. equities by 2.6% annualized.
NTSX is not the first product to try to exploit the idea of diversification and leverage. These ideas have been the backbone of managed futures and risk parity strategies for decades. The entire PIMCO’s StocksPLUS suite – which traces its history back to 1986 – is built on these foundations. The core strategy combines an actively managed portfolio of fixed income with 100% notional exposure in S&P 500 futures to create a 2x levered 50/50 portfolio.
The concept traces its roots back to the earliest eras of modern financial theory. Finding the maximum Sharpe ratio portfolio and gearing it to the appropriate risk level has always been considered to be the theoretically optimal solution for investors.
Nevertheless, after 2008, the words “leverage” and “derivatives” have largely been terms non gratisin the realm of investment products. But that may be to the detriment of investors.
90/60 Through the Decades
While we are proponents of the foundational concepts of the 90/60 portfolio, frequent readers of our commentary will not be surprised to learn that we believe there may be opportunities to enhance the idea through tactical asset allocation. After all, while a 90/60 may have out-performed over the long run, the short-run opportunities available to investors can deviate significantly. The prudent allocation at the top of the dot-com bubble may have looked quite different than that at the bottom of the 2008 crisis.
To broadly demonstrate this idea, we can examine the how the realized efficient frontier of stock/bond mixes has changed shape over time. In the table below, we calculate the Sharpe ratio for different stock/bond mixes realized in each decade from the 1920s through present.
Source: Global Financial Data. Calculations by Newfound Research. Returns are hypothetical and backtested. Returns are gross of all fees, transaction costs, and taxes. Returns assume the reinvestment of all distributions. Bonds are the GFD Indices USA 10-Year Government Bond Total Return Index and Stocks are the S&P 500 Total Return Index (with GFD Extension). Sharpe ratios are calculated with returns excess of the GFD Indices USA Total Return T-Bill Index. You cannot invest in an index. 2010s reflect a partial decade through 4/2019.
We should note here that the original research proposed by Asness (1996) assumed a bond allocation to an Ibbotson corporate bond series while we employ a constant maturity 10-year U.S. Treasury index. While this leads to lower total returns in our bond series, we do not believe it meaningfully changes the conclusions of our analysis.
We can see that while the 60/40 portfolio has a higher realized Sharpe ratio than the 100% equity portfolio in eight of ten decades, it has a lower Sharpe ratio in two consecutive decades from 1950 – 1960. And the 1970s were not a ringing endorsement.
In theory, a higher Sharpe ratio for a 60/40 portfolio would imply that an appropriately levered version would lead to higher realized returns than equities at the same risk level. Knowing the appropriate leverage level, however, is non-trivial, requiring an estimate of equity volatility. Furthermore, leverage requires margin collateral and the application of borrowing rates, which can create a drag on returns.
Even if we conveniently ignore these points and assume a constant 90/60, we can still see that such an approach can go through lengthy periods of relative under-performance compared to buy-and-hold equity. Below we plot the annualized rolling 3-year returns of a 90/60 portfolio (assuming U.S. T-Bill rates for leverage costs) minus 100% equity returns. We can clearly see that the 1950s through the 1980s were largely a period where applying such an approach would have been frustrating.
Source: Global Financial Data. Calculations by Newfound Research. Returns are hypothetical and backtested. Returns are gross of all fees, transaction costs, and taxes. Bonds are the GFD Indices USA 10-Year Government Bond Total Return Index and Stocks are the S&P 500 Total Return Index (with GFD Extension). The 90/60 portfolio invests 150% each month in the 60/40 portfolio and -50% in the GFD Indices USA Total Return T-Bill Index. You cannot invest in an index.
Poor performance of the 90/60 portfolio in this era is due to two effects.
First, 10-year U.S. Treasury rates rose from approximately 4% to north of 15%. While a constant maturity index would constantly roll into higher interest bonds, it would have to do so by selling old holdings at a loss. Constantly harvesting price losses created a headwind for the index.
This is compounded in the 90/60 by the fact that the yield curve over this period spent significant time in an inverted state, meaning that the cost of leverage exceeded the yield earned on 40% of the portfolio, leading to negative carry. This is illustrated in the chart below, with –T-Bills– realizing a higher total return over the period than –Bonds–.
Source: Global Financial Data. Calculations by Newfound Research. Returns are hypothetical and backtested. Returns are gross of all fees, transaction costs, and taxes. Returns assume the reinvestment of all distributions. T-Bills are the GFD Indices USA Total Return T-Bill Index, Bonds are the GFD Indices USA 10-Year Government Bond Total Return Index, and Stocks are the S&P 500 Total Return Index (with GFD Extension). You cannot invest in an index.
This is all arguably further complicated by the fact that while a 1.5x levered 60/40 may closely approximate the risk level of a 100% equity portfolio over the long run, it may be a far cry from it over the short-run. This may be particularly true during periods where stocks and bonds exhibit positive realized correlations as they did during the 1960s through 1980s. This can occur when markets are more pre-occupied with inflation risk than economic risk. As inflationary fears abated and economic risk become the foremost concern in the 1990s, correlations between stocks and bonds flipped.
Thus, during the 1960s-1980s, a 90/60 portfolio exhibited realized volatility levels in excess of an all-equity portfolio, while in the 2000s it has been below.
This all invites the question: should our levered allocation necessarily be static?
Getting Tactical with a 90/60
We might consider two approaches to creating a tactical 90/60.
The first is to abandon the 90/60 model outright for a more theoretically sound approach. Specifically, we could attempt to estimate the maximum Sharpe ratio portfolio, and then apply the appropriate leverage such that we either hit a (1) constant target volatility or (2) the volatility of equities. This would require us to not only accurately estimate the expected excess returns of stocks and bonds, but also their volatilities and correlations. Furthermore, when the Sharpe optimal portfolio is highly conservative, notional exposure far exceeding 200% may be necessary to hit target volatility levels.
In the second approach, equity and bond exposure would each be adjusted tactically, without regard for the other exposure. While less theoretically sound, one might interpret this approach as saying, “we generally want exposure to the equity and bond risk premia over the long run, and we like the 60/40 framework, but there might be certain scenarios whereby we believe the expected return does not justify the risk.” The downside to this approach is that it may sacrifice potential diversification benefits between stocks and bonds.
Given the original concept of portable beta is to increase exposure to the risk premia we’re already exposed to, we prefer the second approach. We believe it more accurately reflects the notion of trying to provide long-term exposure to return-generating risk premia while trying to avoid the significant and prolonged drawdowns that can be realized with buy-and-hold approaches.
Equity Signals
To manage exposure to the equity risk premium, our preferred method is the application of trend following signals in an approach we call trend equity. We will approximate this class of strategies with our Newfound Research U.S. Trend Equity Index.
To determine whether our signals are able to achieve their goal of “protect and participate” with the underlying risk premia, we will plot their regime-conditional betas. To do this, we construct a simple linear model:
We define a bear regime as the worst 16% of monthly returns, a bull regime as the best 16% of monthly returns, and a normal regime as the remaining 68% of months. Note that the bottom and top 16thpercentiles are selected to reflect one standard deviation.
Below we plot the strategy conditional betas relative to U.S. equity
We can see that trend equity has a normal regime beta to U.S. equities of approximately 0.75 and a bear market beta of 0.5, in-line with expectations that such a strategy might capture 70-80% of the upside of U.S. equities in a bull market and 40-50% of the downside in a prolonged bear market. Trend equity beta of U.S. equities in a bull regime is close to the bear market beta, which is consistent with the idea that trend equity as a style has historically sacrificed the best returns to avoid the worst.
Bond Signals
To govern exposure to the bond risk premium, we prefer an approach based upon a combination of quantitative, factor-based signals. We’ve written about many of these signals over the last two years; specifically in Duration Timing with Style Premia (June 2017), Timing Bonds with Value, Momentum, and Carry (January 2018), and A Carry-Trend-Hedge Approach to Duration Timing (October 2018). In these three articles we explore various mixes of value, momentum, carry, flight-to-safety, and bond risk premium measures as potential signals for timing duration exposure.
We will not belabor this commentary unnecessarily by repeating past research. Suffice it to say that we believe there is sufficient evidence that value (deviation in real yield), momentum (prior returns), and carry (term spread) can be utilized as effective timing signals and in this commentary are used to construct bond indices where allocations are varied between 0-100%. Curious readers can pursue further details of how we construct these signals in the commentaries above.
As before, we can determine conditional regime betas for strategies based upon our signals.
We can see that our value, momentum, and carry signals all exhibit an asymmetric beta profile with respect to 10-year U.S. Treasury returns. Carry and momentum exhibit an increase in bull market betas while value exhibits a decrease in bear market beta.
Combining Equity and Bond Signals into a Tactical 90/60
Given these signals, we will construct a tactical 90/60 portfolio as being comprised of 90% trend equity, 20% bond value, 20% bond momentum, and 20% bond carry. When notional exposure exceeds 100%, leverage cost is assumed to be U.S. T-Bills. Taken together, the portfolio has a large breadth of potential configurations, ranging from 100% T-Bills to a 1.5x levered 60/40 portfolio.
But what is the appropriate benchmark for such a model?
In the past, we have argued that the appropriate benchmark for trend equity is a 50% stock / 50% cash benchmark, as it not only reflects the strategic allocation to equities empirically seen in return decompositions, but it also allows both positive and negative trend calls to contribute to active returns.
Similarly, we would argue that the appropriate benchmark for our tactical 90/60 model is not a 90/60 itself – which reflects the upper limit of potential capital allocation – but rather a 45% stock / 30% bond / 25% cash mix. Though, for good measure we might also consider a bit of hand-waving and just use a 60/40 as a generic benchmark as well.
Below we plot the annualized returns versus maximum drawdown for different passive and active portfolio combinations from 1974 to present (reflecting the full period of time when strategy data is available for all tactical signals). We can see that not only does the tactical 90/60 model (with both trend equity and tactical bonds) offer a return in line with U.S. equities over the period, it does so with significantly less drawdown (approximately half). Furthermore, the tactical 90/60 exceeded trend equity and 60/40 annualized returns by 102 and 161 basis points respectively.
These improvements to the return and risk were achieved with the same amount of capital commitment as in the other allocations. That’s the beauty of portable beta.
Source: Federal Reserve of St. Louis, Kenneth French Data Library, and Newfound Research. Calculations by Newfound Research. Returns are hypothetical and backtested. Returns are gross of all fees, transaction costs, and taxes. Returns assume the reinvestment of all distributions. You cannot invest in an index.
Of course, full-period metrics can deceive what an investor’s experience may actually be like. Below we plot rolling 3-year annualized returns of U.S. equities, the 60/40 mix, trend equity, and the tactical 90/60.
Source: Federal Reserve of St. Louis, Kenneth French Data Library, and Newfound Research. Calculations by Newfound Research. Returns are hypothetical and backtested. Returns are gross of all fees, transaction costs, and taxes. Returns assume the reinvestment of all distributions. You cannot invest in an index.
The tactical 90/60 model out-performed a 60/40 in 68% of rolling 3-year periods and the trend equity model in 71% of rolling 3-year periods. The tactical 90/60, however, only out-performs U.S. equities in 35% of rolling 3-year periods, with the vast majority of relative out-performance emerging during significant equity drawdown periods.
For investors already allocated to trend equity strategies, portable beta – or portable tactical beta – may represent an alternative source of potential return enhancement. Rather than seeking opportunities for alpha, portable beta allows for an overlay of more traditional risk premia, which may be more reliable from an empirical and academic standpoint.
The potential for increased returns is illustrated below in the rolling 3-year annualized return difference between the tactical 90/60 model and the Newfound U.S. Trend Equity Index.
Source: Federal Reserve of St. Louis, Kenneth French Data Library, and Newfound Research. Calculations by Newfound Research. Returns are hypothetical and backtested. Returns are gross of all fees, transaction costs, and taxes. Returns assume the reinvestment of all distributions. You cannot invest in an index.
From Theory to Implementation
In practice, it may be easier to acquire leverage through the use of futures contracts. For example, applying portable bond beta on-top of an existing trend equity strategy may be achieved through the use of 10-year U.S. Treasury futures.
Below we plot the growth of $1 in the Newfound U.S. Trend Equity Index and a tactical 90/60 model implemented with Treasury futures. Annualized return increases from 7.7% to 8.9% and annualized volatility declines from 9.7% to 8.5%. Finally, maximum drawdown decreases from 18.1% to 14.3%.
We believe the increased return reflects the potential return enhancement benefits from introducing further exposure to traditional risk premia, while the reduction in risk reflects the benefit achieved through greater portfolio diversification.
Source: Quandl and Newfound Research. Calculations by Newfound Research. Returns are hypothetical and backtested. Returns are gross of all fees, transaction costs, and taxes. Returns assume the reinvestment of all distributions. You cannot invest in an index.
It should be noted, however, that a levered constant maturity 10-year U.S. Treasury index and 10-year U.S. Treasury futures are not the same. The futures contracts are specified such that eligible securities for delivery include Treasury notes with a remaining term to maturity of between 6.5 and 10 years. This means that the investor short the futures contract has the option of which Treasury note to deliver across a wide spectrum of securities with potentially varying characteristics.
In theory, this investor will always choose to deliver the bond that is cheapest. Thus, Treasury futures prices will reflect price changes of this so-calledcheapest-to-deliver bond, which often does not reflect an actual on-the-run 10-year Treasury note.
Treasury futures therefore utilize a “conversion factor” invoicing system referenced to the 6% futures contract standard. Pricing also reflects a basis adjustment that reflects the coupon income a cash bond holder would receive minus financing costs (i.e. the cost of carry) as well as the value of optionality provided to the futures seller.
Below we plot monthly returns of 10-year U.S. Treasury futures versus the excess returns of a constant maturity 10-year U.S. Treasury index. We can see that the futures had a beta of approximately 0.76 over the nearly 20-year period, which closely aligns with the conversion factor over the period.
Source: Quandl and the Federal Reserve of St. Louis. Calculations by Newfound Research.
Despite these differences, futures can represent a highly liquid and cost-effective means of implementing a portable beta strategy. It should be further noted that having a lower “beta” over the last two decades has not necessarily implied a lower return as the basis adjustment can have a considerable impact. We demonstrate this in the graph below by plotting the returns of continuously-rolled 10-year U.S. Treasury futures (rolled on open interest) and the excess return of a constant maturity 10-year U.S. Treasury index.
Source: Quandl and Newfound Research. Calculations by Newfound Research. Returns are hypothetical and backtested. Returns are gross of all fees, transaction costs, and taxes. Returns assume the reinvestment of all distributions. You cannot invest in an index.
Conclusion
In a low return environment, portable beta may be a necessary tool for investors to generate the returns they need to hit their financial goals and reduce their risk of failing slow.
Historically, a 90/60 portfolio has outperformed equities with a similar level of risk. However, the short-term dynamics between stocks and bonds can make the volatility of a 90/60 portfolio significantly higher than a simple buy-and-hold equity portfolio. Rising interest rates and inverted yield curves can further confound the potential benefits versus an all-equity portfolio.
Since constant leverage is not a guarantee and we do not know how the future will play out, moving beyond standard portable beta implementations to tactical solutions may augment the potential for risk management and lead to a smoother ride over the short-term.
Getting over the fear of using leverage and derivatives may be an uphill battle for investors, but when used appropriately, these tools can make portfolios work harder. Risks that are known and compensated with premiums can be prudent to take for those willing to venture out and bear them.
If you are interested in learning how Newfound applies the concepts of tactical portable beta to its mandates, please reach out (info@thinknewfound.com).