In this book, we demonstrate the performance of various strategies, which can require only a single input: historical prices. In this section, we will begin our journey to the world of price-based investing with a short description of how we both calculated and tested these strategies on real historical data. All the strategies have been implemented in a consistent and identical way so as to assure their comparability. Below, we describe three major aspects of our examinations: (1) the data we use, (2) the method we form the portfolios, and (3) the method we evaluate their performance.

What Data We Use?

Today’s financial markets know almost no borders. Sitting in his living room in Berlin an investor can access equity markets in London, New York, or even Tokyo with a single mouse-click. The world of investing has become more interconnected and accessible than ever before. As a result, we do not test our strategies in a single market, even if it’s as large as the American market, but instead, we test them in a robust sample of 24 developed countries with extensive and well-established stock markets—that is, Australia, Austria, Belgium, Canada, Denmark, Finland, France, Germany, Greece, Hong Kong, Ireland, Israel, Italy, Japan, the Netherlands, New Zealand, Norway, Portugal, Singapore, Spain, Sweden, Switzerland, the UK, and the USA. These markets span across many continents and cultures and account for the majority of capitalization in global equity markets. We have based our computations on the price data sourced from FactSet. Naturally, our tests could be further extended to include the emerging or frontier markets, but our focus on the developed economies guarantees the strategies to be accessible to most of the developed-market investors.

As we have focused on the period from January 1995 to June 2017, our sample is fresh and timely, reflecting the recent changes and developments in financial markets. We also used older data, for instance, when forming a strategy for January 1995 requires data from the earlier periods as, for example, a momentum strategy which relies on past performance. At times, the return data for some or all of the countries is available for the shorter periods, in which case we use them. We calculate all of the strategies separately for individual countries.

We collected the initial data in local currencies as comparisons based on various currencies could be misleading (Liew and Vassalou 2000; Bali et al. 2013). This is especially reasonable for countries where inflation and risk-free rates are very high and differ significantly across the markets. As most studies adopt the dollar-denominated approach (Waszczuk 2014a), we also denominated all the data in US dollars to obtain comparable results on an international scale.Footnote 1 For consistency, whenever we needed to use the risk-free rate (e.g., to calculate excess returns), we used the benchmark returns on the US three-month Treasury bills. Throughout the book, we have used gross returns, that is, returns unadjusted for tax (whether income taxes or taxes on dividends), and rely on monthly returns, which is probably most prevalent among such studies, although most of the accounting data would change only quarterly.Footnote 2

Finally, being aware that not all stocks in equity market are tradable, for example, stocks of companies with extremely low liquidity and market capitalization would be very difficult to trade freely, we applied a series of various static and dynamic filters to the common stocks within our calculations at the beginning of each month when forming the investment portfolio. We took account of only companies with the total stock market capitalization exceeding $100 million and the average daily trailing six-month turnover beyond $100,000. As a very low price may also lead to practical difficulties with trading, due to a wide bid-ask spread, we discarded stocks with the trading price below $1.00 at the beginning of a given month.Footnote 3

Portfolios Structure

As in our study we have reviewed a lot of different strategies, to make them easily comparable, we investigated the strategies using portfolios designed in an identical fashion. To test various investment approaches, we applied the so-called one-way sorted portfolios by ranking all the stocks in our universe on a characteristic which in academia is called the “return-predictive variable” for it helps forecast future price changes. Naturally, for our purposes, we used price-based return-predictive variables . Having thus sorted the securities, we formed a long portfolio of stocks ranked with the highest predicted return and a short portfolio of securities with the lowest predicted returns.

In order to calculate returns in a given month, typically called month t, we sorted the stocks within the sample at the end of the previous month (month t−1) according to the investigated characteristic, for example, short-run return and long-run return. Having ranked the markets by the investigated characteristics, we then determined the 20th and 80th percentile breakpoints for each measure. In other words, by focusing only on the 20% of the securities with the highest expected returns and the 20% of the stocks with the lowest predicted future returns, we consequently arrived at two quintile subgroups.Footnote 4

Subsequently, we weighted the respective equities from portfolios. For simplicity, we used a straightforward weighting method—equal weighting, under which each of the best (or worst) stocks from the top (or bottom) quintiles of the ranking was assigned the same weight, that is, a fraction of the portfolio. In other words, we divided the portfolio into equal parts and bought the same amount of every stock. In practice, many methods are used, and all of them has some pros and cons.

Equal Weighting

Among various methods, this is perhaps the simplest way of weighting portfolio components, giving identical weights to all securities. Importantly, we are likely to rebalance such portfolio frequently as stock prices rise and fall every month, changing thus the share in the portfolio. To hold equal stocks, the investor needs to rebalance it on a systematic basis. The more frequent the rebalancing , the more frequent the trading. Whereas the more trades we do, the higher rise the total transaction costs. As a result, a frequently rebalanced equal-weighted portfolio might finally prove costly for investors. In contrast, for portfolios constructed from one-way sorts, the cost drag may not significantly exceed other types of weightings, for example, the value weighting as the portfolio turnover comes not only from rebalancing but mostly from stocks entering and leaving the portfolio, which is common across all weighting schemes. To its advantage, this approach generates no overweight of any type of stocks making equally weighted portfolios exhibit decent exposure to small companies, which tend to yield high anomaly returns.

Capitalization Weighting

Weighting on stock market capitalization, as an alternative to equal-weighting scheme, assigns bigger weights to stock market companies with large market values. As this approach concentrates in particular on large and liquid companies, it may result in lower trading costs (Novy-Marx and Velikov 2016; Zaremba and Nikorowski 2017), although the differences are moderate (Zaremba and Andreu Sánchez 2017), because a large part of the turnover stems from stocks entering and leaving the portfolio rather than from the rebalancing . To its disadvantage, capitalization weighting returns tend to appear the strongest in small caps and this type of portfolio formation underweights small caps diminishing the portfolio benefits from cross-sectional patterns.

Liquidity Weighting

Liquidity weighting is a good candidate for an even more realistic approach to weighting portfolio constituents as it grants a higher share in the portfolio to the most liquid securities ranked by, for example, turnover; its unquestionable advantage is the low-trading cost: the investor concentrates on stocks that are highly liquid, which as a rule also display narrow bid-ask spreads. Unfortunately, such portfolios give also preference to the most efficient market segments, making the stocks less likely to display strong anomalous behavior.

Factor Weighting

Following the factor-weighting approach, we weight the stocks neither according to their capitalization or liquidity but rather by their expected return proxied by an additional variable. For instance, when building a portfolio on the book-to-market ratio, you can weigh the components by the standardized book-to-market ratio; strictly speaking , the weights could be tied to either the raw variables (see, e.g., Zaremba and Umutlu 2018) or the ranking values (Asness et al. 2017).

This approach guarantees the portfolio share be closely linked to the expected performance. Unfortunately, the weights might also prove quite volatile, especially in the case of dynamic strategies, like momentum, leading to a high turnover and, in consequence, high trading costs.

Enhanced Indexing and Other Methods

There are numerous other techniques of weighting the components of quantitatively managed portfolios. Some rely on sophisticated optimization algorithms while others are rule based (Narang 2013). One of the increasingly popular methods includes fundamental weighting based on weighting portfolio components on fundamental variables: for example, sales or the book-to-market ratio . This approach delivers decent returns at the level of both individual stocks and whole countries or indices.Footnote 5

Evaluation of the Strategies

To present the performance of various strategies, we have facilitated an array of statistical data: mean returns, volatilities , or skewness , using the following both simple and popular ratios to assess the returns and strategy risk.

Sharpe Ratio

The Sharpe ratio originates from William Sharpe, a Nobel Prize laureate, who in his research entitled “Mutual Fund Performance” (Sharpe 1966) formulated the index, which was later named after him. Undoubtedly, the ratio is still the most popular investment performance measurement tool, which accounts for not only profit but also risk.

Under the most traditional definition, the Sharpe ratio measures the excess rate of return per unit of risk taken by the investor (Sharpe 1966). The ratio is calculated by dividing the excess return and the risk understood as the volatility (standard deviation) of these excess returns.Footnote 6 By excess return, we mean the difference between the return on the investigated portfolio and the return of the risk-free instrument.Footnote 7 Throughout this book, it is represented by benchmark returns on the US three-month Treasury bills.

The Sharpe ratio is a simple measure and could be expressed with the following formula:

$$ SR\kern0.56em =\kern0.56em \frac{\overline{R}}{\sigma } $$
(1.1)

whereby \( \overline{R} \) represents the mean excess return on the investigated portfolio over the examined period, and σ is its standard deviation of excess returns. The ratio is usually presented on an annual basis, that is, with yearly excess returns.Footnote 8 Although our computations are based on monthly intervals, we also adopted an annualized version of the ratio by simply multiplying the monthly Sharpe ratio by the square root of 12.

While an unquestionable virtue of the Sharpe ratio is its simplicity, it performs poorly in the environment of negative excess returns. For this reason, we facilitated the Sharpe ratio with the so-called Jensen’s alpha.

Jensen’s Alpha

The Jensen’s alpha is a measure derived from the capital asset pricing model (CAPM, Sharpe 1964).Footnote 9 The CAPM is a simple model that was invented by the famous researcher—William Sharpe—for three main purposes: to explain the reasons for portfolio diversification , to create a framework for valuating assets in a risky environment, and to explain differences in the long-term returns of various assets.Footnote 10 The CAPM laid the foundation for many other methods of performance evaluation in investment portfolio management.

The fundamental assumption of the model states that volatility of a financial instrument can be broken down into two parts: a systematic and specific risk. The systematic risk stems from general changes in the market conditions and relates to the volatility of the market portfolio, whereas the specific risk relates to volatility which is, however, driven not by the market but by the internal situation in the company. In other words, losses ensuing a market crash are rather of a systematic nature while losses due to an employee strike belong to the specific risk category.

The CAPM model bears some vital implications for both portfolio construction and diversification . When building a portfolio, systematic risks of individual stock simply add up; however, specific risks, not being correlated, set each other off. Therefore, in a well-diversified portfolio, the influence of the specific risk is generally negligible, and in a well-functioning market, a rational investor may ignore the specific risk and concentrate solely on the systematic part. After all, would the investor even consider the specific risk if it could be easily diversified away at no cost?

This important implication of the CAPM model—stating that the investors should be only compensated for the systematic risk because the specific risk can be easily eliminated—is

$$ {R}_{i,t}={\alpha}_i+{R}_{f,t}+{\beta}_{rm,i}\cdot \kern0.28em \left({R}_{mt}-{R}_{f,t}\right)+{\varepsilon}_{i,t}, $$
(1.2)

where R i,t , R m,t , and R f,t are returns on the analyzed security or portfolio; i, the market portfolio and risk-free returns at time t; and α i and β rm,i are regression parameters. β rm,i is the measure of the systematic risk which tells us how aggressively the stock reacts to the price changes in the broad market. Fundamentally, the CAPM formula implies that the excess returns on the investigated security or portfolio should increase linearly with the systematic risk measured with beta : the higher the risk, the higher the expected return.

Finally, the α i intercept measures the average abnormal return: the so-called Jensen’s alpha. It is defined as the rate of return earned by the portfolio or a strategy in excess of the expected return from the CAPM model. The Eq. 1.3 could be easily rewritten to be used to evaluate past returns on a portfolio:

$$ {\alpha}_i=\overline{R_i^E}-{\beta}_i\cdot \left(\overline{R_m^E}\right), $$
(1.3)

where α i is the Jensen’s alpha on the investigated portfolio, \( \overline{R_i^E} \) is its mean excess return over the examined period, β i is the market beta , and \( \overline{R_m^E} \) is the mean excess return on the market portfolio.Footnote 11 Throughout the book, we have used the capitalization-weighted return as the proxy for the market portfolio, which we calculated based on either gross or the risk-free rate , consequently represented by the US three-month T-bills.Footnote 12 Importantly, as far as a zero-investment portfolio is concerned, there is no need to subtract any risk-free rate.

The decisive rule for the Jensen’s alpha states that when alpha from the CAPM model turns negative, it signals the investment in the analyzed strategy, or portfolio, to become unreasonable as a higher return at a comparable risk level could be achieved via investments in the risk-free asset and market portfolio.

Statistical Significance

One important challenge in examining investment strategies is to distinguish when seemingly abnormal returns are truly abnormal and when it is pure coincidence. If a trader earned 10% annually for five consecutive years, how can we tell whether he has followed a superior investment strategy or he just got lucky? For this purpose, whenever we reported any mean returns or alphas, we simultaneously reported their statistical significance which at least to some extent helps us statistically differentiate real return patterns from mere luck. When some mean return, or alpha, exceeds zero at the 5% level, it indicates a 5% risk of no real pattern in the returns, even though we have identified it in the historical data. In other words, the returns could turn positive only in our specific sample, and this result may not be replicated in another sample. Thus, this 5% threshold could also be interpreted as the probability of the returns plunging below zero when implementing this strategy to another sample.

The statistical significance test may be one sided, that is, informing us whether the returns are significantly higher than zero, or two sided, that is, informing us whether the returns depart from zero (either below or above).

Throughout this book, we presented the significance of both the mean and abnormal returns of the tested strategiesFootnote 13 aiming to provide a better view on how compelling the performance of the strategies really is. If the abnormal returns remain significant at the level of 1% or 5%, we can be fairly sure that the strategy is no random return pattern. At 10%, the evidence is still firm, but less convincing. Once the significance plunges below 10%, the probability that the abnormal returns result from pure chance is considerable, thus it would be risky to assume it would continue in the future.Footnote 14