Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

Entropy-Based Financial Asset Pricing

Entropy-Based Financial Asset Pricing We investigate entropy as a financial risk measure. Entropy explains the equity premium of securities and portfolios in a simpler way and, at the same time, with higher explanatory power than the beta parameter of the capital asset pricing model. For asset pricing we define the continuous entropy as an alternative measure of risk. Our results show that entropy decreases in the function of the number of securities involved in a portfolio in a similar way to the standard deviation, and that efficient portfolios are situated on a hyperbola in the expected return – entropy system. For empirical investigation we use daily returns of 150 randomly selected securities for a period of 27 years. Our regression results show that entropy has a higher explanatory power for the expected return than the capital OPEN ACCESS asset pricing model beta. Furthermore we show the time varying behavior of the Citation: Ormos M, Zibriczky D (2014) Entropy- Based Financial Asset Pricing. PLoS ONE 9(12): beta along with entropy. e115742. doi:10.1371/journal.pone.0115742 Editor: Giampiero Favato, Kingston University London, United Kingdom Received: August 8, 2014 Accepted: November 26, 2014 Introduction Published: December 29, 2014 We build an equilibrium capital asset pricing model by applying a novel risk Copyright:  2014 Ormos, Zibriczky. This is an measure, the entropy. Entropy characterizes the uncertainty or measures the open-access article distributed under the terms of the Creative Commons Attribution License, which dispersion of a random variable. In our particular case, it characterizes the permits unrestricted use, distribution, and repro- uncertainty of stock and portfolio returns. In modern Markowitz [1] portfolio duction in any medium, provided the original author and source are credited. theory and equilibrium asset pricing models [2] we apply linear regressions. This Data Availability: The authors confirm that all data methodology supposes that the returns are stationary and normally distributed; underlying the findings are fully available without however, this is not actually the case [3]. Entropy, on the other hand, does not restriction. Data are from the Center for Research in Security Prices (http://www.crsp.com/). Access to have this kind of boundary condition. The main goal of this paper is to apply CRSP data requires a subscription. For subscrip- entropy as a novel risk measure. As a starting point even the density function itself tion in formation please contact subscrip- tions@crsp.chicagobooth.edu. has to be estimated. In the traditional asset pricing model there is equilibrium between expected return the beta parameter, which is the covariance–variance Funding: The authors have no support or funding to report. ratio between the market portfolio and the investigated investment opportunity. If Competing Interests: The authors have declared the random variable is normally distributed then the entropy follows its standard that no competing interests exist. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 1/21 Entropy-Based Financial Asset Pricing deviation; thus in the ideal case there is no difference between the two risk measures. However; our results show that there is a significant difference between the standard deviation, or beta, and the entropy of a given security or portfolio. In this paper we show that entropy offers an ideal alternative for capturing the risk of an investment opportunity. If we explain the return of a wide sample of securities and portfolios with different risk measures then on an ordinary least squares (OLS) regression setting the explanatory power is much higher in the case of the entropy measure of risk than in the case of the traditional measures, both in- sample and out-of-sample. We show that entropy reduction in line with diversification behaves similarly to standard deviation; however at the same time it captures a beta-like systematic risk of single securities or non-efficient portfolios as well. For well-diversified portfolios the explanatory power of entropy is 1.5 times higher than that of the capital asset pricing model (CAPM) beta. We also test and compare entropy with standard risk measures for market circumstances that are increasing and decreasing, and find that the explanatory power of entropy is significantly higher in a bullish market, but lower for a bearish market. Our results for bullish and bearish regimes show that the different risk measures behave similarly in terms of the positive and negative relationship between risk and return. This behavior underlines the fact that the entropy-based risk measure can give contradictory results in the same way as traditional risk estimations in upward and downward regimes. We also compare the entropy-based risk measures with the CAPM beta in and out of sample, which gives information on the predictive power of the different methods. As the CAPM beta measures the systematic risk only, while entropy based risk measures and the standard deviation captures the total risk of the investment our results are shocking, that entropy gives almost twice as high an average explanatory power as the beta, with an average of 40% less standard deviation. A further contribution of the paper is that we introduce a simple method to estimate the entropy of a security or portfolio return. Data In our empirical analysis we apply daily returns from the Center for Research in Security Prices (CRSP) database for the period from 1985 to the end of 2011. We randomly select 150 securities from the S&P500 index components that are available for the full period. The market return is the CRSP value-weighted index return premium above the risk-free rate. The index tracks the return of the New York Stock Exchange (NYSE), the American Stock Exchange (AMEX) and NASDAQ stocks. The risk-free rate is the return of the one-month Treasury bill from the CRSP. We use daily returns because they are not normally distributed (see S1 Table). Erdo ˝s and Ormos (2009) [3] and Erdo ˝s et al. (2011) [4] describe the main difficulties of modeling asset prices with non-normal returns. The daily return calculation enables us to compare different risk measures. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 2/21 Entropy-Based Financial Asset Pricing Methodology Entropy is a mathematically-defined quantity that is generally used for characterizing the probability of outcomes in a system that is undergoing a process. It was originally introduced in thermodynamics by Rudolf Clausius [5]to measure the ratio of transferred heat through a reversible process in an isolated system. In statistical mechanics the interpretation of entropy is the measure of uncertainty about the system that remains after observing its macroscopic properties (pressure, temperature or volume). The application of entropy in this perspective was introduced by Ludwig Boltzmann [6]. He defined the configuration entropy as the diversity of specific ways in which the components of the system may be arranged. He found a strong relationship between the thermodynamic and the statistical aspects of entropy: the formulae for thermodynamic entropy and configuration entropy only differ in the so-called Boltzmann constant. There is an important application of entropy in information theory as well, and this is often called Shannon [7] entropy. The information provider system operates as a stochastic cybernetic system, in which the message can be considered as a random variable. The entropy quantifies the expected value of the information in a message or, in other words, the amount of information that is missing before the message is received. The more unpredictable (uncertain) the message that is provided by the system, the greater the expected value of the information contained in the message. Consequently, greater uncertainty in the messages of the system means higher entropy. Because the entropy equals the amount of expected information in a message, it measures the maximum compression ratio that can be applied without losing information. In financial applications, Philippatos and Wilson [8] find that entropy is more general and has some advantages over standard deviation; in their paper they compare the behaviors of standard deviation and entropy in portfolio manage- ment. Kirchner and Zunckel [9] argue that in financial economics entropy is a better tool for capturing the reduction of risk by diversification; however, in their study they suppose that the assets are Gaussian. Dionisio et al. [10] argue that entropy observes the effect of diversification and is a more general measure of uncertainty than variance, since it uses more information about the probability distribution. The mutual information and the conditional entropy perform well when compared with the systematic risk and the specific risk estimated through the linear equilibrium model. Regarding the predictability of stock market returns, Maasoumi and Racine [11] point out that entropy has several desirable properties and is capable of efficiently capturing nonlinear dependencies in return time series. Nawrocki and Harding [12] propose applying state-value weighted entropy as a measure of investment risk; however, they are dealing with the discrete case. All the above academic papers recognize that entropy could be a good measure of risk; however, it seems to be difficult to use this measure. Our main motivation is to show that an entropy-based risk measure is, on the one hand, more precise, PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 3/21 Entropy-Based Financial Asset Pricing and, on the other hand, no more complicated to use than variance equilibrium models. Discrete entropy function Entropy functions can be divided into two main types, discrete and differential entropy functions. Let X be a discrete random variable. The possible outcomes of this variable are denoted by o ,o ,::,o , and the corresponding probabilities by p 5Pr(X 5o ), p $0 1 2 n i i i and p~1. The generalized discrete entropy function [13] for the variable X is i~1 defined as: H ðÞ X ~ log p , ð1Þ 1{a i~1 where a is the order of entropy, a$0 and a?1, and the base of the logarithm is 2. The order of entropy expresses the weight taken into account in each outcome; if the order of entropy is lower, the more likely outcomes are underweighted, and vice versa. The most widely used orders are a51 and a52. a51 is a special case of generalized entropy. However the substitution of a51 into (1) results in a division by zero. It can be shown, using l’Hopital’s rule for the limit of a51, that H converges to the Shannon entropy: H ðÞ X ~{ p logðÞ p ð2Þ 1 i i i~1 The case of a52 is called collision entropy and similarly to the literature we refer to this special case as ‘‘Re ´ nyi entropy’’ further in the paper: H ðÞ X ~{log p ð3Þ i~1 H (X) is a non-increasing function in a, and both entropy measures are greater than zero provided that there is a finite number of possible outcomes: 0vH ðÞ X ƒH ðÞ X ð4Þ 2 1 Differential entropy function Let X be a continuous random variable taking values from R with a probability density function f(x). Analogously to (1), the continuous entropy is defined as: PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 4/21 Entropy-Based Financial Asset Pricing H ðÞ X ~ ln fxðÞ dx ð5Þ 1{a One can see that the bases of the logarithms in (1) and (5) are different. Although the entropy depends on the base, it can be shown that the value of the entropy changes only by a constant coefficient for different bases. We use the natural logarithm for all differential entropy functions. The formulas for the special cases (a51 and a52) are the following: H ðÞ X ~{ fxðÞln fxðÞdx ð6Þ H ðÞ X ~{ln fxðÞ dx ð7Þ An important difference between discrete and continuous entropy is that while discrete entropy takes only non-negative values, continuous entropy can also take negative values: H ðÞ X [R ð8Þ In practice, standard risk measures like the CAPM beta or standard deviation are calculated on daily or monthly return data. We also follow this practice, and use a formula that is able to capture risk using this kind of data. Since the return on securities can take values from a continuous codomain, we primarily focus on the differential entropy function. However, by grouping return values into bins the discrete entropy function may also be used; this solution is outside the scope of this paper. Entropy estimation For the estimation of differential entropy, the probability density function of the return values needs to be estimated. Let x ,x ,:::,x be the observations of the 1 2 n continuous random variable X, and H (X) the sample-based estimation of a,n H (X). The plug-in estimations of entropy are calculated on the basis of the density function estimation. The probability density function f(x) is estimated by f (x), the integral estimate of entropy, in the following way: H ðÞ X ~ ln f ðÞ x dx, ð9Þ a,n n 1{a where A is the range of integration, which may exclude small and tail values of f (x). We propose to select A 5(min(x), max(x)). n n PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 5/21 Entropy-Based Financial Asset Pricing Histogram One of the simplest methods of density estimation is the histogram-based density estimation. Let b 5(max(x), min(x)) be the range of sample values; partition the range into k bins of equal width and denote the cutting points by t . The width of a bin is constant: h~ ~t {t . The density function is estimated by using the jz1 j following formula: f ðÞ x ~ , ð10Þ nh th if x(t , t ), where n is the number of data points falling in the j bin. j j+1 j Based on the properties of the histogram, a simpler non plug-in estimation formula can be deduced for Shannon and Renyi entropy using (6), (7), (9) and (10): 1 n H ðÞ X ~ v ln ð11Þ 1,n j n nh j~1 X 2 H ðÞ X ~{ln h ð12Þ 2,n nh j~1 The parameter of this method is the number of equal width bins (k). However, there are several methods for choosing this parameter (e.g. the square root choice, Scott’s normal reference rule [14], or the Freedman-Diaconis rule [15]); the detailed descriptions of these are outside the scope of this paper. Kernel density estimation The kernel-based density estimation is another commonly used method. It applies the following formula: 1 x{x f ðÞ x ~ K , ð13Þ nh h i~1 where KðÞ is the kernel function, and h is the bandwidth parameter. There are several kernel functions that can be used (see S2 Table); for practical reasons (computational time), we propose using the indicator-based Epanechnikov kernel function: KzðÞ~ 1{z I , ð14Þ fg jj z ƒ1 where I is the indicator function. Ha ¨rdle [16] shows that the choice of the kernel function is only of secondary importance, so the focus is rather on the right choice of bandwidth (h). One of the most widely used simple formulas for the estimation of h is Silverman’s rule of thumb [17]: PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 6/21 Entropy-Based Financial Asset Pricing sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi () 1 IQRðÞ x 1 h ~1:06 min ðÞ x{x , n , ð15Þ rot i n{1 1:34 i~1 where IQR(x) is the interquartile range of x. As the formula assumes a normal distribution for X it gives an approximation for optimal bandwidth; despite this, Silverman’s rule of thumb can be used for a good initial value for more sophisticated optimization methods [18]. Sample spacing estimation Let x #x #…#x be the corresponding order of x , x ,…,x , assuming that n,1 n,2 n,n 1 2 n this is a sample of i.i.d. real-valued random variables. x 2x is called a n,i+m n,i spacing of order m (1#i,i+m,n). The simple sample spacing density estimate is the following [19]: m 1 f ðÞ x ~ , ð16Þ n x {x n,im n,(i{1)m if x[x , x ). n,(i-1)m n,im Wachowiak et al. [20] introduced another variation of the sample spacing density estimation, called the Correa estimator: izm=2 x{x ðÞ j{i j i j~i{m=2 f ðÞ x ~ , ð17Þ izm=2 x{x j i j~i{m=2 izm=2 if i:x [[x , x ); x~ x , and 1#j#n. n,i n,i+1 i j mz1 j~i{m=2 The parameter for sample spacing methods is the fixed order m. For practical reasons (e.g. different sizes of samples) we suggest using m , which depends on the size of the sample and is calculated by the following formula: m ~q r, ð18Þ where k is the number of bins, and the braces indicate the ceiling function. Beirlant et al. [19] overview several additional entropy estimation methods, such as resubstitution, splitting-data and cross-validation; however, our paper focuses on the applications that are used most often. Risk estimation Let the following be a given set of data: D :fg S,R,R ,R ð19Þ M F PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 7/21 Entropy-Based Financial Asset Pricing The elements are the set of securities S:{S , S ,…,S }, with the corresponding 1 2 l observations being R:{R , R ,…,R }, where R 5(r , r ,…,r ). The observation for 1 2 l i i1 i2 in the market return is R 5(r , r ,…,r ), and the observation for the risk free M M1 M2 Mn return is R 5(r , r ,…,r ) where l is the number of securities and n is the F F1 F2 Fn number of samples. Let us recall that the main goal of this paper is to apply entropy as a novel risk measure. In order to handle the risk measure uniformly, we introduce k as a unified property for securities. Let k(S ) be the risk estimate for the security i. In the economic literature the most widely used risk measures are the standard deviation and the CAPM beta. Let us denote these by k and k , respectively. The s b estimation of these risk measures for the security i is the following: k ^ ðÞ S ~sðÞ R{R ð20Þ s i i F and covðÞ R{R ,R {R i F M F k ðÞ S ~b~ , ð21Þ b i sðÞ R {R M F where b is the CAPM beta, covðÞ is the covariance of the arguments and s is the standard deviation. Our hypothesis is that uncertainty about the observation values can be interpreted as a risk of the security, and for this reason we apply entropy as a risk measure. Because the differential entropy function can also take negative values, for better interpretability we apply the exponential function to the entropy, and we define the entropy-based risk measure by the following formula: H R {R nðÞ i F k ðÞ S ~e ð22Þ H i One can see that k takes values from the non-negative real numbers, k [[0,+‘). Explanatory and predictive power In order to compare the efficiency of the risk estimation methods, we introduce two basic evaluation approaches, the measurement of in-sample explanatory power and the measurement of out-of-sample predictive power. In-sample Let V be a target variable, with sample v~ðÞ v ,v ,:::,v , and let U be a single 1 2 l explanatory variable with sample u~ðÞ u ,u ,:::,u . To estimate the explanatory 1 2 l power of the variable U for the variable V, we use the following method. The linear relationship between the two variables can be described using the linear regression model: V~a za Uz". 0 1 The parameters of the model (a and a ) are estimated by ordinary least squares 0 1 (OLS), and the estimation for the target value is the following: ^v~^a z^a u where i 0 1 i ^a and ^a are the estimations of a and a , respectively. One of the most 0 1 0 1 PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 8/21 Entropy-Based Financial Asset Pricing commonly applied estimations of the explanatory power is the R (goodness of fit, or coefficient of determination) of the linear regression: ðÞ v{ðÞ ^a z^a u i 0 1 i i~1 RðÞ v,u ~1{ ð23Þ ðÞ v{v i~1 We are curious as to how efficiently the different risk measures describe the expected return of a security, and we denote this measure by g(k). Let the explanatory variable U be the risk measure of the securities, where the sample is: u ~ðÞ k ^ðÞ S ,k ^ðÞ S ,:::,k ^ðÞ S , ð24Þ k 1 2 l and the target variable T is the expected risk premium of the securities, where the sample is: v ~ðÞ ER½ {R ,ER½ {R ,:::,ER½ {R , ð25Þ m 1 F 2 F l F where k is the unified risk measure function, and E½ is the expected value of the argument. We define the estimation of the in-sample explanatory power (efficiency) as the R of the previously defined variables (24) and (25): ^gkðÞ~R v ,u ð26Þ m k Out of sample Let us create a split of samples for a given D:{S, R, R , R } data set (19): M F I I I I I O O O O O D : S ,R ,R ,R , D : S ,R ,R ,R , ð27Þ M F M F I I I I where the corresponding samples for the securities are R : R ,R ,:::,R , 1 2 l I O O O O O R ~ðÞ r ,r ,:::,r and R : R ,R ,:::,R , R ~ r ,r ,:::,r , the i1 i2 im imðÞ z1 imðÞ z2 imðÞ zp i 1 2 i split for market returns is R ~ðÞ r ,r ,:::,r and M1 M2 Mm R ~ r ,r ,:::,r , and the split for the risk free rates is MmðÞ z1 MmðÞ z2 MmðÞ zp I O I O R ~ r ,r ,:::,r and R ~ r ,r ,:::,r , where jj S ~jj S , F1 F2 FmðÞ zp FmðÞ z1 FmðÞ z2 FmðÞ zp F F I O R ~m, R ~p,(1#i#l), and m+p5n. i i The explanatory values contain the risk estimates for the set of securities based on the data set D : I I I I u ~ k ^ S ,k ^ S ,:::,k ^ S , ð28Þ k 1 2 l and the target values are the expected risk premium of the securities based on D : O O O O O O O v ~ ER {R ,ER {R ,:::,ER {R ð29Þ m 1 F 2 F l F Based on (26), (28) and (29), the estimation of the out-of-sample explanatory (predictive) power is the following: PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 9/21 Entropy-Based Financial Asset Pricing 2 O I g ðÞ k ~R v ,u ð30Þ m k Both in- and out of sample we test whether the difference between the explanatory power of the investigated risk measures (standard deviation, CAPM beta, Shannon- and Renyi entropy) are significant by applying bootstrapping method. In our bootstrap iteration we remove 25 random stocks from the investigated 150 ones and measure the R s of the four different models. We apply 1000 iterations to approximate the distribution of R values on random selection, and we test the equality of means of R s by applying t-test on the generated samples. Results and Discussion We present the empirical results in four parts. First, we show how the entropy behaves in the function of securities involved into the portfolio. Second, we present the long-term explanatory power of the investigated models. Third we examine and compare the performance of different risk measures in in upward and downward market trends. Fourth we apply the different risk parameters to predict future returns, thus we test the out of sample explanatory power of the well-known risk parameters and compare their efficiency to the entropy based risk measures. Characterizing the diversification effect We investigate whether entropy is able to measure the reduction of risk by diversification. We generate 10 million random equally-weighted portfolios with different numbers of securities involved (at most 100,000 for each size), based on the 150 randomly selected securities from the S&P500. The risk of portfolios is estimated by standard deviation, and by the Shannon and Re ´ nyi entropies using risk premiums for the full period. Because the CAPM beta measures the systematic risk only, we exclude it from the investigation of risk reduction. Both types of entropy functions are calculated by the histogram-based density function estimation, with 175 bins for the Shannon entropy and 50 bins for the Renyi entropy. (We tested the histogram, sample spacing and kernel density estimation methods, and the histogram-based method proved to be the most efficient in terms of explanatory and predictive power and simplicity. See our results in S3 Table.) Fig. 1 shows the diversification effects that are characterized by the entropic risk measures and by the standard deviation. For 10 random securities involved in the portfolio, approximately 40% of risk reduction can be achieved compared to a single random security, based on all of the three risk estimators under investigation. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 10 / 21 Entropy-Based Financial Asset Pricing Fig. 1. Average value of risk and risk reduction vs. number of securities in portfolio. Note: We generate 10 million random equally weighted portfolios with different number of securities involved (at most 100,000 for each size) based on 150 randomly selected securities from S&P500. The risk of portfolios is estimated by standard deviation (gray continuous curve), Shannon- (black continuous curve) and Re´nyi entropy (black dashed curve) in the period from 1985 to the end of 2011. Both types of entropy functions are calculated by histogram based density function estimation. The left chart shows the average risk estimates for each portfolio size, and the right chart shows the risk reduction compared to an average risk of single security portfolio. doi:10.1371/journal.pone.0115742.g001 Fig. 1 suggests that entropy shows behavior that is similar to but not the same as standard deviation, so it can serve as a good measure of risk. We also investigate how the different portfolios behave in the expected return – risk coordinate system in the function of diversification. We generate 200-200 random equally- weighted portfolios with 2, 5 and 10 securities involved, and compare these to single securities using standard deviation, the CAPM beta, the Shannon entropy and the Re ´ nyi entropy as risk measures; the results are presented in Fig. 2. Fig. 2 shows the performance of random portfolios by diversification using different risk estimation methods. One can see that the characteristics of standard deviation and entropy are quite similar, with the portfolios being situated on a hyperbola as in the portfolio theory of Markowitz [1]. Different characteristics can be observed by using the CAPM beta; the more securities that are involved in a portfolio, the closer they are situated in the center of the coordinate system. Long term explanatory power In order to evaluate how efficiently the risk measures explain the expected risk premium over a long period, we estimate the risk for each security using standard deviation, the CAPM beta, and the Shannon and Re ´ nyi entropies based on the full period (denoted by P1). The single explanatory variable is the risk measure; the target variable is the expected risk premium of the security. We apply the explanatory power estimation by calculating ^gkðÞ (R ) for each risk measure. Fig. 3 shows the efficiency of explaining the expected risk premium by the different risk measures; the expected daily risk premium is presented as a function of risk measure. The CAPM beta performs the worst, with 6.17% efficiency. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 11 / 21 Entropy-Based Financial Asset Pricing Fig. 2. Portfolios with different number of securities involved in E(r) – risk system. Note: The panels show the expected risk premium of the portfolios (calculated by the average of daily risk premiums) versus the estimated risk using different methods; the number of securities involved is indicated by the different markers. We generate a sample of 750 random portfolios by using 150 randomly selected securities and 200-200 random equally weighted portfolios with 2, 5 and 10 securities. The risk of portfolios is estimated by standard deviation, CAPM beta, Shannon- and Re´nyi entropy by using daily returns in the period from 1985 to the end of 2011. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. doi:10.1371/journal.pone.0115742.g002 However, the explanatory power of standard deviation (7.83%) is higher than that of the CAPM beta, and both entropies perform significantly better, with efficiency of 12.98% for the Shannon entropy and 15.71% for the Renyi entropy. Based on the equation of linear regressions, the average unexplained risk premium (intersect on the Y-axis or Jensen alpha [21]) for the entropy methods (0.0091, 0.0059) is lower than that for the standard methods (0.0170 for standard deviation and 0.0209 for the CAPM beta). PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 12 / 21 Entropy-Based Financial Asset Pricing Fig. 3. Explanatory power of risk measures in long term. Note: The four panels show the relationship between risk premium and risk (standard deviation, CAPM beta, Shannon- and Re´nyi entropy) of 150 randomly selected securities by using different estimation methods. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. The equation and the explanatory power (R ) of the linear regressions are presented using expected risk premium as target variable and risk as explanatory variable. Under the OLS regression equations in brackets the p-values can be seen for each parameter estimations. The R s of the models applying entropy based risk measures are significantly different form standard deviation and CAPM beta at 1% level. doi:10.1371/journal.pone.0115742.g003 We also measure the explanatory power for different numbers of securities involved in the portfolio, by generating at most 100,000 samples for each; we present these results in Fig. 4. Fig. 4 illustrates how the explanatory power changes with diversification. One can see that the explanatory power of standard deviation and entropy decreases with an increase in the number of securities involved in the portfolio, while the performance of the CAPM beta is nearly constant. While the CAPM beta models the systematic risk only, the standard deviation and entropy are capable of measuring specific risk, which gives additional explanatory power for less- diversified portfolios. Despite the decreased explanatory power of both entropy PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 13 / 21 Entropy-Based Financial Asset Pricing Fig. 4. Explanatory power of risk measures in long term by diversification. Note: This figure shows the explanatory power (R ) of portfolios with different number of securities involved and different risk estimate methods. We generate 10 million random equally weighted portfolios with different number of securities involved (at most 100,000 for each size) using daily risk premiums of 150 randomly selected securities. The risk of portfolios is estimated by standard deviation (light gray curve), CAPM beta (black dotted curve), Shannon- (gray) and Re´nyi entropy (black). Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. doi:10.1371/journal.pone.0115742.g004 functions, they perform better than the CAPM beta in all the cases that were investigated. For well-diversified portfolios the explanatory power of the Renyi entropy is 1.5 times higher than that of the CAPM beta. Explanatory power by primary market trends We split the original 27-year sample by primary market trend into a ‘‘bullish’’ and a ‘‘bearish’’ sample (denoted by P1+ and P1-), containing returns for upward and downward periods, respectively (for the labels of the periods see S4 Table). For these two sample sets we investigate the explanatory power for standard deviation, the CAPM beta, and the Shannon and Re ´ nyi entropies using the same parameter for the histogram-based entropy estimation as for the previous experiments. Fig. 5 and Fig. 6 show the results in the expected risk premium – risk coordinate system. Our results for the bullish and bearish regimes show that the different risk measures behave similarly in terms of the positive and negative relationships between risk and return. This behavior underlines the fact that an entropy-based risk measure can give contradictory results in a similar way to traditional risk estimations in different regimes. In bullish market circumstances we find a very high explanatory power for all kinds of risk measures: 33.90%, 36.67%, 43.45% and 42.36% with standard deviation, the CAPM beta, the Shannon entropy and the Renyi entropy, respectively. As for the full sample tests, the slopes of the regression lines are positive, meaning that higher risk-taking promises higher returns. In contrast to the bullish market, during downward trends higher risk- taking does not result in higher returns and, indeed, the higher the risk the higher the negative premium achieved by the investor. We have to mention that the explanatory power of the CAPM beta is higher than that of the entropy-based risk measures. Our entropy results are in line with those for the CAPM beta; and the regime dependency is clear as well. On the other hand, the explanatory power is PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 14 / 21 Entropy-Based Financial Asset Pricing Fig. 5. Explanatory power of risk measures in bullish sample. Note: The panels show the relationship between the expected risk premium of securities and risk by using different estimation methods. We present the equation of linear regression and the goodness of fit (R ).We estimated the risk of 150 random securities in upward trend periods (bull market) from 1985 to the end of 2011 using standard deviation, CAPM beta, Shannon- and Re´nyi entropy risk estimation methods. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. Under the OLS regression equations in brackets the p-values can be seen for each parameter estimations. The R s of the models applying entropy based risk measures are significantly higher than the models with standard deviation and CAPM beta at 1% level. doi:10.1371/journal.pone.0115742.g005 again much higher for this regime than for the full sample. Altogether, we argue that the test results for the full sample give a better comparison opportunity, as the sample sizes of the bullish and bearish markets are different and at the present moment the investor cannot decide whether there is an upward or a downward trend. Short term explanatory and predictive power Although attractive results are achieved within the sample, this does not necessarily mean high efficiency outside the sample. Therefore we took several ten-year periods, shifting the starting year by one year for each, with the first PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 15 / 21 Entropy-Based Financial Asset Pricing Fig. 6. Explanatory power of risk measures in bearish sample. Note: The panels show the relationship between the expected risk premium of securities and risk by using different estimation methods. We present the equation of linear regression and the goodness of fit (R ).We estimated the risk of 150 random securities in downward trend periods (bear market) from 1985 to the end of 2011 using standard deviation, CAPM beta, Shannon- and Re´nyi entropy risk estimation methods. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. Under the OLS regression equations in brackets the p-values can be seen for each parameter estimations. The R of the models applying entropy based risk measures are significantly higher than the model with standard deviation at 1% level. doi:10.1371/journal.pone.0115742.g006 period being 1985 to 1994 and the last 2002 to 2011. As the full data set covers 27 complete years, we used 18 ten-year periods. We split each ten-year period into two shorter five-year periods (P2i and P2o), with the risk measures being estimated based on the first period and the predictive efficiency being measured in the second period. In the previous sections, we have presented the results for in- sample for the full sample and for the different regimes, and here we summarize these and we also compare the long-term in-sample results with the short-term in- sample and out-of–sample results. Table 1 summarizes the explanatory power of the investigated risk measures for the different samples. ^g ,^g z , and ^g show the results of the long-term analysis P1 P1 P1 PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 16 / 21 Entropy-Based Financial Asset Pricing PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 17 / 21 Table 1. Efficiency of explaining and predicting risk premium in different samples. Risk measure ^g ^g ^g ^g ^g s ðÞ ^g sðÞ ^g z { P1 P1 P1 P2i P2o R P2i R P2o Standard deviation 7.83% 33.9% 36.7% 7.94% 9.7% 0.75 0.65 Beta 6.17% 36.7% 43.7% 13.31% 6.45% 0.98 1.02 Shannon entropy 12.98% 43.5% 39.6% 13.38% 10.15% 0.69 0.64 Re´nyi entropy 15.71% 42.4% 38.6% 12.82% 9.34% 0.63 0.62 Note: The table summarizes the explanatory power (in sample R ) of the investigated risk measures in different samples. We estimate risk measures of 150 random securities using standard deviation, CAPM beta, Shannon- and Re´nyi entropy risk estimation methods for (1) long term, from 1985 to the end of 2011 (1985–2011); (2) long term on upward trends (bull market), (3) long term on downward trends (bear market), (4) 18 10-year periods shifting by one year from period (1985–1994) to period (2002–2011), split into two 5-5 year periods for each. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. The ^g shows P1 the explanatory power of risk measures for long term, the ^g z, and ^g { summarizes the explanatory power on upward and downward trends, respectively, ^g stands for the average P1 P1 P2i explanatory power of risk measured in the first 5 years of 10-year shorter periods in sample, and ^g shows the average predicting power of risk measures (out P2o of sample R ) calculated by estimating risk in the first 5 years and evaluating them on the other 5 years in each 10-year periods. The last two columns show the relative standard deviation of the explanatory and predicting power based on the 18 shorter periods for the investigated risk measures. doi:10.1371/journal.pone.0115742.t001 Entropy-Based Financial Asset Pricing Table 2. Explanatory power in short period samples. T-test T-test P2i H H H ,s T-test H ,b H ,s T-test H ,b sig H (s/b) sig H (s/b) sb 1 2 1 1 2 2 1 2 1985–1989 3.2% 9.4% 5.1% 3.7% 19.3 232.8 6.7 246.0 ***/ ***/ 1986–1990 1.7% 3.6% 2.6% 3.6% 11.8 212.8 22.7 0.4 ***/ ***/ 1987–1991 4.1% 4.8% 6.0% 7.9% 13.9 7.6 27.8 21.8 ***/*** ***/*** 1988–1992 5.5% 5.0% 6.7% 6.8% 7.7 8.5 8.1 8.8 ***/*** ***/*** 1989–1993 3.5% 4.2% 8.9% 7.2% 40.6 32.9 28.9 21.4 ***/*** ***/*** 1990–1994 9.6% 7.1% 23.4% 20.1% 66.1 75.8 48.2 57.7 ***/*** ***/*** 1991–1995 16.0% 13.6% 28.1% 21.9% 67.7 56.3 32.9 30.8 ***/*** ***/*** 1992–1996 16.3% 17.8% 24.4% 20.5% 41.8 29.1 24.4 12.4 ***/*** ***/*** 1993–1997 7.5% 24.9% 15.4% 13.6% 38.1 240.0 32.7 255.2 ***/ ***/ 1994–1998 7.0% 30.1% 15.8% 12.6% 86.1 2101.1 55.6 2124.3 ***/ ***/ 1995–1999 16.6% 51.2% 28.3% 27.3% 88.3 2166.4 77.8 2167.8 ***/ ***/ 1996–2000 8.6% 28.2% 18.0% 20.8% 63.7 267.7 85.0 249.7 ***/ ***/ 1997–2001 2.1% 15.3% 7.2% 9.7% 39.3 260.0 58.2 241.7 ***/ ***/ 1998–2002 0.2% 2.5% 1.5% 2.4% 28.2 216.0 41.8 20.3 ***/ ***/ 1999–2003 6.1% 7.5% 8.1% 9.8% 16.9 6.2 29.2 18.8 ***/*** ***/*** 2000–2004 1.9% 0.1% 1.5% 1.4% 28.0 35.5 29.2 34.8 /*** /*** 2001–2005 15.1% 5.6% 17.5% 18.4% 16.0 93.5 22.4 102.2 ***/*** ***/*** 2002–2006 17.9% 8.9% 22.3% 23.1% 24.6 89.6 30.2 98.8 ***/*** ***/*** Average 7.94% 13.31% 13.37% 12.82% Rel. dev 0.75 0.98 0.69 0.63 Note: This table summarizes the explanatory power of the different risk measures for expected risk premium in the first 5 years of 18 10-year periods (P2i) shifting by one year from period (1985–1994) to period (2002–2011). We estimate and evaluate risk measures of 150 randomly selected securities from the S&P500 index using standard deviation (s), CAPM beta (b), Shannon entropy (H ) and Re´nyi entropy (H ) risk estimation methods by daily risk premiums. 1 2 Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. We apply t-statistics by bootstrapping method to measure whether differences in R s are significant. We use *s to designate that the entropy based risk measure is significantly higher than the standard deviation and CAPM beta; ***, ** and * stands for 1%, 5% and 10% significance level respectively. doi:10.1371/journal.pone.0115742.t002 for the full period and during the upward and downward trends, respectively; ^g ,and ^g stand for the average efficiency measured for short-term in-sample P2i P2o and out-of-sample, respectively; and sðÞ ^g , and sðÞ ^g measure the relative R R P2i P2o standard deviation of the efficiency when applying the in-sample and out-of- sample test for short periods (For the detailed results for all periods see Table 2 and Table 3). While the standard deviation risk measure performs almost the same in the long and the short run (7.83% vs. 7.94), its predictive efficiency is surprisingly good (9.70%). The explanatory power of the CAPM beta in the long period is low (6.17%), while the average efficiency in the short periods is more than twice as high (13.31%). We use arithmetic averages [22]. Comparing the results for in-sample and the out-of-sample, the predictive power of the beta is relatively low (6.45%), which suggests that the model may be over-fitted for the training sample. The Shannon entropy performs better than the standard deviation and the CAPM beta in each sample. The Renyi entropy shows the highest explanatory power in the long run; however, in short periods the Renyi PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 18 / 21 Entropy-Based Financial Asset Pricing Table 3. Predicting power in short periods out of sample. T-test T-test T-test T-test sig H P2i P2o H H H ,s H ,b H ,s H ,b sig H (s/b) (s/b) sb 1 2 1 1 2 2 1 1985–1989 1990–1994 7.3% 2.8% 13.0% 10.0% 35.2 80.4 15.4 52.0 ***/*** ***/*** 1986–1990 1991–1995 17.0% 4.1% 19.3% 18.1% 12.8 101.4 5.6 85.8 ***/*** ***/*** 1987–1991 1992–1996 21.5% 5.9% 22.6% 17.5% 7.1 99.8 223.0 71.8 ***/*** /*** 1988–1992 1993–1997 9.8% 7.9% 14.6% 13.2% 34.9 43.9 27.8 38.1 ***/*** ***/*** 1989–1993 1994–1998 7.9% 16.5% 13.5% 11.6% 65.8 221.3 45.5 235.8 ***/ ***/ 1990–1994 1995–1999 10.0% 23.9% 16.6% 14.9% 63.2 257.1 48.0 270.1 ***/ ***/ 1991–1995 1996–2000 9.0% 14.1% 9.1% 9.0% 0.1 244.0 20.7 245.0 / / 1992–1996 1997–2001 11.3% 14.7% 11.7% 11.8% 1.7 220.8 2.0 221.0 */ **/ 1993–1997 1998–2002 14.2% 4.7% 12.7% 10.8% 211.0 66.8 226.3 54.4 /*** /*** 1994–1998 1999–2003 24.7% 2.7% 17.5% 19.8% 254.6 154.7 236.3 173.0 /*** /*** 1995–1999 2000–2004 3.6% 6.8% 0.3% 0.5% 259.0 290.3 255.0 287.1 / / 1996–2000 2001–2005 8.0 0.0% 3.6% 3.0% 247.1 67.9 254.5 62.7 /*** /*** 1997–2001 2002–2006 10.3% 0.4% 6.1% 4.5% 238.9 91.9 256.1 78.8 /*** /*** 1998–2002 2003–2007 7.8% 3.2% 6.4% 5.8% 213.4 40.8 219.7 35.0 /*** /*** 1999–2003 2004–2008 1.5% 3.2% 1.9% 2.1% 5.5 218.5 7.7 216.3 ***/ ***/ 2000–2004 2005–2009 4.7% 1.2% 5.0% 5.1% 4.8 62.9 5.9 63.1 ***/*** ***/*** 2001–2005 2006–2010 2.2% 1.9% 3.2% 4.0% 17.8 20.9 28.9 31.5 ***/*** ***/*** 2002–2006 2007–2011 4.0% 2.3% 5.4% 6.5% 21.5 50.1 35.2 62.2 ***/*** ***/*** Average 9.70% 6.45% 10.14% 9.34% Relative deviation 0.65 1.02 0.64 0.62 Note: This table summarizes the predicting power of the investigated risk measures for expected risk premium in the last 5 years of 18 10-year periods shifting by one year from period (1985–1994) to period (2002–2011). We estimate risk measures of 150 randomly selected securities from the S&P500 index using standard deviation (s), CAPM beta (b), Shannon entropy (H ) and Re´nyi entropy (H ) risk estimation methods by daily risk premiums in the first 5 years 1 2 (P2i) and measure the predicting power on the next 5 years (P2o) by estimating the goodness of fit of linear regression (R ). Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. We apply t-statistics by bootstrapping method to measure whether differences in R s are significant. We use *s to designate that the entropy based risk measure is significantly higher than the standard deviation and CAPM beta; ***, ** and * stands for 1%, 5% and 10% significance level respectively. doi:10.1371/journal.pone.0115742.t003 entropy performs worse than the Shannon entropy. Comparing the reliability of the risk estimators, the standard deviation of the in-sample and out-of-sample results is the lowest for the entropy risk measures, and the highest for the CAPM beta. Summarizing our results, we state that the beta can beat the entropy only in the case of bearish market circumstances. In any other situation, entropy seems to be a better and more reliable risk measure. Conclusions Entropy as a novel risk measure combines the advantages of the CAPM’s risk parameter (beta) and the standard deviation. It captures risk without using any information about the market, and it is capable of measuring the risk reduction effect of diversification. The explanatory power for the expected return within the sample is better than the beta, especially in the long run covering bullish and PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 19 / 21 Entropy-Based Financial Asset Pricing bearish periods; the predictive power for the expected return is higher than for standard deviation. Both the Shannon and the Renyi entropies give more reliable risk estimation; their explanatory power exhibits significantly lower variance compared to the beta or the standard deviation. If upward and downward trends are distinguished, the regime dependency of entropy can be recognized: this result is similar to that for the beta. Among the entropy estimation methods reviewed, the histogram-based method proved to be the most efficient in terms of explanatory and predictive power; we propose a simple estimation formula for the Shannon and the Renyi entropy functions, which facilitates the application of an entropy-based risk measure. Supporting Information S1 Table. Descriptive statistics. doi:10.1371/journal.pone.0115742.s001 (DOCX) S2 Table. The most often used Kernel functions. doi:10.1371/journal.pone.0115742.s002 (DOCX) S3 Table. Explanatory power of Shannon entropy by different density estimation methods. doi:10.1371/journal.pone.0115742.s003 (DOCX) S4 Table. Labeling periods by market trend. doi:10.1371/journal.pone.0115742.s004 (DOCX) Author Contributions Conceived and designed the experiments: MO DZ. Performed the experiments: MO DZ. Analyzed the data: MO DZ. Contributed reagents/materials/analysis tools: MO DZ. Wrote the paper: MO DZ. References 1. Markowitz H (1952) Portfolio selection*. The Journal of Finance 7: 77–91. DOI: 10.1111/j.1540- 6261.1952.tb01525.x. 2. Sharpe WF (1964) Capital asset prices: A theory of market equilibrium under conditions of risk*. The Journal of Finance 19: 425–442. DOI: 10.1111/j.1540-6261.1964.tb02865.x. 3. Erdo˝s P, Ormos M (2009) Return calculation methodology: Evidence from the Hungarian mutual fund industry. Acta Oeconomica 59: 391–409. DOI: 10.1556/AOecon.59.2009.4.2. 4. Erdo˝s P, Ormos M, Zibriczky D (2011) Non-parametric and semi-parametric asset pricing. Economic Modelling 28: 1150–1162. DOI: 10.1016/j.econmod.2010.12.008. 5. Clausius R (1870) XVI. On a mechanical theorem applicable to heat. The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science 40: 122–127. DOI: 10.1080/147864470 6. Boltzmann L, Haseno¨hrl F (2012) Weitere Studien u¨ber das Wa¨rmegleichgewicht unter Gas-moleku¨len Wissenschaftliche Abhandlungen: Cambridge University Press. DOI: 10.1017/CBO9781139381420.023. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 20 / 21 Entropy-Based Financial Asset Pricing 7. Shannon CE (1948) A Mathematical Theory of Communication. Bell System Technical Journal 27: 379– 423. DOI: 10.1002/j.1538-7305.1948.tb00917.x. 8. Philippatos GC, Wilson CJ (1972) Entropy, market risk, and the selection of efficient portfolios. Applied Economics 4: 209–220. DOI: 10.1080/00036847200000017. 9. Kirchner U, Zunckel C (2011) Measuring Portfolio Diversification. arXiv preprint arXiv:11024722. 10. Dionisio A, Menezes R, Mendes DA (2006) An econophysics approach to analyse uncertainty in financial markets: an application to the Portuguese stock market. The European Physical Journal B 50: 161–164. DOI: 10.1140/epjb/e2006-00113-2. 11. Maasoumi E, Racine J (2002) Entropy and predictability of stock market returns. Journal of Econometrics 107: 291–312. DOI: 10.1.1.27.1423. 12. Nawrocki DN, Harding WH (1986) State-value weighted entropy as a measure of investment risk. Applied Economics 18: 411–419. DOI: 10.1080/00036848600000038. 13. Renyi A (1961) On Measures of Entropy and Information. Fourth Berkeley Symposium on Mathematical Statistics and Probability; Berkeley, Calif. University of California Press. pp. 547–561. 14. Scott DW (1979) On optimal and data-based histograms. Biometrika 66: 605–610. DOI: 10.1093/biomet/ 66.3.605. 15. Freedman D, Diaconis P (1981) On the histogram as a density estimator: L2 theory. Probability theory and related fields 57: 453–476. DOI: 10.1007/BF01025868. 16. Ha¨rdle W (2004) Nonparametric and semiparametric models: Springer. DOI: 10.1007/978-3-642-17146-8. 17. Silverman BW (1986) Density estimation for statistics and data analysis: CRC press. Monographs on Statistics and Applied Probability 26. DOI: 10.1007/978-1-4899-3324-9. 18. Turlach BA (1993) Bandwidth selection in kernel density estimation: A review: Universite´ catholique de Louvain. DOI: 10.1.1.44.6770. 19. Beirlant J, Dudewicz EJ, Gyo¨ rfi L, Van der Meulen EC (1997) Nonparametric entropy estimation: An overview. International Journal of Mathematical and Statistical Sciences 6: 17–40. 20. Wachowiak MP, Smolikova R, Tourassi GD, Elmaghraby AS (2005) Estimation of generalized entropies with sample spacing. Pattern Analysis and Applications 8: 95–101. DOI: 10.1007/s10044-005- 0247-4. 21. Jensen MC (1968) The performance of mutual funds in the period 1945–1964. The Journal of finance 23: 389–416. DOI: 10.1111/j.1540-6261.1968.tb00815.x. 22. Andor G, Du¨lk M (2013) Harmonic mean as an approximation for discounting intraperiod cash flows. The Engineering Economist 58: 3–18. DOI: 10.1080/0013791X.2012.742607. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 21 / 21 http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png PLoS ONE Pubmed Central

Entropy-Based Financial Asset Pricing

PLoS ONE , Volume 9 (12) – Dec 29, 2014

Loading next page...
 
/lp/pubmed-central/entropy-based-financial-asset-pricing-onu7KcU3Mn

References (33)

Publisher
Pubmed Central
Copyright
© 2014 Ormos, Zibriczky
eISSN
1932-6203
DOI
10.1371/journal.pone.0115742
Publisher site
See Article on Publisher Site

Abstract

We investigate entropy as a financial risk measure. Entropy explains the equity premium of securities and portfolios in a simpler way and, at the same time, with higher explanatory power than the beta parameter of the capital asset pricing model. For asset pricing we define the continuous entropy as an alternative measure of risk. Our results show that entropy decreases in the function of the number of securities involved in a portfolio in a similar way to the standard deviation, and that efficient portfolios are situated on a hyperbola in the expected return – entropy system. For empirical investigation we use daily returns of 150 randomly selected securities for a period of 27 years. Our regression results show that entropy has a higher explanatory power for the expected return than the capital OPEN ACCESS asset pricing model beta. Furthermore we show the time varying behavior of the Citation: Ormos M, Zibriczky D (2014) Entropy- Based Financial Asset Pricing. PLoS ONE 9(12): beta along with entropy. e115742. doi:10.1371/journal.pone.0115742 Editor: Giampiero Favato, Kingston University London, United Kingdom Received: August 8, 2014 Accepted: November 26, 2014 Introduction Published: December 29, 2014 We build an equilibrium capital asset pricing model by applying a novel risk Copyright:  2014 Ormos, Zibriczky. This is an measure, the entropy. Entropy characterizes the uncertainty or measures the open-access article distributed under the terms of the Creative Commons Attribution License, which dispersion of a random variable. In our particular case, it characterizes the permits unrestricted use, distribution, and repro- uncertainty of stock and portfolio returns. In modern Markowitz [1] portfolio duction in any medium, provided the original author and source are credited. theory and equilibrium asset pricing models [2] we apply linear regressions. This Data Availability: The authors confirm that all data methodology supposes that the returns are stationary and normally distributed; underlying the findings are fully available without however, this is not actually the case [3]. Entropy, on the other hand, does not restriction. Data are from the Center for Research in Security Prices (http://www.crsp.com/). Access to have this kind of boundary condition. The main goal of this paper is to apply CRSP data requires a subscription. For subscrip- entropy as a novel risk measure. As a starting point even the density function itself tion in formation please contact subscrip- tions@crsp.chicagobooth.edu. has to be estimated. In the traditional asset pricing model there is equilibrium between expected return the beta parameter, which is the covariance–variance Funding: The authors have no support or funding to report. ratio between the market portfolio and the investigated investment opportunity. If Competing Interests: The authors have declared the random variable is normally distributed then the entropy follows its standard that no competing interests exist. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 1/21 Entropy-Based Financial Asset Pricing deviation; thus in the ideal case there is no difference between the two risk measures. However; our results show that there is a significant difference between the standard deviation, or beta, and the entropy of a given security or portfolio. In this paper we show that entropy offers an ideal alternative for capturing the risk of an investment opportunity. If we explain the return of a wide sample of securities and portfolios with different risk measures then on an ordinary least squares (OLS) regression setting the explanatory power is much higher in the case of the entropy measure of risk than in the case of the traditional measures, both in- sample and out-of-sample. We show that entropy reduction in line with diversification behaves similarly to standard deviation; however at the same time it captures a beta-like systematic risk of single securities or non-efficient portfolios as well. For well-diversified portfolios the explanatory power of entropy is 1.5 times higher than that of the capital asset pricing model (CAPM) beta. We also test and compare entropy with standard risk measures for market circumstances that are increasing and decreasing, and find that the explanatory power of entropy is significantly higher in a bullish market, but lower for a bearish market. Our results for bullish and bearish regimes show that the different risk measures behave similarly in terms of the positive and negative relationship between risk and return. This behavior underlines the fact that the entropy-based risk measure can give contradictory results in the same way as traditional risk estimations in upward and downward regimes. We also compare the entropy-based risk measures with the CAPM beta in and out of sample, which gives information on the predictive power of the different methods. As the CAPM beta measures the systematic risk only, while entropy based risk measures and the standard deviation captures the total risk of the investment our results are shocking, that entropy gives almost twice as high an average explanatory power as the beta, with an average of 40% less standard deviation. A further contribution of the paper is that we introduce a simple method to estimate the entropy of a security or portfolio return. Data In our empirical analysis we apply daily returns from the Center for Research in Security Prices (CRSP) database for the period from 1985 to the end of 2011. We randomly select 150 securities from the S&P500 index components that are available for the full period. The market return is the CRSP value-weighted index return premium above the risk-free rate. The index tracks the return of the New York Stock Exchange (NYSE), the American Stock Exchange (AMEX) and NASDAQ stocks. The risk-free rate is the return of the one-month Treasury bill from the CRSP. We use daily returns because they are not normally distributed (see S1 Table). Erdo ˝s and Ormos (2009) [3] and Erdo ˝s et al. (2011) [4] describe the main difficulties of modeling asset prices with non-normal returns. The daily return calculation enables us to compare different risk measures. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 2/21 Entropy-Based Financial Asset Pricing Methodology Entropy is a mathematically-defined quantity that is generally used for characterizing the probability of outcomes in a system that is undergoing a process. It was originally introduced in thermodynamics by Rudolf Clausius [5]to measure the ratio of transferred heat through a reversible process in an isolated system. In statistical mechanics the interpretation of entropy is the measure of uncertainty about the system that remains after observing its macroscopic properties (pressure, temperature or volume). The application of entropy in this perspective was introduced by Ludwig Boltzmann [6]. He defined the configuration entropy as the diversity of specific ways in which the components of the system may be arranged. He found a strong relationship between the thermodynamic and the statistical aspects of entropy: the formulae for thermodynamic entropy and configuration entropy only differ in the so-called Boltzmann constant. There is an important application of entropy in information theory as well, and this is often called Shannon [7] entropy. The information provider system operates as a stochastic cybernetic system, in which the message can be considered as a random variable. The entropy quantifies the expected value of the information in a message or, in other words, the amount of information that is missing before the message is received. The more unpredictable (uncertain) the message that is provided by the system, the greater the expected value of the information contained in the message. Consequently, greater uncertainty in the messages of the system means higher entropy. Because the entropy equals the amount of expected information in a message, it measures the maximum compression ratio that can be applied without losing information. In financial applications, Philippatos and Wilson [8] find that entropy is more general and has some advantages over standard deviation; in their paper they compare the behaviors of standard deviation and entropy in portfolio manage- ment. Kirchner and Zunckel [9] argue that in financial economics entropy is a better tool for capturing the reduction of risk by diversification; however, in their study they suppose that the assets are Gaussian. Dionisio et al. [10] argue that entropy observes the effect of diversification and is a more general measure of uncertainty than variance, since it uses more information about the probability distribution. The mutual information and the conditional entropy perform well when compared with the systematic risk and the specific risk estimated through the linear equilibrium model. Regarding the predictability of stock market returns, Maasoumi and Racine [11] point out that entropy has several desirable properties and is capable of efficiently capturing nonlinear dependencies in return time series. Nawrocki and Harding [12] propose applying state-value weighted entropy as a measure of investment risk; however, they are dealing with the discrete case. All the above academic papers recognize that entropy could be a good measure of risk; however, it seems to be difficult to use this measure. Our main motivation is to show that an entropy-based risk measure is, on the one hand, more precise, PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 3/21 Entropy-Based Financial Asset Pricing and, on the other hand, no more complicated to use than variance equilibrium models. Discrete entropy function Entropy functions can be divided into two main types, discrete and differential entropy functions. Let X be a discrete random variable. The possible outcomes of this variable are denoted by o ,o ,::,o , and the corresponding probabilities by p 5Pr(X 5o ), p $0 1 2 n i i i and p~1. The generalized discrete entropy function [13] for the variable X is i~1 defined as: H ðÞ X ~ log p , ð1Þ 1{a i~1 where a is the order of entropy, a$0 and a?1, and the base of the logarithm is 2. The order of entropy expresses the weight taken into account in each outcome; if the order of entropy is lower, the more likely outcomes are underweighted, and vice versa. The most widely used orders are a51 and a52. a51 is a special case of generalized entropy. However the substitution of a51 into (1) results in a division by zero. It can be shown, using l’Hopital’s rule for the limit of a51, that H converges to the Shannon entropy: H ðÞ X ~{ p logðÞ p ð2Þ 1 i i i~1 The case of a52 is called collision entropy and similarly to the literature we refer to this special case as ‘‘Re ´ nyi entropy’’ further in the paper: H ðÞ X ~{log p ð3Þ i~1 H (X) is a non-increasing function in a, and both entropy measures are greater than zero provided that there is a finite number of possible outcomes: 0vH ðÞ X ƒH ðÞ X ð4Þ 2 1 Differential entropy function Let X be a continuous random variable taking values from R with a probability density function f(x). Analogously to (1), the continuous entropy is defined as: PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 4/21 Entropy-Based Financial Asset Pricing H ðÞ X ~ ln fxðÞ dx ð5Þ 1{a One can see that the bases of the logarithms in (1) and (5) are different. Although the entropy depends on the base, it can be shown that the value of the entropy changes only by a constant coefficient for different bases. We use the natural logarithm for all differential entropy functions. The formulas for the special cases (a51 and a52) are the following: H ðÞ X ~{ fxðÞln fxðÞdx ð6Þ H ðÞ X ~{ln fxðÞ dx ð7Þ An important difference between discrete and continuous entropy is that while discrete entropy takes only non-negative values, continuous entropy can also take negative values: H ðÞ X [R ð8Þ In practice, standard risk measures like the CAPM beta or standard deviation are calculated on daily or monthly return data. We also follow this practice, and use a formula that is able to capture risk using this kind of data. Since the return on securities can take values from a continuous codomain, we primarily focus on the differential entropy function. However, by grouping return values into bins the discrete entropy function may also be used; this solution is outside the scope of this paper. Entropy estimation For the estimation of differential entropy, the probability density function of the return values needs to be estimated. Let x ,x ,:::,x be the observations of the 1 2 n continuous random variable X, and H (X) the sample-based estimation of a,n H (X). The plug-in estimations of entropy are calculated on the basis of the density function estimation. The probability density function f(x) is estimated by f (x), the integral estimate of entropy, in the following way: H ðÞ X ~ ln f ðÞ x dx, ð9Þ a,n n 1{a where A is the range of integration, which may exclude small and tail values of f (x). We propose to select A 5(min(x), max(x)). n n PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 5/21 Entropy-Based Financial Asset Pricing Histogram One of the simplest methods of density estimation is the histogram-based density estimation. Let b 5(max(x), min(x)) be the range of sample values; partition the range into k bins of equal width and denote the cutting points by t . The width of a bin is constant: h~ ~t {t . The density function is estimated by using the jz1 j following formula: f ðÞ x ~ , ð10Þ nh th if x(t , t ), where n is the number of data points falling in the j bin. j j+1 j Based on the properties of the histogram, a simpler non plug-in estimation formula can be deduced for Shannon and Renyi entropy using (6), (7), (9) and (10): 1 n H ðÞ X ~ v ln ð11Þ 1,n j n nh j~1 X 2 H ðÞ X ~{ln h ð12Þ 2,n nh j~1 The parameter of this method is the number of equal width bins (k). However, there are several methods for choosing this parameter (e.g. the square root choice, Scott’s normal reference rule [14], or the Freedman-Diaconis rule [15]); the detailed descriptions of these are outside the scope of this paper. Kernel density estimation The kernel-based density estimation is another commonly used method. It applies the following formula: 1 x{x f ðÞ x ~ K , ð13Þ nh h i~1 where KðÞ is the kernel function, and h is the bandwidth parameter. There are several kernel functions that can be used (see S2 Table); for practical reasons (computational time), we propose using the indicator-based Epanechnikov kernel function: KzðÞ~ 1{z I , ð14Þ fg jj z ƒ1 where I is the indicator function. Ha ¨rdle [16] shows that the choice of the kernel function is only of secondary importance, so the focus is rather on the right choice of bandwidth (h). One of the most widely used simple formulas for the estimation of h is Silverman’s rule of thumb [17]: PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 6/21 Entropy-Based Financial Asset Pricing sffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi () 1 IQRðÞ x 1 h ~1:06 min ðÞ x{x , n , ð15Þ rot i n{1 1:34 i~1 where IQR(x) is the interquartile range of x. As the formula assumes a normal distribution for X it gives an approximation for optimal bandwidth; despite this, Silverman’s rule of thumb can be used for a good initial value for more sophisticated optimization methods [18]. Sample spacing estimation Let x #x #…#x be the corresponding order of x , x ,…,x , assuming that n,1 n,2 n,n 1 2 n this is a sample of i.i.d. real-valued random variables. x 2x is called a n,i+m n,i spacing of order m (1#i,i+m,n). The simple sample spacing density estimate is the following [19]: m 1 f ðÞ x ~ , ð16Þ n x {x n,im n,(i{1)m if x[x , x ). n,(i-1)m n,im Wachowiak et al. [20] introduced another variation of the sample spacing density estimation, called the Correa estimator: izm=2 x{x ðÞ j{i j i j~i{m=2 f ðÞ x ~ , ð17Þ izm=2 x{x j i j~i{m=2 izm=2 if i:x [[x , x ); x~ x , and 1#j#n. n,i n,i+1 i j mz1 j~i{m=2 The parameter for sample spacing methods is the fixed order m. For practical reasons (e.g. different sizes of samples) we suggest using m , which depends on the size of the sample and is calculated by the following formula: m ~q r, ð18Þ where k is the number of bins, and the braces indicate the ceiling function. Beirlant et al. [19] overview several additional entropy estimation methods, such as resubstitution, splitting-data and cross-validation; however, our paper focuses on the applications that are used most often. Risk estimation Let the following be a given set of data: D :fg S,R,R ,R ð19Þ M F PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 7/21 Entropy-Based Financial Asset Pricing The elements are the set of securities S:{S , S ,…,S }, with the corresponding 1 2 l observations being R:{R , R ,…,R }, where R 5(r , r ,…,r ). The observation for 1 2 l i i1 i2 in the market return is R 5(r , r ,…,r ), and the observation for the risk free M M1 M2 Mn return is R 5(r , r ,…,r ) where l is the number of securities and n is the F F1 F2 Fn number of samples. Let us recall that the main goal of this paper is to apply entropy as a novel risk measure. In order to handle the risk measure uniformly, we introduce k as a unified property for securities. Let k(S ) be the risk estimate for the security i. In the economic literature the most widely used risk measures are the standard deviation and the CAPM beta. Let us denote these by k and k , respectively. The s b estimation of these risk measures for the security i is the following: k ^ ðÞ S ~sðÞ R{R ð20Þ s i i F and covðÞ R{R ,R {R i F M F k ðÞ S ~b~ , ð21Þ b i sðÞ R {R M F where b is the CAPM beta, covðÞ is the covariance of the arguments and s is the standard deviation. Our hypothesis is that uncertainty about the observation values can be interpreted as a risk of the security, and for this reason we apply entropy as a risk measure. Because the differential entropy function can also take negative values, for better interpretability we apply the exponential function to the entropy, and we define the entropy-based risk measure by the following formula: H R {R nðÞ i F k ðÞ S ~e ð22Þ H i One can see that k takes values from the non-negative real numbers, k [[0,+‘). Explanatory and predictive power In order to compare the efficiency of the risk estimation methods, we introduce two basic evaluation approaches, the measurement of in-sample explanatory power and the measurement of out-of-sample predictive power. In-sample Let V be a target variable, with sample v~ðÞ v ,v ,:::,v , and let U be a single 1 2 l explanatory variable with sample u~ðÞ u ,u ,:::,u . To estimate the explanatory 1 2 l power of the variable U for the variable V, we use the following method. The linear relationship between the two variables can be described using the linear regression model: V~a za Uz". 0 1 The parameters of the model (a and a ) are estimated by ordinary least squares 0 1 (OLS), and the estimation for the target value is the following: ^v~^a z^a u where i 0 1 i ^a and ^a are the estimations of a and a , respectively. One of the most 0 1 0 1 PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 8/21 Entropy-Based Financial Asset Pricing commonly applied estimations of the explanatory power is the R (goodness of fit, or coefficient of determination) of the linear regression: ðÞ v{ðÞ ^a z^a u i 0 1 i i~1 RðÞ v,u ~1{ ð23Þ ðÞ v{v i~1 We are curious as to how efficiently the different risk measures describe the expected return of a security, and we denote this measure by g(k). Let the explanatory variable U be the risk measure of the securities, where the sample is: u ~ðÞ k ^ðÞ S ,k ^ðÞ S ,:::,k ^ðÞ S , ð24Þ k 1 2 l and the target variable T is the expected risk premium of the securities, where the sample is: v ~ðÞ ER½ {R ,ER½ {R ,:::,ER½ {R , ð25Þ m 1 F 2 F l F where k is the unified risk measure function, and E½ is the expected value of the argument. We define the estimation of the in-sample explanatory power (efficiency) as the R of the previously defined variables (24) and (25): ^gkðÞ~R v ,u ð26Þ m k Out of sample Let us create a split of samples for a given D:{S, R, R , R } data set (19): M F I I I I I O O O O O D : S ,R ,R ,R , D : S ,R ,R ,R , ð27Þ M F M F I I I I where the corresponding samples for the securities are R : R ,R ,:::,R , 1 2 l I O O O O O R ~ðÞ r ,r ,:::,r and R : R ,R ,:::,R , R ~ r ,r ,:::,r , the i1 i2 im imðÞ z1 imðÞ z2 imðÞ zp i 1 2 i split for market returns is R ~ðÞ r ,r ,:::,r and M1 M2 Mm R ~ r ,r ,:::,r , and the split for the risk free rates is MmðÞ z1 MmðÞ z2 MmðÞ zp I O I O R ~ r ,r ,:::,r and R ~ r ,r ,:::,r , where jj S ~jj S , F1 F2 FmðÞ zp FmðÞ z1 FmðÞ z2 FmðÞ zp F F I O R ~m, R ~p,(1#i#l), and m+p5n. i i The explanatory values contain the risk estimates for the set of securities based on the data set D : I I I I u ~ k ^ S ,k ^ S ,:::,k ^ S , ð28Þ k 1 2 l and the target values are the expected risk premium of the securities based on D : O O O O O O O v ~ ER {R ,ER {R ,:::,ER {R ð29Þ m 1 F 2 F l F Based on (26), (28) and (29), the estimation of the out-of-sample explanatory (predictive) power is the following: PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 9/21 Entropy-Based Financial Asset Pricing 2 O I g ðÞ k ~R v ,u ð30Þ m k Both in- and out of sample we test whether the difference between the explanatory power of the investigated risk measures (standard deviation, CAPM beta, Shannon- and Renyi entropy) are significant by applying bootstrapping method. In our bootstrap iteration we remove 25 random stocks from the investigated 150 ones and measure the R s of the four different models. We apply 1000 iterations to approximate the distribution of R values on random selection, and we test the equality of means of R s by applying t-test on the generated samples. Results and Discussion We present the empirical results in four parts. First, we show how the entropy behaves in the function of securities involved into the portfolio. Second, we present the long-term explanatory power of the investigated models. Third we examine and compare the performance of different risk measures in in upward and downward market trends. Fourth we apply the different risk parameters to predict future returns, thus we test the out of sample explanatory power of the well-known risk parameters and compare their efficiency to the entropy based risk measures. Characterizing the diversification effect We investigate whether entropy is able to measure the reduction of risk by diversification. We generate 10 million random equally-weighted portfolios with different numbers of securities involved (at most 100,000 for each size), based on the 150 randomly selected securities from the S&P500. The risk of portfolios is estimated by standard deviation, and by the Shannon and Re ´ nyi entropies using risk premiums for the full period. Because the CAPM beta measures the systematic risk only, we exclude it from the investigation of risk reduction. Both types of entropy functions are calculated by the histogram-based density function estimation, with 175 bins for the Shannon entropy and 50 bins for the Renyi entropy. (We tested the histogram, sample spacing and kernel density estimation methods, and the histogram-based method proved to be the most efficient in terms of explanatory and predictive power and simplicity. See our results in S3 Table.) Fig. 1 shows the diversification effects that are characterized by the entropic risk measures and by the standard deviation. For 10 random securities involved in the portfolio, approximately 40% of risk reduction can be achieved compared to a single random security, based on all of the three risk estimators under investigation. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 10 / 21 Entropy-Based Financial Asset Pricing Fig. 1. Average value of risk and risk reduction vs. number of securities in portfolio. Note: We generate 10 million random equally weighted portfolios with different number of securities involved (at most 100,000 for each size) based on 150 randomly selected securities from S&P500. The risk of portfolios is estimated by standard deviation (gray continuous curve), Shannon- (black continuous curve) and Re´nyi entropy (black dashed curve) in the period from 1985 to the end of 2011. Both types of entropy functions are calculated by histogram based density function estimation. The left chart shows the average risk estimates for each portfolio size, and the right chart shows the risk reduction compared to an average risk of single security portfolio. doi:10.1371/journal.pone.0115742.g001 Fig. 1 suggests that entropy shows behavior that is similar to but not the same as standard deviation, so it can serve as a good measure of risk. We also investigate how the different portfolios behave in the expected return – risk coordinate system in the function of diversification. We generate 200-200 random equally- weighted portfolios with 2, 5 and 10 securities involved, and compare these to single securities using standard deviation, the CAPM beta, the Shannon entropy and the Re ´ nyi entropy as risk measures; the results are presented in Fig. 2. Fig. 2 shows the performance of random portfolios by diversification using different risk estimation methods. One can see that the characteristics of standard deviation and entropy are quite similar, with the portfolios being situated on a hyperbola as in the portfolio theory of Markowitz [1]. Different characteristics can be observed by using the CAPM beta; the more securities that are involved in a portfolio, the closer they are situated in the center of the coordinate system. Long term explanatory power In order to evaluate how efficiently the risk measures explain the expected risk premium over a long period, we estimate the risk for each security using standard deviation, the CAPM beta, and the Shannon and Re ´ nyi entropies based on the full period (denoted by P1). The single explanatory variable is the risk measure; the target variable is the expected risk premium of the security. We apply the explanatory power estimation by calculating ^gkðÞ (R ) for each risk measure. Fig. 3 shows the efficiency of explaining the expected risk premium by the different risk measures; the expected daily risk premium is presented as a function of risk measure. The CAPM beta performs the worst, with 6.17% efficiency. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 11 / 21 Entropy-Based Financial Asset Pricing Fig. 2. Portfolios with different number of securities involved in E(r) – risk system. Note: The panels show the expected risk premium of the portfolios (calculated by the average of daily risk premiums) versus the estimated risk using different methods; the number of securities involved is indicated by the different markers. We generate a sample of 750 random portfolios by using 150 randomly selected securities and 200-200 random equally weighted portfolios with 2, 5 and 10 securities. The risk of portfolios is estimated by standard deviation, CAPM beta, Shannon- and Re´nyi entropy by using daily returns in the period from 1985 to the end of 2011. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. doi:10.1371/journal.pone.0115742.g002 However, the explanatory power of standard deviation (7.83%) is higher than that of the CAPM beta, and both entropies perform significantly better, with efficiency of 12.98% for the Shannon entropy and 15.71% for the Renyi entropy. Based on the equation of linear regressions, the average unexplained risk premium (intersect on the Y-axis or Jensen alpha [21]) for the entropy methods (0.0091, 0.0059) is lower than that for the standard methods (0.0170 for standard deviation and 0.0209 for the CAPM beta). PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 12 / 21 Entropy-Based Financial Asset Pricing Fig. 3. Explanatory power of risk measures in long term. Note: The four panels show the relationship between risk premium and risk (standard deviation, CAPM beta, Shannon- and Re´nyi entropy) of 150 randomly selected securities by using different estimation methods. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. The equation and the explanatory power (R ) of the linear regressions are presented using expected risk premium as target variable and risk as explanatory variable. Under the OLS regression equations in brackets the p-values can be seen for each parameter estimations. The R s of the models applying entropy based risk measures are significantly different form standard deviation and CAPM beta at 1% level. doi:10.1371/journal.pone.0115742.g003 We also measure the explanatory power for different numbers of securities involved in the portfolio, by generating at most 100,000 samples for each; we present these results in Fig. 4. Fig. 4 illustrates how the explanatory power changes with diversification. One can see that the explanatory power of standard deviation and entropy decreases with an increase in the number of securities involved in the portfolio, while the performance of the CAPM beta is nearly constant. While the CAPM beta models the systematic risk only, the standard deviation and entropy are capable of measuring specific risk, which gives additional explanatory power for less- diversified portfolios. Despite the decreased explanatory power of both entropy PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 13 / 21 Entropy-Based Financial Asset Pricing Fig. 4. Explanatory power of risk measures in long term by diversification. Note: This figure shows the explanatory power (R ) of portfolios with different number of securities involved and different risk estimate methods. We generate 10 million random equally weighted portfolios with different number of securities involved (at most 100,000 for each size) using daily risk premiums of 150 randomly selected securities. The risk of portfolios is estimated by standard deviation (light gray curve), CAPM beta (black dotted curve), Shannon- (gray) and Re´nyi entropy (black). Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. doi:10.1371/journal.pone.0115742.g004 functions, they perform better than the CAPM beta in all the cases that were investigated. For well-diversified portfolios the explanatory power of the Renyi entropy is 1.5 times higher than that of the CAPM beta. Explanatory power by primary market trends We split the original 27-year sample by primary market trend into a ‘‘bullish’’ and a ‘‘bearish’’ sample (denoted by P1+ and P1-), containing returns for upward and downward periods, respectively (for the labels of the periods see S4 Table). For these two sample sets we investigate the explanatory power for standard deviation, the CAPM beta, and the Shannon and Re ´ nyi entropies using the same parameter for the histogram-based entropy estimation as for the previous experiments. Fig. 5 and Fig. 6 show the results in the expected risk premium – risk coordinate system. Our results for the bullish and bearish regimes show that the different risk measures behave similarly in terms of the positive and negative relationships between risk and return. This behavior underlines the fact that an entropy-based risk measure can give contradictory results in a similar way to traditional risk estimations in different regimes. In bullish market circumstances we find a very high explanatory power for all kinds of risk measures: 33.90%, 36.67%, 43.45% and 42.36% with standard deviation, the CAPM beta, the Shannon entropy and the Renyi entropy, respectively. As for the full sample tests, the slopes of the regression lines are positive, meaning that higher risk-taking promises higher returns. In contrast to the bullish market, during downward trends higher risk- taking does not result in higher returns and, indeed, the higher the risk the higher the negative premium achieved by the investor. We have to mention that the explanatory power of the CAPM beta is higher than that of the entropy-based risk measures. Our entropy results are in line with those for the CAPM beta; and the regime dependency is clear as well. On the other hand, the explanatory power is PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 14 / 21 Entropy-Based Financial Asset Pricing Fig. 5. Explanatory power of risk measures in bullish sample. Note: The panels show the relationship between the expected risk premium of securities and risk by using different estimation methods. We present the equation of linear regression and the goodness of fit (R ).We estimated the risk of 150 random securities in upward trend periods (bull market) from 1985 to the end of 2011 using standard deviation, CAPM beta, Shannon- and Re´nyi entropy risk estimation methods. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. Under the OLS regression equations in brackets the p-values can be seen for each parameter estimations. The R s of the models applying entropy based risk measures are significantly higher than the models with standard deviation and CAPM beta at 1% level. doi:10.1371/journal.pone.0115742.g005 again much higher for this regime than for the full sample. Altogether, we argue that the test results for the full sample give a better comparison opportunity, as the sample sizes of the bullish and bearish markets are different and at the present moment the investor cannot decide whether there is an upward or a downward trend. Short term explanatory and predictive power Although attractive results are achieved within the sample, this does not necessarily mean high efficiency outside the sample. Therefore we took several ten-year periods, shifting the starting year by one year for each, with the first PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 15 / 21 Entropy-Based Financial Asset Pricing Fig. 6. Explanatory power of risk measures in bearish sample. Note: The panels show the relationship between the expected risk premium of securities and risk by using different estimation methods. We present the equation of linear regression and the goodness of fit (R ).We estimated the risk of 150 random securities in downward trend periods (bear market) from 1985 to the end of 2011 using standard deviation, CAPM beta, Shannon- and Re´nyi entropy risk estimation methods. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. Under the OLS regression equations in brackets the p-values can be seen for each parameter estimations. The R of the models applying entropy based risk measures are significantly higher than the model with standard deviation at 1% level. doi:10.1371/journal.pone.0115742.g006 period being 1985 to 1994 and the last 2002 to 2011. As the full data set covers 27 complete years, we used 18 ten-year periods. We split each ten-year period into two shorter five-year periods (P2i and P2o), with the risk measures being estimated based on the first period and the predictive efficiency being measured in the second period. In the previous sections, we have presented the results for in- sample for the full sample and for the different regimes, and here we summarize these and we also compare the long-term in-sample results with the short-term in- sample and out-of–sample results. Table 1 summarizes the explanatory power of the investigated risk measures for the different samples. ^g ,^g z , and ^g show the results of the long-term analysis P1 P1 P1 PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 16 / 21 Entropy-Based Financial Asset Pricing PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 17 / 21 Table 1. Efficiency of explaining and predicting risk premium in different samples. Risk measure ^g ^g ^g ^g ^g s ðÞ ^g sðÞ ^g z { P1 P1 P1 P2i P2o R P2i R P2o Standard deviation 7.83% 33.9% 36.7% 7.94% 9.7% 0.75 0.65 Beta 6.17% 36.7% 43.7% 13.31% 6.45% 0.98 1.02 Shannon entropy 12.98% 43.5% 39.6% 13.38% 10.15% 0.69 0.64 Re´nyi entropy 15.71% 42.4% 38.6% 12.82% 9.34% 0.63 0.62 Note: The table summarizes the explanatory power (in sample R ) of the investigated risk measures in different samples. We estimate risk measures of 150 random securities using standard deviation, CAPM beta, Shannon- and Re´nyi entropy risk estimation methods for (1) long term, from 1985 to the end of 2011 (1985–2011); (2) long term on upward trends (bull market), (3) long term on downward trends (bear market), (4) 18 10-year periods shifting by one year from period (1985–1994) to period (2002–2011), split into two 5-5 year periods for each. Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. The ^g shows P1 the explanatory power of risk measures for long term, the ^g z, and ^g { summarizes the explanatory power on upward and downward trends, respectively, ^g stands for the average P1 P1 P2i explanatory power of risk measured in the first 5 years of 10-year shorter periods in sample, and ^g shows the average predicting power of risk measures (out P2o of sample R ) calculated by estimating risk in the first 5 years and evaluating them on the other 5 years in each 10-year periods. The last two columns show the relative standard deviation of the explanatory and predicting power based on the 18 shorter periods for the investigated risk measures. doi:10.1371/journal.pone.0115742.t001 Entropy-Based Financial Asset Pricing Table 2. Explanatory power in short period samples. T-test T-test P2i H H H ,s T-test H ,b H ,s T-test H ,b sig H (s/b) sig H (s/b) sb 1 2 1 1 2 2 1 2 1985–1989 3.2% 9.4% 5.1% 3.7% 19.3 232.8 6.7 246.0 ***/ ***/ 1986–1990 1.7% 3.6% 2.6% 3.6% 11.8 212.8 22.7 0.4 ***/ ***/ 1987–1991 4.1% 4.8% 6.0% 7.9% 13.9 7.6 27.8 21.8 ***/*** ***/*** 1988–1992 5.5% 5.0% 6.7% 6.8% 7.7 8.5 8.1 8.8 ***/*** ***/*** 1989–1993 3.5% 4.2% 8.9% 7.2% 40.6 32.9 28.9 21.4 ***/*** ***/*** 1990–1994 9.6% 7.1% 23.4% 20.1% 66.1 75.8 48.2 57.7 ***/*** ***/*** 1991–1995 16.0% 13.6% 28.1% 21.9% 67.7 56.3 32.9 30.8 ***/*** ***/*** 1992–1996 16.3% 17.8% 24.4% 20.5% 41.8 29.1 24.4 12.4 ***/*** ***/*** 1993–1997 7.5% 24.9% 15.4% 13.6% 38.1 240.0 32.7 255.2 ***/ ***/ 1994–1998 7.0% 30.1% 15.8% 12.6% 86.1 2101.1 55.6 2124.3 ***/ ***/ 1995–1999 16.6% 51.2% 28.3% 27.3% 88.3 2166.4 77.8 2167.8 ***/ ***/ 1996–2000 8.6% 28.2% 18.0% 20.8% 63.7 267.7 85.0 249.7 ***/ ***/ 1997–2001 2.1% 15.3% 7.2% 9.7% 39.3 260.0 58.2 241.7 ***/ ***/ 1998–2002 0.2% 2.5% 1.5% 2.4% 28.2 216.0 41.8 20.3 ***/ ***/ 1999–2003 6.1% 7.5% 8.1% 9.8% 16.9 6.2 29.2 18.8 ***/*** ***/*** 2000–2004 1.9% 0.1% 1.5% 1.4% 28.0 35.5 29.2 34.8 /*** /*** 2001–2005 15.1% 5.6% 17.5% 18.4% 16.0 93.5 22.4 102.2 ***/*** ***/*** 2002–2006 17.9% 8.9% 22.3% 23.1% 24.6 89.6 30.2 98.8 ***/*** ***/*** Average 7.94% 13.31% 13.37% 12.82% Rel. dev 0.75 0.98 0.69 0.63 Note: This table summarizes the explanatory power of the different risk measures for expected risk premium in the first 5 years of 18 10-year periods (P2i) shifting by one year from period (1985–1994) to period (2002–2011). We estimate and evaluate risk measures of 150 randomly selected securities from the S&P500 index using standard deviation (s), CAPM beta (b), Shannon entropy (H ) and Re´nyi entropy (H ) risk estimation methods by daily risk premiums. 1 2 Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. We apply t-statistics by bootstrapping method to measure whether differences in R s are significant. We use *s to designate that the entropy based risk measure is significantly higher than the standard deviation and CAPM beta; ***, ** and * stands for 1%, 5% and 10% significance level respectively. doi:10.1371/journal.pone.0115742.t002 for the full period and during the upward and downward trends, respectively; ^g ,and ^g stand for the average efficiency measured for short-term in-sample P2i P2o and out-of-sample, respectively; and sðÞ ^g , and sðÞ ^g measure the relative R R P2i P2o standard deviation of the efficiency when applying the in-sample and out-of- sample test for short periods (For the detailed results for all periods see Table 2 and Table 3). While the standard deviation risk measure performs almost the same in the long and the short run (7.83% vs. 7.94), its predictive efficiency is surprisingly good (9.70%). The explanatory power of the CAPM beta in the long period is low (6.17%), while the average efficiency in the short periods is more than twice as high (13.31%). We use arithmetic averages [22]. Comparing the results for in-sample and the out-of-sample, the predictive power of the beta is relatively low (6.45%), which suggests that the model may be over-fitted for the training sample. The Shannon entropy performs better than the standard deviation and the CAPM beta in each sample. The Renyi entropy shows the highest explanatory power in the long run; however, in short periods the Renyi PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 18 / 21 Entropy-Based Financial Asset Pricing Table 3. Predicting power in short periods out of sample. T-test T-test T-test T-test sig H P2i P2o H H H ,s H ,b H ,s H ,b sig H (s/b) (s/b) sb 1 2 1 1 2 2 1 1985–1989 1990–1994 7.3% 2.8% 13.0% 10.0% 35.2 80.4 15.4 52.0 ***/*** ***/*** 1986–1990 1991–1995 17.0% 4.1% 19.3% 18.1% 12.8 101.4 5.6 85.8 ***/*** ***/*** 1987–1991 1992–1996 21.5% 5.9% 22.6% 17.5% 7.1 99.8 223.0 71.8 ***/*** /*** 1988–1992 1993–1997 9.8% 7.9% 14.6% 13.2% 34.9 43.9 27.8 38.1 ***/*** ***/*** 1989–1993 1994–1998 7.9% 16.5% 13.5% 11.6% 65.8 221.3 45.5 235.8 ***/ ***/ 1990–1994 1995–1999 10.0% 23.9% 16.6% 14.9% 63.2 257.1 48.0 270.1 ***/ ***/ 1991–1995 1996–2000 9.0% 14.1% 9.1% 9.0% 0.1 244.0 20.7 245.0 / / 1992–1996 1997–2001 11.3% 14.7% 11.7% 11.8% 1.7 220.8 2.0 221.0 */ **/ 1993–1997 1998–2002 14.2% 4.7% 12.7% 10.8% 211.0 66.8 226.3 54.4 /*** /*** 1994–1998 1999–2003 24.7% 2.7% 17.5% 19.8% 254.6 154.7 236.3 173.0 /*** /*** 1995–1999 2000–2004 3.6% 6.8% 0.3% 0.5% 259.0 290.3 255.0 287.1 / / 1996–2000 2001–2005 8.0 0.0% 3.6% 3.0% 247.1 67.9 254.5 62.7 /*** /*** 1997–2001 2002–2006 10.3% 0.4% 6.1% 4.5% 238.9 91.9 256.1 78.8 /*** /*** 1998–2002 2003–2007 7.8% 3.2% 6.4% 5.8% 213.4 40.8 219.7 35.0 /*** /*** 1999–2003 2004–2008 1.5% 3.2% 1.9% 2.1% 5.5 218.5 7.7 216.3 ***/ ***/ 2000–2004 2005–2009 4.7% 1.2% 5.0% 5.1% 4.8 62.9 5.9 63.1 ***/*** ***/*** 2001–2005 2006–2010 2.2% 1.9% 3.2% 4.0% 17.8 20.9 28.9 31.5 ***/*** ***/*** 2002–2006 2007–2011 4.0% 2.3% 5.4% 6.5% 21.5 50.1 35.2 62.2 ***/*** ***/*** Average 9.70% 6.45% 10.14% 9.34% Relative deviation 0.65 1.02 0.64 0.62 Note: This table summarizes the predicting power of the investigated risk measures for expected risk premium in the last 5 years of 18 10-year periods shifting by one year from period (1985–1994) to period (2002–2011). We estimate risk measures of 150 randomly selected securities from the S&P500 index using standard deviation (s), CAPM beta (b), Shannon entropy (H ) and Re´nyi entropy (H ) risk estimation methods by daily risk premiums in the first 5 years 1 2 (P2i) and measure the predicting power on the next 5 years (P2o) by estimating the goodness of fit of linear regression (R ). Both types of entropy functions are calculated by histogram based density function estimation, with 175 bins for Shannon entropy and 50 bins for Re´nyi entropy. We apply t-statistics by bootstrapping method to measure whether differences in R s are significant. We use *s to designate that the entropy based risk measure is significantly higher than the standard deviation and CAPM beta; ***, ** and * stands for 1%, 5% and 10% significance level respectively. doi:10.1371/journal.pone.0115742.t003 entropy performs worse than the Shannon entropy. Comparing the reliability of the risk estimators, the standard deviation of the in-sample and out-of-sample results is the lowest for the entropy risk measures, and the highest for the CAPM beta. Summarizing our results, we state that the beta can beat the entropy only in the case of bearish market circumstances. In any other situation, entropy seems to be a better and more reliable risk measure. Conclusions Entropy as a novel risk measure combines the advantages of the CAPM’s risk parameter (beta) and the standard deviation. It captures risk without using any information about the market, and it is capable of measuring the risk reduction effect of diversification. The explanatory power for the expected return within the sample is better than the beta, especially in the long run covering bullish and PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 19 / 21 Entropy-Based Financial Asset Pricing bearish periods; the predictive power for the expected return is higher than for standard deviation. Both the Shannon and the Renyi entropies give more reliable risk estimation; their explanatory power exhibits significantly lower variance compared to the beta or the standard deviation. If upward and downward trends are distinguished, the regime dependency of entropy can be recognized: this result is similar to that for the beta. Among the entropy estimation methods reviewed, the histogram-based method proved to be the most efficient in terms of explanatory and predictive power; we propose a simple estimation formula for the Shannon and the Renyi entropy functions, which facilitates the application of an entropy-based risk measure. Supporting Information S1 Table. Descriptive statistics. doi:10.1371/journal.pone.0115742.s001 (DOCX) S2 Table. The most often used Kernel functions. doi:10.1371/journal.pone.0115742.s002 (DOCX) S3 Table. Explanatory power of Shannon entropy by different density estimation methods. doi:10.1371/journal.pone.0115742.s003 (DOCX) S4 Table. Labeling periods by market trend. doi:10.1371/journal.pone.0115742.s004 (DOCX) Author Contributions Conceived and designed the experiments: MO DZ. Performed the experiments: MO DZ. Analyzed the data: MO DZ. Contributed reagents/materials/analysis tools: MO DZ. Wrote the paper: MO DZ. References 1. Markowitz H (1952) Portfolio selection*. The Journal of Finance 7: 77–91. DOI: 10.1111/j.1540- 6261.1952.tb01525.x. 2. Sharpe WF (1964) Capital asset prices: A theory of market equilibrium under conditions of risk*. The Journal of Finance 19: 425–442. DOI: 10.1111/j.1540-6261.1964.tb02865.x. 3. Erdo˝s P, Ormos M (2009) Return calculation methodology: Evidence from the Hungarian mutual fund industry. Acta Oeconomica 59: 391–409. DOI: 10.1556/AOecon.59.2009.4.2. 4. Erdo˝s P, Ormos M, Zibriczky D (2011) Non-parametric and semi-parametric asset pricing. Economic Modelling 28: 1150–1162. DOI: 10.1016/j.econmod.2010.12.008. 5. Clausius R (1870) XVI. On a mechanical theorem applicable to heat. The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science 40: 122–127. DOI: 10.1080/147864470 6. Boltzmann L, Haseno¨hrl F (2012) Weitere Studien u¨ber das Wa¨rmegleichgewicht unter Gas-moleku¨len Wissenschaftliche Abhandlungen: Cambridge University Press. DOI: 10.1017/CBO9781139381420.023. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 20 / 21 Entropy-Based Financial Asset Pricing 7. Shannon CE (1948) A Mathematical Theory of Communication. Bell System Technical Journal 27: 379– 423. DOI: 10.1002/j.1538-7305.1948.tb00917.x. 8. Philippatos GC, Wilson CJ (1972) Entropy, market risk, and the selection of efficient portfolios. Applied Economics 4: 209–220. DOI: 10.1080/00036847200000017. 9. Kirchner U, Zunckel C (2011) Measuring Portfolio Diversification. arXiv preprint arXiv:11024722. 10. Dionisio A, Menezes R, Mendes DA (2006) An econophysics approach to analyse uncertainty in financial markets: an application to the Portuguese stock market. The European Physical Journal B 50: 161–164. DOI: 10.1140/epjb/e2006-00113-2. 11. Maasoumi E, Racine J (2002) Entropy and predictability of stock market returns. Journal of Econometrics 107: 291–312. DOI: 10.1.1.27.1423. 12. Nawrocki DN, Harding WH (1986) State-value weighted entropy as a measure of investment risk. Applied Economics 18: 411–419. DOI: 10.1080/00036848600000038. 13. Renyi A (1961) On Measures of Entropy and Information. Fourth Berkeley Symposium on Mathematical Statistics and Probability; Berkeley, Calif. University of California Press. pp. 547–561. 14. Scott DW (1979) On optimal and data-based histograms. Biometrika 66: 605–610. DOI: 10.1093/biomet/ 66.3.605. 15. Freedman D, Diaconis P (1981) On the histogram as a density estimator: L2 theory. Probability theory and related fields 57: 453–476. DOI: 10.1007/BF01025868. 16. Ha¨rdle W (2004) Nonparametric and semiparametric models: Springer. DOI: 10.1007/978-3-642-17146-8. 17. Silverman BW (1986) Density estimation for statistics and data analysis: CRC press. Monographs on Statistics and Applied Probability 26. DOI: 10.1007/978-1-4899-3324-9. 18. Turlach BA (1993) Bandwidth selection in kernel density estimation: A review: Universite´ catholique de Louvain. DOI: 10.1.1.44.6770. 19. Beirlant J, Dudewicz EJ, Gyo¨ rfi L, Van der Meulen EC (1997) Nonparametric entropy estimation: An overview. International Journal of Mathematical and Statistical Sciences 6: 17–40. 20. Wachowiak MP, Smolikova R, Tourassi GD, Elmaghraby AS (2005) Estimation of generalized entropies with sample spacing. Pattern Analysis and Applications 8: 95–101. DOI: 10.1007/s10044-005- 0247-4. 21. Jensen MC (1968) The performance of mutual funds in the period 1945–1964. The Journal of finance 23: 389–416. DOI: 10.1111/j.1540-6261.1968.tb00815.x. 22. Andor G, Du¨lk M (2013) Harmonic mean as an approximation for discounting intraperiod cash flows. The Engineering Economist 58: 3–18. DOI: 10.1080/0013791X.2012.742607. PLOS ONE | DOI:10.1371/journal.pone.0115742 December 29, 2014 21 / 21

Journal

PLoS ONEPubmed Central

Published: Dec 29, 2014

There are no references for this article.