Abstract
This chapter proposes a method to quantify the structure of a bipartite graph using a network entropy per link. The network entropy of a bipartite graph with random links is calculated both numerically and theoretically. As an application of the proposed method to analyze collective behavior, the affairs in which participants quote and trade in the foreign exchange market are quantified. The network entropy per link is found to correspond to the macroeconomic situation. A finite mixture of Gumbel distributions is used to fit the empirical distribution for the minimum values of network entropy per link in each week. The mixture of Gumbel distributions with parameter estimates by segmentation procedure is verified by the Kolmogorov–Smirnov test. The finite mixture of Gumbel distributions that extrapolate the empirical probability of extreme events has explanatory power at a statistically significant level. This method is applicable to detecting extreme synchrony in various types of socioeconomic systems.
Access provided by Autonomous University of Puebla. Download chapter PDF
Similar content being viewed by others
1 Introduction
Recently a comprehensive measurement of agent behavior can be conducted based on high-resolution data on socioeconomic systems due to the development of information and communication technology. In particular, we want to consider a problem that we need to deal with the collective behavior of millions of nodes contribute to the observable dynamical features of such a complex system (de Menezes and Barabási 2004).
To do so, it is necessary to develop an adequate model both which considers states of agent behavior from the comprehensive point of view and which is as simple as one can estimate its parameters from actual data under assumptions. Specifically in socioeconomic activities, if we regard the situation where people exchange things, money, and information with each other as a network, then it further seems to be fruitful to study such social or economic systems from a network point of view.
To make advances in this direction, we need to treat structure on the basis of information transmission among heterogeneous agents in a given socioeconomic system from a limited amount of available data without precise knowledge on communication networks.
Several networks relating to human activity can be described as bipartite graphs with two kinds of nodes as shown in Fig. 1. For example, it is known that financial markets (financial commodities and participants), blog systems (blogs and bloggers), and economic systems (firms and goods/consumers) can be described as a directed bipartite graph (Lambiotte et al. 2007; Chmiel et al. 2007; Sato 2007; Sato and Hołyst 2008).
Such a bipartite network is recognized by observers when constituents are transmitted between two kinds of nodes. Links are recognized by observers when a constituent moves from one node to another node. Therefore bipartite network representation also seems to be one of our cognitive categories such as causality, time and space, intensity, quantity, and so on.
This chapter considers a model-based comparative measurement of collective behavior of groups based on their activities (Sato 2017). Specifically, a descriptive multivariate model of a financial market is proposed from a comprehensive point of view. Using comprehensive high-resolution data on the behavior of market participants, correlations of log returns and quotations (transactions) are analyzed on the basis of theoretical insights from the proposed model.
As a subject, we focus on financial markets which are attracting numerous researchers in various fields since they are of complex systems consisting of various types of heterogeneous agents. In the context of finance, the normal mixture hypothesis (or more generally “mixture of distribution hypothesis”) has been proposed as an alternative explanation for the description of return distribution of financial assets by several studies (Mandelbrot and Taylor 1967; Clark 1973; Tauchen and Pitts 1983; Richardson and Smith 1994). They have considered trading volume or the number of transactions (or quotations) as a proxy of the latent number of information arrivals. The proposed model may lead to better understanding of dynamics for return-volume relationship including volatility persistence (Lamoureux and Lastrapes 1990; Andersen 1996; Liesenfeld 1998; Watanabe 2000).
Cloud-based services can be used to provide various kinds of real-world applications. In fact, rich data on both mobility and activity of users have been accumulated by the cloud service providers. If secondary usage of the data collected in the cloud-based servers is permitted both legally and adequately and data analytics procedures for anonymized data can be developed, then we will be able to quantify risk and chance of our society based on rich data on human mobility and activity with high resolution.
The chapter aims to propose a method to detect change points of temporally evolutionary bipartite network regarding anonymity of personal information. Moreover, in order to confirm an ability of the proposed method empirically, the proposed method is applied to data of trading activity collected in a cloud-based service of the foreign exchange market and quantified trading activity of the foreign exchange market from a comprehensive point of view.
The network structure of various kinds of physical and social systems has attracted considerable research attention. A many-body system can be described as a network, and the nature of growing networks has been examined well (Albert and Barabási 2002; Miura et al. 2012). Power-law properties can be found in the growing networks, which are called complex networks. These properties are related to the growth of elements and preferential attachment (Albert and Barabási 2002).
A network consists of several nodes and links that connect nodes. In the literature on the physics of socioeconomic systems (Carbone et al. 2007), nodes are assumed to represent agents, goods, and computers, while links express the relationships between nodes (Milaković et al. 2010; Lämmer et al. 2006). The network structure is perceived in many cases through the conveyance of information, knowledge, and energy, among others.
In statistical physics, the number of combinations of possible configurations under given energy constraints is related to “entropy.” Entropy is a measure that quantifies the states of thermodynamic systems. In physical systems, entropy naturally increases because of the thermal fluctuations on elements. Boltzmann proposed that entropy S is computed from the possible number of ensembles g by \(S = \log g\). For a system that consists of two sub-systems whose respective entropies are S 1 and S 2, the total entropy S is calculated as the sum of one of two subsystems S 1 + S 2. This case is attributed to the possible number of ensembles g 1 g 2. Entropy in statistical physics is also related to the degree of complexity of a physical system. If the entropy is low (high), then the physical configuration is rarely (often) realized. Energy injection or work in an observed system may be assumed to represent rare situations. Shannon entropy is also used to measure the uncertainty of time series (Carbone and Stanley 2007).
The concept of statistical–physical entropy was applied by Bianconi (2009) to measure network structure. She considered that the complexity of a network is related to the number of possible configurations of nodes and links under some constraints determined by observations. She calculated the network entropy of an arbitrary network in several cases of constraints.
Researchers have used a methodology to characterize network structure with information-theoretic entropy (Dehmer and Mowshowitz 2011; Wilhelm and Hollunder 2007; Rashevsky 1955; Trucco 1956; Mowshowitz 1968; Sato 2009). Several graph invariants such as the number of vertices, vertex degree sequence, and extended degree sequences have been used in the construction of entropy-based measures (Wilhelm and Hollunder 2007; Sato 2009).
2 An Entropy Measure on a Bipartite Network
The number of elements in socioeconomic systems is usually very large, and several restrictions or finiteness of observations can be found. Therefore, we need to develop a method to infer or quantify the affairs of the entire network structure from partial observations. Specifically, many affiliation relationships of socioeconomic systems can be expressed as a bipartite network. Describing the network structure of complex systems that consist of two types of nodes by using the bipartite network is important. A bipartite graph model also can be used as a general model for complex networks (Guillaume and Latapy 2006; Chmiel et al. 2007; Tumminello et al. 2011). Tumminello et al. proposed a statistical method to validate the heterogeneity of bipartite networks (Tumminello et al. 2011).
Suppose a symmetric binary two-mode network can be constructed by linking K groups (A node) and M participants (B node) if the participants belong to groups (Fig. 1). Assume that we can count the number of participants in each group within the time window [tδ, (t + 1)δ] (t = 1, 2, 3, …), which is defined as m i(t) (i = 1, 2, …, K).
Let us assume a bipartite graph consisting of A nodes and B nodes, of which the structure at time t is described as an adjacency matrix C ij(t). We also assume that A nodes are observable and B nodes are unobservable. That is, we only know the number of participants (B node) belonging to A nodes m i(t). We do not know the correct number of B nodes, but we assume that it is M. In this setting, how do we measure the complexity of the bipartite graph from m i(t) at each observation time t?
The network entropy is defined as a logarithmic form of the number of possible configurations of a network under a constraint (Bianconi 2009). We can introduce the network entropy at time t as a measure to quantify the complexity of a bipartite network structure. The number of possible configurations under the constraint \(m_i(t)=\sum _{j=1}^MC_{ij}(t)\) may be counted as
Then, the network entropy is defined as \(\varSigma (t)=\ln N(t)\). Inserting Eq. (1) into this definition, we have
Note that because 0! = 1, \(\sum _{n=1}^{0} \ln n =0\). Obviously, if m i(t) = M for any i, then Σ(t) = 0. If m i(t) = 0 for any i, then Σ(t) = 0. The lower number of combinations gives a lower value of Σ(t). To eliminate a difference in the number of links, we consider the network entropy per link defined as
This quantity shows the degree of complexity of the bipartite network structure. We may capture the temporal development of the network structure from the value of σ(t). The network entropy per link σ(t) is also an approximation of the ratio of the entropy rate for m i(t) to its mean so that
where the entropy rate and the mean are, respectively, defined as
The ratio of the entropy rate to the mean tells us the uncertainty of the mean from a different point of view from the coefficient of variation (C.V. = standard deviation∕mean).
To understand the fundamental properties of Eq. (3), we compute σ(t) in simple cases. Consider values of entropy for several cases at K = 100 with different M. We assume that the total number of links is fixed at 100, which is the same as the number of A nodes, and we confirm the dependence of σ(t) on the degree of monopolization. We assign the same number of links at each A node. That is, we set
where k can be set as 1, 2, 4, 5, 10, 20, 50, or 100. In this case, we can calculate σ(t) as follows:
Figure 2a shows the relationship between σ(t) and the degree of monopolization at M = 1000, 2000, 3000, and 4000. The network entropy per link σ(t) is small if a small population of nodes occupies a large number of links. The multiplication regime gives a large value of σ(t). The value of σ(t) is a monotonically increasing function in terms of k. As M increases, the value of σ(t) increases. From this instance, we confirmed that σ(t) decreases with the degree of monopolization at A nodes.
Next, we confirm the dependency of σ(t) on the density of links. We assume that each element of an adjacency matrix C ij(t) is given by an i.i.d. Bernoulli random variable with a successful probability of p. Then, \(m_i(t)=\sum _{j=1}^MC_{ij}(t)\) is sampled from an i.i.d. binomial distribution Bin(p, M). In this case, one can approximate σ(t) as
Figure 2b shows the plots of σ(t) versus p obtained from both Monte Carlo simulation with random links drawn from Bernoulli trials and Eq. (9). The number of links at each A node monotonically increases as p increases. σ(t) decreases as the density of links decreases. The dependence of the entropy per link on p is independent of M.
3 Empirical Analysis
The application of network analysis to financial time series has been advancing. Several researchers have investigated the network structure of financial markets (Bonanno et al. 2003; Gworek et al. 2010; Podobnik et al. 2009; Iori et al. 2008). Bonanno et al. examined the topological characterization of the correlation-based minimum spanning tree (MST) of real data (Bonanno et al. 2003). Gworek et al. analyzed the exchange rate returns of 38 currencies (including gold) and computed the characteristic path length and average weighted clustering coefficient of the MST topology of the graph extracted from the cross-correlations for several base currencies (Gworek et al. 2010). Podobnik et al. (2009) examined the cross-correlations between volume changes and price changes for the New York Stock Exchange, Standard and Poor’s 500 index, and 28 worldwide financial indices. Iori et al. (2008) analyzed the network topology of the Italian segment of the European overnight money market and investigated the evolution of these banks’ connectivity structure over the maintenance period. These studies collectively aimed to detect the susceptibility of network structures to macroeconomic situations.
Data collected from the ICAP EBS platform were used. The data period spanned May 28, 2007 to November 30, 2012 (ICAP 2013). The data included records for orders (BID/OFFER) and transactions for currencies and precious metals with a 1-s resolution. The total number of orders included in the data set is 520,973,843, and the total number of transactions is 58,679,809. The data set involved 94 currency pairs consisting of 39 currencies, 11 precious metals, and 2 basket currencies (AUD, NZD, USD, CHF, JPY, EUR, CZK, DKK, GBP, HUF, ISK, NOK, PLN, SEK, SKK, ZAR, CAD, HKD, MXC, MXN, MXT, RUB, SGD, XAG, XAU, XPD, XPT, TRY, THB, RON, BKT, ILS, SAU, DLR, KES, KET, AED, BHD, KWD, SAR, EUQ, USQ, CNH, AUQ, GBQ, KZA, KZT, BAG, BAU, BKQ, LPD, and LPT).Footnote 1 Figure 3 shows a network representation of 94 currency pairs.
3.1 The Total Number
The number of quotations and transactions in each currency pair was extracted from the raw data. Let m X,i(t) (t = 0, …;i = 1, …, K) be the number of quotations (X = P) or transactions (X = D) within every minute (δ = 1 [min]) for a currency pair i (K = 94) at time t. Let c X(t) be denoted as the total number of quotations (X = P) and transactions (X = D), which is defined as
Let us consider the maximum value of c X(t) in each week:
where W(s) (s = 1, …, T) represents a set of times included in the s-th week. A total of 288 weeks are included in the data set (T = 288). Figure 4 shows the maximum values c X(t) for the period from May 28, 2007 to November 30, 2012.
According to the extreme value theorem, the probability density for maximum values can be assumed to be a Gumbel density:
where μ X and ρ X are the location and scale parameters, respectively. Under the assumption of the Gumbel density, these parameters are estimated with the maximum likelihood procedure. The log-likelihood function for T observations w X(s′) (s′ = 1, …, T) under Eq. (12) is defined as
The maximum likelihood estimators are obtained by maximizing the log-likelihood function. Partially differentiating l(μ X, ρ X) in terms of μ X and ρ X and setting them to zero, one has its maximum likelihood estimators as
The parameters are estimated as \(\hat {\mu }_P = 772.179499\), \(\hat {\rho }_P = 281.741815\), \(\hat {\mu }_D = 206.454884\), and \(\hat {\rho }_D = 35.984804\).
The Kolmogorov–Smirnov (KS) test is conducted to determine the statistical significance of the estimated distributions. The KS test is a popular statistical method of assessing the difference between observations and its assumed distribution by p-value, which is a measure of probability where a difference between the two distributions happens by chance. Large p-values imply that the observations are sampled from the assumed distribution in the null hypothesis with high significance. Let w X(s) (s = 1, …, T) be T observations, and let K T be a test statistic
where 0 ≤ F(v) ≤ 1 is an assumed cumulative distribution in a null hypothesis and F T(v) an empirical one based on T observations such that F T(v) = k∕T, in which k represents the number of observations satisfying v X(s) ≤ v(s = 1, …, T). The p-value is computed from the Kolmogorov–Smirnov distribution.
The KS test is conducted under the assumption of the Gumbel distribution for the maximum value corresponding to Eq. (19):
The p-values of the KS test are shown in Table 1. The stationary Gumbel assumption cannot explain the maximum values for quotes with a 5% significance level in the KS test. The stationary Gumbel assumption may not be accepted in the case of the block maximum number of quotes. The dominant reason is the strong nonstationarity of the maximum number of quotes. During the last 5 years, the currencies and pairs quoted in the electronic brokerage market increased. The mean value of the total number constantly increased. In fact, the maximum number of quotations w P(t) reached the maximum value on 30 July, 2012. The nonstationarity breaks the assumption of the extreme value theorem.
It is confirmed that the stationary Gumbel assumption can be accepted for the block maxima of transactions in each week using the KS test with a 5% significance level. The maximum number of transactions w D(t) was reached on on January 30, 2012. This period seems to be related to the extreme synchrony.
3.2 Network Entropy Per Link
The proposed method based on statistical–physical entropy is applied to measure the states of the foreign exchange market. The relationship between a bipartite network structure and macroeconomic shocks or crises was investigated, and the occurrence probabilities of extreme synchrony were inferred. We compute a statistical–physical entropy per link from m X,i(t)(X ∈{P, D}) with Eqs. (2) and (3), which are denoted as σ X(t). σ P(t) and σ D(t).
Since small values of σ X(t) correspond to a concentration of links at a few nodes or a dense network structure, let us consider the minimum value of σ X(t) every week:
where W(s) (s = 1, …, T) represents a set of times included in the s-th week. A total of 288 weeks are included in the data set (T = 288). According to the extreme value theorem, the probability density for minimum values can be assumed to be the Gumbel density:
where μ X and ρ X are the location and scale parameters, respectively. Under the assumption of the Gumbel density, these parameters are estimated with the maximum likelihood procedure. The log-likelihood function for T observations v X(s′) (s′ = 1, …, T) under Eq. (19) is defined as
Partially differentiating l(μ X, ρ X) in terms of μ X and ρ X and setting them to zero yields its maximum likelihood estimators as
The parameter estimates are computed as \(\hat {\mu }_P=-4.865382\), \(\hat {\rho }_P=0.110136\), \(\hat {\mu }_D=-5.010175\), and \(\hat {\rho }_D=0.120809\) with Eqs. (21) and (22).
The KS test is conducted for the Gumbel distribution for the minimum values corresponding to Eq. (19):
The p-value of the distribution is shown in Table 2. The stationary Gumbel assumption cannot explain the synchronizations observed in both quotes and transactions completely with a 5% significance level. The stationary Gumbel assumption is rejected because there is a stationary assumption to derive the extreme value distribution. If we can weaken this assumption, then the goodness of fit may be improved.
4 Probability of Extreme Synchrony
The literature detecting structural breaks or change points in an economic time series (Goldfeld and Quandt 1973; Preis et al. 2011; Scalas 2007; Cheong et al. 2012) points out that nonstationary time series are constructed from locally stationary segments sampled from different distributions. Goldfeld and Quandt conducted a pioneering work on the separation of stationary segments (Goldfeld and Quandt 1973). Recently, a hierarchical segmentation procedure was also proposed by Choeng et al. under the Gaussian assumption (Cheong et al. 2012). We applied this concept to define the segments for v X(s′) (s′ = 1, …, T).
Let us consider the null model L 1, which assumes that all the observations v X(s′) (s′ = 1, …, T) are sampled from a stationary Gumbel density parameterized as μ and ρ. An alternative model L 2(s) assumes that the left observations v X(s′) (s′ = 1, …, s) are sampled from a stationary Gumbel density parameterized as μ L and ρ L and that the right observations v X(s′) (s′ = s + 1, …, T) are sampled from a stationary Gumbel density parameterized as μ R and ρ R.
Denoting likelihood functions as
the difference between the log-likelihood functions can be defined as
Δ(s) can be approximated as the Shannon entropy \(H[p]=-\int _{-\infty }^{\infty }\mbox{d}v \log p(v)p(v)\):
Since the Shannon entropy of the Gumbel density expressed in Eq. (12) is calculated as
where γ E represents Euler’s constant, defined as
we obtain
In the context of model selection, several information criteria are proposed. The information criterion provides both goodness of fit of the model to the data and model complexity. For the sake of simplicity, we use the Akaike information criterion (AIC) to determine the adequate model. The AIC for a model with the number of parameters K and the maximum likelihood of L is defined as
We can compute the difference in AIC between model L 2 and model L 1(s) as
since the number of parameters of L 1 is 2, that of L 2(s) is 4, and the maximum likelihood is obtained by using their maximum likelihood estimators calculated from
Therefore, P(v X;μ L, ρ L) is maximally different from P(v X;μ R, ρ R) when Δ(s) assumes a maximal value. This spectrum has a maximum at some time s ∗, which is denoted as
The segmentation can be used recursively to separate the time series into further smaller segments. We do this iteratively until all segment boundaries have converged onto their optimal segment, defined by a stopping (termination) condition.
Several termination conditions were discussed in previous studies (Cheong et al. 2012). Assuming that Δ 0 > 0, we terminate the iteration if \(\varDelta _{AIC}^{*}\) is less than a typical conservative threshold of Δ 0 = 10, while the procedure is recursively conducted if \(\varDelta _{AIC}^*\) is larger than Δ 0. We checked the robustness of this segmentation procedure for Δ 0. Δ 0 gives a statistical significance level of termination. The value of Δ 0 is related to statistical significance. According to Wilks theorem, − 2Δ(s) follows a chi-squared distribution with a degree of freedom r, where r is given by the difference between the number of parameters assumed in the null hypothesis and one in the alternative hypothesis. In this case, r = 2. Hence, the cumulative distribution function of \(\varDelta ^*_{AIC}\) may follow
where γ(x, a) is the regularized incomplete gamma function defined as
Therefore, setting the threshold Δ 0 = 10 implies that the segmentation procedure is tuned as a 4.928% statistical significance level.
Let the number of segments be L X, the parameter estimates be {μ X,j, ρ X,j} at the j-th segment, and the length of the j-th segment be τ X,j, where \(\sum _{j=1}^{R_X}\tau _{X,j}=T\). The cumulative probability distribution for v X(s) (s = 1, …, T) may be assumed to be a finite mixture of Gumbel distributions:
Tables 3 and 4 show parameter estimates of v P(s) and v D(s) using the recursive segmentation procedure. Figure 5 shows the temporal development of v P(s) and v D(s) from May 28, 2007 to November 30, 2012. R P = 6 and R D = 6 are obtained from v X(s) using the proposed segmentation procedure. During the observation period, the global financial system suffered from the following significant macroeconomic shocks and crises: (I) the BNP Paribas shock (August 2007), (II) the Bear Stearns shock (February 2008), (III) the Lehman shock (September 2008 to March 2009), (IV) the European sovereign debt crisis (April to May 2010), (V) the East Japan tsunami (March 2011), (VI) the United States debt-ceiling crisis (May 2011), and (VII) the Bank of Japan’s 10 trillion JPY gift on Valentine’s Day (February 2012).
Before entering these global affairs, both v P(s) and v D(s) took large values. Note that, during the (I) Paribas shock, the (II) Bear Stearns and the (III) Lehman shock v P(s) and v D(s) took smaller values than they did during the previous term. This implies that a global shock may drive many participants and that these participants may trade the same currencies at the same time. The smallest values v P(s) and v D(t) correspond to the days of the (II) Bear Stearns shock, the (III) Lehman shock, and the (VI) Euro crisis. These days are generally related to the start or the end of macroeconomic shocks or crises. The period from December 2011 to March 2012 shows that the values of v D(s) are smaller than they were during other periods. This result implies that, during period, singular patterns appeared in the transactions.
Figure 6 shows both the empirical and estimated cumulative distribution functions of v P(s) and v D(s). The estimated cumulative distributions are drawn from Eq. (39) with parameter estimates. The KS test verifies this mixing assumption. The distribution estimated by the finite mixture of Gumbel distributions for quotes is well fitted, as shown in Table 5. From the p-values, the mixture of Gumbel distributions for quotations accepts the null hypothesis that v P(s) is sampled from the mixing distribution with a 5% significance level. The mixture of Gumbel distributions for transactions also accepts the null hypothesis that v D(s) is sampled from the mixing distribution with a 5% significance level. Extrapolation of cumulative distribution function also provides a guideline of the future probability of extreme events. The finite mixture Gumbel distributions with parameter estimates may be used as an inference of probable extreme synchrony.
5 Conclusion
A method based on the concept of “entropy” in statistical physics was proposed to quantify states of a bipartite network under constraints. The statistical–physical network entropy of a bipartite network was derived under the constraints for the number of links at each group node. Both numerical and theoretical calculations for a binary bipartite graph with random links showed that the network entropy per link can capture both the density and the concentration of links in the bipartite network. The proposed method was applied to measure the structure of bipartite networks consisting of currency pairs and participants in the foreign exchange market.
An empirical investigation of the total number of quotes and transactions was conducted. The nonstationarity of the number of quotes and transactions strongly affected the extreme value distributions. The empirical investigation confirmed that the entropy per link decreased before and after the latest global shocks that have influenced the world economy. A method was proposed to determine segments with recursive segmentation based on the Akaike information criterion between Gumbel distributions with different parameters. Under the assumption of a finite mixture of Gumbel distributions, the estimated distributions were verified by the Kolmogorov–Smirnov test. The finite mixture of Gumbel distributions can estimate the occurrence probabilities of extreme synchrony of a nonstationary system extracted as a bipartite network. The extrapolation of the extreme synchrony can be done based on the estimated mixture of Gumbel distributions.
Notes
- 1.
AED, United Arab Emirates dirham; AUD, Australian dollar; AUQ, Australian dollar (small amount); BAG, gold (bank); BAU, silver (bank); BHD, Bahraini dinar; BKT, basket of USD/EUR; BKQ, basket of USD/EUR (small amount); CAD, Canadian dollar; CHF, Swiss franc; CNH, Chinese yuan; CZK, Czech koruna; DKK, Danish krone; EUR, EU euro; EUQ, EU euro (small amount); GBP, British sterling; GBQ, British sterling (small amount); HKD, Hong Kong dollar; HUF, Hungarian forint; ILS, Israeli new shekel; ISK, Iceland krona; JPY, Japanese yen; KES, Kenyan shilling; KET, Kenyan shilling (small amount); KZA, Kazakhstani tenge (small amount); KZT, Kazakhstani tenge; LPD, palladium (London); LPT, platium (London); MXN, Mexican peso; MXQ, Mexican peso (small amount); MXT, Mexican peso (special deals); NOK, Norwegian krone; NZD, New Zealand dollar; PLN, Poland zloty; RON, Romanian leu; RUB, Russian ruble; SAR, Saudi Arabian riyal; SGD, Singapore dollar; SEK, Swedish krona; SKK, Slovak koruna; SAU, silver (small amount); TRY, Turkish lira; THB, Thai baht; USD/DLR, US dollar; USQ, US dollar (small amount); ZAR, South African rand; XAU, gold; SAU, gold (small amount); XAG, silver; XPD, palladium; and XPT, platinum.
References
Albert R, Barabási A-L (2002) Statistical mechanics of complex networks. Rev Mod Phys 74:47–97
Andersen TG (1996) Return volatility and trading volume: an information flow interpretation of stochastic volatility. J Financ 51:169–204
Bianconi G (2009) Entropy of network ensembles. Phys Rev E 79:036114; Anand K, Bianconi G (2009) Entropy measures for networks: toward an information theory of complex topologies. Phys Rev E 80:045102
Bonanno G, Caldarelli G, Lillo F, Mantegna RN (2003) Topology of correlation-based minimal spanning trees in real and model markets. Phys Rev E 68:046130
Carbone A, Stanley HE (2007) Scaling properties and entropy of long-range correlated time series. Physica A 304:21–24
Carbone A, Kaniadakis G, Scarfone AM (2007) Eur Phys J B 57:121
Cheong SA, Fornia RP, Lee GHT, Kok JL, Yim WS, Xu DY, Zhang Y (2012) The Japanese economy in crises: a time series segmentation study. Econ E-J 2012-5. http://www.economics-ejournal.org
Chmiel AM, Sienkiewicz J, Suchecki K, Hołyst JA (2007) Networks of companies and branches in Poland. Physica A 383:134–138
Clark P (1973) A subordinated stochastic process model with finite variance for speculative prices. Econometrica 41:135–155
Dehmer M, Mowshowitz A (2011) A history of graph entropy measures. Inf Sci 181:57–78
de Menezes MA, Barabási A-L, Fluctuations in network dynamics. Phys Rev Lett 92 (2004) 028701.
Goldfeld SM, Quandt RE (1973) A Markov model for switching regressions. J Econometrics 1:3–15
Guillaume J-L, Latapy M (2006) Bipartite graphs as models of complex networks. Physica A 371:795–813
Gworek S, Kwapień J, Drożdż S (2010) Sign and amplitude representation of the forex networks. Acta Phys Pol A 117:681–687
ICAP (2013). The data is purchased from ICAP EBS: http://www.icap.com
Iori G, Masi GD, Precup OV, Gabbi G, Caldarelli G (2008) A network analysis of the Italian overnight money market. J Econ Dyn Control 32:259–278
Lambiotte R, Ausloos M, Thelwall M (2007) Word statistics in Blogs and RSS feeds: towards empirical universal evidence. J Informetrics 1:277–286
Lämmer S, Gehlsen B, Helbing D (2006) Scaling laws in the spatial structure of urban road networks. Physica A 363:89–95
Lamoureux CG, Lastrapes WD (1990) Heteroskedasticity in stock return data: volume versus GARCH effects. J Financ 45:221–229
Liesenfeld R (1998) Dynamic bivariate mixture models: modeling the behavior of prices and trading volume. J Bus Econ Stat 16:101–109
Mandelbrot BB, Taylor H (1967) On the distribution of stock price differences. Oper Res 15:1057–1062
Milaković M, Alfrano S, Lux T (2010) The small core of the German corporate board network. Comput Math Organ Theory 16:201–215
Miura W, Takayasu H, Takayasu M (2012) Effect of coagulation of nodes in an evolving complex network. Phys Rev Lett 108:168701
Mowshowitz A (1968) Entropy and the complexity of graphs: I. An index of the relative complexity of a graph. Bull Math Biophys 30:175–204
Podobnik B, Horvatic D, Petersen AM, Stanley HE (2009) Cross-correlations between volume change and price change. Proc Natl Acad Sci U S A 106:22079–22084
Preis T, Schneider JJ, Stanley HE (2011) Switching processes in financial markets. Proc Natl Acad Sci U S A 108:7674–7678
Rashevsky N (1955) Life, information theory, and topology. Bull Math Biophys 17:229–235
Richardson M, Smith T (1994) A direct test of the mixture of distributions hypothesis: measuring the daily flow of information. J Financ Quant Anal 29:101–116
Sato A-H (2007) Frequency analysis of tick quotes on the foreign exchange market and agent-based modeling: a spectral distance approach. Physica A 382:258–270
Sato A-H (2010) Comprehensive analysis of information transmission among agents: similarity and heterogeneity of collective behavior. In: Chen S-H et al (eds) Agent-based in economic and social systems VI: post-proceedings of the AESCS international workshop 2009, agent-based social systems, vol 8. Springer, Tokyo, pp 1–17
Sato A-H (2017) Inference of extreme synchrony with an entropy measure on a bipartite network. In: 2017 IEEE 41st annual computer software and applications conference (COMPSAC), pp. 766–771
Sato A-H, Hołyst JA (2008) Characteristic periodicities of collective behavior at the foreign exchange market. Eur Phys J B 62:373–380
Scalas E (2007) Mixtures of compound Poisson processes as models of tick-by-tick financial data. Chaos, Solitons Fractals 34:33–40
Tauchen T, Pitts M (1983) The price variability-volume relationship on speculative markets. Econometrica 51:485–505
Trucco E (1956) A note on the information content of graphs. Bull Math Biophys 18:129–135
Tumminello M, Miccichè S, Lillo F, Pillo J, Mantegna RN (2011) Statistically validated networks in bipartite complex systems. PLoS One 6:e17994
Watanabe T (2000) A nonlinear filtering approach to stochastic volatility models with an application to daily stock returns. J Bus Econ Stat 18:199–210
Wilhelm T, Hollunder J (2007) Information theoretic description of networks. Physica A 385:385–396
Acknowledgements
This work is supported by Japan Science and Technology Agency (JST) PRESTO Grant Number JPMJPR1504, Japan. The work was also financially supported by the Grant-in-Aid for Young Scientists (B) by the Japan Society for the Promotion of Science (JSPS) KAKENHI (#23760074).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this chapter
Cite this chapter
Sato, AH. (2019). On Measuring Extreme Synchrony with Network Entropy of Bipartite Graphs. In: Sato, AH. (eds) Applications of Data-Centric Science to Social Design. Agent-Based Social Systems, vol 14. Springer, Singapore. https://doi.org/10.1007/978-981-10-7194-2_16
Download citation
DOI: https://doi.org/10.1007/978-981-10-7194-2_16
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-7193-5
Online ISBN: 978-981-10-7194-2
eBook Packages: Business and ManagementBusiness and Management (R0)