Abstract
Understanding the brain function requires investigating information transfer across brain regions. Shannon began the remarkable new field of information theory in 1948. It basically can be divided into two categories: directed and undirected information-theoretical approaches. As we all know, neural signals are typically nonlinear and directed flow between brain regions. We can use directed information to quantify feed-forward information flow, feedback information, and instantaneous influence in the high-level visual cortex. Moreover, neural signals have bidirectional information flow properties and are not captured by the transfer entropy approach. Therefore, we used directed information to quantify bidirectional information flow in this study. We found that there has information flow between the scene-selective areas, e.g., OPA, PPA, RSC, and object-selective areas, e.g., LOC. Specifically, strong information flow exists between RSC and LOC. It explained that functionally coupled between RSC and LOC plays a vital role in visual scenes/object categories or recognition in our daily lives. Meanwhile, we also found weak reverse-directed information flow in the visual scenes and objects neural networks.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Understanding how information flow in the brain conducts a specific cognitive task is a major scientific challenge. In the field of neuroscience, studying brain function entails learning about how the brain processes information. Researchers can figure out how information flows through different parts of the brain to gain such knowledge [8, 9, 12]. In most cases, information flow in the brain is random and directional. Granger causality [1, 14], or Transfer Entropy [4, 19], can be used to calculate directional information flow. However, the methods described above can only calculate one-way information flow from one area to another rather than bidirectional information and instantaneous information flow between functional brain regions. Directed information was designed to characterize channels with feedback but the way that we applied it to neuroscience to solve the aforementioned problems [10]. It’s mainly a tool for inference, some causality inference, and more along the lines of Granger causality in spirit [14]. Granger causality says that X causes Y because time series X causes a time series Y if you can predict Y if you also condition on the observations of X. Directed information encapsulates this by prediction reduction attribute so a decrease in randomness. Comparison of other information-theoretic methods applied in neuroscience, DI not only quantify measure feed-forward information but also feedback information. The brain is a complex system, and each region is densely connected to infer specific cognitive tasks. Therefore, directed information can help us understand how neural information flows among brain regions. This study uses a directed information method to measure information communication among visual scene category neural networks with fMRI dataset.
According to fMRI studies, some scene-selective regions in the human visual cortex have been discovered and linked to higher-order functions like scene perception or category, such as Primary Visual Cortex (V1), Fusiform Face Area (FFA) [7], Occipital Face Area (OFA), Occipital Place Area (OPA), Parahippocampal Place Area (PPA) [3], and retrosplenial cortex (RSC) [13]. Although all these regions respond well to scenes and objects, less research has been done on how these regions communicate with one another during the experience of natural scenes, specific bidirected information flow among these regions. That’s also our primary motivation in this study.
2 Methods
2.1 Definition
Shannon Entropy. Assuming a random variable X, which can get a value like x as probability \(\mathrm {P}(\mathrm {X}=\mathrm {x})\), entropy of this variable can be expressed as:
Conditional Entropy. The conditional entropy of X given Y is the average uncertainty that remains about x when y is known:
Mutual Information. Given two random variables X and Y, the mutual information can be calculate as the difference of sum of individual entropy and the entropy of the variables considered jointly as a single system. It can be mathematically formula expressed as:
Conditional Mutual Information. The conditional mutual information of X and Y given Z is the uncertainty that remains about x and y when z is known:
Granger Causality. The Granger causality (GC) idea firstly proposed by Granger in 1969 [5]. The basic idea is, if two signal \(\mathbf {X}\) and \(\mathbf {Y}\) have caucal relationship, instead of history value of \(\mathbf {Y}\), then \(\mathbf {Y}\) also can be predicted given \(\mathbf {X}\) information. Assuming \(\mathbf {X}^{n}=\left[ X_{1}, X_{2}, \ldots , X_{n}\right] \) and \(\mathbf {Y}^{n}=\left[ Y_{1}, Y_{2}, \ldots , Y_{n}\right] \) are two continue time series. The GC analysis can be expressed as a auto-regressive or line prediction model as follows:
where \(e_{i}\) indicate error of prediction \(Y_{i}\) given only past valye of Y, (\(Y_{i-1}, ..., Y_{i-P}\)), and \(\tilde{e}_{i}\) is the error of prediction \(Y_{i}\) given both history value of Y (\(Y_{i-1}, ..., Y_{i-P}\)) and previous value of X (\(X_{i-1}, ..., X_{i-P}\)). Based on GC properties above described, GC analysis gradually applied in the neuroscience disincline.
Transfer Entropy. Another widely applied causal measurement in neuroscience is Transfer Entropy (TE) [17]. How the prior knowledge affects the next state or predicts future state can use TE to address this question. TE can be defined as:
The TE can be used measured the directed information flow from Y \(\rightarrow \) X or X \(\rightarrow \) Y. The basic theory of TE can be shown graphically in Fig. 1. In neuroscience studies, it is very used for defined the directed causality effects between neural signals. However, the real neural activity not just single directed information flow. Neurons can use resonance at the same time, that means X \(\leftrightarrow \) Y. The pitfall of TE is that it cannot measure the bi-directed information flow at the same time. A comprehensive review on TE estimate directed information flow could be found in [16].
2.2 Directed Information
In this section, we are going to describe directed information from mathematics view. Assuming uppercase letters X and Y denoted random variables, and denote n-tuple \(\left( X_{1}, X_{2}, \ldots , X_{n}\right) \) as \(X^{n}\). The information flow from \(X^{n}\) to \(Y^{n}\) can be formula as,
Where \(H\left( Y^{n}\right) -H\left( Y^{n} \Vert X^{n}\right) \) is causally conditional entropy [15], and it can be defined as,
Comparison of mutual information,
The condition entropy instead of causally conditional entropy in the directed information. Meanwhile, directed information is not symmetric, e.g., \(I\left( Y^{n} \rightarrow X^{n}\right) \ne I\left( X^{n} \rightarrow Y^{n}\right) \) in general. On the contrary, reverse information flow can be defined as,
It has a number of significant properties, some can be found in [1, 15]. For the sake of brevity, we’ll just reveal two enlightening conservation rules. Based on Massey and Massey [11] the conservation law,
Equation 10 is particularly enlightening in settings where \(X_{i}\) and \(Y_{i}\) appear alternately, as shown in Fig. 2.
In some cases, \(X^{n}\), \(Y^{n}\) may happen simultaneously, such as neural network in the brain. The following is another conservation law stated in [1] which, in such situations, may be more insightful than that in Eq. 10. The instantaneous influence can be calculated through directed information and reverse directed information as shown in Fig. 3(c),
If \(\sum _{i=1}^{n} I\left( X_{i} ; Y_{i} \mid X^{i-1}, Y^{i-1}\right) =0\). That is, they do not have an instantaneous effect on one another. In this study, the Context-Tree Weighting (CTW) algorithm proposed by Willims [18] was used to estimate DI [6] flow among visual scene neural networks and its powerful algorithms to compress data.
2.3 FMRI Dataset
The public BOLD5000 datasetFootnote 1 [2] used in this study when we estimated bidirected information flow among visual scene neural networks (see Fig. 4). The fMRI experiments used a dataset obtained from 4 subjects (aged 24 to 27 years) with normal or corrected-to-normal vision, who each viewed 5254 images over 15 scanning sessions. The stimuli images were selected from three classical computer vision datasets. They are SUN dataset (1000 Scene Images)Footnote 2, COCO datasetFootnote 3, and ImageNet dataset (1916 images)Footnote 4, respectively.
3 Result
In this section, we experimentally compared correlation, mutual information, transfer entropy, and directed information for quantifying information flow in the visual scenes neural networks. In Fig. 5, we have shown functional connectivity among visual scenes neural networks with correlation, mutual information, and transfer entropy approaches. We found that mutual information can capture more information than correlation, but both methods do not quantify information flow direction. In Fig. 6, the graphs depict the functional connectivity between visual scenes neural networks, and the strength of edges color represents connectivity weights in which consistence of functional connectivity matrix in Fig. 5. We found that information flow between the scene-selective areas, e.g., OPA, PPA, RSC, and object-selective areas, e.g., LOC, plays an important role in visual scene/object categories or recognitions. Nevertheless, we are interested in whether information feedback and resonance information flow in the high-level visual cortex. Therefore, in Fig. 7 and Fig. 8, we found consistent results with correlation and mutual information methods. However, we also found some unknown results that are resonance information exists in the high-level visual cortex.
4 Discussion and Conclusions
This paper begins with an information-theoretic perspective and quantifies information flow in the high-level visual neural networks. It collects directed and reverse directed information, mutual information, and resonance information between the brain’s left and right regions of interest. It opens up a new avenue for us to investigate what happens when pairwise neural signal entanglement occurs. It means a lot in understanding neural signal flow in the brain. However, there are some limitations in which we should point out in the following contents.
First, we got directed information through CTW estimator, and it needs to satisfy that input data should be binary value. In other words, we need to convert the BOLD signal into a binary value that means we will lose some information when we estimate functional connectivity between ROIs. Therefore accuracy estimate binary BOLD signal is a crucial problem when we are going to quantify directed information flow. Second, we directly used in visual scenes/objects selective-ROIs in which defined via t-statistics. Considered effect size and functional overlap problems, the estimated information flow through information-theoretical methods are not accurate. In the following study, on the one hand, we need to consider how to avoid or solve the problems mentioned above. On the other hand, we can reconstruct nature images from the BOLD signal to confirm the information flow in the high-level visual cortex.
Nevertheless, we still found some interesting results through estimated directed information. First, we found that there has information flow within the scene-selective areas, e.g., OPA, PPA, and RSC. Second, we also found that information flow between the scene-selective areas, e.g., RSC, and the object-selective areas, e.g., LOC. Third, we found that there has weak reverse-directed information flow in the high-level visual cortex.
5 Code Availability
The code used to reproduce result can be available under the request author.
References
Amblard, P., Michel, O.: Relating granger causality to directed information theory for networks of stochastic processes. IEEE Trans. Inform. Theory 53 (2009)
Chang, N., Pyles, J., Marcus, A., Mulam, H., Tarr, M., Aminoff, E.: BOLD5000, a public fMRI dataset while viewing 5000 visual images. Sci. Data 6 (2019). https://doi.org/10.1038/s41597-019-0052-3
Epstein, R., Kanwisher, N.: The parahippocampal place area: a cortical representation of the local visual environment. J. Cogn. Neurosci. 7 (1998). https://doi.org/10.1016/S1053-8119(18)31174-1
Gencaga, D.: Transfer entropy. Entropy 20, 288 (2018). https://doi.org/10.3390/e20040288
Granger, C.: Investigating causal relations by econometric models and cross-spectral methods. Econometrica 37, 424–38 (1969). https://doi.org/10.2307/1912791
Jiao, J., Permuter, H.H., Zhao, L., Kim, Y., Weissman, T.: Universal estimation of directed information. CoRR abs/1201.2334 (2012). http://arxiv.org/abs/1201.2334
Kanwisher, N., Mcdermott, J., Chun, M.: The fusiform face area: a module in human extrastriate cortex specialized for face perception. J. Neurosci. Official J. Soc. Neurosci. 17, 4302–11 (1997). https://doi.org/10.3410/f.717989828.793472998
Mahmud, M., Kaiser, M.S., Hussain, A., Vassanelli, S.: Applications of deep learning and reinforcement learning to biological data. IEEE Trans. Neural Netw. Learn. Syst. PP (2017). https://doi.org/10.1109/TNNLS.2018.2790388
Mahmud, M., Kaiser, M.S., Mcginnity, T., Hussain, A.: Deep learning in mining biological data. Cogn. Comput. 13 (2021). https://doi.org/10.1007/s12559-020-09773-x
Massey, J.: Causality, feedback and directed information (1990)
Massey, J., Massey, P.: Conservation of mutual and directed information, pp. 157–158 (2005). https://doi.org/10.1109/ISIT.2005.1523313
Noor, M., Zenia, N.Z., Kaiser, M.S., Al Mamun, S., Mahmud, M.: Application of deep learning in detecting neurological disorders from magnetic resonance images: a survey on the detection of Alzheimer’s disease, Parkinson’s disease and schizophrenia. Brain Inform. 7, 11 (2020). https://doi.org/10.1186/s40708-020-00112-2
Park, S., Chun, M.: Different roles of the parahippocampal place area (PPA) and retrosplenial cortex (RSC) in panoramic scene perception. NeuroImage 47, 1747–56 (2009). https://doi.org/10.1016/j.neuroimage.2009.04.058
Tank, A., Covert, I., Foti, N., Shojaie, A., Fox, E.: Neural granger causality. IEEE Trans. Pattern Anal. Mach. Intell. PP, 1–1 (2021). https://doi.org/10.1109/TPAMI.2021.3065601
Tatikonda, S., Mitter, S.: The capacity of channels with feedback. IEEE Trans. Inf. Theory 55, 323–349 (2009). https://doi.org/10.1109/TIT.2008.2008147
Vicente, R., Wibral, M., Lindner, M., Pipa, G.: Transfer entropy—a model-free measure of effective connectivity for the neurosciences. J. Comput. Neurosci. 30, 45–67 (2011). https://doi.org/10.1007/s10827-010-0262-3
Wibral, M., Vicente, R., Lindner, M.: Transfer entropy in neuroscience. Underst. Complex Syst. 3–36 (2014). https://doi.org/10.1007/978-3-642-54474-3-1
Willems, F.M.J., Shtarkov, Y.M., Tjalkens, T.J.: The context-tree weighting method: basic properties. IEEE Trans. Inf. Theory 41(3), 653–664 (1995). https://doi.org/10.1109/18.382012
Wollstadt, P., Martínez Zarzuela, M., Vicente, R., Díaz-Pernas, F., Wibral, M.: Efficient transfer entropy analysis of non-stationary neural time series. PLoS ONE 9 (2014). https://doi.org/10.1371/journal.pone.0102833
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this paper
Cite this paper
Li, Q. (2021). Bidirected Information Flow in the High-Level Visual Cortex. In: Mahmud, M., Kaiser, M.S., Vassanelli, S., Dai, Q., Zhong, N. (eds) Brain Informatics. BI 2021. Lecture Notes in Computer Science(), vol 12960. Springer, Cham. https://doi.org/10.1007/978-3-030-86993-9_6
Download citation
DOI: https://doi.org/10.1007/978-3-030-86993-9_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-86992-2
Online ISBN: 978-3-030-86993-9
eBook Packages: Computer ScienceComputer Science (R0)