1 Introduction

The last few decades have seen rapid developments of first-principles-based mathematical models to study living systems. Based on a collection of a priori physiological and physical principles, the evolution of mathematical models can offer significant advantages in understanding, reproducing, and predicting complex biological phenomena. However, first-principle-based models can be prohibitively expensive to build due to the large number of parameters and variables characterizing the complexity of biological details, e.g., multiple time scales, complicated interactions between biological elements, among others. Alternatively, modern data-driven models focusing on phenomenological or empirical observations are gaining ground in mathematical biology, in that they are designed to deal with high dimensional and noisy data (Janes and Yaffe 2006; Hasenauer 2015; Solle 2017; Jack et al. 2018; AlQuraishi and Sorger 2021). However, one is still faced with the daunting task of making sense of the coordinates and parameters of the data-driven models to identify interpretable and biologically meaningful features.

In this study, we investigate how the combination of the two classes of methods can be used to study spiking neuronal networks (SNNs). SNNs are capable of producing highly nonlinear, high-dimensional, and multi-timescale dynamics, and have been widely used to investigate cognitive functions and their computation principles (see, e.g., Tao 2006; Ghosh-Dastidar and Adeli 2009; Ponulak and Kasinski 2011; Nobukawa et al. 2017; Börgers and Kopell 2003; Chariker et al. 2016). First-principles-based model reduction methods such as coarse-graining and mean-field theories have been developed to better understand SNN dynamics (Wilson and Cowan 1972; Brunel and Hakim 1999; Buice and Cowan 2007; Cai 2006; Cai et al. 2021; Li and Hui 2019). On the other hand, artificial neural networks (ANN, and its offspring, deep neural networks, or DNN) are a modern data-driven method inspired by the nervous system. DNN has been extremely successful in both engineering applications (image processing, natural language processing, etc.) and applied mathematics (parameter estimation, numerical ordinary/partial differential equations, inverse problems, etc.). See Aggarwal et al. (2018); Schmidhuber (2015); Chon and Cohen (1997); Li (2020); Raissi et al. (2019), for instance. In particular, it has been shown recently that DNN can well approximate both finite and infinite-dimensional operators (Barron 1994; Kovachki et al. 2021). The idea of using DNN surrogates in models to replace the firing rate of SNN was first explored by Zhang and Young (2020). This motivates us to propose a first-principle-informed deep-learning framework that replaces the crucial and complex SNN dynamics by the representation of artificial neurons.

The neuroscience problem we address in this paper is the \(\gamma \)-band oscillations, a type of 30–90 Hz oscillatory physiological dynamics prominent in many brain regions (Andrew Henrie and Shapley 2005; Brosch et al. 2002; Bauer 2006; Buschman and Miller 2007; Pieter Medendorp 2007; van Wingerden 2010; Csicsvari 2003; Popescu et al. 2009; Van Der Meer and David Redish 2009). Remarkably, in previous studies, \(\gamma \)-oscillations can be produced in simple, idealized SNN models involving only two neural populations, excitatory (E) and inhibitory (I) (Chariker et al. 2018; Zhang 2014; Rangan and Young 2013; Li et al. 2019). More specifically, due to transient noise and/or external stimulus, highly correlated spiking patterns (previously termed multiple-firing events, or MFEs) are repeatedly produced from the competitions between E/I populations, involving the interplay between multiple timescales. MFEs are a type of stochastic, high-dimensional emergent phenomena, with rapid and transient dynamical features that are sensitive to the biophysical parameters of the network. The strong fluctuation from input and recurrent neuronal interactions hinders the ability of previous mean-field approaches to trace such transitions. Therefore, it is a very challenging task to build model reductions that can provide biological insights for \(\gamma \)-oscillations in a wide range of parameter regimes.

This paper explores learning the complex \(\gamma \)-oscillations with first-principle-informed DNNs. Our previous study revealed that the complex \(\gamma \)-oscillatory dynamics can be captured by a Poincare mapping F projecting the network state at one initiation of an MFE to the next initiation (Cai et al. 2021). Therefore, F is a high-dimensional mapping subjected to biophysical parameters of SNNs and is thus very hard to analyze. Some early theoretical work about sampling MFEs from integrate-and-fire neuronal networks with all-to-all connection are available in Zhang (2014), Zhang (2014), Zhang and Rangan (2015). However, to pursue analytical results, the scopes of studies were limited to very specialized situations. Instead of continue working on model-specific MFE sampling method, we approximate F by A. using coarse-graining (CG) and discrete cosine transform (DCT) to reduce the dimensionality of the state space, and B. benefiting from the representation power of DNNs. Specifically, DNNs provide a unified data-driven framework for varying SNN model parameters, revealing the potential of generalization to different dynamical regimes of the emergent network oscillations. Despite the significant underlying noise and the drastic dimensional reductions, our DNNs successfully capture the main feature of the \(\gamma \)-oscillations in SNNs. This effectively makes the DNN a surrogate of the SNNs. In principle, this work could be easily extended to partial synchrony and oscillatory dynamics in more biologically realistic networks using, e.g., Hodgkin-Huxley models or neuronal compartmental models (Hodgkin and Huxley 1952; Bressloff 1994).

The organization of this paper is as follows. Section 2 introduces the neuronal network model that serves as the ground truth. The descriptions and capturing algorithms of MFEs are depicted in Sect. 3. Section 4 discusses how to set up the training set for artificial networks. The main results are demonstrated in Sect. 5. Section 6 is the conclusion and discussion.

2 Neuronal network model description

Throughout this manuscript, we study SNN dynamics with a Markovian integrate-and-fire (MIF) neuronal network model. This model imitates a small, local circuit of the brain and shares many features of local circuits in living brains, including extensive recurrent interactions between neurons, leading to the emergence of \(\gamma \)-oscillations. We will evaluate the performance of DNNs based on their predictive power of dynamics produced by the MIF model.

2.1 An Markovian spiking neuronal network

We consider a simple spiking network consisting of \(N_E\) excitatory (E) neurons and \(N_I\) inhibitory (I) neurons, all homogeneously connected. The membrane potential (V) of each neuron is governed by Markovian dynamics, with the following assumptions:

  1. 1.

    V takes value in a finite, discrete state space;

  2. 2.

    For neuron i, \(V_i\) is driven by both external and recurrent E/I inputs from other neurons, through the effect of the arrival of spikes (or action potentials);

  3. 3.

    A spike is released from neuron i when \(V_i\) is driven to the firing threshold. Immediately after that, neuron i enters the refractory state before resetting to the rest state;

  4. 4.

    For a spike released by neuron i, a set of post-synaptic neurons is chosen randomly. Their membrane potentials are driven by this spike.

We now explain these assumptions in detail.

Single neuron dynamics. Let us index the \(N_E\) excitatory neurons from \(1,\ldots ,N_E\), and the \(N_I\) inhibitory neurons from \(N_E+1,\ldots ,N_E+N_I\). For neuron i (\(i=1,\ldots ,N_E+N_I\)), the membrane potential \(V_i\) lies in a discrete state space \(\varGamma \)

$$\begin{aligned} V_i\in \varGamma :=\{-M_r, -M_r+1,\ldots ,-1,0,1,\ldots , M\}\cup \{\mathcal {R}\}, \end{aligned}$$

where the states \(M_r\), M, and \(\mathcal {R}\) are the inhibitory reversal potential, the spiking threshold, and the refractory state, respectively. Once a neuron is driven to the threshold M, its membrane potential \(V_i\) enters the refractory state \(\mathcal {R}\). After an exponentially distributed waiting time with mean \(\tau _\mathcal {R}\), \(V_i\) is reset to the rest state 0. \(V_i\) is driven by the external (i.e., from outside the network itself) and recurrent inputs to neuron i. It is worth noting that, while a neuron is in the refractory state \(\mathcal {R}\), \(V_i\) does not respond to any stimuli.

The external stimulus serves as an analog of feedforward sensory input, e.g., from the thalamus or from other brain regions. In this paper, the external inputs to individual neurons are modeled as series of impulsive kicks, whose arrival times are drawn from independent & identical Poisson processes. The rates of the Poisson processes, \(\lambda ^{E,I}\), are taken to be constants across the E/I populations. Each kick received by neuron i increases \(V_i\) by 1.

The recurrent inputs to a neuron are the spikes from other neurons. That is, an E/I neuron will spike when its membrane potential \(V_i\) reaches threshold M, sending an E/I kick to its postsynaptic neurons (the choice of which will be discussed momentarily). Each recurrent E spike received by neuron i takes effect on \(V_i\) after an independent, exponentially distributed time delay \(\varvec{\tau }\sim \text {Exp}(\tau ^{E})\). The excitatory spikes received by neuron i that have not yet taken effect form a pending E-spike pool, with size \(H_i^{E}\). Therefore, \(H_i^{E}\) increases by 1 when an E kick arrives at neuron i, and drops by 1 when a pending spike takes effect. This discussion applies to the I spikes as well: the size of the pending I-spike pool is \(H_i^I\), and the waiting time of the pending spikes is subjected to \(\varvec{\tau }\sim \text {Exp}(\tau ^{I})\).

In summary, the state of neuron i is therefore described by a triplet

$$\begin{aligned} (V_i, H_i^E, H_i^I). \end{aligned}$$

We note that the pool sizes \(H_i^E\) and \(H_i^I\) may be viewed as E and I synaptic currents of neuron i in the classical leaky integrate-and-fire neuron model (Gerstner 2014).

Impacts of spikes. The effects of the recurrent E/I spikes on membrane potentials are different. When a pending E-spike takes effect, \(V_i\) is increased by \([S^{Q,E}]+u^E\), where

$$\begin{aligned} u^E\sim \text {Bernoulli}(p)\quad \text {and} \quad p&=S^{Q,E}-[S^{Q,E}], \end{aligned}$$

where \(Q\in \{E,I\}\) and \([\ldots ]\) denotes the floor integer function. Likewise, when a pending I-spike takes effect, \(V_i\) is decreased by \([S^{Q,I}]+u^I\), where

$$\begin{aligned} u^I\sim \text {Bernoulli}(q)\quad \text {and} \quad q&=S^{Q,I}-[S^{Q,I}], \end{aligned}$$

\(V_i\) is strictly bounded to the state space \(\varGamma \). Should \(V_i\) exceed M after an E-spike increment, it will reset to \(\mathcal {R}\), and neuron i spikes immediately. On the other hand, should \(V_i\) go below \(-M_r\) due to an I-spike, it will stay at \(-M_r\) instead.

A homogeneous network architecture. Instead of having a predetermined network architecture with fixed synapses, the set of postsynaptic neurons of each spike are decided on-the-fly. That is to say, a new set of postsynaptic neurons is chosen independently for each spike. More specifically, when a type-\(Q'\) neuron spikes, the targeted postsynaptic neurons in the Q populations, excluding the spiking neuron itself, are chosen with probabilities \(P^{QQ'}\) (\(Q,Q'\in \{E,I\}\)). We point out that the motivation of this simplification is for analytical and computational convenience by making neurons interchangeable within each subtype, and is standard in many previous theoretical studies (Cai 2006; Brunel and Hakim 1999; Wilson and Cowan 1972; Buice and Cowan 2007; Cai et al. 2021; Gerstner 2014).

To summarize, the state space of the network is denoted as \(\varvec{\Omega }\). A network state \(\omega \in \varvec{\Omega }\) consists of \(3(N_E+N_I)\) components

$$\begin{aligned} \omega =(&V_1,\ldots ,V_{N_E},V_{N_E+1},\ldots ,V_{N_E+N_I},\nonumber \\&H^E_1,\ldots ,H^E_{N_E},H^E_{N_E+1},\ldots ,H^E_{N_E+N_I},\nonumber \\&H^I_1,\ldots ,H^I_{N_E},H^I_{N_E+1},\ldots ,H^I_{N_E+N_I}). \end{aligned}$$
(1)

2.2 Parameters used for simulations

The choices of parameters are adopted from our previous studies (Li and Hui 2019; Cai et al. 2021; Wu et al. 2022). For all SNN parameters used in the simulations, we list their definitions and values in Table 1. Here, we remark that the projection probability between neurons \(P^{Q'Q}\) (\(Q,Q'\in \{E,I\}\)) are chosen to match the anatomical data in the macaque visual cortex, see (Chariker et al. 2016) for reference. Also, \(\tau ^E<\tau ^I\), since it is known that the Glu-AMPA receptors act faster than the GABA-GABA receptors, with both on a time scale of milliseconds (Koch 1999).

On the other hand, four parameters concerning recurrent synaptic coupling strength will be tested and varied in this study. This is because they are sensitive to SNN dynamics, and yet hard to directly measure by current experiments methods (Chariker et al. 2016; Xiao et al. 2021). The range of tested parameter set \(\varvec{\Theta }=\{S^{EE},S^{EI},S^{IE},S^{II}\}\subset \mathbb {R}^4\) is also given by Table 1.

Table 1 Parameters regarding the network architecture (first row) and individual neuronal physiology (second row)

3 Multiple-firing events

In the studies of \(\gamma \)-oscillations. In many brain regions, electrophysiological recordings of the local field potentials reveal temporal oscillations with power peaking in the \(\gamma \)-frequency band (30–90 Hz) (Andrew Henrie and Shapley 2005; Brosch et al. 2002; Bauer 2006; Buschman and Miller 2007; Pieter Medendorp 2007; van Wingerden 2010; Csicsvari 2003; Popescu et al. 2009; Van Der Meer and David Redish 2009). Because of the belief that these coherent rhythms play a crucial role in cognitive computations, there has been much work on understanding their mechanisms in different brain areas, in disparate dynamical regimes, and within various brain states (Azouz and Gray 2000, 2003; Frien 2000; Womelsdorf 2012; Liu and Newsome 2006; Fries 2001, 2008; Bauer et al. 2007; Pesaran 2002; Womelsdorf 2007; Başar 2013; Krystal 2017; Mably and Colgin 2018).

To explain the neural basis of \(\gamma \)-oscillations, a series of theoretical models have found transient, nearly synchronous collective spiking patterns that emerge from the tight competitions between E and I populations (Whittington 2000; Rangan and Young 2013; Traub 2005; Chariker and Young 2015; Chariker et al. 2018). More specifically, rapid & coherent firing of neurons occurs within a short interval, and such spiking patterns recur with \(\gamma \)-band frequencies in a stochastic fashion. This phenomenon is termed multiple-firing events (MFEs). MFEs are triggered by a chain reaction initiated by recurrent excitation, and terminated by the accumulation of inhibitory synaptic currents. In this scenario, the \(\gamma \)-oscillations appear in electrophysiological recordings as the local change of the electric field generated by MFEs. We refer readers to Rangan and Young 2013; Chariker and Young 2015; Chariker et al. 2018; Li and Hui 2019; Cai et al. 2021; Wu et al. 2022 for further discussions of MFEs.

The alternation of fast and slow phases is one of the most significant dynamical features of MFEs (Fig. 1). Namely, at the beginning of an MFE, a chain reaction leads to a transient state where many neurons fire collectively in a short time interval (the fast phase). On the other hand, during the interval between two successive MFEs (an inter-MFE interval, or IMI), the neuronal network exhibits low firing rates while the system recovers from the extensive firings via a relatively quiescent period (the slow phase). These two phases can be discriminated by the temporal coherence of spikes in the raster plot where the time and location of spikes are indicated (Fig. 1A, blue dots). Here, MFEs and IMIs are separated by vertical lines - we will discuss the method of capturing MFEs momentarily in Sect. 3.2.

Fig. 1
figure 1

Divergence between SSA and tau-leaping methods. A Raster plot from SSA (blue) and tau-leaping method (orange) using same seed. B HEE trajectories from SSA and tau-leaping method. C-D Voltage trajectories of an E/I neuron from SSA and tau-leaping method. Vertical red/black lines indicate the beginnings/endings of MFEs. (solid: SSA, dashed: tau-leaping) (color figure online)

The complexity of MFEs is partially reflected by their sensitivity to spike-timing fluctuations (Xiao and Lin 2022). Specifically, during an MFE, the missing or misplacement of even a small number of spikes can significantly alter the path of network dynamics. Furthermore, the high dimensionality of state space \(\varvec{\Omega }\) and the high degree of nonlinearity make MFE hard to analyze. On the other hand, the network dynamics during IMIs are more robust to perturbations and exhibit low-dimensional features (Cai et al. 2021; Wu et al. 2022).

3.1 Spike-timing sensitiveness of transient dynamics

We illustrate this slow-fast dichotomy through a comparison between two different numerical simulations, one using the stochastic simulation algorithm (SSA) and another one using tau-leaping (Fig. 1).

Stochastic simulation algorithms vs. tau-leaping. In SSA, the timings of state transitions in phase space are exact, since the next state transition is generated by the minimum of finitely many independent, exponentially distributed random variables (see Appendix). On the other hand, algorithms simulating random processes with fixed time steps, such as tau-leaping, introduce errors to event times at every step. In a tau-leaping simulation with a fixed step size of \(\varDelta t\), the effect of spikes (interactions among neurons) is processed after each time step. Hence, within each time step, all events are uncorrelated, and changes are held off until the next update. Therefore, the precision of SSA is determined by the computational precision of the C++ code, which is much higher than tau-leaping.

Comparison. Intuitively, tau-leaping methods can well capture the slow network dynamics during IMIs, but not the fast chain reactions during MFEs. The detailed comparison is depicted in Fig. 1, where we choose \(\varDelta t = 0.5\) ms for the tau-leaping method. To constrain differences induced by stochastic fluctuations, we couple the two simulations with the same external noise, leaving the intrinsic noise (the random waiting times of pending spikes and refractory states) generated within the algorithms themselves. The raster plots depict spiking events produced by both methods and diverge rapidly during MFEs (Fig. 1A. blue: SSA; orange: tau-leaping), since the crucial coherence between firing events is not accurately captured by tau-leaping. This point is also supported by the comparison between voltage traces of single-neurons (Fig. 1C, D): Although well aligned at the beginning, the spike timings of the neurons are strongly affected by the transient fluctuation during MFEs.

In addition to the firing events, we also illustrate the comparison between pending spikes. Here we define

$$\begin{aligned} H^{Q'Q} = \sum _{i\in Q} H_i^{Q'}, \quad Q,\,Q'\in \{E,I\}, \end{aligned}$$

i.e., the total number of pending \(Q'\)-spikes for type-Q neurons. Figure 1B demonstrates the divergence of \(H^{{EE}}\) collected from different simulations. Because of the identical external random noise, trajectories from the two methods are similar in the first hundred milliseconds. However, very rapidly, the accumulation of errors in the first fast phase causes a large divergence. Other pending spikes statistics yield similar disagreement between the two methods (data not shown).

Therefore, in the rest of this paper, we use SSA to simulate the SNN dynamics, which is treated as the “ground truth".

Fig. 2
figure 2

The amount of recurrent excitation serves as the indicator of MFEs. A-B Number of E spikes caused by recurrent excitation within each 0.25 ms interval. \((S^{EE},S^{IE},S^{EI},S^{II}) = (4, 3, -2.2, -2)\) for A; \((4.44, 2.88, -2.22, -1.70)\) for B C-D Raster plot of the two simulations. Vertical red and black lines indicate the beginnings and endings of MFEs respectively

3.2 Capturing MFEs from network dynamics

The first step of investigating MFE is the accurate detection of MFEs patterns from the temporal dynamics of SNN. Due to the lack of a rigorous definition in previous studies, we develop an algorithm to capture MFEs based on the indication of recurrent excitation (Algorithm 1), splitting the spiking patterns into consecutive phases of MFEs and IMIs.

The algorithm capturing MFEs. Our basic assumption is that the existence of the cascade of recurrent excitation has a causal relation with MFEs. Therefore, given a temporal spiking pattern generated by the SNN, Algorithm 1 detects the initiation of a candidate MFE with the following criteria:

  • Two E-to-E spikes take place consecutively, where the second one is triggered due to the influence of the first one;

  • The second E-to-E spike should occur within a 4 ms interval following the first one (twice the excitatory synaptic timescale \(\tau _E=2\)ms).

After the cascade of spiking events, the candidate MFE is deemed terminated if no additional E spike takes place within a 4 ms time-window. Furthermore, to exclude isolated firing events clustering by chance, we apply merging and filtering processes to each of the MFE candidates. More specifically, consecutive MFE candidates are merged into one if they occur within 2 ms. Candidates are eliminated if their duration is less than 5 ms or the number of spikes involved is less than 5. We comment that the filtering threshold is chosen based on the size of 400-neuron networks. A different filtering standard is employed for larger networks, presented in Sect. 4.3.

Algorithm 3.2, based on the timing of recurrent E-spikes, is employed throughout this manuscript to detect MFEs. Two examples of different choices of parameters are illustrated in Fig. 2, where the initiation and termination of MFEs are indicated by red and black vertical lines. In the rest of the paper, we denote the time sections of initiations and terminations of the m-th MFE as \(t_m\) and \(s_m\). Therefore, the m-th MFE takes place during the interval \([t_m, s_m]\), whereas the IMI after that is \([s_m, t_{m+1}]\).

Algorithm 1
figure a

Capturing MFE

4 Learning MFEs with artificial neural networks

Viewed within the framework of random dynamical systems, SNN dynamics exhibiting \(\gamma \)-oscillation can be effectively represented by a Poincare map theory (Wu et al. 2022). Consider a solution map of the SNN dynamics depicted in Sect. 2,

$$\begin{aligned} \omega (t) = \varPhi ^{\theta }_t(\omega _0, \xi _0^t), \end{aligned}$$

where \(\omega _0\in \varvec{\Omega }\) is the initial condition of the SNN, \(\theta \in \varvec{\Theta }\) indicates SNN parameters and \(\xi _0^t\) is the realization of all external/internal noises during the simulation from time 0 to t. (The solution map satisfies \(\varPhi _{t+\tau } = \varPhi _t\circ \varPhi _\tau \).) To study the recurrence of MFEs, we focus on the time sections \(\{t_m\}\) on which \(\omega (t_m)\) returns to the initiation of the mth MFE. Therefore, we define a sudo-Poincare map:

$$\begin{aligned} \omega (t_{m+1}) = F^{\theta }(\omega (t_{m})) = \varPhi ^\theta _{t_{m+1}-t_m}(\omega (t_m), \xi _{t_m}^{t_{m+1}}). \end{aligned}$$
(2)

We note that \(F^{\theta }\) is not a Poincare map in the rigorous sense, since the initiation of MFEs depends on the temporal dynamics in a short interval (Algorithm 3.2).

This paper aims at investigating \(F^{\theta }\). It is generally impractical to study \(F^{\theta }\) in an analytical manner due to the high dimensionality of \(\varvec{\Omega }\). To overcome the difficulty, we propose a first-principle-informed deep learning framework.

4.1 Dissecting \(F^{\theta }\) into slow/fast phases

We first dissect \(F^{\theta }\) guided by SNN dynamics. In the regime of \(\gamma \)-oscillations, the SNN dynamics is composed of the regular alternation of slow/fast phases, though the duration and detailed dynamics of each phase may vary. Therefore, the pseudo-Poincare map \(F^{\theta }\) is equivalent to the composition of two maps:

$$\begin{aligned} F^{\theta } = F^{\theta }_1 \circ F^{\theta }_2, \quad F^{\theta }_{1,2}: \varvec{\Omega }\rightarrow \varvec{\Omega }. \end{aligned}$$
(3)

More specifically, \(F^{\theta }_1\) maps the network state at the initiation of an MFE to its termination, while \(F^{\theta }_2\) maps the network state from the beginning of an IMI to the initiation of the next MFE, i.e.,

$$\begin{aligned} F^{\theta }_1(\omega (t_m)) = \omega (s_m), \quad F^{\theta }_2(\omega (s_m)) = \omega (t_{m+1})\, . \end{aligned}$$

We denote \(F^{\theta }_{1,2}\) as MFE and IMI mappings, respectively. To summarize, the dynamical flow of \(\varPhi ^t\) is equivalent to:

$$\begin{aligned}...\,\omega (t_m)\xrightarrow [\text {MFE}]{F^{\theta }_1} \omega (s_m) \xrightarrow [\text {IMI}]{\ ~~~~F^{\theta }_2~~~~\ } \omega (t_{m+1})\xrightarrow [\text {MFE}]{F^{\theta }_1} \omega _{s_{m+1}}... \end{aligned}$$

Our previous studies demonstrated that the slow and relatively quiescent dynamics during IMIs can be well captured by classical coarse-graining methods, i.e., the IMI mapping \(F^{\theta }_2\) may be represented by the evolution of certain Fokker-Planck equations (Cai 2006). However, this is not the case for the MFE mapping \(F^{\theta }_1\) due to the highly transient and nonlinear dynamics. Instead, we turn to deep neural networks (DNN) due to their success in tackling high-dimensional and nonlinear problems. In the rest of this section, our goal is to train a DNN with relevant data and generate surrogates of \(F^{\theta }_1\).

4.2 First-principle-based reductions of the problem

Instead of requiring the DNN to learn the full map \(F^{\theta }_1\) and directly link the network state from \(\omega (t_m)\) to \(\omega (s_m)\), we prepare a training set \(\mathcal {T}^\theta _{\text {train}}\) with first-principle-based dimensional reduction, noise elimination, and enlargement for robustness to facilitate the training process.

The DNN used in this paper has a feedforward structure with 4 layers, consisting of 512, 512, 512, and 128 neurons, respectively. We leave the detailed information of the DNN architecture to Appendix.

4.2.1 Coarse-graining SNN states

Approximating the features of MFE mapping \(F^{\theta }_1\) with a DNN (or any statistical machine learning method) immediately faces the curse of dimensionality. Namely, \(F^{\theta }_1\) maps a 3N-dimensional space \(\varvec{\Omega }\) to itself. To cope with N, we propose a coarse-graining model reduction with a priori physiological information.

Instead of enumerating the actual state of every neuron, we assume the E and I neuronal populations form two ensembles. That is, the state of a type-Q neuron can be viewed as randomly drawn from a distribution \(\varvec{\rho }^{Q}(v,H^E, H^I)\) of the Q-ensemble, where \(Q \in \{E,I\}\). Furthermore, note that there is no fixed architecture of synaptic connection in SNN (see Sect. 2)—any neurons in the same population has the same probability to be projected when a spike occurs. Therefore, it is reasonable to decorrelate v and the \((H^E, H^I)\), i.e.,

$$\begin{aligned} \varvec{\rho }^{Q}(v,H^E, H^I) \sim \varvec{p}^{Q}(v)\cdot \varvec{q}^{Q}(H^E, H^I), \end{aligned}$$
(4)

where \(\varvec{p}^{Q}\) and \(\varvec{q}^{Q}\) are marginal distributions of \(\varvec{\rho }^{Q}\).

More specifically, \(\varvec{p}^{E}(v),\varvec{p}^{I}(v)\) yield the distributions of neuronal voltages for both populations on a partition of the voltage space \(\varGamma \):

$$\begin{aligned} \varGamma&= \varGamma _1 \cup \varGamma _2 \cup ... \cup \varGamma _{22} \cup \varGamma _{\mathcal {R}}\nonumber \\&= [-M_r,0) \cup [0,5) \cup ... \cup [M-5,M=100) \cup \{\mathcal {R}\}. \end{aligned}$$
(5)

On the other hand, the distribution of pending spike \(\varvec{q}^{Q}(H^E, H^I)\) is effectively represented by the total number of pending spikes summed over each population, \(H^{QE}\) and \(H^{QI}\). To summarize, we define a coarse-grained mapping \(\mathcal {C}:{\varvec{\Omega }} \mapsto {\tilde{\varvec{\Omega }}}\) by projecting the network state \(\omega \) onto a 50-dimensional network state \(\tilde{\omega }\):

$$\begin{aligned} \mathcal {C}(\omega )=&\,\tilde{\omega } \nonumber \\ =&\left( \varvec{p}^{E}(v),\varvec{p}^{I}(v), H^{EE}, H^{EI}, H^{IE}, H^{II}\right) \nonumber \\ =&\left( n^E_1, n^E_2, \cdots , n^E_{22}, n^E_R, n^I_1, n^I_2, \cdots , n^I_{22}, n^I_R, \right. \nonumber \\&\left. H^{EE}, H^{EI}, H^{IE}, H^{II}\right) . \end{aligned}$$
(6)

Here, \(n^Q_k\) denotes the number of type-Q neurons whose potentials lie in \(\varGamma _k\), and \({\tilde{\varvec{\Omega }}}\subset \mathbb {R}^{50}\) represents the coarse-grained state space. The precise definition of \(n^Q_k\) is given in the Appendix.

4.2.2 Training DNN

We now train a DNN to learn the coarse-grained mapping \(\widetilde{F}_1^\theta \) with the coarse-grained network states, \(\tilde{\omega }_{s_m} = \widetilde{F}_1^\theta (\tilde{\omega }_{t_m})\), where \(\tilde{\omega }_{t} = \mathcal {C}(\omega (t))\). That is, for a set of fixed SNN parameters (\(\theta \)), the DNN forms a mapping

$$\begin{aligned} \widehat{F}^{\theta ,\vartheta }_1: {\tilde{\varvec{\Omega }}} \mapsto {\tilde{\varvec{\Omega }}} \end{aligned}$$

where \(\vartheta \) is the hyperparameter for the trained DNN. From a simulation of SNN dynamics producing K MFEs, each piece of data of the training set is composed by

  • Input: \(x_m = \tilde{\omega }_{t_m}\), the coarse-grained network states at the initiation of the m-th MFE, and

  • Output: \(y_m = (\tilde{\omega }_{s_m},\text {Sp}_E, \text {Sp}_I)\), the coarse-grained state at the m-th MFE termination and the numbers of E/I-spikes during MFEs (\(m \in \{1,2,..., { K}\}\)).

The training process aims for the optimal \(\vartheta \) to minimize the following \(L^2\) loss function

$$\begin{aligned} \mathcal {L}^\theta (\vartheta ) = \frac{1}{{ K}} \sum _{m=1}^{ K} \Vert y_m - \widehat{F}^{\theta ,\vartheta }_1 ( x_m ) \Vert ^2_{L^2} \, , \end{aligned}$$
(7)

and we hope to obtain the optimal coarse-grained MFE mapping \(\widehat{F}^{\theta }_1\) as an effective surrogate of \(\widetilde{F}^{\theta }_1\).

Fig. 3
figure 3

Pre-processing and enlarging the training data-set. A An example of pre-MFE E-neuron voltage distribution \(\varvec{p}^{E}(v)\) before DCT + iDCT. For clarity, the number of refractory neurons is depicted by the rightmost bin; B The pre-processed \(\widehat{\varvec{p}^{E}}(v)\) after DCT + iDCT; C The distribution of the 2nd DCT mode of \(\widehat{\varvec{p}^{E}}(v)\); D The distribution of the second DCT mode of \(\widehat{\varvec{p}^{I}}(v)\)

4.2.3 Pre-processing: Eliminating high-frequency noises

Using a discrete cosine transform (DCT), we pre-process the training data (xy) by eliminating the noisy dimensions in potential voltage distributions \(\varvec{p}^{Q}\), i.e., the high-frequency modes.

The raw distributions of membrane potentials \(\varvec{p}^{Q}(v)\) contain significant high-frequency components (Fig. 3A). This is partially due to the stochasticity and the small number of neurons in the MIF model. Unfortunately, the high-frequency components could lead to “conservative" predictions, i.e., the training of DNN converges to the averages of post-MFE voltage distributions to minimize \(\mathcal {L}^\theta (\vartheta )\). Therefore, different inputs would yield similar outputs.

To resolve this issue, we apply a DCT method to remove the high-frequency components from training data (xy), only preserving the first eight modes. This is sufficient to discriminate a pair of neurons whose membrane potentials \(|v_1-v_2|>5\), where \(v_1, v_2 \in \varGamma \). For \(|v_1-v_2|<5\), the two neurons are either placed in the same or consecutive two bins after coarse-graining. Therefore, the E spikes needed to involve them in the upcoming MFE differ by at most 1.

More specifically, in the training set, \(1.86\times 10^5\) pairs of network state (xy) are collected from a 5000-second simulation of the SNN network (see Eq. 6). For the voltage distribution components of each state, \(\varvec{p}^{E}(v)\) and \(\varvec{p}^{I}(v)\), we remove the high-frequency components by

$$\begin{aligned} \widehat{\varvec{p}^{Q}} = \mathcal {F}_c^{-1} \circ T \circ \mathcal {F}_c\{\varvec{p}^{Q}\}, \quad Q\in \{E,I\}. \end{aligned}$$
(8)

Here, \(\mathcal {F}_c\) indicates the DCT operator and T is a truncation matrix preserving the first eight components. Finally, in the training data (xy), the coarse-grained network state \(\tilde{\omega }\) is replaced as

$$\begin{aligned} \tilde{\omega } = \left( \widehat{\varvec{p}^{E}}(v),\widehat{\varvec{p}^{I}}(v), H^{EE}, H^{EI}, H^{IE}, H^{II}\right) . \end{aligned}$$

Fig. 3A, B depicts an example of pre-processing a voltage distribution. We leave the rest of the details regarding the DCT methods to the Appendix.

4.2.4 Robustness of training

The robustness issue of DNN has been addressed in many previous studies (Goodfellow et al. 2014; Yuan 2019). One approach to improving robustness is via data augmentation, in which the training set is enlarged by adding modified copies of existing data (Shorten and Khoshgoftaar 2019). Motivated by this idea, we propose a training set \(\mathcal {T}^\theta _{\text {train}}\) to account for more irregular inputs than SNN-produced network states, helping the trained DNN generate realistic and robust predictions for MFE dynamics. Based on the pre-processed network states \(\tilde{\omega }\), we increase the variability of the first eight frequency modes of \(\widehat{\varvec{p}^{Q}}(v)\) as they are the most salient information revealed by DCT.

\(\mathcal {T}^\theta _{\text {train}}\) consists of pairs of pre/post-MFE network states \((x_m, y_m)\). A candidate initial state close to a pre-MFE state \(x_m\) in \(\mathcal {T}^\theta _{\text {train}}\) is generated as follows:

  1. 1.

    The voltage distributions \(\widehat{\varvec{p}^{Q}}(v)\). We first collect the empirical distributions of all frequency modes (by concerning each entry of \(\mathcal {F}_c\{\varvec{p}^{Q}\}\)) from \(1.86\times 10^5\) pre-processed pre-MFE network states. The empirical distributions are fitted by direct combinations of Gaussian and exponential distributions (see, e.g., Fig. 3C, D). We then increase the variances of fitted distributions by a factor of three, from which the 2nd-8th frequency modes are sampled. On the other hand, the first mode comes from the distributions with the original variances, since it indicates the number of neurons outside the refractory state. Finally, the higher order of DCT frequency modes are treated as noises and truncated.

  2. 2.

    The pending spikes \(H^{Q'Q}\). Likewise, the “fit-expand-sample" operations similar to how we treated the 2nd-8th frequency modes in \(\widehat{\varvec{p}^{Q}}(v)\) are applied to sample the number of pending spikes.

It is also important to make sure all training data are authentic abstractions of SNN states in MFE dynamics, i.e., an MFE can be triggered in a short-time simulation (5ms) from a pre-MFE state. Therefore, we perform a simulation that begins with each initial state generated by the enlargement above, and collect the corresponding pre-MFE states (x) and post-MFE states (y) if an MFE emerges. The enlargement of distributions of MFE magnitudes is shown in Fig. 18

In summary, the enlarged \(\mathcal {T}^\theta _{\text {train}}\) consists of \(3\times 10^5\) pairs of \((x_m,y_m)\):

$$\begin{aligned} \mathcal {T}^\theta _{\text {train}} = \left\{ (x_m, y_m): m = 1,2,...,3\times 10^5\right\} , \end{aligned}$$

where \(50\%\) of the data comes directly from network simulation (after pre-processing of DCT), and the rest comes from the data augmentation process.

4.3 Generalization for different SNN parameters

Varying the synaptic coupling strengths. Trained with data from a dynamical system parameterized by \(\theta \), DNN produces a surrogate mapping \(\widehat{F}^{\theta }_1\). Recall that \(\theta \) is a parameter point in a 4D cube of recurrent synaptic coupling strength

$$\begin{aligned} \varvec{\Theta }&= \left\{ (S^{EE}, S^{IE}, S^{EI}, S^{II} )\in \right. \\&\left. [3.5, 4.5] \times [2.5, 3.5] \times [-2.5, -1.5] \times [-2.5, -1.5]\right\} . \end{aligned}$$

We treat \(\widehat{F}^{\theta }_1\) as a milestone, and further propose a parameter-generic MFE mapping

$$\begin{aligned} \widehat{F}_1: {\tilde{\varvec{\Omega }}} \times \varvec{\Theta } \mapsto {\tilde{\varvec{\Omega }}}. \end{aligned}$$

That is, given any point \(\theta \in \varvec{\Theta }\) and a pre-MFE state, the DNN predicts the post-MFE state. The optimization problem and loss function are analogous to Eq. 7.

The training set \(\mathcal {T}_{\text {train}}\) for parameter-generic MFE mapping consists of SNN states from 20000 different of parameters points, which are randomly drawn from \(\varvec{\Theta }\). To ensure reasonable SNN dynamics, each parameter point \(\theta \) is tested by the following criteria

$$\begin{aligned}&(f_E, f_I) = L(\theta ) \\&f_E\le 50 \, \text{ Hz } \text{ and } \, f_I\le 100 \text{ Hz, } \end{aligned}$$

where \(L(\theta )\) is a simple linear formula estimating firing rates of E/I populations based on recurrent synaptic weights (for details see Appendix and Li et al. 2019). For each accepted point in parameter space, we perform a 500 ms simulation of SNN and collect 20 pairs of pre-processed, coarse-grained pre- and post-MFEs states (see Sect. 4.2). The chosen parameter space is wide to generate various network dynamics and various MFEs. The firing rates of networks with parameters that can produce accepted MFEs are shown in Fig. 19, and the various sizes of produced MFEs are shown in Fig. 17.

Varying the network size. Our method is generic to \(\gamma \)-oscillations produced by SNNs of different sizes. We demonstrate this on a 4000-neuron SNN (3000 E and 1000 I neurons) sharing all parameters with the previous 400-neuron SNN, except that the synaptic weights (\(S^{QQ'}\)) are 1/10 of the values listed in Table 1. This change of synaptic weights aims to control the total recurrent E/I synaptic drives received by each neuron and allows the different network models to have the same mean-field limit. While deferring DNN predictions of SNN dynamics to Sect. 5, we here note the minor modifications to adopt our methods to the 4000-neuron SNN.

First, the filtering threshold of MFE is increased from 5 to 50 spikes when collecting MFE-related data (see Sect. 3.2), since the MFEs are larger. Second, because of the central limit theorem, less intrinsic stochasticity are observed in the SNN dynamics. This leads to an immediate side effect: The span of pre/post-MFE network states collected from SNN simulations is relatively narrower within \(\varvec{\Omega }\), i.e., the training data is less “general". To compensate, we expand the spans of \(\mathcal {T}^\theta _{\text {train}}\) more courageously. That is, during the enlarging step of the training set, the components of pre-MFE states are sampled from distributions with more significant variance expansions.

4.4 Training result

The trained DNNs provide faithful surrogate MFE mappings for both \(\widehat{F}^{\theta }_1\) and \(\widehat{F}_1\).

Parameter-specific MFE mappings. We first illustrate \(\widehat{F}^{\theta }_1\) at a particular parameter point

$$\begin{aligned} \theta = (S^{EE}, S^{EI}, S^{IE}, S^{II}) = (4, 3, -2.2, -2). \end{aligned}$$

We test the predictive power of \(\widehat{F}^{\theta }_1\) on a testing set

$$\begin{aligned} \mathcal {T}^{\theta }_{\text {test}} = \left\{ (x'_m, y'_m): m = 1,2,..., M=6\times 10^4\right\} , \end{aligned}$$

where \((x'_m, y'_m)\) are coarse-grained pre/post-MFE states without pre-processing collected from SNN simulations. In Fig. 4A, one example of DNN prediction to post-MFE \(\varvec{p}^E(v)\) is compared to the SNN simulation results starting from the same \(x'_m\). Also, the comparison between predicted vs. simulated E/I-spike numbers during MFEs is depicted in Fig. 4B. To demonstrate the accuracy of \(\widehat{F}^{\theta }_1\), Fig. 4C depicts the \(L^2\) losses of different components of post-MFE states. The \(L^2\) loss of predicted \(\varvec{p}^Q(v)\) is \(\sim \)4, while the averaged \(L^2\) difference between the post-MFE voltage distributions in the testing set is

$$\begin{aligned} \underset{1\le m,\ell \le M}{\text {mean}} \, \Vert \varvec{p}^{E}_m(v) - \varvec{p}^{E}_\ell (v) \Vert ^2 + \Vert \varvec{p}^{I}_m(v) - \varvec{p}^{I}_\ell (v) \Vert ^2 \approx 20\, , \end{aligned}$$

Notably, DCT pre-processing effectively improves DNN predictions of voltage distributions by reducing the \(L^2\) loss from \(\sim \)10 to \(\sim \)4. A similar comparison is observed for the prediction of pending spike numbers in the post-MFE states (Fig. 4C inset).

Parameter-generic MFE mappings. Likewise, \(\widehat{F}_1\) also provides faithful predictions following SNN dynamics in different parameter regimes by labeling training data with synaptic strength parameters. Figure 4D shows a similar comparison between \(L^2\) losses of different components of post-MFE states.

Fig. 4
figure 4

DNNs predictions of post-MFE states. A Mapping \(\widehat{F}^{\theta }_1\) for \(\theta = (S^{EE}, S^{EI}, S^{IE}, S^{II})= (4, 3, -2.2, -2)\). A Left: a pre-MFE \(\varvec{p}^E(v)\); Right: post-MFE \(\varvec{p}^E(v)\) produced by ANN (blue) vs. spiking network simulations (orange). B Comparison of E and I spike number during MFEs, ANN predictions vs. SSA simulations. The distributions are depicted by 10th contours of max in ks-density estimation; C \(L^2\) losses of predicted post-MFE \(\varvec{p}^Q(v)\) and pending spike number Hs. Blue: auto-\(L^2\) differences between training data; orange: averaged \(L^2\)-loss without pre-processing; yellow: averaged \(L^2\)-loss with pre-processing. D \(L^2\) loss for the parameter-generic mapping \(\widehat{F}_1\). The meaning of all bins are similar to B (color figure online)

5 Producing a surrogate for SNN

Here, we depict how our ANN predictions provide a surrogate of the spiking network dynamics. We focus on the algorithm to replace \(F^\theta \) for a fixed \(\theta \). The algorithm for parameter-generic F is analogous.

Recall that SNN dynamics is divided into a fast phase (MFEs) and a slow phase (IMIs). Our first-principle–informed DNN framework produces a surrogate mapping \(\widehat{F}^\theta _1\) and replaces the pseudo-Poincare mapping \(F^\theta \) if complemented by \(F^\theta _2\). We approximate the IMI mapping \(F^\theta _2\) by evolving SNN with a tau-leaping method, thereby producing the surrogate to the full SNN dynamics by alternating between the two phases.

To resemble the IMI dynamics, we first initialize the network state \(\omega \) from the coarse-grained post-MFE state \(\tilde{\omega }\) predicted by \(F^\theta _1\). Since neurons in the MIF model are exchangeable, \(\omega \) can be randomly sampled from \(\mathcal {C}^{-1}( \tilde{\omega })\). Specifically, we evenly assign voltage to each type-Q neuron based on \(\widehat{p^{Q}}\). On the other hand, same-category pending spikes stay “pooled" and are assigned to each neuron interchangeably. After that, the network state \(\omega \) is evolved by a tau-leaping method with 1-ms timesteps. This process is terminated if more than three E-to-E spikes or six E spikes occur within 1 ms, after which the next MFE is deemed to start and the network state \(\omega \) is fed to \(F^\theta _1\) for another round of prediction.Footnote 1 The loop \(F^\theta \) alternating between \(F^\theta _1\) and \(F^\theta _2\) is thus closed. The mathematical descriptions of the SNN surrogate algorithm are summarized as follows:

  1. (i)

    \(( \tilde{\omega }_{s_m}, \text {Sp}_E, \text {Sp}_I) = \widehat{F}^\theta _1(x_m)\), where \(\tilde{\omega }_{s_m}\) is the coarse-grained post-MFE state.

  2. (ii)

    Sample \(\omega (s_m)\) from \(\mathcal {C}^{-1}( \tilde{\omega }_{s_m} )\).

  3. (iii)

    Evolve the network dynamics with the tau-leaping method and initial condition \(\omega (s_m)\). Stop the simulation with terminal value \(\omega _{s_{m+1}}\).

  4. (iv)

    \(x_{m+1} = \mathcal {C}(\omega _{s_{m+1}})\).

  5. (v)

    Repeat (i)-(iv) with \(m = m+1\).

We demonstrate the resembling power by producing the raster plot of SNN dynamics labeling all spiking events with neuron index vs. time in Figs. 5 and 6. While the spiking events during IMIs are given by tau-leaping simulation, the spiking patterns during MFEs consist of events uniformly randomly assigned to neurons and times within the MFE interval (with the total number of spikes \((\text {Sp}_E, \text {Sp}_I)\) predicted by our DNN). The durations of the MFEs are randomly sampled from the empirical distribution collected from the SNN simulations. In the rest of the section, we give the details of our surrogate SNN dynamics constructed upon \(\widehat{F}^\theta _1\) and \(\widehat{F}_1\).

Fig. 5
figure 5

Surrogates of spiking network dynamics produced by the parameter-specific MFE mapping \(\widehat{F}^\theta _1\). A, C, E resembling a 400-neuron spiking network; B, D, F resembling a 4000-neuron network. A, B Example of pre and post-MFE voltage distributions \(\varvec{p}^E\) and \(\varvec{p}^I\) in the surrogate dynamics. C, D 10th level curves of the first two principal components of \(\varvec{p}^E\) and \(\varvec{p}^I\). (Blue: 20k examples from the enlarged training set, red: 2k examples from the surrogate dynamics.) E, F Raster plots of simulated surrogate dynamics and the real dynamics starting from the same initial profiles (color figure online)

Fig. 6
figure 6

Surrogates of spiking network dynamics produced by the parameter-generic MFE mapping \(\widehat{F}_1\). A-F are in parallel to Fig. 5

Parameter-specific predictions. We use the enlarged training set as described in Section 4 to generate \(\widehat{F}^\theta _1(x_m)\). Here, \(\theta = (S^{EE}, S^{IE}, S^{EI}, S^{II}) = (4, 3, -2.2, -2)\) for the 400-neuron SNN, and the synaptic weights are normalized by 10 times in the 4000-neuron SNN.

We first focus on the surrogate dynamics of the 400-neuron SNN. Figure 5A gives examples of pre-MFE voltage distributions \(\varvec{p}^Q\) generated by the tau-leaping simulation and post-MFE \(\varvec{p}^Q\) generated by DNN predictions. We further compare the distributions of the first two principal components of \(\varvec{p}^Q\) occurring in the surrogate dynamics and the training set (Fig. 5B. red: surrogate dynamics; blue: the training set). On the plane of the first two principal components, voltage distributions produced by surrogate dynamics distribute consistently with training sets, suggesting that the surrogate dynamics align very well with the ground-truth network dynamics. The red/blue contours indicate each tenth of the level curves of the ks-density of the distributions. We also compare the raster plots of the two dynamics, where the initial SNN conditions are the same (Fig. 5E).

The right half of Fig. 5 depicts the surrogate dynamics to the 4000-neuron SNN model. The biased distribution of the principal components of \(\varvec{p}^Q\) is probably due to A. the relatively more narrow training sets (see discussions in Sect. 4.3), and B. the large 1 ms timestep in tau-leaping.

Parameter-generic predictions. The surrogate dynamics generated by the parameter-generic surrogate MFE mapping \(\widehat{F}_1\) and tau-leaping are depicted in Fig. 6, whose panels are analogous to Fig. 5. (The “real" raster plots of SNN dynamics in Fig. 6E is fixed the same as Fig. 6E.) The comparable results demonstrate that the DNNs produce faithful surrogates to the SNN dynamics. Interestingly, by comparing panels C and D of Figs. 5 and 6, we find that the principal components of \(\varvec{p}^Q\) in the surrogate dynamics generated by \(\widehat{F}_1\) are much less biased than \(\widehat{F}^\theta _1\). While a detailed explanation is beyond the current scope of this study, our current conjecture is that the more general training set of \(\widehat{F}_1\) helps it deal with the less regular pre-MFE network states in the surrogate dynamics.

6 Discussions and conclusions

In this paper, we build an artificial neural network (ANN) surrogate of \(\gamma \)-dynamics arising from a biological neuronal circuit. The neuronal circuit model is a stochastic integrate-and-fire network that has been well-studied. Similar to many other models (Li et al. 2019; Zhang 2014; Rangan and Young 2013), it can exhibit semi-synchronous spiking activities called the multiple-firing events (MFEs), which are transient and highly nonlinear emergent phenomena of spiking neuronal networks.

In our study, the sensitive & transient MFE dynamics are represented by the MFE mappings that project the pre-MFE network states to post-MFE network states. The MFE mappings are faithfully approximated by ANNs, despite the significant intrinsic noise in the model. On the other hand, the slower and quieter dynamics between consecutive MFEs are evolved by standard tau-leaping simulations. Remarkably, a surrogate of spiking network dynamics is produced by combining ANN approximations and tau-leaping simulations, generating firing patterns consistent with the spiking network dynamics. Furthermore, the ANN surrogate can be generalized to a wide range of synaptic coupling strengths.

This paper explores the methodology of learning biological neural circuits with ANNs. In this study, the biggest challenges of developing a successful ANN surrogate are A. processing the high-dimensional, noisy data and B. building a representative training set. Both challenges are addressed by the first-principle-based model reduction techniques, i.e., coarse-graining and discrete cosine transform. The model reductions remove the excessive intrinsic noise. The training set collects network states from simulations of SNNs and is enlarged to represent a broader class of voltage distributions. Therefore, the training set covers the “rare" voltage distributions occurring in spiking network dynamics with low probabilities.

Future work. The idea of ANN surrogates elaborated in this paper can be extended and applied to other network models. First, many models of brain dynamics share the difficulties of dimensionality, robustness, and generalizability. Therefore, we propose to extend our ideas to model more sophisticated dynamical phenomena of the brain, such as other types of neural oscillations, and to neural circuits with more complicated network architecture. Furthermore, the power of ANNs to handle large data sets may allow us to extend our framework to deal with experimental data directly. In general, we are motivated by the recent success demonstrating the capability of deep neural networking in representing infinite-dimensional maps (Li et al. 2020; Lu 2021; Wang et al. 2021). Therefore, in future work, we suggest exploring more complex network structures (e.g., the DeepONet Lu 2021) to build mappings between the states of neural circuits.

Another interesting but challenging issue is the interpretability of ANN surrogates, e.g., relating statistics, dynamical features, and architectures of the spiking network models to the ANNs. A potentially viable approach is to map neurons in the spiking networks to artificial neurons, then examine the connection weight after training. However, it is likely that this idea may need ANNs more complicated than the simple feed-forward DNN we considered here. To achieve this goal, one may consider different ANNs with different architectures, such as ResNet or LSTM (He et al. 2016; Hochreiter and Schmidhuber 1997). These studies may shed some light on how the dynamics and information flow in neural systems are represented in ANNs.