Keywords

1 Introduction

Analytic models have limitations when applied to biological applications, because the of both the simplicity and the complexity of the biological systems [1, 2]. Neural networks have provided a functional alternatives, since they can allow for processing large data systems [3]. Neural networks can be considered as descriptive modeling much like more traditional analytic tools, but neural networks can reduce the need for computer power for every parameter that is introduced by analytic models [4]. In this way they can be used for undertaking enormous challenges such as artificial modeling o of the brain [5, 6].

Metabolic pathways can be considered as a series of enzyme-catalyzed reactions where the product of one reaction becomes the substrate for subsequent reactions and where all reactions are branched and interconnected as a graph network. Metabolic pathways have traditionally been modeled by using Michaelis-Menten kinetics [7]. Although the equation’s stable predictions work well in a homogenous system, it becomes very limited when the model is complex, and especially for reverse reactions, and hence the Michaelis-Menten kinetics does not necessarily describe enzyme biology accurately. To overcome this limitation, particularly for complex system interactions in metabolic pathways [8], we propose to use neural networks via Cell Communication Protocol© by using:

  1. 1.

    Self-Organizing Map (SOM) or Self-Organizing Feature Map (SOFM): a type of artificial neural network that is trained using unsupervised learning to produce a low-dimensional (typically two-dimensional), discretized representation of the input space of the training samples, and called a map. Self-organizing maps are different than other artificial neural networks in the sense that they use a neighborhood function to preserve the topological properties of the input space [9, 10].

  2. 2.

    Neural Gas: The neural gas is a biologically inspired adaptive algorithm. It sorts for the input signal according to how far away they are. A certain number of them are selected by distance in order, and then the number of adaptation units and strength are decreased according to a fixed schedule. Neural gas is a special kind of SOM, but here the fixed neighborhood of the SOMs does not exist. As such, even high dimensioned feature spaces can be analyzed or clustered; this could only be done by SOMs in special cases as they are 2 dimensioned [11].

  3. 3.

    Back Propagation (BPP): Propagation of error is a common method of teaching artificial neural networks how to perform a given task. It is a supervised learning method, and is an implementation of the delta rule. It requires a teacher that knows, or can calculate, the desired output for any given input. It is most useful for feed-forward networks (networks that have no feedback, or simply, that have no connections that loop). Back propagation requires that the activation function used by the artificial neurons (or “nodes”) is differentiable [12, 13].

The model can be tested against Kyoto Encyclopaedia of Genes and Genomes (KEGG) [14, 15], BioCyc database collection, MetaboLights, Molecular Ancestry Network (MANET) [16, 17] and Reactome [18].

2 Model

The neural network model (Fig. 1a) can be applied to metabolic pathways [19], and diffusion models. We considered three approaches to maximize the effectiveness of prediction by the neural network via Cell Communication Protocol.

Fig. 1
figure 1

a Neural network modelling allows inputs with different weights to be processed differently (via a sigmoid function). b Neural network modelling can allow for multiple connections to give a solution (output). c Detailed diagram of different weights reduced to a sigmoid function to give output activity

The first approach was by training the self-organizing map (SOM) by using competitive learning. When a training example is fed into the network, its Euclidian distance to all weight vectors is computed. The ‘neuron’, whose weight is most similar to the best matching input (BMU) and the neuron closest to the SOM lattice, are adjusted towards the input vector; the update is given by [20]:

$$Wv(s + 1) = Wv(s) + \theta \left( {u,v,s} \right)\alpha \left( s \right)\left( {D\left( i \right) - Wv\left( s \right)} \right)$$
(1)

where, \(s\) is the step index, \(i\) index for the training set, \(u\) is the weight for BMU, for \(D\left( i \right)\), \(\theta \left( {u,v,s} \right)\) is the neighborhood function depending on the lattice, \(\alpha \left( t \right)\) is a monotonically learning function, \(D\left( i \right)\) is the input vector, \(v\) is all the visit vectors. The SOM option shows the developed network [21].

The second approach uses neural gas which considers a probability distribution, \(P\left( i \right)\) of data vectors \(i\). Every time the \(s\) step is a random choice, with the \(V_{j}\), \(j = 1, \ldots ,N\):

$$V_{{j_{k} }}^{s + 1} = V_{{j_{k} }}^{s} + \varepsilon e^{{{{ - k} \mathord{\left/ {\vphantom {{ - k} \lambda }} \right. \kern-0pt} \lambda }}} \left( {i - V_{{j_{k} }}^{s} } \right)$$
(2)

The second approach is useful to limit the pathway distances, i.e. it tells us the limitations of the metabolic pathway [22, 23].

The third approach is back propagation based on the delta rule (Fig. 1b) where the initial weight is random. During the training, the weights are obtained from the input layers.

The input is a set of tuplets \(\left( {i_{1} ,i_{2} , \ldots ,i_{n} } \right)\), and \(\left( t \right)\) corrects the output. Because the initial weights are random, the weight from the training set is:\(y = i_{1} w_{1} + i{}_{2}w_{2}\) (NB because back propagation treats each neuron in isolation, gradient calculation considers only one input at a time); therefore the square error is \(E = \frac{1}{2}\left( {t - y} \right)^{2}\)[24].

The neuron activation is given by

$$a = \sum\limits_{j} {i_{j} *w_{j} } \quad \left( {{\text{Figure }}1{\text{c}}} \right)$$
(3)

The third approach allows handling of reverse catalysis which is one of the limitations of the Michaelis-Menten Kinetics [25].

3 Method

Metabolic pathways are a series of reactions where one reaction produces the substrate for the subsequent reactions; neural networks can match the model. The initial random weight entry incorporates the constant parameters and particle density from an enzyme assay. Specifically, the input vector is given by weight obtained from density of the enzyme assay. To develop the chemical pathway the number of clusters representing an assumed network for the metabolic pathway, and a maximal range of clusters nodes and possible nodes for the pathway are entered. Figure 2, shows input vectors for the training set that sets up the actual training of the network.

Fig. 2
figure 2

Computer screen for input vector, which represent the input space of the training sample, the data come from the database being analysed, and the data is defined as input vector (SOM Modus)

4 Results

The diffusion across the training set is shown in Fig. 3a, where the upper cells are the initial vectors from the weight given by the density of the enzyme. The reference vector is the upper part; in this example the simulation considered 100 input neurons. The input vector is obtained by the weight of the close nodes (neurons) to the SOM lattices with adjustment [26].

Fig. 3
figure 3figure 3

a The cluster consisted of vectors show in red, the green are the input vectors, the yellow is the reference vector (SOM Modus). b The activity card of the Neuron in 2-Dimensions and 3-Dimensions (show in the Torus) in the SOM Modus. c The number of Neurons and the option for 3-Dimensional and 2-Dimensional data visualization in the Neural Gas Modus. d The data being train in the Neural Gas Modus and the data visualization of the training as 3-Dimensional visualization via torus representation

Figure 3c shows the activity card, where the lattices are followed in real time. The second screen is the development of the network and shows the data point of the nodes which are also evolving in time: the red points are clusters or nodes representing products of enzyme catalysis, and the green lights are the references neurons where the pathway is being developed. The yellow is the reference input vector. In Cell Communication Protocol there is an option for changing the positions to alter the graph, the direction and the catalysis, all in real time.

Figure 3d shows the data representing the evolving network in two-dimensions; each cell is represented by the intensity of the color of the nodes, clusters and chemical reactions in the development of the metabolic pathway. In three-dimensions, the increased data improves the rate of accuracy. The 3-D visualization in Cell Communication Protocol is a torus, and the changes of color during the simulation gives a better handle of the data points of the metabolic pathway [27]. When the edges of the cells are joining, the topology is of the torus.

For the neural gas approach, an input vector is introduced to the Cell Communication Protocol, and similar to SOMs, the learning parameters are calculated from the input vector (see Fig. 3a) [28]. The input vector parameter is the density of the enzyme as per the previous version. SOMs are under constraints [29] of a predefined neighborhood between the neurons and the distances between the vectors that are not directly represented in the map [9, 30]. The output space is not defined in neural gas and the neural gas algorithm is based on a stochastic gradient decent cost function and is connected to a Hebbian competitive rule which forms the connections between the neighborhoods [31].

Figure 3b shows the neural gas data representation in two-dimensions. At the base, the neural gas prototype clusters move around in the data space similar to the Brownian movement of gas molecule in a closed container. The visualization shows the data structures, clusters, and outliers in 2-dimensions as a line and in 3-dimensions as an iris [32,33,34].

The back propagation is included in Cell Communication Protocol© to better mimic reverse catalysis of real biological metabolic pathways which are not considered by Michaelis-Menten Kinetics [35, 36].

Figure 4a shows the learning rule, which considers the desired output for any given input, and the momentum term which is a component that dampens the oscillations of the learning rule to provide direction. Figure 4b shows the overall error, where the rate of learning is given by the probability (y-axis) and the interactions in the x-axis.

Fig. 4
figure 4

a The classification for back propagation in process of learn in the data base. b The error versus the interactions and the error having a solution in the back propagation modus

5 Discussion

The particle diffusion (i.e. mobility) impacts the reaction of the enzyme kinetics. Collision theory gives the enzyme catalysis reaction, and the bimolecular reaction can be expressed as collision [37]. The original Stokes-Einstein diffusion equation was developed for small particles, but can also be used to predict for sizes up to micrometer [38,39,40]. The diffusion equation is a partial equation and used in concentrations where viscous forces are dominated by low Reynolds numbers [41]. From the low Reynolds number, the particle moves at a velocity proportional to the force applied:

$$v_{s} = \sigma F$$
(4)

where, \(\sigma\) is the proportionality constant (\(\sigma \propto {1 \mathord{\left/ {\vphantom {1 {r\eta }}} \right. \kern-0pt} {r\eta }}\)).

$$v_{s} = \sigma \frac{\text{Re}}{\rho }$$
(5)

where, \(\text{Re}\) is the Reynolds number, \(\eta\) is the dynamical viscosity, \(\rho\) is density, and \(r\) is the radius of the particle.

The diffusion coefficient is given by:

$$D = \frac{{k_{B} T}}{\sigma }$$
(6)

where \(k_{B}\) is the Boltzmann constant and \(T\) is temperature.

From collision theory we can be relate to Michaelis-Menten Kinetics.

The limitation is given by the Michaelis-Menten kinetics:

$$v = \frac{d\left[ p \right]}{dt} = \frac{{V_{\hbox{max} } \left[ S \right]}}{{K{}_{m} + \left[ S \right]}} = \frac{D\left[ S \right]}{{K_{m} + \left[ S \right]}}$$
(7)

where \(V_{\hbox{max} }\) is the maximum rate archive by the system, \(S\) is concentration of the substrate and \(K_{m}\) is the Michael constant, which is half of the \(V_{\hbox{max} }\).

The initial weights (e.g. by Michaelis-Menten kinetics) for the SOM, produces a graph for a particular network, where each cell shows a catalyst reaction, Fig. 5a.

Fig. 5
figure 5figure 5

a Training set of the data bases in the SOM modus, the yellow point shows the best solution and red, model feature points. b The interaction that has a solution in the data base in the SOM modus, given the distances to solution at 0.4228, the winner took 49 interaction and the location of the best solution. c The output of the data base in the SOM modus, and solution is to be finding in the upper part of the simulation. d The Neural Gas Modus in a 2-Dimesion data representation with 13 interactions to obtain the best solution, with coordinates locating the solution. e The back propagation output neuron with a solution of the data base, the x coordinated shows the interaction and the y axis the error approximation

Figure 5b shows the winner neuron and its distance (i.e. the initial catalysis reaction), and also shows how the SOM has learnt, as well as how far the last reaction is from the initial catalysis reaction [42]. Figure 5c shows the actual reaction with the red being the initial input vector and the white marks are the number of the neighboring neurons to the reference neuron and their arrangement.

Figure 5d is the same representation for neural gas, where X-Pos are the reference neuron, and where the weight is from Michaelis-Menten kinetics. The winner cells took 13 iterations from the initial catalysis.

Figure 5e shows the iteration and the probability error of the back propagation. The Michaelis-Menten Kinetics weights are derived from the relationship between the Probability of diffusion and Michaelis-Menten Kinetics. The probability is obtained from Ficks First Law.

For proofing the theoretical assumptions above we used pathway data sets from The Kyoto Encyclopaedias of Genes and Genomes (KEGG) [43, 44] and trained the SOM with these data. We selected two different features of the pathway data. The first feature vector coded the number and distances of the pathways [45] and the second vector the metabolism characterized by the chemical substances involved and their concentrations. In both cases we expect that the trained neural classifier would be able to separate/identify different path-ways constellations if data sets from different experiments were presented to the neural net [46]. The theory of the neural nets (and therefore for every kind of SOM) every constellation of measured data can be combined (used) to form an input or training vector [47].

Figure 6a shows an untrained trained SOM with 100 neurons (green crosses) in a 2 dimensional feature spaced which was normalized in an [0, 1] interval. This normalization is used very often, to prevent numerical overflows in the calculation during the training phase of the SOM. The shown 10 red point in Fig. 6a represent the constellation of a pathway metabolism.

Fig. 6
figure 6figure 6

a Structure of a so far untrained SOM, of the data base distances between the data is between 0 and 1 and the best solution is the yellow point, the green points are data distribution and the red point’s model feature points. b Trained SOM, shows smooth distribution of the data in green, with the red points are model feature points. c Evaluation of similarity/identical structure of variable data set by a trained SOM. d Not sufficient trained SOM, the distribution of data and the data is not smooth, the red point, are feature points with irregular distribution. e Sufficient trained SOM, the green points a smooth data and the red point shows minimize model feature points. f Common and modified visualisation of the activity of a trained SOM, the two Dimensions data representations shows, the data distribution and the smoothness of the data

After 15,000 training iterations (requiring about 4 min of computational time) the green neurons adapted to the model noted by localization of the green neurons near the red pathway model feature points as shown in Fig. 6b.

In Fig. 6c similar data sets from other measurements has been presented to the net. As can be pointed out, only one of them met the classifiers structure (the yellow neuron in the middle of Fig. 6c), the other data sets—represented by the 2 white and the 10 purple neurons—present similar but not the same structure as the trained ones.

One of the major advantages of neural nets as be able to chosen method allows defining (e.g. out of theoretical consideration) an arbitrary input vector which can be validated by SOM regarding its fitness [48]. Occasionally, especially if the feature vector and/or the model was more or less complicated, it was be necessary to enlarge the number of neurons of the SOM and/or the number of training cycles [49]. This case is shown in Fig. 6d and e.

Figure 6d shows—due to the fact that now 44 features (red points (model feature points) in Fig. 6d and e) represent the model—the SOM based classifier now bears 400 neurons, but even after 5.426 training cycles the net structure has not adapted the red points (the system forming the classifier).

It took 7,400 training cycles more until the net structure fits the system structure, as shown in Fig. 6e.

Neural nets should not be relied upon for automatic supervision and deciding [50] if a net structure fits or if the model or system will adapt. But, this model allows a visualisation routine to aid the user to find and define a satisfactory decision if a neural net met a desired classification or not [51].

Figure 6f shows the common visualisation of a SOM activity. Here the so called winner neuron (the yellow neuron, left hand side to the SOM activity card) is represented in orange, limited by a black square at the top. The activity of the other neuron—from light orange over yellow to dark green are surrounding the winner, but—and this is most important for our discussing—no information about similarity or identity can by pointed out by this visualisation. So it is the special way of visualisation we discuss here, which enables a deeper model analysis and/or competition of measured data set.

6 Conclusion

Metabolic pathways are difficult simulated with Michaelis-Menten kinetics that do not consider the reverse catalysis. By using a neural network, Self-Organizing Maps, and back propagation, this methodology can overcome the Michaelis-Menten limitations. Self-Organizing Maps are used to graphically describe the catalysis reaction, with the constraint of the weight space being overcome by the neural gas algorithms [52,53,54,55,56], and by doing so achieves a visual display of the catalysis and expands the weight space to improve the accuracy of the modeled reaction [57,58,59,60]. With the use of the back propagation the model can overcome the inherent limitation of Michaelis-Menten, and the Cell Communication Protocol© provides a set of solutions for a descriptive and analytical analysis of metabolic pathways.