Keywords

1 Introduction

A challenging issue for dynamic models is to predict what patterns of behaviour will emerge, and how their emergence depends on the structure of the model, including chosen values for model characteristics or parameters. This applies in particular to network models, where behaviour depends in some way on the network structure, defined by network characteristics such as connections and their weights. It can be an even more challenging issue when adaptive networks are considered, where the network characteristics also change over time, according to certain adaptation principles which themselves depend on certain adaptation characteristics represented by their own particular parameters. It is this latter issue what is the topic of the current paper: how does emerging behaviour of adaptive networks relate to the characteristics of the network and of the adaptation principles used. More in particular, the focus is on adaptive mental networks based on Hebbian learning [1, 3, 4, 6,7,8]. Hebbian learning is, roughly stated, based on the principle ‘neurons that fire together, wire together’ from Neuroscience.

To address the issue, as a vehicle the Network-Oriented Modeling approach based on temporal-causal networks [10] will be used. For temporal-causal networks, parameters characterising the network structure are connection weights, combination functions and speed factors. For the type of adaptive networks considered, the connection weights are dynamic based on Hebbian learning, so they are not part of the characteristics of the network structure anymore. Instead, characteristics of Hebbian learning have been identified that play an important role. In this paper, results will be discussed that have been proven mathematically for this relation between structure and behavior for such adaptive network models, in particular, for the result of Hebbian learning in relation to the connected network states. These results have been proven not for one specific model or function, but for classes of functions that fulfill certain properties. More specifically, it has been found how for the classes of functions considered within an emerging equilibrium state the connection weight and the connected states satisfy a fixed functional relation that can be expressed mathematically.

In this paper, in Sect. 2 the temporal-causal networks that are used as vehicle are briefly introduced. In Sect. 3 the properties of Hebbian learning functions are introduced that define the adaptation principle of the network. Section 4 focuses in particular on the class of functions for which a form of variable separation can be applied, In Sect. 5 a number of examples are discussed. Finally, Sect. 6 is a discussion.

2 Temporal-Causal Networks

For the perspective on networks used in the current paper, the interpretation of connections based on causality and dynamics forms a basis of the structure and semantics of the considered networks. More specifically, the nodes in a network are interpreted here as states (or state variables) that vary over time, and the connections are interpreted as causal relations that define how each state can affect other states over time. This type of network has been called a temporal-causal network [10]. A conceptual representation of a temporal-causal network model by a labeled graph provides a fundamental basis. Such a conceptual representation includes representing in a declarative manner states and connections between them that represent (causal) impacts of states on each other. This part of a conceptual representation is often depicted in a conceptual picture by a graph with nodes and directed connections. However, a complete conceptual representation of a temporal-causal network model also includes a number of labels for such a graph. A notion of strength of a connection is used as a label for connections, some way to aggregate multiple causal impacts on a state is used, and a notion of speed of change of a state is used for timing of the processes. These three notions, called connection weight, combination function, and speed factor, make the graph of states and connections a labeled graph (e.g., see Fig. 1), and form the defining structure of a temporal-causal network model in the form of a conceptual representation; see Table 1, first 5 rows.

Fig. 1.
figure 1

An adaptive temporal-causal network model for adaptive decision making.

Table 1. Concepts of conceptual and numerical representations of a temporal-causal network.

There are many different approaches possible to address the issue of combining multiple impacts. To provide sufficient flexibility, the Network-Oriented Modelling approach based on temporal-causal networks incorporates for each state a way to specify how multiple causal impacts on this state are aggregated by a combination function. For this aggregation a library with a number of standard combination functions are available as options, but also own-defined functions can be added.

Next, this conceptual interpretation is expressed in a formal-numerical way, thus associating semantics to any temporal-causal network specification in a detailed numerical-mathematically defined manner.

This is done by showing how a conceptual representation based on states and connections enriched with labels for connection weights, combination functions and speed factors, can get an associated numerical representation [10], Ch. 2; see Table 1, last five rows. The difference equations in the last row in Table 1 constitute the numerical representation of the temporal-causal network model and can be used for simulation and mathematical analysis; it can also be written in differential equation format:

$$ \begin{aligned} Y(t + \Delta t) = Y(t) + {\varvec{\upeta}}_{Y} [{\mathbf{c}}_{Y} ({\varvec{\upomega}}_{{X_{1} ,Y}} X_{1} (t), \ldots ,{\varvec{\upomega}}_{{X_{k} ,Y}} X_{k} (t)) - Y(t)]\Delta t \hfill \\ {\mathbf{d}}Y(t)/{\mathbf{d}}(t) = {\varvec{\upeta}}_{Y} [{\mathbf{c}}_{Y} ({\varvec{\upomega}}_{{X_{1} ,Y}} X_{1} (t), \ldots ,{\varvec{\upomega}}_{{X_{k} ,Y}} X_{k} (t)) - Y(t)] \hfill \\ \end{aligned} $$
(1)

In adaptive networks connection weights ω are treated in the same way as states, and are defined by combination functions cω(…) in a similar manner (with suitable arguments refering to relevant states and connection weights):

$$ \begin{aligned}\upomega(t + \Delta t) \, = \upomega \left( t \right) \, + {\varvec{\upeta}}_{\upomega} [{\mathbf{c}}_{\upomega} \left( \ldots \right) -\upomega\left( t \right)]\Delta t \hfill \\ {\mathbf{d}}\upomega\left( t \right)/{\mathbf{d}}t = {\varvec{\upeta}}_{\upomega} [{\mathbf{c}}_{\upomega} \left( \ldots \right) -\upomega\left( t \right)] \hfill \\ \end{aligned} $$
(2)

3 Adaptive Networks Based on Hebbian Learning

In this section it is discussed how specific combination functions for Hebbian learning can be defined, and it will be analysed what equilibrium values can emerge for the learnt connections. First a basic definition; see also [2, 5, 9].

Definition 1 (stationary point and equilibrium)

A state Y or connection weight ω has a stationary point at t if dY(t)/dt = 0 or dω(t)/dt. The network is in equilibrium a t if every state Y and connection weight of the model has a stationary point at t. A state Y has is increasing at t if dY(t)/dt > 0; it is decreasing at t if dY(t)/dt < 0. Similar for adaptive connections based on dω(t)/dt.

Considering the specific type of differential equation for a temporal-causal network model, and assuming a nonzero speed factor, from (1) and (2) more specific criteria can be found:

Lemma 1 (Criteria for a stationary, increasing and decreasing)

Let Y be a state and X1, …, Xk the states with outgoing connections to state Y. Then

$$ \eqalign{&Y\text{ has a stationary point at }t &\Leftrightarrow& {\mathbf{c}}_{Y} (\upomega_{{{X}_{_{1}} ,Y}} X_{1} (t),\ldots ,\upomega_{{{X_{k} ,Y}}} X_{k} (t))\, = \,Y(t)\\ &Y\text{ is increasing at }t &\Leftrightarrow& {\mathbf{c}}_{Y} (\upomega_{{{X}_{_{1}} ,Y}} X_{1} (t),\ldots ,\upomega_{{{X_{k} ,Y}}} X_{k} (t))\, > \,Y(t)\\ &Y\text{ is decreasing at }t &\Leftrightarrow& {\mathbf{c}}_{Y} (\upomega_{{{X}_{_{1}} ,Y}} X_{1} (t),\ldots ,\upomega_{{{X_{k} ,Y}}} X_{k} (t))\, < \,Y(t)}$$

Similar criteria are applied to adaptive connection weights:

$$ \eqalign{&\upomega\text{ has a stationary point at }t &\Leftrightarrow& {\mathbf{c}}_{\upomega}(\ldots)\, = \,\upomega(t)\\ & \upomega\text{ is increasing at }t &\Leftrightarrow& {\mathbf{c}}_{\upomega}(\ldots)\, > \,\upomega(t)\\ &\upomega\text{ is decreasing at }t &\Leftrightarrow& {\mathbf{c}}_{\upomega}(\ldots)\, < \,\upomega(t)}$$

The Hebbian learning principle for the connection between two mental states is sometimes formulated as ‘neurons that fire together, wire together’; e.g., [1, 3, 4, 6,7,8, 11]. This is modelled by using the activation values the two mental states X(t) and Y(t) have at time t. Then the weight ωX,Y of the connection from X to Y is changing over time dynamically, depending on these levels X(t) and Y(t). As this connection weight is dynamic, following the Network-Oriented Modeling approach outlined in Sect. 2 it is handled as a state with its own combination function \( \text{c}_{{\upomega _{{X,Y}} }} (\text{V}_{1} ,{\text{V}}_{2},) \), and using the standard difference and differential equation format as shown in (2) in Sect. 2

$$ \begin{aligned} &\upomega_{X,Y} (t + \Delta t) =\upomega_{X,Y} (t) +\upeta_{{\upomega_{X,Y}}} [{\rm c}_{{\upomega_{X,Y}}} (X(t),Y(t),\upomega_{X,Y} (t)) -\upomega_{X,Y} (t)]\Delta t \\ & {\mathbf{d}}\upomega_{X,Y} /{\mathbf{d}}(t) =\upeta_{{\upomega_{X,Y}}} [{\rm c}_{{\upomega_{X,Y}}} (X,Y,\upomega_{X,Y} ) -\upomega_{X,Y} ] \\ \end{aligned} $$
(3)

The parameter \(\upeta_{{\upomega_{X,Y}}}\) is the speed parameter of connection weight ωX,Y, in this case interpreted as learning rate. Note that by the above criteria ωX,Y increases if and only if \(\text{c}_{{\upomega_{X,Y}}}\)(X, Y, ωX,Y) > ωX,Y, and ωX,Y decreases if and only if \(\text{c}_{{\upomega_{X,Y}}}\)(X, Y, ωX,Y) < ωX,Y, and ωX,Y is stationary if and only if \(\text{c}_{{\upomega_{X,Y}}}\)(X, Y, ωX,Y) = ωX,Y.

An example of a mental network model using Hebbian learning is shown in Fig. 1 (adopted from [10], Ch 6, p. 163). It describes adaptive decision making as affected by direct triggering of decision options ai (via weights ω1,i) in combination with emotion-related valuing of the options by an as-if prediction loop (via weights ω3,i and ω2,i). For the weights of the adaptive connections the bending arrows show that they are affected by the states they connect. Here wsw are world states, ssw sensor states, srsw and \( {\text{srs}}_{{e_{i} }}\) sensory representations states for stimulus w and action effect ei, \( {\text{ps}}_{{a_{i} }}\) preparation states for ai, \( {\text{fs}}_{{e_{i} }}\) feeling states for action effect ei, and \( {\text{es}}_{{a_{i} }}\) execution states for ai. A relatively simple example, also used in [10] in a number of applications (including in Ch 6 for the model shown in Fig. 1) is the following combination function:

$$ \eqalign{ & {{\text{c}}_{\upomega_{X,Y}}}({V_1},{V_2},W) = {V_1}{V_2}(1 - W) + \upmu W \cr & {\text{or}}\,\,\,\,{{\text{c}}_{\upomega_{X,Y}}}(X(t),Y(t),{\upomega_{X,Y}}(t)) = X(t)Y(t)(1 - {\upomega_{X, Y}}(t)) + \upmu {\upomega_{X,Y}}(t) \cr} $$
(4)

Here μ is a persistence parameter. In an emerging equilibrium state it turns out that the equilibrium value for ωX,Y functionally depends on the equilibrium values of X and Y according to some formula that has been determined for this case in [10], Ch 12. For some example patterns, see Fig. 2.

It is shown that when the equilibrium values of X and Y are 1, the equilibrium value for ωX,Y is 0.83 (top row), when the equilibrium values of X and Y are 0.6, the equilibrium value for ωX,Y is 0.64 (middle row), and when the equilibrium values of X and Y are 0, the equilibrium value for ωX,Y is 0 (bottom row). This equilibrium value of ωX,Y is always attracting. The three different rows in Fig. 1 illustrate how the equilibrium value of ωX,Y varies with the equilibrium values of X and Y. It is this relation that is analysed in a more general setting in some depth in this paper. In Example 1 in Sect. 5 below, this case is analysed and more precise numbers will be derived for the equilibrium values. In [10], Ch. 12 a mathematical analysis was made for the equilibria of the specific example combination function above. In the current paper a much more general analysis is made which applies to a wide class of functions.

The following plausible assumptions are made for a Hebbian learning function: one set for fully persistent Hebbian learning and one set for Hebbian learning with extinction described by a persistence parameter μ; here V1 is the argument of the function cωX,Y(..) used for X(t), V2 for Y(t), and W for ωX,Y(t).

Definition 2 (Hebbian learning function)

A function c: [0, 1] x [0, 1] x [0, 1] → [0, 1] is called a fully persistent Hebbian learning function if the following hold:

  1. (a)

    c(V1, V2, W) is a monotonically increasing function of V1 and V2

  2. (b)

    c(V1, V2, W) − W is a monotonically decreasing function of W

  3. (c)

    c(V1, V2, W) ≥ W

  4. (d)

    c(V1, V2, W) = W if and only if one of V1 and V2 is 0 (or both), or W = 1

A function c: [0, 1] x [0, 1] x [0, 1] → [0, 1] is called a Hebbian learning function with persistence parameter μ if the following hold:

  1. (a)

    c(V1, V2, W) is a monotonically increasing function of V1 and V2

  2. (b)

    c(V1, V2, W) - μW is a monotonically decreasing function of W

  3. (c)

    c(V1, V2, W) ≥ μW

  4. (d)

    c(V1, V2, W) = μW if and only if one of V1 and V2 is 0 (or both), or W = 1

Note that for μ = 1 the function is fully persistent. The following proposition shows that for any Hebbian learning function with persistence parameter μ there exists a monotonically increasing function fμ(V1, V2) which is implicitly defined for given V1, V2 by the equation \({{\text{c}}_{\upomega_{X,Y}}}\)(V1, V2, W) = W in W. When applied to an equilibrium state of an adaptive temporal-causal network, the existence of this function fμ(V1, V2) reveals that in equilibrium states there is a direct and monotonically increasing functional relation of the equilibrium value ωX,Y of ωX,Y with the equilibrium values X, Y of the states X and Y. This is described in Theorem 1 below. Proposition 1 describes the functional relation needed for that. For proofs of Propositions 1 and 2, see the Appendix.

Proposition 1 (functional relation for W )

Suppose that c(V1, V2, W) is a Hebbian learning function with persistence parameter μ.

(a) Suppose μ < 1. Then the following hold:

(i) The function W → c(V1, V2, W) − W on [0, 1] is strictly monotonically decreasing

(ii) There is a unique function fμ: [0, 1] x [0, 1] → [0, 1] such for any V1, V2 it holds

$$ {\text{c}}(V_{1}, \, V_{2}, \, {{\text{f}}_\upmu }(V_{1}, \, V_{2})) = {{\text{f}}_\upmu }(V_{1}, \, V_{2}) $$

This function fμ is a monotonically increasing function of V1, V2, and is implicitly defined by the above equation. Its maximal value is fμ(1, 1) and minimum fμ(0, 0) = 0.

(b) Suppose μ = 1. Then there is a unique function f1: (0, 1] x (0, 1] → [0, 1], such for any V1, V2 it holds

$$ {\text{c}}(V_{1} , \, V_{2} ,{\text{ f}}_{1} (V_{1} , \, V_{2} )) = {\text{f}}_{1} (V_{1} , \, V_{2} ) $$

This function f1 is a constant function of V1, V2 with f1(V1, V2) = 1 for all V1, V2 > 0 and is implicitly defined on (0, 1] x (0, 1] by the above equation.

If one of V1, V2 is 0, then any value of W satisfies the equation c(V1, V2, W) = W, so no unique function value for f1(V1, V2) can be defined then.

When applied to an equilibrium state of an adaptive temporal-causal network, this proposition entails the following Theorem 1. For μ < 1 this follows from Proposition 1a) applied to the function \({{\text{c}}_{\upomega_{X,Y}}}\)(.). From (a)(i) it follows that the equilibrium value is attracting: suppose ω(t) < ωX,Y, then from \({{\text{c}}_{\upomega_{X,Y}}}\)(X, Y, ωX,Y) − ωX,Y = 0 and the decreasing monotonicity of W → c(V1, V2, W) − W it follows that \({{\text{c}}_{\upomega_{X,Y}}}\)(X, Y, ω(t)) − ω(t) > 0, and therefore by Lemma 1 ω(t) is increasing. Similarly, when ω(t) > ωX,Y, it is decreasing.

For μ = 1 the statement follows from Proposition 1b) applied to the function \({{\text{c}}_{\upomega_{X,Y}}}\)(.).

Theorem 1 (functional relation for equilibrium values of ω X,Y )

Suppose in a temporal-causal network \({{\text{c}}_{\upomega_{X,Y}}}\)(V1, V2, W) is the combination function for connection weight ωX,Y and is a Hebbian learning function with persistence parameter μ, with fμ the function defined by Proposition 1. In an achieved equilibrium state the following hold.

  1. (a)

    Suppose μ < 1. For any equilibrium values X, Y ∈ [0, 1] of states X and Y the value fμ(X, Y) provides the unique equilibrium value ωX,Y for ωX,Y. This ωX,Y monotonically depends on X, Y: it is higher when X, Y are higher. The maximal equilibrium value ωX,Y of ωX,Y is fμ(1, 1) and the minimal equilibrium value is 0. Moreover, the equilibrium value ωX,Y is attracting.

  2. (b)

    Suppose μ = 1. If for the equilibrium values X, Y ∈ [0, 1] of states X and Y it holds X, Y > 0, then ωX,Y = 1. If one of X, Y is 0, then ωX,Y can be any value in [0, 1]: it does not depend on X, Y. So, for μ = 1 the maximal value of ωX,Y in an equilibrium state is 1 and the minimal value is 0.

4 Variable Separation for Hebbian Learning Functions

There is a specific subclass of Hebbian learning functions that is often used. Relatively simple functions c(V1, V2, W) that satisfy the requirements from Definition 2 are obtained when the arguments V1 and V2 and W can be separated as follows.

Definition 3 (variable separation)

The Hebbian learning function c(V1, V2, W) with persistence parameter μ enables variable separation by functions cs: [0, 1] x [0, 1] → [0, 1] monotonically increasing and cs: [0, 1] → [0, 1] monotonically decreasing if

$$ {\text{c}}(V_{1}, \, V_{2}, \, W) \, = {\text{ cs}}(V_{1}, \, V_{2}){\text{ cc}}\left( W \right) \, + \upmu W $$

where cs(V1, V2) = 0 if and only if one of V1, V2 is 0, and cc(1) = 0 and cc(W) > 0 when W < 1

Note that the s in cs stands for states and the second c in cc for connection. When variable separation holds, the following proposition can be obtained. For this type of function the indicated functional relation can be defined.

Proposition 2 (functional relation for W based on variable separation)

Assume the Hebbian function c(V1, V2, W) with persistence parameter μ enables variable separation by the two functions cs(V1, V2) monotonically increasing and cc(W) monotonically decreasing:

$$ {\text{c}}(V_{1}, \, V_{2}, \, W) \, = {\text{ cs}}(V_{1}, \, V_{2}){\text{ cc}}\left( W \right) \, + \upmu W $$

Let hμ(W) be the function defined for W ∈ [0, 1) by

$$ {\text{h}}_{\upmu} \left( W \right) \, = \, ( 1-\upmu)W/{\text{cc}}\left( W \right) $$

Then the following hold.

  1. (a)

    When μ < 1 the function hμ(W) is strictly monotonically increasing, and has a strictly monotonically increasing inverse gμ on the range hμ([0, 1)) of hμ with W = gμ(hμ(W)) for all W∈ [0, 1).

  2. (b)

    When μ < 1 and c(V1, V2, W) = W, then gμ(cs(V1, V2)) < 1 and W < 1, and it holds

    $$ \begin{aligned} {\text{h}}_{\upmu} \left( W \right) \, = {\text{ cs}}(V_{ 1} , \, V_{ 2} ) \hfill \\ W = {\text{ g}}_{\upmu} ({\text{cs}}(V_{ 1} , \, V_{ 2} )) \hfill \\ \end{aligned} $$

    So, in this case the function fμ from Theorem 1 is the function composition gμ o cs of cs followed by gμ; it holds:

    $$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = {\text{ g}}_{\upmu} ({\text{cs}}(V_{ 1} , \, V_{ 2} )) $$
  3. (c)

    For μ = 1 it holds c(V1, V2, W) = W if and only if V1 = 0 or V2 = 0 or W = 1.

  4. (d)

    For μ < 1 the maximal value W with c(V1, V2, W) = W is gμ(cs(1, 1)), and the minimal equilibrium value W is 0. For μ = 1 the maximal value W is 1 (always when V1, V2 > 0 holds) and the minimal value is 0 (occurs when one of V1, V2 is 0).

Note that by Proposition 2 the function fμ(V1, V2) can be determined by inverting the function hμ(W) = (1 − μ)W/cc(W) to find gμ and composing the inverse with the function cs(V1, V2). This will be shown below for some cases. For the case of an equilibrium state of an adaptive temporal network model Proposition 2 entails Theorem 2.

Theorem 2 (functional relation for equilibrium values of ω X,Y : variable separation)

Assume in a temporal-causal network the Hebbian learning combination function \({{\text{c}}_{\upomega_{X,Y}}}\)(V1, V2, W) with persistence parameter μ for ωX,Y enables variable separation by the two functions \({{\text{cs}}_{\upomega_{X,Y}}}\)(V1, V2) monotonically increasing and \({{\text{cc}}_{\upomega_{X,Y}}}\)(W) monotonically decreasing, and the functions fμ and gμ are defined as in Propositions 1 and 2. Then the following hold.

  1. (a)

    When μ < 1 in an achieved equilibrium state with equilibrium values X, Y for states X and Y and ωX,Y for ωX,Y it holds

    $$ \underline\upomega_{X,Y} = {\text{ f}}_{\upmu} (\underline{X}, \underline{Y}) \, = {\text{ g}}_{\upmu} ({{\text{cs}}_{\upomega_{X,Y}}} ({\underline{X}},{\underline{Y}})) \, < { 1} $$
  2. (b)

    For μ = 1 in an equilibrium state with equilibrium values X, Y for states X and Y and ωX,Y for ωX,Y it holds X = 0 or Y = 0 or ωX,Y = 1.

  3. (c)

    For μ < 1 in an equilibrium state the maximal equilibrium value ωX,Y for ωX,Y is gμ(\({{\text{cs}}_{\upomega_{X,Y}}}\)(1, 1)) < 1, and the minimal equilibrium value ωX,Y is 0. For μ = 1 the maximal value is 1 (always when X, Y > 0 holds for the equilibrium values for the states X and Y) and the minimal value is 0 (which occurs when one of X, Y is 0

5 Analysis of Different Cases of Hebbian Learning Functions

In this section some cases are analysed as corollaries of Theorem 2. First the specific class of Hebbian learning functions enabling variable separation with cc(W) = 1−W is considered. Then

$$ {\text{h}}_{\upmu} \left( W \right) \, = \, ( 1-\upmu)W/{\text{cc}}\left( W \right) \, = \, ( 1-\upmu)W/\left( { 1- W} \right) $$
(5)

and the inverse gμ(W′) of hμ(W) can be determined from (4) algebraically as follows.

$$ \begin{aligned} W' = \, ( 1-\upmu)W/\left( { 1- W} \right) \hfill \\ W'\left( { 1- W} \right) \, = \, ( 1-\upmu)W \hfill \\ W' - W'W = \, ( 1-\upmu)W \hfill \\ W' = (W' \, + ( 1-\upmu))W \hfill \\ W = W' \, /[W' \, + ( 1-\upmu)] \hfill \\ \end{aligned} $$

So

$$ {\text{g}}_{\upmu} \left( {W'} \right) \, = W' \, /[W' \, + ( 1-\upmu)] $$
(6)

Substitute W′  = cs(V1, V2) in (6) and it is obtained:

$$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = {\text{ g}}_{\upmu} \left( {{\text{cs}}\left( {V_{ 1} , \, V_{ 2} } \right)} \right) \, = {\text{ cs}}\left( {V_{ 1} , \, V_{ 2} } \right)/[( 1-\upmu) \, + {\text{ cs}}\left( {V_{ 1} , \, V_{ 2} } \right)] $$
(7)

and this is less than 1 because 1 − μ > 0. From this and Theorem 2b) and c) it follows.

Corollary 1 (cases for function \({{\mathbf{cc}}_{\boldsymbol{\upomega}_{X,Y}}}\) ( W ) = 1 – W )

Assume in a temporal-causal network the Hebbian learning combination function \({{\text{c}}_{\upomega_{X,Y}}}\)(V1, V2, W) for ωX,Y with persistence parameter μ enables variable separation by the two functions \({{\text{cs}}_{\upomega_{X,Y}}}\)(V1, V2) monotonically increasing and \({{\text{cc}}_{\upomega_{X,Y}}}\)(W) monotonically decreasing, where \({{\text{cc}}_{\upomega_{X,Y}}}\)(W) = 1 – W. Then the following hold.

  1. (a)

    When μ < 1 in an equilibrium state with equilibrium values X, Y for states X and Y and ωX,Y for ωX,Y it holds

    $$ \upomega_{X,Y} = {\text{ f}}_{\upmu} ({\text{X}},{\text{Y}}) \, = {\text{ cs}}({\text{X}},{\text{Y}})/[( 1-\upmu) \, + {\text{ cs}}({\text{X}},{\text{Y}})] \, < { 1} $$
  2. (b)

    For μ = 1 in an equilibrium state with equilibrium values X, Y for states X and Y and ωX,Y for ωX,Y it holds X = 0 or Y = 0 or ωX,Y = 1.

  3. (c)

    For μ < 1 in an equilibrium state the maximal equilibrium value ωX,Y for ωX,Y is

    $$ {\text{cs}}\left( { 1,{ 1}} \right)/[( 1-\upmu) \, + {\text{ cs}}\left( { 1,{ 1}} \right)] \, < { 1} $$

    and the minimal equilibrium value ωX,Y is 0. For μ = 1 the maximal value is 1 (when X, Y > 0 holds for the equilibrium values for the states X and Y) and the minimal value is 0 (which occurs when one of X, Y is 0).

Corollary 1 is illustrated in the following three examples.

Example 1.

\( {\text{c}}(V_{1}, \, V_{2}, \, W) \, = {V_{1}}{V_{2}}\left( { 1- W} \right) \, + \upmu W \)

$$ {\text{cs}}(V_{ 1} , \, V_{ 2} ) \, = V_{ 1} V_{ 2} \,\,\,\,\,\,\,{\text{cc}}\left( W \right) \, = { 1} - W $$

This is the example shown in Fig. 2

Fig. 2.
figure 2

Hebbian learning η = 0.4, μ = 0.8, ∆t = 0.1; adopted from [10], pp. 339–340. a) Top row: activation levels X1 = 1 and X2 = 1; equilibrium value 0.83 b) Middle row activation levels X1 = 0.6 and X2 = 0.6; equilibrium value 0.64 c) Bottom row: activation levels X1 = X2 = 0; equilibrium value 0 (pure extinction)

$$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = {\text{ cs}}\left( {V_{ 1} , \, V_{ 2} } \right)/[( 1-\upmu) \, + {\text{ cs}}\left( {V_{ 1} , \, V_{ 2} } \right)] $$
(8)

Substitute cs(V1, V2) = V1 V2 in (7) then fμ(V1, V2) = V1 V2 / [(1−μ) + V1 V2].

Maximal W is W = fμ(1, 1) = 1/[2 − μ], which for μ = 1 is 1; minimal W is 0. The equilibrium values shown in Fig. 2 can immediately derived from this (recall μ = 0.8):

Top row V1 = 1, V2 = 1, then fμ(1, 1) = 1/[2 − μ] = 0.833333

Middle row V1 = 0.6, V2 = 0.6, then fμ(0.6, 0.6) = 0.36 / [(1 − 0.8) +0.36] = 0.642857

Bottom row V1 = 0, V2 = 0, then fμ(0, 0) = 0

Example 2.

\( {\text{c}}(V_{1}, \, V_{2}, \, W) \, = \, (\sqrt {{V_{1}}{V_{2}}} )\left( { 1- W} \right) \, + \upmu W \)

$$ {\text{cs}}(V_{ 1} , \, V_{ 2} ) \, = \sqrt {{V_{ 1} V_{ 2} }} \, \,\,\,\,\,\,{\text{cc}}\left( W \right) \, = { 1} - W $$
$$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = {\text{ cs}}\left( {V_{ 1} , \, V_{ 2} } \right)/[( 1-\upmu) \, + {\text{ cs}}\left( {V_{ 1} , \, V_{ 2} } \right)] $$
(9)

Substitute cs(V1, V2) = √(V1V2) in (8) to obtain

$$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = \sqrt { {V_{ 1} V_{ 2} } } \, / \, [( 1-\upmu) + \sqrt {{V_{ 1} V_{ 2} }} ] $$
(10)

Maximal W is W = fμ(1, 1) = 1/[2 − μ], which for μ = 1 is 1; minimal W is 0.

In a similar case as in Fig. 2, but the using this function the following equilibrium values would be found

Top row V1 = 1, V2 = 1, then fμ(1, 1) = 1/[2 − μ] = 0.833333

Middle row V1 = 0.6, V2 = 0.6, then fμ(0.6, 0.6) = 0.6 / [(1 − 0.8) + 0.6] = 0.75

Bottom row V1 = 0, V2 = 0, then fμ(0, 0) = 0

Example 3.

\( {{\text{c}}_{\upomega_{X,Y}}}(V_ 1, \, V_ 2, \, W) \, = {V_ 1}{V_ 2}\left( {{V_ 1} + {V_ 2}} \right)\left( { 1- W} \right) \, + \upmu W \)

$$ {\text{cs}}_{{\upomega_{X,Y}}} (V_{ 1} , \, V_{ 2} ) \, = V_{ 1} V_{ 2} \left( {V_{ 1} + V_{ 2} } \right){\text{ cc}}_{{\upomega_{X,Y}}} \left( W \right) \, = { 1} - W $$
$$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = {\text{ cs}}\left( {V_{ 1} , \, V_{ 2} } \right)/[( 1-\upmu) \, + {\text{ cs}}\left( {V_{ 1} , \, V_{ 2} } \right)] $$
(11)

Substitute cs(V1, V2) = V1V2(V1 + V2) in (10) to obtain

$$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = V_{ 1} V_{ 2} \left( {V_{ 1} + V_{ 2} } \right)/ \, [( 1-\upmu) + \, V_{ 1} V_{ 2} \left( {V_{ 1} + V_{ 2} } \right)] $$
(12)

Maximal W is fμ(1, 1) = 2/[(1 − μ) + 2] = 2/[3 − μ], which for μ = 1 is 1; minimal W is 0

In a similar case as in Fig. 2, but the using this function the following equilibrium values would be found

Top row V1 = 1, V2 = 1, then fμ(1, 1) = 2/[3 − μ] = 0.909090

Middle row V1 = 0.6, V2 = 0.6, then fμ(0.6, 0.6) = 0.36 * 1.2/ [(1 − 0.8) + 0.36 * 1.2] = 0.632

Bottom row V1 = 0, V2 = 0, then fμ(0, 0) = 0

Next the specific class of Hebbian learning functions enabling variable separation with cc(W) = 1 − W2 is considered. Then

$$ {\text{h}}_{\upmu} \left( W \right) \, = \, ( 1-\upmu)W/{\text{cc}}\left( W \right) \, = \, ( 1-\upmu)W/\left( { 1- W^{ 2} } \right) $$
(13)

and the inverse of hμ can be determined algebraically as shown in Corollary 2. Inverting hμ(W) to get inverse gμ(W′) now can be done as follows:

$$ \begin{aligned} W' = \, ( 1-\upmu)W/\left( { 1- W^{ 2} } \right) \hfill \\ \left( { 1- W^{ 2} } \right)W' = \, ( 1-\upmu)W \hfill \\ - W' + \, ( 1-\upmu)W + \, W^{ 2} W' = \, 0 \hfill \\ \end{aligned} $$

This is a quadratic equation in W:

$$ W'W^{ 2} + \, ( 1-\upmu)W - W' = \, 0 $$
(14)

As W ≥ 0 the solution is

$$ W = \, ( - ( 1-\upmu) \, + \sqrt {(( 1-\upmu)^{ 2} + { 4}W'^{ 2})})/\left( { 2W'} \right) $$
(15)
$$ W = \, - ( 1-\upmu)/ 2W' + \sqrt {(( 1-\upmu)/ 2W')^{ 2} + { 1}} $$

So

$$ {\text{g}}_{\upmu} \left( {W'} \right) \, = \, - ( 1-\upmu)/ 2W' + \sqrt {(( 1-\upmu)/ 2W')^{ 2} + { 1}} $$
(16)

By substituting W′ = cs(V1, V2) it follows

$$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = {\text{ g}}_{\upmu} ({\text{cs}}(V_{ 1} , \, V_{ 2} )) \, = \, - ( 1-\upmu)/ 2 {\text{ cs}}(V_{ 1} , \, V_{ 2} ) \, + \sqrt {(( 1-\upmu)/ 2 {\text{ cs}}(V_{ 1} , \, V_{ 2} ))^{ 2} + { 1}} $$
(17)

All this is summarised in the following:

Corollary 2 (cases for function \({{\mathbf{cc}}_{\boldsymbol{\upomega}_{\mathbf{X,Y}}}}\) ( W ) = 1 – W 2 )

Assume in a temporal-causal network the Hebbian learning combination function \({{\text{c}}_{\upomega_{X,Y}}}\)(V1, V2, W) for ωX,Y with persistence parameter μ enables variable separation by the two functions \({{\text{cs}}_{\upomega_{X,Y}}}\)(V1, V2) monotonically increasing and \({{\text{cc}}_{\upomega_{X,Y}}}\)(W) monotonically decreasing, where \({{\text{cc}}_{\upomega_{X,Y}}}\)(W) = 1 – W2. Then the following hold.

  1. (a)

    When μ < 1 in an equilibrium state with equilibrium values X, Y for states X and Y and ωX,Y for ωX,Y it holds

    $$ \upomega_{X,Y} = {\text{ f}}_{\upmu} ({\underline{{\rm X}}},{\underline{{\rm Y}}}) \, = \, - ( 1-\upmu)/ 2 {\text{ cs}}({\underline{{\rm X}}},{\underline{{\rm Y}}}) \, + \sqrt {(( 1-\upmu)/ 2 {\text{ cs}}({\underline{{\rm X}}},{\underline{{\rm Y}}}))^{ 2} + { 1} } < { 1} $$
  2. (b)

    For μ = 1 in an equilibrium state with equilibrium values X, Y for states X and Y and ωX,Y for ωX,Y it holds X  = 0 or Y = 0 or ωX,Y =  1.

  3. (c)

    For μ < 1 in an equilibrium state the maximal equilibrium value ωX,Y for ωX,Y is

    $$ - ( 1-\upmu)/ 2 {\text{ cs}}\left( { 1,{ 1}} \right) \, + \sqrt {(( 1-\upmu)/ 2 {\text{ cs}}\left( { 1,{ 1}} \right))^{ 2} + { 1}} < { 1} $$

    and the minimal equilibrium value ωX,Y is 0. For μ = 1 the maximal value is 1 (when X, Y > 0 holds for the equilibrium values for the states X and Y) and the minimal value is 0 (which occurs when one of X, Y is 0).

Corollary 2 is illustrated in Example 4.

Example 4.

\( {{\text{c}}_{\upomega_{X,Y}}}(V_ 1, \, V_ 2, \, W) \, = {V_ 1}{V_ 2}\left( {{V_ 1} + {V_ 2}} \right)\left( { 1- {W^ 2}} \right) \, + \upmu W \)

$$ {\text{cs}}_{{\upomega_{X,Y}}} (V_{ 1} , \, V_{ 2} ) \, = V_{ 1} V_{ 2} (V_{ 1} + V_{ 2} ){\text{cc}}_{{\upomega_{X,Y}}} \left( W \right) \, = { 1} - W^{ 2} $$
$$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = \, - ( 1-\upmu)/ 2 {\text{ cs}}(V_{ 1} , \, V_{ 2} ) \, + \sqrt {(( 1-\upmu)/ 2 {\text{ cs}}(V_{ 1} , \, V_{ 2} ))^{ 2} + { 1}} $$
(18)

Substitute cs(V1, V2) = V1V2(V1 + V2)

$$ {\text{f}}_{\upmu} (V_{ 1} , \, V_{ 2} ) \, = \, - ( 1-\upmu)/ 2V_{ 1} V_{ 2} (V_{ 1} + V_{ 2} ) + \sqrt {(( 1-\upmu)/ 2V_{ 1} V_{ 2} (V_{ 1} + V_{ 2} ))^{ 2} + { 1}} $$
(19)

Maximal W is W = fμ(1, 1) = −(1 − μ)/4+ √(((1 − μ)/4)2 + 1) = [−(1 − μ)+ √((1 − μ)2 + 16)]/4 = 4/[(1 − μ) + √((1 − μ)2 + 16)], which for μ = 1 is 1; minimal W is 0. In a similar case as in Fig. 2, using this function the equilibrium values can be found by applying (18).

6 Discussion

In this paper it was analysed how emerging behaviour of an adaptive network can be related to characteristics of network structure and adaptation principles. In particular this was addressed for an adaptive mental network based on Hebbian learning [1, 3, 4, 6,7,8, 11]. To this end relevant properties of the functions defining the Hebbian adaptation principle have been identified. For different classes of functions emerging equilibrium values for the connection weight have been expressed as a function of the emerging equilibrium values of the connected states. The presented results do not concern results for just one type of network or function, as more often is found, but were formulated and proven at a more general level and therefore can be applied not just to specific networks but to classes of networks satisfying the identified relevant properties of network structure and adaptation characteristics.