Abstract
In this article we study the degree of approximation of multivariate pointwise and uniform convergences in the q-mean to the Fuzzy-Random unit operator of multivariate Fuzzy-Random Quasi-Interpolation arctangent, algebraic, Gudermannian and generalized symmetric activation functions based neural network operators. These multivariate Fuzzy-Random operators arise in a natural way among multivariate Fuzzy-Random neural networks. The rates are given through multivariate Probabilistic-Jackson type inequalities involving the multivariate Fuzzy-Random modulus of continuity of the engaged multivariate Fuzzy-Random function. The plain stochastic extreme analog of this theory is also met in detail for the stochastic analogs of the operators: the stochastic full quasi-interpolation operators, the stochastic Kantorovich type operators and the stochastic quadrature type operators.
Access provided by Autonomous University of Puebla. Download conference paper PDF
Similar content being viewed by others
Keywords
- Fuzzy-Random analysis
- Fuzzy-Random neural networks and operators
- Fuzzy-Random modulus of continuity
- Fuzzy-Random functions
- Stochastic processes
- Jackson type fuzzy and probabilistic inequalities
1 Fuzzy-Random Functions and Stochastic Processes Background
See also [18], Ch. 22, pp. 497–501.
We start with
Definition 1
(see [35]). Let \(\mu :\mathbb {R}\rightarrow \left[ 0,1\right] \) with the following properties:
-
(i)
is normal, i.e., \(\exists \) \(x_{0}\in \mathbb {R}:\mu \left( x_{0}\right) =1.\)
-
(ii)
\(\mu \left( \lambda x+\left( 1-\lambda \right) y\right) \ge \min \{ \mu \left( x\right) ,\mu \left( y\right) \}\), \(\forall \) \(x,y\in \mathbb {R},\) \(\forall \) \(\lambda \in \left[ 0,1\right] \) (\(\mu \) is called a convex fuzzy subset).
-
(iii)
\(\mu \) is upper semicontinuous on \(\mathbb {R}\), i.e., \(\forall \) \(x_{0}\in \mathbb {R}\) and \(\forall \) \(\varepsilon >0\), \(\exists \) neighborhood \(V\left( x_{0}\right) :\mu \left( x\right) \le \mu \left( x_{0}\right) +\varepsilon \), \(\forall \) \(x\in V\left( x_{0}\right) .\)
-
(iv)
the set \(\overline{\text {supp}\left( \mu \right) }\) is compact in \( \mathbb {R}\) (where supp\(\left( \mu \right) :=\{x\in \mathbb {R};\mu \left( x\right) >0\}\)).
We call \(\mu \) a fuzzy real number. Denote the set of all \(\mu \) with \( \mathbb {R}_{\mathcal {F}}.\)
E.g., \(\chi _{\{x_{0}\}}\in \mathbb {R}_{\mathcal {F}}\), for any \(x_{0}\in \mathbb {R},\) where \(\chi _{\{x_{0}\}}\) is the characteristic function at \( x_{0}\).
For \(0<r\le 1\) and \(\mu \in \mathbb {R}_{\mathcal {F}}\) define \(\left[ \mu \right] ^{r}:=\{x\in \mathbb {R}:\mu \left( x\right) \ge r\}\) and \(\left[ \mu \right] ^{0}:=\overline{\{x\in \mathbb {R}:\mu \left( x\right) >0\}}.\)
Then it is well known that for each \(r\in \left[ 0,1\right] \), \(\left[ \mu \right] ^{r}\) is a closed and bounded interval of \(\mathbb {R}\). For \(u,v\in \mathbb {R}_{\mathcal {F}}\) and \(\lambda \in \mathbb {R}\), we define uniquely the sum \(u\oplus v\) and the product \(\lambda \odot u\) by
where \(\left[ u\right] ^{r}+\left[ v\right] ^{r}\) means the usual addition of two intervals (as subsets of \(\mathbb {R}\)) and \(\lambda \left[ u\right] ^{r}\) means the usual product between a scalar and a subset of \(\mathbb {R}\) (see, e.g., [35]). Notice \(1\odot u=u\) and it holds \(u\oplus v=v\oplus u \), \(\lambda \odot u=u\odot \lambda \). If \(0\le r_{1}\le r_{2}\le 1\) then \(\left[ u\right] ^{r_{2}}\subseteq \left[ u\right] ^{r_{1}}\). Actually \( \left[ u\right] ^{r}=\left[ u_{-}^{\left( r\right) },u_{+}^{\left( r\right) } \right] \), where \(u_{-}^{\left( r\right) }<u_{+}^{\left( r\right) }\), \( u_{-}^{\left( r\right) },u_{+}^{\left( r\right) }\in \mathbb {R}\), \(\forall \) \(r\in \left[ 0,1\right] .\)
Define
by
where \(\left[ v\right] ^{r}=\left[ v_{-}^{\left( r\right) },v_{+}^{\left( r\right) }\right] ;\) \(u,v\in \mathbb {R}_{\mathcal {F}}\). We have that D is a metric on \(\mathbb {R}_{\mathcal {F}}\). Then \(\left( \mathbb {R}_{\mathcal {F} },D\right) \) is a complete metric space, see [35], with the properties
Let \(\left( M,d\right) \) metric space and \(f,g:M\rightarrow \mathbb {R}_{ \mathcal {F}}\) be fuzzy real number valued functions. The distance between f, g is defined by
On \(\mathbb {R}_{\mathcal {F}}\) we define a partial order by “\(\le \)”: \(u,v\in \mathbb {R}_{\mathcal {F}}\), \(u\le v\) iff \(u_{-}^{\left( r\right) }\le v_{-}^{\left( r\right) }\) and \(u_{+}^{\left( r\right) }\le v_{+}^{\left( r\right) }\), \(\forall \) \(r\in \left[ 0,1\right] .\)
\(\sum \limits ^{*}\) denotes the fuzzy summation, \(\widetilde{o}:=\chi _{\{0\}}\in \mathbb {R}_{\mathcal {F}}\) the neutral element with respect to \( \oplus \). For more see also [36, 37].
We need
Definition 2
(see also [30], Definition 13.16, p. 654). Let \(\left( X, \mathcal {B},P\right) \) be a probability space. A fuzzy-random variable is a \( \mathcal {B}\)-measurable mapping \(g:X\rightarrow \mathbb {R}_{\mathcal {F}}\) (i.e., for any open set \(U\subseteq \mathbb {R}_{\mathcal {F}},\) in the topology of \(\mathbb {R}_{\mathcal {F}}\) generated by the metric D, we have
The set of all fuzzy-random variables is denoted by \(\mathcal {L}_{\mathcal {F} }\left( X,\mathcal {B},P\right) \). Let \(g_{n},g\in \mathcal {L}_{\mathcal {F} }\left( X,\mathcal {B},P\right) \), \(n\in \mathbb {N}\) and \(0<q<+\infty \). We say if
Remark 1
(see [30], p. 654). If \(f,g\in \mathcal {L}_{\mathcal {F}}\left( X,\mathcal {B},P\right) \), let us denote \(F:X\rightarrow \mathbb {R}_{+}\cup \{0\}\) by \(F\left( s\right) =D\left( f\left( s\right) ,g\left( s\right) \right) \), \(s\in X\). Here, F is \(\mathcal {B}\)-measurable, because \( F=G\circ H\), where \(G\left( u,v\right) =D\left( u,v\right) \) is continuous on \(\mathbb {R}_{\mathcal {F}}\times \mathbb {R}_{\mathcal {F}}\), and \( H:X\rightarrow \mathbb {R}_{\mathcal {F}}\times \mathbb {R}_{\mathcal {F}}\), \( H\left( s\right) =\left( f\left( s\right) ,g\left( s\right) \right) \), \(s\in X\), is \(\mathcal {B}\)-measurable. This shows that the above convergence in q -mean makes sense.
Definition 3
(see [30], p. 654, Definition 13.17). Let \(\left( T, \mathcal {T}\right) \) be a topological space. A mapping \(f:T\rightarrow \mathcal {L}_{\mathcal {F}}\left( X,\mathcal {B},P\right) \) will be called fuzzy-random function (or fuzzy-stochastic process) on T. We denote \( f\left( t\right) \left( s\right) =f\left( t,s\right) \), \(t\in T\), \(s\in X\).
Remark 2
(see [30], p. 655). Any usual fuzzy real function \( f:T\rightarrow \mathbb {R}_{\mathcal {F}}\) can be identified with the degenerate fuzzy-random function \(f\left( t,s\right) =f\left( t\right) \), \( \forall \) \(t\in T\), \(s\in X\).
Remark 3
(see [30], p. 655). Fuzzy-random functions that coincide with probability one for each \(t\in T\) will be consider equivalent.
Remark 4
(see [30], p. 655). Let \(f,g:T\rightarrow \mathcal {L}_{ \mathcal {F}}\left( X,\mathcal {B},P\right) \). Then \(f\oplus g\) and \(k\odot f\) are defined pointwise, i.e.,
Definition 4
(see also Definition 13.18, pp. 655–656, [30]). For a fuzzy-random function \(f:W\subseteq \mathbb {R}^{N}\rightarrow \mathcal {L}_{ \mathcal {F}}\left( X,\mathcal {B},P\right) \), \(N\in \mathbb {N}\), we define the (first) fuzzy-random modulus of continuity
\(0<\delta ,\) \(1\le q<\infty .\)
Definition 5
[16]. Here \(1\le q<+\infty \). Let \(f:W\subseteq \mathbb {R} ^{N}\rightarrow \mathcal {L}_{\mathcal {F}}\left( X,\mathcal {B},P\right) \), \( N\in \mathbb {N}\), be a fuzzy random function. We call f a (q-mean) uniformly continuous fuzzy random function over W, iff \(\forall \) \( \varepsilon >0\) \(\exists \) \(\delta >0:\)whenever \(\left\| x-y\right\| _{\infty }\le \delta ,\) \(x,y\in W,\) implies that
We denote it as \(f\in C_{FR}^{U_{q}}\left( W\right) .\)
Proposition 1
[16]. Let \(f\in C_{FR}^{U_{q}}\left( W\right) ,\) where \( W\subseteq \mathbb {R}^{N}\) is convex.
Then \(\varOmega _{1}^{\left( \mathcal {F}\right) }\left( f,\delta \right) _{L^{q}}<\infty \), any \(\delta >0.\)
Proposition 2
[16]. Let \(f,g:W\subseteq \mathbb {R}^{N}\rightarrow \mathcal {L}_{\mathcal {F}}\left( X,\mathcal {B},P\right) \), \(N\in \mathbb {N}\), be fuzzy random functions. It holds
-
(i)
\(\varOmega _{1}^{\left( \mathcal {F}\right) }\left( f,\delta \right) _{L^{q}} \) is nonnegative and nondecreasing in \(\delta >0.\)
-
(ii)
\(\underset{\delta \downarrow 0}{\lim }\varOmega _{1}^{\left( \mathcal {F} \right) }\left( f,\delta \right) _{L^{q}}=\varOmega _{1}^{\left( \mathcal {F} \right) }\left( f,0\right) _{L^{q}}=0\), iff \(f\in C_{FR}^{U_{q}}\left( W\right) .\)
We mention
Definition 6
(see also [6]). Let \(f\left( t,s\right) \) be a random function (stochastic process) from \(W\times \left( X,\mathcal {B},P\right) ,\) \(W\subseteq \mathbb {R}^{N},\) into \(\mathbb {R}\), where \(\left( X,\mathcal {B} ,P\right) \) is a probability space. We define the q-mean multivariate first modulus of continuity of f by
\(\delta >0,\) \(1\le q<\infty \).
The concept of f being (q-mean) uniformly continuous random function is defined the same way as in Definition 5, just replace D by \( \left| \cdot \right| \), etc. We denote it as \(f\in C_{\mathbb {R} }^{U_{q}}\left( W\right) .\)
Similar properties as in Propositions 1, 2 are valid for \( \varOmega _{1}\left( f,\delta \right) _{L^{q}}.\)
Also we have
Proposition 3
[3]. Let \(Y\left( t,\omega \right) \) be a real valued stochastic process such that Y is continuous in \(t\in \left[ a,b\right] \). Then Y is jointly measurable in \(\left( t,\omega \right) .\)
According to [28], p. 94 we have the following
Definition 7
Let \(\left( Y,\mathcal {T}\right) \) be a topological space, with its \(\sigma \)-algebra of Borel sets \(\mathcal {B}:=\mathcal {B}\left( Y,\mathcal {T}\right) \) generated by \(\mathcal {T}\). If \(\left( X,\mathcal {S} \right) \) is a measurable space, a function \(f:X\rightarrow Y\) is called measurable iff \(f^{-1}\left( B\right) \in \mathcal {S}\) for all \(B\in \mathcal {B}\).
By Theorem 4.1.6 of [28], p. 89 f as above is measurable iff
We mention
Theorem 1
(see [28], p. 95). Let \(\left( X,\mathcal {S}\right) \) be a measurable space and \(\left( Y,d\right) \) be a metric space. Let \(f_{n}\) be measurable functions from X into Y such that for all \(x\in X\), \( f_{n}\left( x\right) \rightarrow f\left( x\right) \) in Y. Then f is measurable. I.e., \(\underset{n\rightarrow \infty }{\lim }f_{n}=f\) is measurable.
We need also
Proposition 4
[16]. Let f, g be fuzzy random variables from \(\mathcal {S}\) into \(\mathbb {R}_{\mathcal {F}}\). Then
-
(i)
Let \(c\in \mathbb {R}\), then \(c\odot f\) is a fuzzy random variable.
-
(ii)
\(f\oplus g\) is a fuzzy random variable.
Proposition 5
Let \(Y\left( \overrightarrow{t},\omega \right) \) be a real valued multivariate random function (stochastic process) such that Y is continuous in \(\overrightarrow{t}\in \prod \limits _{i=1}^{N}\left[ a_{i},b_{i}\right] \). Then Y is jointly measurable in \(\left( \overrightarrow{t},\omega \right) \) and \(\int _{\prod \limits _{i=1}^{N}\left[ a_{i},b_{i}\right] }Y\left( \overrightarrow{t},\omega \right) d \overrightarrow{t}\) is a real valued random variable.
Proof
Similar to Proposition 18.14, p. 353 of [7].
2 About Neural Networks Background
2.1 About the Arctangent Activation Function
We consider the
We will be using
which is a sigmoid type function and it is strictly increasing. We have that
and
We consider the activation function
and we notice that
it is an even function.
Since \(x+1>x-1\), then \(h\left( x+1\right) >h\left( x-1\right) \), and \(\psi _{1}\left( x\right) >0\), all \(x\in \mathbb {R}\).
We see that
Let \(x>0\), we have that
That is
That is \(\psi _{1}\) is strictly decreasing on \([0,\infty )\) and clearly is strictly increasing on \((-\infty ,0]\), and \(\psi _{1}^{\prime }\left( 0\right) =0.\)
Observe that
That is the x-axis is the horizontal asymptote on \(\psi _{1}\).
All in all, \(\psi _{1}\) is a bell symmetric function with maximum \(\psi _{1}\left( 0\right) \cong 18.31.\)
We need
Theorem 2
([19], p. 286). We have that
Theorem 3
([19], p. 287). It holds
So that \(\psi _{1}\left( x\right) \) is a density function on \(\mathbb {R}.\)
We mention
Theorem 4
([19], p. 288). Let \(0<\alpha <1\), and \(n\in \mathbb {N}\) with \(n^{1-\alpha }>2\). It holds
Denote by \(\left\lfloor \cdot \right\rfloor \) the integral part of the number and by \(\left\lceil \cdot \right\rceil \) the ceiling of the number.
We need
Theorem 5
([19], p. 289). Let \(x\in \left[ a,b\right] \subset \mathbb {R} \) and \(n\in \mathbb {N}\) so that \(\left\lceil na\right\rceil \le \left\lfloor nb\right\rfloor \). It holds
Note 1
([19], pp. 290–291).
-
i)
We have that
$$\begin{aligned} \underset{n\rightarrow \infty }{\lim }\sum _{k=\left\lceil na\right\rceil }^{\left\lfloor nb\right\rfloor }\psi _{1}\left( nx-k\right) \ne 1, \end{aligned}$$(18)for at least some \(x\in \left[ a,b\right] .\)
-
ii)
For large enough \(n\in \mathbb {N}\) we always obtain \(\left\lceil na\right\rceil \le \left\lfloor nb\right\rfloor \). Also \(a\le \frac{k}{n}\le b\), iff \(\left\lceil na\right\rceil \le k\le \left\lfloor nb\right\rfloor \).
In general, by Theorem 2, it holds
We introduce (see [24])
Denote by \(a=\left( a_{1},...,a_{N}\right) \) and \(b=\left( b_{1},...,b_{N}\right) .\)
It has the properties:
-
(i)
\(Z_{1}\left( x\right) >0\), \(\forall \) \(x\in \mathbb {R}^{N},\)
-
(ii)
$$\begin{aligned} \sum _{k=-\infty }^{\infty }Z_{1}\left( x-k\right) :=\sum _{k_{1}=-\infty }^{\infty }\sum _{k_{2}=-\infty }^{\infty }...\sum _{k_{N}=-\infty }^{\infty }Z_{1}\left( x_{1}-k_{1},...,x_{N}-k_{N}\right) =1,\text { } \end{aligned}$$(21)
where \(k:=\left( k_{1},...,k_{n}\right) \in \mathbb {Z}^{N}\), \(\forall \) \( x\in \mathbb {R}^{N},\)
hence
-
(iii)
$$\begin{aligned} \sum _{k=-\infty }^{\infty }Z_{1}\left( nx-k\right) =1, \end{aligned}$$(22)
\(\forall \) \(x\in \mathbb {R}^{N};\) \(n\in \mathbb {N}\),
and
-
(iv)
$$\begin{aligned} \int _{\mathbb {R}^{N}}Z_{1}\left( x\right) dx=1, \end{aligned}$$(23)
that is \(Z_{1}\) is a multivariate density function.
-
(v)
It is clear that
$$\begin{aligned} \sum _{\left\{ \begin{array}{c} k=-\infty \\ \left\| \frac{k}{n}-x\right\| _{\infty }>\frac{1}{n^{\beta }} \end{array} \right. }^{\infty }Z_{1}\left( nx-k\right) <\frac{2}{\pi ^{2}\left( n^{1-\beta }-2\right) }=c_{1}\left( \beta ,n\right) \text {, } \end{aligned}$$(24)\(0<\beta <1,\) \(n\in \mathbb {N}:n^{1-\beta }>2\), \(x\in \mathbb {R}^{N}.\)
-
(vi)
By Theorem 5 we get that
$$\begin{aligned} 0<\frac{1}{\sum _{k=\left\lceil na\right\rceil }^{\left\lfloor nb\right\rfloor }Z_{1}\left( nx-k\right) }<\frac{1}{\left( \psi _{1}\left( 1\right) \right) ^{N}}\cong \left( 4.9737\right) ^{N}=:\gamma _{1}\left( N\right) , \end{aligned}$$(25)\(\forall \) \(x\in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) \), \(n\in \mathbb {N}\).
Furthermore it holds
for at least some \(x\in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) .\)
Above it is \(\left\| x\right\| _{\infty }:=\max \left\{ \left| x_{1}\right| ,...,\left| x_{N}\right| \right\} \), \(x\in \mathbb {R}^{N}\), also set \(\infty :=\left( \infty ,...,\infty \right) \), \( -\infty =\left( -\infty ,...-\infty \right) \) upon the multivariate context.
2.2 About the Algebraic Activation Function
Here see also [20].
We consider the generator algebraic function
which is a sigmoidal type of function and is a strictly increasing function.
We see that \(\varphi \left( -x\right) =-\varphi \left( x\right) \) with \( \varphi \left( 0\right) =0\). We get that
proving \(\varphi \) as strictly increasing over \(\mathbb {R},\varphi ^{\prime }\left( x\right) =\varphi ^{\prime }\left( -x\right) .\) We easily find that \( \underset{x\rightarrow +\infty }{\lim }\varphi \left( x\right) =1\), \(\varphi \left( +\infty \right) =1\), and \(\underset{x\rightarrow -\infty }{\lim } \varphi \left( x\right) =-1\), \(\varphi \left( -\infty \right) =-1.\)
We consider the activation function
Clearly it is \(\psi _{2}\left( x\right) =\psi _{2}\left( -x\right) ,\) \( \forall \) \(x\in \mathbb {R}\), so that \(\psi _{2}\) is an even function and symmetric with respect to the y-axis. Clearly \(\psi _{2}\left( x\right) >0\) , \(\forall \) \(x\in \mathbb {R}\).
Also it is
By [20], we have that \(\psi _{2}^{\prime }\left( x\right) <0\) for \(x>0\). That is \(\psi _{2}\) is strictly decreasing over \(\left( 0,+\infty \right) . \)
Clearly, \(\psi _{2}\) is strictly increasing over \(\left( -\infty ,0\right) \) and \(\psi _{2}^{\prime }\left( 0\right) =0\).
Furthermore we obtain that
and
That is the x-axis is the horizontal asymptote of \(\psi _{2}\).
Conclusion, \(\psi _{2}\) is a bell shape symmetric function with maximum
We need
Theorem 6
[20]. We have that
Theorem 7
[20]. It holds
Theorem 8
[20]. Let \(0<\alpha <1\), and \(n\in \mathbb {N}\) with \( n^{1-\alpha }>2\). It holds
We need
Theorem 9
[20]. Let \(\left[ a,b\right] \subset \mathbb {R}\) and \(n\in \mathbb {N}\) so that \(\left\lceil na\right\rceil \le \left\lfloor nb\right\rfloor \). It holds
\(\forall \) \(x\in \left[ a,b\right] \), \(m\in \mathbb {N}.\)
Note 2
-
1)
By [20] we have that
$$\begin{aligned} \underset{n\rightarrow \infty }{\lim }\sum \limits _{k=\left\lceil na\right\rceil }^{\left\lfloor nb\right\rfloor }\psi _{2}\left( nx-k\right) \ne 1, \text { } \end{aligned}$$(38)for at least some \(x\in \left[ a,b\right] .\)
-
2)
Let \(\left[ a,b\right] \subset \mathbb {R}\). For large \(n\in \mathbb {N}\) we always have \(\left\lceil na\right\rceil \le \left\lfloor nb\right\rfloor \). Also \(a\le \frac{k}{n}\le b\), iff \(\left\lceil na\right\rceil \le k\le \left\lfloor nb\right\rfloor \).
In general it holds that
We introduce (see also [25])
It has the properties:
-
(i)
\(Z_{2}\left( x\right) >0\), \(\forall \) \(x\in \mathbb {R}^{N},\)
-
(ii)
$$\begin{aligned} \sum _{k=-\infty }^{\infty }Z_{2}\left( x-k\right) :=\sum _{k_{1}=-\infty }^{\infty }\sum _{k_{2}=-\infty }^{\infty }...\sum _{k_{N}=-\infty }^{\infty }Z_{2}\left( x_{1}-k_{1},...,x_{N}-k_{N}\right) =1,\text { } \end{aligned}$$(41)
where \(k:=\left( k_{1},...,k_{n}\right) \in \mathbb {Z}^{N}\), \(\forall \) \( x\in \mathbb {R}^{N},\)
hence
-
(iii)
$$\begin{aligned} \sum _{k=-\infty }^{\infty }Z_{2}\left( nx-k\right) =1, \end{aligned}$$(42)
\(\forall \) \(x\in \mathbb {R}^{N};\) \(n\in \mathbb {N}\),
and
-
(iv)
$$\begin{aligned} \int _{\mathbb {R}^{N}}Z_{2}\left( x\right) dx=1, \end{aligned}$$(43)
that is \(Z_{2}\) is a multivariate density function.
-
(v)
It is clear that
$$\begin{aligned} \sum _{\left\{ \begin{array}{c} k=-\infty \\ \left\| \frac{k}{n}-x\right\| _{\infty }>\frac{1}{n^{\beta }} \end{array} \right. }^{\infty }Z_{2}\left( nx-k\right) <\frac{1}{4m\left( n^{1-\beta }-2\right) ^{2m}}=c_{2}\left( \beta ,n\right) \text {, } \end{aligned}$$(44)\(0<\beta <1,\) \(n\in \mathbb {N}:n^{1-\beta }>2\), \(x\in \mathbb {R}^{N}\), \(m\in \mathbb {N}\mathbf {.}\)
-
(vi)
By Theorem 9 we get that
$$\begin{aligned} 0<\frac{1}{\sum _{k=\left\lceil na\right\rceil }^{\left\lfloor nb\right\rfloor }Z_{2}\left( nx-k\right) }<\frac{1}{\left( \psi _{2}\left( 1\right) \right) ^{N}}\cong \left[ 2\left( \root 2m \of {1+4^{m}}\right) \right] ^{N}:=\gamma _{2}\left( N\right) , \end{aligned}$$(45)\(\forall \) \(x\in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) \), \(n\in \mathbb {N}\).
Furthermore it holds
for at least some \(x\in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) .\)
2.3 About the Gudermannian Activation Function
Here we consider \(gd\left( x\right) \) the Gudermannian function [34], which is a sigmoid function, as a generator function:
Let the normalized generator sigmoid function
Here
hence f is strictly increasing on \(\mathbb {R}.\)
Notice that \(\tanh \left( -x\right) =-\tanh x\) and \(\arctan \left( -x\right) =-\arctan x\), \(x\in \mathbb {R}.\)
So, here the neural network activation function will be:
By [21], we get that
i.e. it is even and symmetric with respect to the y-axis. Here we have \( f\left( +\infty \right) =1\), \(f\left( -\infty \right) =-1\) and \(f\left( 0\right) =0\). Clearly it is
an odd function, symmetric with respect to the origin. Since \(x+1>x-1\), and \( f\left( x+1\right) >f\left( x-1\right) \), we obtain \(\psi _{3}\left( x\right) >0\), \(\forall \) \(x\in \mathbb {R}.\)
By [21], we have that
By [21] \(\psi _{3}\) is strictly decreasing on \(\left( 0,+\infty \right) \), and strictly increasing on \(\left( -\infty ,0\right) \), and \(\psi _{3}^{\prime }\left( 0\right) =0\).
Also we have that
that is the x-axis is the horizontal asymptote for \(\psi _{3}\).
Conclusion, \(\psi _{3}\) is a bell shaped symmetric function with maximum \( \psi _{3}\left( 0\right) \cong 0.551\).
We need
Theorem 10
[21]. It holds that
Theorem 11
[21]. We have that
So \(\psi _{3}\left( x\right) \) is a density function.
Theorem 12
[21]. Let \(0<\alpha <1\), and \(n\in \mathbb {N}\) with \( n^{1-\alpha }>2\). It holds
Theorem 13
[21]. Let \(\left[ a,b\right] \subset \mathbb {R}\) and \(n\in \mathbb {N},\) so that \(\left\lceil na\right\rceil \le \left\lfloor nb\right\rfloor \). It holds
\(\forall \) \(x\in \left[ a,b\right] .\)
We make
Remark 5
[21].
-
(i)
We have that
$$\begin{aligned} \underset{n\rightarrow \infty }{\lim }\sum \limits _{k=\left\lceil na\right\rceil }^{\left\lfloor nb\right\rfloor }\psi _{3}\left( nx-k\right) \ne 1, \text { } \end{aligned}$$(58)for at least some \(x\in \left[ a,b\right] .\)
-
(ii)
Let \(\left[ a,b\right] \subset \mathbb {R}\). For large n we always have \(\left\lceil na\right\rceil \le \left\lfloor nb\right\rfloor \). Also \(a\le \frac{k}{n}\le b\), iff \(\left\lceil na\right\rceil \le k\le \left\lfloor nb\right\rfloor \).
In general it holds
We introduce (see also [23])
It has the properties:
-
(i)
\(Z_{3}\left( x\right) >0\), \(\forall \) \(x\in \mathbb {R}^{N},\)
-
(ii)
$$\begin{aligned} \sum _{k=-\infty }^{\infty }Z_{3}\left( x-k\right) :=\sum _{k_{1}=-\infty }^{\infty }\sum _{k_{2}=-\infty }^{\infty }...\sum _{k_{N}=-\infty }^{\infty }Z_{3}\left( x_{1}-k_{1},...,x_{N}-k_{N}\right) =1,\text { } \end{aligned}$$(61)
where \(k:=\left( k_{1},...,k_{n}\right) \in \mathbb {Z}^{N}\), \(\forall \) \( x\in \mathbb {R}^{N},\)
hence
-
(iii)
$$\begin{aligned} \sum _{k=-\infty }^{\infty }Z_{3}\left( nx-k\right) =1, \end{aligned}$$(62)
\(\forall \) \(x\in \mathbb {R}^{N};\) \(n\in \mathbb {N}\),
and
-
(iv)
$$\begin{aligned} \int _{\mathbb {R}^{N}}Z_{3}\left( x\right) dx=1, \end{aligned}$$(63)
that is \(Z_{3}\) is a multivariate density function.
-
(v)
It is also clear that
$$\begin{aligned} \sum _{\left\{ \begin{array}{c} k=-\infty \\ \left\| \frac{k}{n}-x\right\| _{\infty }>\frac{1}{n^{\beta }} \end{array} \right. }^{\infty }Z_{3}\left( nx-k\right) <\frac{2e^{2}}{\pi e^{n^{1-\beta }}}=c_{3}\left( \beta ,n\right) , \end{aligned}$$(64)\(0<\beta <1\), \(n\in \mathbb {N}:n^{1-\beta }>2\), \(x\in \mathbb {R}^{N},\) \(m\in \mathbb {N}.\)
-
(vi)
By Theorem 13 we get that
$$\begin{aligned} 0<\frac{1}{\sum _{k=\left\lceil na\right\rceil }^{\left\lfloor nb\right\rfloor }Z_{3}\left( nx-k\right) }<\left( \frac{2\pi }{gd\left( 2\right) }\right) ^{N}\cong \left( 4.824\right) ^{N}=:\gamma _{3}\left( N\right) , \end{aligned}$$(65)\(\forall \) \(x\in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) \), \(n\in \mathbb {N}\).
Furthermore it holds
for at least some \(x\in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) .\)
2.4 About the Generalized Symmetrical Activation Function
Here we consider the generalized symmetrical sigmoid function [22, 29]
This has applications in immunology and protection from disease together with probability theory. It is also called a symmetrical protection curve.
The parameter \(\mu \) is a shape parameter controling how fast the curve approaches the asymptotes for a given slope at the inflection point. When \( \mu =1\) \(f_{1}\) is the absolute sigmoid function, and when \(\mu =2,\) \(f_{1}\) is the square root sigmoid function. When \(\mu =1.5\) the function approximates the arctangent function, when \(\mu =2.9\) it approximates the logistic function, and when \(\mu =3.4\) it approximates the error function. Parameter \(\mu \) is estimated in the likelihood maximization [29]. For more see [29].
Next we study the particular generator sigmoid function
We have that \(f_{2}\left( 0\right) =0\), and
so \(f_{2}\) is symmetric with respect to zero.
When \(x\ge 0\), we get that [22]
that is \(f_{2}\) is strictly increasing on \([0,+\infty )\) and \(f_{2}\) is strictly increasing on \((-\infty ,0]\). Hence \(f_{2}\) is strictly increasing on \(\mathbb {R}\).
We also have \(f_{2}\left( +\infty \right) =f_{2}\left( -\infty \right) =1.\)
Let us consider the activation function [22]:
Clearly it holds [22]
and
and \(\psi _{4}\left( x\right) >0\), \(\forall \) \(x\in \mathbb {R}\).
Following [22], we have that \(\psi _{4}\) is strictly decreasing over \( [0,+\infty )\), and \(\psi _{4}\) is strictly increasing on \((-\infty ,0]\), by \( \psi _{4}\)-symmetry with respect to y-axis, and \(\psi _{4}^{\prime }\left( 0\right) =0.\)
Clearly it is
therefore the x-axis is the horizontal asymptote of \(\psi _{4}\left( x\right) .\)
The value
is the maximum of \(\psi _{4}\), which is a bell shaped function.
We need
Theorem 14
[22]. It holds
Theorem 15
[22]. We have that
So that \(\psi _{4}\left( x\right) \) is a density function on \(\mathbb {R}.\)
We need
Theorem 16
[22]. Let \(0<\alpha <1\), and \(n\in \mathbb {N}\) with \( n^{1-\alpha }>2\). It holds
where \(\lambda \in \mathbb {N}\) is an odd number.
We also need
Theorem 17
[22]. Let \(\left[ a,b\right] \subset \mathbb {R}\) and \(n\in \mathbb {N}\) so that \(\left\lceil na\right\rceil \le \left\lfloor nb\right\rfloor \). Then
where \(\lambda \) is an odd number, \(\forall \) \(x\in \left[ a,b\right] .\)
We make
Remark 6
[22]. (1) We have that
(2) Let \(\left[ a,b\right] \subset \mathbb {R}\). For large enough n we always obtain \(\left\lceil na\right\rceil \le \left\lfloor nb\right\rfloor \). Also \(a\le \frac{k}{n}\le b\), iff \(\left\lceil na\right\rceil \le k\le \left\lfloor nb\right\rfloor \).
In general it holds that
We introduce (see also [26])
It has the properties:
-
(i)
\(Z_{4}\left( x\right) >0\), \(\forall \) \(x\in \mathbb {R}^{N},\)
-
(ii)
$$\begin{aligned} \sum _{k=-\infty }^{\infty }Z_{4}\left( x-k\right) :=\sum _{k_{1}=-\infty }^{\infty }\sum _{k_{2}=-\infty }^{\infty }...\sum _{k_{N}=-\infty }^{\infty }Z_{4}\left( x_{1}-k_{1},...,x_{N}-k_{N}\right) =1,\text { } \end{aligned}$$(83)
where \(k:=\left( k_{1},...,k_{n}\right) \in \mathbb {Z}^{N}\), \(\forall \) \( x\in \mathbb {R}^{N},\)
hence
-
(iii)
$$\begin{aligned} \sum _{k=-\infty }^{\infty }Z_{4}\left( nx-k\right) =1, \end{aligned}$$(84)
\(\forall \) \(x\in \mathbb {R}^{N};\) \(n\in \mathbb {N}\),
and
-
(iv)
$$\begin{aligned} \int _{\mathbb {R}^{N}}Z_{4}\left( x\right) dx=1, \end{aligned}$$(85)
that is \(Z_{4}\) is a multivariate density function.
-
(v)
It is clear that
$$\begin{aligned} \sum _{\left\{ \begin{array}{c} k=-\infty \\ \left\| \frac{k}{n}-x\right\| _{\infty }>\frac{1}{n^{\beta }} \end{array} \right. }^{\infty }Z_{4}\left( nx-k\right) <\frac{1}{2\lambda \left( n^{1-\beta }-2\right) ^{\lambda }}=c_{4}\left( \beta ,n\right) , \end{aligned}$$(86)\(0<\beta <1\), \(n\in \mathbb {N}:n^{1-\beta }>2\), \(x\in \mathbb {R}^{N},\) \( \lambda \) is odd.
-
(vi)
By Theorem 17 we get that
$$\begin{aligned} 0<\frac{1}{\sum _{k=\left\lceil na\right\rceil }^{\left\lfloor nb\right\rfloor }Z_{4}\left( nx-k\right) }<\left( 2\root \lambda \of {1+2^{\lambda }} \right) ^{N}=:\gamma _{4}\left( N\right) , \end{aligned}$$(87)\(\forall \) \(x\in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) \), \(n\in \mathbb {N}\), \(\lambda \) is odd.
Furthermore it holds
for at least some \(x\in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) .\)
Set
where \(a:=\left( a_{1},...,a_{N}\right) \), \(b:=\left( b_{1},...,b_{N}\right) \), \(k:=\left( k_{1},...,k_{N}\right) .\)
Let \(f\in C\left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) ,\) and \( n\in \mathbb {N}\) such that \(\left\lceil na_{i}\right\rceil \le \left\lfloor nb_{i}\right\rfloor \), \(i=1,...,N.\)
We define the multivariate averaged positive linear quasi-interpolation neural network operators (\(x:=\left( x_{1},...,x_{N}\right) \in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) \)); \(j=1,2,3,4\):
For large enough \(n\in \mathbb {N}\) we always obtain \(\left\lceil na_{i}\right\rceil \le \left\lfloor nb_{i}\right\rfloor \), \(i=1,...,N\). Also \(a_{i}\le \frac{k_{i}}{n}\le b_{i}\), iff \(\left\lceil na_{i}\right\rceil \le k_{i}\le \left\lfloor nb_{i}\right\rfloor \), \( i=1,...,N\).
When \(f\in C_{B}\left( \mathbb {R}^{N}\right) \) we define (\(j=1,2,3,4\))
\(n\in \mathbb {N}\), \(\forall \) \(x\in \mathbb {R}^{N},\) \(N\in \mathbb {N}\), the multivariate full quasi-interpolation neural network operators.
Also for \(f\in C_{B}\left( \mathbb {R}^{N}\right) \) we define the multivariate Kantorovich type neural network operators (\(j=1,2,3,4\))
\(n\in \mathbb {N},\ \forall \) \(x\in \mathbb {R}^{N}.\)
Again for \(f\in C_{B}\left( \mathbb {R}^{N}\right) ,\) \(N\in \mathbb {N},\) we define the multivariate neural network operators of quadrature type \( _{j}D_{n}\left( f,x\right) \), \(n\in \mathbb {N},\) as follows. Let \(\theta =\left( \theta _{1},...,\theta _{N}\right) \in \mathbb {N}^{N},\) \(\overline{r}=\left( r_{1},...,r_{N}\right) \in \mathbb {Z}_{+}^{N}\), \(w_{\overline{r}}=w_{r_{1},r_{2},...r_{N}}\ge 0\), such that \(\sum \limits _{\overline{r}=0}^{\theta }w_{\overline{r}}=\sum \limits _{r_{1}=0}^{\theta _{1}}\sum \limits _{r_{2}=0}^{\theta _{2}}...\sum \limits _{r_{N}=0}^{\theta _{N}}w_{r_{1},r_{2},...r_{N}}=1;\) \(k\in \mathbb {Z}^{N}\) and
where \(\frac{\overline{r}}{\theta }:=\left( \frac{r_{1}}{\theta _{1}},\frac{ r_{2}}{\theta _{2}},...,\frac{r_{N}}{\theta _{N}}\right) \); \(j=1,2,3,4.\)
We put
\(\forall \) \(x\in \mathbb {R}^{N}.\)
For the next we need, for \(f\in C\left( \prod _{i=1}^{N}\left[ a_{i},b_{i} \right] \right) \) the first multivariate modulus of continuity
It holds that
Similarly it is defined for \(f\in C_{B}\left( \mathbb {R}^{N}\right) \) (continuous and bounded functions on \(\mathbb {R}^{N}\)) the \(\omega _{1}\left( f,h\right) \), and it has the property (95), given that \( f\in C_{U}\left( \mathbb {R}^{N}\right) \) (uniformly continuous functions on \( \mathbb {R}^{N}\)).
We mention
Theorem 18
(see [23,24,25,26]). Let \(f\in C\left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) ,\) \(0<\beta <1\), \( x\in \left( \prod _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) ,\) \(N,n\in \mathbb {N}\) with \(n^{1-\beta }>2;\) \(j=1,2,3,4\). Then
-
1)
$$\begin{aligned} \left| _{j}A_{n}\left( f,x\right) -f\left( x\right) \right| \le \gamma _{j}\left( N\right) \left[ \omega _{1}\left( f,\frac{1}{n^{\beta }} \right) +2c_{j}\left( \beta ,n\right) \left\| f\right\| _{\infty } \right] =:\lambda _{j1}, \end{aligned}$$(96)
and
-
2)
$$\begin{aligned} \left\| _{j}A_{n}\left( f\right) -f\right\| _{\infty }\le \lambda _{j1}. \end{aligned}$$(97)
We notice that \(\underset{n\rightarrow \infty }{\lim }\) \(_{j}A_{n}\left( f\right) =f\), pointwise and uniformly.
In this article we extend Theorem 18 to the fuzzy-random level.
We mention
Theorem 19
(see [23,24,25,26]). Let \(f\in C_{B}\left( \mathbb {R}^{N}\right) ,\) \(0<\beta <1\), \(x\in \mathbb {R}^{N},\) \( N,n\in \mathbb {N}\) with \(n^{1-\beta }>2;\) \(j=1,2,3,4\). Then
-
1)
$$\begin{aligned} \left| _{j}B_{n}\left( f,x\right) -f\left( x\right) \right| \le \omega _{1}\left( f,\frac{1}{n^{\beta }}\right) +2c_{j}\left( \beta ,n\right) \left\| f\right\| _{\infty }=:\lambda _{j2}, \end{aligned}$$(98)
-
2)
$$\begin{aligned} \left\| _{j}B_{n}\left( f\right) -f\right\| _{\infty }\le \lambda _{j2}. \end{aligned}$$(99)
Given that \(f\in \left( C_{U}\left( \mathbb {R}^{N}\right) \cap C_{B}\left( \mathbb {R}^{N}\right) \right) \), we obtain \(\underset{n\rightarrow \infty }{ \lim }\) \(_{j}B_{n}\left( f\right) =f\), uniformly.
We also need
Theorem 20
(see [23,24,25,26]). Let \(f\in C_{B}\left( \mathbb {R}^{N}\right) \), \(0<\beta <1\), \(x\in \mathbb {R}^{N},\) \(N,n\in \mathbb {N}\) with \(n^{1-\beta }>2;\) \(j=1,2,3,4\). Then
-
1)
$$\begin{aligned} \left| _{j}C_{n}\left( f,x\right) -f\left( x\right) \right| \le \omega _{1}\left( f,\frac{1}{n}+\frac{1}{n^{\beta }}\right) +2c_{j}\left( \beta ,n\right) \left\| f\right\| _{\infty }=:\lambda _{j3}, \end{aligned}$$(100)
-
2)
$$\begin{aligned} \left\| _{j}C_{n}\left( f\right) -f\right\| _{\infty }\le \lambda _{j3}. \end{aligned}$$(101)
Given that \(f\in \left( C_{U}\left( \mathbb {R}^{N}\right) \cap C_{B}\left( \mathbb {R}^{N}\right) \right) ,\) we obtain \(\underset{n\rightarrow \infty }{ \lim }\) \(_{j}C_{n}\left( f\right) =f\), uniformly.
We also need
Theorem 21
(see [23,24,25,26]). Let \(f\in C_{B}\left( \mathbb {R}^{N}\right) ,\) \(0<\beta <1\), \(x\in \mathbb {R}^{N},\) \( N,n\in \mathbb {N}\) with \(n^{1-\beta }>2;\) \(j=1,2,3,4\). Then
-
1)
$$\begin{aligned} \left| _{j}D_{n}\left( f,x\right) -f\left( x\right) \right| \le \omega _{1}\left( f,\frac{1}{n}+\frac{1}{n^{\beta }}\right) +2c_{j}\left( \beta ,n\right) \left\| f\right\| _{\infty }=\lambda _{j3}, \end{aligned}$$(102)
-
2)
$$\begin{aligned} \left\| _{j}D_{n}\left( f\right) -f\right\| _{\infty }\le \lambda _{j3}. \end{aligned}$$(103)
Given that \(f\in \left( C_{U}\left( \mathbb {R}^{N}\right) \cap C_{B}\left( \mathbb {R}^{N}\right) \right) ,\) we obtain \(\underset{n\rightarrow \infty }{ \lim }\) \(_{j}D_{n}\left( f\right) =f\), uniformly.
In this article we extend Theorems 19, 20, 21 to the random level.
We are also motivated by [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16] and continuing [17]. For general knowledge on neural networks we recommend [31,32,33].
3 Main Results
I) q -mean Approximation by Fuzzy-Random arctangent, algebraic, Gudermannian and generalized symmetric activation functions based Quasi-Interpolation Neural Network Operators
All terms and assumptions here as in Sects. 1, 2.
Let \(f\in C_{\mathcal{F}\mathcal{R}}^{U_{q}}\left( \prod \limits _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) \), \(1\le q<+\infty \), \(n,N\in \mathbb {N}\), \( 0<\beta <1,\) \(\overrightarrow{x}\in \left( \prod \limits _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) \), \(\left( X,\mathcal {B},P\right) \) probability space, \(s\in X\); \(j=1,2,3,4.\)
We define the following multivariate fuzzy random arctangent, algebraic, Gudermannian and generalized symmetric activation functions based quasi-interpolation linear neural network operators
(see also (89).
We present
Theorem 22
Let \(f\in C_{\mathcal{F}\mathcal{R}}^{U_{q}}\left( \prod \limits _{i=1}^{N} \left[ a_{i},b_{i}\right] \right) ,\) \(0<\beta <1,\) \(\overrightarrow{x}\in \left( \prod \limits _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) \), \(n,N\in \mathbb {N},\) with \(n^{1-\beta }>2,\) \(1\le q<+\infty .\) Assume that \(\int _{X}\left( D^{*}\left( f\left( \cdot ,s\right) ,\right. \right. \left. \left. \widetilde{o}\right) \right) ^{q}P\left( ds\right) <\infty ;\) \(j=1,2,3,4.\) Then
-
1)
$$\begin{aligned} \left( \int _{X}D^{q}\left( \left( _{j}A_{n}^{\mathcal{F}\mathcal{R}}\left( f\right) \right) \left( \overrightarrow{x},s\right) ,f\left( \overrightarrow{x} ,s\right) \right) P\left( ds\right) \right) ^{\frac{1}{q}}\le \end{aligned}$$(105)$$\begin{aligned} \gamma _{j}\left( N\right) \left\{ \varOmega _{1}\left( f,\frac{1}{n^{\beta }} \right) _{L^{q}}+2c_{j}\left( \beta ,n\right) \left( \int _{X}\left( D^{*}\left( f\left( \cdot ,s\right) ,\widetilde{o}\right) \right) ^{q}P\left( ds\right) \right) ^{\frac{1}{q}}\right\} =:\lambda _{j1}^{\left( \mathcal {FR }\right) }, \end{aligned}$$
-
2)
$$\begin{aligned} \left\| \left( \int _{X}D^{q}\left( \left( _{j}A_{n}^{\mathcal{F}\mathcal{R}}\left( f\right) \right) \left( \overrightarrow{x},s\right) ,f\left( \overrightarrow{ x},s\right) \right) P\left( ds\right) \right) ^{\frac{1}{q}}\right\| _{\infty ,\left( \prod \limits _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) }\le \lambda _{j1}^{\left( \mathcal{F}\mathcal{R}\right) }, \end{aligned}$$(106)
where \(\gamma _{j}\left( N\right) \) as in (25), (45), (65), (87) and \(c_{j}\left( \beta ,n\right) \) as in (24), (44), (64), (86).
Proof
We notice that
Hence
and
We observe that
So that
Hence it holds
(by (24), (25); (44), (45); (64), (65); (86), (87))
We have proved claim.
Conclusion 6
By Theorem 22 we obtain the pointwise and uniform convergences with rates in the q-mean and D-metric of the operator \( _{j}A_{n}^{\mathcal{F}\mathcal{R}}\) to the unit operator for \(f\in C_{\mathcal{F}\mathcal{R} }^{U_{q}}\left( \prod \limits _{i=1}^{N}\left[ a_{i},b_{i}\right] \right) ,\) \( j=1,2,3,4.\)
II) 1-mean Approximation by Stochastic arctangent, algebraic, Gudermannian and generalized symmetric activation functions based full Quasi-Interpolation Neural Network Operators
Let \(g\in C_{\mathcal {R}}^{U_{1}}\left( \mathbb {R}^{N}\right) \), \(0<\beta <1\) , \(\overrightarrow{x}\in \mathbb {R}^{N}\), \(n,N\in \mathbb {N}\), with \( \left\| g\right\| _{\infty ,\mathbb {R}^{N},X}<\infty \), \(\left( X, \mathcal {B},P\right) \) probability space, \(s\in X.\)
We define
(see also (90)).
We give
Theorem 23
Let \(g\in C_{\mathcal {R}}^{U_{1}}\left( \mathbb {R}^{N}\right) ,\) \( 0<\beta <1\), \(\overrightarrow{x}\in \mathbb {R}^{N}\), \(n,N\in \mathbb {N}\), with \(n^{1-\beta }>2,\) \(\left\| g\right\| _{\infty ,\mathbb {R} ^{N},X}<\infty ;\) \(j=1,2,3,4.\) Then
-
1)
$$\begin{aligned} \int _{X}\left| \left( _{j}B_{n}^{\left( \mathcal {R}\right) }\left( g\right) \right) \left( \overrightarrow{x},s\right) -g\left( \overrightarrow{ x},s\right) \right| P\left( ds\right) \le \end{aligned}$$(118)$$\begin{aligned} \left\{ \varOmega _{1}\left( g,\frac{1}{n^{\beta }}\right) _{L^{1}}+2c_{j}\left( \beta ,n\right) \left\| g\right\| _{\infty ,\mathbb {R}^{N},X}\right\} =:\mu _{j1}^{\left( \mathcal {R}\right) }, \end{aligned}$$
-
2)
$$\begin{aligned} \left\| \int _{X}\left| \left( _{j}B_{n}^{\left( \mathcal {R}\right) }\left( g\right) \right) \left( \overrightarrow{x},s\right) -g\left( \overrightarrow{x},s\right) \right| P\left( ds\right) \right\| _{\infty ,\mathbb {R}^{N}}\le \mu _{j1}^{\left( \mathcal {R}\right) }. \end{aligned}$$(119)
Proof
Since \(\left\| g\right\| _{\infty ,\mathbb {R}^{N},X}<\infty \), then
Hence
We observe that
However it holds
Hence
Furthermore it holds
proving the claim.
Conclusion 7
By Theorem 23 we obtain pointwise and uniform convergences with rates in the 1-mean of random operators \( _{j}B_{n}^{\left( \mathcal {R}\right) }\) to the unit operator for \(g\in C_{ \mathcal {R}}^{U_{1}}\left( \mathbb {R}^{N}\right) \), \(j=1,2,3,4.\)
III) 1-mean Approximation by Stochastic arctangent, algebraic, Gudermannian and generalized symmetric activation functions based multivariate Kantorovich type neural network operator
Let \(g\in C_{\mathcal {R}}^{U_{1}}\left( \mathbb {R}^{N}\right) \), \(0<\beta <1\) , \(\overrightarrow{x}\in \mathbb {R}^{N}\), \(n,N\in \mathbb {N}\), with \( \left\| g\right\| _{\infty ,\mathbb {R}^{N},X}<\infty \), \(\left( X,\mathcal {B},P\right) \) probability space, \(s\in X.\)
We define (\(j=1,2,3,4\)):
(see also (91).
We present
Theorem 24
Let \(g\in C_{\mathcal {R}}^{U_{1}}\left( \mathbb {R}^{N}\right) ,\) \( 0<\beta <1\), \(\overrightarrow{x}\in \mathbb {R}^{N}\), \(n,N\in \mathbb {N}\), with \(n^{1-\beta }>2;\) \(j=1,2,3,4,\) \(\left\| g\right\| _{\infty ,\mathbb {R}^{N},X}<\infty .\) Then
-
1)
$$\begin{aligned} \int _{X}\left| \left( _{j}C_{n}^{\left( \mathcal {R}\right) }\left( g\right) \right) \left( \overrightarrow{x},s\right) -g\left( \overrightarrow{ x},s\right) \right| P\left( ds\right) \le \end{aligned}$$$$\begin{aligned} \left[ \varOmega _{1}\left( g,\frac{1}{n}+\frac{1}{n^{\beta }}\right) _{L^{1}}+2c_{j}\left( \beta ,n\right) \left\| g\right\| _{\infty , \mathbb {R}^{N},X}\right] =:\gamma _{j1}^{\left( \mathcal {R}\right) }, \end{aligned}$$(127)
-
2)
$$\begin{aligned} \left\| \int _{X}\left| \left( _{j}C_{n}^{\left( \mathcal {R}\right) }\left( g\right) \right) \left( \overrightarrow{x},s\right) -g\left( \overrightarrow{x},s\right) \right| P\left( ds\right) \right\| _{\infty ,\mathbb {R}^{N}}\le \gamma _{j1}^{\left( \mathcal {R}\right) }. \end{aligned}$$(128)
Proof
Since \(\left\| g\right\| _{\infty ,\mathbb {R}^{N},X}<\infty \), then
Hence
We observe that
However it holds
Hence
Furthermore it holds
proving the claim.
Conclusion 8
By Theorem 24 we obtain pointwise and uniform convergences with rates in the 1-mean of random operators \( _{j}C_{n}^{\left( \mathcal {R}\right) }\) to the unit operator for \(g\in C_{ \mathcal {R}}^{U_{1}}\left( \mathbb {R}^{N}\right) \), \(j=1,2,3,4.\)
IV) 1-mean Approximation by Stochastic arctangent, algebraic, Gudermannian and generalized symmetric activation functions based multivariate quadrature type neural network operator
Let \(g\in C_{\mathcal {R}}^{U_{1}}\left( \mathbb {R}^{N}\right) \), \(0<\beta <1\) , \(\overrightarrow{x}\in \mathbb {R}^{N}\), \(n,N\in \mathbb {N}\), with \( \left\| g\right\| _{\infty ,\mathbb {R}^{N},X}<\infty \), \(\left( X, \mathcal {B},P\right) \) probability space, \(s\in X\), \(j=1,2,3,4.\)
We define
where
We finally give
Theorem 25
Let \(g\in C_{\mathcal {R}}^{U_{1}}\left( \mathbb {R}^{N}\right) ,\) \( 0<\beta <1\), \(\overrightarrow{x}\in \mathbb {R}^{N}\), \(n,N\in \mathbb {N}\), with \(n^{1-\beta }>2;\) \(j=1,2,3,4,\) \(\left\| g\right\| _{\infty ,\mathbb {R}^{N},X}<\infty .\) Then
-
1)
$$\begin{aligned} \int _{X}\left| \left( _{j}D_{n}^{\left( \mathcal {R}\right) }\left( g\right) \right) \left( \overrightarrow{x},s\right) -g\left( \overrightarrow{ x},s\right) \right| P\left( ds\right) \le \end{aligned}$$$$\begin{aligned} \left\{ \varOmega _{1}\left( g,\frac{1}{n}+\frac{1}{n^{\beta }}\right) _{L^{1}}+2c_{j}\left( \beta ,n\right) \left\| g\right\| _{\infty , \mathbb {R}^{N},X}\right\} =:\gamma _{j1}^{\left( \mathcal {R}\right) }, \end{aligned}$$(140)
-
2)
$$\begin{aligned} \left\| \int _{X}\left| \left( _{j}D_{n}^{\left( \mathcal {R}\right) }\left( g\right) \right) \left( \overrightarrow{x},s\right) -g\left( \overrightarrow{x},s\right) \right| P\left( ds\right) \right\| _{\infty ,\mathbb {R}^{N}}\le \gamma _{j1}^{\left( \mathcal {R}\right) }. \end{aligned}$$(141)
Proof
Notice that
Hence
We observe that
Thus
Hence it holds
Furthermore we derive
proving the claim.
Conclusion 9
From Theorem 25 we obtain pointwise and uniform convergences with rates in the 1-mean of random operators \(_{j}D_{n}^{\left( \mathcal {R}\right) }\) to the unit operator for \(g\in C_{ \mathcal {R}}^{U_{1}}\left( \mathbb {R}^{N}\right) \), \(j=1,2,3,4\).
References
Anastassiou, G.A.: Rate of convergence of fuzzy neural network operators, univariate case. J. Fuzzy Math. 10(3), 755–780 (2002)
Anastassiou, G.A.: Higher order fuzzy approximation by fuzzy wavelet type and neural network operators. Comput. Math. 48, 1387–1401 (2004)
Anastassiou, G.A.: Univariate fuzzy-random neural network approximation operators. Computers and Mathematics with Applications, Special issue Proceedings edited by G. Anastassiou of special session “Computational Methods in Analysis”, AMS meeting in Orlando. Florida, November 2002, vol. 48, pp. 1263–1283 (2004)
Anastassiou, G.A.: Higher order fuzzy Korovkin theory via inequalities. Commun. Appl. Anal. 10(2), 359–392 (2006)
Anastassiou, G.A.: Fuzzy Korovkin theorems and inequalities. J. Fuzzy Math. 15(1), 169–205 (2007)
Anastassiou, G.A.: Multivariate stochastic Korovkin theory given quantitatively. Math. Comput. Model. 48, 558–580 (2008)
Anastassiou, G.A.: Fuzzy Mathematics: Approximation Theory. Springer, Heidelberg (2010)
Anastassiou, G.A.: Intelligent Systems: Approximation by Artificial Neural Networks. Springer, Heidelberg (2011)
Anastassiou, G.A.: Univariate hyperbolic tangent neural network approximation. Math. Comput. Model. 53, 1111–1132 (2011)
Anastassiou, G.A.: Multivariate hyperbolic tangent neural network approximation. Comput. Math. 61, 809–821 (2011)
Anastassiou, G.A.: Multivariate sigmoidal neural network approximation. Neural Netw. 24, 378–386 (2011)
Anastassiou, G.A.: Higher order multivariate fuzzy approximation by multivariate fuzzy wavelet type and neural network operators. J. Fuzzy Math. 19(3), 601–618 (2011)
Anastassiou, G.A.: Univariate sigmoidal neural network approximation. J. Comput. Anal. Appl. 14(4), 659–690 (2012)
Anastassiou, G.A.: Rate of convergence of some multivariate neural network operators to the unit, revisited. J. Comput. Anal. Appl. 15(7), 1300–1309 (2013)
Anastassiou, G.A.: Higher order multivariate fuzzy approximation by basic neural network operators. CUBO 16(3), 21–35 (2014)
Anastassiou, G.A.: Multivariate fuzzy-random Quasi-interpolation neural network approximation operators. J. Fuzzy Math. 22(1), 167–184 (2014)
Anastassiou, G.: Multivariate fuzzy-random error function based neural network approximation. J. Fuzzy Math. 23(4), 917–935 (2015)
Anastassiou, G.A.: Intelligent Systems II: Complete Approximation by Neural Network Operators. SCI, vol. 608. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-20505-2
Anastassiou, G.A.: Intelligent Computations: Abstract Fractional Calculus, Inequalities, Approximations. Springer, Heidelberg (2018). https://doi.org/10.1007/978-3-319-66936-6
Anastassiou, G.A.: Algebraic function based Banach space valued ordinary and fractional neural network approximations. New Trends Math. Sci. 10(1), 100–125 (2022)
Anastassiou, G.A.: Gudermannian function activated Banach space valued ordinary and fractional neural network approximation. Adv. Nonlinear Var. Inequal. 25(2), 27–64 (2022)
Anastassiou, G.A.: Generalized symmetrical sigmoid function activated Banach space valued ordinary and fractional neural network approximation. Analele Universităţii Oradea, Fasc. Matematica (2022, accepted for publication)
Anastassiou, G.A.: Abstract multivariate Gudermannian function activated neural network approximations. Panamerican Math. J. (2022, accepted)
Anastassiou, G.A.: General multivariate arctangent function activated neural network approximations (2022, submitted)
Anastassiou, G.A., Kozma, R.: Abstract multivariate algebraic function activated neural network approximations (2022, submitted)
Anastassiou, G.A.: Generalized symmetrical sigmoid function activated neural network multivariate approximation (2022, submitted)
Chen, Z., Cao, F.: The approximation operators with sigmoidal functions. Comput. Math. Appl. 58, 758–765 (2009)
Dudley, R.M.: Real Analysis and Probability. Wadsworth & Brooks/Cole Mathematics Series. Pacific Grove, California (1989)
Dunning, A.J., Kensler, J., Goudeville, L., Bailleux, F.: Some extensions in continuous methods for immunological correlates of protection. BMC Med. Res. Methodol. 15(107) (2015). https://doi.org/10.1186/s12874-015-0096-9
Gal, S.: Approximation theory in fuzzy setting, Chapter 13. In: Anastassiou, G. (ed.) Handbook of Analytic-Computational Methods in Applied Mathematics, pp. 617–666. Chapman & Hall/CRC, Boca Raton (2000)
Haykin, S.: Neural Networks: A Comprehensive Foundation, 2nd edn. Prentice Hall, New York (1998)
Mitchell, T.M.: Machine Learning. WCB-McGraw-Hill, New York (1997)
McCulloch, W., Pitts, W.: A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 7, 115–133 (1943)
Weisstein, E.W.: Gudermannian, MathWorld
Congxin, W., Zengtai, G.: On Henstock integral of interval-valued functions and fuzzy valued functions. Fuzzy Sets Syst. 115(3), 377–391 (2000)
Wu, C., Gong, Z.: On Henstock integral of fuzzy-number-valued functions (I). Fuzzy Sets Syst. 120(3), 523–532 (2001)
Wu, C., Ma, M.: On embedding problem of fuzzy number space: part 1. Fuzzy Sets Syst. 44, 33–38 (1991)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Anastassiou, G.A. (2023). Multivariate Fuzzy-Random and Stochastic Arctangent, Algebraic, Gudermannian and Generalized Symmetric Activation Functions Induced Neural Network Approximations. In: Singh, J., Anastassiou, G.A., Baleanu, D., Kumar, D. (eds) Advances in Mathematical Modelling, Applied Analysis and Computation . ICMMAAC 2022. Lecture Notes in Networks and Systems, vol 666. Springer, Cham. https://doi.org/10.1007/978-3-031-29959-9_1
Download citation
DOI: https://doi.org/10.1007/978-3-031-29959-9_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-29958-2
Online ISBN: 978-3-031-29959-9
eBook Packages: EngineeringEngineering (R0)