Multistability of delayed complex-valued recurrent neural networks with discontinuous real-imaginary-type activation functions
Huang Yu-Jiao†, Hu Hai-Gen
College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China

Corresponding author. E-mail: hyj0507@zjut.edu.cn

*Project supported by the National Natural Science Foundation of China (Grant Nos. 61374094 and 61503338) and the Natural Science Foundation of Zhejiang Province, China (Grant No. LQ15F030005).

Abstract

In this paper, the multistability issue is discussed for delayed complex-valued recurrent neural networks with discontinuous real-imaginary-type activation functions. Based on a fixed theorem and stability definition, sufficient criteria are established for the existence and stability of multiple equilibria of complex-valued recurrent neural networks. The number of stable equilibria is larger than that of real-valued recurrent neural networks, which can be used to achieve high-capacity associative memories. One numerical example is provided to show the effectiveness and superiority of the presented results.

PACS: 07.05.Mh; 02.30.Ks
Keyword: complex-valued recurrent neural network; discontinuous real-imaginary-type activation function; multistability; delay
1.Introduction

In the past few decades, recurrent neural networks with their various generalizations have been extensively investigated due to their potential applications in pattern classification, associative memory, reconstruction of moving images, signal processing, parallel computation, and solving some classes of optimization problems.[15] Associative memories are brain-style devices, which are designed to memorize a set of patterns such that the stored patterns can be retrieved with the initial probes containing sufficient information about the patterns. In the associative memory based on recurrent neural networks, the stable output vectors are designed as the desired patterns. Hence, coexistence of many equilibria is a necessary feature in the applications of neural networks to associative memory storage and pattern recognition. Multistability of real-valued recurrent neural networks has been extensively investigated.[613]

As the extension of real-valued recurrent neural networks, complex-valued recurrent neural networks have been investigated.[1420] Complex-valued recurrent neural networks have specific applications different from those in real-valued recurrent neural networks. Complex-valued recurrent neural networks are most suitable for robotics having rotational articulation, wave information processing where the phase carries primary information, etc. Although a complex number is presented as an ordered pair of two real numbers, the most significant advantage of complex-valued neural networks originates not from the two-dimensionality of the complex plane, but from the fact that the multiplication of the synaptic weights, which is the elemental process at the synapses of neurons that construct the whole network.[21] This fact reduces the degree of freedom in learning and self-organization, in comparison with a double-dimensional real-valued neural network. Complex-valued recurrent neural networks extend the application fields steadily. General associative memories are also making progress in their improvement.[22] However, to the best of the authors’ knowledge, multistability for complex-valued recurrent neural networks was seldom considered.

It is well known that activation functions play an important role in the dynamical analysis of recurrent neural networks. Dynamics of neural networks rely heavily on the structures of activation functions. Two kinds of activation functions have been considered for recurrent neural networks, that is, the continuous activation functions and the discontinuous activation functions, respectively. As is well known, recurrent neural networks with discontinuous activation functions are important and do frequently arise in applications when dealing with dynamical systems possessing high-slope nonlinear elements. For this reason, much effort has been devoted to analyzing the dynamical behavior of neural networks with discontinuous activation functions.[2326] However, multistability for complex-valued recurrent neural networks with discontinuous real-imaginary-type activation functions was seldom considered.

It is well known that time delays of neural networks are unavoidable because of the finite speed of a signal’ s switch and transmission. Therefore, the studies of delayed systems have become a hot topic of great theoretical and practical importance.[2730]

Multistability of recurrent neural networks is an important issue when it comes to associative memories. Increasing storage capacity is a fundamental problem. The stable output vectors of neural networks are designed as the desired patterns. Hence, the number of stale equilibria is expected to be very large. Because the parameters of complex-valued recurrent neural networks are composed of real parts and imaginary parts, which can have more stable equilibria than real-valued recurrent neural networks. Meanwhile, discontinuous complex-valued activation functions, in which the real parts and imaginary parts are piecewise constants, can make neural networks have more stable equilibria and larger attractive basins than continuous complex-valued activation functions, in which the real parts and imaginary parts are piecewise linear. Therefore, it is important to explore the existence and stability of multiple equilibria for complex-valued recurrent neural networks with discontinuous real-imaginary-type activation functions.

Motivated by the above consideration, the purpose of this paper is to explore the existence and stability of multiple equilibria for delayed complex-valued recurrent neural networks with discontinuous real-imaginary-type activation functions. The main contributions of this paper are depicted as follows. (i) A special class of discontinuous complex-valued activation functions is proposed. The real part and imaginary part of the activation function are piecewise constants. (ii) Sufficient conditions for the existence of stable equilibria for complex-valued recurrent neural networks with discontinuous real-imaginary-type activation functions are established. Based on the discussion of this paper, more stable equilibria of neural networks will be derived by using complex-valued parameters and discontinuous activation functions. Such a neural network is suitable for synthesizing high-capacity associative memories.

The remainder of this paper consists of the following sections. Section 2 describes some preliminaries and problem formulations. The main results are stated in Section 3. One example is used to show the effectiveness of the obtained results in Section 4. Finally, in Section 5, we draw our conclusions.

2.Preliminaries and problem formulations

In this paper, we consider the following n-neuron complex-valued recurrent neural networks with constant delays

where i = 1, 2, … , n; z(t) = (z1(t), z2(t), … , zn(t))T is the neuron state vector; D = diag(d1, d2, … , dn)∈ with di > 0; A = (aij)n× n and B = (bij)n× n are the connection weight matrices; τ ij is the constant delay with 0 < τ ijτ (τ = max1i, jn{τ ij}); f (z(t)) = (f1(z1(t)), f2(z2(t)), … , fn(zn(t)))T: denotes the neuron activation function; u = (u1, u2, … , un)T stands for the external constant input.

In the complex domain, according to Liouville’ s theorem, [31] if f (z) is bounded and analytic at all z, then f (z) is a constant function, which is clearly not suitable. Therefore, in place of the analytic function, we construct the following real-imaginary-type activation function,

where

in Eq.  (2) i denotes the imaginary unit, i.e. are integers which satisfy are constants with .

Since the real part and imaginary part of activation function  (2) have individual properties, we can separate system (1) into its real and imaginary parts. By separating the state, the connection weight, the activation function, and the external input, we get a real part xi(t) and an imaginary part yi(t) for zi(t), respectively; and are the real part and imaginary part of aij, respectively; and are the real part and imaginary part of bij, respectively; and are the real part and imaginary part of ui, respectively. Therefore, system  (1) can be rewritten as follows for i = 1, 2, … , n,

Since and are allowed to have points of discontinuity, we need to specify what we mean by the solution of the system having a discontinuous right-side. For this purpose, we consider the solution of delayed complex-valued recurrent neural networks  (1) in the sense of Filippov.[32]

Definition 1 A function

T ∈ (t0, + ∞ ] is a solution of system (1) on [t0τ , T) if,

i) z(t) is continuous on [t0τ , T] and absolutely continuous on [t0, T);

ii) there exist measurable functions

such that and for almost all (a.a.) t ∈ [t0τ , T), and

for a.a. t ∈ [t0, T), where xi(t) = Re(zi(t)), yi(t) = Im(zi(t)), K[E] represents closure of the convex hull of E.

Definition 2 (Initial value problem) For any continuous functions

and any measurable selections

such that , for a.a. θ ∈ [t0τ , t0] by an initial value problem associated to system (1) with initial condition φ (θ ), ψ (θ ), γ (θ ), δ (θ ), we mean the following problem: find a couple of functions , such that z(t) = x(t) + y(t) i is a solution of system (1) on [t0τ , T) for some T > t0, λ (t) is an output associated to x(t), μ (t) is an output associated to y(t) and

where K[E] represents closure of the convex hull of E.

Definition 3 The equilibrium point of system  (1) is said to be locally exponentially stable in region ⅅ , if there exist constants α > 0, β > 0 such that for any tt0 ≥ 0,

where z(t; ϕ ) is the solution of system  (1) with any initial condition ϕ (s), s ∈ 𝒞 ([t0τ , t0], ⅅ ).

Lemma 1[33] Let ⅅ be a bounded and closed set in ℝ n, and H be a mapping on complete matric space (ⅅ , ‖ · ‖ ), where for any x, y ∈ ⅅ , | | xy| | = max1in {| xiyi| } is measurement in ⅅ . If H(ⅅ ) ⊂ ⅅ and there exists a positive constant α < 1 such that for any x, y ∈ ⅅ , | | H(x) − H(y)| | ≤ α | | xy| | , then there exists a unique ∈ ⅅ such that H() = .

For the convenience of description, denote . Let

for i = 1, 2, … , n. Denote

where

It is easy to see that Φ is composed of (α β )n subregions. In particular, when n = 1, α = β = 3, Φ is composed of 9 subregions (see Fig.  1).

Fig.  1. The phase plane diagram of ℂ 1 with α = β = 3, Φ is composed of Ω 10101, Ω 10112, Ω 10123, Ω 11201, Ω 11212, Ω 11223, Ω 12301, Ω 12312, and Ω 12323.

3.Main results

In this section, we will investigate the existence and stability of multiple equilibria for complex-valued recurrent neural network (1) with discontinuous activation function (2). Sufficient criteria will be proposed to ensure system (1) can have (α β )n locally exponentially stable equilibria.

By the partition of the state space and Lemma 1, the following theorem can be derived.

Theorem 1 There exist (α β )n equilibria for system  (1) with activation function  (2), if the following conditions hold,

where i = 1, 2, … , n, m = 1, 2, … , α , = 1, 2, … , β .

Proof According to conditions (4)– (7), there exists a small constant

where i = 1, 2, … , n, m = 1, 2, … , α , = 1, 2, … , β .

For the convenience of description, denote

let

for i = 1, 2, … , n. Denote

where

It is obvious that is made up of (α β )n subregions and is bounded in Φ .

For any subregion Ω , we will show that there exists an equilibrium point for system  (1) in Ω . Let Ω 1 = {x| (x, y) ∈ Ω }, Ω 2 = {y| (x, y) ∈ Ω }. Denote

where

We will show that H(Ω 1) ⊂ Ω 1, G(Ω 2) ⊂ Ω 2. For any x = (x1, x2, … , xn)TΩ 1, xi belongs to one of the following three cases.

Case 1

Case 2 There exists m ∈ {2, 3, … , α − 1} such that .

Case 3.

When , it follows from Eq.  (8) that

Therefore, .

When , m ∈ {2, 3, … , α − 1}, it follows from Eqs.  (8) and (9) that

Therefore, .

When , it follows from Eq.  (9) that

Therefore, .

Based on the above discussion and the arbitrariness of x, we can derive H(Ω 1) ⊂ Ω 1. Similarly, it also holds that G(Ω 2) ⊂ Ω 2. For any , it holds that for i = 1, 2, … , n. Since we only need to discuss the existence of equilibrium points in Ω , when fixing i ∈ {1, 2, … , n}, we have for any. Choose α = 0.5. We have | Hi(x(1)) − Hi(x(2))| ≤ α | | x(1)x(2)| | . Therefore, | | H(x(1)) − H(x(2))| | ≤ α | | x(1)x(2)| | . According to Lemma  1, there exists a unique fixed point such that H() = . By the similar analysis, there exists a unique fixed point such that G() = . Based on the above discussion, we can derive that y = (y1, y2, … , yn)T in Eq.  (12) and x = (x1, x2, … , xn)T in Eq.  (13) are allowed to be arbitrary. So, we make y = in Eq.  (12) and x = in Eq.  (13). Then, = + i is a unique equilibrium point of system  (1) in Ω . Due to the arbitrariness of Ω and ε , there exist (α β )n equilibria in ℂ n, which locate in each subregion of Φ . The proof is completed.

Remark 1 Conditions (4)– (7) assure each subregion of Φ has one equilibrium point. These equilibria exist at the same time. The number of equilibria replies on α and β of the activation function. Therefore, activation functions have an important influence for the existence of multiple equilibria.

Next, we investigate the stability of multiple equilibria for system (1) with activation function  (2).

Theorem 2 There exist (α β )n locally exponentially stable equilibria for system  (1) with activation function  (2), if conditions (4)– (7) hold.

Proof For any subregion Ω Φ , according to Theorem 1, there exists an equilibrium point = + i, (, ) ∈ Ω . Let v(t) = z(t) − . When (x(t), y(t)) ∈ Ω , and (j = 1, 2, … , n) are constants. That is, , . Therefore, system  (3) can be rewritten as

which implies is locally exponentially stable. Due to the arbitrariness of Ω , we conclude that system  (1) has (α β )n locally exponentially stable equilibria, which are located in each invariant subregion of Φ . The proof is completed.

Remark 2 It can be seen that all equilibria in Φ are locally exponentially stable. Each subregion of Φ is an attractive basin of a stable equilibrium point. Stable output vectors can be designed as desired patterns in associative memory. In order to achieve high-capacity associative memory, the number of stable equilibria is expected to be large. In Theorem  2, conditions  (4)– (7) ensure each equilibrium point of Φ is stable. Therefore, such complex-valued recurrent neural networks with discontinuous real-imaginary-type activation functions are suitable for synthesizing high-capacity associative memories.

For the case of system  (1) without delay, i.e., bij = 0, i, j = 1, 2, … , n, we have the following corollary by using Theorem 1 and Theorem 2.

Corollary 1 There exist (α β )n locally exponentially stable equilibria for system  (1) without delay and with activation function  (2), if the following conditions hold,

where i = 1, 2, … , n, m = 1, 2, … , α , = 1, 2, … , β .

In particular, when yi = 0, , , , , system (1) can be rewritten as the following real-valued recurrent neural network,

where i = 1, 2, … , n.. By using Theorem 1 and Theorem 2, we can derive the following corollary.

Corollary 2 There exist α n locally exponentially stable equilibria for system (18) with activation function (2), if the following conditions hold,

where i = 1, 2, … , n, m = 1, 2, … , α .

Remark 3 Due to the two-dimensionality of the complex plane, real-valued recurrent neural networks can only be seen as one special case of complex-valued recurrent neural networks. Multistability of system  (18) has been investigated in Refs.  [9] and [10]. Hence, Theorem  1 and Theorem  2 are the improvement and extension of the existing stability results about real-valued recurrent neural networks in the literature.

Remark 4 In this paper, the number of locally exponentially stable equilibria is (α β )n, which is larger than kn in Ref.  [25], where α ≥ 1, β ≥ 1, k ≥ 1. Hence, complex-valued neural networks  (1) are more suitable for high-capacity associative memories than real-valued neural networks.

4.Illustrative examples

In this section, we will give one example to illustrate the effectiveness of our results.

Consider the following 2-neuron complex-valued neural network,

where the discontinuous activation function f1(· ) = f2(· ) = fR(· ) + fI(· )i,

System  (21) satisfies conditions (4)– (7). According to Theorems  1 and 2, system  (21) has 36 locally exponentially stable equilibria. The dynamics of system  (21) are illustrated in Figs.  2– 5, where evolutions of 110 initial conditions have been tracked.

Fig.  2. State trajectories of Re(z1(t)) of system  (21).

Fig.  3. State trajectories of Im(z1(t)) of system  (21).

From Figs.  2– 5, it can be seen that the real part and the imaginary part of each neuron state of system  (21) converge into one of 36 locally stable states, respectively. Stable states corresponding to the real parts of neuron states locate in (− ∞ , − 1), (− 1, 1), and (1, + ∞ ). Each interval has 12 stable states. Stable states corresponding to imaginary parts of neuron states locate in (− ∞ , 1) and (1, + ∞ ). Each interval has 18 stable states. The partition of these intervals is related with the segmented interval of the real part and the imaginary part of the activation function. Each stable state is related with the value of the activation function. For example, according to the following stationary equation of system (21),

the stable state of Re(z1) with the initial state located in (− ∞ , − 1) is decided by fI(Im(z1)), fR(Re(z2)), fI(Im(z2)). The combination of these three functions has 12 results. Therefore, there exist 12 stable states of Re(z1) in (− ∞ , − 1).

Fig.  4. State trajectories of Re(z2(t)) of system  (21).

Fig.  5. State trajectories of Im(z2(t)) of system  (21).

When all delays are zero, system  (21) becomes the following system,

System (23) satisfies conditions (14)– (17). According to the Corollary  1, system  (23) still has 36 locally exponentially stable equilibria. It can be seen clearly that delays cannot affect the number of equilibria,

The authors investigated the multistability of 2-neuron real-valued neural networks with a 3-level discontinuous activation function in the simulation example of Ref.  [25]. There were 9 locally exponentially stable equilibrium points in Ref.  [25], which is less than 36 locally exponentially stable equilibria in the example of this paper. So, complex-valued neural networks are more suitable for high-capacity associative memories than real-valued neural networks.

5.Conclusion

In this paper, we have discussed the multistability issue of delayed complex-valued recurrent neural networks. A special class of discontinuous real-imaginary-type activation functions has been proposed. By using this neural network model, sufficient criteria have been established to ensure the existence of (α β )n locally exponentially stable equilibria. The number of stable equilibria is larger than that of real-valued recurrent neural networks. Several extensions would be welcome in future works: such as, studying the influence of delay on the attractive basins of stable equilibria of complex-valued recurrent neural networks, and studying the design procedures of achieving associative memories by using multistability results of delayed complex-valued recurrent neural networks in this paper.

Reference
1 Kwon O M, Kwon J W and Kim S H 2011 Chin. Phys. B 20 050505 DOI:10.1088/1674-1056/20/5/050505 [Cited within:1]
2 Zhang H G, Yang F S, Liu X D and Zhang Q L 2013 IEEE Trans. Neural Netw. Learn. Syst. 24 513 DOI:10.1109/TNNLS.2012.2236571 [Cited within:1]
3 Zhang H G, Wang Z S and Liu D R 2014 IEEE Trans. Neural Netw. Learn. Syst. 25 1229 DOI:10.1109/TNNLS.2014.2317880 [Cited within:1]
4 Lakshmanan S and Balasubramaniam P 2011 Chin. Phys. B 20 040204 DOI:10.1088/1674-1056/20/4/040204 [Cited within:1]
5 Lakshmanan S, Park J H, Rihan F A and Rakkiyappan R 2014 Chin. Phys. B 23 070205 DOI:10.1088/1674-1056/23/7/070205 [Cited within:1]
6 Cheng C Y, Lin K H and Shih C W 2007 Physica D 225 61 DOI:10.1016/j.physd.2006.10.003 [Cited within:1]
7 Cao J D, Feng G and Wang Y Y 2008 Physica D 237 1734 DOI:10.1016/j.physd.2008.01.012 [Cited within:1]
8 Huang Y J, Zhang H G and Wang Z S 2012 Chin. Phys. B 21 070701 DOI:10.1088/1674-1056/21/7/070701 [Cited within:1]
9 Zeng Z G, Huang T W and Zheng W X 2010 IEEE Trans. Neural Netw. 21 1371 DOI:10.1109/TNN.2010.2054106 [Cited within:1]
10 Zeng Z G and Zheng W X 2013 IEEE Trans. Neural Netw. Learn. Syst. 24 1749 DOI:10.1109/TNNLS.2013.2262638 [Cited within:1]
11 Wang L L and Chen T P 2014 Neural Netw. 53 109 DOI:10.1016/j.neunet.2014.02.001 [Cited within:1]
12 Wang L L and Chen T P 2015 Neurocomputing 152 222 DOI:10.1016/j.neucom.2014.10.075 [Cited within:1]
13 Zhang H G, Huang Y J, Cai T Y and Wang Z S 2014 Int. J. Mod. Phys. B 28 1450118 DOI:10.1142/S0217979214501185 [Cited within:1]
14 Jankowski S, Lozowski A and Zurada J M 1996 IEEE Trans. Neural Netw. 7 1491 DOI:10.1109/72.548176 [Cited within:1]
15 Lee D L 2006 IEEE Trans. Neural Netw. 17 1341 DOI:10.1109/TNN.2006.878786 [Cited within:1]
16 Hu J and Wang J 2012 IEEE Trans. Neural Netw. Learn. Syst. 23 853 DOI:10.1109/TNNLS.2012.2195028 [Cited within:1]
17 Hirose A and Yoshida S 2012 IEEE Trans. Neural Netw. Learning Syst. 23 541 DOI:10.1109/TNNLS.2012.2183613 [Cited within:1]
18 Zhou B and Song Q 2013 IEEE Trans. Neural Netw. Learning Syst. 24 1227 DOI:10.1109/TNNLS.2013.2247626 [Cited within:1]
19 Chen X and Song Q 2013 Neurocomputing 121 254 DOI:10.1016/j.neucom.2013.04.040 [Cited within:1]
20 Huang Y J, Zhang H G and Wang Z S 2014 Appl. Math. Comput. 229 187 DOI:10.1016/j.amc.2013.12.027 [Cited within:1]
21 Hirose A 2011 Front. Electr. Electron. Eng. China 6 171 DOI:10.1007/s11460-011-0125-3 [Cited within:1]
22 Lee D 2006 IEEE Trans. Neural Netw. 17 1341 DOI:10.1109/TNN.2006.878786 [Cited within:1]
23 Forti M 2007 Int. J. Circuit Theory Appl. 35 105 DOI:10.1002/(ISSN)1097-007X [Cited within:1]
24 Bao G and Zeng Z G 2012 Neurocomputing 77 101 DOI:10.1016/j.neucom.2011.08.026 [Cited within:1]
25 Huang Y J, Zhang H G and Wang Z S 2012 Neurocomputing 91 21 DOI:10.1016/j.neucom.2012.02.016 [Cited within:3]
26 Duan L and Huang L 2014 Neurocomputing 123 318 DOI:10.1016/j.neucom.2013.07.026 [Cited within:1]
27 Wen S P, Zeng Z G, Huang T W and Zhang Y D 2014 IEEE Trans. Fuzzy Syst. 22 1704 DOI:10.1109/TFUZZ.2013.2294855 [Cited within:1]
28 Wen S P, Zeng Z G, Huang T W, Meng Q G and Yao W 2015 IEEE Trans. Neural Netw. Learn. Syst. 26 1493 DOI:10.1109/TNNLS.2014.2387355 [Cited within:1]
29 Wen S P, Yu X H, Zeng Z G and Wang J J 2015 IEEE Trans. Ind. Electron. DOI:10.1109/TIE.2015.2399394 [Cited within:1]
30 Wen S P, Zeng Z G, Huang T W, Yu X H and Xiao M Q 2015 IEEE Trans. Fuzzy Syst. DOI:10.1109/TFUZZ.2015.2417913 [Cited within:1]
31 Ahlfors L 1979 Complex Analysis New York McGraw-Hill [Cited within:1]
32 Filippov A 1988 Differential Equations with Discontinuous Righthand Sides. Mathematics and Its Applications (Soviet Series) Boston MA Kluwer Academic [Cited within:1]
33 Yosida K 1978 Functional Analysis Berlin Springer-Verlag [Cited within:1]