- Research Article
- Open Access

# Dynamic Analysis of Stochastic Reaction-Diffusion Cohen-Grossberg Neural Networks with Delays

- Jie Pan
^{1, 2}Email author and - Shouming Zhong
^{1}

**2009**:410823

https://doi.org/10.1155/2009/410823

© J. Pan and S. Zhong 2009

**Received:**13 June 2009**Accepted:**2 September 2009**Published:**11 October 2009

## Abstract

Stochastic effects on convergence dynamics of reaction-diffusion Cohen-Grossberg neural networks (CGNNs) with delays are studied. By utilizing Poincaré inequality, constructing suitable Lyapunov functionals, and employing the method of stochastic analysis and nonnegative semimartingale convergence theorem, some sufficient conditions ensuring almost sure exponential stability and mean square exponential stability are derived. Diffusion term has played an important role in the sufficient conditions, which is a preeminent feature that distinguishes the present research from the previous. Two numerical examples and comparison are given to illustrate our results.

## Keywords

- Dirichlet Boundary Condition
- Equilibrium Solution
- Exponential Stability
- Neumann Boundary Condition
- Diffusion Term

## 1. Introduction

In the recent years, the problems of stability of delayed neural networks have received much attention due to its potential application in associative memories, pattern recognition and optimization. A large number of results have appeared in literature, see, for example, [1–14]. As is well known, a real system is usually affected by external perturbations which in many cases are of great uncertainty and hence may be treated as random [15–17]. As pointed out by Haykin [18] that in real nervous systems synaptic transmission is a noisy process brought on by random fluctuations from the release of neurotransmitters and other probabilistic causes, it is of significant importance to consider stochastic effects for neural networks. In recent years, the dynamic behavior of stochastic neural networks, especially the stability of stochastic neural networks, has become a hot study topic. Many interesting results on stochastic effects to the stability of delayed neural networks have been reported (see [16–23]).

In the factual operations, on other hand, diffusion phenomena could not be ignored in neural networks and electric circuits once electrons transport in a nonuniform electromagnetic field. Thus, it is essential to consider state variables varying with time and space variables. The delayed neural networks with diffusion terms can commonly be expressed by partial functional differential equation (PFDE). To study the stability of delayed reaction-diffusion neural networks, for instance, see [24–31], and references therein.

Based on the above discussion, it is significant and of prime importance to consider the stochastic effects on the stability property of the delayed reaction-diffusion networks. Recently, Sun et al. [32, 33] have studied the problem of the almost sure exponential stability and the moment exponential stability of an equilibrium solution for stochastic reaction-diffusion recurrent neural networks with continuously distributed delays and constant delays, respectively. Wan et al. have derived the sufficient condition of exponential stability of stochastic reaction-diffusion CGNNs with delay [34, 35]. In [36], the problem of stochastic exponential stability of the delayed reaction-diffusion recurrent neural networks with Markovian jumping parameters have been investigated. In [32–36], unfortunately, reaction-diffusion terms were omitted in the deductions, which result in that the criteria of obtained stability do not contain the diffusion terms. In other words, the diffusion terms do not take effect in their results. The same cases appear also in other research literatures on the stability of reaction-diffusion neural network [24–31].

Motivated by the above discussions, in this paper, we will further investigate the convergence dynamics of stochastic reaction-diffusion CGNNs with delays, where the activation functions are not necessarily bounded, monotonic, and differentiable. Utilizing Poincaré inequality and constructing appropriate Lyapunov functionals, some sufficient conditions on the almost surely and mean square exponential stability for the equilibrium point are established. The results show that diffusion terms have contributed to the almost surely and mean square exponential stability criteria. Two examples have been provided to illustrate the effectiveness of the obtained results.

The rest of this paper is organized as follows. In Section 2, a stochastic delayed reaction-diffusion CGNNs model is described, and some preliminaries are given. In Section 3, some sufficient conditions to guarantee the mean square and almost surely exponential stability of equilibrium point for the reaction-diffusion delayed CGNNs are derived. Examples and comparisons are given in Section 4. Finally, in Section 5, conclusions are given.

## 2. Model Description and Preliminaries

- (i)
be an open bounded domain in with smooth boundary , and denotes the measure of . ;

- (ii)
is the space of real Lebesgue measurable functions on which is a Banach space for the -norm , ;

- (iii)
, where , . the closure of in ;

- (iv)
is the space of continuous functions which map into with the norm , for any ;

- (v)
and be an -measurable -valued random variable, where, for example, restricted on , and be the Banach space of continuous and bounded functions with the norm , where , for any , ;

- (vi)
is the gradient operator, for . . is the Laplace operator, for .

where , corresponds to the number of units in a neural network; is a space variable, corresponds to the state of the th unit at time and in space ; corresponds to the transmission diffusion coefficient along the th neuron; represents an amplification function; is an appropriately behavior function; , denote the connection strengths of the th neuron on the th neuron, respectively; , denote the activation functions of th neuron at time and in space ; corresponds to the transmission delay and satisfies ( is a positive constant); is the constant input from outside of the network. Moreover, is an -dimensional Brownian motion defined on a complete probability space with the natural filtration generated by the process , where we associate with the canonical space generated by all , and denote by the associated -algebra generated by with the probability measure . The boundary condition is given by (Dirichlet type) or (Neumann type), where denotes the outward normal derivative on .

for .

for .

- (H1)
each function is bounded, positive and continuous, that is, there exist constants , such that , for , ,

- (H2)
and , for ,

- (H3)
, are bounded, and , , are Lipschitz continuous with Lipschitz constant , , , for ,

- (H4)
, for .

Suppose that system (2.1) satisfies assumptions (H1)–(H4), then equilibrium point of model (2.3) is also a unique equilibrium point of system (2.1).

By the theory of stochastic differential equations, see [15, 37], it is known that under the conditions (H1)–(H4), model (2.1) has a global solution denoted by or simply , or , if no confusion should occur. For the effects of stochastic forces to the stability property of delayed CGNNs model (2.1), we will study the almost sure exponential stability and the mean square exponential stability of their equilibrium solution in the following sections. For completeness, we give the following definitions [33], in which denotes expectation with respect to .

Definition 2.1.

Definition 2.2.

When and , it is usually called the exponential stability in mean value and mean square, respectively.

The following lemmas are important in our approach.

Lemma 2.3 (nonnegative semimartingale convergence theorem [16]).

where a.s. denotes . In particular, if a.s., then for almost all and , that is, both and converge to finite random variables.

Lemma 2.4 (Poincaré inequality).

Proof.

We just give a simple sketch here.

Case 1.

Case 2.

Under the Dirichlet boundary condition, that is, . By the same may, we can obtained the inequality.

This completes the proof.

Remark 2.5. (i) The lowest positive eigenvalue in the boundary problem (2.9) is sometimes known as the first eigenvalue. (ii) The magnitude of is determined by domain . For example, let Laplacian on , if and , respectively, then and [38, 39]. (iii) Although the eigenvalue of the laplacian with the Dirichlet boundary condition on a generally bounded domain cannot be determined exactly, a lower bound of it may nevertheless be estimated by , where is a surface area of the unit ball in , is a volume of domain [40].

In Section 4, we will compare the results between this paper and previous literatures. To this end, we recall some previous results as follows (according to the symbols in this paper).

In [35], Wan and Zhou have studied the problem of convergence dynamics of model (2.1) with the Neumann boundary condition and obtained the following result (see [35, Theorem 3.1]).

Proposition 2.6.

- (A)
, , where , , , , , , , . Also, denotes the spectral radius of a square matrix .

Then model (2.1) is mean value exponentially stable.

Remark 2.7.

It should be noted that condition (A) in Proposition 2.6 is equivalent to is a nonsingular -matrix, where . Thus, the following result is treated as a special case of Proposition 2.6.

Proposition 2.8 (see [33, Theorem 3.1]).

- (B)
is a nonsingular -matrix, where , , for .

Then model (2.2) is almost surely exponentially stable.

Remark 2.9.

It is obvious that conditions (A) and (B) are irrelevant to the diffusion term. In other words, the diffusion term does not take effect in Propositions 2.6 and 2.8.

## 3. Main Results

Theorem 3.1.

- (H5)
, for any ,

where is the lowest positive eigenvalue of problem (2.9), , .

Then model (2.1) is almost surely exponentially stable.

Proof.

for .

for .

The proof is complete.

Theorem 3.2.

Under the conditions of Theorem 3.1, model (2.1) is mean square exponentially stable.

Proof.

Let .

The proof is completed.

By the similar way of the proof of Theorem 3.1, it is easy to prove the following results.

Theorem 3.3.

- (H6)
, .

Then model (2.2) is almost surely exponentially stable and mean square exponentially stable.

Remark 3.4.

In the proof of Theorem 3.1, by Poincaré inequality, we have obtained (see (3.6)). This is an important step that results in the condition of Theorem 3.1 including the diffusion terms.

Remark 3.5.

respectively, which cannot guarantee the mean square exponential stability of the equilibrium solution of models (2.1) and (2.2). Thus, as we can see form Theorems 3.1, 3.2, and 3.3, reaction-diffusion terms do contribute the almost surely exponential stability and the mean square exponential stability of models (2.1) and (2.2), respectively. However, as we can see from Propositions 2.6 and 2.8, the diffusion term do not take effect in the convergence dynamics of delayed stochastic reaction-diffusion neural networks. Thus, the criteria what we proposed are less conservative and restrictive than Propositions 2.6 and 2.8.

Theorem 3.6.

- (H7)
, for any ,

holds, the equilibrium point of system (2.2) is globally exponentially stable.

Remark 3.7.

Theorem 3.6 shows that the globally exponential stability criteria on reaction-diffusion CGNNs with delays depend on the diffusion term. In exact words, diffusion terms have contributed to exponentially stabilization of reaction-diffusion CGNNs with delays. It should be noted that the authors in [24–28] have studied reaction-diffusion neural networks (including CGNNs and RNNs) with delays and obtained the sufficient condition of exponential stability. However, those sufficient condition are independent of the diffusion term. Obviously, the criteria what we proposed are less conservative and restrictive than those in [24–28].

## 4. Examples and Comparison

In order to illustrate the feasibility of our above established criteria in the preceding sections, we provide two concrete examples. Although the selection of the coefficients and functions in the examples is somewhat artificial, the possible application of our theoretical theory is clearly expressed.

Example 4.1.

It follows from Theorem 3.3 that the equilibrium solution of such system is almost surely exponentially stable and mean square exponential stable.

Remark 4.2.

it is well known, which cannot guarantee the mean square exponential stability of the equilibrium solution of model (4.1). Thus, as we can see in Example 4.1, the reaction-diffusion terms have contributed to the almost surely and mean square exponential stability of this model.

Example 4.3.

respectively. The remainder parameters and functions unchanged. According to Remark 2.5, we see that . By the same way of Example 4.1, equilibrium solution of model (4.5) is almost surely exponentially stable and mean square.

Now, we compare the results in this paper with Propositions 2.6 and 2.8.

The authors in [33, 35] have considered the stochastic delayed reaction-diffusion neural networks with Neumann boundary condition and obtained the sufficient conditions to guarantee the almost surely or mean value exponential stability. We notice that the conditions of Propositions 2.6 and 2.8 do not include the diffusion terms, hence, in principal, Propositions 2.6 and 2.8 could be applied to analyze the exponential stability of stochastic system (4.1), but could not be model (4.5) for its the Dirichlet boundary condition. Unfortunately, Propositions 2.6 and 2.8 are not applicable to ascertain the exponential stability of model (4.1).

That is, condition (A) of Proposition 2.6 does not hold.

is not a nonsingular -matrix. This implies that condition (A) of Proposition 2.6 is not satisfied.

Remark 4.4.

The above comparison shows that reaction-diffusion term contributes to the exponentially stabilization of a stochastic reaction-diffusion neural network and the previous results have been improved.

## 5. Conclusion

The problem of the convergence dynamics for the stochastic reaction-diffusion CGNNs with delays has been studied in this paper. This neural networks is quite general, and can be used to describe some well-known neural networks, including Hopfield neural networks, cellular neural networks, and generalized CGNNs. By Poincaré inequality and constructing suitable Lyapunov functional, we obtain some sufficient condition to ensure the almost sure and mean square exponential stability of the system. It is worth noting that the diffusion term has played an important role in the obtained conditions, a significant feature that distinguishes the results in this paper from the previous. Two examples are given to show the effectiveness of the results. Moreover, the methods in this paper can been used to consider other stochastic delayed reaction-diffusion neural network model with the Neumann or Dirichlet boundary condition.

## Declarations

### Acknowledgments

The authors would like to thank the editor and the reviewers for their detailed comments and valuable suggestions which have led to a much improved paper. This paper is supported by National Basic Research Program of China (2010CB732501).

## Authors’ Affiliations

## References

- Arik S, Orman Z:
**Global stability analysis of Cohen-Grossberg neural networks with time varying delays.***Physics Letters A*2005,**341**(5-6):410-421. 10.1016/j.physleta.2005.04.095MATHView ArticleGoogle Scholar - Chen Z, Ruan J:
**Global stability analysis of impulsive Cohen-Grossberg neural networks with delay.***Physics Letters A*2005,**345**(1–3):101-111.MATHView ArticleGoogle Scholar - Chen Z, Ruan J:
**Global dynamic analysis of general Cohen-Grossberg neural networks with impulse.***Chaos, Solitons & Fractals*2007,**32**(5):1830-1837. 10.1016/j.chaos.2005.12.018MATHMathSciNetView ArticleGoogle Scholar - Cohen MA, Grossberg S:
**Absolute stability of global pattern formation and parallel memory storage by competitive neural networks.***IEEE Transactions on Systems, Man, and Cybernetics*1983,**13**(5):815-826.MATHMathSciNetView ArticleGoogle Scholar - Huang T, Chan A, Huang Y, Cao J:
**Stability of Cohen-Grossberg neural networks with time-varying delays.***Neural Networks*2007,**20**(8):868-873. 10.1016/j.neunet.2007.07.005MATHView ArticleGoogle Scholar - Liao X, Li C, Wong K-W:
**Criteria for exponential stability of Cohen-Grossberg neural networks.***Neural Networks*2004,**17**(10):1401-1414. 10.1016/j.neunet.2004.08.007MATHMathSciNetView ArticleGoogle Scholar - Liu X, Wang Q:
**Impulsive stabilization of high-order hopfield-type neural networks with time-varying delays.***IEEE Transactions on Neural Networks*2008,**19**(1):71-79.View ArticleGoogle Scholar - Yang Z, Xu D:
**Impulsive effects on stability of Cohen-Grossberg neural networks with variable delays.***Applied Mathematics and Computation*2006,**177**(1):63-78. 10.1016/j.amc.2005.10.032MATHMathSciNetView ArticleGoogle Scholar - Zhang J, Suda Y, Komine H:
**Global exponential stability of Cohen-Grossberg neural networks with variable delays.***Physics Letters A*2005,**338**(1):44-50. 10.1016/j.physleta.2005.02.005MATHView ArticleGoogle Scholar - Zhou Q:
**Global exponential stability for a class of impulsive integro-differential equation.***International Journal of Bifurcation and Chaos*2008,**18**(3):735-743. 10.1142/S0218127408020616MATHMathSciNetView ArticleGoogle Scholar - Park JH, Kwon OM:
**Synchronization of neural networks of neutral type with stochastic perturbation.***Modern Physics Letters B*2009,**23**(14):1743-1751. 10.1142/S0217984909019909MATHView ArticleGoogle Scholar - Park JH, Kwon OM:
**Delay-dependent stability criterion for bidirectional associative memory neural networks with interval time-varying delays.***Modern Physics Letters B*2009,**23**(1):35-46. 10.1142/S0217984909017807MATHView ArticleGoogle Scholar - Park JH, Kwon OM, Lee SM:
**LMI optimization approach on stability for delayed neural networks of neutral-type.***Applied Mathematics and Computation*2008,**196**(1):236-244. 10.1016/j.amc.2007.05.047MATHMathSciNetView ArticleGoogle Scholar - Meng Y, Guo S, Huang L:
**Convergence dynamics of Cohen-Grossberg neural networks with continuously distributed delays.***Applied Mathematics and Computation*2008,**202**(1-2):188-199.MATHMathSciNetView ArticleGoogle Scholar - Arnold L:
*Stochastic Differential Equations: Theory and Applications*. John Wiley & Sons, New York, NY, USA; 1972.Google Scholar - Blythe S, Mao X, Liao X:
**Stability of stochastic delay neural networks.***Journal of the Franklin Institute*2001,**338**(4):481-495. 10.1016/S0016-0032(01)00016-3MATHMathSciNetView ArticleGoogle Scholar - Buhmann J, Schulten K:
**Influence of noise on the function of a "physiological" neural network.***Biological Cybernetics*1987,**56**(5-6):313-327. 10.1007/BF00319512MATHMathSciNetView ArticleGoogle Scholar - Haykin S:
*Neural Networks*. Prentice-Hall, Upper Saddle River, NJ, USA; 1994.MATHGoogle Scholar - Sun Y, Cao J:
**th moment exponential stability of stochastic recurrent neural networks with time-varying delays.***Nonlinear Analysis: Real World Applications*2007,**8**(4):1171-1185. 10.1016/j.nonrwa.2006.06.009MATHMathSciNetView ArticleGoogle Scholar - Wan L, Sun J:
**Mean square exponential stability of stochastic delayed Hopfield neural networks.***Physics Letters A*2005,**343**(4):306-318. 10.1016/j.physleta.2005.06.024MATHView ArticleGoogle Scholar - Wan L, Zhou Q:
**Convergence analysis of stochastic hybrid bidirectional associative memory neural networks with delays.***Physics Letters A*2007,**370**(5-6):423-432. 10.1016/j.physleta.2007.05.095View ArticleGoogle Scholar - Zhao H, Ding N:
**Dynamic analysis of stochastic bidirectional associative memory neural networks with delays.***Chaos, Solitons & Fractals*2007,**32**(5):1692-1702. 10.1016/j.chaos.2005.12.010MATHMathSciNetView ArticleGoogle Scholar - Zhou Q, Wan L:
**Exponential stability of stochastic delayed Hopfield neural networks.***Applied Mathematics and Computation*2008,**199**(1):84-89. 10.1016/j.amc.2007.09.025MATHMathSciNetView ArticleGoogle Scholar - Zhao H, Ding N:
**Dynamic analysis of stochastic Cohen-Grossberg neural networks with time delays.***Applied Mathematics and Computation*2006,**183**(1):464-470. 10.1016/j.amc.2006.05.087MATHMathSciNetView ArticleGoogle Scholar - Song Q, Cao J:
**Global exponential robust stability of Cohen-Grossberg neural network with time-varying delays and reaction-diffusion terms.***Journal of the Franklin Institute*2006,**343**(7):705-719. 10.1016/j.jfranklin.2006.07.001MATHMathSciNetView ArticleGoogle Scholar - Song Q, Cao J:
**Exponential stability for impulsive BAM neural networks with time-varying delays and reaction-diffusion terms.***Advances in Difference Equations*2007,**2007:**-18.Google Scholar - Liang J, Cao J:
**Global exponential stability of reaction-diffusion recurrent neural networks with time-varying delays.***Physics Letters A*2003,**314**(5-6):434-442. 10.1016/S0375-9601(03)00945-9MATHMathSciNetView ArticleGoogle Scholar - Wang L, Xu D:
**Global exponential stability of Hopfield reaction-diffusion neural networks with time-varying delays.***Science in China. Series F*2003,**46**(6):466-474. 10.1360/02yf0146MATHMathSciNetView ArticleGoogle Scholar - Yang J, Zhong S, Luo W:
**Mean square stability analysis of impulsive stochastic differential equations with delays.***Journal of Computational and Applied Mathematics*2008,**216**(2):474-483. 10.1016/j.cam.2007.05.022MATHMathSciNetView ArticleGoogle Scholar - Zhao H, Wang K:
**Dynamical behaviors of Cohen-Grossberg neural networks with delays and reaction-diffusion terms.***Neurocomputing*2006,**70**(1–3):536-543.View ArticleGoogle Scholar - Zhou Q, Wan L, Sun J:
**Exponential stability of reaction-diffusion generalized Cohen-Grossberg neural networks with time-varying delays.***Chaos, Solitons & Fractals*2007,**32**(5):1713-1719. 10.1016/j.chaos.2005.12.003MATHMathSciNetView ArticleGoogle Scholar - Lv Y, Lv W, Sun J:
**Convergence dynamics of stochastic reaction-diffusion recurrent neural networks in continuously distributed delays.***Nonlinear Analysis: Real World Applications*2008,**9**(4):1590-1606. 10.1016/j.nonrwa.2007.04.003MATHMathSciNetView ArticleGoogle Scholar - Sun J, Wan L:
**Convergence dynamics of stochastic reaction-diffusion recurrent neural networks with delays.***International Journal of Bifurcation and Chaos*2005,**15**(7):2131-2144. 10.1142/S0218127405013332MATHMathSciNetView ArticleGoogle Scholar - Wan L, Zhou Q, Sun J:
**Mean value exponential stability of stochastic reaction-diffusion generalized Cohen-Grossberg neural networks with time-varying delay.***International Journal of Bifurcation and Chaos*2007,**17**(9):3219-3227. 10.1142/S021812740701897XMATHMathSciNetView ArticleGoogle Scholar - Wan L, Zhou Q:
**Exponential stability of stochastic reaction-diffusion Cohen-Grossberg neural networks with delays.***Applied Mathematics and Computation*2008,**206**(2):818-824. 10.1016/j.amc.2008.10.002MATHMathSciNetView ArticleGoogle Scholar - Wang L, Zhang Z, Wang Y:
**Stochastic exponential stability of the delayed reaction-diffusion recurrent neural networks with Markovian jumping parameters.***Physics Letters A*2008,**372**(18):3201-3209. 10.1016/j.physleta.2007.07.090MATHMathSciNetView ArticleGoogle Scholar - Mao X:
*Stochastic Differential Equations and Applications*. Horwood, Chichester, UK; 1997.MATHGoogle Scholar - Temam R:
*Infinite-Dimensional Dynamical Systems in Mechanics and Physics, Applied Mathematical Sciences*.*Volume 68*. Springer, New York, NY, USA; 1988:xvi+500.View ArticleGoogle Scholar - Ye Q, Li Z:
*Introduction of Reaction-Diffusion Equation*. Science Press, Beijing, China; 1999.Google Scholar - Niu P, Qu J, Han J:
**Estimation of the eigenvalue of Laplace operator and generalization.***Journal of Baoji College of Arts and Science. Natural Science*2003,**23**(1):85-87.MATHGoogle Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.