Copyright © 2012 Guobao Zhang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Abstract
Together with Lyapunov-Krasovskii functional theory and reciprocal
convex technique, a new sufficient condition is derived to guarantee the global stability for recurrent neural networks with both time-varying and continuously distributed delays, in which one improved delay-partitioning technique is employed. The LMI-based criterion heavily depends on both the upper and lower bounds on state delay and its derivative, which is different from the existent ones and has more application areas as the lower bound of delay derivative is available. Finally, some numerical examples can illustrate the reduced conservatism of the derived results by thinning the delay interval.