Finite time convergent learning law for continuous neural networks

Research output: Contribution to journalArticlepeer-review

4 Scopus citations

Abstract

This paper addresses the design of a discontinuous finite time convergent learning law for neural networks with continuous dynamics. The neural network was used here to obtain a non-parametric model for uncertain systems described by a set of ordinary differential equations. The source of uncertainties was the presence of some external perturbations and poor knowledge of the nonlinear function describing the system dynamics. A new adaptive algorithm based on discontinuous algorithms was used to adjust the weights of the neural network. The adaptive algorithm was derived by means of a non-standard Lyapunov function that is lower semi-continuous and differentiable in almost the whole space. A compensator term was included in the identifier to reject some specific perturbations using a nonlinear robust algorithm. Two numerical examples demonstrated the improvements achieved by the learning algorithm introduced in this paper compared to classical schemes with continuous learning methods. The first one dealt with a benchmark problem used in the paper to explain how the discontinuous learning law works. The second one used the methane production model to show the benefits in engineering applications of the learning law proposed in this paper.

Original languageEnglish
Pages (from-to)175-182
Number of pages8
JournalNeural Networks
Volume50
DOIs
StatePublished - Feb 2014

Keywords

  • Discontinuous systems
  • Learning algorithms
  • Lyapunov method
  • Neural networks
  • Super-twisting

Fingerprint

Dive into the research topics of 'Finite time convergent learning law for continuous neural networks'. Together they form a unique fingerprint.

Cite this