Cover image for Recurrent neural networks for prediction : learning algorithms, architectures and stability
Title:
Recurrent neural networks for prediction : learning algorithms, architectures and stability
Personal Author:
Series:
Wiley series in adaptive and learning systems for signal processing, communications, and control
Publication Information:
Chichester : John Wiley & Sons, 2001
ISBN:
9780471495178
Added Author:

Available:*

Library
Item Barcode
Call Number
Material Type
Item Category 1
Status
Searching...
30000010067754 Q325.5 M36 2001 Open Access Book Book
Searching...

On Order

Summary

Summary

New technologies in engineering, physics and biomedicine are demanding increasingly complex methods of digital signal processing. By presenting the latest research work the authors demonstrate how real-time recurrent neural networks (RNNs) can be implemented to expand the range of traditional signal processing techniques and to help combat the problem of prediction. Within this text neural networks are considered as massively interconnected nonlinear adaptive filters. Analyses the relationships between RNNs and various nonlinear models and filters, and introduces spatio-temporal architectures together with the concepts of modularity and nesting Examines stability and relaxation within RNNsPresents on-line learning algorithms for nonlinear adaptive filters and introduces new paradigms which exploit the concepts of a priori and a posteriori errors, data-reusing adaptation, and normalisation Studies convergence and stability of on-line learning algorithms based upon optimisation techniques such as contraction mapping and fixed point iteration Describes strategies for the exploitation of inherent relationships between parameters in RNNs Discusses practical issues such as predictability and nonlinearity detecting and includes several practical applications in areas such as air pollutant modelling and prediction, attractor discovery and chaos, ECG signal processing, and speech processing

Recurrent Neural Networks for Prediction offers a new insight into the learning algorithms, architectures and stability of recurrent neural networks and, consequently, will have instant appeal. It provides an extensive background for researchers, academics and postgraduates enabling them to apply such networks in new applications.

VISIT OUR COMMUNICATIONS TECHNOLOGY WEBSITE!
http://www.wiley.co.uk/commstech/

VISIT OUR WEB PAGE!
http://www.wiley.co.uk/


Author Notes

Danilo Mandic from the Imperial College London, London, UK was named Fellow of the Institute of Electrical and Electronics Engineers in 2013 for contributions to multivariate and nonlinear learning systems.

Jonathon A. Chambers is the author of Recurrent Neural Networks for Prediction: Learning Algorithms, Architectures and Stability , published by Wiley.


Table of Contents

Preface
Introduction
Fundamentals
Network Architectures for Prediction
Activation Functions Used in Neural Networks
Recurrent Neural Networks Architectures
Neural Networks as Nonlinear Adaptive Filters
Stability Issues in RNN Architectures
Data-Reusing Adaptive Learning Algorithms
A Class of Normalised Algorithms for Online Training of Recurrent Neural Networks
Convergence of Online Learning Algorithms in Neural Networks
Some Practical Considerations of Predictability and Learning Algorithms for Various Signals
Exploiting Inherent Relationships Between Parameters in Recurrent Neural Networks
Appendix A The O Notation and Vector and Matrix Differentiation
Appendix B Concepts from the Approximation Theory
Appendix C Complex Sigmoid Activation Functions, Holomorphic Mappings and Modular Groups
Appendix D Learning Algorithms for RNNs
Appendix E Terminology Used in the Field of Neural Networks
Appendix F On the A Posteriori Approach in Science and Engineering
Appendix G Contraction Mapping Theorems
Appendix H Linear GAS Relaxation
Appendix I The Main Notions in Stability Theory
Appendix J Deasonsonalising Time Series
References
Index