Many researchers have recently focused their efforts on devising efficient algorithms, mainly based on optimization schemes, for learning the weights of recurrent neural networks. Like for feedforward networks, however, these learning algorithms may get stuck in local minima during gradient descent, thus discovering sub-optimal solutions. In this paper, we give sufficient conditions which guarantee local minima free error surfaces. Moreover, we provide an example which shows the constructive role of the proposed theory in designing networks suitable for solving a given task.
Bianchini, M., Gori, M., Maggini, M. (1994). The Decoupling Network Assumptions for Optimal Learning in Recurrent Neural Networks. In Proceedings of Neuro-Mimetiques ‘94, Neural Networks & Their Applications (pp.255-264).
The Decoupling Network Assumptions for Optimal Learning in Recurrent Neural Networks
BIANCHINI, MONICA;GORI, MARCO;MAGGINI, MARCO
1994-01-01
Abstract
Many researchers have recently focused their efforts on devising efficient algorithms, mainly based on optimization schemes, for learning the weights of recurrent neural networks. Like for feedforward networks, however, these learning algorithms may get stuck in local minima during gradient descent, thus discovering sub-optimal solutions. In this paper, we give sufficient conditions which guarantee local minima free error surfaces. Moreover, we provide an example which shows the constructive role of the proposed theory in designing networks suitable for solving a given task.File | Dimensione | Formato | |
---|---|---|---|
NM94.pdf
non disponibili
Tipologia:
Post-print
Licenza:
NON PUBBLICO - Accesso privato/ristretto
Dimensione
726.39 kB
Formato
Adobe PDF
|
726.39 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11365/30483
Attenzione
Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo