Return to search

Reinforcement learning in neural networks with multiple outputs

Reinforcement learning algorithms comprise a class of learning algorithms for neural networks. Reinforcement learning is distinguished from other classes by the type of problems that it is intended to solve. It is used for learning input-output mappings where the desired outputs are not known and only a scalar reinforcement value is available. Primary Reinforcement Learning (PRL) is a core component of the most actively researched form of reinforcement learning. The issues surrounding the convergence characteristics of PRL are considered in this thesis. There have been no convergence proofs for any kind of networks learning under PRL.
A convergence theorem is proved in this thesis, showing that under some conditions, a particular reinforcement learning algorithm, the A[formula omitted] algorithm, will train a single-layer network correctly. The theorem is demonstrated with a series of simulations.
A new PRL algorithm is proposed to deal with the training of multiple layer, binary output networks with continuous inputs. This is a more difficult learning problem than with binary inputs. The new algorithm is shown to be able to successfully train a network with multiple outputs when the environment conforms to the conditions of the convergence theorem for a single-layer network. / Applied Science, Faculty of / Electrical and Computer Engineering, Department of / Graduate

Identiferoai:union.ndltd.org:UBC/oai:circle.library.ubc.ca:2429/29624
Date January 1990
CreatorsIp, John Chong Ching
PublisherUniversity of British Columbia
Source SetsUniversity of British Columbia
LanguageEnglish
Detected LanguageEnglish
TypeText, Thesis/Dissertation
RightsFor non-commercial purposes only, such as research, private study and education. Additional conditions apply, see Terms of Use https://open.library.ubc.ca/terms_of_use.

Page generated in 0.0023 seconds