Variable Hidden Layer Sizing in Elman Recurrent Neuro-Evolution

dc.contributor.authorKaikhah, Khosrow
dc.contributor.authorGarlick, Ryan
dc.date.accessioned2012-02-24T10:17:49Z
dc.date.available2012-02-24T10:17:49Z
dc.date.issued2000-05
dc.description.abstractThe relationship between the size of the hidden layer in a neural network and performance in a particular domain is currently an open research issue. Often, the number of neurons in the hidden layer is chosen empirically and subsequently fixed for the training of the network. Fixing the size of the hidden layer limits an inherent strength of neural networks the ability to generalize experiences from one situation to another, to adapt to new situations, and to overcome the brittleness often associated with traditional artificial intelligence techniques. This paper proposes an evolutionary algorithm to search for network sizes along with weights and connections between neurons. The size of the networks simply becomes another search parameter for the evolutionary algorithm. This research builds upon the neuro-evolution tool SANE, developed by David Moriarty. SANE evolves neurons and networks simultaneously, and is modified in this work in several ways, including varying the hidden layer size, and evolving Elman recurrent neural networks for non-Markovian tasks. These modifications allow the evolution of better performing and more consistent networks, and do so more efficiently and faster. SANE, modified with variable network sizing, learns to play modified casino blackjack and develops a successful card counting strategy. The contributions of this research are up to 8.34% performance increases over fixed hidden layer size models while reducing hidden layer processing time by almost 10%, and a faster, more autonomous approach to the scaling of neuro-evolutionary techniques to solving larger and more difficult problems.
dc.description.departmentComputer Science
dc.formatText
dc.format.extent15 pages
dc.format.medium1 file (.pdf)
dc.identifier.citationKaikhah, K., & Garlick, R. (2000). Variable hidden layer sizing in Elman recurrent neuro-evolution. Applied Intelligence, 12(3), pp. 193-205.
dc.identifier.doihttps://doi.org/10.1023/A:1008315023738
dc.identifier.urihttps://hdl.handle.net/10877/3812
dc.language.isoen
dc.publisherKluwer Academic Publishers
dc.sourceApplied Intelligence, 2000, Vol. 12, No. 3, pp. 193-205.
dc.subjectneuro-evaluatory
dc.subjectElman recurrent neural networks
dc.subjectComputer Science
dc.titleVariable Hidden Layer Sizing in Elman Recurrent Neuro-Evolution
dc.typeArticle

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
fulltext.pdf
Size:
100.08 KB
Format:
Adobe Portable Document Format