S-learning: a reinforcement learning method without parameter tuning

  • Hown Wen Chen*
  • , Von Wun Soo
  • *Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Recently, literature reported progresses on two reinforcement learning algorithms, AHC [1] and Q-learning [2]. We discussed issues raised in these conventional formulations from the aspects of convergence, parameter-tuning, over-training, computational and storage efficiency, and then proposed two new reinforcement learning mechanisms: S-learning and S&AHC learning. Particularly, the representation of the final cost map formed by S-learning series can be explicitly interpreted as the number of minimum movements to the goal state from each individual state. In addition, an adaptive (incremental) S-learning was proposed which incorporated S-learning and the technique of incremental learning [3] to facilitate the practical implementation of neural reinforcement learning. All of S-learning series showed promising performances in exploring Sutton's task [4] of navigating in a maze.

Original languageEnglish
Title of host publicationProceedings of the International Joint Conference on Neural Networks
PublisherPubl by IEEE
Pages557-560
Number of pages4
ISBN (Print)0780314212, 9780780314214
StatePublished - 1993
Externally publishedYes
EventProceedings of 1993 International Joint Conference on Neural Networks. Part 1 (of 3) - Nagoya, Jpn
Duration: 25 10 199329 10 1993

Publication series

NameProceedings of the International Joint Conference on Neural Networks
Volume1

Conference

ConferenceProceedings of 1993 International Joint Conference on Neural Networks. Part 1 (of 3)
CityNagoya, Jpn
Period25/10/9329/10/93

Fingerprint

Dive into the research topics of 'S-learning: a reinforcement learning method without parameter tuning'. Together they form a unique fingerprint.

Cite this