The matter of deep reinforcement learning towards practical ai applications

Tipajin Thaipisutikul, Yi Cheng Chen, Lin Hui, Sheng Chih Chen, Pattanasak Mongkolwat, Timothy K. Shih

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

4 Scopus citations

Abstract

Reinforcement Learning (RL) is an extraordinarily paradigm that aims to solve a complex problem. This technique leverages the traditional feedforward networks with temporal-difference learning to overcome supervised and unsupervised real-world problems. However, RL is one of state-of-the-art topic due to the opaque aspects in design and implementation. Also, in which situation we will get performance gain from RL is still unclear. Therefore, This study firstly examines the impact of Experience Replay in Deep Q-Learning agent with Self-Driving Car application. Secondly, The impact of Eligibility Trace in RNN A3C agents with Breakout AI game application is studied. Our results indicated that these two techniques enhance RL performance by more than 20 percent as compared with traditional RL methods.

Original languageEnglish
Title of host publicationProceedings - 2019 12th International Conference on Ubi-Media Computing, Ubi-Media 2019
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages24-29
Number of pages6
ISBN (Electronic)9781728128207
DOIs
StatePublished - Aug 2019
Event12th International Conference on Ubi-Media Computing, Ubi-Media 2019 - Bali, Indonesia
Duration: 6 Aug 20199 Aug 2019

Publication series

NameProceedings - 2019 12th International Conference on Ubi-Media Computing, Ubi-Media 2019

Conference

Conference12th International Conference on Ubi-Media Computing, Ubi-Media 2019
Country/TerritoryIndonesia
CityBali
Period6/08/199/08/19

Keywords

  • A3C
  • Deep Q-learning
  • Deep reinforcement learning
  • Q-learning

Fingerprint

Dive into the research topics of 'The matter of deep reinforcement learning towards practical ai applications'. Together they form a unique fingerprint.

Cite this