Repository logo
Institutional Digital Repository
Shreenivas Deshpande Library, IIT (BHU), Varanasi

A Tractable Algorithm for Finite-Horizon Continuous Reinforcement Learning

dc.contributor.authorGampa, P.
dc.contributor.authorKondamudi, S.S.
dc.contributor.authorKailasam, L.
dc.date.accessioned2021-01-05T05:22:41Z
dc.date.available2021-01-05T05:22:41Z
dc.date.issued2019-02
dc.description.abstractWe consider the finite horizon continuous reinforcement learning problem. Our contribution is three-fold. First,we give a tractable algorithm based on optimistic value iteration for the problem. Next,we give a lower bound on regret of order Ω(T2/3) for any algorithm discretizes the state space, improving the previous regret bound of Ω(T1/2) of Ortner and Ryabko [1] for the same problem. Next,under the assumption that the rewards and transitions are Hölder Continuous we show that the upper bound on the discretization error is const.Ln-α T. Finally, we give some simple experiments to validate our propositions. © 2019 IEEE.en_US
dc.identifier.issn978-172812662-3
dc.identifier.urihttps://idr-sdlib.iitbhu.ac.in/handle/123456789/1234
dc.language.isoen_USen_US
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_US
dc.relation.ispartofseriesProceedings - 2019 2nd International Conference on Intelligent Autonomous Systems, ICoIAS 2019;
dc.subjectReinforcement Learningen_US
dc.subjectMarkov Decision Process(MDP)en_US
dc.subjectRegreten_US
dc.subjectContinuous State Spaceen_US
dc.subjectBonusen_US
dc.subjectFinite Horizonen_US
dc.titleA Tractable Algorithm for Finite-Horizon Continuous Reinforcement Learningen_US
dc.typeArticleen_US

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
08782442.pdf
Size:
1.35 MB
Format:
Adobe Portable Document Format
Description:
Open Access

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: