Repository logo
Institutional Digital Repository
Shreenivas Deshpande Library, IIT (BHU), Varanasi

A Tractable Algorithm for Finite-Horizon Continuous Reinforcement Learning

dc.contributor.authorGampa P.; Kondamudi S.S.; Kailasam L.
dc.date.accessioned2025-05-24T09:40:22Z
dc.description.abstractWe consider the finite horizon continuous reinforcement learning problem. Our contribution is three-fold. First,we give a tractable algorithm based on optimistic value iteration for the problem. Next,we give a lower bound on regret of order Ω(T2/3) for any algorithm discretizes the state space, improving the previous regret bound of Ω(T1/2) of Ortner and Ryabko [1] for the same problem. Next,under the assumption that the rewards and transitions are Hölder Continuous we show that the upper bound on the discretization error is const.Ln-α T. Finally, we give some simple experiments to validate our propositions. © 2019 IEEE.
dc.identifier.doihttps://doi.org/10.1109/ICoIAS.2019.00018
dc.identifier.urihttp://172.23.0.11:4000/handle/123456789/19134
dc.relation.ispartofseriesProceedings - 2019 2nd International Conference on Intelligent Autonomous Systems, ICoIAS 2019
dc.titleA Tractable Algorithm for Finite-Horizon Continuous Reinforcement Learning

Files

Collections