A Tractable Algorithm for Finite-Horizon Continuous Reinforcement Learning
| dc.contributor.author | Gampa P.; Kondamudi S.S.; Kailasam L. | |
| dc.date.accessioned | 2025-05-24T09:40:22Z | |
| dc.description.abstract | We consider the finite horizon continuous reinforcement learning problem. Our contribution is three-fold. First,we give a tractable algorithm based on optimistic value iteration for the problem. Next,we give a lower bound on regret of order Ω(T2/3) for any algorithm discretizes the state space, improving the previous regret bound of Ω(T1/2) of Ortner and Ryabko [1] for the same problem. Next,under the assumption that the rewards and transitions are Hölder Continuous we show that the upper bound on the discretization error is const.Ln-α T. Finally, we give some simple experiments to validate our propositions. © 2019 IEEE. | |
| dc.identifier.doi | https://doi.org/10.1109/ICoIAS.2019.00018 | |
| dc.identifier.uri | http://172.23.0.11:4000/handle/123456789/19134 | |
| dc.relation.ispartofseries | Proceedings - 2019 2nd International Conference on Intelligent Autonomous Systems, ICoIAS 2019 | |
| dc.title | A Tractable Algorithm for Finite-Horizon Continuous Reinforcement Learning |