dc.creatorFerreira L.A.
dc.creatorC. Bianchi R.A.
dc.creatorSantos P.E.
dc.creatorde Mantaras R.L.
dc.date.accessioned2019-08-19T23:45:19Z
dc.date.accessioned2022-09-09T15:47:34Z
dc.date.accessioned2023-03-13T18:46:43Z
dc.date.available2019-08-19T23:45:19Z
dc.date.available2022-09-09T15:47:34Z
dc.date.available2023-03-13T18:46:43Z
dc.date.created2019-08-19T23:45:19Z
dc.date.created2022-09-09T15:47:34Z
dc.date.issued2017
dc.identifierAnjoletto, L.; Bianchi; Santos, Paulo; De MANTARAS, R. L.. Answer Set Programming for Non-Stationary Markov Decision Processes. APPLIED INTELLIGENCE, v. 1, p. 1, 2017.
dc.identifier1573-7497
dc.identifierhttp://148.201.128.228:8080/xmlui/handle/20.500.12032/8030
dc.identifier.urihttps://repositorioslatinoamericanos.uchile.cl/handle/2250/6138233
dc.description.abstract© 2017, Springer Science+Business Media New York.Non-stationary domains, where unforeseen changes happen, present a challenge for agents to find an optimal policy for a sequential decision making problem. This work investigates a solution to this problem that combines Markov Decision Processes (MDP) and Reinforcement Learning (RL) with Answer Set Programming (ASP) in a method we call ASP(RL). In this method, Answer Set Programming is used to find the possible trajectories of an MDP, from where Reinforcement Learning is applied to learn the optimal policy of the problem. Results show that ASP(RL) is capable of efficiently finding the optimal solution of an MDP representing non-stationary domains.
dc.relationApplied Intelligence
dc.rightsAcesso Restrito
dc.titleAnswer set programming for non-stationary Markov decision processes
dc.typeArtigo


Este ítem pertenece a la siguiente institución