dc.creatorCelemin, Carlos
dc.creatorRuiz del Solar, Javier
dc.creatorKober, Jens
dc.date.accessioned2019-10-14T15:41:02Z
dc.date.available2019-10-14T15:41:02Z
dc.date.created2019-10-14T15:41:02Z
dc.date.issued2019
dc.identifierAutonomous Robots (2019) 43:1173–1186
dc.identifier15737527
dc.identifier09295593
dc.identifier10.1007/s10514-018-9786-6
dc.identifierhttps://repositorio.uchile.cl/handle/2250/171513
dc.description.abstractReinforcement Learning agents can be supported by feedback from human teachers in the learning loop that guides the learning process. In this work we propose two hybrid strategies of Policy Search Reinforcement Learning and Interactive Machine Learning that benefit from both sources of information, the cost function and the human corrective feedback, for accelerating the convergence and improving the final performance of the learning process. Experiments with simulated and real systems of balancing tasks and a 3 DoF robot arm validate the advantages of the proposed learning strategies: (i) they speed up the convergence of the learning process between 3 and 30 times, saving considerable time during the agent adaptation, and (ii) they allow including non-expert feedback because they have low sensibility to erroneous human advice.
dc.languageen
dc.publisherSpringer
dc.rightshttp://creativecommons.org/licenses/by-nc-nd/3.0/cl/
dc.rightsAttribution-NonCommercial-NoDerivs 3.0 Chile
dc.sourceAutonomous Robots
dc.subjectInteractive machine learning
dc.subjectLearning from demonstration
dc.subjectPolicy search
dc.subjectReinforcement learning
dc.titleA fast hybrid reinforcement learning framework with human corrective feedback
dc.typeArtículo de revista


Este ítem pertenece a la siguiente institución