Show simple item record

dc.contributor.authorMannion, Patrick
dc.contributor.authorDevlin, Sam
dc.contributor.authorKarl, Mannion
dc.contributor.authorDuggan, Jim
dc.date.accessioned2018-12-19T16:32:02Z
dc.date.available2018-12-19T16:32:02Z
dc.date.copyright2017-05
dc.date.issued2017-05
dc.identifier.urihttps://research.thea.ie/handle/20.500.12065/2391
dc.description.abstractReward shaping is a well-established family of techniques that have been successfully used to improve the performance and learning speed of Reinforcement Learning agents in singleobjective problems. Here we extend the guarantees of Potential- Based Reward Shaping (PBRS) by providing theoretical proof that PBRS does not alter the true Pareto front in MORL domains. We also contribute the rst empirical studies of the e ect of PBRS in MORL problems.en_US
dc.formatPdfen_US
dc.language.isoenen_US
dc.rightsAttribution-NonCommercial-NoDerivs 3.0 Ireland*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/3.0/ie/*
dc.subjectMulti-Objectiveen_US
dc.subjectReinforcement Learningen_US
dc.subjectReward Shapingen_US
dc.titlePotential-Based Reward Shaping Preserves Pareto Optimal Policiesen_US
dc.description.peerreviewyesen_US
dc.rights.accessCopyrighten_US
dc.subject.departmentDepartment of Computer Science & Applied Physicsen_US


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record

Attribution-NonCommercial-NoDerivs 3.0 Ireland
Except where otherwise noted, this item's license is described as Attribution-NonCommercial-NoDerivs 3.0 Ireland