References of "Felten, Florian 50043968"
     in
Bookmark and Share    
Full Text
Peer Reviewed
See detailMO-Gym: A Library of Multi-Objective Reinforcement Learning Environments
Alegre, Lucas Nunes; Felten, Florian UL; Talbi, El-Ghazali UL et al

Scientific Conference (2022, November)

Detailed reference viewed: 192 (92 UL)
Full Text
Peer Reviewed
See detailMORL/D: Multi-Objective Reinforcement Learning based on Decomposition
Felten, Florian UL; Talbi, El-Ghazali UL; Danoy, Grégoire UL

in International Conference in Optimization and Learning (OLA2022) (2022)

Detailed reference viewed: 127 (23 UL)
Full Text
Peer Reviewed
See detailMetaheuristics-based Exploration Strategies for Multi-Objective Reinforcement Learning
Felten, Florian UL; Danoy, Grégoire; Talbi, El-Ghazali UL et al

in Proceedings of the 14th International Conference on Agents and Artificial Intelligence (2022)

The fields of Reinforcement Learning (RL) and Optimization aim at finding an optimal solution to a problem, characterized by an objective function. The exploration-exploitation dilemma (EED) is a well ... [more ▼]

The fields of Reinforcement Learning (RL) and Optimization aim at finding an optimal solution to a problem, characterized by an objective function. The exploration-exploitation dilemma (EED) is a well known subject in those fields. Indeed, a consequent amount of literature has already been proposed on the subject and shown it is a non-negligible topic to consider to achieve good performances. Yet, many problems in real life involve the optimization of multiple objectives. Multi-Policy Multi-Objective Reinforcement Learning (MPMORL) offers a way to learn various optimised behaviours for the agent in such problems. This work introduces a modular framework for the learning phase of such algorithms, allowing to ease the study of the EED in Inner- Loop MPMORL algorithms. We present three new exploration strategies inspired from the metaheuristics domain. To assess the performance of our methods on various environments, we use a classical benchmark - the Deep Sea Treasure (DST) - as well as propose a harder version of it. Our experiments show all of the proposed strategies outperform the current state-of-the-art ε-greedy based methods on the studied benchmarks. [less ▲]

Detailed reference viewed: 243 (66 UL)