Connectionists: CFP: "Emerging techniques and applications in Multi-objective Reinforcement Learning" at ESANN-2015 in Bruges, Beligum
M.A.Wiering
m.a.wiering at rug.nl
Sun Oct 19 12:28:08 EDT 2014
Apologies for any cross-postings:
Emerging techniques and applications in Multi-Objective Reinforcement Learning (MORL)
Direct links:
https://www.elen.ucl.ac.be/esann/
https://ai.vub.ac.be/ESANN_2015_MORL_special_session
Multi-objective optimization (MOO) and Reinforcement Learning (RL) are two well-established research fields in the area of learning, optimization, and control. RL addresses sequential decision making problems in initially unknown stochastic environments, involving stochastic policies and unknown temporal delays between actions and observable effects. Multi-objective optimization (MOO), which is a sub-area of multi-criteria decision making (MCDM), considers the optimization of more than one objective simultaneously and a decision maker, i.e. an algorithm or a technique, decides either which solutions are important for the user or when to present these solutions to the user for further consideration. Currently, MOO algorithms are seldomly used for stochastic optimization, which makes it an unexplored but promising research area.
State of the art
Examples of algorithms that combine the two techniques MOO and RL are:
Multi-objective reinforcement learning is an extension of RL to multi-criteria stochastic rewards (also called utilities in decision theory). Techniques from multi-objective evolutionary computation have been used for multi-objective RL in order to improve the exploration-exploitation tradeoff. The resulting algorithms are hybrids between MCDM and stochastic optimization. The RL algorithms are enriched with the intuition and efficiency of MOO in handing multi-objective problems.
Preference based reinforcement learning combines reinforcement learning and preference learning that extend RL with qualitative reward vectors, e.g. ranking functions, that can be directly used by the user. Like MORL algorithms, RL is extended with new order relationships to order the policies.
Some multi-objective evolutionary algorithms use also method inspired by reinforcement learning to cope with noisy and uncertain environments.
Aim and scope
The main goal of this special session is to solicit research and potential synergies between multi-objective optimization, evolutionary computation and reinforcement learning. We encourage submissions describing applications of MOO for agents acting in difficult environments that are possibly dynamic, uncertain and partially observable, e.g. in games, multi-agent applications such as scheduling, and other real-world applications.
Topics of interests
Novel frameworks combining both MOO and RL
Multi-objective optimization algorithms such as meta-heuristics and evolutionary algorithms for dynamic and uncertain environments
Theoretical results on learnability in multi-objective dynamic and uncertain environments
On-line self-adapting systems or automatic configuration systems
Solving multi-objective sequential decision making problems with RL
Real-world multi-objective applications in engineering, business, computer science, biological sciences, scientific computation
Organizers
Madalina M. Drugan (mdrugan at vub.ac.be), Bernard Manderick (Bernard.Manderick at vub.ac.be) and
Ann Nowe (anowe at vub.ac.be), Artificial Intelligence Lab, Vrije Universiteit Brussel, Pleinlaan 2, 1050, Brussels, Belgium
Dates
Submission of papers: 21 November 2014
Notification of acceptance: 31 January 2015
ESANN conference:2 2 - 24 April 2015 in Bruges, Belgium
Author guidelines
Papers must not exceed 6 pages, including figures and references.
More information https://www.elen.ucl.ac.be/esann/index.php?pg=guidelines
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/connectionists/attachments/20141019/ae36bc0e/attachment.html>
More information about the Connectionists
mailing list