Connectionists: CFP: "Emerging techniques and applications in Multi-objective Reinforcement Learning" at ESANN-2015 in Bruges, Beligum

M.A.Wiering m.a.wiering at rug.nl
Sun Oct 19 12:28:08 EDT 2014






Apologies for any cross-postings:

Emerging techniques and applications in Multi-Objective Reinforcement Learning (MORL)

Direct links:

https://www.elen.ucl.ac.be/esann/

https://ai.vub.ac.be/ESANN_2015_MORL_special_session

Multi-objective optimization (MOO) and Reinforcement Learning (RL) are two well-established research fields in the area of learning, optimization, and control. RL addresses sequential decision making problems in initially unknown stochastic environments, involving stochastic policies and unknown temporal delays between actions and observable effects. Multi-objective optimization (MOO), which is a sub-area of multi-criteria decision making (MCDM), considers the optimization of more than one objective simultaneously and a decision maker, i.e. an algorithm or a technique, decides either which solutions are important for the user or when to present these solutions to the user for further consideration. Currently, MOO algorithms are seldomly used for stochastic optimization, which makes it an unexplored but promising research area.

State of the art

Examples of algorithms that combine the two techniques MOO and RL are:

Multi-objective reinforcement learning is an extension of RL to multi-criteria stochastic rewards (also called utilities in decision theory). Techniques from multi-objective evolutionary computation have been used for multi-objective RL in order to improve the exploration-exploitation tradeoff. The resulting algorithms are hybrids between MCDM and stochastic optimization. The RL algorithms are enriched with the intuition and efficiency of MOO in handing multi-objective problems.

Preference based reinforcement learning combines reinforcement learning and preference learning that extend RL with qualitative reward vectors, e.g. ranking functions, that can be directly used by the user. Like MORL algorithms, RL is extended with new order relationships to order the policies.

Some multi-objective evolutionary algorithms use also method inspired by reinforcement learning to cope with noisy and uncertain environments.




Aim and scope

The main goal of this special session is to solicit research and potential synergies between multi-objective optimization, evolutionary computation and reinforcement learning. We encourage submissions describing applications of MOO for agents acting in difficult environments that are possibly dynamic, uncertain and partially observable, e.g. in games, multi-agent applications such as scheduling, and other real-world applications.




Topics of interests

	Novel frameworks combining both MOO and RL
	Multi-objective optimization algorithms such as meta-heuristics and evolutionary algorithms for dynamic and uncertain environments
	Theoretical results on learnability in multi-objective dynamic and uncertain environments
	On-line self-adapting systems or automatic configuration systems
	Solving multi-objective sequential decision making problems with RL
	Real-world multi-objective applications in engineering, business, computer science, biological sciences, scientific computation


Organizers

Madalina M. Drugan (mdrugan at vub.ac.be), Bernard Manderick (Bernard.Manderick at vub.ac.be) and

Ann Nowe (anowe at vub.ac.be), Artificial Intelligence Lab, Vrije Universiteit Brussel, Pleinlaan 2, 1050, Brussels, Belgium




Dates

Submission of papers: 21 November 2014

Notification of acceptance: 31 January 2015

ESANN conference:2 2 - 24 April 2015 in Bruges, Belgium




Author guidelines

	Papers must not exceed 6 pages, including figures and references.
	More information https://www.elen.ucl.ac.be/esann/index.php?pg=guidelines
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.srv.cs.cmu.edu/pipermail/connectionists/attachments/20141019/ae36bc0e/attachment.html>


More information about the Connectionists mailing list