Final Call For Papers: Machine Learning Journal Special Issue Empirical Evaluations in Reinforcement Learning Submission Deadline: February 26, 2010 Guest Editors: Shimon Whiteson and Michael Littman
The continuing development of a field requires a healthy exchange between theoretical advances and experimental observations. The purpose of this special issue is to assess progress in empirical evaluations of reinforcement-learning algorithms and to encourage the adoption of effective experimental methodologies. The last several years have seen new trends in uniform software interfaces between environments and learning algorithms, community comparisons and competitions, and an increased interest in experimenting with reinforcement learning in embedded systems. We enthusiastically solicit papers on relevant topics such as:
* The design and dissemination of standardized frameworks and repositories for algorithms, methods, and/or results. * Experience of organizers and participants in reinforcement-learning competitions and bake-offs.
* Novel evaluation methodologies or metrics. * Careful empirical comparisons of existing methods.* Novel methods validated with strong empirical results on existing benchmarks, especially those used in recent RL Competitions (see http://www.rl-competition.org/) . * Applications of reinforcement-learning approaches to real-life environments such as computer networks, system management and robotics. * Theoretical work such as sample complexity bounds that can be used to guide the design of benchmarks and evaluations.
The emphasis of the special issue is not on the development of novel algorithms. Instead, papers will be assessed in terms of the insights they provide about how best to assess performance in reinforcement learning, i.e., the "meta" problem of evaluating the evaluation methodologies themselves. In particular, papers presenting empirical results should also discuss what those results reveal about the strengths and weaknesses of the evaluation methodology. Similarly, papers describing real-life applications should make clear what limitations the application exposes in 'off-the-shelf' methods, how the employed method had to be modified to address real-world complications, and what the results show that could not be learned from experiments in 'toy' domains. Papers proposing new evaluation methodologies should include illustrative empirical results offering insights that would be difficult to obtain with conventional methodologies. Finally, papers proposing new evaluation methodologies should also compare and contrast with methodologies in related areas, e.g. supervised learning, explaining why such methodologies are not adequate and what ideas, if any, can be borrowed from them.
For more details see: http://www.springer.com/cda/content/document/cda_downloaddocument/CFP_10994_2009826.pdf?SGWID=0-0-45-791198-p35726603
_______________________________________________ uai mailing list uai@ENGR.ORST.EDU https://secure.engr.oregonstate.edu/mailman/listinfo/uai