• Classified by Topic • Classified by Publication Type • Sorted by Date • Sorted by First Author Last Name • Classified by Funding Source •
Characterizing Reinforcement Learning Methods through Parameterized Learning Problems.
Shivaram
Kalyanakrishnan and Peter Stone.
Machine Learning (MLJ), 84(1--2):205–247,
July 2011.
Publisher's
on-line version
[PDF]858.5kB [postscript]2.4MB
The field of reinforcement learning (RL) has been energized in the past few decades by elegant theoretical results indicating under what conditions, and how quickly, certain algorithms are guaranteed to converge to optimal policies. However, in practical problems, these conditions are seldom met. When we cannot achieve optimality, the performance of RL algorithms on these tasks must be measured empirically. Consequently, in order to differentiate among the algorithms, it becomes necessary to characterize the performance of different learning methods on different problems, taking into account factors such as state estimation, exploration, function approximation, and constraints on computation and memory. To this end, this article introduces parameterized learning problems, in which such factors can be controlled systematically and their effects on learning methods characterized through targeted studies. Based on a survey of existing RL applications, we focus our attention on two predominant, ``first order'' factors: partial observability and function approximation. We design an appropriate parameterized learning problem, through which we compare two qualitatively distinct classes of algorithms: on-line value function-based methods and policy search methods. Empirical comparisons among various methods within each class project Sarsa($\lambda$) and CMA-ES as the respective winners. Comparing these methods further on relevant problem instances, our study highlights regions of the problem space favoring their contrasting approaches. We obtain additional insights on relationships between method-specific parameters --- such as eligibility traces and initial weights --- and problem instances.
@article(MLJ11-shivaram,
author="Shivaram Kalyanakrishnan and Peter Stone",
title="Characterizing Reinforcement Learning Methods through Parameterized Learning Problems",
journal="Machine Learning (MLJ)",
year="2011",
volume = "84",
number = "1--2",
pages = "205--247",
month = "July",
abstract="
The field of reinforcement learning (RL) has been
energized in the past few decades by elegant
theoretical results indicating under what
conditions, and how quickly, certain algorithms are
guaranteed to converge to \emph{optimal} policies.
However, in practical problems, these conditions are
seldom met. When we cannot achieve optimality, the
performance of RL algorithms on these tasks must be
measured empirically. Consequently, in order to
differentiate among the algorithms, it becomes
necessary to characterize the performance of
different learning \textit{methods} on different
\textit{problems}, taking into account factors such
as state estimation, exploration, function
approximation, and constraints on computation and
memory. To this end, this article introduces
\textit{parameterized learning problems}, in which
such factors can be controlled systematically and
their effects on learning methods characterized
through targeted studies.
Based on a survey of existing RL applications, we
focus our attention on two predominant, ``first
order'' factors: partial observability and function
approximation. We design an appropriate
parameterized learning problem, through which we
compare two qualitatively distinct classes of
algorithms: on-line value function-based methods and
policy search methods. Empirical comparisons among
various methods within each class project
Sarsa($\lambda$) and CMA-ES as the respective
winners. Comparing these methods further on relevant
problem instances, our study highlights regions of
the problem space favoring their contrasting
approaches. We obtain additional insights on
relationships between method-specific parameters ---
such as eligibility traces and initial weights ---
and problem instances.",
wwwnote={<a href="http://www.springerlink.com/openurl.asp?genre=article&id=doi:10.1007/s10994-011-5251-x">Publisher's on-line version</a>},
)
Generated by bib2html.pl (written by Patrick Riley ) on Sat Nov 15, 2025 21:30:07