RL-Glue In Practice
This page is dedicated to list the projects, papers, and classes that have used RL-Glue. Please contact us if you have an example of RL-Glue in practice that we can put on this page!
- 2009 Reinforcement Learning Competition
- 2008 Reinforcement Learning Competition
- The First Annual Reinforcement Learning Competition
- Reinforcement Learning Workshop: Benchmarks and Bakeoffs
- Reinforcement Learning Workshop: Benchmarks and Bakeoffs II
- ICML Reinforcement Learning and Benchmarking Event
- W. Bradley Knox and Peter Stone. Combining Manual Feedback with Subsequent MDP Reward Signals for Reinforcement Learning. In Proc. of 9th Int. Conf. on Autonomous Agents and Multiagent Systems (AAMAS 2010), May 2010. Winner of the Pragnesh Jay Modi Best Student Paper Award.
- Shimon Whiteson, Brian Tanner, and Adam White. The Reinforcement Learning Competitions. AI Magazine, 31(2):81–94, 2010.
- Brian Tanner and Adam White. RL-Glue: Language-Independent Software for Reinforcement-Learning Experiments. Journal of Machine Learning Research, 10(Sep):2133--2136, 2009. (BibTex)
- Lihong Li, Michael L. Littman, and Christopher R. Mansley: Online exploration in least-squares policy iteration. To appear in AAMAS-09, Budapest, Hungary, May, 2009.
- John Asmuth, Lihong Li, Michael L. Littman, Ali Nouri, and David Wingate: A Bayesian Sampling Approach to Exploration in Reinforcement Learning . UAI-09; Montreal, Canada; June 2009.
- W. Bradley Knox and Peter Stone. Interactively Shaping Agents via Human Reinforcement: The TAMER Framework. In Proceedings of The Fifth International Conference on Knowledge Capture, September 2009.
- John Asmuth, Michael L. Littman and Robert Zinkov: Potential-based Shaping in Model-based Reinforcement Learning. AAAI 2008.
- W. Bradley Knox, Ian Fasel and Peter Stone. Design Principles for Creating Human-Shapable Agents. In AAAI Spring 2009 Symposium on Agents that Learn from Human Teachers, March 2009. AAAI Spring 2009 Symposium: Agents that Learn from Human Teachers
- W. Bradley Knox and Peter Stone. TAMER: Training an Agent Manually via Evaluative Reinforcement. In IEEE 7th International Conference on Development and Learning, August 2008. ICDL-2008
- Csáji and Monostori: Value Function Based Reinforcement Learning in Changing Markovian Environments. Journal of Machine Learning Research 9 (2008) 1679-1709
- Steffen Nissen: Large Scale Reinforcement Learning using Q-SARSA(λ) and Cascading Neural Networks. MSc Thesis Department of Computer Science University of Copenhagen Denmark. 2007.
- Thomas J. Walsh, Ali Nouri, Lihong Li, and Michael L. Littman: Planning and learning in environments with delayed feedback. In the Journal of Autonomous Agents and Multi-Agent Systems, 18(1):83-105, 2009. A preliminary version appeared in ECML-07.
- Thomas J. Walsh, Ali Nouri, Lihong Li, and Michael L. Littman: Planning and learning in environments with delayed feedback. In ECML-07, Warsaw, Poland, September, 2007. Also appears in LNCS 4701.
The mission of the Reinforcement-Learning Library (RL-Library) is to create a centralized place for the reinforcement-learning community to share their RL-Glue compatible software projects.
The RL-Library serves two distinct needs. First, to provide standardized, trusted implementations of agents and environments from the reinforcement-learning literature. Second, as a repository for other RL-Glue compatible software and tools, including "up and coming" agents and environments.
The TAMER project seeks to create agents which can be effectively taught behaviors by lay people using positive and negative feedback signals (akin to reward and punishment in animal training). This training scenario is formally defined as the Shaping Problem. The TAMER framework makes use of established supervised learning techniques to model a human's reinforcement function and bases its action selection on the learned model. Videos of TAMER agents being trained can be found here.
rosglue is a framework that allows robots running ROS, Willow Garage's middleware system, to be environments for RL-Glue agents. The goal is to facilitate increased communication between the robotics and RL fields and open further collaborations.
Classes and Course Projects
- Nitish Srivastava and T Sudhamsh Goutham. Exploring Effects of Intrinsic Motivation in Reinforcement Learning Agents. Course Project. 2010.