Sparse Distributed Memories for On-Line Value-Based Reinforcement Learning

نویسندگان

  • Bohdana Ratitch
  • Doina Precup
چکیده

In this paper, we advocate the use of Sparse Distributed Memories (SDMs) for on-line, value-based reinforcement learning (RL). SDMs provide a linear, local function approximation scheme, designed to work when a very large/ high-dimensional input (address) space has to be mapped into a much smaller physical memory. We present an implementation of the SDM architecture for on-line, value-based RL in continuous state spaces. An important contribution of this paper is an algorithm for dynamic on-line allocation and adjustment of memory resources for SDMs, which eliminates the need for choosing the memory size and structure a priori. In our experiments, this algorithm provides very good performance while efficiently managing the memory resources.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Dynamic Obstacle Avoidance by Distributed Algorithm based on Reinforcement Learning (RESEARCH NOTE)

In this paper we focus on the application of reinforcement learning to obstacle avoidance in dynamic Environments in wireless sensor networks. A distributed algorithm based on reinforcement learning is developed for sensor networks to guide mobile robot through the dynamic obstacles. The sensor network models the danger of the area under coverage as obstacles, and has the property of adoption o...

متن کامل

Sparse Memory Structures Detection

Exploiting lower dimensional structures of the state space is often considered as a potential cure for the lack of dimensional scalability of reinforcement learning algorithms. Indeed, some approximation architectures, notably the Sparse Distributed Memories architecture (SDM[3]), attempt to locate regions of the state space that are “more interesting” and allocate proportionally more memory re...

متن کامل

Sparse Distributed Memories in a Bounded Metric State Space: Some Theoretical and Empirical Results

Sparse Distributed Memories (SDM) [7] is a linear, local function approximation architecture that can be used to represent cost-to-go or stateaction value functions of reinforcement learning (RL) problems. It offers a possibility to reconcile the convergence guarantees of linear approximators and the potential to scale to higher dimensionality typically exclusive to nonlinear architectures. We ...

متن کامل

Sparse Approximations to Value Functions in Reinforcement Learning

We present a novel sparsification and value function approximation method for on-line reinforcement learning in continuous state and action spaces. Our approach is based on the kernel least squares temporal difference learning algorithm. We derive a recursive version and enhance the algorithm with a new sparsification mechanism based on the topology obtained from proximity graphs. The sparsific...

متن کامل

An Application of Importance-based Feature Extraction in Reinforcement Learning

|The sparse feedback in reinforcement learning problems makes feature extraction diicult. We present importance-based feature extraction, which guides a bottom-up self-organization of feature detectors according to top-down information as to the importance of the features; we deene importance in terms of the reinforcement values expected as a result of taking diierent actions when a feature is ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2004