ePrivacy and GPDR Cookie Consent by Cookie Consent

What to read after Rollout, Policy Iteration, and Distributed Reinforcement Learning?

Hello there! I go by the name Robo Ratel, your very own AI librarian, and I'm excited to assist you in discovering your next fantastic read after "Rollout, Policy Iteration, and Distributed Reinforcement Learning" by Dimitri Bertsekas! 😉 Simply click on the button below, and witness what I have discovered for you.

Exciting news! I've found some fantastic books for you! 📚✨ Check below to see your tailored recommendations. Happy reading! 📖😊

Rollout, Policy Iteration, and Distributed Reinforcement Learning

Dimitri Bertsekas

Computers / Artificial Intelligence / General

The purpose of this book is to develop in greater depth some of the methods from the author's Reinforcement Learning and Optimal Control recently published textbook (Athena Scientific, 2019). In particular, we present new research, relating to systems involving multiple agents, partitioned architectures, and distributed asynchronous computation. We pay special attention to the contexts of dynamic programming/policy iteration and control theory/model predictive control. We also discuss in some detail the application of the methodology to challenging discrete/combinatorial optimization problems, such as routing, scheduling, assignment, and mixed integer programming, including the use of neural network approximations within these contexts.

The book focuses on the fundamental idea of policy iteration, i.e., start from some policy, and successively generate one or more improved policies. If just one improved policy is generated, this is called rollout, which, based on broad and consistent computational experience, appears to be one of the most versatile and reliable of all reinforcement learning methods. In this book, rollout algorithms are developed for both discrete deterministic and stochastic DP problems, and the development of distributed implementations in both multiagent and multiprocessor settings, aiming to take advantage of parallelism.

Approximate policy iteration is more ambitious than rollout, but it is a strictly off-line method, and it is generally far more computationally intensive. This motivates the use of parallel and distributed computation. One of the purposes of the monograph is to discuss distributed (possibly asynchronous) methods that relate to rollout and policy iteration, both in the context of an exact and an approximate implementation involving neural networks or other approximation architectures.

Much of the new research is inspired by the remarkable AlphaZero chess program, where policy iteration, value and policy networks, approximate lookahead minimization, and parallel computation all play an important role.

Do you want to read this book? 😳
Buy it now!

Are you curious to discover the likelihood of your enjoyment of "Rollout, Policy Iteration, and Distributed Reinforcement Learning" by Dimitri Bertsekas? Allow me to assist you! However, to better understand your reading preferences, it would greatly help if you could rate at least two books.