“Deep reinforcement training. AlphaGo and other technologies ": the announcement of the book

Hello!



We have one of the best books on reinforcement training available for preorder, originally called " Deep Reinforcement Learning Hands-on " by Maxim Lapan. Here is the cover of the Russian translation :







So that you can appreciate the summary of the book, we offer you a translation of the review written by the author to the release of the original.





Hello!



I am a self-taught enthusiast, keen on deep learning. Therefore, when representatives of the Packt publishing house contacted me and suggested writing a practical book about the current state of deep learning with reinforcement, I was a little scared, but after some hesitation I agreed, optimistically assuming: “Oh, there will be an interesting experience.”

I will not say that this work was given to me as an easy walk, of course not. You have no days off, no free time, constant fear of "freezing stupidity" and the pursuit of deadlines for each chapter (two weeks per chapter and example code). However, in general, everything went positively and very interestingly.



Before briefly describing the contents of each chapter, let us describe the idea of ​​the whole book .

When I started experimenting in RL more than four years ago, I had at my disposal the following sources of information:







Maybe there was something else, but these were the most important sources of information. All of them are very far from practice:







At the same time, I was deeply hooked on the article DeepMind (“A neural network can learn to play Atari games in pixels! WOW!”), And I felt that this dry theory hides enormous practical value. So, I spent a lot of time studying the theory, implementing various methods and debugging them. As you probably guessed, it was not easy: you can spend a couple of weeks honing the method and then discover that your implementation is incorrect (or, even worse, you misunderstood the formula). I do not consider such training a waste of time - on the contrary, I think that this is the most correct way to learn something. However, this takes a lot of time.



Two years later, when I started working on the text, my main goal was: to give thorough practical information on RL methods to a reader who is only acquainted with this fascinating discipline - as I once did.



Now a little about the book. It is focused primarily on practice, and I tried to minimize the volume of theory and formulas. It contains key formulas, but no evidence is given. Basically, I try to give an intuitive understanding of what is happening, not seeking the maximum rigor of presentation.



At the same time, it is assumed that the reader has basic knowledge of deep learning and statistics. There is a chapter in the book with an overview of the PyTorch library (since all examples are given using PyTorch), but this chapter cannot be considered a self-sufficient source of information on neural networks. If you have never heard of the loss and activation functions before, start by studying other books, today there are many. (Note: for example, the book " Deep Learning ").



In my book you will find a lot of examples of varying complexity, starting with the simplest ones (the CrossEntropy



method in the CartPole



environment contains ~ 100 lines in python), ending with rather big projects, for example, learning AlphGo Zero or an RL agent for trading on the exchange. Sample code is fully uploaded to GitHub , there are more than 14k lines of Python code in total.



The book consists of 18 chapters covering the most important aspects of modern deep learning with reinforcement:









That's all! I hope you enjoy the book.



All Articles