Keras Reinforcement Learning Projects
eBook - ePub

Keras Reinforcement Learning Projects

9 projects exploring popular reinforcement learning techniques to build self-learning agents

Giuseppe Ciaburro

  1. 288 pages
  2. English
  3. ePUB (adapté aux mobiles)
  4. Disponible sur iOS et Android
eBook - ePub

Keras Reinforcement Learning Projects

9 projects exploring popular reinforcement learning techniques to build self-learning agents

Giuseppe Ciaburro

DĂ©tails du livre
Aperçu du livre
Table des matiĂšres
Citations

À propos de ce livre

A practical guide to mastering reinforcement learning algorithms using Keras

Key Features

  • Build projects across robotics, gaming, and finance fields, putting reinforcement learning (RL) into action
  • Get to grips with Keras and practice on real-world unstructured datasets
  • Uncover advanced deep learning algorithms such as Monte Carlo, Markov Decision, and Q-learning

Book Description

Reinforcement learning has evolved a lot in the last couple of years and proven to be a successful technique in building smart and intelligent AI networks. Keras Reinforcement Learning Projects installs human-level performance into your applications using algorithms and techniques of reinforcement learning, coupled with Keras, a faster experimental library.

The book begins with getting you up and running with the concepts of reinforcement learning using Keras. You'll learn how to simulate a random walk using Markov chains and select the best portfolio using dynamic programming (DP) and Python. You'll also explore projects such as forecasting stock prices using Monte Carlo methods, delivering vehicle routing application using Temporal Distance (TD) learning algorithms, and balancing a Rotating Mechanical System using Markov decision processes.

Once you've understood the basics, you'll move on to Modeling of a Segway, running a robot control system using deep reinforcement learning, and building a handwritten digit recognition model in Python using an image dataset. Finally, you'll excel in playing the board game Go with the help of Q-Learning and reinforcement learning algorithms.

By the end of this book, you'll not only have developed hands-on training on concepts, algorithms, and techniques of reinforcement learning but also be all set to explore the world of AI.

What you will learn

  • Practice the Markov decision process in prediction and betting evaluations
  • Implement Monte Carlo methods to forecast environment behaviors
  • Explore TD learning algorithms to manage warehouse operations
  • Construct a Deep Q-Network using Python and Keras to control robot movements
  • Apply reinforcement concepts to build a handwritten digit recognition model using an image dataset
  • Address a game theory problem using Q-Learning and OpenAI Gym

Who this book is for

Keras Reinforcement Learning Projects is for you if you are data scientist, machine learning developer, or AI engineer who wants to understand the fundamentals of reinforcement learning by developing practical projects. Sound knowledge of machine learning and basic familiarity with Keras is useful to get the most out of this book

Foire aux questions

Comment puis-je résilier mon abonnement ?
Il vous suffit de vous rendre dans la section compte dans paramĂštres et de cliquer sur « RĂ©silier l’abonnement ». C’est aussi simple que cela ! Une fois que vous aurez rĂ©siliĂ© votre abonnement, il restera actif pour le reste de la pĂ©riode pour laquelle vous avez payĂ©. DĂ©couvrez-en plus ici.
Puis-je / comment puis-je télécharger des livres ?
Pour le moment, tous nos livres en format ePub adaptĂ©s aux mobiles peuvent ĂȘtre tĂ©lĂ©chargĂ©s via l’application. La plupart de nos PDF sont Ă©galement disponibles en tĂ©lĂ©chargement et les autres seront tĂ©lĂ©chargeables trĂšs prochainement. DĂ©couvrez-en plus ici.
Quelle est la différence entre les formules tarifaires ?
Les deux abonnements vous donnent un accĂšs complet Ă  la bibliothĂšque et Ă  toutes les fonctionnalitĂ©s de Perlego. Les seules diffĂ©rences sont les tarifs ainsi que la pĂ©riode d’abonnement : avec l’abonnement annuel, vous Ă©conomiserez environ 30 % par rapport Ă  12 mois d’abonnement mensuel.
Qu’est-ce que Perlego ?
Nous sommes un service d’abonnement Ă  des ouvrages universitaires en ligne, oĂč vous pouvez accĂ©der Ă  toute une bibliothĂšque pour un prix infĂ©rieur Ă  celui d’un seul livre par mois. Avec plus d’un million de livres sur plus de 1 000 sujets, nous avons ce qu’il vous faut ! DĂ©couvrez-en plus ici.
Prenez-vous en charge la synthÚse vocale ?
Recherchez le symbole Écouter sur votre prochain livre pour voir si vous pouvez l’écouter. L’outil Écouter lit le texte Ă  haute voix pour vous, en surlignant le passage qui est en cours de lecture. Vous pouvez le mettre sur pause, l’accĂ©lĂ©rer ou le ralentir. DĂ©couvrez-en plus ici.
Est-ce que Keras Reinforcement Learning Projects est un PDF/ePUB en ligne ?
Oui, vous pouvez accĂ©der Ă  Keras Reinforcement Learning Projects par Giuseppe Ciaburro en format PDF et/ou ePUB ainsi qu’à d’autres livres populaires dans Computer Science et Artificial Intelligence (AI) & Semantics. Nous disposons de plus d’un million d’ouvrages Ă  dĂ©couvrir dans notre catalogue.

Informations

Année
2018
ISBN
9781789347975

Simulating Random Walks

Stochastic processes involve systems that evolve over time (but also more generally in space) according to probabilistic laws. Such systems or models describe the complex phenomena of the real world that have the possibility of being random. These phenomena are more frequent than we can believe. We encounter these phenomena when the quantities we are interested in are not predictable with absolute certainty. However, when such phenomena show a variability of possible outcomes that can be somehow explained or described, then we can introduce a probabilistic model of the phenomenon.
For example, say that we are examining the motion involved in a random walking movement. We study the motion of an object that is constrained to move along a straight line in the two directions allowed. At each movement, it moves randomly to the right or left, each step being of equal length and independent of the other steps. A Markov chain is a stochastic process whereby the evolution of a system depends only on its present state and not on its past state. A Markov chain is characterized by a set of states and by the probability of transition between states. Think of a point that can move randomly forward or backward along a line at discrete intervals of time, covering a certain distance at each interval. This is an example of a random walk. In this chapter, we will simulate a random walk using Markov chains through a Python code implementation.
In this chapter, we will cover the following topics:
  • Random walk
  • Random walk simulation
  • Basic probability concepts
  • Markov chain
  • Forecasting using a Markov chain
  • Markov chain text generator
At the end of the chapter, the reader will know the basic concepts of the Markov process, the basic concepts of random walks, how the random walk algorithms work, know how to use a Markov chain to forecast the weather, and how to simulate random walks using Markov chains.

Random walks

Random walks are a mathematical model that is used to describe a path that is given by a succession of random steps, which, depending on the system that we want to describe, may have a certain number of degrees of freedom or direction. The term random walk was introduced by Karl Pearson in 1905. In a random walk, each step has a random direction and possibly also a random dimension. It represents a theoretical model to describe any random process through the evolution of known quantities that follow a precise statistical distribution. Physically speaking, the path that we are going to draw over time will not necessarily describe a real motion, but rather indicate more generally the evolution of features over time. This means that random walks find applications in physics, chemistry, and biology, but also in other fields, such as computer science, economics, and sociology.

One-dimensional random walk

In a one-dimensional random walk, we study the motion of a point-like particle that is constrained to move along a straight line in one of only two directions (right and left). For each (random) movement, it can move one step to the right with a fixed probability p or to the left with a q probability. Each step is of equal length, and is independent of the others, as shown in the following diagram:
The position of the point after n steps—identified by its abscissa, X(n)—obviously contains a random term. We want to calculate the probability after n movements that the particle will return to the starting point (it should be noted that nothing assures us with any certainty that the point will actually return to that position). To do this, we will use the X(n) variable, which gives the abscissa of the straight line after the particle has moved n steps to the left. Obviously, this is a discrete random variable with a binomial distribution.
This variable takes the following scheme: at every instant, n takes a step to the right or left according to the result of a random variable, Z(n), which takes on +1 value with probability of p> 0 and a -1 value with a probability of q, with p + q = 1, as shown in the previous diagram. Suppose that the random Zn variable with n = 1, 2,... are independent, and all have the same distribution. Then the position of the particle at the n instant is given by the following equation:
The Xn variable represents a Markov chain because, to determine the probability that the particle is in a certain position in the next moment, we only need to know where it is at the current moment, even if we are aware of where it was in all moments prior to the current one.

Simulating 1D random walk

We have previously said that a random walk contains a ...

Table des matiĂšres

  1. Title Page
  2. Copyright and Credits
  3. Packt Upsell
  4. Contributors
  5. Preface
  6. Overview of Keras Reinforcement Learning
  7. Simulating Random Walks
  8. Optimal Portfolio Selection
  9. Forecasting Stock Market Prices
  10. Delivery Vehicle Routing Application
  11. Continuous Balancing of a Rotating Mechanical System
  12. Dynamic Modeling of a Segway as an Inverted Pendulum System
  13. Robot Control System Using Deep Reinforcement Learning
  14. Handwritten Digit Recognizer
  15. Playing the Board Game Go
  16. What's Next?
  17. Other Books You May Enjoy
Normes de citation pour Keras Reinforcement Learning Projects

APA 6 Citation

Ciaburro, G. (2018). Keras Reinforcement Learning Projects (1st ed.). Packt Publishing. Retrieved from https://www.perlego.com/book/825780/keras-reinforcement-learning-projects-9-projects-exploring-popular-reinforcement-learning-techniques-to-build-selflearning-agents-pdf (Original work published 2018)

Chicago Citation

Ciaburro, Giuseppe. (2018) 2018. Keras Reinforcement Learning Projects. 1st ed. Packt Publishing. https://www.perlego.com/book/825780/keras-reinforcement-learning-projects-9-projects-exploring-popular-reinforcement-learning-techniques-to-build-selflearning-agents-pdf.

Harvard Citation

Ciaburro, G. (2018) Keras Reinforcement Learning Projects. 1st edn. Packt Publishing. Available at: https://www.perlego.com/book/825780/keras-reinforcement-learning-projects-9-projects-exploring-popular-reinforcement-learning-techniques-to-build-selflearning-agents-pdf (Accessed: 14 October 2022).

MLA 7 Citation

Ciaburro, Giuseppe. Keras Reinforcement Learning Projects. 1st ed. Packt Publishing, 2018. Web. 14 Oct. 2022.