Deep Learning with R for Beginners
eBook - ePub

Deep Learning with R for Beginners

Design neural network models in R 3.5 using TensorFlow, Keras, and MXNet

Mark Hodnett, Joshua F. Wiley, Yuxi (Hayden) Liu, Pablo Maldonado

  1. 612 pages
  2. English
  3. ePUB (adapté aux mobiles)
  4. Disponible sur iOS et Android
eBook - ePub

Deep Learning with R for Beginners

Design neural network models in R 3.5 using TensorFlow, Keras, and MXNet

Mark Hodnett, Joshua F. Wiley, Yuxi (Hayden) Liu, Pablo Maldonado

DĂ©tails du livre
Aperçu du livre
Table des matiĂšres
Citations

À propos de ce livre

Explore the world of neural networks by building powerful deep learning models using the R ecosystem

Key Features

  • Get to grips with the fundamentals of deep learning and neural networks
  • Use R 3.5 and its libraries and APIs to build deep learning models for computer vision and text processing
  • Implement effective deep learning systems in R with the help of end-to-end projects

Book Description

Deep learning finds practical applications in several domains, while R is the preferred language for designing and deploying deep learning models.

This Learning Path introduces you to the basics of deep learning and even teaches you to build a neural network model from scratch. As you make your way through the chapters, you'll explore deep learning libraries and understand how to create deep learning models for a variety of challenges, right from anomaly detection to recommendation systems. The book will then help you cover advanced topics, such as generative adversarial networks (GANs), transfer learning, and large-scale deep learning in the cloud, in addition to model optimization, overfitting, and data augmentation. Through real-world projects, you'll also get up to speed with training convolutional neural networks (CNNs), recurrent neural networks (RNNs), and long short-term memory networks (LSTMs) in R.

By the end of this Learning Path, you'll be well versed with deep learning and have the skills you need to implement a number of deep learning concepts in your research work or projects.

This Learning Path includes content from the following Packt products:

  • R Deep Learning Essentials - Second Edition by Joshua F. Wiley and Mark Hodnett
  • R Deep Learning Projects by Yuxi (Hayden) Liu and Pablo Maldonado

What you will learn

  • Implement credit card fraud detection with autoencoders
  • Train neural networks to perform handwritten digit recognition using MXNet
  • Reconstruct images using variational autoencoders
  • Explore the applications of autoencoder neural networks in clustering and dimensionality reduction
  • Create natural language processing (NLP) models using Keras and TensorFlow in R
  • Prevent models from overfitting the data to improve generalizability
  • Build shallow neural network prediction models

Who this book is for

This Learning Path is for aspiring data scientists, data analysts, machine learning developers, and deep learning enthusiasts who are well versed in machine learning concepts and are looking to explore the deep learning paradigm using R. A fundamental understanding of R programming and familiarity with the basic concepts of deep learning are necessary to get the most out of this Learning Path.

Foire aux questions

Comment puis-je résilier mon abonnement ?
Il vous suffit de vous rendre dans la section compte dans paramĂštres et de cliquer sur « RĂ©silier l’abonnement ». C’est aussi simple que cela ! Une fois que vous aurez rĂ©siliĂ© votre abonnement, il restera actif pour le reste de la pĂ©riode pour laquelle vous avez payĂ©. DĂ©couvrez-en plus ici.
Puis-je / comment puis-je télécharger des livres ?
Pour le moment, tous nos livres en format ePub adaptĂ©s aux mobiles peuvent ĂȘtre tĂ©lĂ©chargĂ©s via l’application. La plupart de nos PDF sont Ă©galement disponibles en tĂ©lĂ©chargement et les autres seront tĂ©lĂ©chargeables trĂšs prochainement. DĂ©couvrez-en plus ici.
Quelle est la différence entre les formules tarifaires ?
Les deux abonnements vous donnent un accĂšs complet Ă  la bibliothĂšque et Ă  toutes les fonctionnalitĂ©s de Perlego. Les seules diffĂ©rences sont les tarifs ainsi que la pĂ©riode d’abonnement : avec l’abonnement annuel, vous Ă©conomiserez environ 30 % par rapport Ă  12 mois d’abonnement mensuel.
Qu’est-ce que Perlego ?
Nous sommes un service d’abonnement Ă  des ouvrages universitaires en ligne, oĂč vous pouvez accĂ©der Ă  toute une bibliothĂšque pour un prix infĂ©rieur Ă  celui d’un seul livre par mois. Avec plus d’un million de livres sur plus de 1 000 sujets, nous avons ce qu’il vous faut ! DĂ©couvrez-en plus ici.
Prenez-vous en charge la synthÚse vocale ?
Recherchez le symbole Écouter sur votre prochain livre pour voir si vous pouvez l’écouter. L’outil Écouter lit le texte Ă  haute voix pour vous, en surlignant le passage qui est en cours de lecture. Vous pouvez le mettre sur pause, l’accĂ©lĂ©rer ou le ralentir. DĂ©couvrez-en plus ici.
Est-ce que Deep Learning with R for Beginners est un PDF/ePUB en ligne ?
Oui, vous pouvez accĂ©der Ă  Deep Learning with R for Beginners par Mark Hodnett, Joshua F. Wiley, Yuxi (Hayden) Liu, Pablo Maldonado en format PDF et/ou ePUB ainsi qu’à d’autres livres populaires dans Computer Science et Neural Networks. Nous disposons de plus d’un million d’ouvrages Ă  dĂ©couvrir dans notre catalogue.

Informations

Année
2019
ISBN
9781838647223
Édition
1
Sous-sujet
Neural Networks

Text Generation using Recurrent Neural Networks

In this chapter, we will describe some of the most exciting techniques in modern (at the time of writing—late 2017) machine learning, recurrent neural networks. They are, however, not new; they have been around since the 1980s, but they have become popular due to the numerous records in language-related tasks in recent years.
Why do we need a different type of architecture for text? Consider the following example:
"I live in Prague since 2015"
and
"Since 2015 I live in Prague"
If we would like to teach a traditional feed-forward network such as a perceptron or a multi-layer perceptron to identify the date I moved to Prague, then this network would have to learn separate parameters for each input feature, which in particular implies that it would have to learn grammar to answer this simple question! This is undesirable in many applications. Similar issues motivated machine learning researchers and statisticians in the 1980s to introduce the idea of sharing parameters across different parts of the model. This idea is the secret sauce of recurrent neural networks, our next deep learning architecture.
By design, recurrent neural networks are well-suited for processing sequential data. In general, machine learning applied to sequential data can be roughly divided into four main areas:
  • Sequence prediction: Given
    , predict the next element of the sequence,
  • Sequence classification: Given
    , predict a category or label for it
  • Sequence generation: Given
    , generate a new element of the sequence,
  • Sequence to sequence prediction: Given
    , generate an equivalent sequence,
Applications of sequence prediction include weather forecasting and stock market prediction. For classification, we can think, for example, of sentiment analysis and document classification. Automatic image captioning or text generation are part of the sequence generation family of problems, whereas machine translation might be the most familiar example of sequence to sequence prediction we see in our everyday lives.
Our focus for this chapter is on applications of recurrent neural networks for text generation. Since, as we saw previously, text generation is part of a much larger set of problems, many of our algorithms are portable to other contexts.
Training deep learning models is often time-consuming, and recurrent neural networks are not the exception. Our focus is on the ideas over the data, which we will illustrate with smaller datasets than those that you might encounter later on in the wild. This is for the purpose of clarity: We want to make it easier for you to get started on any standard laptop. Once you grasp the basics, you can spin off your own cluster in your favorite cloud provider.

What is so exciting about recurrent neural networks?

Coming from a mathematics background, in my rather hectic career I have seen many different trends, particularly during the last few years, which all sound very similar to me: "you have a problem? wavelets can save you!", "finite elements are the solution to everything", and similar over-enthusiastic claims.
Of course, each tool has its time and place and, more importantly, an application domain where it excels. I find recurrent neural networks quite interesting for the many features they can achieve:
  • Produce consistent markup text (opening and closing tags, recognizing timestamp-like data)
  • Write Wikipedia articles with references, and create URLs from non-existing addresses, by learning what a URL should look like
  • Create credible-looking scientific papers from LaTeX
All these amazing features are possible without the network having any context information or metadata. In particular, without knowing English, nor what a URL or a bit of LaTeX syntax looks like.
These and even more interesting capabilities of neural networks are superbly described by Andrej Karpathy in The Unreasonable Effectiveness of Recurrent Neural Networks: http://karpathy.github.io/2015/05/21/rnn-effectiveness/.
What makes recurrent neural networks exciting? Instead of a constrained fixed-input size to fixed-output size, we can operate over sequences of vectors instead.
A limitation of many machine learning algorithms, including standard feed-forward neural networks, is that they accept a fixed size vector as input and produce a fixed size vector as output. For instance, if we want to classify text, we receive a corpus of documents from which we create a vocabulary to vectorize each document and the output is a vector with class probabilities. Recurrent neural networks instead allow us to take sequences of vectors as input. So, from a one-to-one correspondence between fixed input size and fixed output size, we have a much richer landscape, one-to-one, one-to-many, many-to-one, many-to-many.
Why is that desirable? Let's look at a few examples:
  • One-to-one: Supervised learning, for instance, text classification
  • One-to-many: Given an input text, generate a summary (a sequence of words with important information)
  • Many-to-one: Sentiment analysis in text
  • Many-to-many: Machine translation
Moreover, as recurrent neural networks maintain an internal state which gets updated according to new information, we can view RNNs as a description of a program. In fact, a paper by Siegelman in 1995 shows that recurrent neural networks are Turing complete, they can simulate arbitrary programs.

But what is a recurrent neural network, really?

How does the network keep track of the previous states? To put it in the context of text generation, think of our training data as a list of character sequences (tokenized words). For each word, from the first character, we will predict the following:
Formally, let's denote a sequence of t+1 characters as x = [x0, x1, x2, ... , xt]. Let s-1 =0.
For k=0,2,...t, we construct the following sequence:
This is summarized in the following diagram, when input x is received, the internal state, s, of the network is modified, and then used to generate an output, o:
Figure 1...

Table des matiĂšres

  1. Title Page
  2. Copyright and Credits
  3. About Packt
  4. Contributors
  5. Preface
  6. Getting Started with Deep Learning
  7. Training a Prediction Model
  8. Deep Learning Fundamentals
  9. Training Deep Prediction Models
  10. Image Classification Using Convolutional Neural Networks
  11. Tuning and Optimizing Models
  12. Natural Language Processing Using Deep Learning
  13. Deep Learning Models Using TensorFlow in R
  14. Anomaly Detection and Recommendation Systems
  15. Running Deep Learning Models in the Cloud
  16. The Next Level in Deep Learning
  17. Handwritten Digit Recognition using Convolutional Neural Networks
  18. Traffic Signs Recognition for Intelligent Vehicles
  19. Fraud Detection with Autoencoders
  20. Text Generation using Recurrent Neural Networks
  21. Sentiment Analysis with Word Embedding
  22. Other Books You May Enjoy
Normes de citation pour Deep Learning with R for Beginners

APA 6 Citation

Hodnett, M., Wiley, J., Liu, Y., & Maldonado, P. (2019). Deep Learning with R for Beginners (1st ed.). Packt Publishing. Retrieved from https://www.perlego.com/book/969829/deep-learning-with-r-for-beginners-design-neural-network-models-in-r-35-using-tensorflow-keras-and-mxnet-pdf (Original work published 2019)

Chicago Citation

Hodnett, Mark, Joshua Wiley, Yuxi Liu, and Pablo Maldonado. (2019) 2019. Deep Learning with R for Beginners. 1st ed. Packt Publishing. https://www.perlego.com/book/969829/deep-learning-with-r-for-beginners-design-neural-network-models-in-r-35-using-tensorflow-keras-and-mxnet-pdf.

Harvard Citation

Hodnett, M. et al. (2019) Deep Learning with R for Beginners. 1st edn. Packt Publishing. Available at: https://www.perlego.com/book/969829/deep-learning-with-r-for-beginners-design-neural-network-models-in-r-35-using-tensorflow-keras-and-mxnet-pdf (Accessed: 14 October 2022).

MLA 7 Citation

Hodnett, Mark et al. Deep Learning with R for Beginners. 1st ed. Packt Publishing, 2019. Web. 14 Oct. 2022.