Log In
Or create an account -> 
Imperial Library
  • Home
  • About
  • News
  • Upload
  • Forum
  • Help
  • Login/SignUp

Index
Title Page Copyright and Credits
Deep Learning with R for Beginners
About Packt
Why subscribe? Packt.com
Contributors
About the authors Packt is searching for authors like you
Preface
Who this book is for What this book covers To get the most out of this book
Download the example code files Download the color images of this book Conventions used
Get in touch
Reviews
Getting Started with Deep Learning
What is deep learning? A conceptual overview of neural networks
Neural networks as an extension of linear regression Neural networks as a network of memory cells
Deep neural networks Some common myths about deep learning Setting up your R environment
Deep learning frameworks for R
MXNet Keras
Do I need a GPU (and what is it, anyway)? Setting up reproducible results
Summary
Training a Prediction Model
Neural networks in R
Building neural network models Generating predictions from a neural network
The problem of overfitting data – the consequences explained Use case – building and applying a neural network Summary
Deep Learning Fundamentals
Building neural networks from scratch in R
Neural network web application Neural network code Back to deep learning
The symbol, X, y, and ctx parameters The num.round and begin.round parameters The optimizer parameter The initializer parameter The eval.metric and eval.data parameters The epoch.end.callback parameter The array.batch.size parameter
Using regularization to overcome overfitting
L1 penalty
L1 penalty in action
L2 penalty
L2 penalty in action Weight decay (L2 penalty in neural networks)
Ensembles and model-averaging
Use case – improving out-of-sample model performance using dropout Summary
Training Deep Prediction Models
Getting started with deep feedforward neural networks Activation functions Introduction to the MXNet deep learning library
Deep learning layers Building a deep learning model
Use case – using MXNet for classification and regression
Data download and exploration Preparing the data for our models The binary classification model The regression model Improving the binary classification model The unreasonable effectiveness of data
Summary
Image Classification Using Convolutional Neural Networks
CNNs Convolutional layers
Pooling layers Dropout Flatten layers, dense layers, and softmax
Image classification using the MXNet library
Base model (no convolutional layers) LeNet Classification using the fashion MNIST dataset
References/further reading Summary
Tuning and Optimizing Models
Evaluation metrics and evaluating performance
Types of evaluation metric Evaluating performance
Data preparation
Different data distributions Data partition between training, test, and validation sets Standardization Data leakage
Data augmentation
Using data augmentation to increase the training data
Test time augmentation
Using data augmentation in deep learning libraries
Tuning hyperparameters
Grid search Random search
Use case—using LIME for interpretability
Model interpretability with LIME
Summary
Natural Language Processing Using Deep Learning
Document classification
The Reuters dataset Traditional text classification Deep learning text classification Word vectors Comparing traditional text classification and deep learning
Advanced deep learning text classification
1D convolutional neural network model Recurrent neural network model Long short term memory model Gated Recurrent Units model Bidirectional LSTM model Stacked bidirectional model Bidirectional with 1D convolutional neural network model Comparing the deep learning NLP architectures
Summary
Deep Learning Models Using TensorFlow in R
Introduction to the TensorFlow library
Using TensorBoard to visualize deep learning networks
TensorFlow models
Linear regression using TensorFlow Convolutional neural networks using TensorFlow
TensorFlow estimators and TensorFlow runs packages
TensorFlow estimators TensorFlow runs package
Summary
Anomaly Detection and Recommendation Systems
What is unsupervised learning? How do auto-encoders work?
Regularized auto-encoders
Penalized auto-encoders Denoising auto-encoders
Training an auto-encoder in R
Accessing the features of the auto-encoder model
Using auto-encoders for anomaly detection Use case – collaborative filtering
Preparing the data Building a collaborative filtering model Building a deep learning collaborative filtering model Applying the deep learning model to a business problem
Summary
Running Deep Learning Models in the Cloud
Setting up a local computer for deep learning
How do I know if my model is training on a GPU?
Using AWS for deep learning
A brief introduction to AWS Creating a deep learning GPU instance in AWS Creating a deep learning AMI in AWS
Using Azure for deep learning Using Google Cloud for deep learning Using Paperspace for deep learning Summary
The Next Level in Deep Learning
Image classification models
Building a complete image classification solution
Creating the image data Building the deep learning model Using the saved deep learning model
The ImageNet dataset Loading an existing model Transfer learning
Deploying TensorFlow models Other deep learning topics
Generative adversarial networks Reinforcement learning
Summary
Handwritten Digit Recognition using Convolutional Neural Networks
What is deep learning and why do we need it?
What makes deep learning special? What are the applications of deep learning?
Handwritten digit recognition using CNNs
Get started with exploring MNIST First attempt – logistic regression Going from logistic regression to single-layer neural networks Adding more hidden layers to the networks Extracting richer representation with CNNs
Summary
Traffic Signs Recognition for Intelligent Vehicles
How is deep learning applied in self-driving cars?
How does deep learning become a state-of-the-art solution?
Traffic sign recognition using CNN
Getting started with exploring GTSRB First solution – convolutional neural networks using MXNet Trying something new – CNNs using Keras with TensorFlow Reducing overfitting with dropout
Dealing with a small training set – data augmentation Reviewing methods to prevent overfitting in CNNs Summary
Fraud Detection with Autoencoders
Getting ready
Installing Keras and TensorFlow for R Installing H2O
Our first examples
A simple 2D example Autoencoders and MNIST Outlier detection in MNIST
Credit card fraud detection with autoencoders
Exploratory data analysis The autoencoder approach – Keras Fraud detection with H2O Exercises
Variational Autoencoders
Image reconstruction using VAEs Outlier detection in MNIST
Text fraud detection
From unstructured text data to a matrix From text to matrix representation — the Enron dataset Autoencoder on the matrix representation Exercises
Summary
Text Generation using Recurrent Neural Networks
What is so exciting about recurrent neural networks?
But what is a recurrent neural network, really? LSTM and GRU networks
LSTM GRU
RNNs from scratch in R
Classes in R with R6
Perceptron as an R6 class Logistic regression Multi-layer perceptron
Implementing a RNN
Implementation as an R6 class Implementation without R6 RNN without derivatives — the cross-entropy method
RNN using Keras
A simple benchmark implementation Generating new text from old Exercises
Summary
Sentiment Analysis with Word Embedding
Warm-up – data exploration
Working with tidy text The more, the merrier – calculating n-grams instead of single words
Bag of words benchmark
Preparing the data Implementing a benchmark – logistic regression  Exercises
Word embeddings
word2vec GloVe
Sentiment analysis from movie reviews
Data preprocessing From words to vectors Sentiment extraction The importance of data cleansing Vector embeddings and neural networks Bi-directional LSTM networks Other LSTM architectures Exercises
Mining sentiment from Twitter
Connecting to the Twitter API Building our model
Exploratory data analysis Using a trained model
Summary
Other Books You May Enjoy
Leave a review - let other readers know what you think
  • ← Prev
  • Back
  • Next →
  • ← Prev
  • Back
  • Next →

Chief Librarian: Las Zenow <zenow@riseup.net>
Fork the source code from gitlab
.

This is a mirror of the Tor onion service:
http://kx5thpx2olielkihfyo4jgjqfb7zx7wxr3sd4xzt26ochei4m6f7tayd.onion