This document provides an overview of deep learning including:
- Deep learning has significantly improved applications in computer vision, speech recognition, and natural language processing over the past 10 years.
- It describes some common deep learning models like convolutional neural networks, LSTMs, and feedforward networks.
- It also shows how to deploy deep learning models on AWS services like EC2 and provides examples of using MXNet on common deep learning tasks like image classification and recommender systems.
5. Significantly improve many applications on multiple
domains
“deep learning” trend in the past 10 years
image understanding speech recognition natural language
processing
…
Deep Learning
autonomy
7. Elastic Compute Cloud (EC2)
• C4 Instances
• c4.8xlarge On-Demand:
• $1.675/hr
• GoogleNet inference:
• batch size 32
• 262 ims/sec = 3.8 ms/im
• 1 million images costs $1.77
• Spot prices are cheaper
OS: Linux version 3.13.0-86-generic (buildd@lgw01-51) (gcc version 4.8.2 (Ubuntu 4.8.2-19ubuntu1) ) #131-Ubuntu SMP Thu May 12 23:33:13 UTC 2016.
MxNet Tip of tree: commit de41c736422d730e7cfad72dd6afc229ce08cf90, Tue Nov 1 11:43:04 2016 +0800. MKL 2017 Gold update 1
7
6.1 2.4 1.2 0.8
679.5
262.5
79.7 73.9
0
200
400
600
800
AlexNet GoogLeNet v1 ResNet-50 Inception v3
Images/Sec
c4.8xlarge MXNet Inference
No MKL MKL
8. deploy Bigdl on aws
Github: github.com/intel-analytics/BigDL
http://software.intel.com/bigdl
BigDL, A Distributed Deep learning framework for Apache
Spark*
Deploying BigDL on AWS is super easy!
Option 1: Install BigDL on Amazon EMR with Bootstrap action
s3://aws-bigdata-blog/artifacts/aws-blog-emr-jupyter/install-jupyter-emr5-
latest.sh –bigdl
Option 2: Launch Public AMI on EC2 w/Xeon E5 v3 or v4
https://github.com/intel-analytics/BigDL/wiki/Running-on-EC2
Learn how to Classify text, train a Convolutional neural
network, visualize the training using Tensorboard using
BigDL on AWS.
https://aws.amazon.com/blogs/ai/running-
bigdl-deep-learning-for-apache-spark-on-
aws/
9. Launch your notebooks
• Follow the instructions on http://bit.ly/MXNet-blog
#!/bin/bash -v
sudo su -
cd ~
export PYTHONPATH=/home/ubuntu/mxnet/python
export LD_LIBRARY_PATH=/usr/local/cuda/lib64/
git clone https://github.com/dmlc/mxnet-notebooks
jupyter notebook
ssh -L 127.0.0.1:8888:127.0.0.1:8888 -i PEM-KEY ec2-user@EC2-DNS
1000000 M * $1.675 / hr * 1 hr/3600 s * 1 s / 262.5 imgs = $1.77
BigDL is a distributed deep learning framework for Apache Spark that was developed by Intel and contributed to the open source community for the purposes of uniting big data processing and deep learning. BigDL helps make deep learning more accessible to the big data community by allowing developers to continue using familiar tools and infrastructure to build deep learning applications.
As shown in the diagram on the right, BigDL is implemented as a library on top of Spark, so that users can write their deep learning applications as standard Spark programs. As a result, BigDL can be seamlessly integrated with other libraries on top of Spark—Spark SQL and DataFrames, Spark ML pipelines, Spark Streaming, Structured Streaming, etc.—and can run directly on top of existing Spark or Hadoop clusters.
It is super easy to deploy BigDL on AWS. There is a public AMI available to launch BigDL on EC2 Spark instances with Xeon E5 v3 or v4 processors. The Project Git has a detailed tutorial on how to launch this AMI and run the BigDL examples for training various models.
The AWS team also partnered with the BigDL project team to develop a tutorial for running BigDL on AWS – we have simplified it with a bootstrap action….you can use the following bootstrap action with the –bigdl argument in an AWS CLI command or in the AWS EMR console…You can follow the tutorial blog we published and try it out.
We have seen major advancements in deep learning in the recent years; while the deep learning community continue to push the technology envelope, the appeal of a project like BigDL on Apache Spark is that it helps make these breakthroughs more accessible and convenient to use for data scientists and data engineers (who are not necessarily experts in deep learning technologies).
NN can take every type of input and map it to every type of classification
The biological neurological model that is able to learn human intelligence such as vision, language, problem solving and every other human abilities.
The brain is based on chemistry models, and it 100 billions of neurons, and it has modes as depression or high adrenaline, which we are still not simulating in NN.
Deep Structured Semantic Model was developed by MSR, and allows the integration of multiple type of NN output and input to be used together.
Deep Structured Semantic Model was developed by MSR, and allows the integration of multiple type of NN output and input to be used together.
Deep Structured Semantic Model was developed by MSR, and allows the integration of multiple type of NN output and input to be used together.