18. online vs. historic
• Pros
• low storage costs
• real-time model update
• Cons
• algorithm support
• software support
• no algorithmic improvement
• compute power to be inline
with data rate
• Pros
• all algorithms
• abundance of software
• model re-scoring / re-
parameterisation (algorithmic
improvement)
• batch processing
• Cons
• high storage costs
• batch model update
19. 1. API
2. pre-trained model
3. existing pipeline with your data
4. create own pipeline / model
abstraction levels
40. •Outperformed traditional methods, such as
•cumulative sum (CUSUM)
•exponentially weighted moving average (EWMA)
•Hidden Markov Models (HMM)
•Learned what “Normal” is
•Raised error if time series pattern haven't been seen
before
41.
42.
43. Learning of an algorithm
A LSTM network is touring complete
44. Problems
• Neural Networks are computationally very complex
•especially during training
•but also during scoring
CPU (2009) GPU (2016) IBM TrueNorth (2017)
46. DeepLearning
the future in cloud based analytics
Storage Layer (OpenStack SWIFT / Hadoop HDFS / IBM GPFS)
Execution Layer (Spark Executor, YARN, Platform Symphony)
Hardware Layer (Bare Metal High Performance Cluster)
GraphXStreaming SQL MLLib BlinkDB
DeepLearning4J
ND4J
R MLBase H2O
Y
O
U
GPUAVX
Intel Xeon E7-4850 v2 48 core, 3 TB RAM, 72 GB HDD, 10Gbps, NVIDIA TESLA M60 GPU
(cu)BLAS
jcuBLAS
S
T
R
E
A
M
S