SlideShare a Scribd company logo
1 of 66
Download to read offline
Terry Taewoong Um (terry.t.um@gmail.com)
University of Waterloo
Department of Electrical & Computer Engineering
Terry Taewoong Um
MACHINE LEARNING,
DEEP LEARNING, AND
MOTION ANALYSIS
1
Terry Taewoong Um (terry.t.um@gmail.com)
CAUTION
โ€ข I cannot explain everything
โ€ข You cannot get every details
2
โ€ข Try to get a big picture
โ€ข Get some useful keywords
โ€ข Connect with your research
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
1. What is Machine Learning?
(Part 1 Q & A)
2. What is Deep Learning?
(Part 2 Q & A)
3. Machine Learning in Motion Analysis
(Part 3 Q & A)
3
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
4
1. What is Machine Learning?
Terry Taewoong Um (terry.t.um@gmail.com)
WHAT IS MACHINE LEARNING?
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience Eโ€œ โ€“ T. Michell (1997)
Example: A program for soccer tactics
5
T : Win the game
P : Goals
E : (x) Playersโ€™ movements
(y) Evaluation
Terry Taewoong Um (terry.t.um@gmail.com)
WHAT IS MACHINE LEARNING?
6
โ€œToward learning robot table tennisโ€, J. Peters et al. (2012)
https://youtu.be/SH3bADiB7uQ
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience Eโ€œ โ€“ T. Michell (1997)
Terry Taewoong Um (terry.t.um@gmail.com)
TASKS
7
classification
discrete target values
x : pixels (28*28)
y : 0,1, 2,3,โ€ฆ,9
regression
real target values
x โˆˆ (0,100)
y : 0,1, 2,3,โ€ฆ,9
clustering
no target values
x โˆˆ (-3,3)ร—(-3,3)
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience Eโ€œ โ€“ T. Michell (1997)
Terry Taewoong Um (terry.t.um@gmail.com)
PERFORMANCE
8
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience Eโ€œ โ€“ T. Michell (1997)
classification
0-1 loss function
regression
L2 loss function
clustering
Terry Taewoong Um (terry.t.um@gmail.com)
EXPERIENCE
9
"A computer program is said to learn from experience E
with respect to some class of tasks T and performance
measure P, if its performance at tasks in T, as measured
by P, improves with experience Eโ€œ โ€“ T. Michell (1997)
classification
labeled data
(pixels)โ†’(number)
regression
labeled data
(x) โ†’ (y)
clustering
unlabeled data
(x1,x2)
Terry Taewoong Um (terry.t.um@gmail.com)
A TOY EXAMPLE
10
? Height(cm)
Weight
(kg)
[Input X]
[Output Y]
Terry Taewoong Um (terry.t.um@gmail.com)
11
180 Height(cm)
Weight
(kg)
80
Y = aX+b
Model : Y = aX+b Parameter : (a, b)
[Goal] Find (a,b) which best fits the given data
A TOY EXAMPLE
Terry Taewoong Um (terry.t.um@gmail.com)
12
[Analytic Solution]
Least square problem
(from AX = b, X=A#b where
A# is Aโ€™s pseudo inverse)
Not always available
[Numerical Solution]
1. Set a cost function
2. Apply an optimization method
(e.g. Gradient Descent (GD) Method)
L
(a,b)
http://www.yaldex.com/game-
development/1592730043_ch18lev1sec4.html
Local minima problem
http://mnemstudio.org/neural-networks-
multilayer-perceptron-design.htm
A TOY EXAMPLE
Terry Taewoong Um (terry.t.um@gmail.com)
13
32 Age(year)
Running
Record
(min)
140
WHAT WOULD BE THE CORRECT MODEL?
Select a model โ†’ Set a cost function โ†’ Optimization
Terry Taewoong Um (terry.t.um@gmail.com)
14
? X
Y
WHAT WOULD BE THE CORRECT MODEL?
1. Regularization 2. Nonparametric model
โ€œoverfittingโ€
Terry Taewoong Um (terry.t.um@gmail.com)
15
L2 REGULARIZATION
(e.g. w=(a,b) where Y=aX+b)
Avoid a complicated model!
โ€ข Another interpretation :
: Maximum a Posteriori (MAP)
http://goo.gl/6GE2ix
http://goo.gl/6GE2ix
Terry Taewoong Um (terry.t.um@gmail.com)
16
L2 REGULARIZATION
โ€ข Another interpretation :
: Maximum a Posteriori (MAP)
http://goo.gl/6GE2ix
http://goo.gl/6GE2ix
โ€ข Bayesian inference
๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ ๐ท๐‘Ž๐‘ก๐‘Ž =
๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ ๐‘ƒ(๐ท๐‘Ž๐‘ก๐‘Ž|๐ต๐‘’๐‘™๐‘–๐‘’๐‘“)
๐‘ƒ(๐ท๐‘Ž๐‘ก๐‘Ž)
posterior
prior likelihood
ex) fair coin : 50% H, 50% T
falsified coin : 80% H, 20% T
Letโ€™s say we observed ten heads consecutively.
Whatโ€™s the probability for being a fair coin?
๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ = 0.2
๐‘ƒ ๐ท๐‘Ž๐‘ก๐‘Ž|๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ = 0.510
โ‰ˆ 0.001
๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“|๐ท๐‘Ž๐‘ก๐‘Ž โˆ 0.2 โˆ— 0.001 = 0.0002
normalization
(you donโ€™t believe this coin is fair)
Fair
coin?
Falsified
coin?
๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ = 0.8
๐‘ƒ ๐ท๐‘Ž๐‘ก๐‘Ž|๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ = 0.810
โ‰ˆ 0.107
๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“|๐ท๐‘Ž๐‘ก๐‘Ž โˆ 0.8 โˆ— 0.107 = 0.0856
Fair =
0.0002
0.0002+0.0856
= 0.23% , Unfair = 99.77%
Terry Taewoong Um (terry.t.um@gmail.com)
17
WHAT WOULD BE THE CORRECT MODEL?
1. Regularization 2. Nonparametric model
training time
error
training error
test error
we should
stop here
training
set
validation
set
test
set
for training
(parameter
optimization)
for early
stopping
(avoid
overfitting)
for evaluation
(measure the
performance)
keep watching the validation error
Terry Taewoong Um (terry.t.um@gmail.com)
18
NONPARAMETRIC MODEL
โ€ข It does not assume any parametric models (e.g. Y = aX+b, Y=aX2+bX+c, etc.)
โ€ข It often requires much more samples
โ€ข Kernel methods are frequently applied for modeling the data
โ€ข Gaussian Process Regression (GPR), a sort of kernel method, is a widely-used
nonparametric regression method
โ€ข Support Vector Machine (SVM), also a sort of kernel method, is a widely-used
nonparametric classification method
kernel function
[Input space] [Feature space]
Terry Taewoong Um (terry.t.um@gmail.com)
19
SUPPORT VECTOR MACHINE (SVM)
โ€œMyoโ€, Thalmic Labs (2013)
https://youtu.be/oWu9TFJjHaM
[Linear classifiers] [Maximum margin]
Support vector Machine Tutorial, J. Weston, http://goo.gl/19ywcj
[Dual formulation] ( )
kernel function
kernel function
Terry Taewoong Um (terry.t.um@gmail.com)
20
GAUSSIAN PROCESS REGRESSION (GPR)
https://youtu.be/YqhLnCm0KXY
https://youtu.be/kvPmArtVoFE
โ€ข Gaussian Distribution
โ€ข Multivariate regression likelihood
posterior
prior
likelihood
prediction conditioning the joint distribution of the observed & predicted values
https://goo.gl/EO54WN
http://goo.gl/XvOOmf
Terry Taewoong Um (terry.t.um@gmail.com)
21
DIMENSION REDUCTION
[Original space] [Feature space]
low dim. high dim.
high dim. low dim.
๐‘‹ โ†’ โˆ…(๐‘‹)
โ€ข Principal Component Analysis
: Find the best orthogonal axes
(=principal components) which
maximize the variance of the data
Y = P X
* The rows in P are m largest eigenvectors
of
1
๐‘
๐‘‹๐‘‹ ๐‘‡
(covariance matrix)
Terry Taewoong Um (terry.t.um@gmail.com)
22
DIMENSION REDUCTION
http://jbhuang0604.blogspot.kr/2013/04/miss-korea-2013-contestants-face.html
Terry Taewoong Um (terry.t.um@gmail.com)
23
SUMMARY - PART 1
โ€ข Machine Learning
- Tasks : Classification, Regression, Clustering, etc.
- Performance : 0-1 loss, L2 loss, etc.
- Experience : labeled data, unlabelled data
โ€ข Machine Learning Process
(1) Select a parametric / nonparametric model
(2) Set a performance measurement including regularization term
(3) Training data (optimizing parameters) until validation error increases
(4) Evaluate the final performance using test set
โ€ข Nonparametric model : Support Vector Machine, Gaussian Process Regression
โ€ข Dimension reduction : used as pre-processing data
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
24
Questions about Part 1?
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
25
2. What is Deep Learning?
Terry Taewoong Um (terry.t.um@gmail.com)
26
PARADIGM CHANGE
PAST
Knowledge
ML
Method
(e.g.
GPR, SVM)
PRESENT
What is the best
ML method for
the target task?
Knowledge
Representation
How can we find a
good representation?
Terry Taewoong Um (terry.t.um@gmail.com)
27
PARADIGM CHANGE
Knowledge
PRESENT
Representation
How can we find a
good representation?
kernel function
Terry Taewoong Um (terry.t.um@gmail.com)
28
PARADIGM CHANGE
Knowledge
PRESENT
Representation
(Features)
How can we find a
good representation?
IMAGE
SPEECH
Hand-Crafted Features
Terry Taewoong Um (terry.t.um@gmail.com)
29
PARADIGM CHANGE
IMAGE
SPEECH
Hand-Crafted Features
Knowledge
PRESENT
Representation
(Features)
Can we learn a good representation
(feature) for the target task as well?
Terry Taewoong Um (terry.t.um@gmail.com)
30
DEEP LEARNING
โ€ข What is Deep Learning (DL) ?
- Learning methods which have deep (not shallow) architecture
- It often allows end-to-end learning
- It automatically finds intermediate representation. Thus,
it can be regarded as a representation learning
- It often contains stacked โ€œneural networkโ€. Thus,
Deep learning usually indicates โ€œdeep neural networkโ€
โ€œDeep Gaussian Processโ€ (2013)
https://youtu.be/NwoGqYsQifg
http://goo.gl/fxmmPE
http://goo.gl/5Ry08S
Terry Taewoong Um (terry.t.um@gmail.com)
31
OUTSTANDING PERFORMANCE OF DL
error rate : 28% โ†’ 15% โ†’ 8%
(2010) (2014)(2012)
- Object recognition (Simonyan et al., 2015)
- Natural machine translation (Bahdanau et al., 2014)
- Speech recognition (Chorowski et al., 2014)
- Face recognition (Taigman et al., 2014)
- Emotion recognition (Ebrahimi-Kahou et al., 2014)
- Human pose estimation (Jain et al., 2014)
- Deep reinforcement learning(mnih et al., 2013)
- Image/Video caption (Xu et al., 2015)
- Particle physics (Baldi et al., 2014)
- Bioinformatics (Leung et al., 2014)
- And so onโ€ฆ.
โ€ข State-of-art results achieved by DL
DL has won most of ML challenges!
K. Cho, https://goo.gl/vdfGpu
Terry Taewoong Um (terry.t.um@gmail.com)
32
BIOLOGICAL EVIDENCE
โ€ข Somatosensory cortex learns to see
โ€ข Why do we need different ML methods
for different task?
Yann LeCun, https://goo.gl/VVQXJG
โ€ข The vental pathway in the visual cortex has multiple stages
โ€ข There exist a lot of intermediate representations
Andrew Ng, https://youtu.be/ZmNOAtZIgIk
Terry Taewoong Um (terry.t.um@gmail.com)
33
BIG MOVEMENT
http://goo.gl/zNbBE2 http://goo.gl/Lk64Q4
Going deeper and deeperโ€ฆ.
Terry Taewoong Um (terry.t.um@gmail.com)
34
NEURAL NETWORK (NN)
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
โ€ข Universal approximation theorem (Hornik, 1991)
- A single hidden layer NN w/ linear output can approximate any cont. func. arbitrarily well,
given enough hidden units
- This does not imply we have learning method to train them
Terry Taewoong Um (terry.t.um@gmail.com)
35
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
โ€ข First, calculate the output using data & initial parameters (W ,b)
โ€ข Activation functions
http://goo.gl/qMQk5H
1
Terry Taewoong Um (terry.t.um@gmail.com)
36
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
โ€ข Then, calculate the error and update the weights from top to bottom
โ€ข Parameter gradients
http://goo.gl/qMQk5H
: Backpropagation algorithm
2
known
Terry Taewoong Um (terry.t.um@gmail.com)
37
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
โ€ข Then, calculate the error and update the weights from top to bottom
โ€ข Parameter gradients
http://goo.gl/qMQk5H
: Backpropagation algorithm
2
known
Terry Taewoong Um (terry.t.um@gmail.com)
38
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
โ€ข Then, calculate the error and update the weights from top to bottom
โ€ข Parameter gradients
http://goo.gl/qMQk5H
: Backpropagation algorithm
2
known
Terry Taewoong Um (terry.t.um@gmail.com)
39
TRAINING NN
Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html
โ€ข Then, calculate the error and update the weights from top to bottom
โ€ข Parameter gradients
http://goo.gl/qMQk5H
: Backpropagation algorithm
2
known
Terry Taewoong Um (terry.t.um@gmail.com)
40
TRAINING NN
โ€ข Repeat this process with different dataset(mini-batches)
http://goo.gl/qMQk5H
- Forward propagation (calculate the output values)
- Evaluate the error
- Backward propagation (update the weights)
- Repeat this process until the error converges
3
โ€ข As you can see here, NN is not a fancy algorithm,
but just a iterative gradient descent method with
huge number of parameters
โ€ข NN is often likely to be
stuck in local minima pitfall
Terry Taewoong Um (terry.t.um@gmail.com)
41
FROM NN TO DEEP NN
โ€ข From NN to deep NN (since 2006)
- NN requires expertโ€™s skill to tune the hyperparameters
- It sometimes gives a good result, but sometimes gives a bad result.
The result is highly depend on the quality of initialization, regularization,
hyperparameters, data, etc.
- Local minima is always problematic
โ€ข A long winter of NN
Yann LeCun
(NYU, Facebook)
Yoshua Bengio
(U. Montreal)
Geoffrey Hinton
(U. Toronto, Google)
Terry Taewoong Um (terry.t.um@gmail.com)
42
WHY IS DL SO SUCCESSFUL?
http://t-robotics.blogspot.kr/2015/05/deep-learning.html
โ€ข Pre-training with unsupervised learning
โ€ข Convolutional Neural Network
โ€ข Recurrent Neural Net
โ€ข GPGPU (parallel processing) & big data
โ€ข Advanced algorithms for optimization,
activation, regularization
โ€ข Huge research society
(Vision, Speech, NLP, Biology, etc.)
Terry Taewoong Um (terry.t.um@gmail.com)
43
UNSUPERVISED LEARNING
โ€ข How can we avoid pathologic local minima cases?
(1) First, pre-train the data with unsupervised learning method
and get a new representation
(2) Stack up this block structures
(3) Training each layer in end-to-end manner
(4) Fine tune the final structure with (ordinary) fully-connected NN
โ€ข Unsupervised learning method
- Restricted Boltzmann Machine (RBM)
โ†’ Deep RBM, Deep Belief Network (DBN)
- Autoencoder
โ†’ Deep Auto-encoder
http://goo.gl/QGJm5k
Autoencoder http://goo.gl/s6kmqY
Terry Taewoong Um (terry.t.um@gmail.com)
44
UNSUPERVISED LEARNING
โ€œConvolutional deep belief networks for scalable unsupervised learning of hierarchical representationโ€, Lee et al., 2012
Terry Taewoong Um (terry.t.um@gmail.com)
45
CONVOLUTIONAL NN
โ€ข How can we deal with real images which is
much bigger than MNIST digit images?
- Use not fully-connected, but locally-connected NN
- Use convolutions to get various feature maps
- Abstract the results into higher layer by using pooling
- Fine tune with fully-connected NN
https://goo.gl/G7kBjI
https://goo.gl/Xswsbd
http://goo.gl/5OR5oH
Terry Taewoong Um (terry.t.um@gmail.com)
46
CONVOLUTIONAL NN
โ€œVisualization and Understanding Convolutional Networkโ€, Zeiler et al., 2012
Terry Taewoong Um (terry.t.um@gmail.com)
47
CONVNET + RNN
โ€œLarge-scale Video Classification with Convolutional Neural Networkโ€,
A. Karpathy 2014, https://youtu.be/qrzQ_AB1DZk
Terry Taewoong Um (terry.t.um@gmail.com)
48
RECURRENT NEURAL NETWORK (RNN)
t-1 t t+1
[Neural Network] [Recurrent Neural Network]
http://www.dmi.usherb.ca/~larocheh/index_en.html
Terry Taewoong Um (terry.t.um@gmail.com)
49
RECURRENT NEURAL NETWORK (RNN)
[Neural Network] [Recurrent Neural Network]
back propagation
back propagation
through time
(BPTT)
โ€ข Vanishing gradient problem : Canโ€™t have long memory!
โ€œTraining Recurrent Neural Networks, I. Sutskever, 2013
Terry Taewoong Um (terry.t.um@gmail.com)
50
RNN + LSTM
โ€ข Long Short-Term Memory (LSTM) (Hochreiter and Schmidhuber, 1997)
โ€œTraining Recurrent Neural Networks, I. Sutskever, 2013
Terry Taewoong Um (terry.t.um@gmail.com)
51
INTERESTING RESULTS FROM RNN
http://pail.unist.ac.kr/carpedm20/poet/
http://cs.stanford.edu/people/karpathy/deepimagesent/
โ€œgenerating sequences with RNNโ€,
A.Graves, 2013
Terry Taewoong Um (terry.t.um@gmail.com)
52
WHY IS DL SO SUCCESSFUL?
http://t-robotics.blogspot.kr/2015/05/deep-learning.html
โ€ข Pre-training with unsupervised learning
โ€ข Convolutional Neural Network
โ€ข Recurrent Neural Net
โ€ข GPGPU (parallel processing) & big data
โ€ข Advanced algorithms for optimization,
activation, regularization
โ€ข Huge research society
(Vision, Speech, NLP, Biology, etc.)
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
53
Questions about Part 2?
Terry Taewoong Um (terry.t.um@gmail.com)
CONTENTS
54
3. Machine Learning in
Motion Analysis
Terry Taewoong Um (terry.t.um@gmail.com)
55
MOTION DATA
โ€œ์ธ„๋ฆฌ๋‹โ€, ์ด์ƒ์‹  ๊ตญ์ค‘๋ก
Terry Taewoong Um (terry.t.um@gmail.com)
56
MOTION DATA
We need to know the state not only at time t
but also at time t-1, t-2, t-3, etc.
๐‘“ = ๐‘“(๐‘ฅ, ๐‘ก)
โ€œ์ธ„๋ฆฌ๋‹โ€, ์ด์ƒ์‹  ๊ตญ์ค‘๋ก
Terry Taewoong Um (terry.t.um@gmail.com)
57
MOTION DATA
โ€ข Why do motion data need special treatment?
- In general, most machine learning techniques assume i.i.d. (independent
& identically distributed) sampling condition.
e.g.) coins tossing
- However, motion data is temporally & spatially correlated
http://goo.gl/LQulvcswing motion manipulability ellipsoid https://goo.gl/dHjFO9
Terry Taewoong Um (terry.t.um@gmail.com)
58
MOTION DATA
http://goo.gl/ll3sq6
We can infer the next state
based on the temporal &
spatial information
But, how can we exploit
those benefits in ML method?
Terry Taewoong Um (terry.t.um@gmail.com)
59
WHAT CAN WE DO WITH MOTION DATA?
โ€ข Learning the kinematic/dynamic model
โ€ข Motion segmentation
โ€ข Motion generation / synthesis
โ€ข Motion imitation (Imitation learning)
โ€ข Activity / Gesture recognition
TASKS
Data
โ€ข Motion capture data
โ€ข Vision Data
โ€ข Dynamic-level data
Applications
โ€ข Biomechanics
โ€ข Humanoid
โ€ข Animation
http://goo.gl/gFOVWL
Terry Taewoong Um (terry.t.um@gmail.com)
60
HIDDEN MARKOV MODEL (HMM)
Prob. of (n+1) state only depends on state at (n+1)
Terry Taewoong Um (terry.t.um@gmail.com)
61
LIMITATIONS OF HMM
1. Extract features (e.g. PCA)
2. Define the HMM structure (e.g. using GMM)
3. Train a separate HMM per class (Baum-Welch algorithm)
4. Evaluate probability under each HMM (Fwd/Bwd algorithm)
or 3. Choose most probable sequence (Viterbi algorithm)
- HMM handle discrete states only!
- HMM has short memory! (using just the previous state)
- HMM has limited expressive power!
- [Trend1] features-GMM โ†’ unsupervised learning methods
- [Trend2] features-GMM-HMM โ†’ recurrent neural network
โ€ข A common procedure of HMM for motion analysis
โ€ข Limitations & trend change in speech recognition area
Terry Taewoong Um (terry.t.um@gmail.com)
62
CAPTURE TEMPORAL INFORMATION
โ€ข 3D ConvNet
- โ€œ3D Convolutional Neural Network for
Human Action Recognitionโ€ (Ji et al., 2010)
- 3D convolution
- Activity recognition / Pose estimation from video
โ€œJoint Training of a Convolutional Network
and a Graphical Model for Human Pose
Estimationโ€, Tompson et al., 2014
Terry Taewoong Um (terry.t.um@gmail.com)
63
CAPTURE TEMPORAL INFORMATION
โ€ข Recurrent Neural Network (RNN)
โ€œHierarchical Recurrent Neural Network for Skeleton Based Action Recognitionโ€, Y. Du et al., 2015
โ€ข However, how can we capture the
spatial information about motions?
Terry Taewoong Um (terry.t.um@gmail.com)
64
CHALLENGES
We should connect the geometric information with deep neural network!
โ€ข The link transformation from the i-1 th link to the i th link
โ€ข Forward Kinematics
constant, Mvariable, ๐œƒ
c.f.)
๐‘‹๐‘–โˆ’1,๐‘– = ๐‘…๐‘œ๐‘ก ๐‘ง, ๐œƒ๐‘– ๐‘‡๐‘Ÿ๐‘Ž๐‘›๐‘  ๐‘ง, ๐‘‘๐‘– ๐‘‡๐‘Ÿ๐‘Ž๐‘›๐‘  ๐‘ฅ, ๐‘Ž๐‘– ๐‘…๐‘œ๐‘ก ๐‘ง, ๐›ผ๐‘– = ๐‘’[๐ด ๐‘–]๐œƒ ๐‘– ๐‘€๐‘–โˆ’1,๐‘–
๐‘‹0,๐‘› = ๐‘’[๐ด1]๐œƒ1 ๐‘€0,1 ๐‘’[๐ด2]๐œƒ2 ๐‘€1,2 โ‹ฏ ๐‘’ ๐ด ๐‘› ๐œƒ ๐‘› ๐‘€ ๐‘›โˆ’1,๐‘›
= ๐‘’[๐‘†1]๐œƒ1 ๐‘’[๐‘†2]๐œƒ2 โ‹ฏ ๐‘’[๐‘† ๐‘›]๐œƒ ๐‘› ๐‘€0,๐‘›
๐‘†๐‘– = ๐ด๐‘‘ ๐‘€01โ‹ฏ๐‘€ ๐‘–โˆ’2,๐‘–โˆ’1
๐ด๐‘– , ๐‘– = 1, โ‹ฏ , ๐‘›
propagated forces
external force acting
on the ith body where
โ€ข Newton-Euler formulation for inverse dynamics
Lie group & Lie algebra,
http://goo.gl/uqilDV
Terry Taewoong Um (terry.t.um@gmail.com)
65
CHALLENGES
https://www.youtube.com/watch?v=oxA2O-tHftI
Terry Taewoong Um (terry.t.um@gmail.com)
66
Thank you

More Related Content

What's hot

What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
Simplilearn
ย 
Deep learning.pptx
Deep learning.pptxDeep learning.pptx
Deep learning.pptx
MdMahfoozAlam5
ย 

What's hot (20)

Image classification using convolutional neural network
Image classification using convolutional neural networkImage classification using convolutional neural network
Image classification using convolutional neural network
ย 
Intro/Overview on Machine Learning Presentation
Intro/Overview on Machine Learning PresentationIntro/Overview on Machine Learning Presentation
Intro/Overview on Machine Learning Presentation
ย 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
ย 
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
ย 
Machine Learning
Machine LearningMachine Learning
Machine Learning
ย 
Convolutional neural network
Convolutional neural networkConvolutional neural network
Convolutional neural network
ย 
Introduction to Deep Learning
Introduction to Deep LearningIntroduction to Deep Learning
Introduction to Deep Learning
ย 
AI for Everyone: Master the Basics
AI for Everyone: Master the BasicsAI for Everyone: Master the Basics
AI for Everyone: Master the Basics
ย 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
ย 
Deep Learning
Deep Learning Deep Learning
Deep Learning
ย 
Deep Learning in Computer Vision
Deep Learning in Computer VisionDeep Learning in Computer Vision
Deep Learning in Computer Vision
ย 
Introduction to ML (Machine Learning)
Introduction to ML (Machine Learning)Introduction to ML (Machine Learning)
Introduction to ML (Machine Learning)
ย 
Deep learning
Deep learningDeep learning
Deep learning
ย 
Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders
ย 
Introduction of Deep Learning
Introduction of Deep LearningIntroduction of Deep Learning
Introduction of Deep Learning
ย 
Machine learning
Machine learningMachine learning
Machine learning
ย 
Machine learning
Machine learningMachine learning
Machine learning
ย 
HML: Historical View and Trends of Deep Learning
HML: Historical View and Trends of Deep LearningHML: Historical View and Trends of Deep Learning
HML: Historical View and Trends of Deep Learning
ย 
Machine learning overview
Machine learning overviewMachine learning overview
Machine learning overview
ย 
Deep learning.pptx
Deep learning.pptxDeep learning.pptx
Deep learning.pptx
ย 

Viewers also liked

์•ŒํŒŒ๊ณ  (๋ฐ”๋‘‘ ์ธ๊ณต์ง€๋Šฅ)์˜ ์ž‘๋™ ์›๋ฆฌ
์•ŒํŒŒ๊ณ  (๋ฐ”๋‘‘ ์ธ๊ณต์ง€๋Šฅ)์˜ ์ž‘๋™ ์›๋ฆฌ์•ŒํŒŒ๊ณ  (๋ฐ”๋‘‘ ์ธ๊ณต์ง€๋Šฅ)์˜ ์ž‘๋™ ์›๋ฆฌ
์•ŒํŒŒ๊ณ  (๋ฐ”๋‘‘ ์ธ๊ณต์ง€๋Šฅ)์˜ ์ž‘๋™ ์›๋ฆฌ
Shane (Seungwhan) Moon
ย 
Deview deep learning-แ„€แ…ตแ†ทแ„Œแ…ฅแ†ผแ„’แ…ด
Deview deep learning-แ„€แ…ตแ†ทแ„Œแ…ฅแ†ผแ„’แ…ดDeview deep learning-แ„€แ…ตแ†ทแ„Œแ…ฅแ†ผแ„’แ…ด
Deview deep learning-แ„€แ…ตแ†ทแ„Œแ…ฅแ†ผแ„’แ…ด
NAVER D2
ย 

Viewers also liked (20)

๊ธฐ๊ณ„ํ•™์Šต(Machine learning) ์ž…๋ฌธํ•˜๊ธฐ
๊ธฐ๊ณ„ํ•™์Šต(Machine learning) ์ž…๋ฌธํ•˜๊ธฐ๊ธฐ๊ณ„ํ•™์Šต(Machine learning) ์ž…๋ฌธํ•˜๊ธฐ
๊ธฐ๊ณ„ํ•™์Šต(Machine learning) ์ž…๋ฌธํ•˜๊ธฐ
ย 
Human Motion Forecasting (Generation) with RNNs
Human Motion Forecasting (Generation) with RNNsHuman Motion Forecasting (Generation) with RNNs
Human Motion Forecasting (Generation) with RNNs
ย 
Understanding Black-box Predictions via Influence Functions (2017)
Understanding Black-box Predictions via Influence Functions (2017)Understanding Black-box Predictions via Influence Functions (2017)
Understanding Black-box Predictions via Influence Functions (2017)
ย 
Learning with side information through modality hallucination (2016)
Learning with side information through modality hallucination (2016)Learning with side information through modality hallucination (2016)
Learning with side information through modality hallucination (2016)
ย 
Deformable Convolutional Network (2017)
Deformable Convolutional Network (2017)Deformable Convolutional Network (2017)
Deformable Convolutional Network (2017)
ย 
๊ธฐ๊ณ„ํ•™์Šต / ๋”ฅ๋Ÿฌ๋‹์ด๋ž€ ๋ฌด์—‡์ธ๊ฐ€
๊ธฐ๊ณ„ํ•™์Šต / ๋”ฅ๋Ÿฌ๋‹์ด๋ž€ ๋ฌด์—‡์ธ๊ฐ€๊ธฐ๊ณ„ํ•™์Šต / ๋”ฅ๋Ÿฌ๋‹์ด๋ž€ ๋ฌด์—‡์ธ๊ฐ€
๊ธฐ๊ณ„ํ•™์Šต / ๋”ฅ๋Ÿฌ๋‹์ด๋ž€ ๋ฌด์—‡์ธ๊ฐ€
ย 
Lie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsLie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot Mechanics
ย 
Lie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot MechanicsLie Group Formulation for Robot Mechanics
Lie Group Formulation for Robot Mechanics
ย 
์ซ„์ง€๋ง์ž๋”ฅ๋Ÿฌ๋‹2 - CNN RNN ํฌํ•จ๋ฒ„์ „
์ซ„์ง€๋ง์ž๋”ฅ๋Ÿฌ๋‹2 - CNN RNN ํฌํ•จ๋ฒ„์ „์ซ„์ง€๋ง์ž๋”ฅ๋Ÿฌ๋‹2 - CNN RNN ํฌํ•จ๋ฒ„์ „
์ซ„์ง€๋ง์ž๋”ฅ๋Ÿฌ๋‹2 - CNN RNN ํฌํ•จ๋ฒ„์ „
ย 
์•ŒํŒŒ๊ณ  (๋ฐ”๋‘‘ ์ธ๊ณต์ง€๋Šฅ)์˜ ์ž‘๋™ ์›๋ฆฌ
์•ŒํŒŒ๊ณ  (๋ฐ”๋‘‘ ์ธ๊ณต์ง€๋Šฅ)์˜ ์ž‘๋™ ์›๋ฆฌ์•ŒํŒŒ๊ณ  (๋ฐ”๋‘‘ ์ธ๊ณต์ง€๋Šฅ)์˜ ์ž‘๋™ ์›๋ฆฌ
์•ŒํŒŒ๊ณ  (๋ฐ”๋‘‘ ์ธ๊ณต์ง€๋Šฅ)์˜ ์ž‘๋™ ์›๋ฆฌ
ย 
[2A4]DeepLearningAtNAVER
[2A4]DeepLearningAtNAVER[2A4]DeepLearningAtNAVER
[2A4]DeepLearningAtNAVER
ย 
์•ŒํŒŒ๊ณ  ํ•ด๋ถ€ํ•˜๊ธฐ 1๋ถ€
์•ŒํŒŒ๊ณ  ํ•ด๋ถ€ํ•˜๊ธฐ 1๋ถ€์•ŒํŒŒ๊ณ  ํ•ด๋ถ€ํ•˜๊ธฐ 1๋ถ€
์•ŒํŒŒ๊ณ  ํ•ด๋ถ€ํ•˜๊ธฐ 1๋ถ€
ย 
[๋ชจ๋‘์˜์—ฐ๊ตฌ์†Œ] ์ซ„์ง€๋ง์ž๋”ฅ๋Ÿฌ๋‹
[๋ชจ๋‘์˜์—ฐ๊ตฌ์†Œ] ์ซ„์ง€๋ง์ž๋”ฅ๋Ÿฌ๋‹[๋ชจ๋‘์˜์—ฐ๊ตฌ์†Œ] ์ซ„์ง€๋ง์ž๋”ฅ๋Ÿฌ๋‹
[๋ชจ๋‘์˜์—ฐ๊ตฌ์†Œ] ์ซ„์ง€๋ง์ž๋”ฅ๋Ÿฌ๋‹
ย 
์ธ๊ณต ์‹ ๊ฒฝ๋ง ๊ตฌํ˜„์— ๊ด€ํ•œ ๊ฐ„๋‹จํ•œ ์„ค๋ช…
์ธ๊ณต ์‹ ๊ฒฝ๋ง ๊ตฌํ˜„์— ๊ด€ํ•œ ๊ฐ„๋‹จํ•œ ์„ค๋ช…์ธ๊ณต ์‹ ๊ฒฝ๋ง ๊ตฌํ˜„์— ๊ด€ํ•œ ๊ฐ„๋‹จํ•œ ์„ค๋ช…
์ธ๊ณต ์‹ ๊ฒฝ๋ง ๊ตฌํ˜„์— ๊ด€ํ•œ ๊ฐ„๋‹จํ•œ ์„ค๋ช…
ย 
R ํ”„๋กœ๊ทธ๋ž˜๋ฐ ๊ธฐ๋ณธ ๋ฌธ๋ฒ•
R ํ”„๋กœ๊ทธ๋ž˜๋ฐ ๊ธฐ๋ณธ ๋ฌธ๋ฒ•R ํ”„๋กœ๊ทธ๋ž˜๋ฐ ๊ธฐ๋ณธ ๋ฌธ๋ฒ•
R ํ”„๋กœ๊ทธ๋ž˜๋ฐ ๊ธฐ๋ณธ ๋ฌธ๋ฒ•
ย 
๋จธ์‹  ๋Ÿฌ๋‹ ์ž…๋ฌธ #1-๋จธ์‹ ๋Ÿฌ๋‹ ์†Œ๊ฐœ์™€ kNN ์†Œ๊ฐœ
๋จธ์‹  ๋Ÿฌ๋‹ ์ž…๋ฌธ #1-๋จธ์‹ ๋Ÿฌ๋‹ ์†Œ๊ฐœ์™€ kNN ์†Œ๊ฐœ๋จธ์‹  ๋Ÿฌ๋‹ ์ž…๋ฌธ #1-๋จธ์‹ ๋Ÿฌ๋‹ ์†Œ๊ฐœ์™€ kNN ์†Œ๊ฐœ
๋จธ์‹  ๋Ÿฌ๋‹ ์ž…๋ฌธ #1-๋จธ์‹ ๋Ÿฌ๋‹ ์†Œ๊ฐœ์™€ kNN ์†Œ๊ฐœ
ย 
Introduction to Deep Learning with TensorFlow
Introduction to Deep Learning with TensorFlowIntroduction to Deep Learning with TensorFlow
Introduction to Deep Learning with TensorFlow
ย 
์ธ๊ณต์ง€๋Šฅ, ๊ธฐ๊ณ„ํ•™์Šต ๊ทธ๋ฆฌ๊ณ  ๋”ฅ๋Ÿฌ๋‹
์ธ๊ณต์ง€๋Šฅ, ๊ธฐ๊ณ„ํ•™์Šต ๊ทธ๋ฆฌ๊ณ  ๋”ฅ๋Ÿฌ๋‹์ธ๊ณต์ง€๋Šฅ, ๊ธฐ๊ณ„ํ•™์Šต ๊ทธ๋ฆฌ๊ณ  ๋”ฅ๋Ÿฌ๋‹
์ธ๊ณต์ง€๋Šฅ, ๊ธฐ๊ณ„ํ•™์Šต ๊ทธ๋ฆฌ๊ณ  ๋”ฅ๋Ÿฌ๋‹
ย 
Large Scale Deep Learning with TensorFlow
Large Scale Deep Learning with TensorFlow Large Scale Deep Learning with TensorFlow
Large Scale Deep Learning with TensorFlow
ย 
Deview deep learning-แ„€แ…ตแ†ทแ„Œแ…ฅแ†ผแ„’แ…ด
Deview deep learning-แ„€แ…ตแ†ทแ„Œแ…ฅแ†ผแ„’แ…ดDeview deep learning-แ„€แ…ตแ†ทแ„Œแ…ฅแ†ผแ„’แ…ด
Deview deep learning-แ„€แ…ตแ†ทแ„Œแ…ฅแ†ผแ„’แ…ด
ย 

Similar to Introduction to Machine Learning and Deep Learning

Lec 1 Ds
Lec 1 DsLec 1 Ds
Lec 1 Ds
Qundeel
ย 
Data Structure
Data StructureData Structure
Data Structure
sheraz1
ย 
Lec 1 Ds
Lec 1 DsLec 1 Ds
Lec 1 Ds
Qundeel
ย 
Machine learning (1)
Machine learning (1)Machine learning (1)
Machine learning (1)
NYversity
ย 

Similar to Introduction to Machine Learning and Deep Learning (20)

Deep learning (Machine learning) tutorial for beginners
Deep learning (Machine learning) tutorial for beginnersDeep learning (Machine learning) tutorial for beginners
Deep learning (Machine learning) tutorial for beginners
ย 
Machine Learning, Financial Engineering and Quantitative Investing
Machine Learning, Financial Engineering and Quantitative InvestingMachine Learning, Financial Engineering and Quantitative Investing
Machine Learning, Financial Engineering and Quantitative Investing
ย 
Introduction to Algorithms
Introduction to AlgorithmsIntroduction to Algorithms
Introduction to Algorithms
ย 
Data Structure: Algorithm and analysis
Data Structure: Algorithm and analysisData Structure: Algorithm and analysis
Data Structure: Algorithm and analysis
ย 
Chapter two
Chapter twoChapter two
Chapter two
ย 
Week 2 - ML models and Linear Regression.pptx
Week 2 - ML models and Linear Regression.pptxWeek 2 - ML models and Linear Regression.pptx
Week 2 - ML models and Linear Regression.pptx
ย 
Pythonic Math
Pythonic MathPythonic Math
Pythonic Math
ย 
Data Structures and Algorithm - Week 5 - AVL Trees
Data Structures and Algorithm - Week 5 - AVL TreesData Structures and Algorithm - Week 5 - AVL Trees
Data Structures and Algorithm - Week 5 - AVL Trees
ย 
Learning to Reconstruct
Learning to ReconstructLearning to Reconstruct
Learning to Reconstruct
ย 
Lec 1 Ds
Lec 1 DsLec 1 Ds
Lec 1 Ds
ย 
Data Structure
Data StructureData Structure
Data Structure
ย 
Lec 1 Ds
Lec 1 DsLec 1 Ds
Lec 1 Ds
ย 
Deep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAI
Deep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAIDeep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAI
Deep Reinforcement Learning Through Policy Optimization, John Schulman, OpenAI
ย 
Machine learning (1)
Machine learning (1)Machine learning (1)
Machine learning (1)
ย 
Scala 3 Is Coming: Martin Odersky Shares What To Know
Scala 3 Is Coming: Martin Odersky Shares What To KnowScala 3 Is Coming: Martin Odersky Shares What To Know
Scala 3 Is Coming: Martin Odersky Shares What To Know
ย 
Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)
ย 
Computer Network Assignment Help
Computer Network Assignment HelpComputer Network Assignment Help
Computer Network Assignment Help
ย 
Introduction to computing Processing and performance.pdf
Introduction to computing Processing and performance.pdfIntroduction to computing Processing and performance.pdf
Introduction to computing Processing and performance.pdf
ย 
Chap 8. Optimization for training deep models
Chap 8. Optimization for training deep modelsChap 8. Optimization for training deep models
Chap 8. Optimization for training deep models
ย 
Abductive commonsense reasoning
Abductive commonsense reasoningAbductive commonsense reasoning
Abductive commonsense reasoning
ย 

More from Terry Taewoong Um

More from Terry Taewoong Um (9)

#44. KAIST์—์„œ "๋Œ€ํ•™ ์œ ์ฃ„"๋ฅผ ์™ธ์น˜๋‹ค: ART Lab์˜ ๋„์ „
#44. KAIST์—์„œ "๋Œ€ํ•™ ์œ ์ฃ„"๋ฅผ ์™ธ์น˜๋‹ค: ART Lab์˜ ๋„์ „#44. KAIST์—์„œ "๋Œ€ํ•™ ์œ ์ฃ„"๋ฅผ ์™ธ์น˜๋‹ค: ART Lab์˜ ๋„์ „
#44. KAIST์—์„œ "๋Œ€ํ•™ ์œ ์ฃ„"๋ฅผ ์™ธ์น˜๋‹ค: ART Lab์˜ ๋„์ „
ย 
A brief introduction to OCR (Optical character recognition)
A brief introduction to OCR (Optical character recognition)A brief introduction to OCR (Optical character recognition)
A brief introduction to OCR (Optical character recognition)
ย 
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...
Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynami...
ย 
์ธ๊ณต์ง€๋Šฅ์˜ ์‚ฌํšŒ์ •์˜์˜ ํŽธ์ด ๋  ์ˆ˜ ์žˆ์„๊นŒ? (์ธ๊ณต์ง€๋Šฅ๊ณผ ๋ฒ•)
์ธ๊ณต์ง€๋Šฅ์˜ ์‚ฌํšŒ์ •์˜์˜ ํŽธ์ด ๋  ์ˆ˜ ์žˆ์„๊นŒ? (์ธ๊ณต์ง€๋Šฅ๊ณผ ๋ฒ•)์ธ๊ณต์ง€๋Šฅ์˜ ์‚ฌํšŒ์ •์˜์˜ ํŽธ์ด ๋  ์ˆ˜ ์žˆ์„๊นŒ? (์ธ๊ณต์ง€๋Šฅ๊ณผ ๋ฒ•)
์ธ๊ณต์ง€๋Šฅ์˜ ์‚ฌํšŒ์ •์˜์˜ ํŽธ์ด ๋  ์ˆ˜ ์žˆ์„๊นŒ? (์ธ๊ณต์ง€๋Šฅ๊ณผ ๋ฒ•)
ย 
Deep Variational Bayes Filters (2017)
Deep Variational Bayes Filters (2017)Deep Variational Bayes Filters (2017)
Deep Variational Bayes Filters (2017)
ย 
On Calibration of Modern Neural Networks (2017)
On Calibration of Modern Neural Networks (2017)On Calibration of Modern Neural Networks (2017)
On Calibration of Modern Neural Networks (2017)
ย 
Deep Learning: A Critical Appraisal (2018)
Deep Learning: A Critical Appraisal (2018)Deep Learning: A Critical Appraisal (2018)
Deep Learning: A Critical Appraisal (2018)
ย 
About Two Motion Planning Papers
About Two Motion Planning PapersAbout Two Motion Planning Papers
About Two Motion Planning Papers
ย 
๋กœ๋ด‡๊ณผ ์ธ๊ณต์ง€๋Šฅ, ๊ทธ๋ฆฌ๊ณ  ๋ฏธ๋ž˜์˜ ๋…ธ๋™
๋กœ๋ด‡๊ณผ ์ธ๊ณต์ง€๋Šฅ, ๊ทธ๋ฆฌ๊ณ  ๋ฏธ๋ž˜์˜ ๋…ธ๋™๋กœ๋ด‡๊ณผ ์ธ๊ณต์ง€๋Šฅ, ๊ทธ๋ฆฌ๊ณ  ๋ฏธ๋ž˜์˜ ๋…ธ๋™
๋กœ๋ด‡๊ณผ ์ธ๊ณต์ง€๋Šฅ, ๊ทธ๋ฆฌ๊ณ  ๋ฏธ๋ž˜์˜ ๋…ธ๋™
ย 

Recently uploaded

VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
dharasingh5698
ย 
Call Girls in Netaji Nagar, Delhi ๐Ÿ’ฏ Call Us ๐Ÿ”9953056974 ๐Ÿ” Escort Service
Call Girls in Netaji Nagar, Delhi ๐Ÿ’ฏ Call Us ๐Ÿ”9953056974 ๐Ÿ” Escort ServiceCall Girls in Netaji Nagar, Delhi ๐Ÿ’ฏ Call Us ๐Ÿ”9953056974 ๐Ÿ” Escort Service
Call Girls in Netaji Nagar, Delhi ๐Ÿ’ฏ Call Us ๐Ÿ”9953056974 ๐Ÿ” Escort Service
9953056974 Low Rate Call Girls In Saket, Delhi NCR
ย 
Integrated Test Rig For HTFE-25 - Neometrix
Integrated Test Rig For HTFE-25 - NeometrixIntegrated Test Rig For HTFE-25 - Neometrix
Integrated Test Rig For HTFE-25 - Neometrix
Neometrix_Engineering_Pvt_Ltd
ย 
notes on Evolution Of Analytic Scalability.ppt
notes on Evolution Of Analytic Scalability.pptnotes on Evolution Of Analytic Scalability.ppt
notes on Evolution Of Analytic Scalability.ppt
MsecMca
ย 

Recently uploaded (20)

Unit 1 - Soil Classification and Compaction.pdf
Unit 1 - Soil Classification and Compaction.pdfUnit 1 - Soil Classification and Compaction.pdf
Unit 1 - Soil Classification and Compaction.pdf
ย 
Employee leave management system project.
Employee leave management system project.Employee leave management system project.
Employee leave management system project.
ย 
KubeKraft presentation @CloudNativeHooghly
KubeKraft presentation @CloudNativeHooghlyKubeKraft presentation @CloudNativeHooghly
KubeKraft presentation @CloudNativeHooghly
ย 
(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7
(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7
(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7
ย 
Work-Permit-Receiver-in-Saudi-Aramco.pptx
Work-Permit-Receiver-in-Saudi-Aramco.pptxWork-Permit-Receiver-in-Saudi-Aramco.pptx
Work-Permit-Receiver-in-Saudi-Aramco.pptx
ย 
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
ย 
VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 BookingVIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
VIP Call Girls Palanpur 7001035870 Whatsapp Number, 24/07 Booking
ย 
Generative AI or GenAI technology based PPT
Generative AI or GenAI technology based PPTGenerative AI or GenAI technology based PPT
Generative AI or GenAI technology based PPT
ย 
22-prompt engineering noted slide shown.pdf
22-prompt engineering noted slide shown.pdf22-prompt engineering noted slide shown.pdf
22-prompt engineering noted slide shown.pdf
ย 
DC MACHINE-Motoring and generation, Armature circuit equation
DC MACHINE-Motoring and generation, Armature circuit equationDC MACHINE-Motoring and generation, Armature circuit equation
DC MACHINE-Motoring and generation, Armature circuit equation
ย 
Call Girls in Netaji Nagar, Delhi ๐Ÿ’ฏ Call Us ๐Ÿ”9953056974 ๐Ÿ” Escort Service
Call Girls in Netaji Nagar, Delhi ๐Ÿ’ฏ Call Us ๐Ÿ”9953056974 ๐Ÿ” Escort ServiceCall Girls in Netaji Nagar, Delhi ๐Ÿ’ฏ Call Us ๐Ÿ”9953056974 ๐Ÿ” Escort Service
Call Girls in Netaji Nagar, Delhi ๐Ÿ’ฏ Call Us ๐Ÿ”9953056974 ๐Ÿ” Escort Service
ย 
Integrated Test Rig For HTFE-25 - Neometrix
Integrated Test Rig For HTFE-25 - NeometrixIntegrated Test Rig For HTFE-25 - Neometrix
Integrated Test Rig For HTFE-25 - Neometrix
ย 
Unleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leapUnleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leap
ย 
notes on Evolution Of Analytic Scalability.ppt
notes on Evolution Of Analytic Scalability.pptnotes on Evolution Of Analytic Scalability.ppt
notes on Evolution Of Analytic Scalability.ppt
ย 
A Study of Urban Area Plan for Pabna Municipality
A Study of Urban Area Plan for Pabna MunicipalityA Study of Urban Area Plan for Pabna Municipality
A Study of Urban Area Plan for Pabna Municipality
ย 
chapter 5.pptx: drainage and irrigation engineering
chapter 5.pptx: drainage and irrigation engineeringchapter 5.pptx: drainage and irrigation engineering
chapter 5.pptx: drainage and irrigation engineering
ย 
Introduction to Serverless with AWS Lambda
Introduction to Serverless with AWS LambdaIntroduction to Serverless with AWS Lambda
Introduction to Serverless with AWS Lambda
ย 
Double Revolving field theory-how the rotor develops torque
Double Revolving field theory-how the rotor develops torqueDouble Revolving field theory-how the rotor develops torque
Double Revolving field theory-how the rotor develops torque
ย 
(INDIRA) Call Girl Meerut Call Now 8617697112 Meerut Escorts 24x7
(INDIRA) Call Girl Meerut Call Now 8617697112 Meerut Escorts 24x7(INDIRA) Call Girl Meerut Call Now 8617697112 Meerut Escorts 24x7
(INDIRA) Call Girl Meerut Call Now 8617697112 Meerut Escorts 24x7
ย 
Block diagram reduction techniques in control systems.ppt
Block diagram reduction techniques in control systems.pptBlock diagram reduction techniques in control systems.ppt
Block diagram reduction techniques in control systems.ppt
ย 

Introduction to Machine Learning and Deep Learning

  • 1. Terry Taewoong Um (terry.t.um@gmail.com) University of Waterloo Department of Electrical & Computer Engineering Terry Taewoong Um MACHINE LEARNING, DEEP LEARNING, AND MOTION ANALYSIS 1
  • 2. Terry Taewoong Um (terry.t.um@gmail.com) CAUTION โ€ข I cannot explain everything โ€ข You cannot get every details 2 โ€ข Try to get a big picture โ€ข Get some useful keywords โ€ข Connect with your research
  • 3. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 1. What is Machine Learning? (Part 1 Q & A) 2. What is Deep Learning? (Part 2 Q & A) 3. Machine Learning in Motion Analysis (Part 3 Q & A) 3
  • 4. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 4 1. What is Machine Learning?
  • 5. Terry Taewoong Um (terry.t.um@gmail.com) WHAT IS MACHINE LEARNING? "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience Eโ€œ โ€“ T. Michell (1997) Example: A program for soccer tactics 5 T : Win the game P : Goals E : (x) Playersโ€™ movements (y) Evaluation
  • 6. Terry Taewoong Um (terry.t.um@gmail.com) WHAT IS MACHINE LEARNING? 6 โ€œToward learning robot table tennisโ€, J. Peters et al. (2012) https://youtu.be/SH3bADiB7uQ "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience Eโ€œ โ€“ T. Michell (1997)
  • 7. Terry Taewoong Um (terry.t.um@gmail.com) TASKS 7 classification discrete target values x : pixels (28*28) y : 0,1, 2,3,โ€ฆ,9 regression real target values x โˆˆ (0,100) y : 0,1, 2,3,โ€ฆ,9 clustering no target values x โˆˆ (-3,3)ร—(-3,3) "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience Eโ€œ โ€“ T. Michell (1997)
  • 8. Terry Taewoong Um (terry.t.um@gmail.com) PERFORMANCE 8 "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience Eโ€œ โ€“ T. Michell (1997) classification 0-1 loss function regression L2 loss function clustering
  • 9. Terry Taewoong Um (terry.t.um@gmail.com) EXPERIENCE 9 "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience Eโ€œ โ€“ T. Michell (1997) classification labeled data (pixels)โ†’(number) regression labeled data (x) โ†’ (y) clustering unlabeled data (x1,x2)
  • 10. Terry Taewoong Um (terry.t.um@gmail.com) A TOY EXAMPLE 10 ? Height(cm) Weight (kg) [Input X] [Output Y]
  • 11. Terry Taewoong Um (terry.t.um@gmail.com) 11 180 Height(cm) Weight (kg) 80 Y = aX+b Model : Y = aX+b Parameter : (a, b) [Goal] Find (a,b) which best fits the given data A TOY EXAMPLE
  • 12. Terry Taewoong Um (terry.t.um@gmail.com) 12 [Analytic Solution] Least square problem (from AX = b, X=A#b where A# is Aโ€™s pseudo inverse) Not always available [Numerical Solution] 1. Set a cost function 2. Apply an optimization method (e.g. Gradient Descent (GD) Method) L (a,b) http://www.yaldex.com/game- development/1592730043_ch18lev1sec4.html Local minima problem http://mnemstudio.org/neural-networks- multilayer-perceptron-design.htm A TOY EXAMPLE
  • 13. Terry Taewoong Um (terry.t.um@gmail.com) 13 32 Age(year) Running Record (min) 140 WHAT WOULD BE THE CORRECT MODEL? Select a model โ†’ Set a cost function โ†’ Optimization
  • 14. Terry Taewoong Um (terry.t.um@gmail.com) 14 ? X Y WHAT WOULD BE THE CORRECT MODEL? 1. Regularization 2. Nonparametric model โ€œoverfittingโ€
  • 15. Terry Taewoong Um (terry.t.um@gmail.com) 15 L2 REGULARIZATION (e.g. w=(a,b) where Y=aX+b) Avoid a complicated model! โ€ข Another interpretation : : Maximum a Posteriori (MAP) http://goo.gl/6GE2ix http://goo.gl/6GE2ix
  • 16. Terry Taewoong Um (terry.t.um@gmail.com) 16 L2 REGULARIZATION โ€ข Another interpretation : : Maximum a Posteriori (MAP) http://goo.gl/6GE2ix http://goo.gl/6GE2ix โ€ข Bayesian inference ๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ ๐ท๐‘Ž๐‘ก๐‘Ž = ๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ ๐‘ƒ(๐ท๐‘Ž๐‘ก๐‘Ž|๐ต๐‘’๐‘™๐‘–๐‘’๐‘“) ๐‘ƒ(๐ท๐‘Ž๐‘ก๐‘Ž) posterior prior likelihood ex) fair coin : 50% H, 50% T falsified coin : 80% H, 20% T Letโ€™s say we observed ten heads consecutively. Whatโ€™s the probability for being a fair coin? ๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ = 0.2 ๐‘ƒ ๐ท๐‘Ž๐‘ก๐‘Ž|๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ = 0.510 โ‰ˆ 0.001 ๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“|๐ท๐‘Ž๐‘ก๐‘Ž โˆ 0.2 โˆ— 0.001 = 0.0002 normalization (you donโ€™t believe this coin is fair) Fair coin? Falsified coin? ๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ = 0.8 ๐‘ƒ ๐ท๐‘Ž๐‘ก๐‘Ž|๐ต๐‘’๐‘™๐‘–๐‘’๐‘“ = 0.810 โ‰ˆ 0.107 ๐‘ƒ ๐ต๐‘’๐‘™๐‘–๐‘’๐‘“|๐ท๐‘Ž๐‘ก๐‘Ž โˆ 0.8 โˆ— 0.107 = 0.0856 Fair = 0.0002 0.0002+0.0856 = 0.23% , Unfair = 99.77%
  • 17. Terry Taewoong Um (terry.t.um@gmail.com) 17 WHAT WOULD BE THE CORRECT MODEL? 1. Regularization 2. Nonparametric model training time error training error test error we should stop here training set validation set test set for training (parameter optimization) for early stopping (avoid overfitting) for evaluation (measure the performance) keep watching the validation error
  • 18. Terry Taewoong Um (terry.t.um@gmail.com) 18 NONPARAMETRIC MODEL โ€ข It does not assume any parametric models (e.g. Y = aX+b, Y=aX2+bX+c, etc.) โ€ข It often requires much more samples โ€ข Kernel methods are frequently applied for modeling the data โ€ข Gaussian Process Regression (GPR), a sort of kernel method, is a widely-used nonparametric regression method โ€ข Support Vector Machine (SVM), also a sort of kernel method, is a widely-used nonparametric classification method kernel function [Input space] [Feature space]
  • 19. Terry Taewoong Um (terry.t.um@gmail.com) 19 SUPPORT VECTOR MACHINE (SVM) โ€œMyoโ€, Thalmic Labs (2013) https://youtu.be/oWu9TFJjHaM [Linear classifiers] [Maximum margin] Support vector Machine Tutorial, J. Weston, http://goo.gl/19ywcj [Dual formulation] ( ) kernel function kernel function
  • 20. Terry Taewoong Um (terry.t.um@gmail.com) 20 GAUSSIAN PROCESS REGRESSION (GPR) https://youtu.be/YqhLnCm0KXY https://youtu.be/kvPmArtVoFE โ€ข Gaussian Distribution โ€ข Multivariate regression likelihood posterior prior likelihood prediction conditioning the joint distribution of the observed & predicted values https://goo.gl/EO54WN http://goo.gl/XvOOmf
  • 21. Terry Taewoong Um (terry.t.um@gmail.com) 21 DIMENSION REDUCTION [Original space] [Feature space] low dim. high dim. high dim. low dim. ๐‘‹ โ†’ โˆ…(๐‘‹) โ€ข Principal Component Analysis : Find the best orthogonal axes (=principal components) which maximize the variance of the data Y = P X * The rows in P are m largest eigenvectors of 1 ๐‘ ๐‘‹๐‘‹ ๐‘‡ (covariance matrix)
  • 22. Terry Taewoong Um (terry.t.um@gmail.com) 22 DIMENSION REDUCTION http://jbhuang0604.blogspot.kr/2013/04/miss-korea-2013-contestants-face.html
  • 23. Terry Taewoong Um (terry.t.um@gmail.com) 23 SUMMARY - PART 1 โ€ข Machine Learning - Tasks : Classification, Regression, Clustering, etc. - Performance : 0-1 loss, L2 loss, etc. - Experience : labeled data, unlabelled data โ€ข Machine Learning Process (1) Select a parametric / nonparametric model (2) Set a performance measurement including regularization term (3) Training data (optimizing parameters) until validation error increases (4) Evaluate the final performance using test set โ€ข Nonparametric model : Support Vector Machine, Gaussian Process Regression โ€ข Dimension reduction : used as pre-processing data
  • 24. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 24 Questions about Part 1?
  • 25. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 25 2. What is Deep Learning?
  • 26. Terry Taewoong Um (terry.t.um@gmail.com) 26 PARADIGM CHANGE PAST Knowledge ML Method (e.g. GPR, SVM) PRESENT What is the best ML method for the target task? Knowledge Representation How can we find a good representation?
  • 27. Terry Taewoong Um (terry.t.um@gmail.com) 27 PARADIGM CHANGE Knowledge PRESENT Representation How can we find a good representation? kernel function
  • 28. Terry Taewoong Um (terry.t.um@gmail.com) 28 PARADIGM CHANGE Knowledge PRESENT Representation (Features) How can we find a good representation? IMAGE SPEECH Hand-Crafted Features
  • 29. Terry Taewoong Um (terry.t.um@gmail.com) 29 PARADIGM CHANGE IMAGE SPEECH Hand-Crafted Features Knowledge PRESENT Representation (Features) Can we learn a good representation (feature) for the target task as well?
  • 30. Terry Taewoong Um (terry.t.um@gmail.com) 30 DEEP LEARNING โ€ข What is Deep Learning (DL) ? - Learning methods which have deep (not shallow) architecture - It often allows end-to-end learning - It automatically finds intermediate representation. Thus, it can be regarded as a representation learning - It often contains stacked โ€œneural networkโ€. Thus, Deep learning usually indicates โ€œdeep neural networkโ€ โ€œDeep Gaussian Processโ€ (2013) https://youtu.be/NwoGqYsQifg http://goo.gl/fxmmPE http://goo.gl/5Ry08S
  • 31. Terry Taewoong Um (terry.t.um@gmail.com) 31 OUTSTANDING PERFORMANCE OF DL error rate : 28% โ†’ 15% โ†’ 8% (2010) (2014)(2012) - Object recognition (Simonyan et al., 2015) - Natural machine translation (Bahdanau et al., 2014) - Speech recognition (Chorowski et al., 2014) - Face recognition (Taigman et al., 2014) - Emotion recognition (Ebrahimi-Kahou et al., 2014) - Human pose estimation (Jain et al., 2014) - Deep reinforcement learning(mnih et al., 2013) - Image/Video caption (Xu et al., 2015) - Particle physics (Baldi et al., 2014) - Bioinformatics (Leung et al., 2014) - And so onโ€ฆ. โ€ข State-of-art results achieved by DL DL has won most of ML challenges! K. Cho, https://goo.gl/vdfGpu
  • 32. Terry Taewoong Um (terry.t.um@gmail.com) 32 BIOLOGICAL EVIDENCE โ€ข Somatosensory cortex learns to see โ€ข Why do we need different ML methods for different task? Yann LeCun, https://goo.gl/VVQXJG โ€ข The vental pathway in the visual cortex has multiple stages โ€ข There exist a lot of intermediate representations Andrew Ng, https://youtu.be/ZmNOAtZIgIk
  • 33. Terry Taewoong Um (terry.t.um@gmail.com) 33 BIG MOVEMENT http://goo.gl/zNbBE2 http://goo.gl/Lk64Q4 Going deeper and deeperโ€ฆ.
  • 34. Terry Taewoong Um (terry.t.um@gmail.com) 34 NEURAL NETWORK (NN) Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html โ€ข Universal approximation theorem (Hornik, 1991) - A single hidden layer NN w/ linear output can approximate any cont. func. arbitrarily well, given enough hidden units - This does not imply we have learning method to train them
  • 35. Terry Taewoong Um (terry.t.um@gmail.com) 35 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html โ€ข First, calculate the output using data & initial parameters (W ,b) โ€ข Activation functions http://goo.gl/qMQk5H 1
  • 36. Terry Taewoong Um (terry.t.um@gmail.com) 36 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html โ€ข Then, calculate the error and update the weights from top to bottom โ€ข Parameter gradients http://goo.gl/qMQk5H : Backpropagation algorithm 2 known
  • 37. Terry Taewoong Um (terry.t.um@gmail.com) 37 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html โ€ข Then, calculate the error and update the weights from top to bottom โ€ข Parameter gradients http://goo.gl/qMQk5H : Backpropagation algorithm 2 known
  • 38. Terry Taewoong Um (terry.t.um@gmail.com) 38 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html โ€ข Then, calculate the error and update the weights from top to bottom โ€ข Parameter gradients http://goo.gl/qMQk5H : Backpropagation algorithm 2 known
  • 39. Terry Taewoong Um (terry.t.um@gmail.com) 39 TRAINING NN Hugo Larochelle, http://www.dmi.usherb.ca/~larocheh/index_en.html โ€ข Then, calculate the error and update the weights from top to bottom โ€ข Parameter gradients http://goo.gl/qMQk5H : Backpropagation algorithm 2 known
  • 40. Terry Taewoong Um (terry.t.um@gmail.com) 40 TRAINING NN โ€ข Repeat this process with different dataset(mini-batches) http://goo.gl/qMQk5H - Forward propagation (calculate the output values) - Evaluate the error - Backward propagation (update the weights) - Repeat this process until the error converges 3 โ€ข As you can see here, NN is not a fancy algorithm, but just a iterative gradient descent method with huge number of parameters โ€ข NN is often likely to be stuck in local minima pitfall
  • 41. Terry Taewoong Um (terry.t.um@gmail.com) 41 FROM NN TO DEEP NN โ€ข From NN to deep NN (since 2006) - NN requires expertโ€™s skill to tune the hyperparameters - It sometimes gives a good result, but sometimes gives a bad result. The result is highly depend on the quality of initialization, regularization, hyperparameters, data, etc. - Local minima is always problematic โ€ข A long winter of NN Yann LeCun (NYU, Facebook) Yoshua Bengio (U. Montreal) Geoffrey Hinton (U. Toronto, Google)
  • 42. Terry Taewoong Um (terry.t.um@gmail.com) 42 WHY IS DL SO SUCCESSFUL? http://t-robotics.blogspot.kr/2015/05/deep-learning.html โ€ข Pre-training with unsupervised learning โ€ข Convolutional Neural Network โ€ข Recurrent Neural Net โ€ข GPGPU (parallel processing) & big data โ€ข Advanced algorithms for optimization, activation, regularization โ€ข Huge research society (Vision, Speech, NLP, Biology, etc.)
  • 43. Terry Taewoong Um (terry.t.um@gmail.com) 43 UNSUPERVISED LEARNING โ€ข How can we avoid pathologic local minima cases? (1) First, pre-train the data with unsupervised learning method and get a new representation (2) Stack up this block structures (3) Training each layer in end-to-end manner (4) Fine tune the final structure with (ordinary) fully-connected NN โ€ข Unsupervised learning method - Restricted Boltzmann Machine (RBM) โ†’ Deep RBM, Deep Belief Network (DBN) - Autoencoder โ†’ Deep Auto-encoder http://goo.gl/QGJm5k Autoencoder http://goo.gl/s6kmqY
  • 44. Terry Taewoong Um (terry.t.um@gmail.com) 44 UNSUPERVISED LEARNING โ€œConvolutional deep belief networks for scalable unsupervised learning of hierarchical representationโ€, Lee et al., 2012
  • 45. Terry Taewoong Um (terry.t.um@gmail.com) 45 CONVOLUTIONAL NN โ€ข How can we deal with real images which is much bigger than MNIST digit images? - Use not fully-connected, but locally-connected NN - Use convolutions to get various feature maps - Abstract the results into higher layer by using pooling - Fine tune with fully-connected NN https://goo.gl/G7kBjI https://goo.gl/Xswsbd http://goo.gl/5OR5oH
  • 46. Terry Taewoong Um (terry.t.um@gmail.com) 46 CONVOLUTIONAL NN โ€œVisualization and Understanding Convolutional Networkโ€, Zeiler et al., 2012
  • 47. Terry Taewoong Um (terry.t.um@gmail.com) 47 CONVNET + RNN โ€œLarge-scale Video Classification with Convolutional Neural Networkโ€, A. Karpathy 2014, https://youtu.be/qrzQ_AB1DZk
  • 48. Terry Taewoong Um (terry.t.um@gmail.com) 48 RECURRENT NEURAL NETWORK (RNN) t-1 t t+1 [Neural Network] [Recurrent Neural Network] http://www.dmi.usherb.ca/~larocheh/index_en.html
  • 49. Terry Taewoong Um (terry.t.um@gmail.com) 49 RECURRENT NEURAL NETWORK (RNN) [Neural Network] [Recurrent Neural Network] back propagation back propagation through time (BPTT) โ€ข Vanishing gradient problem : Canโ€™t have long memory! โ€œTraining Recurrent Neural Networks, I. Sutskever, 2013
  • 50. Terry Taewoong Um (terry.t.um@gmail.com) 50 RNN + LSTM โ€ข Long Short-Term Memory (LSTM) (Hochreiter and Schmidhuber, 1997) โ€œTraining Recurrent Neural Networks, I. Sutskever, 2013
  • 51. Terry Taewoong Um (terry.t.um@gmail.com) 51 INTERESTING RESULTS FROM RNN http://pail.unist.ac.kr/carpedm20/poet/ http://cs.stanford.edu/people/karpathy/deepimagesent/ โ€œgenerating sequences with RNNโ€, A.Graves, 2013
  • 52. Terry Taewoong Um (terry.t.um@gmail.com) 52 WHY IS DL SO SUCCESSFUL? http://t-robotics.blogspot.kr/2015/05/deep-learning.html โ€ข Pre-training with unsupervised learning โ€ข Convolutional Neural Network โ€ข Recurrent Neural Net โ€ข GPGPU (parallel processing) & big data โ€ข Advanced algorithms for optimization, activation, regularization โ€ข Huge research society (Vision, Speech, NLP, Biology, etc.)
  • 53. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 53 Questions about Part 2?
  • 54. Terry Taewoong Um (terry.t.um@gmail.com) CONTENTS 54 3. Machine Learning in Motion Analysis
  • 55. Terry Taewoong Um (terry.t.um@gmail.com) 55 MOTION DATA โ€œ์ธ„๋ฆฌ๋‹โ€, ์ด์ƒ์‹  ๊ตญ์ค‘๋ก
  • 56. Terry Taewoong Um (terry.t.um@gmail.com) 56 MOTION DATA We need to know the state not only at time t but also at time t-1, t-2, t-3, etc. ๐‘“ = ๐‘“(๐‘ฅ, ๐‘ก) โ€œ์ธ„๋ฆฌ๋‹โ€, ์ด์ƒ์‹  ๊ตญ์ค‘๋ก
  • 57. Terry Taewoong Um (terry.t.um@gmail.com) 57 MOTION DATA โ€ข Why do motion data need special treatment? - In general, most machine learning techniques assume i.i.d. (independent & identically distributed) sampling condition. e.g.) coins tossing - However, motion data is temporally & spatially correlated http://goo.gl/LQulvcswing motion manipulability ellipsoid https://goo.gl/dHjFO9
  • 58. Terry Taewoong Um (terry.t.um@gmail.com) 58 MOTION DATA http://goo.gl/ll3sq6 We can infer the next state based on the temporal & spatial information But, how can we exploit those benefits in ML method?
  • 59. Terry Taewoong Um (terry.t.um@gmail.com) 59 WHAT CAN WE DO WITH MOTION DATA? โ€ข Learning the kinematic/dynamic model โ€ข Motion segmentation โ€ข Motion generation / synthesis โ€ข Motion imitation (Imitation learning) โ€ข Activity / Gesture recognition TASKS Data โ€ข Motion capture data โ€ข Vision Data โ€ข Dynamic-level data Applications โ€ข Biomechanics โ€ข Humanoid โ€ข Animation http://goo.gl/gFOVWL
  • 60. Terry Taewoong Um (terry.t.um@gmail.com) 60 HIDDEN MARKOV MODEL (HMM) Prob. of (n+1) state only depends on state at (n+1)
  • 61. Terry Taewoong Um (terry.t.um@gmail.com) 61 LIMITATIONS OF HMM 1. Extract features (e.g. PCA) 2. Define the HMM structure (e.g. using GMM) 3. Train a separate HMM per class (Baum-Welch algorithm) 4. Evaluate probability under each HMM (Fwd/Bwd algorithm) or 3. Choose most probable sequence (Viterbi algorithm) - HMM handle discrete states only! - HMM has short memory! (using just the previous state) - HMM has limited expressive power! - [Trend1] features-GMM โ†’ unsupervised learning methods - [Trend2] features-GMM-HMM โ†’ recurrent neural network โ€ข A common procedure of HMM for motion analysis โ€ข Limitations & trend change in speech recognition area
  • 62. Terry Taewoong Um (terry.t.um@gmail.com) 62 CAPTURE TEMPORAL INFORMATION โ€ข 3D ConvNet - โ€œ3D Convolutional Neural Network for Human Action Recognitionโ€ (Ji et al., 2010) - 3D convolution - Activity recognition / Pose estimation from video โ€œJoint Training of a Convolutional Network and a Graphical Model for Human Pose Estimationโ€, Tompson et al., 2014
  • 63. Terry Taewoong Um (terry.t.um@gmail.com) 63 CAPTURE TEMPORAL INFORMATION โ€ข Recurrent Neural Network (RNN) โ€œHierarchical Recurrent Neural Network for Skeleton Based Action Recognitionโ€, Y. Du et al., 2015 โ€ข However, how can we capture the spatial information about motions?
  • 64. Terry Taewoong Um (terry.t.um@gmail.com) 64 CHALLENGES We should connect the geometric information with deep neural network! โ€ข The link transformation from the i-1 th link to the i th link โ€ข Forward Kinematics constant, Mvariable, ๐œƒ c.f.) ๐‘‹๐‘–โˆ’1,๐‘– = ๐‘…๐‘œ๐‘ก ๐‘ง, ๐œƒ๐‘– ๐‘‡๐‘Ÿ๐‘Ž๐‘›๐‘  ๐‘ง, ๐‘‘๐‘– ๐‘‡๐‘Ÿ๐‘Ž๐‘›๐‘  ๐‘ฅ, ๐‘Ž๐‘– ๐‘…๐‘œ๐‘ก ๐‘ง, ๐›ผ๐‘– = ๐‘’[๐ด ๐‘–]๐œƒ ๐‘– ๐‘€๐‘–โˆ’1,๐‘– ๐‘‹0,๐‘› = ๐‘’[๐ด1]๐œƒ1 ๐‘€0,1 ๐‘’[๐ด2]๐œƒ2 ๐‘€1,2 โ‹ฏ ๐‘’ ๐ด ๐‘› ๐œƒ ๐‘› ๐‘€ ๐‘›โˆ’1,๐‘› = ๐‘’[๐‘†1]๐œƒ1 ๐‘’[๐‘†2]๐œƒ2 โ‹ฏ ๐‘’[๐‘† ๐‘›]๐œƒ ๐‘› ๐‘€0,๐‘› ๐‘†๐‘– = ๐ด๐‘‘ ๐‘€01โ‹ฏ๐‘€ ๐‘–โˆ’2,๐‘–โˆ’1 ๐ด๐‘– , ๐‘– = 1, โ‹ฏ , ๐‘› propagated forces external force acting on the ith body where โ€ข Newton-Euler formulation for inverse dynamics Lie group & Lie algebra, http://goo.gl/uqilDV
  • 65. Terry Taewoong Um (terry.t.um@gmail.com) 65 CHALLENGES https://www.youtube.com/watch?v=oxA2O-tHftI
  • 66. Terry Taewoong Um (terry.t.um@gmail.com) 66 Thank you