SlideShare uma empresa Scribd logo
1 de 26
Baixar para ler offline
.
.
. ..
.
.
Introduction to Blind Source Separation and Non-negative
Matrix Factorization
Tatsuya Yokota
Tokyo Institute of Technology
November 29, 2011
November 29, 2011 1/26
Outline
.
..1 Blind Source Separation
.
..2 Non-negative Matrix Factrization
.
..3 Experiments
.
..4 Summary
November 29, 2011 2/26
What’s a Blind Source Separation
Blind Source Separation is a method to estimate original signals from observed
signals which consist of mixed original signals and noise.
November 29, 2011 3/26
Example of BSS
BSS is often used for Speech analysis and Image analysis.
November 29, 2011 4/26
Example of BSS (cont’d)
BSS is also very important for brain signal analysis.
November 29, 2011 5/26
Model Formalization
The problem of BSS is formalized as follow:
The matrix
X ∈ Rm×d
(1)
denotes original signals, where m is number of original signals, and d is dimension
of one signal.
We consider that the observed signals Y ∈ Rn×d
are given by linear mixing system
as
Y = AX + E, (2)
where A ∈ Rn×m
is the unknown mixing matrix and E ∈ Rn×d
denotes a noise.
Basically, n ≥ m.
The goal of BSS is to estimate ˆA and ˆX so that ˆX provides unknown original
signal as possible.
November 29, 2011 6/26
Kinds of BSS Methods
Actually, degree of freedom of BSS model is very high to estimate A and X.
Bcause there are a huge number of combinations (A, X) which satisfy
Y = AX + E.
Therefore, we need some constraint to solve the BSS problem such as:
PCA : orthogonal constraint
SCA : sparsity constraint
NMF : non-negativity constraint
ICA : independency constraint
In this way, there are many methods to solve the BSS problem depending on the
constraints. What we use is depend on subject matter.
I will introduce about Non-negative Matrix Factrization(NMF) based
technique for BSS.
November 29, 2011 7/26
Non-negativity
Non-negativity means that value is 0 or positive (i.e. at least 0). Non-negative
matrix is a matrix which its all elements are at least 0. For example
1 10
−10 1
/∈ R2×2
+ ,
1 10
0 1
∈ R2×2
+ . (3)
November 29, 2011 8/26
Why nonnegativity constraints?
This is the key-point of this method.
We can say that because many real-world data are nonnegative and the
corresponding hidden components have a physical meaning only when
nonnegative.
Example of nonnegative data,
image data (luminance)
power spectol of signal (speech, EEG etc.)
In general, observed data can be considered as a linear combination of such
nonnegative data with noise.(i.e. Y = AX + E)
November 29, 2011 9/26
Standard NMF Model
Consider the standard NMF model, given by:
Y = AX + E, (4)
A ≥ 0, X ≥ 0. (5)
For simply, we denote A ≥ 0 as nonnegativity of matrix A.
The NMF problem is given by:
.
NMF problem
..
.
. ..
.
.
minimize
1
2
||Y − AX||2
F (6)
subject to A ≥ 0, X ≥ 0 (7)
I will introduce the alternating least square(ALS) algorithm to solve this
problem.
November 29, 2011 10/26
ALS algorithm
Our goal is to estimate ( ˆA, ˆX) = minA,X
1
2 ||Y − AX||2
F , s.t. A ≥ 0, X ≥ 0.
However, it is very difficult to solve such a multivariate problem.
The ALS strategy is to solve following sub-problems alternately.
.
sub-problem for A
..
.
. ..
.
.
min
1
2
||Y − AX||2
F (8)
s.t. A ≥ 0 (9)
.
sub-problem for X
..
.
. ..
.
.
min
1
2
||Y − AX||2
F (10)
s.t. X ≥ 0 (11)
November 29, 2011 11/26
How to Solve the sub-problem
Solving procedure of sub-problem for A is only two steps as follow:
.
sub-procedure for A
..
.
. ..
.
.
...1 A ← argminA
1
2 ||Y − AX||2
F
...2 A ← [A]+
where [a]+ = max(ε, a) rectify to enforce nonnegativity.
In a similar way, we can solve sub-problem for X as follow:
.
sub-prodecure for X
..
.
. ..
.
.
...1 X ← argminX
1
2 ||Y − AX||2
F
...2 X ← [X]+
November 29, 2011 12/26
Solution of Least Squares
Here, I show basic calculation of least squares. First, objective function can be
transformed as
1
2
||Y − AX||2
F (12)
=
1
2
tr(Y − AX)(Y − AX)T
(13)
=
1
2
tr(Y Y T
− 2AXY T
+ AXXT
AT
) (14)
The stationary points ˆA can be found by equating the gradient components to 0:
∂
∂AT
1
2
tr(Y Y T
− 2AXY T
+ AXXT
AT
) (15)
= −XY T
+ XXT
AT
= 0 (16)
Therefore,
ˆAT
= (XXT
)−1
XY T
. (17)
November 29, 2011 13/26
Solution of Least Squares (cont’d)
In similar way, we can also obtain a solution of X. The objective function can be
transformed as
1
2
||Y − AX||2
F (18)
=
1
2
tr(Y − AX)T
(Y − AX) (19)
=
1
2
tr(Y T
Y − 2XT
AT
Y + XT
AT
AX) (20)
The stationary points ˆX can be found by equating the gradient components to 0:
∂
∂X
1
2
tr(Y T
Y − 2XT
AT
Y + XT
AT
AX) (21)
= −AT
Y + AT
AX = 0 (22)
Therefore,
ˆX = (AT
A)−1
AT
Y. (23)
November 29, 2011 14/26
Implimentation
Implimentation in “octave” is very simple:
i=0;
while 1
i++;
printf("%d: %f n",i,sqrt(sum(sum(E.^2))));
if sqrt(sum(sum(E.^2))) < er || i > maxiter
break;
end
At=(X*X’+ 1e-5*eye(J))(X*Y’);A=At’;
A(A < 0)=1e-16;
X=(A’*A + 1e-5*eye(J))(A’*Y);
X(X < 0)=1e-16;
E=Y-A*X;
end
November 29, 2011 15/26
Experiments: Artificial Data 1
-0.5
0
0.5
1
1.5
0 50 100 150 200
(a) sin2(t)
-0.5
0
0.5
1
1.5
0 50 100 150 200
(b) cos2(t)
Figure: Original Signals
-0.5
0
0.5
1
1.5
0 50 100 150 200
(a) mixed (observed) signal 1
-0.5
0
0.5
1
1.5
0 50 100 150 200
(b) mixed (observed) signal 2
Figure: Observed Signals
-0.5
0
0.5
1
1.5
0 50 100 150 200
(a) estimated signal 1
-0.5
0
0.5
1
1.5
0 50 100 150 200
(b) estimated signal 2
Figure: Estimated Signals
We can see that results are almost separated.
November 29, 2011 16/26
Experiments: Artificial Data 2
-0.5
0
0.5
1
1.5
0 50 100 150 200
(a) sin2(t)
-0.5
0
0.5
1
1.5
0 50 100 150 200
(b) sin2(2t)
Figure: Original Signals
-0.5
0
0.5
1
1.5
0 50 100 150 200
(a) mixed (observed) signal 1
-0.5
0
0.5
1
1.5
0 50 100 150 200
(b) mixed (observed) signal 2
Figure: Observed Signals
-0.5
0
0.5
1
1.5
0 50 100 150 200
(a) estimated signal 1
-0.5
0
0.5
1
1.5
0 50 100 150 200
(b) estimated signal 2
Figure: Estimated Signals
We can see that results are not so separated.
November 29, 2011 17/26
Experiments: Real Image 1
(a) newyork
(b) shanghai
Figure: Original Signals
(a) mixed (observed) signal 1
(b) mixed (observed) signal 2
Figure: Observed Signals
(a) estimated signal 1
(b) estimated signal 2
Figure: Estimated Signals
We can see that results are almost separated.
November 29, 2011 18/26
Experiments: Real Image 2
(a) rock
(b) pig
Figure: Original Signals
(a) mixed (observed) signal 1
(b) mixed (observed) signal 2
Figure: Observed Signals
(a) estimated signal 1
(b) estimated signal 2
Figure: Estimated Signals
We can see that estimate 2 is not so separated.
November 29, 2011 19/26
Experiments: Real Image 3
(a) nyc (b) sha
(c) rock (d) pig
(e) obs1 (f) obs2
(g) obs3 (h) obs4
Figure: Ori. & Obs.
(a) estimated signal 1 (b) estimated signal 2
(c) estimated signal 3 (d) estimated signal 4
Figure: Estimated Signals: Results are not good.
Signals are still mixed.
November 29, 2011 20/26
Issues
There are some issues of this method as follow:
Solution is non-unique (i.e. result is depend on starting point)
It is not enough only by non-negativity constraint (shoud consider
independency of each components)
When number of estimate signals is large, features of components are
overlapped.
And we should consider more general issues of BSS as follow:
Number of original signals is unknown (How can we decide its number?)
November 29, 2011 21/26
About this research area
In this research area, many method for BSS are studied and proposed as follow:
KL-divergence based NMF [Honkela et al., 2011]
Alpha-Beta divergences based NMF [Cichocki et al., 2011]
Non-Gaussianity based ICA [Hyv¨arinen et al., 2001]
Kurtosis based ICA
Negentropy based ICA
Solving method for ICA
gradient method
fast fixed-point algorithm [Hyv¨arinen and Oja, 1997]
MLE based ICA
Mutual information based ICA
Non-linear ICA
Tensor ICA
In this way, this research area is very broad!!
November 29, 2011 22/26
Summary
I introduced about BSS problem and basic NMF tequnique from
[Cichocki et al., 2009].
EEG classification is very difficult problem.
.
Future Work
..
.
. ..
.
.
To study about extension of NMF and Independent Component Analysis
(ICA).
To apply the BSS problem, EEG analysis, and some pattern recognition
problem.
November 29, 2011 23/26
Bibliography I
[Cichocki et al., 2011] Cichocki, A., Cruces, S., and Amari, S.-i. (2011).
Generalized alpha-beta divergences and their application to robust nonnegative
matrix factorization.
Entropy, 13(1):134–170.
[Cichocki et al., 2009] Cichocki, A., Zdunek, R., Phan, A. H., and Amari, S.
(2009).
Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory
Multi-way Data Analysis.
Wiley.
[Honkela et al., 2011] Honkela, T., Duch, W., Girolami, M., and Kaski, S.,
editors (2011).
Kullback-Leibler Divergence for Nonnegative Matrix Factorization, volume
6791 of Lecture Notes in Computer Science. Springer Berlin / Heidelberg.
[Hyv¨arinen et al., 2001] Hyv¨arinen, A., Karhunen, J., and Oja, E. (2001).
Independent Component Analysis.
Wiley.
November 29, 2011 24/26
Bibliography II
[Hyv¨arinen and Oja, 1997] Hyv¨arinen, A. and Oja, E. (1997).
A fast fixed-point algorithm for independent component analysis.
Neural Computation, 9:1483–1492.
November 29, 2011 25/26
Thank you for listening
November 29, 2011 26/26

Mais conteúdo relacionado

Mais procurados

Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Simplilearn
 
Introduction to bayesian_networks[1]
Introduction to bayesian_networks[1]Introduction to bayesian_networks[1]
Introduction to bayesian_networks[1]
JULIO GONZALEZ SANZ
 

Mais procurados (20)

Recurrent Neural Networks
Recurrent Neural NetworksRecurrent Neural Networks
Recurrent Neural Networks
 
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
 
Introduction to bayesian_networks[1]
Introduction to bayesian_networks[1]Introduction to bayesian_networks[1]
Introduction to bayesian_networks[1]
 
Independent Component Analysis
Independent Component AnalysisIndependent Component Analysis
Independent Component Analysis
 
Anomaly detection
Anomaly detectionAnomaly detection
Anomaly detection
 
Csc446: Pattern Recognition
Csc446: Pattern Recognition Csc446: Pattern Recognition
Csc446: Pattern Recognition
 
Vanishing & Exploding Gradients
Vanishing & Exploding GradientsVanishing & Exploding Gradients
Vanishing & Exploding Gradients
 
Feature selection
Feature selectionFeature selection
Feature selection
 
ResNet basics (Deep Residual Network for Image Recognition)
ResNet basics (Deep Residual Network for Image Recognition)ResNet basics (Deep Residual Network for Image Recognition)
ResNet basics (Deep Residual Network for Image Recognition)
 
Predictive coding
Predictive codingPredictive coding
Predictive coding
 
Convolutional Neural Networks
Convolutional Neural NetworksConvolutional Neural Networks
Convolutional Neural Networks
 
Fuzzy Clustering(C-means, K-means)
Fuzzy Clustering(C-means, K-means)Fuzzy Clustering(C-means, K-means)
Fuzzy Clustering(C-means, K-means)
 
Probabilistic Reasoning
Probabilistic ReasoningProbabilistic Reasoning
Probabilistic Reasoning
 
Neural network final NWU 4.3 Graphics Course
Neural network final NWU 4.3 Graphics CourseNeural network final NWU 4.3 Graphics Course
Neural network final NWU 4.3 Graphics Course
 
5.3 mining sequential patterns
5.3 mining sequential patterns5.3 mining sequential patterns
5.3 mining sequential patterns
 
An introduction on normalizing flows
An introduction on normalizing flowsAn introduction on normalizing flows
An introduction on normalizing flows
 
Introduction to Artificial Neural Network
Introduction to Artificial Neural Network Introduction to Artificial Neural Network
Introduction to Artificial Neural Network
 
Recurrent neural networks rnn
Recurrent neural networks   rnnRecurrent neural networks   rnn
Recurrent neural networks rnn
 
Unsupervised visual representation learning overview: Toward Self-Supervision
Unsupervised visual representation learning overview: Toward Self-SupervisionUnsupervised visual representation learning overview: Toward Self-Supervision
Unsupervised visual representation learning overview: Toward Self-Supervision
 
K-Nearest Neighbor Classifier
K-Nearest Neighbor ClassifierK-Nearest Neighbor Classifier
K-Nearest Neighbor Classifier
 

Destaque

Factorization Machines with libFM
Factorization Machines with libFMFactorization Machines with libFM
Factorization Machines with libFM
Liangjie Hong
 
Matrix factorization
Matrix factorizationMatrix factorization
Matrix factorization
rubyyc
 
Lecture 6 lu factorization & determinants - section 2-5 2-7 3-1 and 3-2
Lecture 6   lu factorization & determinants - section 2-5 2-7 3-1 and 3-2Lecture 6   lu factorization & determinants - section 2-5 2-7 3-1 and 3-2
Lecture 6 lu factorization & determinants - section 2-5 2-7 3-1 and 3-2
njit-ronbrown
 
Recommender system introduction
Recommender system   introductionRecommender system   introduction
Recommender system introduction
Liang Xiang
 
Building a Recommendation Engine - An example of a product recommendation engine
Building a Recommendation Engine - An example of a product recommendation engineBuilding a Recommendation Engine - An example of a product recommendation engine
Building a Recommendation Engine - An example of a product recommendation engine
NYC Predictive Analytics
 
Recommender system algorithm and architecture
Recommender system algorithm and architectureRecommender system algorithm and architecture
Recommender system algorithm and architecture
Liang Xiang
 

Destaque (16)

Matrix Factorization Technique for Recommender Systems
Matrix Factorization Technique for Recommender SystemsMatrix Factorization Technique for Recommender Systems
Matrix Factorization Technique for Recommender Systems
 
Factorization Machines with libFM
Factorization Machines with libFMFactorization Machines with libFM
Factorization Machines with libFM
 
Matrix factorization
Matrix factorizationMatrix factorization
Matrix factorization
 
Collaborative Filtering with Spark
Collaborative Filtering with SparkCollaborative Filtering with Spark
Collaborative Filtering with Spark
 
Neighbor methods vs matrix factorization - case studies of real-life recommen...
Neighbor methods vs matrix factorization - case studies of real-life recommen...Neighbor methods vs matrix factorization - case studies of real-life recommen...
Neighbor methods vs matrix factorization - case studies of real-life recommen...
 
Intro to Factorization Machines
Intro to Factorization MachinesIntro to Factorization Machines
Intro to Factorization Machines
 
Lecture 6 lu factorization & determinants - section 2-5 2-7 3-1 and 3-2
Lecture 6   lu factorization & determinants - section 2-5 2-7 3-1 and 3-2Lecture 6   lu factorization & determinants - section 2-5 2-7 3-1 and 3-2
Lecture 6 lu factorization & determinants - section 2-5 2-7 3-1 and 3-2
 
آموزش محاسبات عددی - بخش دوم
آموزش محاسبات عددی - بخش دومآموزش محاسبات عددی - بخش دوم
آموزش محاسبات عددی - بخش دوم
 
Recommender system introduction
Recommender system   introductionRecommender system   introduction
Recommender system introduction
 
Recommender Systems
Recommender SystemsRecommender Systems
Recommender Systems
 
Introduction to Matrix Factorization Methods Collaborative Filtering
Introduction to Matrix Factorization Methods Collaborative FilteringIntroduction to Matrix Factorization Methods Collaborative Filtering
Introduction to Matrix Factorization Methods Collaborative Filtering
 
Recommender Systems
Recommender SystemsRecommender Systems
Recommender Systems
 
Recommendation system
Recommendation system Recommendation system
Recommendation system
 
Collaborative Filtering Recommendation System
Collaborative Filtering Recommendation SystemCollaborative Filtering Recommendation System
Collaborative Filtering Recommendation System
 
Building a Recommendation Engine - An example of a product recommendation engine
Building a Recommendation Engine - An example of a product recommendation engineBuilding a Recommendation Engine - An example of a product recommendation engine
Building a Recommendation Engine - An example of a product recommendation engine
 
Recommender system algorithm and architecture
Recommender system algorithm and architectureRecommender system algorithm and architecture
Recommender system algorithm and architecture
 

Semelhante a Nonnegative Matrix Factorization

2012 mdsp pr10 ica
2012 mdsp pr10 ica2012 mdsp pr10 ica
2012 mdsp pr10 ica
nozomuhamada
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
ijceronline
 

Semelhante a Nonnegative Matrix Factorization (20)

2012 mdsp pr10 ica
2012 mdsp pr10 ica2012 mdsp pr10 ica
2012 mdsp pr10 ica
 
ICAoverview
ICAoverviewICAoverview
ICAoverview
 
Single Channel Speech De-noising Using Kernel Independent Component Analysis...
	Single Channel Speech De-noising Using Kernel Independent Component Analysis...	Single Channel Speech De-noising Using Kernel Independent Component Analysis...
Single Channel Speech De-noising Using Kernel Independent Component Analysis...
 
Section9 stochastic
Section9 stochasticSection9 stochastic
Section9 stochastic
 
Blind Audio Source Separation (Bass): An Unsuperwised Approach
Blind Audio Source Separation (Bass): An Unsuperwised Approach Blind Audio Source Separation (Bass): An Unsuperwised Approach
Blind Audio Source Separation (Bass): An Unsuperwised Approach
 
Independent Component Analysis
Independent Component Analysis Independent Component Analysis
Independent Component Analysis
 
Principal Component Analysis for Tensor Analysis and EEG classification
Principal Component Analysis for Tensor Analysis and EEG classificationPrincipal Component Analysis for Tensor Analysis and EEG classification
Principal Component Analysis for Tensor Analysis and EEG classification
 
MATHEMATICAL EXPLANATION TO SOLUTION FOR EX-NOR PROBLEM USING MLFFN
MATHEMATICAL EXPLANATION TO SOLUTION FOR EX-NOR PROBLEM USING MLFFNMATHEMATICAL EXPLANATION TO SOLUTION FOR EX-NOR PROBLEM USING MLFFN
MATHEMATICAL EXPLANATION TO SOLUTION FOR EX-NOR PROBLEM USING MLFFN
 
Mixed Spectra for Stable Signals from Discrete Observations
Mixed Spectra for Stable Signals from Discrete ObservationsMixed Spectra for Stable Signals from Discrete Observations
Mixed Spectra for Stable Signals from Discrete Observations
 
Mixed Spectra for Stable Signals from Discrete Observations
Mixed Spectra for Stable Signals from Discrete ObservationsMixed Spectra for Stable Signals from Discrete Observations
Mixed Spectra for Stable Signals from Discrete Observations
 
MIXED SPECTRA FOR STABLE SIGNALS FROM DISCRETE OBSERVATIONS
MIXED SPECTRA FOR STABLE SIGNALS FROM DISCRETE OBSERVATIONSMIXED SPECTRA FOR STABLE SIGNALS FROM DISCRETE OBSERVATIONS
MIXED SPECTRA FOR STABLE SIGNALS FROM DISCRETE OBSERVATIONS
 
Mixed Spectra for Stable Signals from Discrete Observations
Mixed Spectra for Stable Signals from Discrete ObservationsMixed Spectra for Stable Signals from Discrete Observations
Mixed Spectra for Stable Signals from Discrete Observations
 
Mixed Spectra for Stable Signals from Discrete Observations
Mixed Spectra for Stable Signals from Discrete ObservationsMixed Spectra for Stable Signals from Discrete Observations
Mixed Spectra for Stable Signals from Discrete Observations
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
 
13005810.ppt
13005810.ppt13005810.ppt
13005810.ppt
 
40120130406002
4012013040600240120130406002
40120130406002
 
DSP_DiscSignals_LinearS_150417.pptx
DSP_DiscSignals_LinearS_150417.pptxDSP_DiscSignals_LinearS_150417.pptx
DSP_DiscSignals_LinearS_150417.pptx
 
StatPhysPerspectives_AMALEA_Cetraro_AnnaCarbone.pdf
StatPhysPerspectives_AMALEA_Cetraro_AnnaCarbone.pdfStatPhysPerspectives_AMALEA_Cetraro_AnnaCarbone.pdf
StatPhysPerspectives_AMALEA_Cetraro_AnnaCarbone.pdf
 
12.pdf
12.pdf12.pdf
12.pdf
 

Mais de Tatsuya Yokota

Linked CP Tensor Decomposition (presented by ICONIP2012)
Linked CP Tensor Decomposition (presented by ICONIP2012)Linked CP Tensor Decomposition (presented by ICONIP2012)
Linked CP Tensor Decomposition (presented by ICONIP2012)
Tatsuya Yokota
 
Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...
Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...
Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...
Tatsuya Yokota
 

Mais de Tatsuya Yokota (9)

Tensor representations in signal processing and machine learning (tutorial ta...
Tensor representations in signal processing and machine learning (tutorial ta...Tensor representations in signal processing and machine learning (tutorial ta...
Tensor representations in signal processing and machine learning (tutorial ta...
 
テンソル多重線形ランクの推定法について(Estimation of Multi-linear Tensor Rank)
テンソル多重線形ランクの推定法について(Estimation of Multi-linear Tensor Rank)テンソル多重線形ランクの推定法について(Estimation of Multi-linear Tensor Rank)
テンソル多重線形ランクの推定法について(Estimation of Multi-linear Tensor Rank)
 
自己相似な情報モデリング
自己相似な情報モデリング自己相似な情報モデリング
自己相似な情報モデリング
 
Priorに基づく画像/テンソルの復元
Priorに基づく画像/テンソルの復元Priorに基づく画像/テンソルの復元
Priorに基づく画像/テンソルの復元
 
低ランク性および平滑性を用いたテンソル補完 (Tensor Completion based on Low-rank and Smooth Structu...
低ランク性および平滑性を用いたテンソル補完 (Tensor Completion based on Low-rank and Smooth Structu...低ランク性および平滑性を用いたテンソル補完 (Tensor Completion based on Low-rank and Smooth Structu...
低ランク性および平滑性を用いたテンソル補完 (Tensor Completion based on Low-rank and Smooth Structu...
 
2013 11 01(fast_grbf-nmf)_for_share
2013 11 01(fast_grbf-nmf)_for_share2013 11 01(fast_grbf-nmf)_for_share
2013 11 01(fast_grbf-nmf)_for_share
 
2014 3 13(テンソル分解の基礎)
2014 3 13(テンソル分解の基礎)2014 3 13(テンソル分解の基礎)
2014 3 13(テンソル分解の基礎)
 
Linked CP Tensor Decomposition (presented by ICONIP2012)
Linked CP Tensor Decomposition (presented by ICONIP2012)Linked CP Tensor Decomposition (presented by ICONIP2012)
Linked CP Tensor Decomposition (presented by ICONIP2012)
 
Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...
Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...
Introduction to Common Spatial Pattern Filters for EEG Motor Imagery Classifi...
 

Último

+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
?#DUbAI#??##{{(☎️+971_581248768%)**%*]'#abortion pills for sale in dubai@
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Safe Software
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
panagenda
 

Último (20)

Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
presentation ICT roal in 21st century education
presentation ICT roal in 21st century educationpresentation ICT roal in 21st century education
presentation ICT roal in 21st century education
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
Apidays Singapore 2024 - Building Digital Trust in a Digital Economy by Veron...
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024Axa Assurance Maroc - Insurer Innovation Award 2024
Axa Assurance Maroc - Insurer Innovation Award 2024
 
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
Apidays New York 2024 - The Good, the Bad and the Governed by David O'Neill, ...
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
Bajaj Allianz Life Insurance Company - Insurer Innovation Award 2024
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
Top 10 Most Downloaded Games on Play Store in 2024
Top 10 Most Downloaded Games on Play Store in 2024Top 10 Most Downloaded Games on Play Store in 2024
Top 10 Most Downloaded Games on Play Store in 2024
 
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers:  A Deep Dive into Serverless Spatial Data and FMECloud Frontiers:  A Deep Dive into Serverless Spatial Data and FME
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FME
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024Tata AIG General Insurance Company - Insurer Innovation Award 2024
Tata AIG General Insurance Company - Insurer Innovation Award 2024
 

Nonnegative Matrix Factorization

  • 1. . . . .. . . Introduction to Blind Source Separation and Non-negative Matrix Factorization Tatsuya Yokota Tokyo Institute of Technology November 29, 2011 November 29, 2011 1/26
  • 2. Outline . ..1 Blind Source Separation . ..2 Non-negative Matrix Factrization . ..3 Experiments . ..4 Summary November 29, 2011 2/26
  • 3. What’s a Blind Source Separation Blind Source Separation is a method to estimate original signals from observed signals which consist of mixed original signals and noise. November 29, 2011 3/26
  • 4. Example of BSS BSS is often used for Speech analysis and Image analysis. November 29, 2011 4/26
  • 5. Example of BSS (cont’d) BSS is also very important for brain signal analysis. November 29, 2011 5/26
  • 6. Model Formalization The problem of BSS is formalized as follow: The matrix X ∈ Rm×d (1) denotes original signals, where m is number of original signals, and d is dimension of one signal. We consider that the observed signals Y ∈ Rn×d are given by linear mixing system as Y = AX + E, (2) where A ∈ Rn×m is the unknown mixing matrix and E ∈ Rn×d denotes a noise. Basically, n ≥ m. The goal of BSS is to estimate ˆA and ˆX so that ˆX provides unknown original signal as possible. November 29, 2011 6/26
  • 7. Kinds of BSS Methods Actually, degree of freedom of BSS model is very high to estimate A and X. Bcause there are a huge number of combinations (A, X) which satisfy Y = AX + E. Therefore, we need some constraint to solve the BSS problem such as: PCA : orthogonal constraint SCA : sparsity constraint NMF : non-negativity constraint ICA : independency constraint In this way, there are many methods to solve the BSS problem depending on the constraints. What we use is depend on subject matter. I will introduce about Non-negative Matrix Factrization(NMF) based technique for BSS. November 29, 2011 7/26
  • 8. Non-negativity Non-negativity means that value is 0 or positive (i.e. at least 0). Non-negative matrix is a matrix which its all elements are at least 0. For example 1 10 −10 1 /∈ R2×2 + , 1 10 0 1 ∈ R2×2 + . (3) November 29, 2011 8/26
  • 9. Why nonnegativity constraints? This is the key-point of this method. We can say that because many real-world data are nonnegative and the corresponding hidden components have a physical meaning only when nonnegative. Example of nonnegative data, image data (luminance) power spectol of signal (speech, EEG etc.) In general, observed data can be considered as a linear combination of such nonnegative data with noise.(i.e. Y = AX + E) November 29, 2011 9/26
  • 10. Standard NMF Model Consider the standard NMF model, given by: Y = AX + E, (4) A ≥ 0, X ≥ 0. (5) For simply, we denote A ≥ 0 as nonnegativity of matrix A. The NMF problem is given by: . NMF problem .. . . .. . . minimize 1 2 ||Y − AX||2 F (6) subject to A ≥ 0, X ≥ 0 (7) I will introduce the alternating least square(ALS) algorithm to solve this problem. November 29, 2011 10/26
  • 11. ALS algorithm Our goal is to estimate ( ˆA, ˆX) = minA,X 1 2 ||Y − AX||2 F , s.t. A ≥ 0, X ≥ 0. However, it is very difficult to solve such a multivariate problem. The ALS strategy is to solve following sub-problems alternately. . sub-problem for A .. . . .. . . min 1 2 ||Y − AX||2 F (8) s.t. A ≥ 0 (9) . sub-problem for X .. . . .. . . min 1 2 ||Y − AX||2 F (10) s.t. X ≥ 0 (11) November 29, 2011 11/26
  • 12. How to Solve the sub-problem Solving procedure of sub-problem for A is only two steps as follow: . sub-procedure for A .. . . .. . . ...1 A ← argminA 1 2 ||Y − AX||2 F ...2 A ← [A]+ where [a]+ = max(ε, a) rectify to enforce nonnegativity. In a similar way, we can solve sub-problem for X as follow: . sub-prodecure for X .. . . .. . . ...1 X ← argminX 1 2 ||Y − AX||2 F ...2 X ← [X]+ November 29, 2011 12/26
  • 13. Solution of Least Squares Here, I show basic calculation of least squares. First, objective function can be transformed as 1 2 ||Y − AX||2 F (12) = 1 2 tr(Y − AX)(Y − AX)T (13) = 1 2 tr(Y Y T − 2AXY T + AXXT AT ) (14) The stationary points ˆA can be found by equating the gradient components to 0: ∂ ∂AT 1 2 tr(Y Y T − 2AXY T + AXXT AT ) (15) = −XY T + XXT AT = 0 (16) Therefore, ˆAT = (XXT )−1 XY T . (17) November 29, 2011 13/26
  • 14. Solution of Least Squares (cont’d) In similar way, we can also obtain a solution of X. The objective function can be transformed as 1 2 ||Y − AX||2 F (18) = 1 2 tr(Y − AX)T (Y − AX) (19) = 1 2 tr(Y T Y − 2XT AT Y + XT AT AX) (20) The stationary points ˆX can be found by equating the gradient components to 0: ∂ ∂X 1 2 tr(Y T Y − 2XT AT Y + XT AT AX) (21) = −AT Y + AT AX = 0 (22) Therefore, ˆX = (AT A)−1 AT Y. (23) November 29, 2011 14/26
  • 15. Implimentation Implimentation in “octave” is very simple: i=0; while 1 i++; printf("%d: %f n",i,sqrt(sum(sum(E.^2)))); if sqrt(sum(sum(E.^2))) < er || i > maxiter break; end At=(X*X’+ 1e-5*eye(J))(X*Y’);A=At’; A(A < 0)=1e-16; X=(A’*A + 1e-5*eye(J))(A’*Y); X(X < 0)=1e-16; E=Y-A*X; end November 29, 2011 15/26
  • 16. Experiments: Artificial Data 1 -0.5 0 0.5 1 1.5 0 50 100 150 200 (a) sin2(t) -0.5 0 0.5 1 1.5 0 50 100 150 200 (b) cos2(t) Figure: Original Signals -0.5 0 0.5 1 1.5 0 50 100 150 200 (a) mixed (observed) signal 1 -0.5 0 0.5 1 1.5 0 50 100 150 200 (b) mixed (observed) signal 2 Figure: Observed Signals -0.5 0 0.5 1 1.5 0 50 100 150 200 (a) estimated signal 1 -0.5 0 0.5 1 1.5 0 50 100 150 200 (b) estimated signal 2 Figure: Estimated Signals We can see that results are almost separated. November 29, 2011 16/26
  • 17. Experiments: Artificial Data 2 -0.5 0 0.5 1 1.5 0 50 100 150 200 (a) sin2(t) -0.5 0 0.5 1 1.5 0 50 100 150 200 (b) sin2(2t) Figure: Original Signals -0.5 0 0.5 1 1.5 0 50 100 150 200 (a) mixed (observed) signal 1 -0.5 0 0.5 1 1.5 0 50 100 150 200 (b) mixed (observed) signal 2 Figure: Observed Signals -0.5 0 0.5 1 1.5 0 50 100 150 200 (a) estimated signal 1 -0.5 0 0.5 1 1.5 0 50 100 150 200 (b) estimated signal 2 Figure: Estimated Signals We can see that results are not so separated. November 29, 2011 17/26
  • 18. Experiments: Real Image 1 (a) newyork (b) shanghai Figure: Original Signals (a) mixed (observed) signal 1 (b) mixed (observed) signal 2 Figure: Observed Signals (a) estimated signal 1 (b) estimated signal 2 Figure: Estimated Signals We can see that results are almost separated. November 29, 2011 18/26
  • 19. Experiments: Real Image 2 (a) rock (b) pig Figure: Original Signals (a) mixed (observed) signal 1 (b) mixed (observed) signal 2 Figure: Observed Signals (a) estimated signal 1 (b) estimated signal 2 Figure: Estimated Signals We can see that estimate 2 is not so separated. November 29, 2011 19/26
  • 20. Experiments: Real Image 3 (a) nyc (b) sha (c) rock (d) pig (e) obs1 (f) obs2 (g) obs3 (h) obs4 Figure: Ori. & Obs. (a) estimated signal 1 (b) estimated signal 2 (c) estimated signal 3 (d) estimated signal 4 Figure: Estimated Signals: Results are not good. Signals are still mixed. November 29, 2011 20/26
  • 21. Issues There are some issues of this method as follow: Solution is non-unique (i.e. result is depend on starting point) It is not enough only by non-negativity constraint (shoud consider independency of each components) When number of estimate signals is large, features of components are overlapped. And we should consider more general issues of BSS as follow: Number of original signals is unknown (How can we decide its number?) November 29, 2011 21/26
  • 22. About this research area In this research area, many method for BSS are studied and proposed as follow: KL-divergence based NMF [Honkela et al., 2011] Alpha-Beta divergences based NMF [Cichocki et al., 2011] Non-Gaussianity based ICA [Hyv¨arinen et al., 2001] Kurtosis based ICA Negentropy based ICA Solving method for ICA gradient method fast fixed-point algorithm [Hyv¨arinen and Oja, 1997] MLE based ICA Mutual information based ICA Non-linear ICA Tensor ICA In this way, this research area is very broad!! November 29, 2011 22/26
  • 23. Summary I introduced about BSS problem and basic NMF tequnique from [Cichocki et al., 2009]. EEG classification is very difficult problem. . Future Work .. . . .. . . To study about extension of NMF and Independent Component Analysis (ICA). To apply the BSS problem, EEG analysis, and some pattern recognition problem. November 29, 2011 23/26
  • 24. Bibliography I [Cichocki et al., 2011] Cichocki, A., Cruces, S., and Amari, S.-i. (2011). Generalized alpha-beta divergences and their application to robust nonnegative matrix factorization. Entropy, 13(1):134–170. [Cichocki et al., 2009] Cichocki, A., Zdunek, R., Phan, A. H., and Amari, S. (2009). Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory Multi-way Data Analysis. Wiley. [Honkela et al., 2011] Honkela, T., Duch, W., Girolami, M., and Kaski, S., editors (2011). Kullback-Leibler Divergence for Nonnegative Matrix Factorization, volume 6791 of Lecture Notes in Computer Science. Springer Berlin / Heidelberg. [Hyv¨arinen et al., 2001] Hyv¨arinen, A., Karhunen, J., and Oja, E. (2001). Independent Component Analysis. Wiley. November 29, 2011 24/26
  • 25. Bibliography II [Hyv¨arinen and Oja, 1997] Hyv¨arinen, A. and Oja, E. (1997). A fast fixed-point algorithm for independent component analysis. Neural Computation, 9:1483–1492. November 29, 2011 25/26
  • 26. Thank you for listening November 29, 2011 26/26