SlideShare uma empresa Scribd logo
1 de 26
An Algorithm for Bayesian Network
     Construction from Data
              by: Jie Cheng
                   David A. Bell
                   Weiru Liu
         University of Ulster, UK


         Presented by: Jian Xu
Outline
• Introduction
• Some basic concepts
• The proposed algorithm for BN
  construction
• Experiment results
• Discussions & comments



4/26/2010        Machine Learning   2
What is a Bayesian Network?
                                                      P(M)
                           Metastatic Cancer
                                                       .20
       M P(S)
       + .80
       - .20                                                 M P(B)
                                                             + .20
                Serum Calcium                 Brain Tumor    - .05


                                                              B P(H)
                S   B   P(C)
                                                              + .80
                +   +   .80
                                                              - .60
                +   -   .80
                -   +   .80       Coma
                                                      Headaches
                -   -   .05

                          Cancer BN Example
4/26/2010                      Machine Learning                        3
Bayesian Network (BN)
   • A Bayesian network is a compact
     graphical representation of a probability
     distribution over a set of domain
     random variables X = {X1, X2, …, Xn}
   • Two components
        – Structure: direct acyclic graph (DAG) over
          nodes, which exploits causal relations in
          the domain
        – CPD: each node has a conditional
          probability distribution associated with it

4/26/2010                Machine Learning               4
BN Learning
   • Structure learning
        – To identify the topology of the network
        – Score based methods
        – Dependency analysis methods
   • Parameter learning
        – To learn the conditional probabilities for a
          given network topology
        – MLE, Bayesian approach, etc

4/26/2010                Machine Learning                5
BN Structure Learning
   • Search & scoring methods:
        – To search for a structure most likely to have
          generated the data
        – Use heuristic search method to construct a model
          and evaluate it using a scoring method, such as
          MDL, Bayesian approach, etc
        – May not find the best solution
        – Random restarts: to avoid getting stuck in a local
          maximum
        – Less time complexity in the worst case, i.e., when
          the underlying DAG is fully connected

4/26/2010                  Machine Learning                    6
BN Learning Algorithms (Cont’d)
   • Dependency analysis methods:
        – Use conditional independency (CI) test to analyze
          dependency relationships among nodes.
        – Usually asymptotically correct when the data is
          DAG-faithful
        – Works efficiently when the underlying network is
          sparse
        – CI tests with large condition sets may be
          unreliable unless the volume of data is enormous.
        – Used in this proposed algorithm


4/26/2010                  Machine Learning                   7
Basic Concepts
   • D-separation: two nodes X and Y are called d-
     separated given C if and only if there exists no
     adjacency path P between X and Y, such that:
        – every collider on P is in C or has a descendant in C
        – no other nodes on path P is in C
        – C is called a condition-set
   • Open path: a path between X and Y is said to
     be open if every node in the path is active.
   • Closed path: if any node in the path is inactive
   • Collider node
   • Non-collider node
4/26/2010                   Machine Learning                     8
Basic Concepts (Cont’d)
   • DAG-faithful: when there exists such a DAG that can
     represent all the conditional independence relations of
     the underlying distribution.
   • D-map: a graph G is a dependency map (D-map) of M
     if every independence relationship in M is true in G. (a
     BN with no edge)
   • I-map: a graph G is an independency map (I-map) of
     M if every independence relationship in G is true in M.
     (fully-connected BN)
   • Minimum I-map: a graph G is an I-map of M, but the
     removal of any arc from G yields a graph that is not an
     I-map of M.
   • P-map: a graph G is a perfect map of M if it is both a
     D-map and an I-map of M.
4/26/2010                 Machine Learning                      9
Mutual Information
   • The mutual information of two nodes Xi ,
     Xj is defined as:


   • The conditional mutual information is
     defined as:




4/26/2010           Machine Learning         10
Assumptions
   • All attributes are discrete
   • No missing values in any record
   • All the records are drawn from a single
     probability model independently
   • The size of dataset is big enough for
     reliable CI tests
   • The ordering of the attributes are
     available before the network
     construction
4/26/2010           Machine Learning           11
An Algorithm for BN Construction
   • Drafting
        – Compute mutual information of each pair
          of nodes, and creates a draft of the model
   • Thickening
        – Adds arcs when the pairs of nodes cannot
          be d-separated, get an I-map of the model
   • Thinning
        – Each arc of the I-map is examined using CI
          tests and will be removed if the two nodes
          are the arc are conditionally independent
4/26/2010               Machine Learning               12
Drafting Phase
  1. Initiate a graph G(V, E) where V={all nodes}, E={ }, Initiate two
      empty lists S, R
  2. For each pair of nodes (vi, vj), i≠j, compute I(vi, vj). Sort all of the
     I(vi, vj) ≥ ε from large to small, and put the corresponding pairs of
     nodes into an ordered set S.
  3. Get the first two pairs of nodes in S, and remove them from S. Add
     the Corresponding arc to E. (the direction of the arcs is
     determined by the available node ordering)
  4. Get the first pair of nodes remained in S and remove it from S. If
     there is no open path between the two nodes (they are d-
     separated given empty set), add the corresponding arc to E.
     Otherwise add the pair of nodes to the end of an ordered set R.
  5. Repeat step 4 until S is empty.



4/26/2010                       Machine Learning                            13
Drafting Example
   • Figure (a) is the
     underlying BN structure
   • I(B,D) ≥ I(C,E) ≥ I(B,E)
     ≥ I(A,B) ≥ I(B,C) ≥
     I(C,D) ≥ I(D,E) ≥ I(A,D)
     ≥ I(A,E) ≥ I(A,C) ≥ ε
   • Figure (b) is the draft
     graph


4/26/2010              Machine Learning   14
Thickening Phase

6. Get the first pair of nodes in R and remove it
  from R
7. Find a block set that blocks each open path
  between these nodes by a set of minimum
  number of nodes. Conduct a CI test, if these two
  nodes are still dependent on each other given
  the block set, connect them by an arc.
8. Go to step 6 until R is empty.


4/26/2010            Machine Learning               15
Thickening Example
   • Figure (b) is the draft
     graph
   • Examine (D,E) pair, find
     the minimum set that
     blocks all the open paths
     between D and E {B}
   • CI test reveal that D and E
     are dependent given {B},
     so arc (D,E) is added
   • (A,C) is not added because
     A and C are independent
     given {B}

4/26/2010                 Machine Learning   16
Thinning Phase

9. For each arc in E, if there are open paths
  between the two nodes besides this arc,
  remove this arc from E temporarily, and
  call procedure find_block_set(current
  graph, node1, node2). Conduct a CI test
  on the condition of the block set. If the
  two nodes are dependent, add this arc
  back to E; otherwise remove the arc
  permanently.

4/26/2010          Machine Learning             17
Thinning Example
   • Figure (c) is the I-map
     of the underlying BN
   • Arc (B,E) is removed
     because B and E are
     independent of each
     other given {C,D}.
   • Figure (d) is the
     perfect I-map of the
     underlying dependency
     model (a).

4/26/2010             Machine Learning   18
Finding Minimum Block Set




4/26/2010      Machine Learning   19
Complexity Analysis
   • For a dataset with N attributes, r
     maximum possible values each, k
     parents at most
        – Phase I: N2 mutual information
          computation, each of which requires O(r2)
          basic operations, O(N2r2)
        – Phase II: at most N2 CI tests, each with at
          most O(rk+2) basic operations, O(N2rk+2),
          worst case O(N2rN)
        – Phase III: same as Phase II.

4/26/2010                Machine Learning               20
ALARM Network Structure




4/26/2010      Machine Learning   21
Experiment setup
   • ALARM BN (A Logical Alarm Reduction
     Mechanism): a medical diagnosis system for
     patient monitoring
        – 37 nodes, 46 arcs
        – 3 versions: same structure, different CPD’s
   • 10000 cases for each dataset
   • Modified conditional mutual information
     calculation by taking the variable’s degree of
     freedom into consideration to make CI tests
     more reliable
   • ε = 0.003
4/26/2010                  Machine Learning             22
Result on ALARM BN




4/26/2010     Machine Learning   23
Discussions & Comments
   • About the assumptions
        – All attributes are discrete
        – No missing values in any record
        – The size of dataset is big enough for
          reliable CI tests
        – The ordering of the attributes are available
          before the network construction



4/26/2010                Machine Learning                24
Discussions & Comments
   • Threshold ε
        – ε = 0.003
        – How do we pick an appropriate ε?
        – How does it affect the accuracy and time by
          choosing different ε?
   • Modification in the experiment part
        – Use Modified conditional mutual information
          calculation by taking the variable’s degree of
          freedom into consideration to make CI tests more
          reliable
        – Does this modification affect the result in any way
          other than increasing the accuracy?
4/26/2010                  Machine Learning                     25
4/26/2010   Machine Learning   26

Mais conteúdo relacionado

Mais procurados

Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
Subspace Indexing on Grassmannian Manifold for Large Scale Visual IdentificationSubspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
Subspace Indexing on Grassmannian Manifold for Large Scale Visual IdentificationUnited States Air Force Academy
 
Differential analyses of structures in HiC data
Differential analyses of structures in HiC dataDifferential analyses of structures in HiC data
Differential analyses of structures in HiC datatuxette
 
Manifold learning with application to object recognition
Manifold learning with application to object recognitionManifold learning with application to object recognition
Manifold learning with application to object recognitionzukun
 
JOSA TechTalks - Machine Learning on Graph-Structured Data
JOSA TechTalks - Machine Learning on Graph-Structured DataJOSA TechTalks - Machine Learning on Graph-Structured Data
JOSA TechTalks - Machine Learning on Graph-Structured DataJordan Open Source Association
 
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)Universitat Politècnica de Catalunya
 
Explanable models for time series with random forest
Explanable models for time series with random forestExplanable models for time series with random forest
Explanable models for time series with random foresttuxette
 
Kernel methods for data integration in systems biology
Kernel methods for data integration in systems biologyKernel methods for data integration in systems biology
Kernel methods for data integration in systems biologytuxette
 
Exact Inference in Bayesian Networks using MapReduce__HadoopSummit2010
Exact Inference in Bayesian Networks using MapReduce__HadoopSummit2010Exact Inference in Bayesian Networks using MapReduce__HadoopSummit2010
Exact Inference in Bayesian Networks using MapReduce__HadoopSummit2010Yahoo Developer Network
 
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)Universitat Politècnica de Catalunya
 
Comparison of Matrix Completion Algorithms for Background Initialization in V...
Comparison of Matrix Completion Algorithms for Background Initialization in V...Comparison of Matrix Completion Algorithms for Background Initialization in V...
Comparison of Matrix Completion Algorithms for Background Initialization in V...ActiveEon
 
Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering
Convolutional Neural Networks on Graphs with Fast Localized Spectral FilteringConvolutional Neural Networks on Graphs with Fast Localized Spectral Filtering
Convolutional Neural Networks on Graphs with Fast Localized Spectral FilteringSOYEON KIM
 
VEHICLE RECOGNITION USING VIBE AND SVM
VEHICLE RECOGNITION USING VIBE AND SVM VEHICLE RECOGNITION USING VIBE AND SVM
VEHICLE RECOGNITION USING VIBE AND SVM cseij
 
教師なし画像特徴表現学習の動向 {Un, Self} supervised representation learning (CVPR 2018 完全読破...
教師なし画像特徴表現学習の動向 {Un, Self} supervised representation learning (CVPR 2018 完全読破...教師なし画像特徴表現学習の動向 {Un, Self} supervised representation learning (CVPR 2018 完全読破...
教師なし画像特徴表現学習の動向 {Un, Self} supervised representation learning (CVPR 2018 完全読破...cvpaper. challenge
 

Mais procurados (18)

Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
Subspace Indexing on Grassmannian Manifold for Large Scale Visual IdentificationSubspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
Subspace Indexing on Grassmannian Manifold for Large Scale Visual Identification
 
Differential analyses of structures in HiC data
Differential analyses of structures in HiC dataDifferential analyses of structures in HiC data
Differential analyses of structures in HiC data
 
Lec16 subspace optimization
Lec16 subspace optimizationLec16 subspace optimization
Lec16 subspace optimization
 
Manifold learning with application to object recognition
Manifold learning with application to object recognitionManifold learning with application to object recognition
Manifold learning with application to object recognition
 
JOSA TechTalks - Machine Learning on Graph-Structured Data
JOSA TechTalks - Machine Learning on Graph-Structured DataJOSA TechTalks - Machine Learning on Graph-Structured Data
JOSA TechTalks - Machine Learning on Graph-Structured Data
 
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)
 
Explanable models for time series with random forest
Explanable models for time series with random forestExplanable models for time series with random forest
Explanable models for time series with random forest
 
Dagstuhl 2013 - Montali - On the Relationship between OBDA and Relational Map...
Dagstuhl 2013 - Montali - On the Relationship between OBDA and Relational Map...Dagstuhl 2013 - Montali - On the Relationship between OBDA and Relational Map...
Dagstuhl 2013 - Montali - On the Relationship between OBDA and Relational Map...
 
Kernel methods for data integration in systems biology
Kernel methods for data integration in systems biologyKernel methods for data integration in systems biology
Kernel methods for data integration in systems biology
 
Exact Inference in Bayesian Networks using MapReduce__HadoopSummit2010
Exact Inference in Bayesian Networks using MapReduce__HadoopSummit2010Exact Inference in Bayesian Networks using MapReduce__HadoopSummit2010
Exact Inference in Bayesian Networks using MapReduce__HadoopSummit2010
 
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)
Convolutional Neural Networks (D1L3 2017 UPC Deep Learning for Computer Vision)
 
Comparison of Matrix Completion Algorithms for Background Initialization in V...
Comparison of Matrix Completion Algorithms for Background Initialization in V...Comparison of Matrix Completion Algorithms for Background Initialization in V...
Comparison of Matrix Completion Algorithms for Background Initialization in V...
 
Deep 3D Visual Analysis - Javier Ruiz-Hidalgo - UPC Barcelona 2017
Deep 3D Visual Analysis - Javier Ruiz-Hidalgo - UPC Barcelona 2017Deep 3D Visual Analysis - Javier Ruiz-Hidalgo - UPC Barcelona 2017
Deep 3D Visual Analysis - Javier Ruiz-Hidalgo - UPC Barcelona 2017
 
Perceptrons (D1L2 2017 UPC Deep Learning for Computer Vision)
Perceptrons (D1L2 2017 UPC Deep Learning for Computer Vision)Perceptrons (D1L2 2017 UPC Deep Learning for Computer Vision)
Perceptrons (D1L2 2017 UPC Deep Learning for Computer Vision)
 
Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering
Convolutional Neural Networks on Graphs with Fast Localized Spectral FilteringConvolutional Neural Networks on Graphs with Fast Localized Spectral Filtering
Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering
 
Attention Models (D3L6 2017 UPC Deep Learning for Computer Vision)
Attention Models (D3L6 2017 UPC Deep Learning for Computer Vision)Attention Models (D3L6 2017 UPC Deep Learning for Computer Vision)
Attention Models (D3L6 2017 UPC Deep Learning for Computer Vision)
 
VEHICLE RECOGNITION USING VIBE AND SVM
VEHICLE RECOGNITION USING VIBE AND SVM VEHICLE RECOGNITION USING VIBE AND SVM
VEHICLE RECOGNITION USING VIBE AND SVM
 
教師なし画像特徴表現学習の動向 {Un, Self} supervised representation learning (CVPR 2018 完全読破...
教師なし画像特徴表現学習の動向 {Un, Self} supervised representation learning (CVPR 2018 完全読破...教師なし画像特徴表現学習の動向 {Un, Self} supervised representation learning (CVPR 2018 完全読破...
教師なし画像特徴表現学習の動向 {Un, Self} supervised representation learning (CVPR 2018 完全読破...
 

Destaque

Seminário redes bayesianas
Seminário redes bayesianasSeminário redes bayesianas
Seminário redes bayesianasiaudesc
 
Bayesian Networks with R and Hadoop
Bayesian Networks with R and HadoopBayesian Networks with R and Hadoop
Bayesian Networks with R and HadoopOfer Mendelevitch
 
Metropolis-Hastings MCMC Short Tutorial
Metropolis-Hastings MCMC Short TutorialMetropolis-Hastings MCMC Short Tutorial
Metropolis-Hastings MCMC Short TutorialRalph Schlosser
 
Enhancing the Status Message Question Asking Process on Facebook
Enhancing the Status Message Question Asking Process on FacebookEnhancing the Status Message Question Asking Process on Facebook
Enhancing the Status Message Question Asking Process on FacebookJonathas Magalhães
 
Design Pembelajaran Matematika dg Konteks GMT
Design Pembelajaran Matematika dg Konteks GMTDesign Pembelajaran Matematika dg Konteks GMT
Design Pembelajaran Matematika dg Konteks GMTSuci Agustina
 
Estatística: Introduçao à Estimacao Bayesiana
Estatística: Introduçao à Estimacao BayesianaEstatística: Introduçao à Estimacao Bayesiana
Estatística: Introduçao à Estimacao BayesianaRenato Vicente
 
Inteligencia financeira II
Inteligencia financeira IIInteligencia financeira II
Inteligencia financeira IIRenato Vicente
 
Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...
Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...
Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...PyData
 

Destaque (10)

Redes Bayesianas
Redes BayesianasRedes Bayesianas
Redes Bayesianas
 
Seminário redes bayesianas
Seminário redes bayesianasSeminário redes bayesianas
Seminário redes bayesianas
 
Bayesian Networks with R and Hadoop
Bayesian Networks with R and HadoopBayesian Networks with R and Hadoop
Bayesian Networks with R and Hadoop
 
Análise bayesiana de decisões aspectos práticos
Análise bayesiana de decisões   aspectos práticosAnálise bayesiana de decisões   aspectos práticos
Análise bayesiana de decisões aspectos práticos
 
Metropolis-Hastings MCMC Short Tutorial
Metropolis-Hastings MCMC Short TutorialMetropolis-Hastings MCMC Short Tutorial
Metropolis-Hastings MCMC Short Tutorial
 
Enhancing the Status Message Question Asking Process on Facebook
Enhancing the Status Message Question Asking Process on FacebookEnhancing the Status Message Question Asking Process on Facebook
Enhancing the Status Message Question Asking Process on Facebook
 
Design Pembelajaran Matematika dg Konteks GMT
Design Pembelajaran Matematika dg Konteks GMTDesign Pembelajaran Matematika dg Konteks GMT
Design Pembelajaran Matematika dg Konteks GMT
 
Estatística: Introduçao à Estimacao Bayesiana
Estatística: Introduçao à Estimacao BayesianaEstatística: Introduçao à Estimacao Bayesiana
Estatística: Introduçao à Estimacao Bayesiana
 
Inteligencia financeira II
Inteligencia financeira IIInteligencia financeira II
Inteligencia financeira II
 
Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...
Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...
Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...
 

Semelhante a Algorithm for Bayesian Network Construction from Data

Reverse Engineering for additive manufacturing
Reverse Engineering for additive manufacturingReverse Engineering for additive manufacturing
Reverse Engineering for additive manufacturingAchmadRifaie4
 
“Understanding DNN-Based Object Detectors,” a Presentation from Au-Zone Techn...
“Understanding DNN-Based Object Detectors,” a Presentation from Au-Zone Techn...“Understanding DNN-Based Object Detectors,” a Presentation from Au-Zone Techn...
“Understanding DNN-Based Object Detectors,” a Presentation from Au-Zone Techn...Edge AI and Vision Alliance
 
SCALABLE PATTERN MATCHING OVER COMPRESSED GRAPHS VIA DE-DENSIFICATION
SCALABLE PATTERN MATCHING OVER COMPRESSED GRAPHS VIA DE-DENSIFICATIONSCALABLE PATTERN MATCHING OVER COMPRESSED GRAPHS VIA DE-DENSIFICATION
SCALABLE PATTERN MATCHING OVER COMPRESSED GRAPHS VIA DE-DENSIFICATIONaftab alam
 
So sánh cấu trúc protein_Protein structure comparison
So sánh cấu trúc protein_Protein structure comparisonSo sánh cấu trúc protein_Protein structure comparison
So sánh cấu trúc protein_Protein structure comparisonbomxuan868
 
Advanced Data Structures 2006
Advanced Data Structures 2006Advanced Data Structures 2006
Advanced Data Structures 2006Sanjay Goel
 
NS-CUK Seminar: H.B.Kim, Review on "Inductive Representation Learning on Lar...
NS-CUK Seminar: H.B.Kim,  Review on "Inductive Representation Learning on Lar...NS-CUK Seminar: H.B.Kim,  Review on "Inductive Representation Learning on Lar...
NS-CUK Seminar: H.B.Kim, Review on "Inductive Representation Learning on Lar...ssuser4b1f48
 
ieee nss mic 2016 poster N30-21
ieee nss mic 2016 poster N30-21ieee nss mic 2016 poster N30-21
ieee nss mic 2016 poster N30-21Dae Woon Kim
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Mostafa G. M. Mostafa
 
Fault tolerance in wireless sensor networks by Constrained Delaunay Triangula...
Fault tolerance in wireless sensor networks by Constrained Delaunay Triangula...Fault tolerance in wireless sensor networks by Constrained Delaunay Triangula...
Fault tolerance in wireless sensor networks by Constrained Delaunay Triangula...Sigma web solutions pvt. ltd.
 
ProbabilisticModeling20080411
ProbabilisticModeling20080411ProbabilisticModeling20080411
ProbabilisticModeling20080411Clay Stanek
 
machine learning.pptx
machine learning.pptxmachine learning.pptx
machine learning.pptxAbdusSadik
 
Design of 8-Bit Comparator Using 45nm CMOS Technology
Design of 8-Bit Comparator Using 45nm CMOS TechnologyDesign of 8-Bit Comparator Using 45nm CMOS Technology
Design of 8-Bit Comparator Using 45nm CMOS TechnologyIJMER
 
Camera-Based Road Lane Detection by Deep Learning II
Camera-Based Road Lane Detection by Deep Learning IICamera-Based Road Lane Detection by Deep Learning II
Camera-Based Road Lane Detection by Deep Learning IIYu Huang
 
UNIT II GEOMETRIC MODELING (COMPUTER AIDED DESIGN AND MANUFACTURING )
UNIT II GEOMETRIC MODELING (COMPUTER AIDED DESIGN AND MANUFACTURING )UNIT II GEOMETRIC MODELING (COMPUTER AIDED DESIGN AND MANUFACTURING )
UNIT II GEOMETRIC MODELING (COMPUTER AIDED DESIGN AND MANUFACTURING )ravis205084
 
Distributed graph summarization
Distributed graph summarizationDistributed graph summarization
Distributed graph summarizationaftab alam
 
Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud
Point-GNN: Graph Neural Network for 3D Object Detection in a Point CloudPoint-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud
Point-GNN: Graph Neural Network for 3D Object Detection in a Point CloudNuwan Sriyantha Bandara
 

Semelhante a Algorithm for Bayesian Network Construction from Data (20)

Bayesian network
Bayesian networkBayesian network
Bayesian network
 
lecture_16.pptx
lecture_16.pptxlecture_16.pptx
lecture_16.pptx
 
Reverse Engineering for additive manufacturing
Reverse Engineering for additive manufacturingReverse Engineering for additive manufacturing
Reverse Engineering for additive manufacturing
 
“Understanding DNN-Based Object Detectors,” a Presentation from Au-Zone Techn...
“Understanding DNN-Based Object Detectors,” a Presentation from Au-Zone Techn...“Understanding DNN-Based Object Detectors,” a Presentation from Au-Zone Techn...
“Understanding DNN-Based Object Detectors,” a Presentation from Au-Zone Techn...
 
SCALABLE PATTERN MATCHING OVER COMPRESSED GRAPHS VIA DE-DENSIFICATION
SCALABLE PATTERN MATCHING OVER COMPRESSED GRAPHS VIA DE-DENSIFICATIONSCALABLE PATTERN MATCHING OVER COMPRESSED GRAPHS VIA DE-DENSIFICATION
SCALABLE PATTERN MATCHING OVER COMPRESSED GRAPHS VIA DE-DENSIFICATION
 
So sánh cấu trúc protein_Protein structure comparison
So sánh cấu trúc protein_Protein structure comparisonSo sánh cấu trúc protein_Protein structure comparison
So sánh cấu trúc protein_Protein structure comparison
 
Advanced Data Structures 2006
Advanced Data Structures 2006Advanced Data Structures 2006
Advanced Data Structures 2006
 
NS-CUK Seminar: H.B.Kim, Review on "Inductive Representation Learning on Lar...
NS-CUK Seminar: H.B.Kim,  Review on "Inductive Representation Learning on Lar...NS-CUK Seminar: H.B.Kim,  Review on "Inductive Representation Learning on Lar...
NS-CUK Seminar: H.B.Kim, Review on "Inductive Representation Learning on Lar...
 
ieee nss mic 2016 poster N30-21
ieee nss mic 2016 poster N30-21ieee nss mic 2016 poster N30-21
ieee nss mic 2016 poster N30-21
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)
 
Fault tolerance in wireless sensor networks by Constrained Delaunay Triangula...
Fault tolerance in wireless sensor networks by Constrained Delaunay Triangula...Fault tolerance in wireless sensor networks by Constrained Delaunay Triangula...
Fault tolerance in wireless sensor networks by Constrained Delaunay Triangula...
 
ProbabilisticModeling20080411
ProbabilisticModeling20080411ProbabilisticModeling20080411
ProbabilisticModeling20080411
 
machine learning.pptx
machine learning.pptxmachine learning.pptx
machine learning.pptx
 
Data Applied:Outliers
Data Applied:OutliersData Applied:Outliers
Data Applied:Outliers
 
Data Applied:Outliers
Data Applied:OutliersData Applied:Outliers
Data Applied:Outliers
 
Design of 8-Bit Comparator Using 45nm CMOS Technology
Design of 8-Bit Comparator Using 45nm CMOS TechnologyDesign of 8-Bit Comparator Using 45nm CMOS Technology
Design of 8-Bit Comparator Using 45nm CMOS Technology
 
Camera-Based Road Lane Detection by Deep Learning II
Camera-Based Road Lane Detection by Deep Learning IICamera-Based Road Lane Detection by Deep Learning II
Camera-Based Road Lane Detection by Deep Learning II
 
UNIT II GEOMETRIC MODELING (COMPUTER AIDED DESIGN AND MANUFACTURING )
UNIT II GEOMETRIC MODELING (COMPUTER AIDED DESIGN AND MANUFACTURING )UNIT II GEOMETRIC MODELING (COMPUTER AIDED DESIGN AND MANUFACTURING )
UNIT II GEOMETRIC MODELING (COMPUTER AIDED DESIGN AND MANUFACTURING )
 
Distributed graph summarization
Distributed graph summarizationDistributed graph summarization
Distributed graph summarization
 
Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud
Point-GNN: Graph Neural Network for 3D Object Detection in a Point CloudPoint-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud
Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud
 

Mais de butest

EL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBEEL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBEbutest
 
1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同butest
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALbutest
 
Timeline: The Life of Michael Jackson
Timeline: The Life of Michael JacksonTimeline: The Life of Michael Jackson
Timeline: The Life of Michael Jacksonbutest
 
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...butest
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALbutest
 
Com 380, Summer II
Com 380, Summer IICom 380, Summer II
Com 380, Summer IIbutest
 
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet JazzThe MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazzbutest
 
MICHAEL JACKSON.doc
MICHAEL JACKSON.docMICHAEL JACKSON.doc
MICHAEL JACKSON.docbutest
 
Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1butest
 
Facebook
Facebook Facebook
Facebook butest
 
Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...butest
 
Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...butest
 
NEWS ANNOUNCEMENT
NEWS ANNOUNCEMENTNEWS ANNOUNCEMENT
NEWS ANNOUNCEMENTbutest
 
C-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.docC-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.docbutest
 
MAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.docMAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.docbutest
 
Mac OS X Guide.doc
Mac OS X Guide.docMac OS X Guide.doc
Mac OS X Guide.docbutest
 
WEB DESIGN!
WEB DESIGN!WEB DESIGN!
WEB DESIGN!butest
 

Mais de butest (20)

EL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBEEL MODELO DE NEGOCIO DE YOUTUBE
EL MODELO DE NEGOCIO DE YOUTUBE
 
1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同1. MPEG I.B.P frame之不同
1. MPEG I.B.P frame之不同
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIAL
 
Timeline: The Life of Michael Jackson
Timeline: The Life of Michael JacksonTimeline: The Life of Michael Jackson
Timeline: The Life of Michael Jackson
 
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
Popular Reading Last Updated April 1, 2010 Adams, Lorraine The ...
 
LESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIALLESSONS FROM THE MICHAEL JACKSON TRIAL
LESSONS FROM THE MICHAEL JACKSON TRIAL
 
Com 380, Summer II
Com 380, Summer IICom 380, Summer II
Com 380, Summer II
 
PPT
PPTPPT
PPT
 
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet JazzThe MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
The MYnstrel Free Press Volume 2: Economic Struggles, Meet Jazz
 
MICHAEL JACKSON.doc
MICHAEL JACKSON.docMICHAEL JACKSON.doc
MICHAEL JACKSON.doc
 
Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1Social Networks: Twitter Facebook SL - Slide 1
Social Networks: Twitter Facebook SL - Slide 1
 
Facebook
Facebook Facebook
Facebook
 
Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...Executive Summary Hare Chevrolet is a General Motors dealership ...
Executive Summary Hare Chevrolet is a General Motors dealership ...
 
Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...Welcome to the Dougherty County Public Library's Facebook and ...
Welcome to the Dougherty County Public Library's Facebook and ...
 
NEWS ANNOUNCEMENT
NEWS ANNOUNCEMENTNEWS ANNOUNCEMENT
NEWS ANNOUNCEMENT
 
C-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.docC-2100 Ultra Zoom.doc
C-2100 Ultra Zoom.doc
 
MAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.docMAC Printing on ITS Printers.doc.doc
MAC Printing on ITS Printers.doc.doc
 
Mac OS X Guide.doc
Mac OS X Guide.docMac OS X Guide.doc
Mac OS X Guide.doc
 
hier
hierhier
hier
 
WEB DESIGN!
WEB DESIGN!WEB DESIGN!
WEB DESIGN!
 

Algorithm for Bayesian Network Construction from Data

  • 1. An Algorithm for Bayesian Network Construction from Data by: Jie Cheng David A. Bell Weiru Liu University of Ulster, UK Presented by: Jian Xu
  • 2. Outline • Introduction • Some basic concepts • The proposed algorithm for BN construction • Experiment results • Discussions & comments 4/26/2010 Machine Learning 2
  • 3. What is a Bayesian Network? P(M) Metastatic Cancer .20 M P(S) + .80 - .20 M P(B) + .20 Serum Calcium Brain Tumor - .05 B P(H) S B P(C) + .80 + + .80 - .60 + - .80 - + .80 Coma Headaches - - .05 Cancer BN Example 4/26/2010 Machine Learning 3
  • 4. Bayesian Network (BN) • A Bayesian network is a compact graphical representation of a probability distribution over a set of domain random variables X = {X1, X2, …, Xn} • Two components – Structure: direct acyclic graph (DAG) over nodes, which exploits causal relations in the domain – CPD: each node has a conditional probability distribution associated with it 4/26/2010 Machine Learning 4
  • 5. BN Learning • Structure learning – To identify the topology of the network – Score based methods – Dependency analysis methods • Parameter learning – To learn the conditional probabilities for a given network topology – MLE, Bayesian approach, etc 4/26/2010 Machine Learning 5
  • 6. BN Structure Learning • Search & scoring methods: – To search for a structure most likely to have generated the data – Use heuristic search method to construct a model and evaluate it using a scoring method, such as MDL, Bayesian approach, etc – May not find the best solution – Random restarts: to avoid getting stuck in a local maximum – Less time complexity in the worst case, i.e., when the underlying DAG is fully connected 4/26/2010 Machine Learning 6
  • 7. BN Learning Algorithms (Cont’d) • Dependency analysis methods: – Use conditional independency (CI) test to analyze dependency relationships among nodes. – Usually asymptotically correct when the data is DAG-faithful – Works efficiently when the underlying network is sparse – CI tests with large condition sets may be unreliable unless the volume of data is enormous. – Used in this proposed algorithm 4/26/2010 Machine Learning 7
  • 8. Basic Concepts • D-separation: two nodes X and Y are called d- separated given C if and only if there exists no adjacency path P between X and Y, such that: – every collider on P is in C or has a descendant in C – no other nodes on path P is in C – C is called a condition-set • Open path: a path between X and Y is said to be open if every node in the path is active. • Closed path: if any node in the path is inactive • Collider node • Non-collider node 4/26/2010 Machine Learning 8
  • 9. Basic Concepts (Cont’d) • DAG-faithful: when there exists such a DAG that can represent all the conditional independence relations of the underlying distribution. • D-map: a graph G is a dependency map (D-map) of M if every independence relationship in M is true in G. (a BN with no edge) • I-map: a graph G is an independency map (I-map) of M if every independence relationship in G is true in M. (fully-connected BN) • Minimum I-map: a graph G is an I-map of M, but the removal of any arc from G yields a graph that is not an I-map of M. • P-map: a graph G is a perfect map of M if it is both a D-map and an I-map of M. 4/26/2010 Machine Learning 9
  • 10. Mutual Information • The mutual information of two nodes Xi , Xj is defined as: • The conditional mutual information is defined as: 4/26/2010 Machine Learning 10
  • 11. Assumptions • All attributes are discrete • No missing values in any record • All the records are drawn from a single probability model independently • The size of dataset is big enough for reliable CI tests • The ordering of the attributes are available before the network construction 4/26/2010 Machine Learning 11
  • 12. An Algorithm for BN Construction • Drafting – Compute mutual information of each pair of nodes, and creates a draft of the model • Thickening – Adds arcs when the pairs of nodes cannot be d-separated, get an I-map of the model • Thinning – Each arc of the I-map is examined using CI tests and will be removed if the two nodes are the arc are conditionally independent 4/26/2010 Machine Learning 12
  • 13. Drafting Phase 1. Initiate a graph G(V, E) where V={all nodes}, E={ }, Initiate two empty lists S, R 2. For each pair of nodes (vi, vj), i≠j, compute I(vi, vj). Sort all of the I(vi, vj) ≥ ε from large to small, and put the corresponding pairs of nodes into an ordered set S. 3. Get the first two pairs of nodes in S, and remove them from S. Add the Corresponding arc to E. (the direction of the arcs is determined by the available node ordering) 4. Get the first pair of nodes remained in S and remove it from S. If there is no open path between the two nodes (they are d- separated given empty set), add the corresponding arc to E. Otherwise add the pair of nodes to the end of an ordered set R. 5. Repeat step 4 until S is empty. 4/26/2010 Machine Learning 13
  • 14. Drafting Example • Figure (a) is the underlying BN structure • I(B,D) ≥ I(C,E) ≥ I(B,E) ≥ I(A,B) ≥ I(B,C) ≥ I(C,D) ≥ I(D,E) ≥ I(A,D) ≥ I(A,E) ≥ I(A,C) ≥ ε • Figure (b) is the draft graph 4/26/2010 Machine Learning 14
  • 15. Thickening Phase 6. Get the first pair of nodes in R and remove it from R 7. Find a block set that blocks each open path between these nodes by a set of minimum number of nodes. Conduct a CI test, if these two nodes are still dependent on each other given the block set, connect them by an arc. 8. Go to step 6 until R is empty. 4/26/2010 Machine Learning 15
  • 16. Thickening Example • Figure (b) is the draft graph • Examine (D,E) pair, find the minimum set that blocks all the open paths between D and E {B} • CI test reveal that D and E are dependent given {B}, so arc (D,E) is added • (A,C) is not added because A and C are independent given {B} 4/26/2010 Machine Learning 16
  • 17. Thinning Phase 9. For each arc in E, if there are open paths between the two nodes besides this arc, remove this arc from E temporarily, and call procedure find_block_set(current graph, node1, node2). Conduct a CI test on the condition of the block set. If the two nodes are dependent, add this arc back to E; otherwise remove the arc permanently. 4/26/2010 Machine Learning 17
  • 18. Thinning Example • Figure (c) is the I-map of the underlying BN • Arc (B,E) is removed because B and E are independent of each other given {C,D}. • Figure (d) is the perfect I-map of the underlying dependency model (a). 4/26/2010 Machine Learning 18
  • 19. Finding Minimum Block Set 4/26/2010 Machine Learning 19
  • 20. Complexity Analysis • For a dataset with N attributes, r maximum possible values each, k parents at most – Phase I: N2 mutual information computation, each of which requires O(r2) basic operations, O(N2r2) – Phase II: at most N2 CI tests, each with at most O(rk+2) basic operations, O(N2rk+2), worst case O(N2rN) – Phase III: same as Phase II. 4/26/2010 Machine Learning 20
  • 21. ALARM Network Structure 4/26/2010 Machine Learning 21
  • 22. Experiment setup • ALARM BN (A Logical Alarm Reduction Mechanism): a medical diagnosis system for patient monitoring – 37 nodes, 46 arcs – 3 versions: same structure, different CPD’s • 10000 cases for each dataset • Modified conditional mutual information calculation by taking the variable’s degree of freedom into consideration to make CI tests more reliable • ε = 0.003 4/26/2010 Machine Learning 22
  • 23. Result on ALARM BN 4/26/2010 Machine Learning 23
  • 24. Discussions & Comments • About the assumptions – All attributes are discrete – No missing values in any record – The size of dataset is big enough for reliable CI tests – The ordering of the attributes are available before the network construction 4/26/2010 Machine Learning 24
  • 25. Discussions & Comments • Threshold ε – ε = 0.003 – How do we pick an appropriate ε? – How does it affect the accuracy and time by choosing different ε? • Modification in the experiment part – Use Modified conditional mutual information calculation by taking the variable’s degree of freedom into consideration to make CI tests more reliable – Does this modification affect the result in any way other than increasing the accuracy? 4/26/2010 Machine Learning 25
  • 26. 4/26/2010 Machine Learning 26