CS 59000 Statistical Machine learning Lecture 20 - PowerPoint PPT Presentation

1 / 34
About This Presentation
Title:

CS 59000 Statistical Machine learning Lecture 20

Description:

CS 59000 Statistical Machine learning Lecture 20 – PowerPoint PPT presentation

Number of Views:81
Avg rating:3.0/5.0
Slides: 35
Provided by: Markus105
Category:

less

Transcript and Presenter's Notes

Title: CS 59000 Statistical Machine learning Lecture 20


1
CS 59000 Statistical Machine learningLecture 20
  • Yuan (Alan) Qi
  • Purdue CS
  • Nov. 6 2008

2
Outline
  • Review of Bayesian networks, conditional
    independence, explaining away effect
  • D-separation, Markov random fields, Markov
    blankets, inference on chain

3
Bayesian Networks
  • Directed Acyclic Graph (DAG)

4
Bayesian Curve Fitting (2)
Plate
5
Bayesian Curve Fitting (3)
  • Input variables and explicit hyperparameters

6
Bayesian Curve Fitting Learning
  • Condition on data

7
Discrete Variables (1)
  • General joint distribution K 2 -1 parameters
  • Independent joint distribution 2(K-1) parameters

8
Discrete Variables Bayesian Parameters (1)
9
Parameterized Conditional Distributions
10
Linear-Gaussian Models
  • Directed Graph
  • Vector-valued Gaussian Nodes

Each node is Gaussian, the mean is a linear
function of the parents.
11
Conditional Independence
  • a is independent of b given c
  • Equivalently
  • Notation

12
Conditional Independence Example 1
13
Conditional Independence Example 1
14
Conditional Independence Example 2
15
Conditional Independence Example 2
16
Conditional Independence Example 3
  • Note this is the opposite of Example 1, with c
    unobserved.

17
Conditional Independence Example 3
  • Note this is the opposite of Example 1, with c
    observed.

18
D-separation
  • A, B, and C are non-intersecting subsets of nodes
    in a directed graph.
  • A path from A to B is blocked if it contains a
    node such that either
  • the arrows on the path meet either head-to-tail
    or tail-to-tail at the node, and the node is in
    the set C, or
  • the arrows meet head-to-head at the node, and
    neither the node, nor any of its descendants, are
    in the set C.
  • If all paths from A to B are blocked, A is said
    to be d-separated from B by C.
  • If A is d-separated from B by C, the joint
    distribution over all variables in the graph
    satisfies .

19
D-separation Example
20
D-separation I.I.D. Data
21
Bayesian Curve Fitting Revisited
D-separation implies that information from
training data is summarized in w.
22
Question
  • The minimal set of nodes that isolates a
    particular node from the rest of graph?

23
The Markov Blanket
Factors independent of xi cancel between
numerator and denominator.
24
Markov Random Fields
25
Cliques and Maximal Cliques
26
Joint Distribution
  • where is the potential over
    clique C and
  • is the normalization coefficient note M K-state
    variables ? KM terms in Z.
  • Energies and the Boltzmann distribution

27
Illustration Image De-Noising (1)
Original Image
Noisy Image
28
Illustration Image De-Noising (2)
29
Illustration Image De-Noising (3)
Noisy Image
Restored Image (ICM)
30
Converting Directed to Undirected Graphs (1)
31
Converting Directed to Undirected Graphs (2)
  • Additional links marrying parents, i.e.,
    moralization

32
Directed vs. Undirected Graphs (2)
33
Inference in Graphical Models
34
Inference on a Chain
Computational time increases exponentially with N.
35
Inference on a Chain
Write a Comment
User Comments (0)
About PowerShow.com