:
:
:
Dynamic
Data
Ba
essing
y
:
esian
and
Net
4.3
w
:
orks
esian
?
net
Zoubin
:
Ghahramani
:
Departmen
:
t
:
of
:
Computer
:
Science
ersion
Univ
Learning
ersit
:
y
:
of
:
T
6
oron
:
to
o
T
:
oron
:
to
:
ON
:
M
ariables
3H5,
:
Canada
smo
httpww
:
cs
Learning
t
:
oro
M
nt
Lecture
o
2
a
tutorial
:
zoubin
:
zoubins
Dynamic
tor
:
on
:
to
:
ed
Statepace
u
:
Octob
:
er
Example
1997
:
Abstract
:
Ba
:
y
:
esian
:
net
:
w
ML
orks
:
are
:
directed
Estimation
acyclic
:
graphs
statepace
that
:
represen
:
t
:
dep
:
endencies
:
b
12
et
v
w
:
een
A
v
in
ariables
daptive
in
or
a
In
probabilisti
:
c
Ba
mo
w
del
:
Man
:
y
:
time
:
se
:
ries
3
mo
y
dels
orks
including
:
the
:
hidden
:
Mark
:
o
:
v
Example
mo
dels
dels
:
MMs
:
used
:
in
:
sp
7
eec
Hidden
h
mo
recognition
:
and
:
Kalman
:
ter
:
mo
4
dels
:
used
:
in
:
tering
:
and
:
con
:
trol
8
ap
with
plications
:
can
:
b
:
e
:
view
4.2
ed
Hidden
as
EM
examples
:
of
1:
dynamic
dels
Ba
:
y
:
esian
:
net
4.4
w
:
orks
:
W
:
e
:
st
:
pro
:
vide
:
a
Example
brief
Mark
tutorial
dels
on
:
learning
:
and
14
Ba
did
y
app
esian
Giles
net
ds
w
o
orks
T
W
Information
e
in
then
Springer
presen
:
t
2
some
A
dynamic
y
Ba
net
y
ork
esian
:
net
:
w
:
orks
:
that
:
can
:
capture
:
m
:
uc
:
h
:
ric
:
her
3
structure
Ba
than
esian
HMMs
w
and
:
Kalman
:
ters
:
including
:
spatial
:
and
:
temp
:
oral
:
m
:
ultiresolutio
:
n
:
structure
3.1
distributed
1:
hidden
mo
state
:
represen
:
ta
:
tions
:
and
:
m
:
ultiple
:
switc
:
hing
:
linear
:
regimes
3.2
While
2:
exact
Mark
probabili
v
stic
dels
inference
:
is
:
in
:
tractable
:
in
:
these
:
net
:
w
:
orks
8
one
Learning
can
Inference
obtain
:
tractable
:
v
:
ari
:
ational
:
appro
:
ximations
:
whic
:
h
:
call
:
as
:
subroutines
:
the
4.1
forw
Estimation
ardac
Complete
kw
:
ard
:
and
:
Kalman
:
ter
:
recursions
:
These
:
appro
:
ximations
9
can
ML
b
with
e
V
used
The
to
algorithm
learn
:
the
10
mo
Example
del
Learning
parameters
mo
b
:
y
:
maximizing
:
a
:
lo
:
w
:
er
:
b
11
ound
Kalman
on
othing
the
:
lik
:
elih
:
o
:
o
:
d
:
T
:
able
:
of
:
Con
:
ten
:
ts
:
1
:
In
4.5
tro
2:
duction
hidden
:
o
:
mo
:
:
:
:
:
:
:
:
:
:
:
:
:
?
:
mo
:
v
:
to
:
ear
:
C
:
and
:
Gori
:
A
:
Pr
:
c
:
of
:
emp
:
al
:
.
:
Notes
:
Artiial
:
telligence
:
erlag
:
:predictions
sev
inference
The
in
forw
probabilit
ardac
learning
kw
et
ard
Y
algorithm
temp
:
h
:
deling
:
series
:
y
:
a
:
prediction
:
a
:
loss
:
w
:
ariables
:
one
:
w
:
and
:
algorithm
:
b
:
as
:
data
:
true
15
expressing
5
can
Bey
via
ond
:
T
e
ractable
that
Mo
probabilistic
dels
mo
:
or
:
h
:
only
:
but
:
probabilities
:
a
:
Ba
:
wn
:
Section
:
net
:
some
:
Inference
:
er
:
ximate
:
is
:
limited
:
y
:
mo
:
y
:
erful
:
and
:
[23
:
t
16
outcome
5.1
densit
Example
Y
3:
t
F
y
actorial
e
HMMs
or
:
to
:
hapter
:
learning
:
W
:
Ba
:
probabilistic
:
w
:
graph
:
et
:
.
:
user
:
whic
:
as
:
computing
:
b
:
next
:
Ba
:
demonstrates
:
w
:
ing
16
as
5.2
o
Example
on
4:
a
T
the
ree
Section
structured
dels
HMMs
m
:
mo
:
tractable
:
w
:
ds
:
can
:
learning
:
uncertain
:
from
:
of
:
and
:
h
:
een
:
and
:
cess
:
pro
:
o
:
ol
18
oth
5.3
t
Example
mo
5:
W
Switc
the
hing
in
State
the
space
t
mo
probabilit
dels
P
:
+1
:
;
:
;
:
Suc
:
y
:
then
:
to
:
oin
:
error
:
e
:
exp
:
ize
:
This
19
ts
6
ork
Inference
dels
and
ral
In
express
tractabili
using
t
esian
y
formalism
:
mo
:
elief
:
marriage
:
theory
:
in
:
endencies
:
een
:
expressed
:
graph
:
ws
:
understand
:
ariables
:
other
:
serv
:
bac
:
eien
:
and
:
ma
:
required
:
learning
:
pro
20
tutorial
6.1
esian
Gibbs
Section
sampling
use
:
esian
:
for
:
series
:
w
:
suc
:
Kalman
:
hidden
:
mo
:
fo
:
problem
:
parameters
:
y
:
ork
:
ectationaximization
:
,
:
describ
:
her
:
for
:
nonlinear
:
tiresolution
:
suc
:
ma
:
computationally
:
w
:
section
:
presen
:
tractable
:
ap
21
whic
6.2
e
V
basis
ariational
4.6
Metho
added
ds
t
:
resulting
:
the
:
size
:
our
:
set
:
an
:
mismatc
:
b
:
w
:
our
:
del
:
the
:
pro
:
Probabilit
:
theory
:
vides
:
p
:
w
:
to
:
for
:
b
:
randomness
:
uncertain
:
y
:
our
22
del
6.3
].
Example
e
Mean
express
ld
uncertain
for
y
factorial
our
HMMs
of
:
future
:
Y
:
+1
:
a
:
y
:
y
:
(
:
t
:
j
:
1
:
:
:
:
:
Y
:
).
23
h
6.4
probabilit
Example
densit
Structured
can
appro
b
ximation
used
for
mak
factorial
p
HMMs
t
:
dee
:
bars
:
mak
:
decisions
:
are
25
ected
6.5
minim
Con
some
v
function
ex
c
dualit
presen
y
a
:
framew
:
for
:
mo
:
of
:
o
:
data
:
e
:
these
:
dels
:
the
:
y
:
net
:
ork
:
(a.k.a.
:
graphical
:
dels
:
b
:
net
:
orksa
:
of
:
y
:
and
:
theory
:
whic
:
dep
:
b
:
w
:
v
27
are
7
graphically
Conclusion
The
:
not
:
allo
:
the
:
to
:
whic
:
v
:
act
:
h
:
ones
:
also
:
es
:
the
:
kb
:
for
:
tly
:
marginal
:
conditional
:
that
:
y
:
e
:
for
:
and
:
The
:
section
:
vides
:
brief
:
of
:
y
:
net
:
orks
:
3
:
the
28
of
1
y
In
net
tro
orks
duction
mo
Supp
time
ose
includ
w
some
e
ellno
wish
examples
to
h
build
the
a
er
mo
the
del
Mark
of
v
data
del
from
4
a
cuses
ite
the
sequence
of
of
the
ordered
of
observ
Ba
ations
esian
f
w
Y
using
1
Exp
;
M
Y
[3
2
10].
;
5
:
es
:
ric
:
mo
;
appropriate
Y
time
t
with
g
or
.
ul
In
structure
most
in
realistic
h
scenarios
dels
from
y
mo
e
deling
in
sto
Ho
c
ev
k
in
prices
6
to
e
ph
t
ysiological
eral
data
metho
the
for
observ
pro
ations
inference
are
h
not
b
related
used
deterministically
the
.
for
F
2
urthermore
thereW
indep
the
A
ted
Ba
B
y
ork
esian
)
net
the
w
at
ork
from
tutorial
suc
A
a
Ba
Y
y
represen
esian
wn
net
(1)
w
is
ork
no
is
c
simply
B
a
A
graphical
of
mo
giv
del
often
for
3
represen
)
ting
ork
conditional
t
in
net
dep
B
endencies
or
b
arc
et
net
w
deitions
een
A
a
arc
set
desc
of
on
random
no
v
in
ariables
An
Consider
des
four
the
random
tics
v
conditionally
ariables
,
W
no
,
b
X
v
,
Y
Y
(
,
j
and
y
Z
w
.
factorization
F
v
rom
no
basic
directed
probabilit
A
y
conditioned
theory
join
w
t
e
ould
kno
Y
w
The
that
ting
w
Figure
e
will
can
t
factor
ar
the
there
join
B
t
of
probabilit
no
y
c
as
cte
a
is
pro
A
duct
eac
of
a
conditional
de
probabilities
d
P
a
(
and
W
h
;
paren
X
no
;
y
Y
h
;
from
Z
ts
)
2
=
b
P
ariables
(
conditional
W
no
)
et
P
with
(
X
X
P
j
=
W
j
)
(
P
;
(
A
Y
net
j
a
W
y
;
a
X
a
)
Eac
P
is
(
y
Z
in
j
ork
W
is
;
no
X
no
;
B
Y
A
)
of
:
distribution
This
to
factorization
factoriza
do
e
es
w
not
W
tell
not
us
Z
an
y
ything
ork
useful
factorization
ab
wn
out
Some
the
graph
join
e
t
p
probabilit
no
y
a
dis
of
tribution
B
eac
a
h
A
v
if
ariable
a
can
.
p
of
oten
are
tially
c
dep
and
end
dir
on
p
ev
to
ery
sequence
other
starting
v
ending
ariable
h
Ho
no
w
sequence
ev
t
er
wing
consider
the
the
e
follo
ath
wing
B
factorization
of
P
from
(
in
W
that
;
de
X
is
;
or
Y
follo
;
The
Z
a
)
net
=
simple
P
de
(
enden
W
nonescenden
)
its
P
More
(
w
X
there
)
corresp
P
w
(
and
Y
e
j
ab
W
endence
)
w
P
meaning
(
relations
Z
een
j
asso
X
no
;
j
Y
;
)
)
:
(
(1)
)
The
P
ab
W
o
Y
v
P
e
Z
factorization
X
implies
Y
a
:
set
Ba
of
esian
conditional
w
indep
is
endence
graphical
relations
a
A
to
v
t
ariable
particular
r
of
set
join
of
distribution
v
h
ariables
ariable
A
represen
is
b
c
a
onditional
de
ly
the
indep
w
endent
A
from
arc
B
dra
giv
from
en
de
C
to
if
de
P
if
(
is
A
on
B
in
j
factorization
C
the
)
t
=
F
P
example
(
represen
A
the
j
tion
C
w
)
w
P
dra
(
an
B
from
j
to
C
but
)
from
for
to
all
.
A
Ba
,
esian
B
w
and
represen
C
the
suc
(1)
h
sho
that
in
P
1.
(
basic
C
from
)
theory
6
b
=
necessary
0.
this
F
oin
rom
The
the
de
ab
is
o
p
v
ent
e
another
factorization
de
w
if
e
is
can
directed
sho
from
w
to
that
;
giv
so
en
is
the
child
v
A
alues
The
of
endents
X
a
and
de
Y
its
,
hildren
Z
hildren
and
hilden
W
so
are
A
indep
e
enden
d
t
ath
P
A
(
B
Z
a
;
of
W
des
j
from
X
and
;
in
Y
suc
)
that
=
h
P
de
(
the
W
is
;
paren
X
of
;
follo
Y
no
;
in
Z
sequence
)
undir
P
cte
(
p
X
from
;
to
Y
is
)
sequence
=
no
P
starting
(
A
W
ending
)
B
P
h
(
eac
X
no
)
in
P
sequence
(
a
Y
t
j
child
W
the
)
wing
P
de
(
seman
Z
of
j
Ba
X
esian
;
w
Y
are
)
eac
R
no
P
is
(
indep
W
t
)
its
P
ts
(
en
X
paren
)
2
P
generally
(
t
Y
o
j
Since
W
is
)
oneone
P
ondence
(
et
Z
een
j
des
X
v
;
w
Y
will
)
talk
dW
out
d
indep
Z
relations
=
et
P
een
(
des
W
conditional
)
endence
P
b
(
w
Y
the
j
ariables
W
ciated
)
the
P
des
(
2
Zare
conditional
Iap
er
orks
pap
diren
this
Iap
Fig
whic
1.
as
A
path
directed
Ba
acyclic
onds
graph
ert
(D
computing
A
b
G
een
consisten
of
t
is
with
t
the
with
conditional
said
indep
d
endence
relation
relations
from
in
a
P
e
(
c
W
lo
;
or
X
more
;
there
Y
,
;
exact
Z
to
).
previous
disjoin
v
t
y
sets
].
of
particular
no
w
des
an
A
P
and
ed
B
alid
are
G
conditionally
can
indep
ving
enden
absence
t
net
giv
whic
en
eien
C
probabilities
,
net
if
undirected
C
a
d
agation
sep
cte
ar
there
ates
path
A
w
and
general
B
e
,
pro
that
propagation
is
throughout
if
and
along
3
ev
of
ery
no
undirected
mo
path
are
b
rep
et
ha
w
seman
een
deal
a
dels
no
t
de
esian
in
G
A
b
and
endency
a
a
no
ev
de
displa
in
G
B
a
there
indep
is
P
a
a
no
no
de
e
D
without
suc
Iap
h
.
that
arcs
(1)
y
D
orks
has
endence
con
can
v
to
erging
algorithms
arro
and
ws
or
3
cte
and
orks
neither
the
D
has
nor
there
its
algorithm
descenden
pr
ts
41].
are
c
in
net
C
whic
,
b
or
one
(2)
et
D
y
do
no
es
a
not
kno
ha
junction
v
algorithm
e
I
con
the
v
b
erging
ince
arro
ds
w
pap
and
on
D
the
is
an
in
is
C
c
[41
oth
].
follo
F
in
rom
Undirected
visual
ark
insp
w
ection
imp
of
ol
the
ting
graphical
ns
mo
e
del
set
it
[5,
is
e
therefore
el
easy
graphical
to
a
infer
in
man
A
y
y
indep
net
endence
ork
relations
is
without
to
explicitly
e
grinding
indep
through
map
Ba
for
y
distribution
es
if
rule
ery
F
eparation
or
y
example
in
W
corresp
is
to
conditionally
v
indep
conditional
enden
endence
t
in
from
.
X
is
giv
minimal
en
if
the
arc
set
b
C
deleted
=
G
f
remo
Y
the
;
prop
Z
y
g
The
,
of
since
in
Y
Ba
2
esian
C
w
is
implies
along
indep
the
relations
only
h
path
b
b
exploited
et
obtain
w
t
een
for
W
marginal
and
conditional
X
F
,
singly
and
onne
Y
d
do
w
es
in
not
h
ha
underlying
v
graph
e
no
con
ops
v
exists
erging
general
arro
called
ws
elief
Ho
op
w
[31,
ev
F
er
multiply
w
onne
e
d
cannot
w
infer
in