Chapter15 2
Chapter15 2
Probabilistic Reasoning
Over Time-02
Sukarna Barua
Assistant Professor, CSE, BUET
Smoothing
Compute the probability distribution over past states.
Compute where
Smoothing
Compute where
Hence,
Where denotes forward probabilities [filtering problem]
Note that is a vector [ ]
Smoothing
Now, how to compute
)
[assume an specific output value at time step ]
Smoothing
Recurrence relation for
- )
- The algorithm for computing starts from the -th state [and go backward up to [Fill
DP tables for
- The algorithm is known as backward algorithm [in contrast to forward algorithm]
Smoothing
Finally
In vector form:
[point-wise multiplication]
Finally, compute:
Smoothing: Example
Smoothed estimate:
Filtered estimate:
Smoothing: Example
Smoothed estimate:
Filtered estimate:
Observation: Smoothed estimate for rain on day 1 is higher than the filtered
estimate (0.818) [Why?]
- This is because the umbrella on day 2 makes it more likely to have rained
on day 2 which in turn, because rain tends to persist, that makes it more
likely to have rained on day 1.
Smoothing: Algorithm Pseudocode
Forward and backward algorithm can be combined to compute posterior
probabilities in linear time
Most Likely Explanation
Suppose that [true, true, false, true, true] is the umbrella sequence for the
security guard’s first five days on the job.
There are possible weather sequences. Is there a way to find the most likely
one?
Most Likely Explanation: Naïve
Approach
Compute the probability distribution .
- [Bayes theorem]
- We will show that instead of maximizing over all sequence of previous states, it is
enough to maximize over only previous state [we will derive a recurrence
relation over the previous state]
Most Likely Explanation
Recurrence relation for computing : we can compute as follows:
- For all possible states at time step
- Compute the probability of most likely state sequence that produce observation
sequence and ends at (this probability is
- Move from state to [with transition probability ]
- Produce observation at state [with emission probability assuming ]
- As we don’t know the previous state leading to the most likely sequence, we just take the
)
Most Likely Explanation
Finally obtain the best possible state sequence:
is the most likely state sequence ending at at time step
We don’t know what is the end state for the most likely state sequence, hence we just
take the maximum over all