Hidden Markov Models Part 2 Algorithms CSE 4309

  • Slides: 77
Download presentation
Hidden Markov Models Part 2: Algorithms CSE 4309 – Machine Learning Vassilis Athitsos Computer

Hidden Markov Models Part 2: Algorithms CSE 4309 – Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1

Hidden Markov Model • 2

Hidden Markov Model • 2

The Basic HMM Problems • 3

The Basic HMM Problems • 3

The Basic HMM Problems • 4

The Basic HMM Problems • 4

Probability of Observations • 5

Probability of Observations • 5

Probability of Observations • 6

Probability of Observations • 6

Probability of Observations • 7

Probability of Observations • 7

The Sum Rule • 8

The Sum Rule • 8

The Sum Rule • 9

The Sum Rule • 9

The Sum Rule • 10

The Sum Rule • 10

The Forward Algorithm - Initialization • 12

The Forward Algorithm - Initialization • 12

The Forward Algorithm - Initialization • 13

The Forward Algorithm - Initialization • 13

The Forward Algorithm - Initialization • 14

The Forward Algorithm - Initialization • 14

The Forward Algorithm - Main Loop • 15

The Forward Algorithm - Main Loop • 15

The Forward Algorithm - Main Loop • 16

The Forward Algorithm - Main Loop • 16

The Forward Algorithm - Main Loop • 17

The Forward Algorithm - Main Loop • 17

The Forward Algorithm - Main Loop • 18

The Forward Algorithm - Main Loop • 18

The Forward Algorithm • 19

The Forward Algorithm • 19

The Viterbi Algorithm • 25

The Viterbi Algorithm • 25

The Viterbi Algorithm - Initialization • 26

The Viterbi Algorithm - Initialization • 26

The Viterbi Algorithm – Main Loop • 27

The Viterbi Algorithm – Main Loop • 27

The Viterbi Algorithm – Main Loop • 28

The Viterbi Algorithm – Main Loop • 28

The Viterbi Algorithm – Main Loop • 29

The Viterbi Algorithm – Main Loop • 29

The Viterbi Algorithm – Output • 30

The Viterbi Algorithm – Output • 30

State Probabilities at Specific Times • 31

State Probabilities at Specific Times • 31

State Probabilities at Specific Times • 32

State Probabilities at Specific Times • 32

State Probabilities at Specific Times • 33

State Probabilities at Specific Times • 33

State Probabilities at Specific Times • 34

State Probabilities at Specific Times • 34

The Backward Algorithm • 35

The Backward Algorithm • 35

Backward Algorithm - Initialization • 36

Backward Algorithm - Initialization • 36

Backward Algorithm - Initialization • 37

Backward Algorithm - Initialization • 37

Backward Algorithm - Initialization • 38

Backward Algorithm - Initialization • 38

Backward Algorithm – Main Loop • 39

Backward Algorithm – Main Loop • 39

Backward Algorithm – Main Loop • We will take a closer look at the

Backward Algorithm – Main Loop • We will take a closer look at the last step… 40

Backward Algorithm – Main Loop • 41

Backward Algorithm – Main Loop • 41

Backward Algorithm – Main Loop • 42

Backward Algorithm – Main Loop • 42

Backward Algorithm – Main Loop • 43

Backward Algorithm – Main Loop • 43

Backward Algorithm – Main Loop • 44

Backward Algorithm – Main Loop • 44

Backward Algorithm – Main Loop • 45

Backward Algorithm – Main Loop • 45

Backward Algorithm – Main Loop • 46

Backward Algorithm – Main Loop • 46

The Forward-Backward Algorithm • 47

The Forward-Backward Algorithm • 47

Problem 1: Training an HMM • 48

Problem 1: Training an HMM • 48

Problem 1: Training an HMM • 49

Problem 1: Training an HMM • 49

Expectation-Maximization • When we wanted to learn a mixture of Gaussians, we had the

Expectation-Maximization • When we wanted to learn a mixture of Gaussians, we had the following problem: – If we knew the probability of each object belonging to each Gaussian, we could estimate the parameters of each Gaussian. – If we knew the parameters of each Gaussian, we could estimate the probability of each object belonging to each Gaussian. – However, we know neither of these pieces of information. • The EM algorithm resolved this problem using: – An initialization of Gaussian parameters to some random or non-random values. – A main loop where: • The current values of the Gaussian parameters are used to estimate new weights of membership of every training object to every Gaussian. • The current estimated membership weights are used to estimate new parameters (mean and covariance matrix) for each Gaussian. 50

Expectation-Maximization • 51

Expectation-Maximization • 51

Baum-Welch: Initialization • 52

Baum-Welch: Initialization • 52

Baum-Welch: Initialization • 53

Baum-Welch: Initialization • 53

Baum-Welch: Expectation Step • 54

Baum-Welch: Expectation Step • 54

Baum-Welch: Expectation Step • 55

Baum-Welch: Expectation Step • 55

Baum-Welch: Summary of E Step • 63

Baum-Welch: Summary of E Step • 63

Baum-Welch: Maximization Step • 64

Baum-Welch: Maximization Step • 64

Baum-Welch: Maximization Step • 65

Baum-Welch: Maximization Step • 65

Baum-Welch: Maximization Step • 66

Baum-Welch: Maximization Step • 66

Baum-Welch: Maximization Step • 67

Baum-Welch: Maximization Step • 67

Baum-Welch: Maximization Step • 68

Baum-Welch: Maximization Step • 68

Baum-Welch: Maximization Step • 69

Baum-Welch: Maximization Step • 69

Baum-Welch: Maximization Step • 70

Baum-Welch: Maximization Step • 70

Baum-Welch: Maximization Step • 71

Baum-Welch: Maximization Step • 71

Baum-Welch: Maximization Step • 72

Baum-Welch: Maximization Step • 72

Baum-Welch: Summary of M Step • 73

Baum-Welch: Summary of M Step • 73

Baum-Welch: Summary of M Step • 74

Baum-Welch: Summary of M Step • 74

Baum-Welch Summary • 75

Baum-Welch Summary • 75

Hidden Markov Models - Recap • 76

Hidden Markov Models - Recap • 76

Hidden Markov Models - Recap • 77

Hidden Markov Models - Recap • 77