1. In a Hidden Markov Model, which component determines how likely an observation is generated from a hidden state?
A. Transition probability
B. Initial state probability
C. Emission probability
D. Posterior probability
Explanation:
Emission probabilities define how observations are generated from hidden states, making them critical in mapping hidden behavior to visible outputs.
What is emission probability in HMM?
Emission probability (also called output probability) in a Hidden Markov Model represents the likelihood of observing a particular symbol or observation given that the model is in a specific hidden state at a particular time step.
In an HMM, you have two types of events happening simultaneously: hidden states that are not directly observable and observations (emissions) that are visible. The emission probability defines the relationship between these hidden states and what we actually observe.
Example: Please refer hereA. The likelihood of the observation sequence
B. The most probable hidden state sequence
C. Transition matrix normalization
D. The number of emission symbols
Explanation:
The Viterbi algorithm finds the single most probable sequence of hidden states that could have produced the given observations.
Viterbi algorithm
The Viterbi algorithm is a dynamic programming algorithm that finds the most likely sequence of hidden states that would explain a sequence of observed events in a Hidden Markov Model. It solves the decoding problem in HMMs: given observations and the HMM model, what sequence of hidden states most likely produced those observations?
When would you need Viterbi algorithm?
You need the Viterbi algorithm whenever you have a decoding problem in a Hidden Markov Model—that is, when you need to infer the most likely sequence of hidden states from a sequence of observations. More specifically, the algorithm is essential when you face problems where hidden states influence observable data, but you only have access to the observations and need to determine what the hidden states were. Some example cases as follows;
- When you need the single most likely state sequence (e.g., transcribing a spoken word), Viterbi gives the exact MAP (maximum a‑posteriori) path.
- When the number of states is modest (tens to a few hundred). Runtime O(N²T) is usually fine.
A. A supervised algorithm for labeled sequences
B. A greedy optimization algorithm
C. An unsupervised EM-based algorithm for parameter estimation
D. A rule-based decoding algorithm
Explanation:
Baum-Welch is an Expectation–Maximization algorithm that updates transition and emission probabilities based on unlabeled data.
Baum-Welch algorithm (forward-backword algorithm)
The Baum-Welch algorithm is a machine learning algorithm used to solve the learning problem in Hidden Markov Models—estimating the unknown parameters of an HMM from observed data. It is also known as the forward-backward algorithm and is a special case of the Expectation-Maximization (EM) algorithm.
It is a method used to train a Hidden Markov Model (HMM) when you don’t know the correct state sequence in your data.
How does Baum-Welch algorith work?
It uses a two-step repeating process called EM (Expectation–Maximization):Expectation Step (E-step): The algorithm guesses the hidden state sequence based on current model parameters. In this step, it uses both forward and backward algorithms.
Maximization Step (M-step): Based on that guess, the algorithm updates the model parameters to better fit the data.
A. 6 × 6
B. 1 × 6
C. 4 × 6
D. 6 × 4
Explanation:
Each state must assign probabilities to all observation symbols, so the matrix is defined as: number of states × number of symbols.
Explanation: In a Hidden Markov Model (HMM), the emission matrix (also called the observation probability matrix) represents the probability of emitting each observation symbol from each hidden state.So its size depends on: Number of hidden states (N) → here: 4 Number of observation symbols (M) → here: 6 Therefore: Emission Matrix Size = 𝑁 × 𝑀 = 4 × 6
A. Removing hidden states
B. Modeling higher-order dependencies between previous states
C. Using equal (uniform) probabilities for all transitions
D. Allowing continuous observations only
Explanation:
The Markov assumption states that a state depends only on the previous state. If violated, the model must incorporate higher-order context. That means, Higher-order HMMs (Second-order, Third-order, etc.), where transitions depend on multiple past states, not just one.
Mathematically:
If we violate this assumption, it means the model must consider more than one previous state — meaning:
A. Maximization over paths
B. Linear rule-based selection
C. Random sampling of hidden sequences
D. Summation over possible hidden paths
Explanation:
The forward algorithm does not find the best path — instead, it computes the total probability of observing the sequence by summing over all possible hidden state paths.
Forward algorithm in HMM
The Forward Algorithm in a Hidden Markov Model (HMM) is a dynamic programming method used to compute the probability of an observation sequence, given the model parameters.
possible hidden-state paths that could produce the observations — too many to compute manually. The forward algorithm solves this efficiently by reusing intermediate results instead of recalculating everything.
A. Uniform state transition probabilities
B. High accuracy on training data but poor performance on new data
C. Low number of hidden states
D. Use of discrete emission probabilities only
Explanation:
Overfitting occurs when an HMM learns noise and memorizes transitions instead of generalizing sequence structure.
HMM can overfit?
An HMM becomes overfitted when it learns the training sequences too specifically, rather than learning general patterns. This often happens when:
-
The model has too many hidden states
-
The emission/transition probabilities become too precise for the training data
-
The dataset is small, but the model is complex
-
The parameters are estimated without regularization
In such cases, the HMM starts modeling noise or rare patterns in the training data, instead of meaningful structure.
A. Forward sums probabilities, Viterbi finds maximum sums path
B. Forward maximizes likelihood, Viterbi sums paths
C. Forward ignores emissions, Viterbi uses emissions
D. Forward is supervised, Viterbi is unsupervised
Explanation:
The forward algorithm computes total likelihood using summation, while Viterbi finds the best hidden sequence using maximization.
Difference between Forward algorithm and Viterbi algorithm in HMM
The Forward Algorithm and Viterbi Algorithm are two fundamental dynamic programming techniques used in Hidden Markov Models, but they solve different problems and employ different mathematical operations.
The Forward Algorithm
- computes the probability of observing a sequence, considering all possible hidden state paths that could have generated that sequence. It answers the question: "What is the likelihood of seeing this observation sequence?"
- used for evaluation problem in HMM.
- uses summation.
- Analogy: Given all possible routes to reach a city B from city A, forward algorithm answers "What is the total chance of reaching a city using any route?"
The Viterbi Algorithm, by contrast,
- finds the single most probable hidden state sequence that could have generated the observations. It answers: "What is the best sequence of hidden states that explains these observations?"
- used for decoding problem in HMM.
- uses maximization.
- Analogy: Given all possible routes to reach a city B from city A, Viterbi algorithm answers "Which single route is the most likely/best?"
A. They eliminate decoding steps
B. Speech signals are continuous-valued
C. They simplify transition probability computation
D. They require no training data
Explanation:
Speech data consists of real-valued acoustic features, making continuous modeling more natural than discrete symbol assignments.
A. Prior probability
B. Posterior state probability
C. Forward likelihood
D. Emission certainty factor
Explanation:
Posterior state probability represents confidence in each state at a specific time, calculated using the Forward-Backward algorithm.
Smoothing in HMMs means estimating the probability of hidden states using all observations — past, present, and future — to make the most accurate prediction.
Example:
Smoothing means deciding what part-of-speech a word most likely is, using the entire sentence — not just the words before it.
No comments:
Post a Comment