✔ Scroll down and test yourself — answers are hidden under the “View Answer” button.
Attempt all questions first.
✔️ Click SUBMIT at the end to unlock VIEW ANSWER buttons.
With uniform transitions, tagging depends only on P(word|tag), i.e., emission probabilities.
With uniform transitions, an HMM POS tagger reduces to a unigram model that tags each word independently using emission probabilities only.
Step-by-step Explanation
An HMM POS tagger assigns part-of-speech tags using two probabilities:
-
Transition probability -
P(ti | ti−1)
→ How likely a tag follows the previous tag -
Emission probability -
P(wi | ti)
→ How likely a word is generated by a tag
During decoding using the Viterbi algorithm, the model maximizes:
P(ti | ti−1) × P(wi | ti)
What does uniform / random transitions mean?
Uniform transitions imply:
P(ti | ti−1) = constant for all tag pairs
- Transition probabilities do not prefer any particular tag sequence
- They contribute the same value for every possible path
Therefore, transition probabilities no longer influence the tagging decision.
What remains?
Only the emission probabilities matter:
arg maxti P(wi | ti)
This is exactly what a unigram POS tagger does:
- Assigns each word the tag with the highest emission probability
- Ignores contextual information entirely
This question is about how POS taggers handle unknown (out-of-vocabulary) words—words that were not seen during training.
In Part-of-Speech (POS) tagging, unknown words present a fundamental challenge—they don't appear in the training corpus, so the model cannot rely on learned word-to-tag associations. The solution lies in morphological features, particularly prefix and suffix distributions linked to grammatical categories. Morphological cues like -ly, -ness, -tion strongly correlate with POS tags.
Prefix/suffix distribution per POS
Why? Many parts of speech follow strong morphological patterns:
- -tion, -ness → Noun
- -ly → Adverb
- -ing, -ed → Verb
- un-, re-, pre- → Verbs / Adjectives
- Infer the POS of new (unknown) words it has never seen
Therefore, unknown word tagging accuracy is highest when the model learns prefix/suffix distributions per POS.
Both words and tags are known, so probabilities are estimated directly.
Baum–Welch uses EM to estimate hidden states from unlabeled data.
This question tests your understanding of the three fundamental HMM problems and when to apply each algorithm.
This question asks about the Learning problem of three HMM problems. As per the Learning problem, we are given only the observation sequence without tags, and we need to find the model parameters with the help of Forward-Backward (Baum-Welch) algorithm. This is unsupervised learning.
A wrong but strong emission can lock Viterbi into an incorrect path.
More information:The question is about Viterbi decoding. Viterbi decoding is used in POS tagging and other sequence labeling tasks. In these tasks, each tag depends on the previous tag.
What is error propagation?In Viterbi decoding: The algorithm selects the best path step by step. If a wrong tag is chosen early, that wrong choice affects the next tags. As a result, more errors occur later. This spreading of mistakes is called error propagation.
Why does error propagation happen in Viterbi algorithm?Viterbi uses dynamic programming. It keeps only one best path, not many alternatives. It depends on: Transition probabilities (from one tag to the next) and Emission probabilities (tag to word). If the model is too confident about a wrong tag, the error continues through the sentence.
Why option B is correct?For a rare or unknown word, the model assigns: Very high probability to one tag and Very low probability to others. If that high-probability tag is wrong, Viterbi commits strongly to it and the alternative paths are discarded. As a result, future tags are forced to follow this wrong tag via transitions
Neural models capture global dependencies beyond Markov assumptions.
Determiner → noun is a common English syntactic pattern.
P(NN | DT) means, "The probability that the next tag is a Noun (NN), given that the current tag is a Determiner (DT)".
High probability indicates that Determiners are usually followed by Nouns in language data. Example: "the book", "a pen", "this idea", etc.
HMM probability is the product over all transition and emission terms.
Sentence probability in HMM = the sum of probabilities of all possible hidden state sequences that could generate the observed sentence, where each path's probability is the product of its transition and emission probabilities.
This probabilistic framework is what makes HMMs powerful for sequence modeling in NLP and speech processing—it elegantly handles the uncertainty of hidden states while maintaining computational efficiency through dynamic programming.
Backpointers reconstruct the optimal tag sequence.
The Viterbi backpointer table is a table that stores where each best probability came from during Viterbi decoding. In simple words, it remembers which previous state (tag) gave the best path to the current state.
Why do we need a backpointer table?
Viterbi decoding has two main steps: (1) Forward pass (Compute the best probability of reaching each state at each time step, Store these probabilities in the Viterbi table.), (2) Backtracking (Recover the best sequence of tags, For this, we must know which state we came from). The backpointer table makes backtracking possible.
Simple Viterbi Example
Sentence: "Time flies"
Possible Tags:
- NN (Noun)
- VB (Verb)
Viterbi Probability Table
| Time | NN | VB |
|---|---|---|
| t = 1 | 0.3 | 0.7 |
| t = 2 | 0.6 | 0.4 |
Backpointer Table
| Time | NN | VB |
|---|---|---|
| t = 2 | VB | NN |
Interpretation
- Best path to NN at t = 2 came from VB at t = 1.
- Best path to VB at t = 2 came from NN at t = 1.
Acoustic signals are continuous-valued, well-modeled by Gaussians.
Fundamental Distinction: Discrete vs. Continuous Observations
The choice between discrete and Gaussian (continuous) HMM emission distributions depends entirely on the nature of the observations being modeled.
Discrete HMMs represent observations as discrete symbols from a finite alphabet—such as words in part-of-speech tagging or written text. When observations are discrete, emission probabilities are modeled as categorical distributions over symbol categories.
Continuous (Gaussian) HMMs represent observations as continuous-valued feature vectors. When observations are real-valued, discrete emission probabilities are not applicable; instead, the probability density is modeled using continuous distributions such as Gaussians or Gaussian Mixture Models (GMMs).
Why Gaussian Emission HMMs Fit Speech Data
In speech tagging or speech recognition, the observed data are acoustic features, not words.
Hidden Markov Models require an emission model to represent:
P(observation | state)
- In text POS tagging → observations are discrete words
- In speech tagging → observations are continuous feature vectors
Therefore, Gaussian (or Gaussian Mixture) distributions are ideal for modeling continuous acoustic data.
Gaussian-emission HMMs model:
P(xt | statet)
where xt is a continuous acoustic feature vector.
HMMs rely on local Markov assumptions, unlike deep contextual models.
HMMs underperform because they rely on short-context Markov assumptions, while neural models capture long-range and global linguistic information.
HMM-based POS taggers rely on the Markov assumption, typically using bigram or trigram tagging. This means the POS tag at position t depends only on a very limited local context.
P(tt | tt−1) or P(tt | tt−1, tt−2)
In other words, Hidden Markov Models (HMMs) assume:
- The current POS tag depends only on a limited number of previous tags (usually one in bigram HMMs, two in trigram HMMs).
- They cannot look far ahead or far behind in a sentence.
Example:
“The book you gave me yesterday was interesting.”
To correctly tag the word “was”, the model benefits from understanding long-distance syntactic relationships. However, HMMs cannot effectively capture such long-range dependencies.
Why neural models perform better
Modern neural POS taggers such as BiLSTM, Transformer, and BERT:
- Capture long-range dependencies across the sentence
- Use bidirectional context (both left and right)
- Learn character-level and subword features
- Handle unknown and rare words more effectively
No comments:
Post a Comment