POS Tagging - Viterbi Decoding
After completing the Viterbi experiment, you understand that the algorithm's core principle is based on which mathematical concept?
In the simulation, when filling the Viterbi table for the second word onwards, what is the correct formula for computing V[i][j]?
Based on your experience with the simulation, what happens when you encounter a word that has zero emission probability for all tags?
What insight does the 'Show Answer' feature provide about your understanding of the Viterbi computation?
From your experience with different corpora, what can you conclude about the relationship between training data and decoding outcomes?
When using the simulation's hint feature, what key algorithmic insight does it emphasize about Viterbi computation?
Consider the sentence 'Book a park' from Corpus A. If you manually changed the emission probability of P(park|verb) from 0.1 to 0.8, how would this likely affect the Viterbi decoding result?
What computational advantage does the Viterbi algorithm provide compared to exhaustively checking all possible tag sequences?
Based on your simulation experience, why is it crucial to understand both the mathematical foundations and practical implementation of the Viterbi algorithm?