POS Tagging - Hidden Markov Model
What is the primary purpose of Part-of-Speech (POS) tagging in natural language processing?
In a Hidden Markov Model, what are the 'hidden' states?
What does the word 'Markov' refer to in Hidden Markov Models?
What are emission probabilities in an HMM for POS tagging?
Why is context important in POS tagging?
In the calculation P(tag₂|tag₁) = count(tag₁, tag₂) / count(tag₁), what does this represent?
What is the main advantage of using statistical methods like HMMs over hand-crafted rules for POS tagging?
In the Viterbi algorithm, what does dynamic programming help achieve?
What is a major limitation of first-order HMMs for POS tagging?