N-Grams Smoothing

What is the main purpose of an N-gram language model?
Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Which of the following is a bigram?
Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Why do we need smoothing in N-gram language models?
Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

What does Add-One (Laplace) Smoothing do to the counts in an N-gram model?
Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

If a bigram never appears in the training corpus, what is its probability in a maximum likelihood estimate (without smoothing)?
Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

What is the effect of smoothing on the probability distribution of N-grams?
Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Which of the following is a limitation of Add-One Smoothing?
Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

How does the vocabulary size (V) affect the denominator in Add-One Smoothing for bigram probabilities?
Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Which of the following is a real-world application of N-gram smoothing?
Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation

Explanation