Major links



Quicklinks


📌 Quick Links
[ DBMS ] [ SQL ] [ DDB ] [ ML ] [ DL ] [ NLP ] [ DSA ] [ PDB ] [ DWDM ] [ Quizzes ]


Wednesday, May 27, 2020

Solved Questions and Answers in NLP 03

Natural language processing quiz questions with answers, NLP true false interview questions, NLP quiz questions for competitive exams



NLP TRUE/FALSE Quiz Questions - SET 03

1. Models that assign probabilities to sequences of words are called Support Vector Models.
(a) TRUE                                          (b) FALSE

View Answer

Answer: (b) FALSE
Models that assign probabilities to sequences of words are called language models or LMs.

2. Number of trigrams in the following sentence is 4; “calculates the similarity between two strings”.
(a) TRUE                                          (b) FALSE

View Answer

Answer: (a) TRUE
There are four trigrams (3-grams) in the given sentence if we do not include the start and end markers. The trigrams are;
“calculates the similarity”, “the similarity between”, “similarity between two” and “between two strings”.

3. We normalize the counts of words in an n-gram model to make the value to fall between 0 and 100.
(a) TRUE                                          (b) FALSE

View Answer

Answer: (b) FALSE
We normalize the counts of words in an n-gram model to make the value to fall between 0 and 1.
We get the maximum likelihood estimation (MLE) for the parameters of an n-gram model by getting counts from a normalize corpus, and normalizing the counts so that they lie between 0 and 1.

4. To calculate the bigram probability of a word wn given the previous word wn-1, we count the occurrence of word sequence “wn-1 wn” and normalize this by the count of wn-1.
(a) TRUE                                          (b) FALSE

View Answer

Answer: (a) TRUE
To compute a particular bigram probability of a word y given a previous word x, we will compute the count of the bigram C(x y) and normalize by the count of unigram C(x).
P(y|x) = count(y x)/count(y)
This is called Maximum Likelihood Estimate (MLE).

5. It is better to compute the probabilities in a language model as log probabilities.
(a) TRUE                                          (b) FALSE

View Answer

Answer: (a) TRUE
Since probabilities are (by definition) less than or equal to probabilities 1, the more probabilities we multiply together, the smaller the product becomes. Multiplying enough n-grams together would result in numerical underflow. By using log probabilities instead of raw probabilities, we get numbers that are not as small.

*************************




Top interview questions in NLP

NLP quiz questions with answers explained

Bigram and trigram language models

Online NLP quiz with solutions




No comments:

Post a Comment

Please visit, subscribe and share 10 Minutes Lectures in Computer Science