Tuesday, May 12, 2020

Machine Learning Exam Questions TRUE or FALSE 03

Machine learning quiz questions TRUE or FALSE with answers, important machine learning interview questions for data science, Top 3 machine learning question set


Machine Learning TRUE / FALSE Questions - SET 03


1. MAP estimates are equivalent to the ML estimates when the prior used in the MAP is a uniform prior over the parameter space.

(a) TRUE                                                   (b) FALSE

View Answer

Answer: TRUE

The only difference between MLE and MAP is, in MAP the prior probability P(θ) is included. In simpler terms, the likelihood is weighted with prior probability to become MAP. We can say MLE is a special case of MAP when prior follows a uniform distribution.

Uniform prior is about assigning equal weights everywhere like a constant.
Hence, MAP estimates are equivalent to the MLE when the prior used in the MAP is a uniform prior over the parameter space.
[Refer for more: MLE Vs MAP] 

2. Because decision trees learn to classify discrete-valued outputs instead of real-valued functions it is impossible for them to overfit.

(a) TRUE                                                   (b) FALSE

View Answer

Answer: FALSE
Overfitting is possible in decision trees. If a decision tree is fully grown, it may lose some generalization capability. This is a phenomenon known as overfitting.
Not just a decision tree, (almost) every ML algorithm is prone to overfitting.

What is overfitting?

Over-fitting is the phenomenon in which the learning system tightly fits the given training data so much that it would be inaccurate in predicting the outcomes of the untrained data. One simple way to understand this is to compare the accuracy of your model w.r.t. to training set and test set. If there is a huge difference between them, then your model has achieved overfitting.

Pruning is the method to solve overfitting in decision trees.


3. If P(A|B) = P(A) then P(A ∩ B) = P(A)P(B).

(a) TRUE                                                   (b) FALSE

View Answer

Answer: TRUE
The joint probability of the random variables A and B together are actually separable into the product of their individual probabilities. This states that the probability of any outcome of A and any outcome of B occurring simultaneously is the product of those individual probabilities.

P(A|B) = P(A) – this implies that A is independent of B.
Conditional probability P(A|B) = P(A ∩ B)/P(B)
Multiplication rule: P(A ∩ B) = P(A|B)P(B).
If P(A|B) = P(A), then P(A ∩ B) = P(A)P(B).


**********************

Related links:


top 5 questions in machine learning

quiz questions for data scientists

data science quiz online

online quiz questions on machine learning

true or false quiz on machine learning and data science

data science interview questions

data science previous question papers

machine learning multiple choice questions

test on machine learning skills

top 5 machine learning interview questions

machine learning exam questions

decision trees, maximum-a-posteriori, bayes theorem


No comments:

Post a Comment

Featured Content

Multiple choice questions in Natural Language Processing Home

MCQ in Natural Language Processing, Quiz questions with answers in NLP, Top interview questions in NLP with answers Multiple Choice Que...

All time most popular contents

data recovery