Top
3 Machine Learning Quiz Questions with Answers explanation, Interview
questions on machine learning, quiz questions for data scientist answers
explained, machine learning exam questions, question bank in machine
learning, overfitting in non-parametric machine learning algorithms, decision tree, linear regression
Machine
learning Quiz Questions - Set 25
1. Which of the
following is a disadvantage of non-parametric machine learning algorithms?
a) Capable of
fitting a large number of functional forms (Flexibility)
b) Very fast to
learn (Speed)
c) More of a risk
to overfit the training data (Overfitting)
d) They do not
require much training data
Answer: (c) More of a risk to overfit the training data
Overfitting
happens when a model learns the detail and noise in the training data to the
extent that it negatively impacts the performance of the model on new data. Overfitting
is more likely with nonparametric and nonlinear models that have more
flexibility when learning a target function. For example, decision trees are
a nonparametric machine learning algorithm that is very flexible and is
subject to overfitting training data.
[For more, please
refer Overfitting and Underfitting With Machine Learning Algorithms]
|
2. A decision tree
has low training error and a large test error. What is the possible problem?
a) Decision tree is
too shallow
b) Learning rate
too high
c) There is too
much training data
d) Decision tree is
overfitting
Answer: (d) Decision tree is overfitting
Overfitting
causes low training error. Overfitting means that the model predicts the
(training) data too well. It is too good to be true. If the new data point
comes in, the prediction may be wrong.
Pruning can help
in reducing the complexity of the final classifier, and hence improves
predictive accuracy by the reduction of overfitting.
|
3. Suppose we have
a regularized linear regression model. What is the effect of increasing λ on
bias and variance?
a) Increases bias,
increases variance
b) Increases bias,
decreases variance
c) Decreases bias,
increases variance
d) Decreases bias,
decreases variance
Answer: (b) Increases bias, decreases variance
Increasing λ
increases bias and decreases variance
Regularized
regression
It is a type of
regression where the coefficient estimates are constrained to zero. The
magnitude (size) of coefficients, as well as the magnitude of the error term
are penalized. Complex models are discouraged, primarily to avoid
overfitting. In other words, this technique discourages learning a more
complex or flexible model, so as to avoid the risk of overfitting. [For more
refer here – regularized regression, ]
and [Refer here - regularization ]
Type of
regularized regression
Ridge regression
(L2 regularization)
Lasso regression
(L1 regularization)
|
**********************
Related links:
List the type of regularized regression
Multiple choice quiz questions in machine learning
What is regularized linear regression
low training error vs large test error in decision tree
What is the disadvantage of non-parametric machine learning algorithms