Top
3 Machine Learning Quiz Questions with Answers explanation, Interview
questions on machine learning, quiz questions for data scientist answers
explained, machine learning exam questions, question bank in machine
learning, k-nearest neighbor, decision tree, linear regression
Machine learning Quiz Questions - Set 32
1. Which of the following is true about regularized linear regression model?
a) Increase in regularization parameter (lambda) will make the model to underfit the data and the validation error will go up.
b) Decrease in regularization parameter (lambda) will make the model to overfit the data and the training error go up
c) Increase in regularization parameter (lambda) will make the model to underfit the data and the training error go down
d) All of the above are true
Answer: (a) Increase in regularization parameter (lambda) will make the model to underfit the data and the validation error will go up. Regularization
parameter (tuning parameter) Î», used in the regularization techniques,
controls the impact on bias and variance. As the value of Î» rises, it reduces
the value of coefficients and thus reducing the variance. Till a point, this
increase in Î» is beneficial as it is only reducing the variance (hence
avoiding overfitting), without loosing any important properties in the data.
But after certain value, the model starts loosing important properties,
giving rise to bias in the model and thus underfitting. [Refer here for more.] |
2. Which of the following is a characteristic of decision tree?
a) High variance
b) High bias
c) Smoothness of prediction surfaces
d) Low variance
Answer: (a) High variance A model has high variance if it is very sensitive to (small) changes in the training data. Decision trees are generally unstable considering that a small change in the data set can result in a very different set of splits. This results in high variance. This is mainly due to the hierarchical nature of decision trees, since a change in split points in the initial stages will affect all the subsequent splits. |
3. Let us consider single-link and complete-link hierarchical clustering. In which of these approaches, it is possible for a point to be closer to points in other clusters than the points in its own cluster?
a) It is possible in single-link clustering
b) It is possible in complete-link clustering
c) Both in single-link and complete-link clustering
d) Neither in single-link nor in complete-link values
Answer: (c) Both in single-link and complete-link clustering This is possible in both single-link and complete-link clustering. In the single-link case, an example would be two parallel chains where many points are closer to points in the other chain/cluster than the points in their own cluster. In the complete-link case, this notion is more intuitive due to the clustering constraint (measuring distance between two clusters by the distance between their farthest points). Both single-link (aka single-linkage clustering) and complete-link (aka complete-linkage clustering) are methods of agglomerative hierarchical clustering. In single-link clustering, the similarity of two clusters is the similarity of their most similar members whereas in complete-linkage clustering, the similarity of two clusters is the similarity of their most dissimilar members. |