Advanced Database Management System - Tutorials and Notes: Machine Learning Multiple Choice Questions and Answers 11

Search Engine

Please visit, subscribe and share 10 Minutes Lectures in Computer Science

Thursday, 28 May 2020

Machine Learning Multiple Choice Questions and Answers 11

Top 3 Machine Learning Quiz Questions with Answers explanation, Interview questions on machine learning, quiz questions for data scientist answers explained, machine learning exam questions



Machine learning MCQ - Set 11


1. Given a Neural Net with N input nodes, no hidden layers, one output node, with Entropy Loss and Sigmoid Activation Functions, which of the following algorithms (with the proper hyper-parameters and initialization) can be used to find the global optimum?

a) Stochastic Gradient Descent

b) Mini-Batch Gradient Descent

c) Batch Gradient Descent

d) All of the above


View Answer

Answer: (d) all of the above

All three can be used to find global optimum.

Batch gradient descent - all the training data is taken into consideration to take a single step (one epoch).

Mini-batch gradient descent - We use a batch of a fixed number of training examples which is less than the actual dataset and call it a mini-batch. (process neural net with each mini-batch is one epoch).

Stochastic gradient descent – we consider 1 example at a time to take a single step. (process neural net with 1 example is 1 epoch)

 

2. Say the incidence of a disease D is about 5 cases per 100 people (i.e., P(D) = 0.05). Let Boolean random variable D mean a patient “has disease D” and let Boolean random variable TP stand for "tests positive." Tests for disease D are known to be very accurate in the sense that the probability of testing positive when you have the disease is 0.99, and the probability of testing negative when you do not have the disease is 0.97. What is P(TP), the prior probability of testing positive.

a) 0.0368

b) 0.473

c) 0.078

d) None of the above


View Answer

Answer: (c) 0.078

Given,

            P(D) = 0.05,

            P(TP | D) = 0.99,

            PTP | ¬D) = 0.97,

P(TP) = P(TP | D) P(D) + P(TP | ¬D) PD)

            = (.99)(.05) + (1 - .97)(1 - .05) = 0.078.

 

3. For a Gaussian Bayes classifier, which one of these structural assumptions is the one that most affects the trade-off between underfitting and overfitting:

a) Whether we learn the class centers by Maximum Likelihood or Gradient Descent

b) Whether we assume full class covariance matrices or diagonal class covariance matrices

c) Whether we have equal class priors or priors estimated from the data.

d) Whether we allow classes to have different mean vectors or we force them to share the same mean vector


View Answer

Answer: (b) Whether we assume full class covariance matrices or diagonal class covariance matrices

Diagonal covariance matrix satisfies the naive Bayes assumption.


**********************

Related links:


top 5 questions in machine learning

quiz questions for data scientists

data science quiz online

online quiz questions on machine learning

MCQs on machine learning and data science

data science interview questions

data science previous question papers

machine learning multiple choice questions

test on machine learning skills

top 5 machine learning interview questions

machine learning exam questions

No comments:

Post a comment

Featured Content

Multiple choice questions in Natural Language Processing Home

MCQ in Natural Language Processing, Quiz questions with answers in NLP, Top interview questions in NLP with answers Multiple Choice Que...

All time most popular contents