Classification_G1

Classification_G1

University

10 Qs

quiz-placeholder

Similar activities

Intro to ML: K-Nearest Neighbours

Intro to ML: K-Nearest Neighbours

University

9 Qs

klustering

klustering

University

15 Qs

Classification

Classification

University

10 Qs

DATA CLUSTERING ALGORITHM TEST 1

DATA CLUSTERING ALGORITHM TEST 1

University

6 Qs

Session 4

Session 4

University

6 Qs

Intro to ML: The ML Revision Quiz

Intro to ML: The ML Revision Quiz

University

11 Qs

Quiz on Car Automation

Quiz on Car Automation

University

15 Qs

Unsupervised Learning

Unsupervised Learning

University

10 Qs

Classification_G1

Classification_G1

Assessment

Quiz

Computers

University

Hard

Created by

Dr Kumar

FREE Resource

10 questions

Show all answers

1.

MULTIPLE SELECT QUESTION

1 min • 1 pt

Why is PCA sometimes used as a preprocessing step before regression?

To reduce overfitting by removing poorly predictive dimensions.

To make computation faster by reducing the dimensionality of the data

For inference and scientific discovery, we prefer features that are not axis-aligned.

To expose information missing from the input data.

2.

MULTIPLE SELECT QUESTION

1 min • 1 pt

What tends to be true about increasing the k in k-nearest neighbors?

The decision boundary tends to get smoother

As the number of sample points approaches infinity (with n/k → ∞), the error rate approaches less than twice the Bayes risk (assuming training and test points are drawn independently from the same distribution).

The variance tends to increase.

The bias tends to increase.

3.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

Which of the following are true about decision trees?

All the leaves must be pure.

Pruning usually achieves better test accuracy than stopping early.

The tree depth never exceeds O(log n) for n sample points.

They can be used only for classification.

4.

MULTIPLE SELECT QUESTION

1 min • 1 pt

Which of the following are true for k-nearest neighbor classification? (Hint: multiple statements may be true)

It is more likely to overfit with k = 1 (1-NN) than with k = 1,000 (1,000-NN)

The optimal running time to classify a point with k-NN grows linearly with k

In very high dimensions, exhaustively checking every training point is often faster than any widely used competing exact k-NN query algorithm

If you have enough training points drawn from the same distribution as the test points, k-NN can achieve accuracy almost as good as the Bayes decision rule

5.

FILL IN THE BLANK QUESTION

1 min • 1 pt

What is the number of parameters needed to represent a Naive Bayes classifier with n Boolean variables and a Boolean label ?

6.

MULTIPLE CHOICE QUESTION

1 min • 1 pt

Are these statements true or false?

(A) When we use PCA, we need data to be labelled.

(B) PCA extracts the variance structure from high dimensional data such that the variance of projected data is minimized.

True, True

True, False

False, True

False, False

7.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

The following statement is true/false?

Gain ratio is less biased towards tests with many outcomes than InfoGain.

True

False

Create a free account and access millions of resources

Create resources
Host any resource
Get auto-graded reports
or continue with
Microsoft
Apple
Others
By signing up, you agree to our Terms of Service & Privacy Policy
Already have an account?