In K-NN, is the query time longer than the training time?
Intro to ML: K-Nearest Neighbours 2024

Quiz
•
Computers
•
University
•
Medium

Josiah Wang
Used 54+ times
FREE Resource
10 questions
Show all answers
1.
MULTIPLE CHOICE QUESTION
1 min • 1 pt
Yes
No
Answer explanation
Recall K-NN algorithms are lazy learners!
2.
MULTIPLE CHOICE QUESTION
1 min • 1 pt
James is trying to decide which machine learning algorithm to use for his project. Which of the following options is true about the k-NN algorithm?
It can only be used for classification
It can only be used for regression
It can be used for both classification and regression
Answer explanation
Regression: considers the VALUE in majority of k neighbours
Classification: considers the CLASS in majority of k neighbours
3.
MULTIPLE CHOICE QUESTION
1 min • 1 pt
Which of the following is true about Manhattan distances?
It can be used for continuous variables
It can be used for categorical variables
It can be used for categorical as well as continuous variables
Answer explanation
Manhattan distance is the L1-norm and therefore cannot be used for categorical variables. Hamming distance would be a good choice for categorical variables.
4.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
Poppy is working on a machine learning project using k-NN and has a dataset with a lot of noise. What are the appropriate things to do with k-NN in this situation?
Increase the value of K
Decrease the value of K
K does not depend on the noise
None of these
Answer explanation
Recall when K=1 the decision boundaries perfectly match the data and its noise. This will result in overfitting. Increasing K makes K-NN more resilient to overfitting.
5.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
In K-NN, what is the effect of increasing/decreasing the value of K?
The boundary becomes smoother with increasing values of K
Smoothness of the boundary does not depend on the value of K
The boundary becomes smoother with decreasing values of K
None of these
Answer explanation
Think about what happens when K=1. Increasing K results in K-NN being less effected by noise as a large number of neighbours are considered when making a decision.
6.
MULTIPLE CHOICE QUESTION
1 min • 1 pt
For embedded applications (i.e. running on a smartphone), what is the most appropriate family of algorithm?
Eager learners
Lazy learners
Answer explanation
Eager learners are desirable as all the heavy computation occurs at training time. Therefore the algorithm is time + compute efficient at inference time.
7.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
Given d the distance between a point of the dataset and the query point, which of the following weight functions is appropriate for Distance-Weighted k-NN?
w = exp( -d )
w = log ( min ( 0.25 * d, 1 ) )
w = -d
Answer explanation
Recall that the weight you assign is inversely correlated with the distance metric. Both '-d' and 'log(min(0.25*d, 1)' yield negative distances which does not make sense. exp(-d) acts as a good distance metric as it increases exponentially with distance. This favours points close by and quickly ignores points far away.
Create a free account and access millions of resources
Similar Resources on Quizizz
10 questions
Java Arrays

Quiz
•
University
15 questions
Quiz 3 - Grade 5 (Filmora)

Quiz
•
5th Grade - University
15 questions
ANALYSIS OF ALGORITHMS

Quiz
•
University
15 questions
Quiz 1

Quiz
•
University
10 questions
Classification_G1

Quiz
•
University
5 questions
Week 10 - Machine Learning and AI

Quiz
•
University
15 questions
WS2324 S2 & S10 Formative Assessment

Quiz
•
University
10 questions
Unit 7 - Formatting Cells and Ranges

Quiz
•
5th Grade - University
Popular Resources on Quizizz
15 questions
Character Analysis

Quiz
•
4th Grade
17 questions
Chapter 12 - Doing the Right Thing

Quiz
•
9th - 12th Grade
10 questions
American Flag

Quiz
•
1st - 2nd Grade
20 questions
Reading Comprehension

Quiz
•
5th Grade
30 questions
Linear Inequalities

Quiz
•
9th - 12th Grade
20 questions
Types of Credit

Quiz
•
9th - 12th Grade
18 questions
Full S.T.E.A.M. Ahead Summer Academy Pre-Test 24-25

Quiz
•
5th Grade
14 questions
Misplaced and Dangling Modifiers

Quiz
•
6th - 8th Grade