
SSL Exam
Authored by Radu ird
Information Technology (IT)
University
Used 7+ times

AI Actions
Add similar questions
Adjust reading levels
Convert to real-world scenario
Translate activity
More...
Content View
Student View
15 questions
Show all answers
1.
MULTIPLE SELECT QUESTION
45 sec • 1 pt
Which are the main differences when applying maximum likelihood for a single Gaussian versus a mixture of Gaussians?
A. Overfitting can only happen for the single Gaussian model
B. The mixture model is prone to singularities
C. ML treatment for the mixture model can be solved in closed form, leading to a simple solution
D. The mixture model may provide a better fit, depending on the dataset
2.
MULTIPLE SELECT QUESTION
45 sec • 1 pt
Regularized least squares for a linear regression model is similar to a Bayesian treatment for regression when:
A. The prior for the weights is a Dirichlet distribution
The prior for the weights is any type of Gaussian
The prior for the weights is a simple zero mean isotropic Gaussian
D. There is no link between regularized regression models and Bayesian models for regression
3.
MULTIPLE SELECT QUESTION
45 sec • 1 pt
Which of the following option(s) is / are true?
A. You need to initialize parameters in PCA
B. You don't need to initialize parameters in PCA
C. PCA can be trapped into local minima problem
D. PCA can't be trapped into local minima problem
4.
MULTIPLE SELECT QUESTION
45 sec • 1 pt
Regarding bias and variance, which of the following statements are true?
A. Models which overfit have a high bias.
B. Models which overfit have a low bias.
C. Models which underfit have a high variance.
D. Models which underfit have a low variance
5.
MULTIPLE SELECT QUESTION
45 sec • 1 pt
Which of the following is true about generative models?
A. They capture the joint probability
B. The perceptron is a generative model
C. Generative models can be used for classification
D. They capture the conditional probability
6.
MULTIPLE SELECT QUESTION
45 sec • 1 pt
Compared to the variance of the Maximum Likelihood Estimate (MLE), the variance of the Maximum A Posteriori (MAP) estimate is:
Higher
Same
Lower
D. it could be any of the above
7.
MULTIPLE SELECT QUESTION
45 sec • 1 pt
How can we demonstrate that a function is a valid kernel?
A. We need to be able to express it as a dot product in a latent feature space, maybe with an infinite number of dimensions
B. We need to be able to express it as a dot product in a latent feature space, but with a finite number of dimensions
C. We show that the Gram matrix contains only positive elements
D. We show that the design matrix is positive semi-definite
Access all questions and much more by creating a free account
Create resources
Host any resource
Get auto-graded reports

Continue with Google

Continue with Email

Continue with Classlink

Continue with Clever
or continue with

Microsoft
%20(1).png)
Apple
Others
Already have an account?