Search Header Logo

Round - 3 ( Technical round )

Authored by Hinata Hyuga

Other

University

Used 1+ times

Round - 3 ( Technical round )
AI

AI Actions

Add similar questions

Adjust reading levels

Convert to real-world scenario

Translate activity

More...

    Content View

    Student View

20 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

45 sec • 1 pt

We can categorise the BBN Butterfly, Cosmic Cube, Cedar, and Hypercube machine as

a. MISD

b. SIMD

c. MIMD

d. SISD

2.

MULTIPLE CHOICE QUESTION

45 sec • 1 pt

What is overfitting in the context of machine learning models?

a) Fitting a model with insufficient data

b)Fitting a model too closely to the training data

c)Fitting a model with too few features

d) Fitting a model to the validation set

3.

MULTIPLE CHOICE QUESTION

45 sec • 5 pts

What is feature importance in the context of tree-based models like Random Forests?

a)The importance of including a feature in the dataset

b)The correlation between features

c)The contribution of each feature to the model's predictions

d)The number of times a feature appears in the dataset

4.

MULTIPLE CHOICE QUESTION

45 sec • 5 pts

The function that assigns the symbol a to b in LISP is:

a. (setq b ‘a’)

b. (setq b = ‘a’)

c. (setq b a)

d. (set b = ‘a’)

5.

MULTIPLE CHOICE QUESTION

45 sec • 1 pt

In a neural network, what is the role of the activation function?

a)Defines the learning rate

b)Sets the number of neurons in each layer

c)Introduces non-linearity to the model

d)Controls the initialization of weights

6.

MULTIPLE CHOICE QUESTION

45 sec • 1 pt

What is the difference between bagging and boosting in ensemble learning?

a)Bagging increases model diversity, boosting decreases it

b)Bagging trains models sequentially, boosting trains them in parallel

c)Bagging combines predictions using voting, boosting combines predictions using weighted averaging

d)Bagging trains each model independently, boosting focuses on examples misclassified by previous models

7.

MULTIPLE CHOICE QUESTION

45 sec • 1 pt

What is the purpose of the Expectation-Maximization (EM) algorithm in unsupervised learning?

a)Maximizing the likelihood of the observed data

b)Minimizing the reconstruction error in autoencoders

c)Imputing missing values in a dataset

d)Iteratively estimating parameters for mixture models

Access all questions and much more by creating a free account

Create resources

Host any resource

Get auto-graded reports

Google

Continue with Google

Email

Continue with Email

Classlink

Continue with Classlink

Clever

Continue with Clever

or continue with

Microsoft

Microsoft

Apple

Apple

Others

Others

Already have an account?