C1M3

C1M3

University

10 Qs

quiz-placeholder

Similar activities

Dependent vs Independent Variables

Dependent vs Independent Variables

University

10 Qs

Computer Programming I Finals

Computer Programming I Finals

University

11 Qs

Linux Basics for Beginners

Linux Basics for Beginners

University

15 Qs

C1M4

C1M4

University

10 Qs

Révision ANN

Révision ANN

University

12 Qs

CLC Unit 2 Lesson 8,9 and 10 Quiz

CLC Unit 2 Lesson 8,9 and 10 Quiz

University

12 Qs

Web Application - PHP Repetition

Web Application - PHP Repetition

University

15 Qs

CMDP2063 Unix and C Programming

CMDP2063 Unix and C Programming

University

15 Qs

C1M3

C1M3

Assessment

Quiz

Information Technology (IT)

University

Medium

Created by

Abylai Aitzhanuly

Used 1+ times

FREE Resource

10 questions

Show all answers

1.

MULTIPLE SELECT QUESTION

45 sec • 1 pt

Which of the following are true? (Check all that apply.) Notice that I only list correct options.

Media Image
Media Image
Media Image
Media Image
Media Image

2.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

The tanh activation usually works better than sigmoid activation function for hidden units because the mean of its output is closer to zero, and so it centers the data better for the next layer. True/False?

TRUE

FALSE

3.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Media Image

Media Image
Media Image
Media Image
Media Image

4.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

You are building a binary classifier for recognizing cucumbers (y=1) vs. watermelons (y=0). Which one of these activation functions would you recommend using for the output layer?

ReLU

Leaky ReLU

sigmoid

tanh

5.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Consider the following code:

A = np.random.randn(4,3) B = np.sum(A, axis = 1, keepdims = True)

What will be B.shape?

B.shape = (1, 4)

B.shape = (3, 2)

B.shape = (4, 1)

B.shape = (4, 0)

6.

MULTIPLE SELECT QUESTION

45 sec • 1 pt

Suppose you have built a neural network. You decide to initialize the weights and biases to be zero. Which of the following statements are True? (Check all that apply)

Each neuron in the first hidden layer will perform the same computation. So even after multiple iterations of gradient descent each neuron in the layer will be computing the same thing as other neurons.

Each neuron in the first hidden layer will perform the same computation in the first iteration. But after one iteration of gradient descent they will learn to compute different things because we have “broken symmetry”.

Each neuron in the first hidden layer will compute the same thing, but neurons in different layers will compute different things, thus we have accomplished “symmetry breaking” as described in lecture.

The first hidden layer’s neurons will perform different computations from each other even in the first iteration; their parameters will thus keep evolving in their own way.

7.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Logistic regression’s weights w should be initialized randomly rather than to all zeros, because if you initialize to all zeros, then logistic regression will fail to learn a useful decision boundary because it will fail to “break symmetry”, True/False?

TRUE

FALSE

Create a free account and access millions of resources

Create resources
Host any resource
Get auto-graded reports
or continue with
Microsoft
Apple
Others
By signing up, you agree to our Terms of Service & Privacy Policy
Already have an account?