DSBC - Attention Is All You Need

DSBC - Attention Is All You Need

University - Professional Development

11 Qs

quiz-placeholder

Similar activities

Quiz on Human Activity Recognition Using Deep Learning

Quiz on Human Activity Recognition Using Deep Learning

University

15 Qs

Isometric Draw

Isometric Draw

9th Grade - University

11 Qs

Math Fractions

Math Fractions

3rd Grade - University

11 Qs

Sqrt Questions

Sqrt Questions

11th Grade - University

11 Qs

L05 - Рекуррентные сети

L05 - Рекуррентные сети

University

10 Qs

Ch 1 part 2

Ch 1 part 2

University

6 Qs

Linear Inequality Graphs

Linear Inequality Graphs

9th Grade - University

10 Qs

DSBC - Attention Is All You Need

DSBC - Attention Is All You Need

Assessment

Quiz

Mathematics

University - Professional Development

Hard

Created by

Oussama Tchita

Used 5+ times

FREE Resource

11 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

For what RNN is used and achieve the best results?

Speech and image recognition

Handwriting and image recognition

Financial predictions

Handwriting and speech recognition

2.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

What is the basic concept of Recurrent Neural Network?

Use recurrent features from dataset to find the best answers.

Use a loop between inputs and outputs in order to achieve the better prediction.

Use loops between the most important features to predict next output.

Use previous inputs to find the next output according to the training set.

3.

MULTIPLE SELECT QUESTION

2 mins • 1 pt

What architecture represents many-to-many RNNs ?

Media Image
Media Image
Media Image
Media Image

4.

MULTIPLE SELECT QUESTION

2 mins • 1 pt

When RNN sequences are too long, what is most likely to happen ?

the model is prone to ignoring the hidden states with the highest gradients.

the model is prone to mixing the content of distant positions it with following positions’ content.

the model is prone to overfitting easily as exploding gradients hinder efficient back-propagation.

the model is prone to forgetting the content of distant positions in sequence.

5.

MULTIPLE CHOICE QUESTION

3 mins • 1 pt

Media Image

The multi-head attention block is fed three matrices named the Values (V), the Keys (K) and the Query (Q). Which of the following statements is correct ?

V and K are outputted from the input embedding and Q from the output embedding.

Q and K are outputted from the input embedding and V from the output embedding.

V is outputted from the input embedding and K and Q from the output embedding.

Q is outputted from the input embedding and K and V from the output embedding.

6.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

The formula summarizing the multi-head attention operations is:

7.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

The self-attention mechanism is permutation invariant.

True

False

Create a free account and access millions of resources

Create resources
Host any resource
Get auto-graded reports
or continue with
Microsoft
Apple
Others
By signing up, you agree to our Terms of Service & Privacy Policy
Already have an account?