DSBC - Attention Is All You Need

DSBC - Attention Is All You Need

University - Professional Development

11 Qs

quiz-placeholder

Similar activities

The AI Shield

The AI Shield

University

15 Qs

G.13 & G.14

G.13 & G.14

9th Grade - University

10 Qs

Numerical Analysis 1

Numerical Analysis 1

University

10 Qs

shuckluckmumppump

shuckluckmumppump

University

10 Qs

GRE Review Exercises

GRE Review Exercises

University

12 Qs

Pembagian Bersusun Dua Digit

Pembagian Bersusun Dua Digit

3rd Grade - University

10 Qs

Quiz 2 First Order DE

Quiz 2 First Order DE

University

10 Qs

NATIONAL MATHEMATICS DAY

NATIONAL MATHEMATICS DAY

5th Grade - Professional Development

10 Qs

DSBC - Attention Is All You Need

DSBC - Attention Is All You Need

Assessment

Quiz

Mathematics

University - Professional Development

Practice Problem

Hard

Created by

Oussama Tchita

Used 5+ times

FREE Resource

11 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

For what RNN is used and achieve the best results?

Speech and image recognition

Handwriting and image recognition

Financial predictions

Handwriting and speech recognition

2.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

What is the basic concept of Recurrent Neural Network?

Use recurrent features from dataset to find the best answers.

Use a loop between inputs and outputs in order to achieve the better prediction.

Use loops between the most important features to predict next output.

Use previous inputs to find the next output according to the training set.

3.

MULTIPLE SELECT QUESTION

2 mins • 1 pt

What architecture represents many-to-many RNNs ?

Media Image
Media Image
Media Image
Media Image

4.

MULTIPLE SELECT QUESTION

2 mins • 1 pt

When RNN sequences are too long, what is most likely to happen ?

the model is prone to ignoring the hidden states with the highest gradients.

the model is prone to mixing the content of distant positions it with following positions’ content.

the model is prone to overfitting easily as exploding gradients hinder efficient back-propagation.

the model is prone to forgetting the content of distant positions in sequence.

5.

MULTIPLE CHOICE QUESTION

3 mins • 1 pt

Media Image

The multi-head attention block is fed three matrices named the Values (V), the Keys (K) and the Query (Q). Which of the following statements is correct ?

V and K are outputted from the input embedding and Q from the output embedding.

Q and K are outputted from the input embedding and V from the output embedding.

V is outputted from the input embedding and K and Q from the output embedding.

Q is outputted from the input embedding and K and V from the output embedding.

6.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

The formula summarizing the multi-head attention operations is:

7.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

The self-attention mechanism is permutation invariant.

True

False

Create a free account and access millions of resources

Create resources

Host any resource

Get auto-graded reports

Google

Continue with Google

Email

Continue with Email

Classlink

Continue with Classlink

Clever

Continue with Clever

or continue with

Microsoft

Microsoft

Apple

Apple

Others

Others

By signing up, you agree to our Terms of Service & Privacy Policy

Already have an account?