Search Header Logo

DSBC - Attention Is All You Need

Authored by Oussama Tchita

Mathematics

University - Professional Development

Used 5+ times

DSBC - Attention Is All You Need
AI

AI Actions

Add similar questions

Adjust reading levels

Convert to real-world scenario

Translate activity

More...

    Content View

    Student View

11 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

For what RNN is used and achieve the best results?

Speech and image recognition

Handwriting and image recognition

Financial predictions

Handwriting and speech recognition

2.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

What is the basic concept of Recurrent Neural Network?

Use recurrent features from dataset to find the best answers.

Use a loop between inputs and outputs in order to achieve the better prediction.

Use loops between the most important features to predict next output.

Use previous inputs to find the next output according to the training set.

3.

MULTIPLE SELECT QUESTION

2 mins • 1 pt

What architecture represents many-to-many RNNs ?

Media Image
Media Image
Media Image
Media Image

4.

MULTIPLE SELECT QUESTION

2 mins • 1 pt

When RNN sequences are too long, what is most likely to happen ?

the model is prone to ignoring the hidden states with the highest gradients.

the model is prone to mixing the content of distant positions it with following positions’ content.

the model is prone to overfitting easily as exploding gradients hinder efficient back-propagation.

the model is prone to forgetting the content of distant positions in sequence.

5.

MULTIPLE CHOICE QUESTION

3 mins • 1 pt

Media Image

The multi-head attention block is fed three matrices named the Values (V), the Keys (K) and the Query (Q). Which of the following statements is correct ?

V and K are outputted from the input embedding and Q from the output embedding.

Q and K are outputted from the input embedding and V from the output embedding.

V is outputted from the input embedding and K and Q from the output embedding.

Q is outputted from the input embedding and K and V from the output embedding.

6.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

The formula summarizing the multi-head attention operations is:

7.

MULTIPLE CHOICE QUESTION

2 mins • 1 pt

The self-attention mechanism is permutation invariant.

True

False

Access all questions and much more by creating a free account

Create resources

Host any resource

Get auto-graded reports

Google

Continue with Google

Email

Continue with Email

Classlink

Continue with Classlink

Clever

Continue with Clever

or continue with

Microsoft

Microsoft

Apple

Apple

Others

Others

Already have an account?