
DSBC - Attention Is All You Need
Authored by Oussama Tchita
Mathematics
University - Professional Development
Used 5+ times

AI Actions
Add similar questions
Adjust reading levels
Convert to real-world scenario
Translate activity
More...
Content View
Student View
11 questions
Show all answers
1.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
For what RNN is used and achieve the best results?
Speech and image recognition
Handwriting and image recognition
Financial predictions
Handwriting and speech recognition
2.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
What is the basic concept of Recurrent Neural Network?
Use recurrent features from dataset to find the best answers.
Use a loop between inputs and outputs in order to achieve the better prediction.
Use loops between the most important features to predict next output.
Use previous inputs to find the next output according to the training set.
3.
MULTIPLE SELECT QUESTION
2 mins • 1 pt
What architecture represents many-to-many RNNs ?
4.
MULTIPLE SELECT QUESTION
2 mins • 1 pt
When RNN sequences are too long, what is most likely to happen ?
the model is prone to ignoring the hidden states with the highest gradients.
the model is prone to mixing the content of distant positions it with following positions’ content.
the model is prone to overfitting easily as exploding gradients hinder efficient back-propagation.
the model is prone to forgetting the content of distant positions in sequence.
5.
MULTIPLE CHOICE QUESTION
3 mins • 1 pt
The multi-head attention block is fed three matrices named the Values (V), the Keys (K) and the Query (Q). Which of the following statements is correct ?
V and K are outputted from the input embedding and Q from the output embedding.
Q and K are outputted from the input embedding and V from the output embedding.
V is outputted from the input embedding and K and Q from the output embedding.
Q is outputted from the input embedding and K and V from the output embedding.
6.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
The formula summarizing the multi-head attention operations is:
7.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
The self-attention mechanism is permutation invariant.
True
False
Access all questions and much more by creating a free account
Create resources
Host any resource
Get auto-graded reports

Continue with Google

Continue with Email

Continue with Classlink

Continue with Clever
or continue with

Microsoft
%20(1).png)
Apple
Others
Already have an account?