
RNN&Transformers Quiz
Authored by Taha rajeh
Computers
12th Grade
Used 19+ times

AI Actions
Add similar questions
Adjust reading levels
Convert to real-world scenario
Translate activity
More...
Content View
Student View
28 questions
Show all answers
1.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
What is the main challenge associated with training vanilla RNNs on long sequences?
Exploding gradients
Vanishing gradients
Both exploding and vanishing gradients
None of the above
2.
MULTIPLE CHOICE QUESTION
30 sec • 1 pt
What is the main limitation of using one-hot encoding for representing words in an RNN?
a) One-hot encoding does not capture semantic relationships between words.
b) One-hot encoding is inefficient for large vocabularies.
c) One-hot encoding can lead to exploding gradients during training.
Both a) and b).
3.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
What is the key idea behind Recurrent Neural Networks (RNNs)?
RNNs have an internal state that is updated as a sequence is processed.
RNNs use a fixed set of parameters for each time step.
RNNs can only process sequential data.
RNNs are feed-forward neural networks.
4.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
What is the purpose of the "output gate" in an LSTM?
It controls what information goes into the output of the LSTM cell.
It determines how much information goes through the cell state.
It decides what information is added to the cell state.
It updates the cell state.
5.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
What is the purpose of the "forget gate" in an LSTM (Long Short-Term Memory) network?
It determines how much information goes through the cell state.
It decides what information is added to the cell state.
It controls what goes into the output.
It updates the cell state.
6.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
What is the purpose of the "attention mechanism" in sequence-to-sequence models?
To improve the model's ability to capture long-term dependencies.
To reduce the computational complexity of the model.
To allow the model to focus on different parts of the input sequence when generating the output sequence.
To prevent vanishing gradients during training.
7.
MULTIPLE CHOICE QUESTION
2 mins • 1 pt
In the context of sequence to sequence (seq2seq) models, what is the purpose of the "encoder" component?
To generate the output sequence from the input sequence.
To encode the input sequence into a fixed-size vector representation.
To decode the fixed-size vector representation into the output sequence.
To combine the input and output sequences into a single sequence.
Access all questions and much more by creating a free account
Create resources
Host any resource
Get auto-graded reports

Continue with Google

Continue with Email

Continue with Classlink

Continue with Clever
or continue with

Microsoft
%20(1).png)
Apple
Others
Already have an account?