
SummerSchool-Quiz8
Authored by Irfan Ahmad
Computers
University
Used 2+ times

AI Actions
Add similar questions
Adjust reading levels
Convert to real-world scenario
Translate activity
More...
Content View
Student View
9 questions
Show all answers
1.
MULTIPLE CHOICE QUESTION
30 sec • 1 pt
Transformer models do not have recurrent units but can still perform sequence modeling.
True
False
2.
MULTIPLE CHOICE QUESTION
45 sec • 1 pt
As the number of training examples goes to infinity, your model will have:
Low bias
High Bias
Same Bias
Depends on the model’s variance
3.
MULTIPLE CHOICE QUESTION
1 min • 1 pt
Compared to the encoder-decoder model which does not use an attention mechanism, we expect the attention model to have the greatest advantage when:
The input sequence length is large.
The input sequence length is small.
The vocabulary size is large.
The vocabulary size is small.
4.
MULTIPLE CHOICE QUESTION
1 min • 1 pt
You have a friend whose mood is heavily dependent on the current and past few days’ weather. You’ve collected data for the past 365 days on the weather, which you represent as a sequence as x<1>, …, x<365>. You’ve also collected data on your friend’s mood, which you represent as y<1>, …, y<365>. You’d like to build a model to map from x→y. Should you use a Unidirectional RNN or Bidirectional RNN for this problem?
Bidirectional RNN, because this allows the prediction of mood on day t to take into account more information
Bidirectional RNN, because this allows backpropagation to compute more accurate gradients
Unidirectional RNN, because the value of y<t> depends only on x<1>,…,x<t>, but not on x<t+1>,…,x<365>
Unidirectional RNN, because the value of y depends only on x<t> , and not other days
5.
MULTIPLE SELECT QUESTION
1 min • 2 pts
In beam search, if you increase the beam width, which of the following would you expect to be true?
Beam search will run more slowly
Beam search will use up more memory
Beam search will generally find better solutions
Beam search will converge after fewer steps
Beam search will run much faster as more options can be considered
6.
MULTIPLE CHOICE QUESTION
45 sec • 1 pt
How does decoder module of the transformer model avoid seeing the tokens that do not appear yet in output sequence?
Multi-head attention
Positional encoding
Self attention
Masking future positions
7.
MULTIPLE CHOICE QUESTION
45 sec • 1 pt
Which concept in transformer allows for inducing the sequence information in input tokens:
Multi-head attention
Positional encoding
Self attention
Masking future positions before the softmax step
Access all questions and much more by creating a free account
Create resources
Host any resource
Get auto-graded reports

Continue with Google

Continue with Email

Continue with Classlink

Continue with Clever
or continue with

Microsoft
%20(1).png)
Apple
Others
Already have an account?
Similar Resources on Wayground
10 questions
Teste python- Unesp-Jaboticabal
Quiz
•
University
14 questions
IOTA : Web Development Bootcamp Quiz 2
Quiz
•
University
11 questions
What is a computer
Quiz
•
KG - University
12 questions
PSSI Chp 1 dan 2
Quiz
•
University
10 questions
Server Administration- Quiz 1
Quiz
•
12th Grade - University
11 questions
Computer Science (1-9) - Identifying & Preventing Threats
Quiz
•
University
14 questions
Semi-Long Quiz#1 - SAM, NW3A-3E
Quiz
•
University
12 questions
CIS1103 Week 9_10 CLO3 Part 1
Quiz
•
University
Popular Resources on Wayground
15 questions
Fractions on a Number Line
Quiz
•
3rd Grade
20 questions
Equivalent Fractions
Quiz
•
3rd Grade
25 questions
Multiplication Facts
Quiz
•
5th Grade
22 questions
fractions
Quiz
•
3rd Grade
20 questions
Main Idea and Details
Quiz
•
5th Grade
20 questions
Context Clues
Quiz
•
6th Grade
15 questions
Equivalent Fractions
Quiz
•
4th Grade
20 questions
Figurative Language Review
Quiz
•
6th Grade