ML B2 CH8

ML B2 CH8

University

10 Qs

quiz-placeholder

Similar activities

Operating Systems Quick Quiz Module 1

Operating Systems Quick Quiz Module 1

11th Grade - University

10 Qs

Scrum 101: The Introduction

Scrum 101: The Introduction

University

10 Qs

BCSC0006 - Quiz 4 - Linked List & Stacks

BCSC0006 - Quiz 4 - Linked List & Stacks

University

10 Qs

C/C++ C Function

C/C++ C Function

University

12 Qs

Engineering ACW Semester 2 - #5 AI Part 2

Engineering ACW Semester 2 - #5 AI Part 2

University

15 Qs

Chapter 1: IT Infrastructure and Emerging Technology

Chapter 1: IT Infrastructure and Emerging Technology

University

11 Qs

PECs

PECs

University

10 Qs

Principles of User Interface Design - Q1

Principles of User Interface Design - Q1

University

14 Qs

ML B2 CH8

ML B2 CH8

Assessment

Quiz

Computers

University

Practice Problem

Medium

Created by

Jhonston Benjumea

Used 1+ times

FREE Resource

AI

Enhance your content in a minute

Add similar questions
Adjust reading levels
Convert to real-world scenario
Translate activity
More...

10 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the main limitation of the traditional seq2seq model?

It cannot translate short sentences
It uses too many output layers
It encodes input into a fixed-length vector regardless of sentence length
It requires bidirectional input only

Answer explanation

Traditional seq2seq models compress all input into a fixed-length vector, which limits performance on longer inputs.

2.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What does the attention mechanism allow the decoder to use?

Only the last hidden state
Only encoder output embeddings
All encoder hidden states with different importance
Static word vectors

Answer explanation

Attention allows the decoder to focus on all encoder hidden states, assigning different weights to each.

3.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

How is the context vector in attention calculated?

Average of all vectors
Maximum value among vectors
Weighted sum of encoder outputs
Concatenation of last and first vectors

Answer explanation

The context vector is a weighted sum of all encoder output vectors based on attention weights.

4.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Which mathematical operation is used to compute attention scores?

Matrix inverse
Subtraction
Dot product
Hadamard product

Answer explanation

Attention scores are calculated using the dot product between decoder state and encoder outputs.

5.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the role of the softmax function in attention?

To generate output probabilities
To normalize attention scores into weights
To embed characters
To reduce vocabulary size

Answer explanation

Softmax normalizes raw attention scores into a probability distribution over encoder outputs.

6.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What does a Time Attention layer do?

Calculates self-attention only
Aggregates attention across all time steps
Forgets context information
Adds noise to encoder outputs

Answer explanation

Time Attention layers apply attention mechanisms at each time step of the sequence.

7.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is bidirectional RNN used for in attention-based models?

To skip decoding steps
To allow input data to flow only backward
To consider context from both directions of the input
To reduce computation time

Answer explanation

Bidirectional RNNs allow the model to gather information from both past and future inputs.

Access all questions and much more by creating a free account

Create resources

Host any resource

Get auto-graded reports

Google

Continue with Google

Email

Continue with Email

Classlink

Continue with Classlink

Clever

Continue with Clever

or continue with

Microsoft

Microsoft

Apple

Apple

Others

Others

Already have an account?