Search Header Logo

CS4417/9117/9647 Final Exam Review Game

Authored by Marwa Elsayed

Computers

University

Used 243+ times

 CS4417/9117/9647 Final Exam Review Game
AI

AI Actions

Add similar questions

Adjust reading levels

Convert to real-world scenario

Translate activity

More...

    Content View

    Student View

40 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Media Image

Suppose you have a 10000 word vocabulary, and are learning 500-dimensional word embeddings. The GloVe model minimizes this objective: True/False: Xij is the number of times word j appears in the context of word i.

True

False

2.

MULTIPLE SELECT QUESTION

45 sec • 1 pt

Which of these equations do you think should hold for a good word embedding? [Check all that apply)

Media Image
Media Image
Media Image
Media Image

3.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

True/False: Suppose you learn a word embedding for a vocabulary of 60000 words. Then the embedding vectors could be 60000 dimensional, so as to capture the full range of variation and meaning in those words.

True

False

4.

MULTIPLE SELECT QUESTION

45 sec • 1 pt

In beam search, if you increase the beam width B, which of the following would you expect to be true?

Beam search will use up more memory.

Beam search will converge after fewer steps.

Beam search will generally find better solutions (i.e. do a better job maximizing P(y∣x)).

Beam search will run more quickly.

5.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Media Image

Consider using this encoder-decoder model for machine translation.

True/False: This model is a “conditional language model” in the sense that the decoder portion (shown in purple) is modeling the probability of the output sentence y given the input sentence x.

True

False

6.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

In machine translation, if we carry out beam search without using sentence normalization, the algorithm will tend to output overly short translations.


True

False

7.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

A Transformer Network, unlike its predecessors RNNs, GRUs and LSTMs, can process entire sentences all at the same time. (Parallel architecture).

True

False

Access all questions and much more by creating a free account

Create resources

Host any resource

Get auto-graded reports

Google

Continue with Google

Email

Continue with Email

Classlink

Continue with Classlink

Clever

Continue with Clever

or continue with

Microsoft

Microsoft

Apple

Apple

Others

Others

Already have an account?