Search Header Logo

Intro to AI Final Exam Review

Authored by Dio Padilla

Education

University

Used 3+ times

Intro to AI Final Exam Review
AI

AI Actions

Add similar questions

Adjust reading levels

Convert to real-world scenario

Translate activity

More...

    Content View

    Student View

30 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the role of the discriminator_loss and generator_loss variables in a typical GAN training loop?

They are hyperparameters that control the learning rate.

They are used to track the accuracy of the discriminator and generator.

They store the model weights of the discriminator and generator respectively.

They represent the loss values to be optimized during training.

2.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Which optimizer is commonly used for training GANs in TensorFlow?

Adagrad

RMSprop

SGD (Stochastic Gradient Descent)

Adam

3.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the purpose of the tf.GradientTape context in TensorFlow when training a GAN?

It visualizes the training progress.

It automatically updates the model weights.

It saves the model checkpoints during training.

It records operations for automatic diff erentiation

4.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Which loss function is commonly used to train the discriminator in a GAN?

Kullback-Leibler Divergence (KLD)

Categorical Cross-Entropy (CCE)

Binary Cross-Entropy (BCE)

Mean Squared Error (MSE)

5.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Which of the following best describes the relationship between the generator and discriminator in a GAN?

The generator and discriminator cooperate to produce high-quality samples.

The generator and discriminator are independent networks with separate objectives.

The generator tries to maximize the discriminator's loss, while the discriminator tries to minimize its loss.

The generator tries to minimize the discriminator's loss, while the discriminator tries to maximize its loss.

6.

MULTIPLE SELECT QUESTION

45 sec • 1 pt

If the deep learning neural network has the problem of gradient disappearance or gradient explosion, our common solutions are: (Multiple answers)

Regularization

Gradient shear

Random under-sampling

Use Relu Activation function

7.

MULTIPLE SELECT QUESTION

30 sec • 1 pt

What are the regularizations in deep learning? (Multiple answers)

Dropout

Data set enhancement

Integration method

L1 norm and L2 norm

Access all questions and much more by creating a free account

Create resources

Host any resource

Get auto-graded reports

Google

Continue with Google

Email

Continue with Email

Classlink

Continue with Classlink

Clever

Continue with Clever

or continue with

Microsoft

Microsoft

Apple

Apple

Others

Others

Already have an account?