Bias and Ethics in Datasets

Bias and Ethics in Datasets

Assessment

Interactive Video

Social Studies

11th - 12th Grade

Hard

Created by

Patricia Brown

FREE Resource

The presentation by Yusuke Hirota discusses gender and racial bias in visual question answering (VQA) datasets. It highlights the challenges of vision language tasks, such as image captioning and VQA, and the societal biases present in these datasets. The analysis reveals gender and racial stereotypes, with men often associated with sports and women with motionless activities. Racial bias is evident in the overrepresentation of white individuals. The presentation suggests solutions like automatic screening and ethical instructions to address these biases.

Read more

10 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is one of the main challenges in vision language tasks?

Generating high-quality images

Ensuring data represents the real world

Improving computational speed

Reducing the size of datasets

2.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

In the analysis of gender bias, what was found about the frequency of questions related to men compared to women?

Questions about women were non-existent

Questions about men were about double those about women

Questions about men and women were equal

Questions about women were more frequent

3.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What stereotype is often associated with men in visual question answering datasets?

Gardening

Sports

Fashion

Cooking

4.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

How are racial biases identified in datasets?

By analyzing the color of images

By counting the number of images

By checking the size of the dataset

By examining questions containing race or ethnicity

5.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is a common issue with racial questions in datasets?

They are too complex

They often lack visual grounding

They are too simple

They are always accurate

6.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is one proposed solution to address harmful samples in datasets?

Use automatic screening to remove ungrounded questions

Ignore the harmful samples

Reduce the number of questions

Increase the dataset size

7.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

Why is it important to incorporate ethical instructions in dataset annotation?

To increase the number of annotators

To reduce the cost of annotation

To ensure annotators are aware of biases

To speed up the annotation process

Create a free account and access millions of resources

Create resources
Host any resource
Get auto-graded reports
or continue with
Microsoft
Apple
Others
By signing up, you agree to our Terms of Service & Privacy Policy
Already have an account?