Jump to content

Wikipedia:Wikipedia Signpost/2024-03-02/Recent research

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by HaeB (talk | contribs) at 21:58, 28 February 2024 (copyedits / this is basically just Neuroscience News quoting the central part of the abstract, I don't think we need to give credit for that - quote them for a more substantial statement instead, and the lead author from the AFP article on potential impact. CC User:Bri). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Recent research

YOUR ARTICLE'S DESCRIPTIVE TITLE HERE

A monthly overview of recent academic research about Wikipedia and other Wikimedia projects, also published as the Wikimedia Research Newsletter.


Online Images Amplify Gender Bias

Reviewed by Bri

A Nature paper titled "Online Images Amplify Gender Bias"[1] studies

"gender associations of 3,495 social categories (such as 'nurse' or 'banker') in more than one million images from Google, Wikipedia and Internet Movie Database (IMDb), and in billions of words from these platforms"

As summarized by Neuroscience News and by AFP:

This pioneering study indicates that online images not only display a stronger bias towards men but also leave a more lasting psychological impact compared to text, with effects still notable after three days.

This was a two-part research paper in which the authors

  • examined text and images from the Internet for gender bias
  • examined the responses of experimental subjects who were exposed to text and images from the Internet

For the first part, images were drawn from Google search results, and tagged with gender by workers recruited via Amazon Mechanical Turk. The reliability of tagging was validated against a "canonical set" of celebrity portraits culled from IMDB and Wikipedia.[supp 1]

The images represented examples of holders of "social categories" (mostly occupations) in a preselected category list; the 22 occupations included immunologist, harpist, hygienist, and intelligence analyst, as examples, all found in WordNet.

Text samples were taken from Google News and gender bias analyzed with word embedding model, a computational natural language processing technique. The news story text was also associated with social categories using automation.

For the second part, an implicit association test (IAT) methodology was used, which supposedly reveals unconscious bias in a timed sorting task. In the researchers' words, "the participant will be fast at sorting in a manner that is consistent with one's latent associations, which is expected to lead to greater cognitive fluency [lower measured sorting times] in one's intuitive reactions." The test measured times when images and text? were presented in sets, whose individuals could be separated both into male/female and into science/liberal arts (based on their Wikipedia biographies). The labeling of text descriptions was performed by other humans recruited via Amazon Mechanical Turk. Both the test subject, and the labelers, were adults from the United States, and the test subjects were screened to be representative of the U.S. population to include a nearly 50/50 male/female split (none self identified as other than those two categories).

Some test subjects were given a task related to occupation-related text prior to the IAT, and some were given a task related to images. The task was either to use Google search to retrieve images of representative individuals in the occupation, or Google search to retrieve a textual description of the occupation. A control group performed an unrelated Google search. Before the IAT was performed, the test subjects were required to indicate on a sliding scale, for each of the occupations, "which gender do you most expect to belong to this category?" The test was performed again a few days later with the same test subjects.

On the second test, subjects exposed to images in the first test had a stronger IAT score for bias than those exposed to text.

The experimental part of the study depends partly on IAT and partly on self-assessment to detect priming, and there are concerns about replicability concerning the priming effect, and the validity and reliability of IAT. Some of the concerns are described at Implicit-association test § Criticism and controversy. It seemed that the authors recognized this in the statement We acknowledge important continuing debate about the reliability of the IAT, and in their own study found that We note, however, that the distribution of participants' implicit bias scores [arrived at with IAT] was less stable across our preregistered studies than the distribution of participants' explicit bias scores, and discounted the implicit bias scores somewhat.

The conclusion drawn by the researchers, based partly but not entirely on the different IAT scores of experimental subjects, was that of the paper title, "images amplify gender bias" – both explicitly as determined by the subject's assignments of occupation to gender on a sliding scale, and implicitly as determined by reaction times measured in the IAT. Combined with the observation that "Each year, people spend less time reading and more time viewing images" that the paper opens with, this forms an "alarming" trend according to the study's lead author (Douglas Guilbeault of UC Berkeley's Haas School of Business), as quoted by AFP on "the potential consequences this can have on reinforcing stereotypes that are harmful, mostly to women, but also to men".

The researchers also determined, apart from experimental subjects, that the Internet – represented singularly by Google News – exhibits a strong gender bias. It was unclear to this reviewer how much of the reported Internet bias is really "Google selection bias". Based on these findings, the authors go on to speculate that "gender biases in multimodal AI may stem in part from the fact that they are trained on public images from platforms such as Google and Wikipedia, which are rifle with gender bias..."

...

Reviewed by ...

...

Reviewed by ...


Briefly

Other recent publications

Other recent publications that could not be covered in time for this issue include the items listed below. Contributions, whether reviewing or summarizing newly published research, are always welcome.

Compiled by ...

===="..."====` From the abstract:

...

"..."

From the abstract:

...

"..."

From the abstract:

...

References

  1. ^ Guilbeault, Douglas; Delecourt, Solène; Hull, Tasker; Desikan, Bhargav Srinivasa; Chu, Mark; Nadler, Ethan (February 14, 2024), "Online Images Amplify Gender Bias", Nature (online ahead of print), doi:10.1038/s41586-024-07068-xOpen access icon
Supplementary references and notes:
  1. ^ the Wikipedia-based Image Text Dataset [1]