Jump to content

Wikipedia:Wikipedia Signpost/2024-03-02/Recent research: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
copyedits / this is basically just Neuroscience News quoting the central part of the abstract, I don't think we need to give credit for that - quote them for a more substantial statement instead, and the lead author from the AFP article on potential impact. CC User:Bri
Line 28: Line 28:
=== Online Images Amplify Gender Bias ===
=== Online Images Amplify Gender Bias ===
:''Reviewed by [[User:Bri|Bri]]''
:''Reviewed by [[User:Bri|Bri]]''
"Online Images Amplify Gender Bias" by Douglas Guilbeault et al. in ''Nature''<ref>{{citation|journal=[[Nature (journal)|Nature]]|type=online ahead of print|title=Online Images Amplify Gender Bias|author1-first=Douglas |author1-last=Guilbeault |author2-first=Solène |author2-last=Delecourt |author3-first=Tasker |author3-last=Hull |author4-first=Bhargav Srinivasa |author4-last=Desikan |author5-first=Mark |author5-last=Chu |author6-first=Ethan |author6-last=Nadler |date= February 14, 2024|doi=10.1038/s41586-024-07068-x}}{{open access}}</ref>
A ''[[Nature (journal)|Nature]]'' paper titled "Online Images Amplify Gender Bias"<ref>{{citation|journal=[[Nature (journal)|Nature]]|type=online ahead of print|title=Online Images Amplify Gender Bias|author1-first=Douglas |author1-last=Guilbeault |author2-first=Solène |author2-last=Delecourt |author3-first=Tasker |author3-last=Hull |author4-first=Bhargav Srinivasa |author4-last=Desikan |author5-first=Mark |author5-last=Chu |author6-first=Ethan |author6-last=Nadler |date= February 14, 2024|doi=10.1038/s41586-024-07068-x}}{{open access}}</ref> studies
<blockquote style="padding-left:1.0em; padding-right:1.0em; background-color:#eaf8f4;">

{{blockquote|examines "gender associations of 3,495 social categories (such as 'nurse' or 'banker') in more than one million images from Google, Wikipedia and Internet Movie Database (IMDb), and in billions of words from these platforms"|source=Neuroscience News}}
"gender associations of 3,495 social categories (such as 'nurse' or 'banker') in more than one million images from Google, Wikipedia and Internet Movie Database (IMDb), and in billions of words from these platforms"
</blockquote>

Reviewed at ''Neuroscience News'': https://neurosciencenews.com/gender-bias-images-25615/ and by AFP: [https://www.todayonline.com/world/online-images-reinforce-gender-stereotypes-more-text-study-2363301] <!-- not (yet) found at AFP.com -->
As summarized by [https://neurosciencenews.com/gender-bias-images-25615/ ''Neuroscience News''] and by AFP:
<blockquote style="padding-left:1.0em; padding-right:1.0em; background-color:#eaf8f4;">
This pioneering study indicates that online images not only display a stronger bias towards men but also leave a more lasting psychological impact compared to text, with effects still notable after three days.
</blockquote>


This was a two-part research paper in which the authors
This was a two-part research paper in which the authors
Line 44: Line 47:
Text samples were taken from Google News and gender bias analyzed with [[word embedding]] model, a computational natural language processing technique. The news story text was also associated with social categories using automation.
Text samples were taken from Google News and gender bias analyzed with [[word embedding]] model, a computational natural language processing technique. The news story text was also associated with social categories using automation.


For the second part, an [[implicit association test]] (IAT) methodology was used, which supposedly reveals unconscious bias in a timed sorting task. In the researchers' words, {{tq|"the participant will be fast at sorting in a manner that is consistent with one's latent associations, which is expected to lead to greater cognitive fluency [lower measured sorting times] in one's intuitive reactions."}} The test measured times when images {{highlight|and text?}} were presented in sets, whose individuals could be separated both into male/female and into science/liberal arts (based on their Wikipedia biographies). The labeling of text descriptions was performed by other humans recruited via [[Amazon Mechanical Turk]]. Both the test subject, and the labelers, were adults from the United States, and the test subjects were screened to be representative of the U.S. population to include a nearly 50/50 male/female split (none self identified as other than those two categories).
For the second part,
an [[implicit association test]] (IAT) methodology was used, which supposedly reveals unconscious bias in a timed sorting task, in the researchers' words "the participant will be fast at sorting in a manner that is consistent with one's latent associations, which is expected to lead to greater cognitive fluency [lower measured sorting times] in one's intuitive reactions." The test measured times when images {{highlight|and text?}} were presented in sets, whose individuals could be separated both into male/female and into science/liberal arts (based on their Wikipedia biographies). The labeling of text descriptions was performed by other humans recruited via [[Amazon Mechanical Turk]]. Both the test subject, and the labelers, were adults from the United States, and the test subjects were screened to be representative of the U.S. population to include a nearly 50/50 male/female split (none self identified as other than those two categories).


Some test subjects were given a task related to occupation-related text prior to the IAT, and some were given a task related to images. The task was either to use Google search to retrieve images of representative individuals in the occupation, or Google search to retrieve a textual description of the occupation. A control group performed an unrelated Google search. Before the IAT was performed, the test subjects were required to indicate on a sliding scale, for each of the occupations, "which gender do you most expect to belong to this category?" The test was performed again a few days later with the same test subjects.
Some test subjects were given a task related to occupation-related text prior to the IAT, and some were given a task related to images. The task was either to use Google search to retrieve images of representative individuals in the occupation, or Google search to retrieve a textual description of the occupation. A control group performed an unrelated Google search. Before the IAT was performed, the test subjects were required to indicate on a sliding scale, for each of the occupations, "which gender do you most expect to belong to this category?" The test was performed again a few days later with the same test subjects.
Line 53: Line 55:
The experimental part of the study depends partly on IAT and partly on self-assessment to detect [[Priming (psychology)|priming]], and there are concerns about replicability concerning the priming effect, and the validity and reliability of IAT. Some of the concerns are described at {{section link|Implicit-association test|Criticism and controversy}}. It seemed that the authors recognized this in the statement {{tq|We acknowledge important continuing debate about the reliability of the IAT}}, and in their own study found that {{tq|We note, however, that the distribution of participants' implicit bias scores [arrived at with IAT] was less stable across our preregistered studies than the distribution of participants' explicit bias scores}}, and discounted the implicit bias scores somewhat.
The experimental part of the study depends partly on IAT and partly on self-assessment to detect [[Priming (psychology)|priming]], and there are concerns about replicability concerning the priming effect, and the validity and reliability of IAT. Some of the concerns are described at {{section link|Implicit-association test|Criticism and controversy}}. It seemed that the authors recognized this in the statement {{tq|We acknowledge important continuing debate about the reliability of the IAT}}, and in their own study found that {{tq|We note, however, that the distribution of participants' implicit bias scores [arrived at with IAT] was less stable across our preregistered studies than the distribution of participants' explicit bias scores}}, and discounted the implicit bias scores somewhat.


The conclusion drawn by the researchers, based partly but not entirely on the different IAT scores of experimental subjects, was that of the paper title, "images amplify gender bias" – both explicitly as determined by the subject's assignments of occupation to gender on a sliding scale, and implicitly as determined by reaction times measured in the IAT. The researchers also determined, apart from experimental subjects, that the Internet represented singularly by Google News exhibits a strong gender bias. It was unclear to this reviewer how much of the reported Internet bias is really "Google selection bias". Based on these findings, the authors go on to speculate that "gender biases in multimodal AI may stem in part from the fact that they are trained on public images from platforms such as Google and Wikipedia, which are rifle with gender bias..."
The conclusion drawn by the researchers, based partly but not entirely on the different IAT scores of experimental subjects, was that of the paper title, "images amplify gender bias" – both explicitly as determined by the subject's assignments of occupation to gender on a sliding scale, and implicitly as determined by reaction times measured in the IAT. Combined with the <!--thinly referenced--> observation that {{tq|"Each year, people spend less time reading and more time viewing images"}} that the paper opens with, this forms an {{tq|"alarming"}} trend according to the study's lead author ([[Douglas Guilbeault]] of UC Berkeley's [[Haas School of Business]]), as quoted by [https://www.todayonline.com/world/online-images-reinforce-gender-stereotypes-more-text-study-2363301 AFP] <!-- not (yet) found at AFP.com --> on {{tq|"the potential consequences this can have on reinforcing stereotypes that are harmful, mostly to women, but also to men"}}.

The researchers also determined, apart from experimental subjects, that the Internet – represented singularly by Google News – exhibits a strong gender bias. It was unclear to this reviewer how much of the reported Internet bias is really "Google selection bias". Based on these findings, the authors go on to speculate that "gender biases in multimodal AI may stem in part from the fact that they are trained on public images from platforms such as Google and Wikipedia, which are rifle with gender bias..."


=== ... ===
=== ... ===
Line 70: Line 74:
:<small>''Compiled by ...''</small>
:<small>''Compiled by ...''</small>


===="..."====
===="..."====`
From the abstract:
From the abstract:
<blockquote style="padding-left:1.0em; padding-right:1.0em; background-color:#eaf8f4;">
<blockquote style="padding-left:1.0em; padding-right:1.0em; background-color:#eaf8f4;">
Line 95: Line 99:
{{Wikipedia:Wikipedia Signpost/Templates/Signpost-block-end-v2}}
{{Wikipedia:Wikipedia Signpost/Templates/Signpost-block-end-v2}}
{{Wikipedia:Wikipedia Signpost/Templates/Signpost-article-end-v2}}
{{Wikipedia:Wikipedia Signpost/Templates/Signpost-article-end-v2}}
<noinclude>{{Wikipedia:Wikipedia Signpost/Templates/Signpost-article-comments-end||2024-02-13|}}</noinclude>
<noinclude>{{Wikipedia:Wikipedia Signpost/Templates/Signpost-article-comments-end||2024-01-31|}}</noinclude>

Revision as of 21:58, 28 February 2024

Recent research

YOUR ARTICLE'S DESCRIPTIVE TITLE HERE

A monthly overview of recent academic research about Wikipedia and other Wikimedia projects, also published as the Wikimedia Research Newsletter.


Online Images Amplify Gender Bias

Reviewed by Bri

A Nature paper titled "Online Images Amplify Gender Bias"[1] studies

"gender associations of 3,495 social categories (such as 'nurse' or 'banker') in more than one million images from Google, Wikipedia and Internet Movie Database (IMDb), and in billions of words from these platforms"

As summarized by Neuroscience News and by AFP:

This pioneering study indicates that online images not only display a stronger bias towards men but also leave a more lasting psychological impact compared to text, with effects still notable after three days.

This was a two-part research paper in which the authors

  • examined text and images from the Internet for gender bias
  • examined the responses of experimental subjects who were exposed to text and images from the Internet

For the first part, images were drawn from Google search results, and tagged with gender by workers recruited via Amazon Mechanical Turk. The reliability of tagging was validated against a "canonical set" of celebrity portraits culled from IMDB and Wikipedia.[supp 1]

The images represented examples of holders of "social categories" (mostly occupations) in a preselected category list; the 22 occupations included immunologist, harpist, hygienist, and intelligence analyst, as examples, all found in WordNet.

Text samples were taken from Google News and gender bias analyzed with word embedding model, a computational natural language processing technique. The news story text was also associated with social categories using automation.

For the second part, an implicit association test (IAT) methodology was used, which supposedly reveals unconscious bias in a timed sorting task. In the researchers' words, "the participant will be fast at sorting in a manner that is consistent with one's latent associations, which is expected to lead to greater cognitive fluency [lower measured sorting times] in one's intuitive reactions." The test measured times when images and text? were presented in sets, whose individuals could be separated both into male/female and into science/liberal arts (based on their Wikipedia biographies). The labeling of text descriptions was performed by other humans recruited via Amazon Mechanical Turk. Both the test subject, and the labelers, were adults from the United States, and the test subjects were screened to be representative of the U.S. population to include a nearly 50/50 male/female split (none self identified as other than those two categories).

Some test subjects were given a task related to occupation-related text prior to the IAT, and some were given a task related to images. The task was either to use Google search to retrieve images of representative individuals in the occupation, or Google search to retrieve a textual description of the occupation. A control group performed an unrelated Google search. Before the IAT was performed, the test subjects were required to indicate on a sliding scale, for each of the occupations, "which gender do you most expect to belong to this category?" The test was performed again a few days later with the same test subjects.

On the second test, subjects exposed to images in the first test had a stronger IAT score for bias than those exposed to text.

The experimental part of the study depends partly on IAT and partly on self-assessment to detect priming, and there are concerns about replicability concerning the priming effect, and the validity and reliability of IAT. Some of the concerns are described at Implicit-association test § Criticism and controversy. It seemed that the authors recognized this in the statement We acknowledge important continuing debate about the reliability of the IAT, and in their own study found that We note, however, that the distribution of participants' implicit bias scores [arrived at with IAT] was less stable across our preregistered studies than the distribution of participants' explicit bias scores, and discounted the implicit bias scores somewhat.

The conclusion drawn by the researchers, based partly but not entirely on the different IAT scores of experimental subjects, was that of the paper title, "images amplify gender bias" – both explicitly as determined by the subject's assignments of occupation to gender on a sliding scale, and implicitly as determined by reaction times measured in the IAT. Combined with the observation that "Each year, people spend less time reading and more time viewing images" that the paper opens with, this forms an "alarming" trend according to the study's lead author (Douglas Guilbeault of UC Berkeley's Haas School of Business), as quoted by AFP on "the potential consequences this can have on reinforcing stereotypes that are harmful, mostly to women, but also to men".

The researchers also determined, apart from experimental subjects, that the Internet – represented singularly by Google News – exhibits a strong gender bias. It was unclear to this reviewer how much of the reported Internet bias is really "Google selection bias". Based on these findings, the authors go on to speculate that "gender biases in multimodal AI may stem in part from the fact that they are trained on public images from platforms such as Google and Wikipedia, which are rifle with gender bias..."

...

Reviewed by ...

...

Reviewed by ...


Briefly

Other recent publications

Other recent publications that could not be covered in time for this issue include the items listed below. Contributions, whether reviewing or summarizing newly published research, are always welcome.

Compiled by ...

===="..."====` From the abstract:

...

"..."

From the abstract:

...

"..."

From the abstract:

...

References

  1. ^ Guilbeault, Douglas; Delecourt, Solène; Hull, Tasker; Desikan, Bhargav Srinivasa; Chu, Mark; Nadler, Ethan (February 14, 2024), "Online Images Amplify Gender Bias", Nature (online ahead of print), doi:10.1038/s41586-024-07068-xOpen access icon
Supplementary references and notes:
  1. ^ the Wikipedia-based Image Text Dataset [1]