End-user image suppression: Difference between revisions

From Meta, a Wikimedia project coordination wiki
Content deleted Content added
Christiaan (talk | contribs)
Line 19: Line 19:


A default setting for new and anonymous users would need to be chosen, although this may be largely redundant if we use a [[Wikipedia:HTTP_cookie|cookie]].
A default setting for new and anonymous users would need to be chosen, although this may be largely redundant if we use a [[Wikipedia:HTTP_cookie|cookie]].

Alternatively, a screen could be presented before the article is shown asking if this "nudity", "gore", "nudity in medicine" (whatever) images should be shown.


=== User preference system ===
=== User preference system ===

Revision as of 23:35, 23 February 2005

This page deals with the technical implications of implementing end-user controlled suppression of potentially offensive images on Wikimedia projects. For discussion of whether it would be desirable, see desirability of end-user image suppression.

This page originated from a discussion on the WikiEN-I mailing list where some level of consensus was reached between the free-speech and content suppression parties that a technical solution is possible that will go a long way towards satisfying most mainstream positions in the debate.

Goal

The primary goal would be to give end-users the choice as to whether or not they see certain potentially offensive images by default, thus avoiding their offence and hence minimising the urge of editors to participate in self-censorship while also placating those who would seek to discredit Wikimedia projects based on the adult nature of some content (disclaimers aside).

Basic idea

It would primarily involve ensuring that potentially offensive images are categorised, coupled with some combination of the following:

  1. a site-based preference option allowing users to remove, hide or show potentially offensive images by default
  2. a cookie-based preference option allowing users to remove, hide or show potentially offensive images by default
  3. browser-based filtration (when supported in the future)
  4. proxy and IP based filtration for organisations such as primary schools, etc. (this option is already possible and wouldn't need any action on our part)

Initial default setting

A default setting for new and anonymous users would need to be chosen, although this may be largely redundant if we use a cookie.

Alternatively, a screen could be presented before the article is shown asking if this "nudity", "gore", "nudity in medicine" (whatever) images should be shown.

User preference system

An Image Suppression preference section would be created. In this would be a choice to show all images by default. Under that would be the ability to choose amongst categories based on potentially offensive images and then choose to have one of the following happen:

  1. Hide images (this would create a place-holder where images would normally be which the user could then choose to display on a case-by-case basis)
  2. Remove images (this would remove any link or placeholder (but not the image from the server of course))

Categorising images

Tagging images would simply involve ensuring that potentially offensive images have all been given categories as per normal procedures.

Relevant links

See also