Perhaps it's blindingly obvious but the best solution is to prevent conflict
happening in the first place. OK, that sounds a little trite, but what it
means is to design your platform so it's harder for conflict to occur. We
routinely "design in" safety into engineering projects. Think about roads.
Having everyone drive on the left (or right, depending on which country you
are in) avoids a lot of road crashes. Traffic lights control busy
interactions. There are many things we "design into" roads to ensure that
people can co-exist on the roads without killing each other. Yet we seem to
completely ignore such principles in the social aspects of projects.
For example, if a message in a forum contains words likely to cause offense
or indicate aggression "you stupid moron, you dumb twat", point this out to
the user and ask if they really mean to send it (possibly reminding them
that your Code of Conduct does not allow personal attacks, racist language
or whatever the words suggest is going on, and suggest they rewrite or cool
off for a while). At least then, if they proceed to send it and there are
complaints, then it's clear that they did use those words deliberately aware
of their potential to offend and knowing that it was contrary to the Code of
Conduct). No more "I didn't mean to offend because I tell everyone, even my
best friends, that they are stupid morons". Or you can just quietly divert
that message to a moderator for consideration before it's posted.
In Facebook, we have the protocol of confirming each other as friends, which
then allows us to see one another's postings etc, as well as the ability to
accept someone as a friend (avoiding the offence of refusing friend status)
and still giving you the ability to avoid reading their posts and
restricting which ones of yours they see. Again, this is an example of the
system of allowing you to be "friends" but putting some boundaries on that
relationship. "Good fences makes good neighbours" as they say.
Obviously, the way to "design in" conflict avoidance depends a bit on the
nature of the platform, e.g. forums are different to wikis. For example,
newspapers and blogs increasing disable comments because of the conflicts
that can arise there. You could do something like limit people to only one
comment per topic, this prevents the to-ing and fro-ing between a couple of
people escalating their dispute and forces people to use their one-comment
wisely, hopefully to address the substantive topic and not to criticise
others. Indeed, if you restricted the ability to mention another user name
in a comment, it would become much harder to make personal attacks (works
best if you don't allow anyone to register the username "the" or other
common words though :-))
Now in Wikipedia, we seem to like to do the exact opposite, we seem to
"design in" conflict. Think about "undo". It's is a quick-and-easy
solution
for dealing with vandalism but it's a terrible tool for dealing with a
difference of opinion between good faith editors. By the time one of them
gets blocked by the three-revert rule, you've got conflict well-escalated.
So long as it is more work for people to "talk" about the issue than
"undo",
what's going to happen? Water and people take the path of least resistance.
So make "undo" just that little bit harder. Pop up a box that asks them to
tick one of "vandalism", "spam", "patent nonsense" (and
whichever other
reasons are appropriate, e.g. unsourced material on BLP if applicable) and
the last is "discuss on the talk page". This means that someone cannot undo
unless they are prepared to declare it's vandalism, etc. This takes away
"undo because I don't agree with you" or "undo because I think my
source is
better than yours". That box could also highlight "This is a new
contributor; remember not to bite". The NOBITE rule is easily overlooked as
it is not immediately visible that the person being reverted is a new user.
Also, think about the old saying "praise in public, criticise in private"
and then remember that a user talk page is visible to everyone and
watchlists bring changes to the attention of goodness-knows-how-many people.
Can you think of a worse place for someone to come and say something
critical? Is there a better way to handle this? Have some private channel?
Allow users to decide which posts on their user talk page will be visible to
all? Allow users to redact anything in their user space? And while Barnstars
and other WikiLove are public (good!), thanks are not. Why not at least
include that information on a User's page or Talk page in some way
("SmellyJockStraps gave thanks 234 times and received thanks 432 times")? We
need to build a platform that makes positive behaviours and positive
interaction easier and more visible than negative behaviour and negative
interaction.
But start with a Code of Conduct. I would stick to the principles not the
specifics. I would have an all-emcompassing "a panel of
administrators/moderators can ban/block/suppress a user whose repeated
behaviour violates these principles" to avoid wikilawyering leading to a
never-ending proliferation of rules . E.g. Rule 1235. Not only is "moron"
disallowed, so is "m0r0n" and "m-o-r-o-n" or .
Kerry
_____
From: gendergap-bounces(a)lists.wikimedia.org
[mailto:gendergap-bounces@lists.wikimedia.org] On Behalf Of Vicky Knox
Sent: Tuesday, 25 November 2014 7:13 AM
To: gendergap(a)lists.wikimedia.org
Subject: Re: [Gendergap] Conflict resolution resources for
onlinecommunities?
Here is the very basic first draft of the article so far:
https://localwiki.org/main/Conflict_resolution If you want to add anything
directly into the article, please do so. :]
2014-11-24 12:20 GMT-08:00 Vicky Knox <vknoxsironi(a)gmail.com>om>:
Hi gendergap folks!
I hope you're well. :]
I'm writing conflict resolution documentation for LocalWiki
(
https://localwiki.org/main/Front_Page), a global local knowledge commons.
Do you have any conflict resolution resources for online communities, or
conflict resolution examples from Wikimedia projects you'd like to
recommend? I'm particularly interested in examples of online nonviolent
communication modalities, and intersectional feminist perspectives on online
conflict resolution in communities of mixed real name and *nym identities.
(This all said, I'm open to all suggestions--I've lurked this list for a
while and highly value the perspectives I've found on it.)
Thank you!
Vicky