Am 18.10.2011 17:23, schrieb Thomas Morton:
That comes down to the two layers of judgment involved in this proposal. At first we give them the option to view anything and we give them the option to view not anything. The problem is that we have to define what "not anything" is. This imposes our judgment to the reader. That means, that even if the reader decides to hide some content, then it was our (and not his) decision what is hidden.
No; because the core functionality of a filter should always present the choice "do you want to see this image or not". Which is specifically not imposing our judgement on the reader :) Whether we then place some optional preset filters for the readers to use is certainly a matter of discussion - but nothing I have seen argues against this core ideas.
Yes; because even the provision of a filter implies that some content is seen as objectionable and treated different from other content. This is only no problem, as long we don't represent default settings, aka categories, which introduce our judgment to the readership. Only the fact that our judgment is visible, is already enough to manipulate the reader in what to see as objectionable or not. This scenario is very much comparable to the unknown man that sits behind you, looking randomly onto your screen, while you want to inform yourself. Just the thought that someone else could be upset is already an issue. Having us to directly show/indicate what we think of as objectionable "by others" is even the stronger.
If we treat nothing as objectionable (no filter), then we don't need to
play the judge. We say: "We accept anything, it's up to you to judge". If we start to add a "category based" filter, then we play the judge over our own content. We say: "We accept anything, but this might not be good to look at. Now it is up to you to trust our opinion or not".
By implementing a graded filter; one which lets you set grades of visibility rather than off/on addresses this concern - because once again it gives the reader ultimate control over the question of what they want to see. If they are seeing "too much" for their preference they can tweak up, and vice versa.
This would imply that we, the ones that are unable to neutrally handle content, would be perfect in categorizing images after a fine degree of nudity. But even having multiple steps would not be a satisfying solution. There are many cultural regions which differentiate strongly between man an woman. While they would have no problem to see a man in just his boxer short, it would be seen as offending to show a woman open hair. I wonder what effort it would need to accomplish this goal (if even possible), compared to the benefits.
The later imposes our judgment to the reader, while the first makes no judgment at all and leaves anything to free mind of the reader. ("free mind" means, that the reader has to find his own answer to this question. He might have objections or could agree.)
And if he objects, we are then just ignoring him?
I disagree with your argument; both points are imposing our judgement on the reader.
If _we_ do the categorization, then we impose our judgment, since it was us, who made the decision. It is not a customized filter where the user decides what is best for himself. Showing anything might not be ideal for all readers. Hiding more then preferred might also no be ideal for all readers. Hiding less then preferred is just another not ideal case. We can't meet everyones taste like no book can meet everyones taste. While Harry Potter seams to be fine in many cultures, in some there might be parts that are seen as offensive. Would you hide/rewrite parts from Harry Potter to make them all happy, or would you go after the majority of the market and ignore the rest?
There is one simple way to deal with it. If someone does not like our content, then he don't need to use it. If someone does not like the content of a book he does not need to buy it. He can complain about it. Thats whats Philip Pullman meant with: "No one has the right to life without being shocked".
Agreed; which is why we allow people to filter based on a sliding scale, rather than a discrete yes or no. So someone who has no objection to such an image, but wants to hide people having sex can do so. And someone who wants to hide that image can have a stricter grade on the filter.
If nothing else the latter case is the more important one to address; because sexual images are largely tied to sexual subjects, and any reasonably person should expect those images to appear. But if culturally you object to seeing people in swimwear then this could be found in almost any article.
We shouldn't judge those cultural objections as invalid. Equally we shouldn't endorse them as valid. There is a balance somewhere between those two extremes.
Yes there is a balance between two extremes. But who ever said that the center between two opinions is seen as an valid option by both parties? If that would be the case and if that would work in practice, then we wouldn't have problems like in Israel. In this case everyone has a viewpoint, but neither party is willing to agree with a median. Both have very different perspectives about what a median should look like. This applies at large scale to situations like in Israel and it also applies to small things like a single line of text or an image.
The result is simple: Neither side is happy with a balance. Evey side has his point of view and they won't back down. At the result we have the so called second battle field aside from the articles itself. As soon we start to categorize it will happen, and I'm sure that even you would shake with the head as you see those differences colliding with each other. The battles inside articles can be described as the mild ones. Here we have arguments and sources. How many sources do belong to our images? What would you cite as the base for your argumentation?
I suggested a way in which we could cover a broad spectrum of views on one key subject without setting discrete categories of visibility.
As explained above, this will be a very very hard job to do. Even in the most simple subject "sexuality" you will need more then one scale to measure content against. Other topics, like the religious or cultural topics, will be even a much harder job.
I belive that the idea dies at the moment as we assume that we can achieve neutrality through filtering. Speaking theoretically there are only three types of neutral filters. The first leaves anything through, the second blocks all and the third is totally random, resulting in an equal 50:50 chance for large numbers. Currently we would ideally have the first filter. Your examples show that this isn't always true. But at least this is the goal. Filter two would equal to don't show anything, or shut down Wikipedia. Not an real option. I know. The third option is a construct out of theory that would not work, since it contains an infinite amount of information, but also nothing at all.
What about the fourth type; that gives you extensive options to filter out (or better description; to collapse) content from initial viewing per your specific preferences.
This is a technical challenge, but in no way unachievable.
This is by far not a technical challenge. It's a new/additional challenge for the authors. A new burden if you will. The finer the categorization, the more effort you will need to put into it. The more exceptions have to be made. Technically you could support thousands of categories with different degrees. But what can be managed by our authors and what by the readers? At which point the error made by us (we are humans, an computers can't judge images) is bigger then thin lines we draw?
In technical theory it sounds nice and handy, but in practice we also have to consider effort vs result. I'm strongly confident that the effort would not justify the result, even if we ignore side effects like third party filtering, based upon our categories, removing the options from the user.
I made an analogy before that some people might prefer to surf Wikipedia with plot summaries collapsed (I would be one of them!). In a perfect world we would have the option to collapse *any* section in a Wikipedia article and have that option stored. Over time the software would notice I was collapsing plot summaries and, so, intelligently collapse summaries on newly visited pages for me. Plus there might even be an option in preferences saying "collapse plot summaries" because it's recognised as a common desire.
In this scenario we keep all of the knowledge present, but optionally hide some aspects of it until the reader pro-actively accesses it. Good stuff.
That would be a solution. But this would not imply any categorization by ourself, since the program on the servers would find out what to do. This works pretty well for simple things like text already. Images are a much bigger problem, which can't be simply handed down to program, since no program at the current time would be able to do this. So we are back again at: effort vs result + gathering of private user data + works only opt-in with an account.
I removed some paragraphs below, since all come down to the effort vs result problem. Additionally we have no way to implement a system like this at the moment. That is something for the future.
The whole problems starts with the intention to spread our knowledge to more people that we currently reach, faster then necessary.
That we might not be reaching certain people due to a potentially fixable problem is certainly something we can/should address :)
Yes we should address it. But we also should start to think about other options then to hide content. There are definitely better and more effective solutions as this quicky called "image filter".
We have a mission, but it is not the mission to entertain as many people as possible. It is not to gain as much money trough donors as possible.
Is this a language barrier? do you mean entertain in the context of having them visit us, or in the context of them having a fun& enjoyable time.
Because in the latter case - of course you are right. I don't see the relevance though because this isn't about entertaining people, just making material accessible.
With "entertain" i meant this: Providing them only with content that will please their mind, causing no bad thoughts or surprise to learn something new or very different.
It isn't our purpose to please the readers by only representing
knowledge they would like to hear of.
Yeh, this is a finicky area to think about... because although we ostensibly report facts, we also record opinions on those facts. Conceivably a conservative reading a topic would prefer to see more conservative opinion on that topic and a liberal more liberal opinion.
Ok, so we have forks that cover this situation - but often they are of poor quality, and present the facts in a biased way. In an ideal future world I see us maintaining a core, netural and broad article that could be extended per reader preference with more commentary from their political/religious/career/interest spectrum.
The point is to inform, after all.
Tom
That is kind of another "drawing the line" case. To be neutral we should represent both (or more) point of views. But showing the reader only that what he want's to read is not real knowledge. Real knowledge is gained by looking over the borders and not by building own temples/territories with huge walls around them, to trap or bait an otherwise free mind. It is always good to have an opposition that thinks different. Making them all happy by dividing them into two (or more) territories, while differences remain or grow, will not help both of them, especially if you can't draw a clean line.
nya~