I was searching Google, and found a website (
http://tramadol.tfres.net/wiki/Main_Page) that, when visited, loads a live
page from Wikipedia, and puts a large number of Google Ads at the top of
each page. According to a relevant page on Wikipedia (
http://en.wikipedia.org/wiki/WP:FORK#Remote_loading), this, and the irc
channel #wikimedia-tech, is the place to report this.
I first reported this on #wikimedia-tech. The first reply said to just "post
to mailing list", while the second said to go to bugzilla. Is this the right
place?
[attempt 2, hope it gets through this time]
At present, Wikimedia operates a global spam URL blacklist, which
lives on meta: and is editable only by Meta admins, which blocks URLs
added to it from being added to any page.
en:wp is also getting an imperial arseload of spammers. The people on
the front lines, in the
#wikipedia-spam and #wikipedia-spam-t channels, are finding things
getting very difficult.
At present, getting a URL added requires finding a meta admin, or
adding it to [[m:Talk:Spam blacklist]] and hoping someone will get to
it soon. And of course meta's admin requirements are very different to
those on other wikis.
Obviously, the present blacklist works globally because if a spam URL
shouldn't be on one wiki, then it shouldn't be on the others either.
But it would be nice to have a wiki-specific list as well as the
global list. The upside of this would be being able to respond to spam
much more quickly. The downside is consolidating the lists regularly
and sensibly, which may turn out to be a damn nuisance - regularly is
easy, sensibly requires thinking about every URL added and whether a
local admin has been overenthusiastic in dealing with a spam problem
when a temporary block would have sufficed.
In the long-term, with our growing problems with spam, it's probably
not very efficient to keep
throwing the task onto a few meta admins, and to continually go
through a two-step process to get sites added to it.
So ...
wikitech-l: Does this sound workable in our present setup, per
http://meta.wikimedia.org/wiki/SpamBlacklist_Documentation ?
foundation-l: How would we make the local <-> global thing workable?
There's an automatic updater for trusted shared lists ... would that
be enough?
- d.
Repost with better formatting:
I am in the process of upgrading my Mediawiki parser for TomeRaider files,
adding conditional template processing, and I also noticed how complicated
templates can become
Here is an example:
For article fr.wikipedia.org/wiki/Paris the following templates were
executed:
#expr: 294
#if: 100
#ifexpr: 350
After disabling geohack code, which yields two numbers only (longitude and
latitude):
#expr: 2
#if: 88
#ifexpr: 2
Also what happens when a switch stateent has hundreds of clauses, each with
new sub-templates: are these all resolved before resolving the switch? Just
wondering.
Erik Zachte
-----------------
Domas:
We had some slowness today, where after some temporary issue cluster ended
up spending >50% of it's time working on Spanish wikipedia template. We
nuked the template (es.wikipedia site was also turned off at some moment),
so now it is all up and running.
The template was used for adding a single line of text, though it took 20
seconds to render. ;-)
We'll probably have to do something about that.
I want to preprocess the wikitext of articles adding
info found by a database lookup based on the article
name.
The best place for preprocessing is in the parser so
I've hooked ParserBeforeStrip but the parser works
in chunks of wikitext rather than in articles. I've been
able to filter out the chunks which are not part of the
article because $parser->getVariableValue('revisionid')
will return 0 for those.
But long artices can cause more than one chunk of
wikitext to be sent to the parser and since I only
want to add to the top of the article I need to detect
these chunks and ignore them.
Another problem is that I also want to add to the top
of the 'noarticletext' page which is displayed when
the user enters a URL for a page which doesn't exist.
There is currently no hook for this. I've tried adding
one which works for me but there are actually 3
places in the code which can result in this page.
I'm not sure that I need to hook the other two since
they appear to be edge cases.
One final question is on sharing data between my
hooks for ArticleAfterFetchContent and
ParserBeforeStrip. I'm using a global variable but
maybe there's a better way. ArticleAfterFetchContent
seems to be the best place for doing the database
lookup based on the article name to work nice with
the cache.
Andrew Dunbar (hippietrail)
--
http://linguaphile.sf.net
On 12.01.2007 05:43, werdna(a)svn.wikimedia.org wrote:
> Revision: 19142
> Author: werdna
> Date: 2007-01-11 20:43:33 -0800 (Thu, 11 Jan 2007)
>
> Log Message:
> -----------
> UI stuff for the cascading protection feature - warnings for sysops editing cascading protected pages (including which page the protection has come from), more useful errors for users (including which page the protection has come from), and a note at the top of the protect page when trying to unprotect a page affected by cascading protection to the effect of 'you can't affect the cascading protection here'.
>
Tried that revision in my local playground wiki install.
*Awesome*
Of course, this is all way cooler than a bot. Thanks a lot for
developing all this!
--Ligulem
Hi all.
The Libresoft Group, at Universidad Rey Juan Carlos (Madrid), has started to generate a bunch of interesting metric graphics about Wikipedia.
These graphics currently include histograms about article size. They will soon also include, for instance, caracterization of several important groups of authors and their evolution in time (much in the way Jimbo Wales exposed a few months ago).
We plan to extend these results to all language versions of Wikipedia.
As these graphics complement current wikimetrics info about the project, generated by Erik Zatche's scripts, we wonder about the better place to release them, in case Wikimedia is interested (under a GPL license).
Could we include them in an "official repository page" with metrics about Wikipedia? Should we place only a link to a server of our own? Which could be the better place to create those links?
Thanks a lot.
Regards.
Felipe Ortega.
__________________________________________________
Correo Yahoo!
Espacio para todos tus mensajes, antivirus y antispam ¡gratis!
Regístrate ya - http://correo.yahoo.es
An automated run of parserTests.php showed the following failures:
This is MediaWiki version 1.10alpha (r19155).
Reading tests from "maintenance/parserTests.txt"...
Reading tests from "extensions/Cite/citeParserTests.txt"...
Reading tests from "extensions/Poem/poemParserTests.txt"...
19 still FAILING test(s) :(
* TODO: Table security: embedded pipes (http://mail.wikipedia.org/pipermail/wikitech-l/2006-April/034637.html)
* TODO: Link containing double-single-quotes '' (bug 4598)
* TODO: message transform: <noinclude> in transcluded template (bug 4926)
* TODO: message transform: <onlyinclude> in transcluded template (bug 4926)
* BUG 1887, part 2: A <math> with a thumbnail- math enabled
* TODO: HTML bullet list, unclosed tags (bug 5497)
* TODO: HTML ordered list, unclosed tags (bug 5497)
* TODO: HTML nested bullet list, open tags (bug 5497)
* TODO: HTML nested ordered list, open tags (bug 5497)
* TODO: Parsing optional HTML elements (Bug 6171)
* TODO: Inline HTML vs wiki block nesting
* TODO: Mixing markup for italics and bold
* TODO: 5 quotes, code coverage +1 line
* TODO: dt/dd/dl test
* TODO: Images with the "|" character in the comment
* TODO: Parents of subpages, two levels up, without trailing slash or name.
* TODO: Parents of subpages, two levels up, with lots of extra trailing slashes.
* TODO: Don't fall for the self-closing div
* TODO: Always escape literal '>' in output, not just after '<'
Passed 484 of 503 tests (96.22%)... FAILED!