I suppose it goes without saying that this could be achieved with an Extension. Perhaps some kind of regex blacklist or whitelisting extension?
So for example you could have a page called [[MediaWiki:TitlesBlacklist]] which could be a newline separated list of regular expressions to block for title creation. (With of course an associated [[MediaWiki:TitlesWhitelist]]). Then have an extension which applies the rules prior to page submission, and also on retrieval (in case some malicious user finds a way through).
Or even better, a page called [[MediaWiki:TitlesFilter]] with each line delimiting and "allow" or "deny" rule. Something like this:
------------------------- SNIP ---------------- # Comments and blank lines are ignored
# Disallow any title like this - /.*index.php?title=.*/i
# Allow anything else + /.*/ ------------------------- SNIP ----------------
This is how Nutch decides whether or not to allow a page.[1]
[1] http://lucene.apache.org/nutch/tutorial8.html#Intranet%3A+Configuration
On 2/7/07, Mark Clements gmane@kennel17.co.uk wrote:
"Matthew Flaschen" matthew.flaschen@gatech.edu wrote in message news:45C963F9.1070803@gatech.edu...
Minh Nguyen wrote:
With spammers, the IP addresses always change (they've gotten good
about
that over the past few years), so I suppose the best way to control these "scout messages" is to grep [[Special:Newpages]] regularly for telltale page titles (like ".php") and delete on sight.
An admin bot should be able to delete those pages automatically, if someone cares to write it.
Is there a way of telling MW to block certain page names, e.g. to disallow new pages that match the regex "/index.php?title=/i"? If not then this seems like a worthwhile config setting to add.
- Mark Clements (HappyDog)
Wikitech-l mailing list Wikitech-l@lists.wikimedia.org http://lists.wikimedia.org/mailman/listinfo/wikitech-l