> "wm-license-information-description" will probably be comming from
E:WikimediaMessages, You will also want a dump containing the locally
one done on the wiki's MediaWiki: namespace.
Thank you for your response. Is there a dump file containing this
information, and if not, is there any way to extract it (sanely) from the
running Wikipedia instance?
I am learning as I go, so my apologies if I am overlooking something
simple. :-)
Thanks!
Tim
Hey,
For the new renderer backend for the Collections Extension we've come up
with a tentative architecture that we would like operations buy in on. The
living document is here [1]. It's worth saying explicitly that whatever
setup we use must be able to handle the greater than 150k requests a day we
serve using the old setup.
Basically we're looking at having
* 'render servers' run node.js
* doing job management in Redis
* rendering content using PhantomJS and/or Latex
* storing rendered files locally on the render servers (and streaming the
rendered results through MediaWiki -- this is how it's done now as well).
* having a garbage collector run routinely on the render servers to cleanup
old stale content
Post comments to the talk page please :)
[1 ]https://www.mediawiki.org/wiki/PDF_rendering/Architecture
~Matt Walker
Wikimedia Foundation
Fundraising Technology Team
Hi all,
I have successfully set up a local wikipedia instance using the English
data dumps, but I cannot seem to find a copy of the MediaWiki translation
files in use on the Wikipedia sites. As a result, tags like
"wm-license-information-description" are not displayed correctly.
Is there a place I can download the Wikipedia-modified system message
files (languages/messages/MessagesEn.php,
languages/messages/MessagesQqq.php, etc.)?
Thanks!
Tim
TL;DR: How can we collaboratively put together a list of non-spammy sites
that wikis may want to add to their interwiki tables for whitelisting
purposes; and how can we arrange for the list to be efficiently distributed
and imported?
Nemo bis points out that
Interwiki<https://www.mediawiki.org/wiki/Extension:Interwiki>is "the
easiest way to manage whitelisting" since nofollow isn't applied to
interwiki links. Should we encourage, then, wikis to make more use of
interwiki links? Usually, MediaWiki installations are configured so that
only sysops can add, remove, or modify interwiki prefixes and URLs. If a
user wants to link to another wiki, but it's not on the list, often he will
just use an external link rather than asking a sysop to add the prefix
(since it's a hassle for both parties and often people don't want to bother
the sysops too much in case they might need their help with something else
later). This defeats much of the point of having the interwiki table
available as a potential whitelist, unless the sysops are pretty on top of
their game when it comes to figuring out what new prefixes should be added.
In most cases, they probably aren't; the experience of Nupedia shows that
elitist, top-down systems tend not to work as well as egalitarian,
bottom-up systems.
Currently, interwiki.sql<https://git.wikimedia.org/blob/mediawiki%2Fcore//maintenance%2Finterwiki.sql>has
100 wikis, and there doesn't seem to be much rhyme or reason to which
ones are included (e.g. Seattlewiki?) I wrote
InterwikiMap<https://www.mediawiki.org/wiki/Extension:InterwikiMap>,
which dumps the contents of the wiki's interwiki table into a backup page
and substitutes in its place the interwiki table of some other wiki (e.g. I
usually use Wikimedia's), with such modifications as the sysops see fit to
make. The extension lets sysops add, remove and modify interwiki prefixes
and URLs in bulk rather than one by one through Special:Interwiki, which is
a pretty tedious endeavor. Unfortunately, as written it is not a very
scalable solution, in that it can't accommodate very many thousand wiki
prefixes before the backup wikitables it generates exceed the capacity of
wiki pages, or it breaks for other reasons.
I was thinking of developing a tool that WikiIndex (or some other wiki
about wikis) could use to manage its own interwiki table via edits to
pages. Users would add interwiki prefixes to the table by adding a
parameter to a template that would in turn use a parser function that, upon
the saving of the page, would add the interwiki prefix to the table.
InterwikiMap could be modified to do incremental updates, polling the API
to find out what changes have recently been made to the interwiki table,
rather than getting the whole table each time. It would then be possible
for WikiIndex (or whatever other site were to be used) to be the
wikisphere's central repository of canconical interwiki prefixes. See
http://wikiindex.org/index.php?title=User_talk%3AMarkDilley&diff=172654&old…
But there's been some question as to whether there would be much demand for
a 200,000-prefix interwiki table, or whether it would be desirable. It
could also provide an incentive for spammers to try to add their sites to
WikiIndex. See
https://www.mediawiki.org/wiki/Talk:Canonical_interwiki_prefixes
It's hard to get stuff added to meta-wiki's interwiki
map<https://meta.wikimedia.org/wiki/Interwiki_map>because one of the
criteria is that the prefix has to be one that would be
used a lot on Wikimedia sites. How can we put together a list of non-spammy
sites that wikis would be likely to want to have as prefixes for nofollow
whitelisting purposes, and distribute that list efficiently? I notice that
people are more likely to put together lists of spammy than non-spammy
sites; see e.g. Freakipedia's
list<http://freakipedia.net/index.php5?title=Spam_Site_List>.
(Hmm, I think I'll pimp my websites to that wiki when I get a chance; the
fact that the spam isn't just removed but put on permanent record in a
public denunciation means it's a potential opportunity to gain exposure for
my content. They say there's no such thing as bad publicity. ;) )
--
Nathan Larson <https://mediawiki.org/wiki/User:Leucosticte>
Distribution of my contributions to this email is hereby authorized
pursuant to the CC0 license<http://creativecommons.org/publicdomain/zero/1.0/>
.
Just got this when trying to read an article. Not sure if anyone is doing
anything, but if you are, you might have broken something.
If you report this error to the Wikimedia System Administrators, please include
the details below.
Request: GET http://en.wikipedia.org/wiki/Johnny_the_Homicidal_Maniac, from
10.64.32.106 via cp1067 cp1067 ([10.64.0.104]:3128), Varnish XID 2127871567
Forwarded for: 162.17.205.153, 208.80.154.76, 10.64.32.106
Error: 503, Service Unavailable at Thu, 14 Nov 2013 18:48:33 GMT
Thank you,
Derric Atzrott
Computer Specialist
Alizee Pathology
Hi,
I'm currently working on the Memento Extension for Mediawiki, as announced earlier today by Herbert Van de Sompel.
The goal of this extension is to work with the Memento framework, which attempts to display web pages as they appeared at a given date and time in the past.
Our goal is for this to be a collaborative effort focusing on solving issues and providing functionality in "the Wikimedia Way" as much as possible.
Without further ado, I have the following technical questions (I apologize in advance for the fire hose):
1. The Memento protocol has a resource called a TimeMap [1] that takes an article name and returns text formatted as application/link-format. This text contains a machine-readable list of all of the prior revisions (mementos) of this page. It is currently implemented as a SpecialPage which can be accessed like http://www.example.com/index.php/Special:TimeMap/Article_Name. Is this the best method, or is it more preferable for us to extend the Action class and add a new action to $wgActions in order to return a TimeMap from the regular page like http://www.example.com/index.php?title=Article_Name&action=gettimemap without using the SpecialPage? Is there another preferred way of solving this problem?
2. We currently make several database calls using the the select method of the Database Object. After some research, we realized that Mediawiki provides some functions that do what we need without making these database calls directly. One of these needs is to acquire the oldid and timestamp of the first revision of a page, which can be done using Title->getFirstRevision()->getId() and Title->getFirstRevision()->getTimestamp() methods. Is there a way to get the latest ID and latest timestamp? I see I can do Title->getLatestRevID() to get the latest revision ID; what is the best way to get the latest timestamp?
3. In order to create the correct headers for use with the Memento protocol, we have to generate URIs. To accomplish this, we use the $wgServer global variable (through a layer of abstraction); how do we correctly handle situations if it isn't set by the installation? Is there an alternative? Is there a better way to construct URIs?
4. We use exceptions to indicate when showErrorPage should be run; should the hooks that catch these exceptions and then run showErrorPage also return false?
5. Is there a way to get previous revisions of embedded content, like images? I tried using the ImageBeforeProduceHTML hook, but found that setting the $time parameter didn't return a previous revision of an image. Am I doing something wrong? Is there a better way?
6. Are there any additional coding standards we should be following besides those on the "Manual:Coding_conventions" and "Manual:Coding Conventions - Mediawiki" pages?
7. We have two styles for serving pages back to the user:
* 302-style[2], which uses a 302 redirect to tell the user's browser to go fetch the old revision of the page (e.g. http://www.example.com/index.php?title=Article&oldid=12345)
* 200-style[3], which actually modifies the page content in place so that it resembles the old revision of the page
Which of these styles is preferable as a default?
8. Some sites don't wish to have their past Talk/Discussion pages accessible via Memento. We have the ability to exclude namespaces (Talk, Template, Category, etc.) via configurable option. By default it excludes nothing. What namespaces should be excluded by default?
Thanks in advance for any advice, assistance, further discussion, and criticism on these and other topics.
Shawn M. Jones
Graduate Research Assistant
Department of Computer Science
Old Dominion University
[1] http://www.mementoweb.org/guide/rfc/ID/#Pattern6
[2] http://www.mementoweb.org/guide/rfc/ID/#Pattern1.1
[3] http://www.mementoweb.org/guide/rfc/ID/#Pattern1.2
Hi everyone!
I'm searching for some report about MediaWiki performance testing and
load testing results. Is such thing available anywhere?
-----
Yury Katkov, WikiVote