I've been putting placeholder images on a lot of articles on en:wp.
e.g. [[Image:Replace this image male.svg]], which goes to
[[Wikipedia:Fromowner]], which asks people to upload an image if they
own one.
I know it's inspired people to add free content images to articles in
several cases. What I'm interested in is numbers. So what I'd need is
a list of edits where one of the SVGs that redirects to
[[Wikipedia:Fromowner]] is replaced with an image. (Checking which of
those are actually free images can come next.)
Is there a tolerably easy way to get this info from a dump? Any
Wikipedia statistics fans who think this'd be easy?
(If the placeholders do work, then it'd also be useful convincing some
wikiprojects to encourage the things. Not that there's ownership of
articles on en:wp, of *course* ...)
- d.
I have a Wikipedia OTRS ticket related to someone who keeps getting
"Download this file" dialogs when they go to Wikipedia pages...
System description is:
>I have Vista and IE 8.0.6001.18813.
Page they keep seeing this on is:
en.wikipedia.org/wiki/W._Edwards_Deming
Has anyone seen behavior like this before? That's a very standard
browser and OS.
Any ideas appreciated.
--
-george william herbert
george.herbert(a)gmail.com
Hi
Now I have constructed a local wiki.And I want to add the data which download
from the internet Wikipedia to the local wiki.I tried to read the source
code,but I coudln’t find the exact thing(Interface) that I want.
So,I want to ask some questions:
when click the save button after edit an article or add a new article, how is
the data stored? Which function/class does it call?
Could you describe the process of data storage ?
What form are articles stored in database?
Thanks
Vanessa
Hi,
I have been importing the English Wikipeida XML Dumps every few
months (last time I did this was in June). I then used xml2sql and it
always worked for me. Now I attempted the import on the latest dump
enwiki-20090920-pages-articles.xml (and on the dump from
enwiki-20090810-pages-articles.xml), both of these have the error:
>$ xml2sql enwiki-20090920-pages-articles.xml
unexpected element <redirect>
xml2sql: parsing aborted at line 33 pos 16.
So then I try mwdumper and after 1.4 M Pages, it craps out:
……
1,423,000 pages (957.283/sec), 1,423,000 revs (957.283/sec)
1,424,000 pages (957.465/sec), 1,424,000 revs (957.465/sec)
Exception in thread "main" java.lang.IllegalArgumentException: Invalid
contributor
at
org.mediawiki.importer.XmlDumpReader.closeContributor(Unknown Source)
at org.mediawiki.importer.XmlDumpReader.endElement(Unknown Source)
at
org.apache.xerces.parsers.AbstractSAXParser.endElement(Unknown Source)
at
org.apache.xerces.parsers.AbstractXMLDocumentParser.emptyElement(Unknown
Source)
at
org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanStartElement(Unknown
Source)
at
org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(Unknown
Source)
at
org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(Unknown
Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown
Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown
Source)
at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
at org.apache.xerces.parsers.AbstractSAXParser.parse(Unknown
Source)
at
org.apache.xerces.jaxp.SAXParserImpl$JAXPSAXParser.parse(Unknown Source)
at javax.xml.parsers.SAXParser.parse(SAXParser.java:395)
at javax.xml.parsers.SAXParser.parse(SAXParser.java:198)
at org.mediawiki.importer.XmlDumpReader.readDump(Unknown Source)
at org.mediawiki.dumper.Dumper.main(Unknown Source)
I tried the importDump.php and I get errors of the kind (MediaWiki 1.14.0)
…
Warning: xml_parse(): Unable to call handler in_() in
/var/www/includes/Import.php on line 437
Warning: xml_parse(): Unable to call handler in_() in
/var/www/includes/Import.php on line 437
Warning: xml_parse(): Unable to call handler out_() in
/var/www/includes/Import.php on line 437
….
(Sorry I don’t know where this error starts, but it processes a few
thousand pages, up till I get sick of looking at it before failing.)
Any ideas if the format of the XML files have changed because I can
swear that as of June or may be May, I had xml2sql working. I know that
I might need to upgrade MediaWiki to 1.15, however importDump.php
usually does not work for the English Wikipedia anyways.
I would be grateful if someone has any ideas?
Thanks guys,
O. O.
P.S. http://download.wikimedia.org/tools/ does not have the source of
MWDumper. I thought this was open source?
The idea is to select edit and submit calls that are relevant to the
usability project and track edit/save ratio of filtered calls over time.
Bots will be filtered, "action=edit&redlink=1,.." will be discarded (as 95%
inadvertent edit calls), and some more.
I would appreciate help in decoding most occurring squid/html statuses:
Here are the relevant html codes from FAQ:
http://wiki.squid-cache.org/SquidFaq/SquidLogs
Also http://www.w3.org/Protocols/rfc2616/rfc2616-sec10.html
000 Used mostly with UDP traffic.
200 OK
206 Partial Content
301 Moved Permanently
302 Moved Temporarily
400 Bad Request
403 Forbidden
404 Not Found
[417 Expectation Failed]
500 Internal Server Error
502 Bad Gateway
503 Service Unavailable
504 Gateway Timeout
The following are frequencies in which the index.php result codes are
found in the 1:1000 sampled squid logs from just over 6 months:
TCP_DENIED/403,action=edit 321390
TCP_DENIED/403,action=submit 33
TCP_MISS/000,action=edit 7352
TCP_MISS/000,action=submit 1186
TCP_MISS/200,action=edit 800200
TCP_MISS/200,action=submit 75768
TCP_MISS/206,action=edit 20
TCP_MISS/206,action=submit 269
TCP_MISS/301,action=edit 662
TCP_MISS/302,action=edit 184217
TCP_MISS/302,action=submit 116141
TCP_MISS/400,action=edit 6
TCP_MISS/403,action=edit 2746
TCP_MISS/404,action=edit 119
TCP_MISS/404,action=submit 206
TCP_MISS/417,action=edit 53
TCP_MISS/417,action=submit 716
TCP_MISS/500,action=edit 362
TCP_MISS/500,action=submit 81
TCP_MISS/502,action=submit 87
TCP_MISS/503,action=edit 7
TCP_MISS/503,action=submit 5878
TCP_MISS/504,action=edit 53
TCP_MISS/504,action=submit 91
Out of these most significant given range and/or frequency are:
TCP_DENIED/403,action=edit 321390
TCP_MISS/000,action=edit 7352
TCP_MISS/200,action=edit 800200
TCP_MISS/302,action=edit 184217
TCP_MISS/000,action=submit 1186
TCP_MISS/200,action=submit 75768
TCP_MISS/302,action=submit 116141
Specific questions:
A
Any idea why there are so many TCP_DENIED/403, are these really failures ?
B
For action=submit the difference between preview and save is in the result
codes right ?
I understood earlier that TCP_MISS/302 is a successful save, right ?
Does that mean TCP_MISS/200 is preview ?
C
For action=edit how to interpret /200 vs /302 ?
D (minor)
Are TCP/000 indeed (invalid) UDP messages ?
Erik Zachte
BTW For all squid status codes from Wikimedia servers see
http://stats.wikimedia.org/wikimedia/squids/SquidReportMethods.htm
Hi,
On Thu, May 7, 2009 at 11:30 PM, Brian <Brian.Mingus(a)colorado.edu> wrote:
>
> Because I am doing a C++ implementation of the mediawiki api [...]
May I ask what the status of your implementation is? Is it still under
development, or is it ready?
Or, does anyone else know of a C++ implementation of the MediaWiki API?
Thanks,
--
Guillaume Paumier
[[m:User:guillom]]
http://www.gpaumier.org
I have a bug report for a user of my Google Wave extension:
http://www.mediawiki.org/wiki/Extension:GoogleWavehttp://code.google.com/p/micke/source/browse/GoogleWave/GoogleWave.php (source)
It seems Parser.php sends the parser object to my function as a value($parser)
rather than as a reference (&$parser):
[Tue Oct 06 16:57:53 2009] [error] [client x.x.x.x.] PHP Warning:
Parameter 3 to waveRender() expected to be a reference, value given in
C:\\Apache2.2\\htdocs\\mediawiki\\includes\\parser\\Parser.php on line
3243, referer: ...
Why is this? I don't get an error like this in any of my test installations
(although I have never tried it in a Windows environment). Googleing gives me
nothing...
My questions are these: Why does this happen? Is it something with the version
of PHP?
What are the benefits in my case of wanting the parser object as a reference
instead of a value? That is, should I do:
function waveRender($input, $argv, $parser)
rather than:
function waveRender($input, $argv, &$parser)
I just set it up the way I did after reading the docs on http://mediawiki.org
where it alwas seems to be passed by reference.
The users bug report can be found here btw:
http://groups.google.com/group/google-wave-api/browse_thread/thread/8597137…
/five messages down)
Erik Zachte, Data Analyst for the Wikimedia Foundation, has just
released the first official public version of the Wikimedia
Foundation's monthly report card on key program metrics. If you
haven't seen it, his blog post about it is here:
http://infodisiac.com/blog/2009/10/the-wikimedia-report-card/
And the public page is here:
http://stats.wikimedia.org/reportcard/
What's being tracked right now is the following:
* number of unique visitors to WMF projects according to comScore,
with breakdown by region, and the percentage reach this represents
among all Internet users;
* number of page requests according to our server log data, with
breakdown by language;
* comScore rank relative to other web properties (we can't release all
the data we have access to from comScore here);
* number of content objects (pages, binaries), with breakdown by
project for pages, and breakdown by filetype for binaries;
* participation rates: new pages per day, edits per month, new editors
per month, active editors making 5+ edits, very active editors making
100+ edits
We've used indexes as a tool to make trends easily understandable over
time -- see the "indexed" tabs in the different sections. The top left
of each section makes some key numbers easily accessible, e.g.
month-to-month comparisons and year-to-year comparisons.
The report card will continue to evolve in response primarily to the
Wikimedia Foundation's organizational priorities and needs - if we
have a specific project designed to achieve impact in one category,
we'll break out a section to look at it. And we're also hoping to have
higher level summary information. Feedback is very much welcome -
probably the best place to post it is in response to Erik's blog post
above so we have it all in one place. :-) Aside from feedback,
however, I also want to take this opportunity to praise Erik for all
his work over many months in putting this together, as well has his
ongoing work to release it on a regular basis.
--
Erik Möller
Deputy Director, Wikimedia Foundation
Support Free Knowledge: http://wikimediafoundation.org/wiki/Donate
http://wikimediafoundation.org/wiki/Job_openings/Code_Maintenance_Engineer
The core purpose of this job is to review and resolve bug reports,
enhancement requests, code changes and extensions relevant to
Wikimedia's free education projects, and to implement systematic
testing procedures. This is a very hands-on role for an experienced
software engineer who is interested in achieving high impact in a
fast-paced environment with a public benefit mission. If you have
existing development experience with MediaWiki, that's a huge plus.
:-)
We will take remote applicants into consideration, and we'll also
consider supporting relocation of the right candidate - our preference
is for this person to be based in the San Francisco office of the
Wikimedia Foundation.
--
Erik Möller
Deputy Director, Wikimedia Foundation
Support Free Knowledge: http://wikimediafoundation.org/wiki/Donate