Hello.
Since just a few hours ago, a new public repository has been created to host WikiXRay database dumps, containing info extracted from public Wikipedia dbdumps. The image is hosted by RedIRIS (in short, the Spanish equivalent of Kennisnet in Netherlands).
http://sunsite.rediris.es/mirror/WKP_research
ftp://ftp.rediris.es/mirror/WKP_research
These new dumps are aimed to save time and effort to other researchers, since they won't need to parse the complete XML dumps to extract all relevant activity metadata. We used mysqldump to create the dumps from our databases..
As of today, only some of the biggest Wikipedias are available. However, in the following days the full set of available languages will be ready for downloading. The files will be updated regularly.
The procedure is as follows:
1. Find the research dump of your interest. Download and decompress it in your local system.
2. Create a local DB to import the information.
3. Load the dump file, using a MySQL user with insert privileges:
$> mysql -u user -p passw myDB < dumpfile.sql
And you're done.
Final warning. 3 fields in the revision table are not reliable yet:
rev_num_inlinks
rev_num_outlinks
rev_num_trans
All remaining fields/values are trustable (in particular rev_len, rev_num_words, and so forth).
Regards,
Felipe.
Hi All,
I have been trying to upload one of the latest version of the XML dumps,
pages-articles.xml.bz2 from
http://download.wikimedia.org/enwiki/20090604/. I dont want the front end
and other things that comes with wikimedia installations, so i thought i
would just create the database and upload the dump.
I tried using mwdumper, but it breaks with error. After searching a bit, I
found there was a related bug filed on that issue.
https://bugzilla.wikimedia.org/show_bug.cgi?id=18328
I made the changes suggested in the thread, but i cudnt build the source,
as I couldnt get all the dependent libraries working in my machine.
I also tried using mwimport, that also failed due to the same problem.
any one have any suggestions to import the XML dump successfully to a
mysql database ?
Thanks
Srini
Wikimedia's copy of MediaWiki has been updated to r51904 (r51864 for
extensions). There were lots of fun bugs, the most exciting of which
are now fixed.
Developers, please check the bug tracker for bugs reported against
code that you maintain, or bugs that you can otherwise fix. Users,
please report any bugs you notice there, after doing a search for
duplicates, even if you also report them in other places such as IRC
or [[WP:VPT]].
https://bugzilla.wikimedia.org/
Please do not discourage users from reporting "known issues", point
them to the bug report so that they can give us a priority hint in a
comment. It's hard to prioritise bug fixes when we don't know how many
people are affected or how severe the issue is.
-- Tim Starling
-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1
Hi Robert et al,
Robert Stojnic <rainmansr(a)gmail.com> wrote:
> ... Just a quick update, a couple of days ago all of this stuff got
> enabled on all WMF projects. ...
Is there a complete and reliable docu of the current search backend and
frontend? In the past I have found a lot of pieces, sometimes not
matching each other, so it is hard (for me) to see the current structure
and really necessary parts to install it in our environment.
What about differential or incremental index update for lucene?
Some good links would be helpful.
Thx.
Uwe (Baumbach)
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.9 (MingW32)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org
iEYEARECAAYFAkomQ6gACgkQFEbayCH8zXlNFgCeIYJeSOwZIQJRQndapZFTE9FQ
b9YAoLCIPdrUvrAdHR/8nnC6AYq38wTK
=yHLC
-----END PGP SIGNATURE-----
Please let me know if you know of any schema changes that need to be
run on Wikimedia wikis apart from the following:
http://wikitech.wikimedia.org/view/Schema_changes
This includes optional changes such as feature blockers and
performance enhancements.
I'd like to run them in the next 24-48 hours. There should be no
disruption to service.
-- Tim Starling
Hoi,
We have been testing the LocalisationUpdate extension for some time now and,
we consider it quite good at the moment. We have been testing it in a test
environment and we would like to expand our testing to MediaWiki wikis that
do not run in English or any of the other languages that are already
completely localised. What we are looking for are Wikis that are/will be
running MediaWiki 1.15 and would like to experience that the localisation
for their Wiki gets updated with later localisations.
Obviously in order for this to work, there have to be people localising for
your language.
What we offer is help with the installation of the extension and support
with the running of the extension on your MediaWiki wiki. We are looking for
five wikis with five different languages. We can make this offer for wikis
where you are able to install new extensions and, where you can add a
chronjob.
Thanks,
GerardM
Hi,
I have just seen two independent instances where people said that they
sent posts to foundation-l in the last +/- 12 hours, which never got
posted on the list. The emails do not show up in the moderation queue
either (nor are these two subscribers, or the entire list, moderated).
Are there any technical problems with the mail(inglist) server that
you are aware of?
Michael
--
Michael Bimmler
mbimmler(a)gmail.com
Hi @all,
Actually I am trying to comprehend how the article text within monobook.php is
filled for the actually chosen article.
I've seen that it is processed by this line of code:
<?php $this->html('bodytext') ?>
Furthermore I found the definition of the function 'html' in SkinTemplate.php
function html( $str ) {
echo $this->data[$str];
}
...the way the filling of the array 'data' is defined (SkinTemplate.php)
function setRef($name, &$value) {
$this->data[$name] =& $value;
}
...and the call of the function setRef (SkinTemplate.php)
$tpl->setRef( 'bodytext', $out->mBodytext );
What I actually don't understand is the way the article content is allocated to
mBodytext.
The only sequence I found where some content is allocated to mBodytext is the
following line, which only contains the footer.
$out->mBodytext .= $printfooter ;
Thanks for your feedback
greets
magggus
Hi all,
It seems to me that there's been sterling work on the 'flagged revisions'
front - with the bulk of the credit due to User:Cenarium over on en, and the
various folk working away over there.
With that in mind, could I please encourage a dev.s attention to;
https://bugzilla.wikimedia.org/show_bug.cgi?id=18244
Hopefully we can enable the extension as soon as possible :-)
best,
Peter,
PM.