Hi Diederik,
I have two questions:
1. Which algorithm you used to get the added/removed content between two
revisions of wikipedia?
2. What is the size of the diffdb dump after extracting? I do not want
to waste wikipedia bandwidth if I know that I can not deal with it ;).
By the way, what you did is exactly what I just started working on to
implement for my project, so thanks a lot :)
Regards.
On Fri, Nov 4, 2011 at 13:19, Diederik van Liere <dvanliere(a)gmail.com>wrote;wrote:
Dear Wiki Researchers,
During the summer we have worked on Wikihadoop [0], a tool that allows us
to create the diffs between two revisions of a Wiki article using Hadoop.
Now I am happy to announce that the entire diffdb is available for
download at
http://http://dumps.wikimedia.org/other/diffdb/
This dataset is based on the English Wikipedia April 2011 XML dump files.
The advantage of this dataset is that:
a) You can search for specific content being added / removed
b) Measure more accurately how much text an editor has added or removed
We are currently working on a Lucene-based application [1] that will allow
us to quickly search for specific strings being added or removed.
If you have any questions, then please let me know!
[0]
https://github.com/whym/wikihadoop
[1]
https://github.com/whym/diffindexer
Best regards,
Diederik van Liere
_______________________________________________
Wiki-research-l mailing list
Wiki-research-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wiki-research-l
--
Rami Al-Rfou'
631-371-3165