once we confirm that a few thousand people using this to suck out WP histories won't thrash the servers.
The charts are beautiful. yet I think it would be overdone to generate them on demand for every article. After you've seen a few you've seen them all. They give a general impression of how fluid popular and/or contested articles are, but are too crowded for detailed analysis.
For me a limited set of pregenerated charts for say 50 selected articles, refreshed once a month, would be sufficient.
Erik Zachte
On Sat, 26 Mar 2005 01:59:30 +0100, Erik Zachte e.p.zachte@chello.nl wrote:
once we confirm that a few thousand people using this to suck out WP histories won't thrash the servers.
The charts are beautiful. yet I think it would be overdone to generate them on demand for every article. After you've seen a few you've seen them all. They give a general impression of how fluid popular and/or contested articles are, but are too crowded for detailed analysis.
Hmm. If you have access to the tool itself, you are able to effectively browse the article text at each revision, in combination with the visual overview, in quite an effective way. So having a local copy of the full history flow, with the interface tool, is certainly useful for all articles.
As for generating snapshots for each article; certainly on-demand wouldn't be very interesting. But see for instance these 100 revisions of the "Evolution" article from 2003, miniaturized to a width of 200px. That suffices to tell you how much of the article was written by one person, when the last sea-change in article content occurred, and whether or not there have been significant reversion-wars or blankings or size changes over the past few dozen edits.
http://en.wikinews.org/wiki/IBM_releases_free_software_for_visualizing_docum...
For me a limited set of pregenerated charts for say 50 selected articles, refreshed once a month, would be sufficient.
Which 50? For what purpose?
Is there a way to export Wikipedia revisions, via the site or database dump, to the text files the tool wants?
There is a way, I think involving a mediawiki plugin which was not released on Friday; but which has been fully tested and exists somewhere. I expect there will be more details and FAQ answers on the site next week.
On Fri, 25 Mar 2005 23:51:44 -0600, Jack Lutz jack-lutz@comcast.net wrote:
Is there a way to export Wikipedia revisions, via the site or database dump, to the text files the tool wants?
Well, I need this tool for my own PhD research... so please let it go public !
Jean-Baptiste Soufron
Le 26 mars 05, à 01:59, Erik Zachte a écrit :
once we confirm that a few thousand people using this to suck out WP histories won't thrash the servers.
The charts are beautiful. yet I think it would be overdone to generate them on demand for every article. After you've seen a few you've seen them all. They give a general impression of how fluid popular and/or contested articles are, but are too crowded for detailed analysis.
For me a limited set of pregenerated charts for say 50 selected articles, refreshed once a month, would be sufficient.
Erik Zachte
Wikipedia-l mailing list Wikipedia-l@Wikimedia.org http://mail.wikipedia.org/mailman/listinfo/wikipedia-l
On Sat, 26 Mar 2005 01:59:30 +0100, Erik Zachte e.p.zachte@chello.nl wrote:
once we confirm that a few thousand people using this to suck out WP histories won't thrash the servers.
The charts are beautiful. yet I think it would be overdone to generate them on demand for every article. After you've seen a few you've seen them all. They give a general impression of how fluid popular and/or contested articles are, but are too crowded for detailed analysis.
I don't agree. They could be used to check whether significant portions of the article have been deleted, to see whether a an article is basically one person's work or patched up from different authors, and to faster get a list of the 'main authors' than can be gotten through the history.
I would thus want this to be available for all articles with certain properties (more than 3 authors or more than 10 revisions, for example, or maybe simply all articles), updated if the last request is more than one week old. It would be even better if we had enough machines to have a backup of the database (updated weekly) on one or two machines, which then could be used for information like this, searching, maintenance page, the more difficult special pages, and perhaps even direct SQL queries.
Andre Engels
wikipedia-l@lists.wikimedia.org