Hello,
Simetrical wrote:
On 8/2/06, Steve Bennett stevage@gmail.com wrote:
I would be interested to hear from the server admins on the feasibility of integrating this tool in Wikipedia - particularly that compilation time.
I strongly suspect that the number of runtime cycles this could take is a much greater concern. Half an hour to compile the list isn't far from the length some special-page indices take to build, IIRC.
Here is the time output of compiler for the analysis of wikipedia english articles : 622,65s user 53,69s system 70% cpu 15:58,91 total 95% of this time (16mins) was to read the files from disk (with a fresh reboot so nothing was in the cache).
The algorithm uses very few cpu power. Imagine for example that all the data needed are in memory, the compiled automaton could be obtained in 1 minutes.
Best Regards. Julien Lemoine