Merlijn van Deen <valhallasw(a)arctus.nl> wrote:
> Personally, I rather we wait for the Pywikipedia
devs to fix that script,
This is not going to happen anytime soon. Considering
the state of the
code base (two hundred exceptions for three hunderd wikis, long
functions and no automated testing - and thus practically untestable),
and the state of the InterLanguage extension ('will be installed
soon'), so-one is really willing to invest a lot of time in tracking
memory usage and reducing it.
The only reasonable action we can take to reduce the
memory
consumption is to let the OS do its job in freeing memory: using one
process to track pages that have to be corrected (using the database,
if possible), and one process to do the actual fixing (interwiki.py).
This should be reasonably easy to implement (i.e. use a pywikibot page
generator to generate a list of pages, use a database layer to track
interlanguage links and popen('interwiki.py <page>') if this is a
fixable situation)
We could also move the pressure: Labs' bot running infra-
structure doesn't seem to be /that/ far from opening. If
interwiki bots were running there, it would allow the foun-
dation to judge whether pushing for the deployment of Inter-
Language isn't worth it in the end.
Meanwhile I think DaB.'s proposal is very adequate.
Tim