Hi Nuno,
Op 25-8-2011 11:41, Nuno Tavares schreef:
Em 25-08-2011 09:28, Maarten Dammers escreveu:
Of course I'm getting it from the wiki lists.
:-)
The basis is at
http://commons.wikimedia.org/wiki/Commons:Wiki_Loves_Monuments_2011/Structu…
The overview of the database is at
http://commons.wikimedia.org/wiki/Commons:Wiki_Loves_Monuments_2011/Monumen…
The source is at
https://svn.toolserver.org/svnroot/p_erfgoed/erfgoedbot/
On the input side I could really need some help with:
* Adding more sources to the database
* Tweaking the current sources
* Adding some converters (for example for different coordinates format)
*Sigh*
Should you have used PERL, I could expand the tool considerably.
It would take me ages to do it in Python...
Only the harvest part is python. Output
(api) is all php.
What I was about to propose would be something like
this flow:
1. grab your database dump
2. use data quality tools to apply corrections (it's far easier for me)
3. stuff into our tools/plist/ to be able to export them into wiki format
4. paste the exported to the wikipedia lists
And then your bot would grab everything fine, I hope.
It will. Did you manage to do
this?
On the output side I really need help:
* Need some clean up functions to be implemented
* Option to filter on with or without images
* KML output needs to be prettified
* (more)
I'm updating the database now. I'm afraid the weekly run failed because
of some toolserver SQL issues. I will schedule it to run every day now.
Did you add the coordinates for Portugal Nuno?
We are in the process of copying
the database into the wiki lists (step
4 above), so you should grab them in the following days.
Anyway, I just converted our table to monuments_all (you probably
noticed in the last index2.php listing), I can do it for monuments_pt as
well if you want an updated version quickly, as this process will take
some time to finish...
I changed the bot to run every night so you have a fresh
copy every morning.
Maarten