On Tue, 8 Feb 2005 23:13:12 +0100, Maik Musall lists@musall.de wrote:
- I have to avoid wasting bandwidth and wikipedia server load as much as possible.
To save server load and bandwidth, I considered using the database dump, but that lacks the images and the layout, right? I even downloaded the wikipedia CDROM image, but discovered that's a Windows software with data stuffed into some database where it's probably difficult to retrieve and make PDFs from.
My current idea is to use the normal web access since I have no other working solution. I would spread the accesses over a week and only use times where normal server load is low.
- What's the best way to get those 1000 articles from the servers without putting too much load on them?
Do you know about Special:Export? http://en.wikipedia.org/wiki/Special:Export/Electricity
All you need to do then is to download the images.
Find out how the wikireaders were made. http://en.wikipedia.org/wiki/Wikipedia:WikiReader