On Tue, 8 Feb 2005 23:13:12 +0100, Maik Musall <lists(a)musall.de> wrote:
1. I have to avoid wasting bandwidth and wikipedia
server load as
much as possible.
To save server load and bandwidth, I considered using the database dump,
but that lacks the images and the layout, right? I even downloaded the
wikipedia CDROM image, but discovered that's a Windows software with
data stuffed into some database where it's probably difficult to
retrieve and make PDFs from.
My current idea is to use the normal web access since I have no other
working solution. I would spread the accesses over a week and only use
times where normal server load is low.
1. What's the best way to get those 1000 articles from the servers
without putting too much load on them?
Do you know about Special:Export?
http://en.wikipedia.org/wiki/Special:Export/Electricity
All you need to do then is to download the images.
Find out how the wikireaders were made.
http://en.wikipedia.org/wiki/Wikipedia:WikiReader