Brian wrote:
It was very much preferable when this was still a
maintenance script, mostly
because it had the useful property of working!
/var/www/emergent/extensions/DumpHTML$ *php dumpHTML.php
> /tmp/emergent_static/ -k monobook*
Now, let's comment out line 137 of dumpHTML.php
and see what happens (I
already created this directory anyway - it doesn't need to do that).
I don't know why this script isn't working. But one thing is clear - it is
thinking *entirely too hard* about what it's supposed to be doing. Whatever
changes have been made since this was a maintenance script have not been
great changes for me.
I tried using wget. Something like: wget -e robots=off --mirror -I emergent
--convert-links --no-parent
http://grey.colorado.edu/emergent
But mediawiki is a dynamic content trap when it comes to wget, so I am
presently without a solution.
Block /emergent/index.php? URLs with robots.txt
Does anyone see whats wrong with this script?
Thanks,
Brian
Do you have write permission to that folder?