Brian wrote:
It was very much preferable when this was still a maintenance script, mostly because it had the useful property of working!
/var/www/emergent/extensions/DumpHTML$ *php dumpHTML.php
/tmp/emergent_static/ -k monobook*
Now, let's comment out line 137 of dumpHTML.php and see what happens (I already created this directory anyway - it doesn't need to do that).
I don't know why this script isn't working. But one thing is clear - it is thinking *entirely too hard* about what it's supposed to be doing. Whatever changes have been made since this was a maintenance script have not been great changes for me.
I tried using wget. Something like: wget -e robots=off --mirror -I emergent --convert-links --no-parent http://grey.colorado.edu/emergent
But mediawiki is a dynamic content trap when it comes to wget, so I am presently without a solution.
Block /emergent/index.php? URLs with robots.txt
Does anyone see whats wrong with this script?
Thanks, Brian
Do you have write permission to that folder?