Hello all:
I was interested in this exchange since I've never really been
absolutely sure how to do this:
But in the interest of short URLs, I serve my
MediaWiki directly from
site / without any /wiki/ or /w/ directories. So above meathod would
not work on my installation.
Any ideas how I can exclude robots from crawling all my wiki's edit,
history, talk, etc, pages *without* excluding its article pages?
Excluding index.php using robots.txt should work if an article link on
your page is
http://mysite.tld/My_Page.
So, what do you do if the wiki is in the root directory and not a
subdir and you're using ugly URLs?
Thanks,
~Tricia
webmaster(a)prwatch.org