Hello all:
I was interested in this exchange since I've never really been absolutely sure how to do this:
But in the interest of short URLs, I serve my MediaWiki directly from site / without any /wiki/ or /w/ directories. So above meathod would not work on my installation.
Any ideas how I can exclude robots from crawling all my wiki's edit, history, talk, etc, pages *without* excluding its article pages?
Excluding index.php using robots.txt should work if an article link on your page is http://mysite.tld/My_Page.
So, what do you do if the wiki is in the root directory and not a subdir and you're using ugly URLs?
Thanks,
~Tricia webmaster@prwatch.org
Fix it! Copy the content to a w directory and change 1 line in localsettings.php. Then use a .htaccess to rewrite urls, its very easy.
http://meta.wikimedia.org/wiki/Eliminating_index.php_from_the_url
Frisoni
On 11/6/06, Patricia Barden webmaster@prwatch.org wrote:
Hello all:
I was interested in this exchange since I've never really been absolutely sure how to do this:
But in the interest of short URLs, I serve my MediaWiki directly from site / without any /wiki/ or /w/ directories. So above meathod would not work on my installation.
Any ideas how I can exclude robots from crawling all my wiki's edit, history, talk, etc, pages *without* excluding its article pages?
Excluding index.php using robots.txt should work if an article link on your page is http://mysite.tld/My_Page.
So, what do you do if the wiki is in the root directory and not a subdir and you're using ugly URLs?
Thanks,
~Tricia webmaster@prwatch.org
MediaWiki-l mailing list MediaWiki-l@Wikimedia.org http://mail.wikipedia.org/mailman/listinfo/mediawiki-l
mediawiki-l@lists.wikimedia.org