It might be useful to include a robots.txt file that'll tell search spiders not to bother with any of the active pages such as 'Edit'. While it isn't hard to make this kind of file it could be useful to include it for the sake of giving people a starting place. As most wiki's would have the same robots.txt file anyway. On most of my sites anyway they get hit by spiders several times a day so keeping spiders from wasting time on pages they don't need to index minimizes the wasted server time.