Although for a different use case, I find myself facing a related problem with my bot on Commons.
https://commons.wikimedia.org/wiki/Commons:Auto-protected_files/wikipedia/zh
Those pages get updated periodically whenever a commons image starts or ends being used on a Main Page. Aside from the fact that this in particular could be a native feature[1], this page will get a lot of revisions.
Now also aside is that these old revisions are quite useless. Even if someone would make an edit in between, the bot will overwrite it. But I don't care so much for the wasted space, since it is relatively small waste.
The problem comes in when these pages need maintenance. I find myself periodically checking up on my bot generated pages to make sure I move them to /Archive_# and delete those and start clean. Because when the revision count reaches 2,500, it can no longer be deleted because of the limit we implemented.
So to keep them mobile and usable, I always fight the limit by moving it to a subpage (without a redirect) before it reaches that limit and delete it there and start a new page on the old name.
Would be nice if these kind of bot-generated pages could avoid that. And if in doing so we saving revision space, that's nice.
Other examples: * https://commons.wikimedia.org/wiki/Commons:Database_reports (and subpages) * https://commons.wikimedia.org/wiki/Commons:Auto-protected_files (and subpages)
-- Krinkle