It is increasingly common to add books to Wikisource by finding a PDF or Djvu file, uploading it to Commons, and then to create an Index: page on Wikisource for proofreading.
But this would be much easier if:
1) The fields (author, title, etc.) of the Index page were filled in from the data already given on Commons. (Yes, those could be wrong or need additional care, but this could always be edited afterwards, if initial values are fetched from Commons.)
2) The <pagelist/> tag was already in the "pages" box.
3) All pages were created in automatically with the OCR text from Commons, instead of leaving a long list of red links. (This would require the text for each page to be extracted, something that pdftotext can do in seconds, but Commons takes weeks to do.)
Could this be automated? Is there already some tool or bot that does this?