On 11/25/12 22:16, vitalif@yourcmc.ru wrote:
Also, I want to ask you and other developers about the idea of packing export XML file along with all exported uploads to ZIP archive (instead of putting them to XML in base64) - what do you think about it? We use it in our Mediawiki installations ("mediawiki4intranet") and find it quite convenient. Actually, ZIP was the idea of Tim Starling, before ZIP we used very strange "multipart/related" archives (I don't know why we did it :))
I want to try to get this change reviewed at last... What do you think about it?
Looks a better solution than base64 files. :)
Other improvements include advanced page selection (based on namespaces, categories, dates, imagelinks, templatelinks and pagelinks) and an advanced import report (including some sort of "conflict detection"). I should probably need to split them to separate patches in Gerrit for the ease of review?
I don't see a need to split eg. templatelinks selection and pagelinks selection. But if you provide a 64K patch, you may have a hard time getting people to review it :) I would probably generate a couple of patches, one with the selection parameters and the other with the advanced report. Depending on how big are those changes, YMMV.
Also, do all the archiving methods (7z) really need to be built in the Export.php as dump filters? (especially when using ZIP?) I.e. with simple XML dumps you could just pipe the output to the compressor.
Or are they really needed to save the temporary disk space during export? I ask because my version of import/export does not build the archive "on-the-fly" - it puts all the contents to a temporary directory and then archives it fully. Is it an acceptable method?
Probably not the best method, but a suboptimal implementation that works is better than no implementation at all. So go ahead and submit it. We can then be picky later in front of the code :)
Regards