On 11/25/12 22:16, vitalif(a)yourcmc.ru wrote:
Also, I want to ask you and other developers about the
idea of packing
export XML file along with all exported uploads to ZIP archive (instead
of putting them to XML in base64) - what do you think about it? We use
it in our Mediawiki installations ("mediawiki4intranet") and find it
quite convenient. Actually, ZIP was the idea of Tim Starling, before ZIP
we used very strange "multipart/related" archives (I don't know why we
did it :))
I want to try to get this change reviewed at last... What do you think
Looks a better solution than base64 files. :)
Other improvements include advanced page selection
(based on namespaces,
categories, dates, imagelinks, templatelinks and pagelinks) and an
advanced import report (including some sort of "conflict detection"). I
should probably need to split them to separate patches in Gerrit for the
ease of review?
I don't see a need to split eg. templatelinks selection and pagelinks
selection. But if you provide a 64K patch, you may have a hard time
getting people to review it :)
I would probably generate a couple of patches, one with the selection
parameters and the other with the advanced report.
Depending on how big are those changes, YMMV.
Also, do all the archiving methods (7z) really need to
be built in the
Export.php as dump filters? (especially when using ZIP?) I.e. with
simple XML dumps you could just pipe the output to the compressor.
Or are they really needed to save the temporary disk space during
export? I ask because my version of import/export does not build the
archive "on-the-fly" - it puts all the contents to a temporary directory
and then archives it fully. Is it an acceptable method?
Probably not the best method, but a suboptimal implementation that works
is better than no implementation at all. So go ahead and submit it. We
can then be picky later in front of the code :)