On Fri, 5 May 2023, at 22:53, Evan Lloyd New-Schmidt wrote:
Hi, I'm starting a project that will involve repeated processing of HTML wikipedia articles.
Using the enterprise dumps seems like it would be much simpler than converting the XML dumps, but I don't know what the "experimental" status really means.
Hi,
From my experience working with the Wiktionary HTML dumps I can say that the data quality is quite poor: there are stale and missing entries (https://phabricator.wikimedia.org/T305407).
There are also entire namespaces excluded from the dumps, and more recently there have been issues with the dumps not getting updated.
So it depends what kind of processing you need to do–in general I find the parsing to be much easier, hopefully they'll manage to sort out the problems.
Jan