There are strange people who make such links (kindof urlencoded?):
[[Második világháború#Partrasz.C3.A1ll.C3.A1s Szic.C3.ADli.C3.A1ban
.28Huskey hadm.C5.B1velet.29|Huskey hadműveletben]]
So the section title must have been copied from the URL.
Do we have a ready tool to fix these?
https://phabricator.wikimedia.org/T161726 for handling this in pywikibot
before Wikidata bots start breaking because they encounter pages with
On 29 March 2017 at 16:10, Daniel Kinzler <daniel.kinzler(a)wikimedia.de>
> Am 29.03.2017 um 15:19 schrieb Luca Martinelli:
> >> One thing to note: We currently do not export statements that use this
> >> datatype to RDF. They can therefore not be queried in the Wikidata Query
> >> Service. The reason is that we are still waiting for geoshapes to get
> >> URIs. This is handled in this ticket.
> This ticket: <https://phabricator.wikimedia.org/T159517>. And more
> The technically inclined of you may be interested in joining the relevant
> discussion on IRC tonight at 21:00 UTC (2pm PDT, 23:00 CEST)
> Daniel Kinzler
> Principal Platform Engineer
> Wikimedia Deutschland
> Gesellschaft zur Förderung Freien Wissens e.V.
> Wikidata mailing list
Dear bot users & developers,
The review of the Code of Conduct for Wikimedia technical spaces has been
completed and now it is time to bootstrap its first committee. The
Technical Collaboration team is looking for five candidates to form the
Committee plus five additional auxiliary members. One of them could be you
or someone you know!
You can propose yourself as a candidate and you can recommend others
techconductcandidates AT wikimedia DOT org
We want to form a very diverse list of candidates reflecting the variety of
people, activities, and spaces in the Wikimedia technical community. We are
also open to other candidates with experience in the field. Diversity in
the Committee is also a way to promote fairness and independence in their
decisions. This means that no matter who you are, where you come from, what
you work on, or for how long, you are a potential good member of this
The main requirements to join the Committee are a will to foster an open
and welcoming community and a commitment to making participation in
Wikimedia technical projects a respectful and harassment-free experience
for everyone. The committee will handle reports of unacceptable behavior,
will analyze the cases, and will resolve on them according to the Code of
Conduct. The Committee will also handle proposals to amend the Code of
Conduct for the purpose of increasing its efficiency. The term of this
first Committee will be one year.
Once we have a list of 5 + 5 candidates, we will announce it here for
review. You can learn more about the Committee and its selection process at
https://www.mediawiki.org/wiki/Code_of_Conduct/FAQ#Committee. You can ask
questions in the related Talk pages (preferred) or here.
You can also track the progress of this bootstrapping process at
Engineering Community Manager @ Wikimedia Foundation
Recently, I'm working on a research project which needs extracting article information from wikipedia.
I managed to get pywikibot work on my computer and was able to pull out a few simple results.
One question is regarding a method called pywikibot.pagegenerators.AllpagesPageGenerator.
By setting the argument "content" to "True", it will return a page generator with current version. But, which version will be returned if setting the argument to False?
Also, is there a way in pywikibot to get a page generator that contains articles/pages up to a certain date?
Maybe, pywikibot is not a right tool to do this.
I was thinking of using wiki dump data instead of using a wiki API.
But, it seems the files are huge. I appreciate it you happen to have any idea to deal with this.
Thanks a lot!