There are strange people who make such links (kindof urlencoded?):
[[Második világháború#Partrasz.C3.A1ll.C3.A1s Szic.C3.ADli.C3.A1ban
.28Huskey hadm.C5.B1velet.29|Huskey hadműveletben]]
So the section title must have been copied from the URL.
Do we have a ready tool to fix these?
>From one of my assignments as a bot operator I have some code which
does template parsing and general text parsing (e.g. Image/File tags).
It is not using regex and thus able to correctly parse nested
templates and other such nasty things. I have written those as library
classes and written tests for them which cover almost all of the code.
I would now really like to contribute that code back to the community.
Would you be interested in adding this code to the pywikibot
framework? If yes, can I send the code to someone for code review or
how do you usually operate?
PS: wiki userpage is http://en.wikipedia.org/wiki/User:Hannes_R%C3%B6st
At first, I am really happy to see the pywikipedia framework is still
alive and very active.
As you may have noticed, whenever a change is send to Gerrit that
triggers Jenkins jobs that run the code style utilities pep8 and pyflakes.
It seems your code is not passing the style checks so whenever they fail
it is not going to prevent you from merging the code.
I am not sure how your community likes pep8/pyflakes. But I think it
would be nice to have the code repositories to pass those tests and
enforce authors to follow them. You can still have some pep8 checks
ignored such as the "line too long".
The questions are:
Is there any interest in making your repositories pep8 compliants?
If so, is there anything I can do to help? :-]
Antoine "hashar" Musso
I'm trying to convert a fairly large set of scripts from compat to
core and I found a significant loss of functionality in getting image
and template info. While writing this, I've noticed that the latest
version of core also has some of these prblems. I will elaborate on
this loss of functionality below, but I would like to know if this
simplification is intended or if this is part of some work in
For the image parsing, the function linkedPages(withImageLinks = True)
used to provide images that were not included through templates, while
imageLinks would provide all the images. In core, the linkedPages
function no longer provides this capability, and I haven't found any
replacement (I ported the old function in my code)
For template parsing, templatesWithParams from class Page used to
provide a pair containing the template name and a list of parameters,
with the full "key=value" string. Nowadays, we're getting a dictionary
instead of that list. Normally there is nothing wrong with that,
except that in Python 2 the dictionary is unordered, which means that:
* the order of the parameters is forever lost
* the original text cannot be reconstructed (because of the above and
the missing whitespace information) - this means there is no easy way
to identify and/or replace a particular instance of the template in a
page with many identical templates. It used to be you could do it with
simple find/replace operations, now it takes some more work.
I personally would like to have the old behavior back, it would save
me and probably others a lot of work.
Unfortunately, tables can't use # symbols for numbering, therefore people
use constant numbers.
Do we have any tool for handling that? I mean I want to insert a new 6th
row after 5th and renumber the rest of the table.
at the moment Pywikibot has four mailing lists.
* pywikipedia-announce – Only two messages this year (one spam, the
other one to say sorry.) no last year
* pywikipedia-l – This list, very active ;-)
* pywikipedia-svn – Not needed anymore because of the move to git.
* pywikipedia-bugs – Used to inform users about bugs reported on
Sourceforge, after the move to Bugzilla this is not needed anymore, right?
So I propose to close at least pywikipedia-svn and pywikipedia-bugs,
maybe also pywikipedia-announce because it is not used.
On 10 August 2013 11:26, Dr. Trigon <dr.trigon(a)surfeu.ch> wrote:
> Pywikibot finished GIT migration. Now a question has anybody ever
> tryed to clone and use the e.g. the pywikibot/compat repo on the TS?
> If I do, I get and .git folder of about 200MB which nearly fills my
> home quota. (this is the plain clone - nothing done to/with it yet!!)
Thanks to Chad:
valhallasw@lisilwen:~/src/pwb$ git clone
Cloning into 'tmp'...
remote: Counting objects: 37472, done
remote: Finding sources: 100% (37472/37472)
remote: Getting sizes: 100% (864/864)
remote: Compressing objects: 100% (600634/600634)
remote: Total 37472 (delta 26599), reused 37439 (delta 26591)
Receiving objects: 100% (37472/37472), 8.05 MiB | 275 KiB/s, done.
Resolving deltas: 100% (26611/26611), done.
8.05MB. Beat that! :-) (the checkout itself accounts for the remaining 6
MB, totaling to 14 MB)