Happy Monday,
There are strange people who make such links (kindof urlencoded?):
[[Második világháború#Partrasz.C3.A1ll.C3.A1s Szic.C3.ADli.C3.A1ban
.28Huskey hadm.C5.B1velet.29|Huskey hadműveletben]]
So the section title must have been copied from the URL.
Do we have a ready tool to fix these?
--
Bináris
Hello all
>From one of my assignments as a bot operator I have some code which
does template parsing and general text parsing (e.g. Image/File tags).
It is not using regex and thus able to correctly parse nested
templates and other such nasty things. I have written those as library
classes and written tests for them which cover almost all of the code.
I would now really like to contribute that code back to the community.
Would you be interested in adding this code to the pywikibot
framework? If yes, can I send the code to someone for code review or
how do you usually operate?
Greetings
Hannes
PS: wiki userpage is http://en.wikipedia.org/wiki/User:Hannes_R%C3%B6st
Hi,
I was just thinking about renaming articles in the way we change text by
replace.py.
I use replace.py for correcting spelling errors. They may appear as well in
titles. Is there a way to correct misspelled titles using the flexibility
and robustness of replace.py and fixes.py?
I can use them to search such articles without renaming, if I replace space
to anything and write the regex to 'require-title' exception list, but it
only allows one regex (being the listed expressions in AND relation), which
is not really comfortable.
--
Bináris
Hello,
Since Wikidata is awesome and now replaced all of Template:Link FA and its
similar templates with badges. We decided
<https://phabricator.wikimedia.org/T109210> to archive featured.py and keep
it for historical reasons. The script is moved to "archive" folder in
scripts directory. You still can run it using "python pwb.py featured" and
some tests will be ran over the script but it won't be maintained anymore.
It would be good to write a new script to add badges to Wikidata Even
though it's a easy task that humans should do ;). This change
<https://gerrit.wikimedia.org/r/267410> is merged in master branch but
we'll port it to 2.0 branch soon so if you still use this script in your
wiki please note this deprecation.
Also it's the first script to be archived, I think we can archive a lot
more now.
Best
'Pywikibot' has this problem. The results dont show which version is doing
it.
---------- Forwarded message ----------
From: "Brad Jorsch (Anomie)" <bjorsch(a)wikimedia.org>
Date: 30 Jan 2016 6:06 am
Subject: [Wikitech-l] *unclear-"now"-timestamp* (was Re: Bot operators:
Please check your configuration)
To: "Wikimedia developers" <wikitech-l(a)lists.wikimedia.org>, "MediaWiki API
announcements & discussion" <mediawiki-api(a)lists.wikimedia.org>
Cc:
On Fri, Jan 29, 2016 at 1:41 PM, Bináris <wikiposta(a)gmail.com> wrote:
>
> > 2016-01-29 18:56 GMT+01:00 Brad Jorsch (Anomie) <bjorsch(a)wikimedia.org>:
> >
> > > by going to
> > > https://www.mediawiki.org/wiki/Special:ApiFeatureUsage, entering your
> > > agent
> > > (or any useful prefix of it), and looking for "https-expected".
> > >
> >
> > What does *unclear-"now"-timestamp* mean here?
> >
>
> For various API timestamp-typed parameters, you can pass unusual values
> such as the empty string or "0" and it will be interpreted as meaning
> "now", which doesn't make very much sense except for the fact that it has
> always done that. If you really mean "now", you should pass that as the
> value instead.
>
> action=edit even has to hack around this to avoid spurious edit conflicts
> if you do it for the 'basetimestamp' parameter. Ideally we'd make empty
> string and '0' be rejected as invalid timestamps, but first people have to
> stop passing them in.
>
>
> --
> Brad Jorsch (Anomie)
> Senior Software Engineer
> Wikimedia Foundation
> _______________________________________________
> Wikitech-l mailing list
> Wikitech-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
hello hello!
about the use of pywikibot:
is it possible to use to parse the xml dump?
I am interested in extracting links from pages (internal, external, with
distinction from ones belonging to category).
I also would like to handle transitive redirect.
I would like to process the dump, without accessing wiki, either access
wiki with proper limits in butch.
Is there maybe something in the package already taking care of this ?
I 've seen in https://www.mediawiki.org/wiki/Manual:Pywikibot/Scripts
there is a "ghost" extracting_links.py" script,
I wonted to ask before re-inventing the wheel, and if pywikibot is suitable
tool for the purpose.
Thank you,
L.
In core master (not 2.0), reflinks has been converted to using
requests instead of urllib2/httplib.
The change is https://gerrit.wikimedia.org/r/#/c/264251/ , by a GCI participant.
As requests does not support FTP, reflinks no longer supports FTP links.
A new task has been created to investigate FTP support:
https://phabricator.wikimedia.org/T124007
There was also a minor performance regression. Where previously only
the first 1Mb of a URL was fetched, now the entire resource is
fetched.
https://phabricator.wikimedia.org/T124138
--
John Vandenberg
I have read the WM messaging about the changes that are taking place in WM
1.27.0/11 with regards to authentication, though not understanding what
that will mean for Pywikibot. Would SKS please cover that information gap.
Thx.
-- billinghurst
Hi.
I have an old problem. When I run replace.py, sometimes the diff is not
coloured with red and green (see the attachment). This makes replacing very
uncomfortable and also dangerous (even if I see a coloured difference, I
can't be sure there are no more).
I use compat so tell me, if the problem is already fixed in core and I am
just speaking for nothing... I couldn't identify the reason, but in the
last times I always see it vith ó and ő, as seen on image; as to my best
knowledge Hungarian is the only language in the discovered part of Universe
to use ő, this issue may have not come up for others.
--
Bináris