Bugs item #2001249, was opened at 2008-06-24 00:11
Message generated for change (Comment added) made by nicdumz
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2001249&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
>Status: Closed
>Resolution: Fixed
Priority: 5
Private: No
Submitted By: Carsrac (carsrac)
Assigned to: Nobody/Anonymous (nobody)
Summary: disambig page not detected in zea
Initial Comment:
With the standard version of /families/wikipedia_family.py disambig was not detected.
After I modified some line in self.disambiguationTempletes into
'zea':[u'dp', u'Deurverwiespagina'],
and in self.disambcatname
'zea': u'Wikipedia:Deurverwiespagina',
After those modification it did work for the first time for disambig pages in zea.
I'm working with version r5619.
----------------------------------------------------------------------
>Comment By: NicDumZ Nicolas Dumazet (nicdumz)
Date: 2008-06-24 01:12
Message:
Logged In: YES
user_id=1963242
Originator: NO
fixed in r5620, thanks for the report !
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2001249&group_…
Bugs item #2001249, was opened at 2008-06-24 00:11
Message generated for change (Tracker Item Submitted) made by Item Submitter
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2001249&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Carsrac (carsrac)
Assigned to: Nobody/Anonymous (nobody)
Summary: disambig page not detected in zea
Initial Comment:
With the standard version of /families/wikipedia_family.py disambig was not detected.
After I modified some line in self.disambiguationTempletes into
'zea':[u'dp', u'Deurverwiespagina'],
and in self.disambcatname
'zea': u'Wikipedia:Deurverwiespagina',
After those modification it did work for the first time for disambig pages in zea.
I'm working with version r5619.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2001249&group_…
Bugs item #1988771, was opened at 2008-06-09 06:17
Message generated for change (Comment added) made by andreasjs
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1988771&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: other
Group: None
Status: Open
Resolution: None
Priority: 6
Private: No
Submitted By: siebrand (siebrand)
Assigned to: Nobody/Anonymous (nobody)
Summary: Encoding issues with Esperanto
Initial Comment:
Output:
siebrand@v05-s341:~/nl$ python nowcommons.py -lang:eo -autonomous -replacealways -replaceloose
WARNING: Configuration variable 'language' is defined but unknown. Misspelled?
Checked for running processes. 1 processes currently running, including the current process.
Getting references to [[ablono:Nun en komunejo]]
NOTE: You have new messages on wikipedia:eo
>>> Dosiero:LieioMapoAmel.png <<<
Traceback (most recent call last):
File "nowcommons.py", line 231, in <module>
main()
File "nowcommons.py", line 227, in main
bot.run()
File "nowcommons.py", line 222, in run
wikipedia.output(u'%s' % e)
UnicodeEncodeError: 'ascii' codec can't encode character u'\u011d' in position 35: ordinal not in range(128)
This is about the page: http://eo.wikipedia.org/wiki/Dosiero:LiegxioMapoFleron.png
Esperanto uses LanguageEo.php with some iconv magic. (http://svn.wikimedia.org/viewvc/mediawiki/trunk/phase3/languages/classes/La…) .
----------------------------------------------------------------------
>Comment By: AndreasJS (andreasjs)
Date: 2008-06-23 17:44
Message:
Logged In: YES
user_id=1738850
Originator: NO
With this chante, bug [ 1879122 ] Double redirect fixes not coding special
characters in EO.WP reappears. A solution has to be found that fixes both
bugs.
----------------------------------------------------------------------
Comment By: siebrand (siebrand)
Date: 2008-06-12 04:39
Message:
Logged In: YES
user_id=1107255
Originator: YES
Ouch. No. It works if I remove those 4 lines, indeed. Now why are they in
there! :)
----------------------------------------------------------------------
Comment By: Francesco Cosoleto (cosoleto)
Date: 2008-06-12 04:23
Message:
Logged In: YES
user_id=181280
Originator: NO
I don't fixed the bug in SVN. Have you edited wikipedia.py in
Page::title() as described?
----------------------------------------------------------------------
Comment By: siebrand (siebrand)
Date: 2008-06-12 04:16
Message:
Logged In: YES
user_id=1107255
Originator: YES
It appears there is another issue related/behind this:
python nowcommons.py -autonomous -replaceloose -replacealways -lang:eo
WARNING: Configuration variable 'language' is defined but unknown.
Misspelled?
Checked for running processes. 1 processes currently running, including
the current process.
Getting references to [[ablono:Nun en komunejo]]
>>> Dosiero:LieioMapoAmel.png <<<
Image file URL for [[eo:Dosiero:LieioMapoAmel.png]] not found.
----------------------------------------------------------------------
Comment By: Francesco Cosoleto (cosoleto)
Date: 2008-06-12 02:38
Message:
Logged In: YES
user_id=181280
Originator: NO
>>> import wikipedia
>>> s=wikipedia.Site('eo')
>>> p=wikipedia.ImagePage(s, 'Dosiero:LiegxioMapoAmel.png')
>>> p.title()
u'Dosiero:Lie\u011dioMapoAmel.png'
>>> p.get()
[...]
wikipedia.NoPage: (wikipedia:eo,
u'[[eo:Dosiero:Lie\u011dioMapoAmel.png]]')
r5305 has changed Page::title() in wikipedia.py ("Page.title(): Esperanto
decoding, fix bug 1425206 and patch 1880140") [*],
If I remove:
if self.site().lang == 'eo':
title = decodeEsperantoX(title)
pages are found and the nowcommons.py seems full working. But maybe there
are problems with redirect.py.
[*]
[ 1880140 ] Decode Esperanto titles
http://sourceforge.net/tracker/index.php?func=detail&aid=1880140&group_id=9…
[ 1425206 ] Esperanto X-Convention and interwiki
http://sourceforge.net/tracker/index.php?func=detail&aid=1425206&group_id=9…
----------------------------------------------------------------------
Comment By: Marc-Etienne Vargenau (vargenau)
Date: 2008-06-11 04:25
Message:
Logged In: YES
user_id=1118700
Originator: NO
See also bug 1988290
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1988771&group_…
Patches item #1880140, was opened at 2008-01-25 21:48
Message generated for change (Comment added) made by andreasjs
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603140&aid=1880140&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: AndreasJS (andreasjs)
Assigned to: Nobody/Anonymous (nobody)
Summary: Decode Esperanto titles
Initial Comment:
Titles have to be decoded from x convention for liks to work properly.
Here is the patch:
Index: wikipedia.py
===================================================================
--- wikipedia.py (revision 4939)
+++ wikipedia.py (working copy)
@@ -454,6 +454,8 @@
pass
if underscore:
title = title.replace(' ', '_')
+ if self.site().lang == 'eo':
+ title = decodeEsperantoX(title)
return title
def titleWithoutNamespace(self, underscore=False):
----------------------------------------------------------------------
>Comment By: AndreasJS (andreasjs)
Date: 2008-06-23 17:41
Message:
Logged In: YES
user_id=1738850
Originator: YES
No, don't close, but try to fix it properly. This patch fixes bug 1879122
(Double redirect fixes not coding special characters in EO.WP) but
provokes bug 1988771. See also bug (Bot wants to remove existing eo:
pages). So instead of declaring this patch "evil", one should try to find a
solution to fix both bugs. For this to happen, there should be a clear
rationale for when to X-code an esperanto text and when not.
----------------------------------------------------------------------
Comment By: Nobody/Anonymous (nobody)
Date: 2008-06-23 16:22
Message:
Logged In: NO
close this **evil** patch?
----------------------------------------------------------------------
Comment By: Francesco Cosoleto (cosoleto)
Date: 2008-06-13 06:19
Message:
Logged In: YES
user_id=181280
Originator: NO
Applied in r5305, undone in r5563.
See [ 1988771 ] Encoding issues with Esperanto
(https://sourceforge.net/tracker/index.php?func=detail&aid=1988771&group_id=…)
----------------------------------------------------------------------
Comment By: Andr Malafaya Baptista (malafaya)
Date: 2008-01-26 09:50
Message:
Logged In: YES
user_id=1037345
Originator: NO
This seems to be working fine using "redirect.py -lang:eo double".
----------------------------------------------------------------------
Comment By: AndreasJS (andreasjs)
Date: 2008-01-26 09:37
Message:
Logged In: YES
user_id=1738850
Originator: YES
File Added: wikipedia.diff
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603140&aid=1880140&group_…
Bugs item #1994441, was opened at 2008-06-15 15:48
Message generated for change (Comment added) made by nobody
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1994441&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Purodha B Blissenbach (purodha)
Assigned to: NicDumZ Nicolas Dumazet (nicdumz)
Summary: -prefixindex page generator strips everything before ":"
Initial Comment:
Here the error:
python /home/.../interwiki.py -v -array:10 -query:20 -prefixindex:Category:Wikipedia:Medmaacher_kan -namespace:14 -lang:ksh -family:wikipedia
Checked for running processes. 1 processes currently running, including the current process.
Pywikipediabot (r5564 (wikipedia.py), Jun 13 2008, 09:43:07)
Python 2.5.2 (r252:60911, May 28 2008, 19:19:25)
[GCC 4.2.4 (Debian 4.2.4-1)]
Retrieving mediawiki messages from Special:Allmessages
WARNING: No character set found.
Not a wiki page.
NOTE: Number of pages queued is 0, trying to add 20 more.
Retrieving Allpages special page for wikipedia:ksh from Medmaacher%20kan, namespace 14
Not a wiki page.
NOTE: Nothing left to do
What I had expected to see:
Retrieving Allpages special page for wikipedia:ksh from Wikipedia:Medmaacher%20kan, namespace 14
Thus, I believe, there are too man "Namepace" prefixes stripped. If I recall it right, another page generator had a similar problem in the past which was fixed.
----------------------------------------------------------------------
Comment By: Nobody/Anonymous (nobody)
Date: 2008-06-23 20:30
Message:
Logged In: NO
close this?
----------------------------------------------------------------------
Comment By: Purodha B Blissenbach (purodha)
Date: 2008-06-15 23:21
Message:
Logged In: YES
user_id=46450
Originator: YES
Thank you for the quick response.
I updated to the new revision, and the situation is not really better. It
seems, prefixindex in not working, somehow.
In fact, there are several categories "User *" from the Babel system in
the cu wikipedia, but:
(1) Try:
python /home/purodha/pywikipedia/interwiki.py -v -lang:cu
-prefixindex:Category:User
Checked for running processes. 1 processes currently running, including
the current process.
Pywikipediabot (r5575 (wikipedia.py), Jun 15 2008, 15:38:19)
Python 2.5.2 (r252:60911, May 28 2008, 19:19:25)
[GCC 4.2.4 (Debian 4.2.4-1)]
Retrieving mediawiki messages from Special:Allmessages
WARNING: No character set found.
Not a wiki page.
NOTE: Number of pages queued is 0, trying to add 60 more.
Retrieving Allpages special page for wikipedia:cu from User, namespace 14
Not a wiki page.
NOTE: Nothing left to do
I am wondering that it says "Retrieving Allpages special page ...", not
"... Prefixindex ..."
(2) Try:
python /home/.../interwiki.py -v -lang:cu -prefixindex:User -namespace:14
Checked for running processes. 1 processes currently running, including
the current process.
Pywikipediabot (r5575 (wikipedia.py), Jun 15 2008, 15:38:19)
Python 2.5.2 (r252:60911, May 28 2008, 19:19:25)
[GCC 4.2.4 (Debian 4.2.4-1)]
Retrieving mediawiki messages from Special:Allmessages
WARNING: No character set found.
Not a wiki page.
NOTE: Number of pages queued is 0, trying to add 60 more.
Retrieving Allpages special page for wikipedia:cu from User, namespace 0
Not a wiki page.
NOTE: Nothing left to do
Namespace:14 seems to be ignored. That should not be happening silently,
at least.
(3) Try:
python /home/..../interwiki.py -v -lang:cu -prefixindex:Category:User
-namespace:14
Checked for running processes. 1 processes currently running, including
the current process.
Pywikipediabot (r5575 (wikipedia.py), Jun 15 2008, 15:38:19)
Python 2.5.2 (r252:60911, May 28 2008, 19:19:25)
[GCC 4.2.4 (Debian 4.2.4-1)]
Retrieving mediawiki messages from Special:Allmessages
WARNING: No character set found.
Not a wiki page.
NOTE: Number of pages queued is 0, trying to add 60 more.
Retrieving Allpages special page for wikipedia:cu from User, namespace 14
Not a wiki page.
NOTE: Nothing left to do
I am not wondering anymore :-) but was expecting that, now,
"Category:User" should be looked for in NS 14, not "User"
(4) Try:
I had used -start:... as a poor replacement for -prefixindex:... before,
it works (partially) by returning pages (but incorrectly claims they were
in another namespace)
python /home/.../interwiki.py -v -lang:cu -start:Category:User
Checked for running processes. 1 processes currently running, including
the current process.
Pywikipediabot (r5575 (wikipedia.py), Jun 15 2008, 15:38:19)
Python 2.5.2 (r252:60911, May 28 2008, 19:19:25)
[GCC 4.2.4 (Debian 4.2.4-1)]
Retrieving mediawiki messages from Special:Allmessages
WARNING: No character set found.
Not a wiki page.
NOTE: Number of pages queued is 0, trying to add 60 more.
Retrieving Allpages special page for wikipedia:cu from User, namespace 14
Not a wiki page.
Getting 60 pages from wikipedia:cu...
Not a wiki page.
WARNING: Outdated family file wikipedia: namespace['cu'][-2] is
'?????????', but should be '?????????'
WARNING: Outdated family file wikipedia: namespace['cu'][-1] is
'????????', but should be '????????'
WARNING: Outdated family file wikipedia: namespace['cu'][1] is '??????',
but should be '??????'
WARNING: Outdated family file wikipedia: namespace['cu'][2] is
'?????????????', but should be '?????????????'
WARNING: Outdated family file wikipedia: namespace['cu'][3] is
'???????????? ??????', but should be '???????????? ??????'
WARNING: Outdated family file wikipedia: namespace['cu'][5] is '?????????
??????', but should be '????????? ??????'
WARNING: Outdated family file wikipedia: namespace['cu'][6] is '????', but
should be '????'
WARNING: Outdated family file wikipedia: namespace['cu'][7] is '????
??????', but should be '???? ??????'
WARNING: Outdated family file wikipedia: namespace['cu'][9] is 'MediaWiki
??????', but should be 'MediaWiki ??????'
WARNING: Outdated family file wikipedia: namespace['cu'][10] is
'????????', but should be '????????'
WARNING: Outdated family file wikipedia: namespace['cu'][11] is '????????
??????', but should be '???????? ??????'
WARNING: Outdated family file wikipedia: namespace['cu'][12] is '??????',
but should be '??????'
WARNING: Outdated family file wikipedia: namespace['cu'][13] is '??????
??????', but should be '?????? ??????'
WARNING: Outdated family file wikipedia: namespace['cu'][14] is
'?????????', but should be '?????????'
WARNING: Outdated family file wikipedia: namespace['cu'][15] is '?????????
??????', but should be '????????? ??????'
NOTE: [[cu:?????????:User bg]] does not have any interwiki links
NOTE: [[cu:?????????:User bg-N]] does not have any interwiki links
WARNING: [[cu:?????????:User cs]] is in namespace 0, but
[[lt:Kategorija:User cs]] is in namespace 14. Follow it anyway? ([y]es,
[n]o)
1000s of such messages follows, so this is also not useable, but that is
likely another subject matter.
P.S.
Should I better have opened another but for this instead of using a reply?
----------------------------------------------------------------------
Comment By: NicDumZ Nicolas Dumazet (nicdumz)
Date: 2008-06-15 16:14
Message:
Logged In: YES
user_id=1963242
Originator: NO
Yup, right !
I fixed that bug in r5578 :)
Thanks for the report !
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1994441&group_…
Feature Requests item #1864162, was opened at 2008-01-04 20:11
Message generated for change (Comment added) made by nobody
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603141&aid=1864162&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Priority: 5
Private: No
Submitted By: Nobody/Anonymous (nobody)
Assigned to: Nobody/Anonymous (nobody)
Summary: Error interwiki.py and sysop account
Initial Comment:
I'm using a sysop bot to make some interwikies in protected pages in pt.wikinews. The bot finds a new interwikies, but the bot doesn't save.
This bot have sysop at pt.wikinews and es.wikinews, because both wikies are protecting pages to prevent spam bot and another things.
Here is a example:
======Post-processing [[pt:Milhares de pessoas fogem do furaco Katrina em Nova
Orlees, nos EUA]]======
Updating links on page [[es:Numerosas personas huyen del huracn Katrina en Nue
va Orleans]].
No changes needed
Updating links on page [[pt:Milhares de pessoas fogem do furaco Katrina em Nov
a Orlees, nos EUA]].
Changes to be made: Modificando: [[en:Mandatory evacuation declared in New Orle
ans, US]]
+ [[de:Angst vor dem Hurrikan: New Orleans wird evakuiert]]
+ [[en:Mandatory evacuation declared in New Orleans, US]]
- [[en:Mandatory Evacuation declared in New Orleans]]
- [[de:Angst vor dem Hurrikan: New Orleans wird evakuiert]]
NOTE: Updating live wiki...
Page is locked, using sysop account.
Sleeping for 3.0 seconds, 2008-01-04 17:05:23
Changing page [[pt:Milhares de pessoas fogem do furaco Katrina em Nova Orlee
s, nos EUA]]
200 OK
WARNING: wikinews: [[pt:Milhares de pessoas fogem do furaco Katrina em Nova Or
lees, nos EUA]] does not link to [[en:Mandatory evacuation declared in New Orl
eans, US]] but to [[en:Mandatory Evacuation declared in New Orleans]]
WARNING: wikinews: [[en:Mandatory evacuation declared in New Orleans, US]] links
to incorrect [[nl:New-Orleans houdt hart vast voor komst Katrina]]
----------------------------------------------------------------------
Comment By: Nobody/Anonymous (nobody)
Date: 2008-06-23 20:27
Message:
Logged In: NO
close this?
----------------------------------------------------------------------
Comment By: Nobody/Anonymous (nobody)
Date: 2008-01-05 17:24
Message:
Logged In: NO
I view the html code and a user in pt.wn don't say me the truth about
sysop bot.
You can close the ticket :$
----------------------------------------------------------------------
Comment By: Rotem Liss (rotemliss)
Date: 2008-01-05 08:57
Message:
Logged In: YES
user_id=1327030
Originator: NO
It seems like an error Page.put didn't detect. Please replace line 1114 in
your copy of interwiki.py:
wikipedia.output(u'%s %s' % (status, reason))
(if this is not the line, update the file first) to:
wikipedia.output("Data: %s" % data)
(make sure to keep the spaces before the line!) and copy the results to
here. This will output the whole HTML data from the server in case of error
(note that it will contain many lines). Make sure to change the line to the
previous text before using interwiki.py again.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603141&aid=1864162&group_…
Feature Requests item #1939195, was opened at 2008-04-10 08:01
Message generated for change (Comment added) made by nobody
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603141&aid=1939195&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: Interface Improvements
Group: None
Status: Open
Priority: 5
Private: No
Submitted By: Nobody/Anonymous (nobody)
Assigned to: Nobody/Anonymous (nobody)
Summary: Translation for simple wiki
Initial Comment:
Please add the language entries for Simple English Wikipedia into the scripts. It can be annoying re-entering them after checking out the files.
I am Chenzw on Wikipedia.
----------------------------------------------------------------------
Comment By: Nobody/Anonymous (nobody)
Date: 2008-06-23 20:25
Message:
Logged In: NO
close this?
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603141&aid=1939195&group_…
Patches item #1880140, was opened at 2008-01-26 02:48
Message generated for change (Comment added) made by nobody
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603140&aid=1880140&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: AndreasJS (andreasjs)
Assigned to: Nobody/Anonymous (nobody)
Summary: Decode Esperanto titles
Initial Comment:
Titles have to be decoded from x convention for liks to work properly.
Here is the patch:
Index: wikipedia.py
===================================================================
--- wikipedia.py (revision 4939)
+++ wikipedia.py (working copy)
@@ -454,6 +454,8 @@
pass
if underscore:
title = title.replace(' ', '_')
+ if self.site().lang == 'eo':
+ title = decodeEsperantoX(title)
return title
def titleWithoutNamespace(self, underscore=False):
----------------------------------------------------------------------
Comment By: Nobody/Anonymous (nobody)
Date: 2008-06-23 20:22
Message:
Logged In: NO
close this **evil** patch?
----------------------------------------------------------------------
Comment By: Francesco Cosoleto (cosoleto)
Date: 2008-06-13 10:19
Message:
Logged In: YES
user_id=181280
Originator: NO
Applied in r5305, undone in r5563.
See [ 1988771 ] Encoding issues with Esperanto
(https://sourceforge.net/tracker/index.php?func=detail&aid=1988771&group_id=…)
----------------------------------------------------------------------
Comment By: Andr Malafaya Baptista (malafaya)
Date: 2008-01-26 14:50
Message:
Logged In: YES
user_id=1037345
Originator: NO
This seems to be working fine using "redirect.py -lang:eo double".
----------------------------------------------------------------------
Comment By: AndreasJS (andreasjs)
Date: 2008-01-26 14:37
Message:
Logged In: YES
user_id=1738850
Originator: YES
File Added: wikipedia.diff
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603140&aid=1880140&group_…
Patches item #1843798, was opened at 2007-12-04 02:45
Message generated for change (Comment added) made by nobody
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603140&aid=1843798&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Pietro Battiston (toobaz)
Assigned to: Nobody/Anonymous (nobody)
Summary: Add capabiliy to remember pages to replace.py
Initial Comment:
When doing very long semi-automatic replacements, it can happen to kill the bot and to start again. So you have to say "no" again to all non wanted replacements. It is even worse if you're using an xml dump: it can be several weeks old, and it will make you download lot of pages that where ALREADY corrected.
This patch consist in two parts:
1) a patch to replace.py that adds a new parameter, "-exclude", and makes it accept a path to a file which will be used both for:
-> knowing which articles to exclude from substitution
-> logging denied replaces' pages and pages already known to be not needing replacements
2) a patch to pagegenerators.py that adds a generator filter, able to yield only pages not appearing in a given list
The only doubt I have is: should the replace.py log in some other way? xml? wikipedia module's predefined functions? log into a given wikipedia userpage (so that logs can easily be shared)?
As I've done it, it needs to import os and codecs modules... don't know if it's a problem.
Anyway, a patch like this is something really needed, if needed I can try to improve it.
----------------------------------------------------------------------
Comment By: Nobody/Anonymous (nobody)
Date: 2008-06-23 20:22
Message:
Logged In: NO
closed this patch?
----------------------------------------------------------------------
Comment By: Nobody/Anonymous (nobody)
Date: 2008-01-16 14:20
Message:
Logged In: NO
replace.py already has the option -xmlstart:page when using an xml dump,
to skip all entries before "page".
----------------------------------------------------------------------
Comment By: Daniel Herding (wikipedian)
Date: 2008-01-16 12:35
Message:
Logged In: YES
user_id=880694
Originator: NO
We already have something very similar for solve_disambiguation.py. When
you run it with the -primary parameter, e.g. on [[en:London]], it saves all
page titles where the user pressed 'N' to the 'disambiguations' directory,
and skips these pages when you run the same command later.
It saves the URL-encoded titles into a text files, one title per line,
without [[brackets]].
It would be nice if some code could be shared, although I'm not sure if
that's possible (I haven't yet looked at your code, but
solve_disambiguation.py is a bit complicated). But we should keep
solve_disambiguation's format because there are probably people who want to
keep using their logs.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603140&aid=1843798&group_…