Bugs item #1853805, was opened at 2007-12-19 04:37
Message generated for change (Comment added) made by russblau
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853805&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Nicolas Dumazet (nicdumz)
Assigned to: Nobody/Anonymous (nobody)
Summary: redirect.py -namespace:X crashes
Initial Comment:
Hi !
Title is quite self explanatory.
It was a simple typo + a wrong initialization.
Patch is included :)
Thanks,
Nicolas.
----------------------------------------------------------------------
Comment By: Russell Blau (russblau)
Date: 2007-12-19 08:57
Message:
Logged In: YES
user_id=855050
Originator: NO
Fixed in r4735; thanks for the patch!
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853805&group_…
Revision: 4735
Author: russblau
Date: 2007-12-19 13:56:39 +0000 (Wed, 19 Dec 2007)
Log Message:
-----------
Fixing bug #1853805; thanks nicdumz for the patch!
Modified Paths:
--------------
trunk/pywikipedia/redirect.py
Modified: trunk/pywikipedia/redirect.py
===================================================================
--- trunk/pywikipedia/redirect.py 2007-12-18 18:09:38 UTC (rev 4734)
+++ trunk/pywikipedia/redirect.py 2007-12-19 13:56:39 UTC (rev 4735)
@@ -338,7 +338,7 @@
xmlFilename = None
# Which namespace should be processed when using a XML dump
# default to -1 which means all namespaces will be processed
- namespace = None
+ namespaces = []
# at which redirect shall we start searching double redirects again (only with dump)
# default to -1 which means all redirects are checked
offset = -1
@@ -368,7 +368,7 @@
if not action:
wikipedia.showHelp('redirect')
else:
- gen = RedirectGenerator(xmlFilename, namespace, offset)
+ gen = RedirectGenerator(xmlFilename, namespaces, offset)
bot = RedirectRobot(action, gen, always)
bot.run()
Bugs item #1853805, was opened at 2007-12-19 10:37
Message generated for change (Tracker Item Submitted) made by Item Submitter
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853805&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Nicolas Dumazet (nicdumz)
Assigned to: Nobody/Anonymous (nobody)
Summary: redirect.py -namespace:X crashes
Initial Comment:
Hi !
Title is quite self explanatory.
It was a simple typo + a wrong initialization.
Patch is included :)
Thanks,
Nicolas.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853805&group_…
Bugs item #1853748, was opened at 2007-12-19 09:16
Message generated for change (Tracker Item Submitted) made by Item Submitter
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853748&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: General
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: pixelFire (pixelfire)
Assigned to: Nobody/Anonymous (nobody)
Summary: Login to hy.wikipedia impossible
Initial Comment:
It's impossible to log on to hy.wikipedia with the Bot.
Password for user XYBot on wikipedia:hy:
Logging in to wikipedia:hy as XYBot
Login failed. Wrong password or CAPTCHA answer?
Username and Password are correct of course.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853748&group_…
Bugs item #1853415, was opened at 2007-12-18 20:52
Message generated for change (Comment added) made by a_engels
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853415&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: General
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Alleborgo (alleborgo)
Assigned to: Nobody/Anonymous (nobody)
Summary: interwiki - the page doesn't exists but exist!
Initial Comment:
When I run interwiki.py on the page [[zh:路易斯·范加爾]] it say me that the page doesn't exist but the page exists!
I don't know if is a problem of wikipedia.py or interwiki.py script.
----------------------------------------------------------------------
>Comment By: Andre Engels (a_engels)
Date: 2007-12-19 09:07
Message:
Logged In: YES
user_id=843018
Originator: NO
When working on a Chinese page, you need to always use the simplified
title. Try to work on [[zh:路易斯·范加尔]] instead, and things
should work just fine. The reason for this is that the bot uses
[[Special:Export]], and there only the 'original' (simplified) title is
kept. See http://bugzilla.wikimedia.org/show_bug.cgi?id=9419.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853415&group_…
Bugs item #1853415, was opened at 2007-12-18 20:52
Message generated for change (Tracker Item Submitted) made by Item Submitter
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853415&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: General
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Alleborgo (alleborgo)
Assigned to: Nobody/Anonymous (nobody)
Summary: interwiki - the page doesn't exists but exist!
Initial Comment:
When I run interwiki.py on the page [[zh:路易斯·范加爾]] it say me that the page doesn't exist but the page exists!
I don't know if is a problem of wikipedia.py or interwiki.py script.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1853415&group_…
Bugs item #1852276, was opened at 2007-12-17 14:50
Message generated for change (Comment added) made by rotemliss
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1852276&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Purodha B Blissenbach (purodha)
Assigned to: Nobody/Anonymous (nobody)
Summary: Endless loop in weblinkchecker.py
Initial Comment:
$ python weblinkchecker.py -start:00f -family:wikipedia -lang:ksh -v
Checked for running processes. 1 processes currently running, including the current process.
Pywikipediabot (r4720 (wikipedia.py), Dec 15 2007, 18:57:27)
Python 2.4.4 (#2, Aug 16 2007, 00:34:54)
[GCC 4.1.3 20070812 (prerelease) (Debian 4.1.2-15)]
Retrieving Allpages special page for wikipedia:ksh from 00f, namespace 0
Retrieving Allpages special page for wikipedia:ksh from 00f, namespace 0
Retrieving Allpages special page for wikipedia:ksh from 00f, namespace 0
... ad infinitum ...
This may be related to bug #1852173 in this tracker.
Note, a page "00f" does not exist. I was expecting to start at the next available page, which, as a matter of fact, does exists.
----------------------------------------------------------------------
Comment By: Rotem Liss (rotemliss)
Date: 2007-12-18 20:18
Message:
Logged In: YES
user_id=1327030
Originator: NO
It was indeed the same problem as bug #1852173, and was fixed with it.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1852276&group_…
Bugs item #1852173, was opened at 2007-12-17 10:57
Message generated for change (Comment added) made by rotemliss
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1852173&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: other
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Purodha B Blissenbach (purodha)
Assigned to: Nobody/Anonymous (nobody)
Summary: utf-8 coding problem, kills weblinkchecker
Initial Comment:
Weblinkchecker chokes on many instances when reading the Special:Allpages of the ksh Wikipedia. It claims to see non-unicode data, which is unlikely to be so. I did not dig the code in detail atm, and I cannot tell the offending byte sequences atm.
Here is a copy of a linux command line, and the output generated by it:
$~> python weblinkchecker.py -putthrottle:300 -start:00er -v -family:wikipedia -lang:ksh
Checked for running processes. 1 processes currently running, including the current process.
Pywikipediabot (r4720 (wikipedia.py), Dec 15 2007, 18:57:27)
Python 2.4.4 (#2, Aug 16 2007, 00:34:54)
[GCC 4.1.3 20070812 (prerelease) (Debian 4.1.2-15)]
Retrieving Allpages special page for wikipedia:ksh from 00er, namespace 0
Retrieving Allpages special page for wikipedia:ksh from 00er%20Joare%20%28Watt%20%C4%97%C3%9F%C3%9F%20datt%3F%29%21, namespace 0
Retrieving Allpages special page for wikipedia:ksh from 00er%2520Joare%2520%2528Watt%2520%25C4%2597%25C3%259F%25C3%259F%2520datt%253F%2529%2521, namespace 0
DBG> BUG: Non-unicode passed to wikipedia.output without decoder!
File "threading.py", line 442, in __bootstrap
self.run()
File "/home/purodha/pywikipedia/pagegenerators.py", line 632, in run
wikipedia.output(str(e))
File "/home/purodha/pywikipedia/wikipedia.py", line 5351, in output
print traceback.print_stack()
None
DBG> Attempting to recover, but please report this problem
Couldn't extract allpages special page. Make sure you're using MonoBook skin.
Saving history...
I had made sure, the user [[:ksh:User:Weblinkchcker]] was logged in, using the monobook skin, and the English interface language.
I could not make sure that weblinkchecker does use his user account while reading only. A test revealed that there is no apparent difference in behaviour when I rename login-data/wikipedia-ksh-Weblinkchecker-login.data to something else.
If there are questions, I am prepared to provide more info, once I know where to look.
----------------------------------------------------------------------
Comment By: Rotem Liss (rotemliss)
Date: 2007-12-18 20:13
Message:
Logged In: YES
user_id=1327030
Originator: NO
The problem was that you found a very special case in Allpages, in which
you didn't get one single page of the kind you requested
(redirect/non-redirect), so you didn't get the next page but double-encoded
the current one. Fixed this in r4734. Note that the UTF-8 problem isn't
related to this bug (the "Couldn't extract allpages special page. Make sure
you're using MonoBook skin." message is related, and the problem was it was
a string, not a unicode, passed to wikipedia.output), but was fixed in
r4733.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=1852173&group_…
Revision: 4734
Author: rotem
Date: 2007-12-18 18:09:38 +0000 (Tue, 18 Dec 2007)
Log Message:
-----------
(bug 1852173) Fix a problem of a very special case in Allpages, when no pages of the requested type (redirect/non-redirect) are found, and the start parameter remains and gets double-encoded.
Modified Paths:
--------------
trunk/pywikipedia/wikipedia.py
Modified: trunk/pywikipedia/wikipedia.py
===================================================================
--- trunk/pywikipedia/wikipedia.py 2007-12-18 17:48:42 UTC (rev 4733)
+++ trunk/pywikipedia/wikipedia.py 2007-12-18 18:09:38 UTC (rev 4734)
@@ -4444,8 +4444,14 @@
if (not includeredirects) or includeredirects == 'only':
# Maybe there were only so few because the rest is or is not a redirect
R = re.compile('title ?=\"(.*?)\"')
- if len(R.findall(returned_html)) < 100:
+ allLinks = R.findall(returned_html)
+ if len(allLinks) < 100:
break
+ elif n == 0:
+ # In this special case, no pages of the requested type
+ # were found, and "start" will remain and be double-encoded.
+ # Use the last page as the start of the next page.
+ start = Page(self, allLinks[-1]).titleWithoutNamespace() + '!'
else:
break