Bugs item #2061186, was opened at 2008-08-20 03:26
Message generated for change (Comment added) made by wikipedian
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: General
Group: None
>Status: Closed
>Resolution: Fixed
Priority: 8
Private: No
Submitted By: Woo-Jin Kim (kwj2772)
Assigned to: Nobody/Anonymous (nobody)
Summary: Special:Allpages now requires "to" parameter
Initial Comment:
I have found serious bug on interwiki.py
I'm using Pywikipedia revision 5816.
Error message:
C:\Python25\pywikipedia\interwiki.py -autonomous -lang:en -start:!
Checked for running processes. 1 process currently running, including currently process.
NOTE:Number of pages queued is 0, trying to add 60 more.
Retreiving Allpages special page for wikipedia:en from %21, namespace 0
NOTE:Nothing to left to do
Why nothing to left to do?
----------------------------------------------------------------------
>Comment By: Daniel Herding (wikipedian)
Date: 2008-08-20 17:24
Message:
Logged In: YES
user_id=880694
Originator: NO
Fixed in SVN, please test the new version on different wikis.
----------------------------------------------------------------------
Comment By: Francesco Cosoleto (cosoleto)
Date: 2008-08-20 11:56
Message:
Logged In: YES
user_id=181280
Originator: NO
We can use API with latest MediaWiki versions. If I am not wrong, it's
enabled by default when you install it. Why do more complicate and limited
things?
----------------------------------------------------------------------
Comment By: Andre Engels (a_engels)
Date: 2008-08-20 11:49
Message:
Logged In: YES
user_id=843018
Originator: NO
Direct quotes of the reactions:
"The User Interface is designed for users. The python Wikipedia
framework, therefore, should use the API, and then it won't have these
problems."
"The API has existed for awhile, and the UI should not be depended on for
any automated things."
In other words, they regard it as our own fault for not using the API.
----------------------------------------------------------------------
Comment By: Daniel Herding (wikipedian)
Date: 2008-08-20 11:43
Message:
Logged In: YES
user_id=880694
Originator: NO
The problem now is: In the new version of Special:Allpages, you need not
only a "from" parameter, but also a "to" parameter to get page titles. This
makes things really complicated for us, as we first need to find out the
title of the "to" page.
The following strategy could work:
* The user runs interwiki.py -start:Foo
* The bot loads
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo
* Using regular expressions or BeautifulSoap, the bot searches for lines
like
"Foo to Forbes Fictional 15", and follows the links.
*
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Forb…
is split again, so we need to do this recursively.
* Now
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Food…
is a normal page that we can parse with our existing code.
The nicer solution, of course, would be to add a "limit" parameter to
Special:Allpages, so that
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo&limit…
would give us the page in a useful format. But from what Andre wrote, I
guess that begging the MediaWiki devs for that won't have any effects.
----------------------------------------------------------------------
Comment By: Andre Engels (a_engels)
Date: 2008-08-20 10:56
Message:
Logged In: YES
user_id=843018
Originator: NO
Yes, Special:Allpages has changed, and we already got the "fuck off, we
don't care about your framework" response when complaining.
----------------------------------------------------------------------
Comment By: Multichill (multichill)
Date: 2008-08-20 10:50
Message:
Logged In: YES
user_id=1777493
Originator: NO
I first noticed this last night. I have this problem on different systems
(WinXP/BSD/Linux), all running the latest version. I noticed it with
imageuncat.py, but interwiki.py didnt work either. I doesnt seem to matter
which wiki you want to work on (nl and commons both didnt work).
Brion told me that Special:AllPages changed recently.
python version.py
Pywikipedia [http] trunk/pywikipedia (r5819, Aug 20 2008, 08:09:06)
Python 2.5.2 (r252:60911, Aug 14 2008, 13:31:58)
[GCC 4.3.1]
python imageuncat.py -start:Image:AChecked for running processes. 1
processes currently running, including the current process.
Retrieving Allpages special page for commons:commons from A, namespace 6
<done>
----------------------------------------------------------------------
Comment By: Mikko Silvonen (silvonen)
Date: 2008-08-20 07:15
Message:
Logged In: YES
user_id=127947
Originator: NO
Yep, the allpages method in wikipedia.py doesn't seem to find any pages,
so the -start parameter doesn't work at all. Has the format of
Special:Allpages changed, or what is causing this problem? But now it's
time for my day job...
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…
Revision: 5820
Author: wikipedian
Date: 2008-08-20 15:23:36 +0000 (Wed, 20 Aug 2008)
Log Message:
-----------
Rewrote allpages() using the MediaWiki API. Rewriting was easier than backporting
from the rewrite branch.
The old allpages() method is still available for older MediaWiki versions.
Modified Paths:
--------------
trunk/pywikipedia/wikipedia.py
Modified: trunk/pywikipedia/wikipedia.py
===================================================================
--- trunk/pywikipedia/wikipedia.py 2008-08-20 08:09:06 UTC (rev 5819)
+++ trunk/pywikipedia/wikipedia.py 2008-08-20 15:23:36 UTC (rev 5820)
@@ -5125,9 +5125,11 @@
if not repeat:
break
+
def allpages(self, start='!', namespace=0, includeredirects=True,
throttle=True):
- """Yield all Pages from Special:Allpages.
+ """
+ Yield all Pages in alphabetical order.
Parameters:
start Start at this page. By default, it starts at '!', and yields
@@ -5137,6 +5139,58 @@
at a time.
If includeredirects is False, redirects will not be found.
+
+ It is advised not to use this directly, but to use the
+ AllpagesPageGenerator from pagegenerators.py instead.
+
+ """
+ try:
+ api_url = self.api_address()
+ except NotImplementedError:
+ for page in self._allpagesOld(start, namespace, includeredirects, throttle):
+ yield page
+
+ rEntry = re.compile('<p pageid="\d+" ns="\d+" title="(?P<title>.*?)" />')
+
+ while True:
+ startEncoded = urllib.quote(start.encode(self.encoding()))
+ api_url += 'action=query&format=xml&list=allpages&apfrom=%s&aplimit=%i&apnamespace=%i' % (startEncoded, config.special_page_limit, namespace)
+ # TODO: support includeredirects="only" like in the old method
+ if not includeredirects:
+ api_url += '&apfilterredir=nonredirects'
+
+ if throttle:
+ get_throttle()
+ text = self.getUrl(api_url)
+
+ soup = BeautifulSoup(text,
+ convertEntities=BeautifulSoup.HTML_ENTITIES)
+
+ for p in soup.api.query.allpages:
+ yield Page(self, p['title'])
+
+ if soup.api.find('query-continue') is None:
+ # Last page reached.
+ break
+ start = soup.api.find('query-continue').allpages['apfrom']
+
+ def _allpagesOld(self, start='!', namespace=0, includeredirects=True,
+ throttle=True):
+ """
+ Yield all Pages from Special:Allpages.
+
+ This method doesn't work with MediaWiki 1.14 because of a change to
+ Special:Allpages. It is only left here for compatibility with older
+ MediaWiki versions, which don't support the API.
+
+ Parameters:
+ start Start at this page. By default, it starts at '!', and yields
+ all pages.
+ namespace Yield all pages in this namespace; defaults to 0.
+ MediaWiki software will only return pages in one namespace
+ at a time.
+
+ If includeredirects is False, redirects will not be found.
If includeredirects equals the string 'only', only redirects
will be found. Note that this has not been tested on older
versions of the MediaWiki code.
Patches item #2062542, was opened at 2008-08-20 16:05
Message generated for change (Tracker Item Submitted) made by Item Submitter
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603140&aid=2062542&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: Jitse Niesen (jitseniesen)
Assigned to: Nobody/Anonymous (nobody)
Summary: Typo in Page.move()
Initial Comment:
There is a typo in Page.move() where it refers to the variable predate instead of predata if fixredirects is False.
(I sent a message about this to the mailing list a few hours ago, but I'm not subscribed and it looks like the message is either rejected or held in a queue)
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603140&aid=2062542&group_…
Bugs item #2053373, was opened at 2008-08-16 00:53
Message generated for change (Comment added) made by ficell
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2053373&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: interwiki
Group: None
>Status: Closed
Resolution: Fixed
Priority: 5
Private: No
Submitted By: Yong-geun Jeong (ficell)
Assigned to: Nobody/Anonymous (nobody)
Summary: Pybot can't recognize translated magicwords.
Initial Comment:
http://ko.wikipedia.org/?diff=2194957&oldid=2194231
Recently, magicwords were translated into Korean language. But pywikipediabot can't recognize that. Pywikipediabot adding interwiki at redirects.
----------------------------------------------------------------------
>Comment By: Yong-geun Jeong (ficell)
Date: 2008-08-20 23:00
Message:
Logged In: YES
user_id=2182228
Originator: YES
It seems works fine.
----------------------------------------------------------------------
Comment By: Nobody/Anonymous (nobody)
Date: 2008-08-16 14:45
Message:
Logged In: NO
Yes. is Korean magicword for REDIRECT.
----------------------------------------------------------------------
Comment By: NicDumZ Nicolas Dumazet (nicdumz)
Date: 2008-08-16 01:38
Message:
Logged In: YES
user_id=1963242
Originator: NO
so if I get it right, is the ko: magic word for REDIRECT ?
I have added support for it in r5800 :)
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2053373&group_…
Feature Requests item #1916496, was opened at 2008-03-17 05:07
Message generated for change (Comment added) made by uberfuzzy
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603141&aid=1916496&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: None
Group: None
>Status: Closed
Priority: 5
Private: No
Submitted By: Uberfuzzy (uberfuzzy)
Assigned to: Nobody/Anonymous (nobody)
Summary: family.py patch to force monobook skin use
Initial Comment:
if the bot needs to the pages to be in monobook, rather then trusting the user to set the skin of the bot account to monobook, why dont we just make the bot get the pages in monobook via url?
this patch adds useskin=monobook into the urls in family.py
----------------------------------------------------------------------
>Comment By: Uberfuzzy (uberfuzzy)
Date: 2008-08-20 08:06
Message:
Logged In: YES
user_id=1976885
Originator: YES
was done in rev 5435
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603141&aid=1916496&group_…
Bugs item #2062179, was opened at 2008-08-20 13:35
Message generated for change (Tracker Item Submitted) made by Item Submitter
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2062179&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: category
Group: None
Status: Open
Resolution: None
Priority: 5
Private: No
Submitted By: stanlekub (stanlekub)
Assigned to: Nobody/Anonymous (nobody)
Summary: catlib.py should handle error pages returned by servers
Initial Comment:
I think there should be a minimal check in catlib.py to be sure that the page the server returned is a valid mediawiki one and not a basic error page.
In the current behaviour (r5811 for me) it silently passes through as if the given category was empty (obviously finding no valid wikilinks).
Maybe just searching for something like the "pt-logout" id in the text just before all the regex stuff could do the trick ?
Don't no if it's better to go into a retrying loop or raise an exception, I let you do this in a consitent manner with other parts of the code... or maybe you'll just think it's a bad idea :)
And sorry for my english, it isn't as good as I wishes it to be. :( Hope you'll understand anyway.
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2062179&group_…
Bugs item #2061186, was opened at 2008-08-20 03:26
Message generated for change (Comment added) made by cosoleto
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: General
Group: None
Status: Open
Resolution: None
Priority: 8
Private: No
Submitted By: Woo-Jin Kim (kwj2772)
Assigned to: Nobody/Anonymous (nobody)
Summary: Special:Allpages now requires "to" parameter
Initial Comment:
I have found serious bug on interwiki.py
I'm using Pywikipedia revision 5816.
Error message:
C:\Python25\pywikipedia\interwiki.py -autonomous -lang:en -start:!
Checked for running processes. 1 process currently running, including currently process.
NOTE:Number of pages queued is 0, trying to add 60 more.
Retreiving Allpages special page for wikipedia:en from %21, namespace 0
NOTE:Nothing to left to do
Why nothing to left to do?
----------------------------------------------------------------------
>Comment By: Francesco Cosoleto (cosoleto)
Date: 2008-08-20 11:56
Message:
Logged In: YES
user_id=181280
Originator: NO
We can use API with latest MediaWiki versions. If I am not wrong, it's
enabled by default when you install it. Why do more complicate and limited
things?
----------------------------------------------------------------------
Comment By: Andre Engels (a_engels)
Date: 2008-08-20 11:49
Message:
Logged In: YES
user_id=843018
Originator: NO
Direct quotes of the reactions:
"The User Interface is designed for users. The python Wikipedia
framework, therefore, should use the API, and then it won't have these
problems."
"The API has existed for awhile, and the UI should not be depended on for
any automated things."
In other words, they regard it as our own fault for not using the API.
----------------------------------------------------------------------
Comment By: Daniel Herding (wikipedian)
Date: 2008-08-20 11:43
Message:
Logged In: YES
user_id=880694
Originator: NO
The problem now is: In the new version of Special:Allpages, you need not
only a "from" parameter, but also a "to" parameter to get page titles. This
makes things really complicated for us, as we first need to find out the
title of the "to" page.
The following strategy could work:
* The user runs interwiki.py -start:Foo
* The bot loads
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo
* Using regular expressions or BeautifulSoap, the bot searches for lines
like
"Foo to Forbes Fictional 15", and follows the links.
*
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Forb…
is split again, so we need to do this recursively.
* Now
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Food…
is a normal page that we can parse with our existing code.
The nicer solution, of course, would be to add a "limit" parameter to
Special:Allpages, so that
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo&limit…
would give us the page in a useful format. But from what Andre wrote, I
guess that begging the MediaWiki devs for that won't have any effects.
----------------------------------------------------------------------
Comment By: Andre Engels (a_engels)
Date: 2008-08-20 10:56
Message:
Logged In: YES
user_id=843018
Originator: NO
Yes, Special:Allpages has changed, and we already got the "fuck off, we
don't care about your framework" response when complaining.
----------------------------------------------------------------------
Comment By: Multichill (multichill)
Date: 2008-08-20 10:50
Message:
Logged In: YES
user_id=1777493
Originator: NO
I first noticed this last night. I have this problem on different systems
(WinXP/BSD/Linux), all running the latest version. I noticed it with
imageuncat.py, but interwiki.py didnt work either. I doesnt seem to matter
which wiki you want to work on (nl and commons both didnt work).
Brion told me that Special:AllPages changed recently.
python version.py
Pywikipedia [http] trunk/pywikipedia (r5819, Aug 20 2008, 08:09:06)
Python 2.5.2 (r252:60911, Aug 14 2008, 13:31:58)
[GCC 4.3.1]
python imageuncat.py -start:Image:AChecked for running processes. 1
processes currently running, including the current process.
Retrieving Allpages special page for commons:commons from A, namespace 6
<done>
----------------------------------------------------------------------
Comment By: Mikko Silvonen (silvonen)
Date: 2008-08-20 07:15
Message:
Logged In: YES
user_id=127947
Originator: NO
Yep, the allpages method in wikipedia.py doesn't seem to find any pages,
so the -start parameter doesn't work at all. Has the format of
Special:Allpages changed, or what is causing this problem? But now it's
time for my day job...
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…
Bugs item #2061186, was opened at 2008-08-20 03:26
Message generated for change (Comment added) made by a_engels
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: General
Group: None
Status: Open
Resolution: None
Priority: 8
Private: No
Submitted By: Woo-Jin Kim (kwj2772)
Assigned to: Nobody/Anonymous (nobody)
Summary: Special:Allpages now requires "to" parameter
Initial Comment:
I have found serious bug on interwiki.py
I'm using Pywikipedia revision 5816.
Error message:
C:\Python25\pywikipedia\interwiki.py -autonomous -lang:en -start:!
Checked for running processes. 1 process currently running, including currently process.
NOTE:Number of pages queued is 0, trying to add 60 more.
Retreiving Allpages special page for wikipedia:en from %21, namespace 0
NOTE:Nothing to left to do
Why nothing to left to do?
----------------------------------------------------------------------
>Comment By: Andre Engels (a_engels)
Date: 2008-08-20 11:49
Message:
Logged In: YES
user_id=843018
Originator: NO
Direct quotes of the reactions:
"The User Interface is designed for users. The python Wikipedia
framework, therefore, should use the API, and then it won't have these
problems."
"The API has existed for awhile, and the UI should not be depended on for
any automated things."
In other words, they regard it as our own fault for not using the API.
----------------------------------------------------------------------
Comment By: Daniel Herding (wikipedian)
Date: 2008-08-20 11:43
Message:
Logged In: YES
user_id=880694
Originator: NO
The problem now is: In the new version of Special:Allpages, you need not
only a "from" parameter, but also a "to" parameter to get page titles. This
makes things really complicated for us, as we first need to find out the
title of the "to" page.
The following strategy could work:
* The user runs interwiki.py -start:Foo
* The bot loads
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo
* Using regular expressions or BeautifulSoap, the bot searches for lines
like
"Foo to Forbes Fictional 15", and follows the links.
*
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Forb…
is split again, so we need to do this recursively.
* Now
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Food…
is a normal page that we can parse with our existing code.
The nicer solution, of course, would be to add a "limit" parameter to
Special:Allpages, so that
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo&limit…
would give us the page in a useful format. But from what Andre wrote, I
guess that begging the MediaWiki devs for that won't have any effects.
----------------------------------------------------------------------
Comment By: Andre Engels (a_engels)
Date: 2008-08-20 10:56
Message:
Logged In: YES
user_id=843018
Originator: NO
Yes, Special:Allpages has changed, and we already got the "fuck off, we
don't care about your framework" response when complaining.
----------------------------------------------------------------------
Comment By: Multichill (multichill)
Date: 2008-08-20 10:50
Message:
Logged In: YES
user_id=1777493
Originator: NO
I first noticed this last night. I have this problem on different systems
(WinXP/BSD/Linux), all running the latest version. I noticed it with
imageuncat.py, but interwiki.py didnt work either. I doesnt seem to matter
which wiki you want to work on (nl and commons both didnt work).
Brion told me that Special:AllPages changed recently.
python version.py
Pywikipedia [http] trunk/pywikipedia (r5819, Aug 20 2008, 08:09:06)
Python 2.5.2 (r252:60911, Aug 14 2008, 13:31:58)
[GCC 4.3.1]
python imageuncat.py -start:Image:AChecked for running processes. 1
processes currently running, including the current process.
Retrieving Allpages special page for commons:commons from A, namespace 6
<done>
----------------------------------------------------------------------
Comment By: Mikko Silvonen (silvonen)
Date: 2008-08-20 07:15
Message:
Logged In: YES
user_id=127947
Originator: NO
Yep, the allpages method in wikipedia.py doesn't seem to find any pages,
so the -start parameter doesn't work at all. Has the format of
Special:Allpages changed, or what is causing this problem? But now it's
time for my day job...
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…
Bugs item #2061186, was opened at 2008-08-20 03:26
Message generated for change (Settings changed) made by wikipedian
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: General
Group: None
Status: Open
Resolution: None
Priority: 8
Private: No
Submitted By: Woo-Jin Kim (kwj2772)
Assigned to: Nobody/Anonymous (nobody)
>Summary: Special:Allpages now requires "to" parameter
Initial Comment:
I have found serious bug on interwiki.py
I'm using Pywikipedia revision 5816.
Error message:
C:\Python25\pywikipedia\interwiki.py -autonomous -lang:en -start:!
Checked for running processes. 1 process currently running, including currently process.
NOTE:Number of pages queued is 0, trying to add 60 more.
Retreiving Allpages special page for wikipedia:en from %21, namespace 0
NOTE:Nothing to left to do
Why nothing to left to do?
----------------------------------------------------------------------
Comment By: Daniel Herding (wikipedian)
Date: 2008-08-20 11:43
Message:
Logged In: YES
user_id=880694
Originator: NO
The problem now is: In the new version of Special:Allpages, you need not
only a "from" parameter, but also a "to" parameter to get page titles. This
makes things really complicated for us, as we first need to find out the
title of the "to" page.
The following strategy could work:
* The user runs interwiki.py -start:Foo
* The bot loads
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo
* Using regular expressions or BeautifulSoap, the bot searches for lines
like
"Foo to Forbes Fictional 15", and follows the links.
*
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Forb…
is split again, so we need to do this recursively.
* Now
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Food…
is a normal page that we can parse with our existing code.
The nicer solution, of course, would be to add a "limit" parameter to
Special:Allpages, so that
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo&limit…
would give us the page in a useful format. But from what Andre wrote, I
guess that begging the MediaWiki devs for that won't have any effects.
----------------------------------------------------------------------
Comment By: Andre Engels (a_engels)
Date: 2008-08-20 10:56
Message:
Logged In: YES
user_id=843018
Originator: NO
Yes, Special:Allpages has changed, and we already got the "fuck off, we
don't care about your framework" response when complaining.
----------------------------------------------------------------------
Comment By: Multichill (multichill)
Date: 2008-08-20 10:50
Message:
Logged In: YES
user_id=1777493
Originator: NO
I first noticed this last night. I have this problem on different systems
(WinXP/BSD/Linux), all running the latest version. I noticed it with
imageuncat.py, but interwiki.py didnt work either. I doesnt seem to matter
which wiki you want to work on (nl and commons both didnt work).
Brion told me that Special:AllPages changed recently.
python version.py
Pywikipedia [http] trunk/pywikipedia (r5819, Aug 20 2008, 08:09:06)
Python 2.5.2 (r252:60911, Aug 14 2008, 13:31:58)
[GCC 4.3.1]
python imageuncat.py -start:Image:AChecked for running processes. 1
processes currently running, including the current process.
Retrieving Allpages special page for commons:commons from A, namespace 6
<done>
----------------------------------------------------------------------
Comment By: Mikko Silvonen (silvonen)
Date: 2008-08-20 07:15
Message:
Logged In: YES
user_id=127947
Originator: NO
Yep, the allpages method in wikipedia.py doesn't seem to find any pages,
so the -start parameter doesn't work at all. Has the format of
Special:Allpages changed, or what is causing this problem? But now it's
time for my day job...
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…
Bugs item #2061186, was opened at 2008-08-20 03:26
Message generated for change (Comment added) made by wikipedian
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…
Please note that this message will contain a full copy of the comment thread,
including the initial issue submission, for this request,
not just the latest update.
Category: General
Group: None
Status: Open
Resolution: None
Priority: 8
Private: No
Submitted By: Woo-Jin Kim (kwj2772)
Assigned to: Nobody/Anonymous (nobody)
Summary: interwiki.py doesn't work well
Initial Comment:
I have found serious bug on interwiki.py
I'm using Pywikipedia revision 5816.
Error message:
C:\Python25\pywikipedia\interwiki.py -autonomous -lang:en -start:!
Checked for running processes. 1 process currently running, including currently process.
NOTE:Number of pages queued is 0, trying to add 60 more.
Retreiving Allpages special page for wikipedia:en from %21, namespace 0
NOTE:Nothing to left to do
Why nothing to left to do?
----------------------------------------------------------------------
>Comment By: Daniel Herding (wikipedian)
Date: 2008-08-20 11:43
Message:
Logged In: YES
user_id=880694
Originator: NO
The problem now is: In the new version of Special:Allpages, you need not
only a "from" parameter, but also a "to" parameter to get page titles. This
makes things really complicated for us, as we first need to find out the
title of the "to" page.
The following strategy could work:
* The user runs interwiki.py -start:Foo
* The bot loads
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo
* Using regular expressions or BeautifulSoap, the bot searches for lines
like
"Foo to Forbes Fictional 15", and follows the links.
*
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Forb…
is split again, so we need to do this recursively.
* Now
http://en.wikipedia.org/w/index.php?title=Special:AllPages&from=Foo&to=Food…
is a normal page that we can parse with our existing code.
The nicer solution, of course, would be to add a "limit" parameter to
Special:Allpages, so that
http://en.wikipedia.org/w/index.php?title=Special%3AAllPages&from=Foo&limit…
would give us the page in a useful format. But from what Andre wrote, I
guess that begging the MediaWiki devs for that won't have any effects.
----------------------------------------------------------------------
Comment By: Andre Engels (a_engels)
Date: 2008-08-20 10:56
Message:
Logged In: YES
user_id=843018
Originator: NO
Yes, Special:Allpages has changed, and we already got the "fuck off, we
don't care about your framework" response when complaining.
----------------------------------------------------------------------
Comment By: Multichill (multichill)
Date: 2008-08-20 10:50
Message:
Logged In: YES
user_id=1777493
Originator: NO
I first noticed this last night. I have this problem on different systems
(WinXP/BSD/Linux), all running the latest version. I noticed it with
imageuncat.py, but interwiki.py didnt work either. I doesnt seem to matter
which wiki you want to work on (nl and commons both didnt work).
Brion told me that Special:AllPages changed recently.
python version.py
Pywikipedia [http] trunk/pywikipedia (r5819, Aug 20 2008, 08:09:06)
Python 2.5.2 (r252:60911, Aug 14 2008, 13:31:58)
[GCC 4.3.1]
python imageuncat.py -start:Image:AChecked for running processes. 1
processes currently running, including the current process.
Retrieving Allpages special page for commons:commons from A, namespace 6
<done>
----------------------------------------------------------------------
Comment By: Mikko Silvonen (silvonen)
Date: 2008-08-20 07:15
Message:
Logged In: YES
user_id=127947
Originator: NO
Yep, the allpages method in wikipedia.py doesn't seem to find any pages,
so the -start parameter doesn't work at all. Has the format of
Special:Allpages changed, or what is causing this problem? But now it's
time for my day job...
----------------------------------------------------------------------
You can respond by visiting:
https://sourceforge.net/tracker/?func=detail&atid=603138&aid=2061186&group_…