https://bugzilla.wikimedia.org/show_bug.cgi?id=55259
Web browser: ---
Bug ID: 55259
Summary: xmlreader.py fails a lot
Product: Pywikibot
Version: unspecified
Hardware: All
OS: All
Status: NEW
Severity: normal
Priority: Unprioritized
Component: General
Assignee: Pywikipedia-bugs(a)lists.wikimedia.org
Reporter: legoktm.wikipedia(a)gmail.com
Classification: Unclassified
Mobile Platform: ---
Originally from: http://sourceforge.net/p/pywikipediabot/bugs/1245/
Reported by: emijrp
Created on: 2010-10-03 13:51:00
Subject: xmlreader.py fails a lot
Original description:
Hi all;
I think that there is an error in xmlreader.py. When parsing a full revision
XML \(in this case\[1\]\), using this code\[2\] \(look at the try-catch, it
writes in console when it fails\) I get correctly username, timestamp and
revisionid, but sometimes, the page title and the page id are None or empty
string.
The first error is:
\['', None, 'QuartierLatin1968', '2004-10-10T04:24:14Z', '4267'\] \#look the
empty string for the title, and the None for pageid
But if we do:
7za e -bd -so kwwiki-20100926-pages-meta-history.xml.7z 2>/dev/null | egrep
-i '2004-10-10T04::14Z' -C20
We get this\[3\], which is OK, the page title and the page id are available in
the XML, but not correctly parsed. And this is not the only page title and page
it that fails.
Perhaps I have missed something, because I'm learning to parsing XML. Sorry in
that case.
Regards,
emijrp
\[1\]
http://download.wikimedia.org/kwwiki/20100926/kwwiki-20100926-pages-meta-hi…
\[2\] http://pastebin.ca/1951930
\[3\] http://pastebin.ca/1951937
--
You are receiving this mail because:
You are the assignee for the bug.
https://bugzilla.wikimedia.org/show_bug.cgi?id=65238
Amir Ladsgroup <ladsgroup(a)gmail.com> changed:
What |Removed |Added
----------------------------------------------------------------------------
Priority|Unprioritized |Normal
CC| |ladsgroup(a)gmail.com
--
You are receiving this mail because:
You are the assignee for the bug.
https://bugzilla.wikimedia.org/show_bug.cgi?id=61783
Bug ID: 61783
Summary: Install fails in virtualenv
Product: Pywikibot
Version: unspecified
Hardware: All
OS: All
Status: UNCONFIRMED
Severity: normal
Priority: Unprioritized
Component: General
Assignee: Pywikipedia-bugs(a)lists.wikimedia.org
Reporter: kikkocristian(a)gmail.com
Web browser: ---
Mobile Platform: ---
Created attachment 14657
--> https://bugzilla.wikimedia.org/attachment.cgi?id=14657&action=edit
Detailed walktrough to reproduce the bug
Steps to reproduce:
* get pywikibot from http://tools.wmflabs.org/pywikibot/core.tar.gz
* tar extract that file
* create a virtualenv (e.g. pywikibottest)
* workon that virtualenv
* install the package following install instructions (basically python setup.py
install)
Import of pywikibot fails.
See attachment for details.
--
You are receiving this mail because:
You are the assignee for the bug.
https://bugzilla.wikimedia.org/show_bug.cgi?id=55243
Web browser: ---
Bug ID: 55243
Summary: replaceCategoryLinks misplaces featured article links
Product: Pywikibot
Version: unspecified
Hardware: All
OS: All
Status: NEW
Severity: normal
Priority: Unprioritized
Component: General
Assignee: Pywikipedia-bugs(a)lists.wikimedia.org
Reporter: legoktm.wikipedia(a)gmail.com
Classification: Unclassified
Mobile Platform: ---
Originally from: http://sourceforge.net/p/pywikipediabot/bugs/1320/
Reported by: silvonen
Created on: 2011-06-12 10:47:24
Subject: replaceCategoryLinks misplaces featured article links
Original description:
I don't know if this is a known problem, but the function replaceCategoryLinks
in pywikibot/textlib.py moves featured/good article templates \(e.g., \{\{Link
FA|en\}\}\) above categories, although these templates are usually placed
between categories and interwiki links. This has caused problems in one of the
bot scripts that we use in the Finnish Wikipedia.
I am attaching a patch proposed by zache.
>python version.py
Pywikipedia \[http\] trunk/pywikipedia \(r9291, 2011/06/10, 04:46:24\)
Python 2.5.4 \(r254:67916, Jan 29 2009, 12:02:11\) \[MSC v.1310 32 bit
\(Intel\)\]
config-settings:
use\_api = True
use\_api\_login = True
unicode test: ok
--
You are receiving this mail because:
You are the assignee for the bug.
https://bugzilla.wikimedia.org/show_bug.cgi?id=55654
Web browser: ---
Bug ID: 55654
Summary: Recursive clone over ssh fails
Product: Pywikibot
Version: core (2.0)
Hardware: All
OS: All
Status: NEW
Severity: normal
Priority: Unprioritized
Component: General
Assignee: Pywikipedia-bugs(a)lists.wikimedia.org
Reporter: maarten(a)mdammers.nl
Classification: Unclassified
Mobile Platform: ---
<~/pywikibot>git clone --recursive
ssh://multichill@gerrit.wikimedia.org:29418/pywikibot/core.git
Cloning into 'core'...
Warning: Permanently added the RSA host key for IP address
'[208.80.154.81]:29418' to the list of known hosts.
Enter passphrase for key 'key':
remote: Counting objects: 10842, done
remote: Finding sources: 100% (10842/10842)
remote: Getting sizes: 100% (1495/1495)
remote: Compressing objects: 100% (2292733/2292733)
remote: Total 10842 (delta 7190), reused 10548 (delta 7165)
Receiving objects: 100% (10842/10842), 23.42 MiB | 326 KiB/s, done.
Resolving deltas: 100% (7337/7337), done.
Submodule 'externals/httplib2'
(https://gerrit.wikimedia.org/r/p/pywikibot/externals/httplib2.git) registered
for path 'externals/httplib2'
Submodule 'i18n' (https://gerrit.wikimedia.org/r/p/pywikibot/i18n.git)
registered for path 'scripts/i18n'
Cloning into 'externals/httplib2'...
error: SSL certificate problem: self signed certificate in certificate chain
while accessing
https://gerrit.wikimedia.org/r/p/pywikibot/externals/httplib2.git/info/refs…
fatal: HTTP request failed
Clone of 'https://gerrit.wikimedia.org/r/p/pywikibot/externals/httplib2.git'
into submodule path 'externals/httplib2' failed
I'm cloning over ssh. Why are the submodules being loaded over https?
--
You are receiving this mail because:
You are the assignee for the bug.
https://bugzilla.wikimedia.org/show_bug.cgi?id=59080
Web browser: ---
Bug ID: 59080
Summary: isCategoryRedirect is very slow
Product: Pywikibot
Version: core (2.0)
Hardware: All
OS: All
Status: NEW
Severity: normal
Priority: Unprioritized
Component: General
Assignee: Pywikipedia-bugs(a)lists.wikimedia.org
Reporter: maarten(a)mdammers.nl
Classification: Unclassified
Mobile Platform: ---
In page.py we have a function "isCategoryRedirect":
def isCategoryRedirect(self):
"""Return True if this is a category redirect page, False otherwise."""
This is rather slow. Maybe some caching should be used.
--
You are receiving this mail because:
You are the assignee for the bug.
https://bugzilla.wikimedia.org/show_bug.cgi?id=58878
Web browser: ---
Bug ID: 58878
Summary: Report sensible error when family file cannot be
loaded
Product: Pywikibot
Version: core (2.0)
Hardware: All
OS: All
Status: NEW
Severity: normal
Priority: Unprioritized
Component: General
Assignee: Pywikipedia-bugs(a)lists.wikimedia.org
Reporter: valhallasw(a)arctus.nl
Classification: Unclassified
Mobile Platform: ---
"pywikibot.exceptions.Error: Family dibaf does not exist" is not a helpful
message if the problem really is
'cannot import family' (because someone tries 'import family' instead of 'from
pywikibot import family'
or something like that.
--
You are receiving this mail because:
You are the assignee for the bug.
https://bugzilla.wikimedia.org/show_bug.cgi?id=56964
Web browser: ---
Bug ID: 56964
Summary: Tests: create a logger at DEBUG level to log to a file
Product: Pywikibot
Version: core (2.0)
Hardware: All
OS: All
Status: NEW
Severity: normal
Priority: Unprioritized
Component: General
Assignee: Pywikipedia-bugs(a)lists.wikimedia.org
Reporter: valhallasw(a)arctus.nl
Classification: Unclassified
Mobile Platform: ---
To solve unit test issues, it can be very useful to have a debug log that
includes all network connectivity et cetera. This should be fairly easy -- just
create a logging hander that catches all logging at level DEBUG (or lower), and
stores these to a file.
Basically, this is what -debug does in bot.py:handleArgs.
--
You are receiving this mail because:
You are the assignee for the bug.
https://bugzilla.wikimedia.org/show_bug.cgi?id=55308
Web browser: ---
Bug ID: 55308
Summary: Fullurl in imageharvest.py does not work
Product: Pywikibot
Version: unspecified
Hardware: All
OS: All
Status: NEW
Severity: normal
Priority: Unprioritized
Component: General
Assignee: Pywikipedia-bugs(a)lists.wikimedia.org
Reporter: legoktm.wikipedia(a)gmail.com
Classification: Unclassified
Mobile Platform: ---
Originally from: http://sourceforge.net/p/pywikipediabot/bugs/897/
Reported by: dasch88
Created on: 2009-04-04 09:54:06
Subject: Fullurl in imageharvest.py does not work
Original description:
When I have a page where the picturelinks are given as fullurl they are
appended to the pageurl
Skript progess:
python imageharvest.py http://commons.wikimedia.org/wiki/User:DaSch/Bot-Export
Checked for running processes. 1 processes currently running, including the
current process.
What text should be added at the end of the description of each image from this
url? \[\[Import::04.04.2009\]\] \[\[Kategorie:Commons-Import\]\]
Include image
http://commons.wikimedia.org/wiki/User:DaSch/http://commons.wikipedia.org/a…
\(\[y\]es, \[N\]o, \[s\]top\) N
Include image
http://commons.wikimedia.org/wiki/User:DaSch/http://upload.wikimedia.org/wi…
\(\[y\]es, \[N\]o, \[s\]top\) y
Give the description of this image: 1789
Reading file
http://commons.wikimedia.org/wiki/User:DaSch/http://upload.wikimedia.org/wi…
Traceback \(most recent call last\):
File "imageharvest.py", line 117, in <module>
main\(url,image\_url\)
File "imageharvest.py", line 95, in main
uploadBot.run\(\)
File "/root/pywecowi/upload.py", line 296, in run
return self.upload\_image\(\)
File "/root/pywecowi/upload.py", line 119, in upload\_image
file = uo.open\(self.url\)
File "/usr/lib/python2.5/urllib.py", line 190, in open
return getattr\(self, name\)\(url\)
File "/usr/lib/python2.5/urllib.py", line 338, in open\_http
return self.http\_error\(url, fp, errcode, errmsg, headers\)
File "/usr/lib/python2.5/urllib.py", line 355, in http\_error
return self.http\_error\_default\(url, fp, errcode, errmsg, headers\)
File "/root/pywecowi/wikipedia.py", line 6954, in http\_error\_default
raise PageNotFound\(u'Page %s could not be retrieved. Check your family file ?'
% url\)
wikipedia.PageNotFound: Page
//commons.wikimedia.org/wiki/User:DaSch/http://upload.wikimedia.org/wikipedia/commons/1/1b/electoralcollege1789.svg
could not be retrieved. Check your family file ?
Version:
Pywikipedia \[http\] trunk/pywikipedia \(r6569, Apr 03 2009, 16:17:41\)
Python 2.5.2 \(r252:60911, Jul 31 2008, 17:28:52\)
\[GCC 4.2.3 \(Ubuntu 4.2.3-2ubuntu7\)\]
--
You are receiving this mail because:
You are the assignee for the bug.
https://bugzilla.wikimedia.org/show_bug.cgi?id=55041
Web browser: ---
Bug ID: 55041
Summary: Cosmetic_changes.py deletes cross wiki-links
Product: Pywikibot
Version: unspecified
Hardware: All
OS: All
Status: NEW
Severity: enhancement
Priority: Unprioritized
Component: interwiki.py
Assignee: Pywikipedia-bugs(a)lists.wikimedia.org
Reporter: legoktm.wikipedia(a)gmail.com
Classification: Unclassified
Mobile Platform: ---
Originally from: http://sourceforge.net/p/pywikipediabot/feature-requests/296/
Reported by: Anonymous user
Created on: 2011-12-26 21:47:33
Subject: Cosmetic_changes.py deletes cross wiki-links
Original description:
Python 2.6.7 \(r267:88850, Sep 19 2011, 13:25:28\)
\[GCC 4.5.2\]
config-settings:
use\_api = True
use\_api\_login = True
unicode test: ok
1\. I ran in RU-wiki one command:
python /home/$USERNAME/pywiki/cosmetic\_changes.py -lang:ru -always
-file:/tmp/somefile
2\. File "/tmp/somefile" contain list of articles for processing
Struthiomimus
QoS
3\. For article "Struthiomimus" in RU-wiki I see, that it was deleted EN
cross-link
\[\[en:Steveville|Steveville\]\]
https://secure.wikimedia.org/wikipedia/ru/w/index.php?title=Struthiomimus&d…
4\. I think, that cosmetic\_changes.py should not deletes cross-links to other
wikis.
--
You are receiving this mail because:
You are the assignee for the bug.