Xqt has submitted this change and it was merged.
Change subject: [PEP8]ify wikipedia.py
......................................................................
[PEP8]ify wikipedia.py
Change-Id: I39bdae87d2f8797bf09709d5b741b4388fc1a1e3
---
M wikipedia.py
1 file changed, 11 insertions(+), 16 deletions(-)
Approvals:
Xqt: Looks good to me, approved
jenkins-bot: Verified
diff --git a/wikipedia.py b/wikipedia.py
index a9d3a58..3179f1c 100644
--- a/wikipedia.py
+++ b/wikipedia.py
@@ -6545,18 +6545,14 @@
def _userIndex(self, sysop=False):
"""Returns the internal index of the user."""
- if sysop:
- return 1
- else:
- return 0
+ return 1 if sysop else 0
def username(self, sysop=False):
return self._userName[self._userIndex(sysop=sysop)]
def sitename(self):
"""Return string representing this Site's name and code."""
-
- return self.family.name+':'+self.code
+ return '%s:%s' % (self.family.name, self.code)
def __repr__(self):
return '%s:%s' % (self.family.name, self.code)
@@ -7029,7 +7025,8 @@
raise
except urllib2.HTTPError, e:
if e.code in [401, 404]:
- debug(u"Got HTTP/%i %s: \n %r" % (e.code, e.message, e.read()))
+ debug(u"Got HTTP/%i %s: \n %r"
+ % (e.code, e.message, e.read()))
raise PageNotFound(u'Page %s could not be retrieved. Check '
u'your family file ?' % url)
# just check for HTTP Status 500 (Internal Server Error)?
@@ -8350,7 +8347,7 @@
fileext = None
title = m.group('title')
if extension:
- fileext = title[len(title)-3:]
+ fileext = title[len(title) - 3:]
if title not in seen and fileext == extension:
## Check whether the media is used in a Proofread page
# code disabled because it slows this method down, and
@@ -8635,7 +8632,6 @@
url += '&namespace=%s' % namespace
parser_text = self.getUrl(url)
while True:
- # <li><a href="/wiki/Pagina_principale" title="Pagina principale">Pagina principale</a> <small>(6.522 byte)</small> (protetta)</li>
m = re.findall(
r'<li><a href=".*?" title=".*?">(.*?)</a>.*?<small>\((.*?)\)'
r'</small>.*?\((.*?)\)</li>', parser_text)
@@ -9613,7 +9609,7 @@
# the parameter moduleName is deprecated and should be left out.
moduleName = moduleName or calledModuleName()
try:
- moduleName = moduleName[moduleName.rindex("\\")+1:]
+ moduleName = moduleName[moduleName.rindex("\\") + 1:]
except ValueError:
# There was no \ in the module name, so presumably no problem
pass
@@ -9695,8 +9691,8 @@
# TEST for bug #3081100
unicode_error = __import__('unicodedata').normalize(
'NFC',
+ u'\u092e\u093e\u0930\u094d\u0915 \u091c\u093c\u0941\u0915\u0947\u0930\u092c\u0930\u094d\u0917') != \
u'\u092e\u093e\u0930\u094d\u0915 \u091c\u093c\u0941\u0915\u0947\u0930\u092c\u0930\u094d\u0917'
- ) != u'\u092e\u093e\u0930\u094d\u0915 \u091c\u093c\u0941\u0915\u0947\u0930\u092c\u0930\u094d\u0917'
if unicode_error:
print u'unicode test: triggers problem #3081100'
@@ -9859,6 +9855,7 @@
CRITICAL - fatal error messages
Accordingly, do ''not'' use print statements in bot code; instead,
use pywikibot.output function.
+
"""
# currently only the logger is initialized here
# the handlers are initialized in setLogfileStatus
@@ -9878,24 +9875,22 @@
nh = NullHandler()
logger.addHandler(nh)
- logger.setLevel(DEBUG+1)
+ logger.setLevel(DEBUG + 1)
if hasattr(logger, 'captureWarnings'):
logger.captureWarnings(True) # introduced in Python >= 2.7
logger = logging.getLogger('pywiki')
logger.addHandler(nh)
- logger.setLevel(DEBUG+1)
-
+ logger.setLevel(DEBUG + 1)
logger.propagate = True
-
- return
def writelogheader():
"""
Save additional version, system and status info to the logfile in use,
so that the user can look it up later to track errors or report bugs.
+
"""
output(u'=== Pywikipediabot framework v1.0 -- Logging header ===')
--
To view, visit https://gerrit.wikimedia.org/r/95207
To unsubscribe, visit https://gerrit.wikimedia.org/r/settings
Gerrit-MessageType: merged
Gerrit-Change-Id: I39bdae87d2f8797bf09709d5b741b4388fc1a1e3
Gerrit-PatchSet: 1
Gerrit-Project: pywikibot/compat
Gerrit-Branch: master
Gerrit-Owner: Xqt <info(a)gno.de>
Gerrit-Reviewer: Xqt <info(a)gno.de>
Gerrit-Reviewer: jenkins-bot
jenkins-bot has submitted this change and it was merged.
Change subject: ignore pep8 for catimages.py for now which is really hard
......................................................................
ignore pep8 for catimages.py for now which is really hard
Change-Id: I3c5d368797a1da864a4f284ba41d7e2c6f7e832e
---
M .pep8
1 file changed, 1 insertion(+), 1 deletion(-)
Approvals:
Xqt: Looks good to me, approved
jenkins-bot: Verified
diff --git a/.pep8 b/.pep8
index 8e83dbb..4f1391e 100644
--- a/.pep8
+++ b/.pep8
@@ -1,4 +1,4 @@
[pep8]
ignore = E122,E127,E241,E502
-exclude = ./wiktionary.py,./wiktionary/*,./tests/*,./userinterfaces/*,./commonsdelinker/*,./families/*,./maintenance/*,./pywikibot/*,./externals/*
+exclude = ./catimages.py,./wiktionary.py,./wiktionary/*,./tests/*,./userinterfaces/*,./commonsdelinker/*,./families/*,./maintenance/*,./pywikibot/*,./externals/*
max_line_length = 256
--
To view, visit https://gerrit.wikimedia.org/r/95182
To unsubscribe, visit https://gerrit.wikimedia.org/r/settings
Gerrit-MessageType: merged
Gerrit-Change-Id: I3c5d368797a1da864a4f284ba41d7e2c6f7e832e
Gerrit-PatchSet: 1
Gerrit-Project: pywikibot/compat
Gerrit-Branch: master
Gerrit-Owner: Xqt <info(a)gno.de>
Gerrit-Reviewer: Xqt <info(a)gno.de>
Gerrit-Reviewer: jenkins-bot
jenkins-bot has submitted this change and it was merged.
Change subject: [PEP8] pep8 changes
......................................................................
[PEP8] pep8 changes
Change-Id: I892ebfcbcabcfcd9ac1e6c899cd75029fc4d2993
---
M casechecker.py
1 file changed, 184 insertions(+), 138 deletions(-)
Approvals:
Xqt: Looks good to me, approved
jenkins-bot: Verified
diff --git a/casechecker.py b/casechecker.py
index 094f5ab..80fede6 100644
--- a/casechecker.py
+++ b/casechecker.py
@@ -4,17 +4,27 @@
with mixed latin and cyrilic alphabets.
"""
#
-# (C) Pywikipedia bot team, 2006-2012
+# (C) Pywikibot team, 2006-2013
#
# Distributed under the terms of the MIT license.
#
__version__ = '$Id$'
+import os
+import sys
+import query
+import re
+import codecs
+import wikipedia as pywikibot
+
+
#
-# Permutations code was taken from http://aspn.activestate.com/ASPN/Cookbook/Python/Recipe/190465
+# Permutations code was taken from
+# http://aspn.activestate.com/ASPN/Cookbook/Python/Recipe/190465
#
def xuniqueCombinations(items, n):
- if n==0: yield []
+ if n == 0:
+ yield []
else:
for i in xrange(len(items)):
for cc in xuniqueCombinations(items[i+1:], n-1):
@@ -29,16 +39,16 @@
#
# Adapted from http://aspn.activestate.com/ASPN/Cookbook/Python/Recipe/496901
#
-STD_OUTPUT_HANDLE= -11
+STD_OUTPUT_HANDLE = -11
-FOREGROUND_BLUE = 0x01 # text color contains blue.
-FOREGROUND_GREEN= 0x02 # text color contains green.
-FOREGROUND_RED = 0x04 # text color contains red.
-FOREGROUND_INTENSITY = 0x08 # text color is intensified.
-BACKGROUND_BLUE = 0x10 # background color contains blue.
-BACKGROUND_GREEN= 0x20 # background color contains green.
-BACKGROUND_RED = 0x40 # background color contains red.
-BACKGROUND_INTENSITY = 0x80 # background color is intensified.
+FOREGROUND_BLUE = 0x01 # text color contains blue.
+FOREGROUND_GREEN = 0x02 # text color contains green.
+FOREGROUND_RED = 0x04 # text color contains red.
+FOREGROUND_INTENSITY = 0x08 # text color is intensified.
+BACKGROUND_BLUE = 0x10 # background color contains blue.
+BACKGROUND_GREEN = 0x20 # background color contains green.
+BACKGROUND_RED = 0x40 # background color contains red.
+BACKGROUND_INTENSITY = 0x80 # background color is intensified.
FOREGROUND_WHITE = FOREGROUND_BLUE | FOREGROUND_GREEN | FOREGROUND_RED
@@ -57,18 +67,17 @@
except:
pass
- if color == FOREGROUND_BLUE: print '(b:',
- if color == FOREGROUND_GREEN: print '(g:',
- if color == FOREGROUND_RED: print '(r:',
+ if color == FOREGROUND_BLUE:
+ print '(b:',
+ if color == FOREGROUND_GREEN:
+ print '(g:',
+ if color == FOREGROUND_RED:
+ print '(r:',
# end of console code
-import os
-import sys, query, re, codecs
-import wikipedia as pywikibot
-
-class CaseChecker( object ):
+class CaseChecker(object):
msgRename = {
'ar': u'تغيير اسم لحالة مخلوطة',
'en': u'mixed case rename',
@@ -90,8 +99,8 @@
# These words are always in one language, even though they could be typed
# in both
- alwaysInLocal = [ u'СССР', u'Как', u'как' ]
- alwaysInLatin = [ u'II', u'III' ]
+ alwaysInLocal = [u'СССР', u'Как', u'как']
+ alwaysInLatin = [u'II', u'III']
localUpperLtr = u'ЁІЇЎАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯҐ'
localLowerLtr = u'ёіїўабвгдежзийклмнопрстуфхцчшщъыьэюяґ'
@@ -113,7 +122,7 @@
whitelists = {
'ru': u'ВП:КЛ/Проверенные',
- }
+ }
latLtr = u'abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ'
@@ -167,7 +176,7 @@
elif arg == '-autonomous' or arg == '-a':
self.autonomous = True
elif arg.startswith('-ns:'):
- self.namespaces.append( int(arg[4:]) )
+ self.namespaces.append(int(arg[4:]))
elif arg.startswith('-wikilog:'):
self.wikilogfile = arg[9:]
elif arg.startswith('-failedlog:'):
@@ -191,13 +200,14 @@
self.aplimit = 200 if self.links else 'max'
if not self.doFailed:
- self.queryParams = { 'action' : 'query',
- 'generator' : 'allpages',
- 'gaplimit' : self.aplimit,
- 'gapfilterredir': self.filterredir}
+ self.queryParams = {'action': 'query',
+ 'generator': 'allpages',
+ 'gaplimit': self.aplimit,
+ 'gapfilterredir': self.filterredir}
else:
- self.queryParams = { 'action' : 'query' }
- if self.apfrom != u'': pywikibot.output(u'Argument "-from" is ignored with "-failed"')
+ self.queryParams = {'action': 'query'}
+ if self.apfrom != u'':
+ pywikibot.output(u'Argument "-from" is ignored with "-failed"')
propParam = 'info'
if self.links:
@@ -262,12 +272,12 @@
wlpage = self.whitelists[self.site.lang]
pywikibot.output(u'Loading whitelist from %s' % wlpage)
wlparams = {
- 'action' : 'query',
- 'prop' : 'links',
- 'titles' : wlpage,
- 'redirects' : '',
- 'indexpageids' : '',
- 'pllimit' : 'max',
+ 'action': 'query',
+ 'prop': 'links',
+ 'titles': wlpage,
+ 'redirects': '',
+ 'indexpageids': '',
+ 'pllimit': 'max',
}
data = query.GetData(wlparams)
@@ -275,21 +285,25 @@
pageid = data['query']['pageids'][0]
links = data['query']['pages'][pageid]['links']
- allWords = [nn for n in links for nn in self.FindBadWords(n['title'])]
+ allWords = [nn for n in links
+ for nn in self.FindBadWords(n['title'])]
self.knownWords = set(allWords)
-# kw = set()
-# for w in allWords:
-# if len(self.ProcessTitle(w)[1]) > 0:
-# kw.add(w)
-# self.knownWords = kw
+## kw = set()
+## for w in allWords:
+## if len(self.ProcessTitle(w)[1]) > 0:
+## kw.add(w)
+## self.knownWords = kw
else:
raise ValueError(u'The number of pageids is not 1')
- pywikibot.output(u'Loaded whitelist with %i items' % len(self.knownWords))
+ pywikibot.output(u'Loaded whitelist with %i items'
+ % len(self.knownWords))
if pywikibot.verbose and len(self.knownWords) > 0:
- pywikibot.output(u'Whitelist: %s' % u', '.join([self.MakeLink(i, False) for i in self.knownWords]))
+ pywikibot.output(u'Whitelist: %s'
+ % u', '.join([self.MakeLink(i, False)
+ for i in self.knownWords]))
else:
pywikibot.output(u'Whitelist is not known for language %s'
% self.site.lang)
@@ -303,18 +317,23 @@
yield data
# Clear any continuations first
- if 'clcontinue' in params: del params['clcontinue']
- if 'plcontinue' in params: del params['plcontinue']
+ if 'clcontinue' in params:
+ del params['clcontinue']
+ if 'plcontinue' in params:
+ del params['plcontinue']
if 'query-continue' not in data:
- if 'gapcontinue' in params: del params['gapcontinue']
+ if 'gapcontinue' in params:
+ del params['gapcontinue']
break
qc = data['query-continue']
# First continue properties only, once done, continue with allpages
if 'categories' in qc or 'links' in qc:
- if 'categories' in qc: params.update(qc['categories'])
- if 'links' in qc: params.update(qc['links'])
+ if 'categories' in qc:
+ params.update(qc['categories'])
+ if 'links' in qc:
+ params.update(qc['links'])
elif 'allpages' in qc:
params.update(qc['allpages'])
else:
@@ -336,14 +355,14 @@
self.currentTitle = None
batchSize = 10
for batchStart in xrange(0, len(self.titleList), batchSize):
- self.queryParams['titles'] = self.titleList[batchStart:batchStart+batchSize]
+ self.queryParams['titles'] = self.titleList[
+ batchStart:batchStart + batchSize]
for data in self.RunQuery(self.queryParams):
self.ProcessDataBlock(data)
-
print "*" * 29, "Done"
-
except:
- pywikibot.output(u'Exception at Title = %s, Next = %s' % (self.currentTitle, self.apfrom))
+ pywikibot.output(u'Exception at Title = %s, Next = %s'
+ % (self.currentTitle, self.apfrom))
try:
import traceback
pywikibot.output(traceback.format_exc())
@@ -357,20 +376,16 @@
firstItem = True
for pageID, page in data['query']['pages'].iteritems():
-
printed = False
title = page['title']
self.currentTitle = title
-
if 'missing' in page:
continue
-
if firstItem:
if self.lastLetter != title[0]:
pywikibot.ui.output('Processing %s\n' % title)
self.lastLetter = title[0]
firstItem = False
-
if self.titles:
err = self.ProcessTitle(title)
if err:
@@ -378,31 +393,46 @@
if self.replace:
if len(err[1]) == 1:
newTitle = err[1][0]
- # choice = pywikibot.inputChoice(u'Move %s to %s?' % (title, newTitle), ['Yes', 'No'], ['y', 'n'])
- editSummary = pywikibot.translate(self.site, self.msgRename)
+## choice = pywikibot.inputChoice(u'Move %s to %s?'
+## % (title, newTitle),
+## ['Yes', 'No'],
+## ['y', 'n'])
+ editSummary = pywikibot.translate(self.site,
+ self.msgRename)
dst = self.Page(newTitle)
if 'redirect' in page:
src = self.Page(title)
redir = src.getRedirectTarget()
- redirTitle = redir.title(asLink = True, textlink = True)
+ redirTitle = redir.title(asLink=True,
+ textlink=True)
if not dst.exists():
- src.move(newTitle, editSummary, movesubpages=True)
+ src.move(newTitle, editSummary,
+ movesubpages=True)
changed = True
replErrors = False
- for p in src.getReferences(follow_redirects = False):
+ for p in src.getReferences(
+ follow_redirects=False):
if p.namespace() == 2:
continue
- oldText = p.get(get_redirect = True)
- newText = self.ReplaceLink(oldText, title, newTitle)
- if not self.PutNewPage(p, newText, [self.MakeMoveSummary(title, newTitle)]):
+ oldText = p.get(get_redirect=True)
+ newText = self.ReplaceLink(oldText, title,
+ newTitle)
+ if not self.PutNewPage(
+ p, newText, [
+ self.MakeMoveSummary(title,
+ newTitle)]):
replErrors = True
if not replErrors:
- editSummary = pywikibot.translate(self.site, self.msgDeleteRedirect)
- newText = pywikibot.translate(self.site, self.textDeleteRedirect) % redirTitle
- src.put(newText, editSummary, minorEdit=False)
+ editSummary = pywikibot.translate(
+ self.site, self.msgDeleteRedirect)
+ newText = pywikibot.translate(
+ self.site,
+ self.textDeleteRedirect) % redirTitle
+ src.put(newText, editSummary,
+ minorEdit=False)
changed = True
elif not dst.exists():
@@ -414,11 +444,13 @@
src.title(withNamespace=False),
dst.title(withNamespace=False),
self.autonomous,
- editSummary + u' ' + self.MakeMoveSummary(title, newTitle),
+ editSummary + u' ' +
+ self.MakeMoveSummary(title, newTitle),
True)
bot.run()
else:
- src.move(newTitle, editSummary, movesubpages=True)
+ src.move(newTitle, editSummary,
+ movesubpages=True)
changed = True
if not changed:
@@ -463,18 +495,19 @@
pageObj = self.Page(title)
pageTxt = pageObj.get()
- msg.append(self.MakeMoveSummary(ltxt, newTitle))
+ msg.append(self.MakeMoveSummary(ltxt,
+ newTitle))
- pageTxt = self.ReplaceLink(pageTxt, ltxt, newTitle)
-
+ pageTxt = self.ReplaceLink(pageTxt, ltxt,
+ newTitle)
if not newTitle:
if not printed:
- self.WikiLog(u"* %s: link to %s" % (self.MakeLink(title, False), err[0]))
+ self.WikiLog(u"* %s: link to %s"
+ % (self.MakeLink(title, False),
+ err[0]))
printed = True
else:
self.WikiLog(u"** link to %s" % err[0])
-
-
if pageObj is not None:
if self.PutNewPage(pageObj, pageTxt, msg):
# done, no need to log anything
@@ -482,8 +515,6 @@
if foundSuggestions:
self.AppendLineToLog(self.failedTitles, title)
-
-
if self.stopAfter > 0:
self.stopAfter -= 1
if self.stopAfter == 0:
@@ -496,19 +527,18 @@
def FindBadWords(self, title):
for m in self.badWordPtrn.finditer(title):
- yield title[m.span()[0] : m.span()[1]]
+ yield title[m.span()[0]:m.span()[1]]
def ProcessTitle(self, title):
-
badWords = list(self.FindBadWords(title))
-
if len(badWords) > 0:
# Allow known words, allow any roman numerals with local suffixes
- badWords = set([i for i in badWords if i not in self.knownWords and self.romanNumSfxPtrn.match(i) is not None])
+ badWords = set([i for i in badWords
+ if i not in self.knownWords and
+ self.romanNumSfxPtrn.match(i) is not None])
if len(badWords) == 0 or self.Page(title).isImage():
- return None
-
+ return
count = 0
ambigBadWords = set()
ambigBadWordsCount = 0
@@ -516,7 +546,6 @@
mapLat = {}
for badWord in badWords:
-
# See if it would make sense to treat the whole word as either
# cyrilic or latin
mightBeLat = mightBeLcl = True
@@ -527,7 +556,8 @@
else:
if mightBeLcl and l not in self.latinSuspects:
mightBeLcl = False
- if l not in self.latLtr: raise ValueError(u'Assert failed')
+ if l not in self.latLtr:
+ raise ValueError(u'Assert failed')
# Some words are well known and frequently mixed-typed
if mightBeLcl and mightBeLat:
@@ -559,9 +589,11 @@
c2 = self.lclToLatDict[co]
else:
c2 = None
- kw = [w for w in kw if p < len(w) and (w[p] == c or (c2 is not None and w[p] == c2))]
+ kw = [w for w in kw if p < len(w) and
+ (w[p] == c or (c2 is not None and w[p] == c2))]
if len(kw) > 1:
- pywikibot.output(u"Word '%s' could be treated as more than one known words" % badWord)
+ pywikibot.output(u"Word '%s' could be treated as more than "
+ u"one known words" % badWord)
elif len(kw) == 1:
mapLcl[badWord] = kw[0]
count += 1
@@ -580,10 +612,9 @@
else:
# Replace all unambiguous bad words
- for k,v in mapLat.items() + mapLcl.items():
+ for k, v in mapLat.items() + mapLcl.items():
if k not in ambigBadWords:
- title = title.replace(k,v)
-
+ title = title.replace(k, v)
if len(ambigBadWords) == 0:
# There are no ambiguity, we can safelly convert
possibleAlternatives.append(title)
@@ -593,7 +624,7 @@
# combinations from the bad words list, and convert just the
# picked words to cyrilic, whereas making all other words as
# latin character.
- for itemCntToPick in xrange(0, len(ambigBadWords)+1):
+ for itemCntToPick in xrange(0, len(ambigBadWords) + 1):
title2 = title
for uc in xuniqueCombinations(list(ambigBadWords),
itemCntToPick):
@@ -610,14 +641,11 @@
[self.MakeLink(t) for t in possibleAlternatives])
else:
infoText += u", no suggestions"
-
return (infoText, possibleAlternatives)
def PickTarget(self, title, original, candidates):
-
if len(candidates) == 0:
- return None
-
+ return
if len(candidates) == 1:
return candidates[0]
@@ -633,16 +661,15 @@
pagesRedir[newTitle] = dst.getRedirectTarget().title()
else:
pagesExist.append(newTitle)
-
if len(pagesExist) == 1:
return pagesExist[0]
elif len(pagesExist) == 0 and len(pagesRedir) > 0:
if len(pagesRedir) == 1:
return pagesRedir.keys()[0]
t = None
- for k,v in pagesRedir.iteritems():
+ for k, v in pagesRedir.iteritems():
if not t:
- t = v # first item
+ t = v # first item
elif t != v:
break
else:
@@ -651,54 +678,66 @@
return pagesRedir.keys()[0]
if not self.autonomous:
- pywikibot.output(u'Could not auto-decide for page %s. Which link should be chosen?' % self.MakeLink(title, False))
+ pywikibot.output(u'Could not auto-decide for page %s. Which link '
+ u'should be chosen?' % self.MakeLink(title, False))
pywikibot.output(u'Original title: ', newline=False)
self.ColorCodeWord(original + "\n", True)
count = 1
for t in candidates:
- if t in pagesDontExist: msg = u'missing'
- elif t in pagesRedir: msg = u'Redirect to ' + pagesRedir[t]
- else: msg = u'page exists'
- self.ColorCodeWord(u' %d: %s (%s)\n'
- % (count, t, msg), True)
+ if t in pagesDontExist:
+ msg = u'missing'
+ elif t in pagesRedir:
+ msg = u'Redirect to ' + pagesRedir[t]
+ else:
+ msg = u'page exists'
+ self.ColorCodeWord(u' %d: %s (%s)\n' % (count, t, msg), True)
count += 1
-
answers = [str(i) for i in xrange(0, count)]
choice = int(pywikibot.inputChoice(
u'Which link to choose? (0 to skip)',
answers, [a[0] for a in answers]))
if choice > 0:
- return candidates[choice-1]
+ return candidates[choice - 1]
- return None
-
- def ColorCodeWord(self, word, toScreen = False):
-
- if not toScreen: res = u"<b>"
+ def ColorCodeWord(self, word, toScreen=False):
+ if not toScreen:
+ res = u"<b>"
lastIsCyr = word[0] in self.localLtr
if lastIsCyr:
- if toScreen: SetColor(FOREGROUND_GREEN)
- else: res += self.lclClrFnt
+ if toScreen:
+ SetColor(FOREGROUND_GREEN)
+ else:
+ res += self.lclClrFnt
else:
- if toScreen: SetColor(FOREGROUND_RED)
- else: res += self.latClrFnt
+ if toScreen:
+ SetColor(FOREGROUND_RED)
+ else:
+ res += self.latClrFnt
for l in word:
if l in self.localLtr:
if not lastIsCyr:
- if toScreen: SetColor(FOREGROUND_GREEN)
- else: res += self.suffixClr + self.lclClrFnt
+ if toScreen:
+ SetColor(FOREGROUND_GREEN)
+ else:
+ res += self.suffixClr + self.lclClrFnt
lastIsCyr = True
elif l in self.latLtr:
if lastIsCyr:
- if toScreen: SetColor(FOREGROUND_RED)
- else: res += self.suffixClr + self.latClrFnt
+ if toScreen:
+ SetColor(FOREGROUND_RED)
+ else:
+ res += self.suffixClr + self.latClrFnt
lastIsCyr = False
- if toScreen: pywikibot.output(l, newline=False)
- else: res += l
+ if toScreen:
+ pywikibot.output(l, newline=False)
+ else:
+ res += l
- if toScreen: SetColor(FOREGROUND_WHITE)
- else: return res + self.suffixClr + u"</b>"
+ if toScreen:
+ SetColor(FOREGROUND_WHITE)
+ else:
+ return res + self.suffixClr + u"</b>"
def AddNoSuggestionTitle(self, title):
if title in self.seenUnresolvedLinks:
@@ -706,10 +745,10 @@
self.seenUnresolvedLinks.add(title)
params = {
- 'action' : 'query',
- 'list' : 'backlinks',
- 'bltitle' : title,
- 'bllimit' : '50',
+ 'action': 'query',
+ 'list': 'backlinks',
+ 'bltitle': title,
+ 'bllimit': '50',
}
data = query.GetData(params)
@@ -725,15 +764,17 @@
else:
count = str(cl if cl > 0 else 'no backlinks')
- self.AppendLineToLog(self.nosuggestions, u'* %s (%s%s)' %
- (self.MakeLink(title), count, u', %d redirects' % redirs if redirs > 0 else u''))
+ self.AppendLineToLog(self.nosuggestions, u'* %s (%s%s)'
+ % (self.MakeLink(title), count, u', %d redirects'
+ % redirs if redirs > 0 else u''))
return False
def PutNewPage(self, pageObj, pageTxt, msg):
- title = pageObj.title(asLink = True, textlink = True)
+ title = pageObj.title(asLink=True, textlink=True)
coloredMsg = u', '.join([self.ColorCodeWord(m) for m in msg])
- if pageObj.get(get_redirect = True) == pageTxt:
- self.WikiLog(u"* Error: Text replacement failed in %s (%s)" % (self.MakeLink(title, False), coloredMsg))
+ if pageObj.get(get_redirect=True) == pageTxt:
+ self.WikiLog(u"* Error: Text replacement failed in %s (%s)"
+ % (self.MakeLink(title, False), coloredMsg))
else:
pywikibot.output(u'Case Replacements: %s' % u', '.join(msg))
try:
@@ -748,7 +789,8 @@
except KeyboardInterrupt:
raise
except:
- self.WikiLog(u"* Error: Could not save updated page %s (%s)" % (self.MakeLink(title, False), coloredMsg))
+ self.WikiLog(u"* Error: Could not save updated page %s (%s)"
+ % (self.MakeLink(title, False), coloredMsg))
return False
def MakeMoveSummary(self, fromTitle, toTitle):
@@ -774,8 +816,10 @@
def ReplaceLink(self, text, oldtxt, newtxt):
- frmParts = [str.strip(self.stripChars) for str in self.wordBreaker.split(oldtxt)]
- toParts = [str.strip(self.stripChars) for str in self.wordBreaker.split(newtxt)]
+ frmParts = [s.strip(self.stripChars)
+ for d in self.wordBreaker.split(oldtxt)]
+ toParts = [s.strip(self.stripChars)
+ for s in self.wordBreaker.split(newtxt)]
if len(frmParts) != len(toParts):
raise ValueError(u'Splitting parts do not match counts')
@@ -783,8 +827,10 @@
if len(frmParts[i]) != len(toParts[i]):
raise ValueError(u'Splitting parts do not match word length')
if len(frmParts[i]) > 0:
- text = text.replace(frmParts[i][0].lower() + frmParts[i][1:], toParts[i][0].lower() + toParts[i][1:])
- text = text.replace(frmParts[i][0].upper() + frmParts[i][1:], toParts[i][0].upper() + toParts[i][1:])
+ text = text.replace(frmParts[i][0].lower() + frmParts[i][1:],
+ toParts[i][0].lower() + toParts[i][1:])
+ text = text.replace(frmParts[i][0].upper() + frmParts[i][1:],
+ toParts[i][0].upper() + toParts[i][1:])
return text
--
To view, visit https://gerrit.wikimedia.org/r/95164
To unsubscribe, visit https://gerrit.wikimedia.org/r/settings
Gerrit-MessageType: merged
Gerrit-Change-Id: I892ebfcbcabcfcd9ac1e6c899cd75029fc4d2993
Gerrit-PatchSet: 1
Gerrit-Project: pywikibot/compat
Gerrit-Branch: master
Gerrit-Owner: Xqt <info(a)gno.de>
Gerrit-Reviewer: Xqt <info(a)gno.de>
Gerrit-Reviewer: jenkins-bot