jenkins-bot has submitted this change and it was merged.
Change subject: pep8-ified scripts/interwiki.py
......................................................................
pep8-ified scripts/interwiki.py
Change-Id: Ic78ca03626ce51fd043afd7d8b8a03c973a544d3
---
M scripts/interwiki.py
1 file changed, 81 insertions(+), 63 deletions(-)
Approvals:
Ladsgroup: Looks good to me, approved
jenkins-bot: Verified
diff --git a/scripts/interwiki.py b/scripts/interwiki.py
index 713d2d7..ff94157 100755
--- a/scripts/interwiki.py
+++ b/scripts/interwiki.py
@@ -667,7 +667,8 @@
index = 1
while True:
path = config.datafilepath('cache', 'pagestore' +
str(index))
- if not os.path.exists(path): break
+ if not os.path.exists(path):
+ break
index += 1
StoredPage.SPpath = path
StoredPage.SPstore = shelve.open(path)
@@ -1031,8 +1032,8 @@
if linkedPage in self.foundIn:
# We have seen this page before, don't ask again.
return False
- elif self.originPage and \
- self.originPage.namespace() != linkedPage.namespace():
+ elif (self.originPage and
+ self.originPage.namespace() != linkedPage.namespace()):
# Allow for a mapping between different namespaces
crossFrom = self.originPage.site.family.crossnamespace.get(
self.originPage.namespace(), {})
@@ -1099,8 +1100,9 @@
pywikibot.output(u"NOTE: Ignoring %s for %s in wiktionary
mode"
% (page, self.originPage))
return True
- elif page.title() != self.originPage.title() and \
- self.originPage.site.nocapitalize and page.site.nocapitalize:
+ elif (page.title() != self.originPage.title() and
+ self.originPage.site.nocapitalize and
+ page.site.nocapitalize):
pywikibot.output(
u"NOTE: Ignoring %s for %s in wiktionary mode because both
"
u"languages are uncapitalized."
@@ -1334,8 +1336,8 @@
if not globalvar.quiet:
pywikibot.output(
u"NOTE: not following static %sredirects." %
redir)
- elif page.site.family == redirectTargetPage.site.family \
- and not self.skipPage(page, redirectTargetPage, counter):
+ elif (page.site.family == redirectTargetPage.site.family and
+ not self.skipPage(page, redirectTargetPage, counter)):
if self.addIfNew(redirectTargetPage, counter, page):
if config.interwiki_shownew:
pywikibot.output(u"%s: %s gives new %sredirect %s"
@@ -1414,21 +1416,21 @@
self.makeForcedStop(counter)
try:
f = codecs.open(
-
pywikibot.config.datafilepath('autonomous_problems.dat'),
- 'a', 'utf-8')
+
pywikibot.config.datafilepath('autonomous_problems.dat'),
+ 'a', 'utf-8')
f.write(u"* %s {Found more than one link for %s}"
% (self.originPage, page.site))
if config.interwiki_graph and config.interwiki_graph_url:
- filename = interwiki_graph.getFilename(self.originPage, extension
= config.interwiki_graph_formats[0])
+ filename = interwiki_graph.getFilename(self.originPage,
extension=config.interwiki_graph_formats[0])
f.write(u" [%s%s graph]" % (config.interwiki_graph_url,
filename))
f.write("\n")
f.close()
# FIXME: What errors are we catching here?
# except: should be avoided!!
except:
- #raise
- pywikibot.output(u'File autonomous_problems.dat open or corrupted!
Try again with -restore.')
- sys.exit()
+ #raise
+ pywikibot.output(u'File autonomous_problems.dat open or
corrupted! Try again with -restore.')
+ sys.exit()
iw = ()
elif page.isEmpty() and not page.isCategory():
globalvar.remove.append(unicode(page))
@@ -1451,7 +1453,7 @@
if self.addIfNew(linkedPage, counter, page):
# It is new. Also verify whether it is the second on the
# same site
- lpsite=linkedPage.site
+ lpsite = linkedPage.site
for prevPage in self.foundIn:
if prevPage != linkedPage and prevPage.site == lpsite:
# Still, this could be "no problem" as
either may be a
@@ -1478,7 +1480,7 @@
"""Return True if all the work for this subject has
completed."""
return len(self.todo) == 0
- def problem(self, txt, createneed = True):
+ def problem(self, txt, createneed=True):
"""Report a problem with the resolution of this
subject."""
pywikibot.output(u"ERROR: %s" % txt)
self.confirm = True
@@ -1488,10 +1490,9 @@
def whereReport(self, page, indent=4):
for page2 in sorted(self.foundIn[page]):
if page2 is None:
- pywikibot.output(u" "*indent + "Given as a hint.")
+ pywikibot.output(u" " * indent + "Given as a hint.")
else:
- pywikibot.output(u" "*indent + unicode(page2))
-
+ pywikibot.output(u" " * indent + unicode(page2))
def assemble(self):
# No errors have been seen so far, except....
@@ -1505,7 +1506,7 @@
site = page.site
if site.family.interwiki_forward:
#TODO: allow these cases to be propagated!
- continue # inhibit the forwarding families pages to be updated.
+ continue # inhibit the forwarding families pages to be updated.
if site == self.originPage.site:
if page != self.originPage:
self.problem(u"Found link to %s" % page)
@@ -1549,7 +1550,7 @@
i += 1
pywikibot.output(u" (%d) Found link to %s in:"
% (i, page2))
- self.whereReport(page2, indent = 8)
+ self.whereReport(page2, indent=8)
while True:
#TODO: allow answer to repeat previous or go back after a mistake
answer = pywikibot.input(u"Which variant should be used?
(<number>, [n]one, [g]ive up) ").lower()
@@ -1577,7 +1578,7 @@
pywikibot.output(u"=" * 30)
page2 = pages[0]
pywikibot.output(u"Found link to %s in:" % page2)
- self.whereReport(page2, indent = 4)
+ self.whereReport(page2, indent=4)
while True:
if acceptall:
answer = 'a'
@@ -1609,12 +1610,13 @@
time1 = str(time1)
if type(time2) is long:
time2 = str(time2)
- t1 = (((int(time1[0:4]) * 12 + int(time1[4:6])) * 30 +
- int(time1[6:8])) * 24 + int(time1[8:10])) * 60 + \
- int(time1[10:12])
- t2 = (((int(time2[0:4]) * 12 + int(time2[4:6])) * 30 +
+
+ t1 = ((((int(time1[0:4]) * 12 + int(time1[4:6])) * 30 +
+ int(time1[6:8])) * 24 + int(time1[8:10])) * 60 +
+ int(time1[10:12]))
+ t2 = ((((int(time2[0:4]) * 12 + int(time2[4:6])) * 30 +
int(time2[6:8])) * 24 + int(time2[8:10])) * 60 + \
- int(time2[10:12])
+ int(time2[10:12]))
return abs(t2 - t1)
if not self.isDone():
@@ -1630,7 +1632,7 @@
return
if not self.untranslated and globalvar.untranslatedonly:
return
- if self.forcedStop: # autonomous with problem
+ if self.forcedStop: # autonomous with problem
pywikibot.output(u"======Aborted processing %s======"
% self.originPage)
return
@@ -1683,8 +1685,9 @@
notUpdatedSites.append(site)
except GiveUpOnPage:
break
- elif not globalvar.strictlimittwo and site in new \
- and site != lclSite:
+ elif (not globalvar.strictlimittwo and
+ site in new and
+ site != lclSite):
old = {}
try:
for link in new[site].iterlanglinks():
@@ -1695,12 +1698,12 @@
% new[site])
continue
mods, mcomment, adding, removing, modifying \
- = compareLanguages(old, new, insite=lclSite)
- if (len(removing) > 0 and not globalvar.autonomous) or \
- (len(modifying) > 0 and self.problemfound) or \
- len(old) == 0 or \
- (globalvar.needlimit and
- len(adding) + len(modifying) >= globalvar.needlimit + 1):
+ = compareLanguages(old, new, insite=lclSite)
+ if ((len(removing) > 0 and not globalvar.autonomous) or
+ (len(modifying) > 0 and self.problemfound) or
+ (len(old) == 0) or
+ (globalvar.needlimit and
+ len(adding) + len(modifying) >= globalvar.needlimit +
1)):
try:
if self.replaceLinks(new[site], new):
updatedSites.append(site)
@@ -1723,10 +1726,12 @@
# allow edits for the same conditions as -whenneeded
# or the last edit wasn't a bot
# or the last edit was 1 month ago
- if smallWikiAllowed and globalvar.autonomous and \
- (page.site.sitename() == 'wikipedia:is' or
- page.site.sitename() == 'wikipedia:zh' and
- page.namespace() == 10):
+ if (smallWikiAllowed and
+ globalvar.autonomous and
+ (page.site.sitename() == 'wikipedia:is' or
+ page.site.sitename() == 'wikipedia:zh' and
+ page.namespace() == 10
+ )):
old = {}
try:
for mypage in new[page.site].interwiki():
@@ -1736,18 +1741,23 @@
% new[site])
continue
mods, mcomment, adding, removing, modifying \
- = compareLanguages(old, new, insite=site)
+ = compareLanguages(old, new, insite=site)
#cannot create userlib.User with IP
- smallWikiAllowed = page.isIpEdit() or \
- len(removing) > 0 or len(old) == 0 or \
- len(adding) + len(modifying) > 2 or \
- len(removing) + len(modifying) == 0 and \
- adding == [page.site]
+ smallWikiAllowed = (
+ page.isIpEdit() or
+ len(removing) > 0 or
+ len(old) == 0 or
+ len(adding) + len(modifying) > 2 or
+ (
+ len(removing) + len(modifying) == 0 and
+ adding == [page.site]
+ )
+ )
if not smallWikiAllowed:
import userlib
user = userlib.User(page.site, page.userName())
if not 'bot' in user.groups() \
- and not 'bot' in page.userName().lower(): #erstmal
auch keine namen mit bot
+ and not 'bot' in page.userName().lower(): # erstmal
auch keine namen mit bot
smallWikiAllowed = True
else:
diff = minutesDiff(page.editTime(),
@@ -1860,7 +1870,8 @@
pltmp = new[page.site]
if pltmp != page:
s = u"None"
- if pltmp is not None: s = pltmp
+ if pltmp is not None:
+ s = pltmp
pywikibot.output(
u"BUG>>> %s is not in the list of new links! Found
%s."
% (page, s))
@@ -1894,10 +1905,14 @@
continue
rmPage = old[rmsite]
#put it to new means don't delete it
- if not globalvar.cleanup or \
- unicode(rmPage) not in globalvar.remove or \
- rmPage.site.sitename() == 'wikipedia:hi' and \
- page.site.sitename() != 'wikipedia:de': #work-arround for bug
#3081100 (do not remove hi-pages)
+ if (
+ not globalvar.cleanup or
+ unicode(rmPage) not in globalvar.remove or
+ (
+ rmPage.site.sitename() == 'wikipedia:hi' and
+ page.site.sitename() != 'wikipedia:de' # work-arround
for bug #3081100 (do not remove hi-pages)
+ )
+ ):
new[rmsite] = rmPage
pywikibot.output(
u"WARNING: %s is either deleted or has a mismatching
disambiguation state."
@@ -1978,7 +1993,7 @@
if answer == 'y':
if not globalvar.quiet:
pywikibot.output(u"NOTE: Updating live wiki...")
- timeout=60
+ timeout = 60
while True:
try:
if globalvar.async:
@@ -2002,7 +2017,7 @@
pywikibot.output(u'ERROR putting page: %s' % (error.args,))
raise SaveError(u'PageNotSaved')
except (socket.error, IOError), error:
- if timeout>3600:
+ if timeout > 3600:
raise
pywikibot.output(u'ERROR putting page: %s' % (error.args,))
pywikibot.output(u'Sleeping %i seconds before trying again.'
@@ -2120,9 +2135,10 @@
def dump(self, append=True):
site = pywikibot.getSite()
dumpfn = pywikibot.config.datafilepath(
- 'data',
- 'interwiki-dumps',
- '%s-%s.pickle' % (site.family.name, site.lang))
+ 'data',
+ 'interwiki-dumps',
+ '%s-%s.pickle' % (site.family.name, site.lang)
+ )
if append:
mode = 'appended'
else:
@@ -2385,7 +2401,8 @@
return mods, mcomment, adding, removing, modifying
-def botMayEdit (page):
+
+def botMayEdit(page):
tmpl = []
try:
tmpl, loc = moved_links[page.site.lang]
@@ -2399,7 +2416,7 @@
pass
tmpl += ignoreTemplates['_default']
if tmpl != []:
- templates = page.templatesWithParams();
+ templates = page.templatesWithParams()
for template in templates:
if template[0].title(withNamespace=False).lower() in tmpl:
return False
@@ -2419,6 +2436,7 @@
hintStrings = ['%s:%s' % (hintedPage.site.language(),
hintedPage.title()) for hintedPage in pagelist]
bot.add(page, hints=hintStrings)
+
def main():
singlePageTitle = []
@@ -2528,10 +2546,10 @@
elif optRestore or optContinue or globalvar.restoreAll:
dumpFileName = pywikibot.config.datafilepath(
- 'data',
- 'interwiki-dumps',
- u'%s-%s.pickle'
- % (site.family.name, site.lang))
+ 'data',
+ 'interwiki-dumps',
+ u'%s-%s.pickle' % (site.family.name, site.lang)
+ )
try:
f = open(dumpFileName, 'r')
dumpedTitles = pickle.load(f)
@@ -2550,7 +2568,7 @@
pywikibot.output(u"Dump file is empty?! Starting at the
beginning.")
nextPage = "!"
namespace = 0
- hintlessPageGen = pagegenerators.CombinedPageGenerator([hintlessPageGen,
pagegenerators.AllpagesPageGenerator(nextPage, namespace, includeredirects = False)])
+ hintlessPageGen = pagegenerators.CombinedPageGenerator([hintlessPageGen,
pagegenerators.AllpagesPageGenerator(nextPage, namespace, includeredirects=False)])
bot = InterwikiBot()
--
To view, visit
https://gerrit.wikimedia.org/r/80851
To unsubscribe, visit
https://gerrit.wikimedia.org/r/settings
Gerrit-MessageType: merged
Gerrit-Change-Id: Ic78ca03626ce51fd043afd7d8b8a03c973a544d3
Gerrit-PatchSet: 1
Gerrit-Project: pywikibot/core
Gerrit-Branch: master
Gerrit-Owner: Merlijn van Deen <valhallasw(a)arctus.nl>
Gerrit-Reviewer: Ladsgroup <ladsgroup(a)gmail.com>
Gerrit-Reviewer: Legoktm <legoktm.wikipedia(a)gmail.com>
Gerrit-Reviewer: jenkins-bot