jenkins-bot has submitted this change and it was merged.
Change subject: New stars templates for gl-wiki, synchronize with core
......................................................................
New stars templates for gl-wiki, synchronize with core
Change-Id: Icb37b0148fc3745435fc51f95d141bbdb210a62b
---
M add_text.py
M cosmetic_changes.py
M featured.py
3 files changed, 57 insertions(+), 42 deletions(-)
Approvals:
Ladsgroup: Looks good to me, approved
jenkins-bot: Verified
diff --git a/add_text.py b/add_text.py
index 4daa02a..ee4b3ab 100644
--- a/add_text.py
+++ b/add_text.py
@@ -66,7 +66,7 @@
#
# (C) Filnik, 2007-2010
-# (C) Pywikibot team, 2007-2013
+# (C) Pywikibot team, 2007-2014
#
# Distributed under the terms of the MIT license.
#
@@ -104,6 +104,7 @@
u'liên[ _]kết[ _]bài[ _]chất[ _]lượng[ _]tốt',
u'liên[ _]kết[ _]chọn[ _]lọc',
u'ligam[ _]adq',
+ u'ligazón[ _]a[bd]',
u'ligoelstara',
u'ligoleginda',
u'link[ _][afgu]a', u'link[ _]adq', u'link[ _]f[lm]',
u'link[ _]km',
diff --git a/cosmetic_changes.py b/cosmetic_changes.py
index 31a1fe1..2085ff0 100644
--- a/cosmetic_changes.py
+++ b/cosmetic_changes.py
@@ -53,7 +53,7 @@
"""
#
# (C) xqt, 2009-2013
-# (C) Pywikipedia bot team, 2006-2013
+# (C) Pywikibot team, 2006-2014
#
# Distributed under the terms of the MIT license.
#
@@ -175,7 +175,7 @@
text = self.fixArabicLetters(text)
try:
text = isbn.hyphenateIsbnNumbers(text)
- except isbn.InvalidIsbnException, error:
+ except isbn.InvalidIsbnException as error:
if pywikibot.verbose:
pywikibot.output(u"ISBN error: %s" % error)
pass
@@ -222,6 +222,7 @@
u'liên[ _]kết[ _]bài[ _]chất[ _]lượng[ _]tốt',
u'liên[ _]kết[ _]chọn[ _]lọc',
u'ligam[ _]adq',
+ u'ligazón[ _]a[bd]',
u'ligoelstara',
u'ligoleginda',
u'link[ _][afgu]a', u'link[ _]adq', u'link[ _]f[lm]',
u'link[ _]km',
@@ -346,9 +347,10 @@
if thisNs and namespaces:
text = pywikibot.replaceExcept(
text,
- r'\[\[\s*(' + '|'.join(namespaces) +
- ') *:(?P<nameAndLabel>.*?)\]\]', r'[[' + thisNs
+
- ':\g<nameAndLabel>]]', exceptions)
+ r'\[\[\s*(%s) *:(?P<nameAndLabel>.*?)\]\]'
+ % '|'.join(namespaces),
+ r'[[%s:\g<nameAndLabel>]]' % thisNs,
+ exceptions)
return text
def translateMagicWords(self, text):
@@ -367,7 +369,7 @@
continue
text = pywikibot.replaceExcept(
text,
- r'\[\[(?P<left>.+?:.+?\..+?\|) *(' +
'|'.join(aliases) + \
+ r'\[\[(?P<left>.+?:.+?\..+?\|) *(' +
'|'.join(aliases) +
') *(?P<right>(\|.*?)?\]\])',
r'[[\g<left>' + aliases[0] + '\g<right>',
exceptions)
return text
@@ -449,11 +451,11 @@
newLink = "[[%s]]" % label
# Check if we can create a link with trailing characters
# instead of a pipelink
- elif self.site.sitename() != 'wikipedia:fa' and \
- len(titleWithSection) <= len(label) and \
- label[:len(titleWithSection)] == titleWithSection and \
- re.sub(trailR, '',
- label[len(titleWithSection):]) == '':
+ elif (self.site.sitename() != 'wikipedia:fa' and
+ len(titleWithSection) <= len(label) and
+ label[:len(titleWithSection)] == titleWithSection and
+ re.sub(trailR, '',
+ label[len(titleWithSection):]) == ''):
newLink = "[[%s]]%s" % (label[:len(titleWithSection)],
label[len(titleWithSection):])
else:
@@ -535,20 +537,21 @@
def removeUselessSpaces(self, text):
multipleSpacesR = re.compile(' +')
spaceAtLineEndR = re.compile(' $')
-
- exceptions = ['comment', 'math', 'nowiki', 'pre',
'startspace', 'table', 'template']
+ exceptions = ['comment', 'math', 'nowiki', 'pre',
'startspace', 'table',
+ 'template']
text = pywikibot.replaceExcept(text, multipleSpacesR, ' ', exceptions)
text = pywikibot.replaceExcept(text, spaceAtLineEndR, '', exceptions)
-
return text
def removeNonBreakingSpaceBeforePercent(self, text):
- '''
+ """
Newer MediaWiki versions automatically place a non-breaking space in
front of a percent sign, so it is no longer required to place it
manually.
- '''
- text = pywikibot.replaceExcept(text, r'(\d) %', r'\1 %',
['timeline'])
+
+ """
+ text = pywikibot.replaceExcept(text, r'(\d) %', r'\1
%',
+ ['timeline'])
return text
def cleanUpSectionHeaders(self, text):
@@ -590,7 +593,8 @@
exceptions = ['comment', 'math', 'nowiki',
'pre']
if self.site.family.name in deprecatedTemplates and \
self.site.lang in deprecatedTemplates[self.site.family.name]:
- for template in deprecatedTemplates[self.site.family.name][self.site.lang]:
+ for template in deprecatedTemplates[
+ self.site.family.name][self.site.lang]:
old = template[0]
new = template[1]
if new is None:
@@ -617,9 +621,10 @@
## % (self.site.lang, self.site.family.name),
## r'[[\g<link>|\g<title>]]',
exceptions)
# external link in double brackets
- text = pywikibot.replaceExcept(text,
-
r'\[\[(?P<url>https?://[^\]]+?)\]\]',
- r'[\g<url>]', exceptions)
+ text = pywikibot.replaceExcept(
+ text,
+ r'\[\[(?P<url>https?://[^\]]+?)\]\]',
+ r'[\g<url>]', exceptions)
# external link starting with double bracket
text = pywikibot.replaceExcept(text,
r'\[\[(?P<url>https?://.+?)\]',
@@ -627,15 +632,17 @@
# external link and description separated by a dash, with
# whitespace in front of the dash, so that it is clear that
# the dash is not a legitimate part of the URL.
- text = pywikibot.replaceExcept(text,
- r'\[(?P<url>https?://[^\|\] \r\n]+?) +\|
*(?P<label>[^\|\]]+?)\]',
- r'[\g<url> \g<label>]',
exceptions)
+ text = pywikibot.replaceExcept(
+ text,
+ r'\[(?P<url>https?://[^\|\] \r\n]+?) +\|
*(?P<label>[^\|\]]+?)\]',
+ r'[\g<url> \g<label>]', exceptions)
# dash in external link, where the correct end of the URL can
# be detected from the file extension. It is very unlikely that
# this will cause mistakes.
- text = pywikibot.replaceExcept(text,
- r'\[(?P<url>https?://[^\|\]
]+?(\.pdf|\.html|\.htm|\.php|\.asp|\.aspx|\.jsp)) *\| *(?P<label>[^\|\]]+?)\]',
- r'[\g<url> \g<label>]',
exceptions)
+ text = pywikibot.replaceExcept(
+ text,
+ r'\[(?P<url>https?://[^\|\]
]+?(\.pdf|\.html|\.htm|\.php|\.asp|\.aspx|\.jsp)) *\| *(?P<label>[^\|\]]+?)\]',
+ r'[\g<url> \g<label>]', exceptions)
return text
def fixHtml(self, text):
@@ -679,9 +686,11 @@
# it should be name = " or name=" NOT name ="
text = re.sub(r'(?i)<ref +name(= *| *=)"', r'<ref
name="', text)
#remove empty <ref/>-tag
- text = pywikibot.replaceExcept(text, r'(?i)(<ref\s*/>|<ref
*>\s*</ref>)',
+ text = pywikibot.replaceExcept(text,
+ r'(?i)(<ref\s*/>|<ref
*>\s*</ref>)',
r'', exceptions)
- text = pywikibot.replaceExcept(text,
r'(?i)<ref\s+([^>]+?)\s*>\s*</ref>',
+ text = pywikibot.replaceExcept(text,
+
r'(?i)<ref\s+([^>]+?)\s*>\s*</ref>',
r'<ref \1/>', exceptions)
return text
@@ -691,17 +700,18 @@
# convert prettytable to wikitable class
if self.site.language in ('de', 'en'):
text = pywikibot.replaceExcept(text,
-
ur'(class="[^"]*)prettytable([^"]*")',
- ur'\1wikitable\2', exceptions)
+
r'(class="[^"]*)prettytable([^"]*")',
+ r'\1wikitable\2', exceptions)
return text
def fixTypo(self, text):
exceptions = ['nowiki', 'comment', 'math', 'pre',
'source',
'startspace', 'gallery', 'hyperlink',
'interwiki', 'link']
# change <number> ccm -> <number> cm³
- text = pywikibot.replaceExcept(text, ur'(\d)\s* ccm',
+ text = pywikibot.replaceExcept(text, r'(\d)\s* ccm',
ur'\1 cm³', exceptions)
- text = pywikibot.replaceExcept(text, ur'(\d)\s*ccm',
ur'\1 cm³',
+ text = pywikibot.replaceExcept(text,
+ r'(\d)\s*ccm',
ur'\1 cm³',
exceptions)
# Solve wrong Nº sign with °C or °F
# additional exception requested on fr-wiki for this stuff
@@ -711,7 +721,8 @@
ur'\1 °\2', exceptions)
text = pywikibot.replaceExcept(text, ur'(\d)\s*[º°]([CF])',
ur'\1 °\2', exceptions)
- text = pywikibot.replaceExcept(text, ur'º([CF])', ur'°\1',
exceptions)
+ text = pywikibot.replaceExcept(text, u'º([CF])', ur'°\1',
+ exceptions)
return text
def fixArabicLetters(self, text):
@@ -741,8 +752,10 @@
old = digits[digits.keys()[0]]
# do not change inside file links
namespaces = list(self.site.namespace(6, all=True))
- pattern = re.compile(u'\[\[(' + '|'.join(namespaces) +
'):.+?\.\w+? *(\|((\[\[.*?\]\])|.)*)?\]\]',
- re.UNICODE)
+ pattern = re.compile(
+ u'\[\[(' + '|'.join(namespaces) +
+ '):.+?\.\w+? *(\|((\[\[.*?\]\])|.)*)?\]\]',
+ re.UNICODE)
#not to let bot edits in latin content
exceptions.append(re.compile(u"[^%(fa)s] *?\"*? *?, *?[^%(fa)s]"
% {'fa': faChrs}))
@@ -755,7 +768,7 @@
text = pywikibot.replaceExcept(text, u'ه', u'ە',
exceptions)
text = pywikibot.replaceExcept(text, u'ه', u'ھ', exceptions)
text = pywikibot.replaceExcept(text, u'ك', u'ک', exceptions)
- text = pywikibot.replaceExcept(text, ur'[ىي]', u'ی', exceptions)
+ text = pywikibot.replaceExcept(text, u'[ىي]', u'ی', exceptions)
return text
# replace persian/arabic digits
## deactivated due to bug #3539407
@@ -821,7 +834,6 @@
text,
r'([\r\n]|^)\=\= *{{int:license-header}} *\=\=(?:[\r\n ]*)\=\=
*{{int:license-header}} *\=\=',
r'\1== {{int:license-header}} ==', exceptions, True)
-
return text
diff --git a/featured.py b/featured.py
index 2f9aceb..230e054 100644
--- a/featured.py
+++ b/featured.py
@@ -52,7 +52,7 @@
# (C) Maxim Razin, 2005
# (C) Leonardo Gregianin, 2005-2008
# (C) xqt, 2009-2013
-# (C) Pywikipedia bot team, 2005-2012
+# (C) Pywikibot team, 2005-2014
#
# Distributed under the terms of the MIT license.
#
@@ -92,7 +92,7 @@
'ast': ['Enllaz AD'],
'az': ['Link FM'],
'br': ['Liamm PuB', 'Lien AdQ'],
- 'ca': [u'Enllaç AD', 'Link FA', 'Destacat'],
+ 'ca': [u'Enllaç AD', 'Destacat'],
'cy': ['Cyswllt erthygl ddethol', 'Dolen ED'],
'eo': ['LigoElstara'],
'en': ['Link FA', 'FA link'],
@@ -101,6 +101,7 @@
'fr': ['Lien AdQ'],
'fur': ['Leam VdC'],
'ga': ['Nasc AR'],
+ 'gl': [u'Ligazón AD', 'Destacado'],
'hi': ['Link FA', 'Lien AdQ'],
'is': [u'Tengill ÚG'],
'it': ['Link V', 'Link AdQ'],
@@ -117,11 +118,12 @@
template_good = {
'_default': ['Link GA'],
'ar': [u'وصلة مقالة جيدة'],
- 'ca': [u'Enllaç AB', 'Lien BA', 'Abo', 'Link
GA'],
+ 'ca': [u'Enllaç AB', 'Lien BA', 'Abo'],
'da': ['Link GA', 'Link AA'],
'eo': ['LigoLeginda'],
'es': ['Bueno'],
'fr': ['Lien BA'],
+ 'gl': [u'Ligazón AB'],
'is': ['Tengill GG'],
'it': ['Link VdQ'],
'nn': ['Link AA'],
--
To view, visit
https://gerrit.wikimedia.org/r/111421
To unsubscribe, visit
https://gerrit.wikimedia.org/r/settings
Gerrit-MessageType: merged
Gerrit-Change-Id: Icb37b0148fc3745435fc51f95d141bbdb210a62b
Gerrit-PatchSet: 1
Gerrit-Project: pywikibot/compat
Gerrit-Branch: master
Gerrit-Owner: Xqt <info(a)gno.de>
Gerrit-Reviewer: Alex S.H. Lin <alexsh(a)mail2000.com.tw>
Gerrit-Reviewer: DrTrigon <dr.trigon(a)surfeu.ch>
Gerrit-Reviewer: Huji <huji.huji(a)gmail.com>
Gerrit-Reviewer: Ladsgroup <ladsgroup(a)gmail.com>
Gerrit-Reviewer: Meno25 <meno25mail(a)gmail.com>
Gerrit-Reviewer: Merlijn van Deen <valhallasw(a)arctus.nl>
Gerrit-Reviewer: Pyfisch <pyfisch(a)gmail.com>
Gerrit-Reviewer: Rotem Liss <rotemliss(a)gmail.com>
Gerrit-Reviewer: Siebrand <siebrand(a)wikimedia.org>
Gerrit-Reviewer: jenkins-bot <>