Happy Monday,
There are strange people who make such links (kindof urlencoded?):
[[Második világháború#Partrasz.C3.A1ll.C3.A1s Szic.C3.ADli.C3.A1ban
.28Huskey hadm.C5.B1velet.29|Huskey hadműveletben]]
So the section title must have been copied from the URL.
Do we have a ready tool to fix these?
--
Bináris
Hello all
>From one of my assignments as a bot operator I have some code which
does template parsing and general text parsing (e.g. Image/File tags).
It is not using regex and thus able to correctly parse nested
templates and other such nasty things. I have written those as library
classes and written tests for them which cover almost all of the code.
I would now really like to contribute that code back to the community.
Would you be interested in adding this code to the pywikibot
framework? If yes, can I send the code to someone for code review or
how do you usually operate?
Greetings
Hannes
PS: wiki userpage is http://en.wikipedia.org/wiki/User:Hannes_R%C3%B6st
Hi,
After a couple of experiments it seems to me that
*replace.py something otherthing -search:thing*
works in ALL BUT article namespace.
Could you please check where is the error?
--
Bináris
Hello there!
At first, I am really happy to see the pywikipedia framework is still
alive and very active.
As you may have noticed, whenever a change is send to Gerrit that
triggers Jenkins jobs that run the code style utilities pep8 and pyflakes.
It seems your code is not passing the style checks so whenever they fail
it is not going to prevent you from merging the code.
I am not sure how your community likes pep8/pyflakes. But I think it
would be nice to have the code repositories to pass those tests and
enforce authors to follow them. You can still have some pep8 checks
ignored such as the "line too long".
The questions are:
Is there any interest in making your repositories pep8 compliants?
If so, is there anything I can do to help? :-]
--
Antoine "hashar" Musso
-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1
Hello all!
Hello amir!
I have created and maintained some other additional SVN repositiories
on TS in the past [1].
[1] https://fisheye.toolserver.org/browse/drtrigon/externals
Since it is not possible to include SVN repos into GIT as submodule
(or else?) I fear I have to convert them (by svn2git?) to GIT repos in
order to re-integrate them again into pywikibot/compat. Would it be
possible to create those repos on gerrit in pywikibot too? E.g.:
pywikibot/haartraining
pywikibot/opencv
pywikibot/pycolorname
pywikibot/archive(s) (for .tar.gz files - other suggestions welcome)
in order to copy them
or pywikibot/other or pywikibot/thirdparty
as general repos to create several such additional codes (also for
future). Though later is not optimal since it can be included just as
1 submodule only.
Thanks a lot for your thought and help
Greetins
DrTrigon
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.13 (GNU/Linux)
Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/
iEYEARECAAYFAlH4420ACgkQAXWvBxzBrDC5KwCghIA4WXfFogvMW9DZA69bfP47
Yg8An2YpPRTR+ANrqzApHw2cbPRMHtmU
=BYOr
-----END PGP SIGNATURE-----
I have written the following script:
# -*- coding: utf-8 -*-
import pywikibot
site = pywikibot.Site("gl", "wiktionary")
page = pywikibot.Page(site, u"𐌰𐌽𐌳𐌰𐌿𐍂𐌰")
print page.get()
It fails with the following output:
[gallaecio@afonso fontes]$ python2 test.py
Traceback (most recent call last):
File "test.py", line 7, in <module>
print page.get()
File "/usr/lib/python2.7/site-packages/pywikibot/__init__.py", line 249, in wrapper
return method(*__args, **__kw)
File "/usr/lib/python2.7/site-packages/pywikibot/__init__.py", line 249, in wrapper
return method(*__args, **__kw)
File "/usr/lib/python2.7/site-packages/pywikibot/page.py", line 278, in get
self._getInternals(sysop)
File "/usr/lib/python2.7/site-packages/pywikibot/page.py", line 302, in _getInternals
self.site.loadrevisions(self, getText=True, sysop=sysop)
File "/usr/lib/python2.7/site-packages/pywikibot/page.py", line 96, in site
return self._link.site
File "/usr/lib/python2.7/site-packages/pywikibot/page.py", line 3080, in site
self.parse()
File "/usr/lib/python2.7/site-packages/pywikibot/page.py", line 3037, in parse
u"contains illegal char(s) '%s'" % m.group(0))
pywikibot.exceptions.InvalidTitle: contains illegal char(s) '𐌰'
Hi,
I saw the existence of this script, and I would like have some informations about the parameters to use it. Does someone know it?
I saw for instance the parameter allrevisions, but is there a parameter to choose the page where we put the reults the page than
match the content)? Parameter for the regexp?
Thanks by advance for your answer
Hi,
\b in a regex treats letter "é" (which is a correct Hungarian letter) as a
word boundary.
Can I prevent this behaviour with some kind of settings?
--
Bináris
I would like to know two things:
• I want to implement a script to rename pages. Do you agree that implementing this feature within replace.py as an option would the the right approach, or should I create a separated script?
• What is the current procedure to provide patches (for the rewrite branch)? Reading http://www.mediawiki.org/wiki/Manual:Pywikipediabot/Development didn’t solve this doubt of mine. I see there is a Patches section in SourceForge, is that the right place?
Hello, I am again asking if it is possible for me to gain SVN access (I know it probably wouldn't last too long, considering that SVN will be read-only in the (near?) future, but that time is still unknown to us). I frequently use the framework, and I'd like to be able to directly make commits, rather than generating patches to put on the bug tracker, when possible. This would include things such as PEP and typo fixes, as well as improving wikipedia.DataPage. Thanks.
Hazard-SJ