As I mentioned in wikitech-l, I've started putting a captcha into production
testing on some of our sites. (At the moment: Meta, Commons, most of the smaller
Wikipedias, and a couple others.)
This is meant to preempt automated spam and, potentially, other automated
vandalism attacks. At the moment only edits which add new URLs to a page will be
stopped by the captcha; potentially other actions may be as well from time to time.
Obviously a captcha form will be a bit annoying to a bot. ;) Any registered bot
account is excepted from the edit captcha, as are sysops and accounts that pass
the 'semi-protection' threshold.
If you're operating a bot on the Wikimedia cluster which isn't logged in, or
uses a brand-new account which hasn't been registered as a bot on that wiki, be
aware that you might get hit by the captcha if your edits add URL links.
-- brion vibber (brion @ pobox.com)
After an absence of somewhat under 3 months from Wikipedia and bot
editing, I have decided to return. I used to be very disappointed at
the bot, but looking back I decided the problem was with their
international character, having to deal with all the languages at the
same time. I therefore have now decided to resurrect Robbot, but only
on the Dutch (and Frisian) Wikipedia. I still don't feel like doing
the heavy interwiki linking I did before.
--
Andre Engels, andreengels(a)gmail.com
ICQ: 6260644 -- Skype: a_engels
Every Monday i public a new weekly botting from from my Interwiki link
checker tool.
After publishing i bot the list with my FlaBot-Bot. At the end of the week
i have an
tool to find out if an articel is now present with both links in both
languages in the wiki.
If this is not, perhaps because my Bot cant do it in autonoumous-mode i will
post
a list with all the still missing entries of my Database.
Here the batch-list for botting :
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:af
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:es
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:fi
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:tr
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:ca
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:da
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:de
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:nds
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:en
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:nl
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:no
python interwiki.py -warnfile:warning_bot_rebot_need.log -lang:sv
The Data you can get here :
http://www.flacus.de/wikipedia/Interwiki-Link-Checker/bot-reb.php
At the moment we have week 49. My bot is botting week 48.
The List above has entries until week 47,
See you next week ;-)
--
[[:de:Benutzer:Flacus]][[:de:Benutzer:FlaBot]]
http://www.flacus.de/wikipedia/Interwiki-Link-Checker/
Hi,
the main directory keeps filling up with scripts. To keep some order, I
would suggest moving some older scripts to an archive directory. That's
why I want to know which of the following scripts are still in use. I
don't want to offend anyone; if you think one of the scripts in the list
is still in use or under development, please simply say so.
are-identical.py
http://tools.wikimedia.de/~flacus/IWLC/start.php works much better
brackethttp.py
I don't think anyone still uses it
check_extern.py
Replaced by weblinkchecker.py
copy_table.py
Too much work to maintain it
editarticle.py
No longer maintained, but we should re-use parts of it for other
scripts.
extract_names.py
Doesn't write the file format expected by most scripts.
find.py
Never worked; we might also consider deleting it.
getimages.py
imagetransfer.py can do everything this can do
pagefromfile.py
should either be updated or moved away
saveHTML.py
No longer maintained and maybe also no longer used
sqldump.py
All scripts have been changed and now only support xmldump.py
translator.py
Part of copy_table.py
us-states.py
Unless someone is still using it
vertexgen.py
Needs commenting, also in interwiki.py. Now it's unclear what's its
purpose.
WdT.py and WdTXMLParser.py
No longer used/maintained.
windows_chars.py
I don't think there are still any more ISO 8859-1 wikis left
anywhere, are they?
Daniel
The getReferences() function needs to be re-written due to the new change
in the What links here page since the addition of '(inclusion)' marking things
as templates. The reason is that the current regex will count that as a
redirect.
I am in the current process of re-writing this function, but in case anyone
wants to beat me to it, I suggest the following all encompassing regular
expression to use:
re.compile('<li><a href=".*?" title=".*?">(.*?)</a> *\(*(inclusion|redirect page)*\)*.*?</li>')
group(1) will give you the title, and group(2) of the search will be either:
'', 'inclusion', 'redirect page'
--
Jason Y. Lee
AKA AllyUnion
[[Wikipedia:Sandbox]] does not function in other projects beyond the
wikipedia.
** 1 ** with [[Wikipedia:Sandbox]]
======Post-processing [[pt:Alexis de Tocqueville]]======
Updating links on page [[he:?????? ??-??????]].
Changes to be made: Adding: it
+ [[it:Alexis de Tocqueville]]
NOTE: Updating live wiki...
Getting a page to check if we're logged in on wikiquote:he
Getting page to get a token.
Getting page [[he:Wikipedia:Sandbox]]
Sleeping for 8.5 seconds
Retrieving MediaWiki messages for wikiquote:he
Parsing MediaWiki messages
WARNING: No text area found on
he.wikiquote.org/w/index.php?title=Wikipedia%3ASandbox&action=edit.
Maybe the server is down. Retrying in 1 minutes...
Traceback (most recent call last):
File "interwiki.py", line 1330, in ?
bot.run()
File "interwiki.py", line 1114, in run
self.queryStep()
File "interwiki.py", line 1093, in queryStep
subj.finish(self)
File "interwiki.py", line 762, in finish
if self.replaceLinks(page, new, sa):
File "interwiki.py", line 858, in replaceLinks
status, reason, data = pl.put(newtext, comment =
wikipedia.translate(pl.site
().lang, msg)[0] + mods)
File "C:\Python24\wikipedia.py", line 677, in put
return self.putPage(newtext, comment, watchArticle, minorEdit,
newPage, self
.site().getToken(sysop = sysop), sysop = sysop)
File "C:\Python24\wikipedia.py", line 2559, in getToken
Page(self, "Wikipedia:Sandbox").get(force = True, sysop = sysop)
File "C:\Python24\wikipedia.py", line 351, in get
self._contents, self._isWatched, self.editRestriction =
self.getEditPage(get
_redirect = get_redirect, throttle = throttle, sysop = sysop)
File "C:\Python24\wikipedia.py", line 448, in getEditPage
i2 = re.search('</textarea>', text).start()
AttributeError: 'NoneType' object has no attribute 'start'
** 2 ** with Non-existing page
======Post-processing [[pt:Alexis de Tocqueville]]======
Updating links on page [[he:?????? ??-??????]].
Changes to be made: Adding: it
+ [[it:Alexis de Tocqueville]]
NOTE: Updating live wiki...
Getting a page to check if we're logged in on wikiquote:he
Getting page to get a token.
Getting page [[he:Non-existing page]]
Changing page [[he:?????? ??-??????]]
Updating links on page [[en:Alexis de Tocqueville]].
Changes to be made: Adding: it
+ [[it:Alexis de Tocqueville]]
NOTE: Performing a recursive query first to save time....
NOTE: Nothing left to do 2
NOTE: Updating live wiki...
Getting a page to check if we're logged in on wikiquote:en
...
Leonardo Gregianin