As some of you might remember, we have been working on author
reputation and text trust systems for wikis; some of you may have seen
our demo at WikiMania 2007, or the on-line demo
http://wiki-trust.cse.ucsc.edu/
Since then, we have been busy at work to build a system that can be
deployed on any wiki, and display the text trust information.
And we finally made it!
We are pleased to announce the release of WikiTrust version 2!
With it, you can compute author reputation and text trust of your
wikis in real-time, as edits to the wiki are made, and you can display
text trust via a new "trust" tab.
The tool can be installed as a MediaWiki extension, and is released
open-source, under the BSD license; the project page is
http://trust.cse.ucsc.edu/WikiTrust
WikiTrust can be deployed both on new, and on existing, wikis.
WikiTrust stores author reputation and text trust in additional
database tables. If deployed on an existing wiki, WikiTrust first
computes the reputation and trust information for the current wiki
content, and then processes new edits as they are made. The
computation is scalable, parallel, and fault-tolerant, in the sense
that WikiTrust adaptively fills in missing trust or reputation
information.
On my MacBook, running under Ubuntu in vmware, WikiTrust can analize
some 10-20 revisions / second of a wiki; so with a little patience,
unless your wiki is truly huge, you can just deploy it and wait a
bit.
Go to http://trust.cse.ucsc.edu/WikiTrust for more information and for
the code!
Feedback, comments, etc are much appreciated!
Luca de Alfaro
(with Ian Pye and Bo Adler)
There are several attempts to make bots that detect copyright
violations. The problem is that there are a lot of such "infringements"
that are legal, quotations for example, and then the writers gets pissed
because they have used the material in a completely legal way.
I have made a Javascript-based solution that seems to solve the problem
by placing a user in the loop. The only thing the script does is to mine
the web for possible similar texts.
Basically the script takes the additional text, extract the plain text,
excludes some of the text, breaks it into sentences, uses the sentences
to build a query, rematches the result to the sentences, accumulates
those and gives some warnings if a match limit is reached.
For the moment I try to extend the system to older edits, and also to
make it a bit more resistant to small changes in the text. It is already
fairly resistive to small reorganizations of the text.
John
In a message dated 8/27/2008 12:02:04 P.M. Pacific Daylight Time,
john.erling.blad(a)jeb.no writes:
Basically the script takes the additional text, extract the plain text,
excludes some of the text, breaks it into sentences, uses the sentences
to build a query, rematches the result to the sentences, accumulates
those and gives some warnings if a match limit is reached.>>
--------------
Wouldn't that flag situations where the author of the work, existing at some
site, is giving the copyright to the Wiki as well? That is they are
giving-up their copyright.
**************It's only a deal if it's where you want to go. Find your travel
deal here.
(http://information.travel.aol.com/deals?ncid=aoltrv00050000000047)
In a message dated 8/27/2008 11:40:01 A.M. Pacific Daylight Time,
luca(a)dealfaro.org writes:
1) are there some statistics, in mediawiki, that allow a process to check
how many times a page has been accessed? I can of course add a new hook and a
db table, but it may not work due to caching, and as this would run for every
hit, I don't want to increase load. >>
-----------
There must be, because my wiki at _http://www.countyhistorian.com_
(http://www.countyhistorian.com)
does have a page counter for each page. The entire list is under
Special/Popular or something like that.
But Wikipedia seems to have that ability turned off.
Will Johnson
**************It's only a deal if it's where you want to go. Find your travel
deal here.
(http://information.travel.aol.com/deals?ncid=aoltrv00050000000047)