Thats the only problem. We can't AWB every single link in the
database, then AWB it back just for the purpose of a study, which is
why we would have to somehow do it at software parser level for a day
or so. And we could easily host the file on toolserver and handle
redirects / logging from there. I have spare low traffic mysql db
server we could use. Then again, it would be simpler to build the
functionality into the Wikipedia toolbar, and thats a violation of
privacy, so it would have to be optional, and there'd be a clearly
rigged demographic left of those who allowed the logging to take
place...
On 9/9/06, Gregory Maxwell <gmaxwell(a)gmail.com> wrote:
On 9/8/06, Akash Mehta <draicone(a)gmail.com>
wrote:
I'm sure it wouldn't be very expensive.
All that needs to be done is
that the external (non-interwiki) link code passes URLs through a
script on a server that counts the number of clicks to that particular
link and then passes the user on to the appropriate site, which would
probably be specified through GET. I'll code it if I have time today,
its very simple in php, and using a database means that statistics can
be generated VERY easily.
Turning every external link follow into a database write on Wikipedia
is simply not going to be acceptable.
We don't currently log HTTP GETs for a reason...
_______________________________________________
WikiEN-l mailing list
WikiEN-l(a)Wikipedia.org
To unsubscribe from this mailing list, visit:
http://mail.wikipedia.org/mailman/listinfo/wikien-l