Hello all,
between 18:00 and 20:00 (CEST) I will do a little maintance on hemlock
(creating a partion for logging and so). I guess, that it do not take
the full 2h and will write an eMail when I'm finish.
Sincerly,
DaB.
--
PGP: 2D3EE2D42B255885
Hello again,
if we have one or more artists at this list, the toolserver could need
an logo for the favicon and the bugzilla.
Sincerly,
DaB.
--
PGP: 2D3EE2D42B255885
Hi.
fyi if you get some odd errors:
-:[~]$> df -h
Filesystem Size Used Avail Use% Mounted on
/dev/sdb1 4.0G 3.8G 0 100% /
Not sure whether the broken bugzilla has to do something with it.
-- Leon.
Note to admins: 'rcbot.pl' script it's _not_ out of control - this is
antivandal bot which listens to recent changes IRC channel. It should
never exit (but it does nothing between two edits, of course). Please
do not kill this process.
--
Edward Chernenko <edwardspec(a)gmail.com>
On Thu, 21 Sep 2006 09:54:31 +0200, <toolserver-l-request(a)Wikipedia.org>
wrote:
> Message: 13
> Date: Thu, 21 Sep 2006 09:54:29 +0200
> From: Stefan K?hn <kuehn-s(a)gmx.net>
> Subject: [Toolserver-l] Five questions
> To: Mailingliste Toolserver <toolserver-l(a)Wikipedia.org>
> Message-ID: <45124535.8050706(a)gmx.net>
> Content-Type: text/plain; charset=ISO-8859-15; format=flowed
> Can I use Perl to get some data from the database? (I work since three
> weeks with Perl and CGI, so I need a small example.)
Yes. I have no experience with it (I use python, which -imo- is easier
than perl), but google showed
http://www.codeproject.com/perl/perldbi.asp
which could be useful. Perl/SQL is - for example- used at slashdot.org
> Question 2:
> Can I use a SQL-Question with perl? For example: I want the category's
> of page "xy" in DE.
See the previous link. Make a SQL query for it; something like
"select * from dewiki_p.templatelinks where tl_namespace=<template
namespace> and tl_title=<template title>".
> Question 3.
> When I scan all articles from the Dump (XML-File) for coordinates I need
> with Perl in EN 45 minutes and in DE 15 minutes. If I use in the future
> hopefully the MySQL-database I think this process will be to long for
> the database. So that all other services have a problem. Is this right?
> Or is the power of the database strong enough for this full-text search?
There is no need for a full-text search. If you get all articles that have
the template on it, you will only need the text from those pages. Getting
the text then regexping it on hemlock (with a low priority) probably is
the best way to do it. You will need the text tables for it though, and
they are not yet available.
> Question 4.
> After the full-text search I will put the results in the database.
> Therefore I need also a Perl-example. Please help me.
See the same link.
> Question 5.
> If I can not do the full-text search at the MySQL-database I think it
> would be very helpful to make one directory for all users with all dumps
> (XML-files). So that every user can use this dumps. At the moment I have
> the current dump (DE+EN) in my home-directory, but I am sure that other
> users also have dumps. What did you think about this?
Wait until the text tables are there ;) - a FULLTEXT index of the text
table would be useful, but probably very space-consuming. However, if
zedler is overloaded by all the database requests, it could be an idea to
do xml-fulltext-searches on hemlock (on low priority). Just my $0.02
-valhallasw
Hello all,
river has move some disc-space from zedler to hemlock. It can be find
at /mnt/archiv/ and is writeable by all users. I had created a directory
"dumps" in it for wikimedia-dumps. Because "home" is a little bit small,
you should move all big files (like dumps) to /mnt/archiv/.
Please note, that the files there have no backup.
Sincerly,
DaB.
--
PGP: 2D3EE2D42B255885
Hi,
I work since two weeks on the tool server. In my directory home I have
extract the English and German dump. So that I can scan the dumps for
geocoordinates. At the moment I have a Perl script for this work. The
results is the KML for Google Earth and a CSV for many other services.
Now I see the most of you use the MySQL-database with all the nice
tables. I have no experience with MySQL, but I know SQL.
Question 1:
Can I use Perl to get some data from the database? (I work since three
weeks with Perl and CGI, so I need a small example.)
Question 2:
Can I use a SQL-Question with perl? For example: I want the category's
of page "xy" in DE.
Question 3.
When I scan all articles from the Dump (XML-File) for coordinates I need
with Perl in EN 45 minutes and in DE 15 minutes. If I use in the future
hopefully the MySQL-database I think this process will be to long for
the database. So that all other services have a problem. Is this right?
Or is the power of the database strong enough for this full-text search?
Question 4.
After the full-text search I will put the results in the database.
Therefore I need also a Perl-example. Please help me.
Question 5.
If I can not do the full-text search at the MySQL-database I think it
would be very helpful to make one directory for all users with all dumps
(XML-files). So that every user can use this dumps. At the moment I have
the current dump (DE+EN) in my home-directory, but I am sure that other
users also have dumps. What did you think about this?
Thanks,
Stefan (sk)
--
Stefan Kühn
Email: kuehn-s(a)gmx.net / Web: http://www.webkuehn.de