My idea for faster, cheaper access to Wikipedia in poorer countries:
http://meta.wikimedia.org/wiki/Reducing_transit_requirements
It requires patching squid and extending the HTCP protocol. The latter
should be no big deal, the protocol is only in the "experimental" phase
after all.
-- Tim Starling
Will the upgrade to MySQL 5 give us (within the limits of Wikipedia per
language) culture correct sorting or just the default Unicode collation?
Peter
--
Lust, ein paar Euro nebenbei zu verdienen? Ohne Kosten, ohne Risiko!
Satte Provisionen f�r GMX Partner: http://www.gmx.net/de/go/partner
Since Domas keeps complaining about the database load from the dumps
(and then killing the dump processes), I've made some changes which
should reduce the load involved.
Dumps are now being generated on a two-pass system. The first pass reads
through page and revision quickly and makes a stub dump, with
rev_text_id references in place of the full page text.
The second pass reads this stub dump, and the previous full dump of the
same database. Existing revision text can be copied directly from the
previous dump (page contents on a given revision ID are immutable). New
revisions not in the old dump are read individually out of the database,
using the rev_text_id to avoid having to hit the page or revision tables.
At the moment I'm doing the full/current/articles split on the first
pass, and the compression to bzip2 and 7zip is done on the second pass
with the final data.
Hopefully this should go a little smoother.
Also, last week the mwdumper dump import tool got a number of optimizations:
* Inserts are batched more efficiently for bulk insert
* Folke Behrens sent a patch to rearrange and properly buffer things to
significantly speed up the XML input and SQL generation
* You can have it connect directly to the MySQL server if you have the
MySQL Connector/J driver in classpath.
* There are some hints in the README on server configuration tweaks for
faster import
A precompiled .jar of the current code is available at:
http://download.wikipedia.org/tools/
Source is in CVS, module mwdumper.
It's known to work with Sun's 1.5 JDK and GNU GCJ 4.0.1. Sun Java 1.4
may have problems with some dumps (known to fail on the last Japanese
Wikipedia dump.)
-- brion vibber (brion @ pobox.com)
I've got a dump of simple loaded.
When revision.rev_delete = 1, is there a way to tell both when the rev
was created and when the rev was deleted?
Thanks!
MySQL 5 is scheduled to come out of beta next month, and we're going to
be looking at upgrading sometime in the coming months. Among other
things we're probably going to want to start making use of the support
for Unicode collation, so we can get better sorting and perhaps use it
for case-insensitive matching.
There is however a compatibility issue: MySQL's Unicode support is
limited to the 16-bit character range (basic multilingual plane), both
for ucs2 and utf8 storage modes.
Characters beyond the BMP are relatively rare, but they do occur. Mostly
in there are ancient/dead scripts, some invented scripts, and a bunch of
rare Han characters which sometimes turn up in Chinese and Japanese.
This won't affect page _contents_; our content is stored in binary blobs
and can have any wacky characters we want. But to support these high
characters in page titles, usernames, and such might require jumping
through a lot of hoops.
It would be relatively simple to disable use of titles and usernames
with these high characters; to assess possible impact I did a check
through all our current wikis and found 99 extant pages:
43 in en.wiktionary.org
31 in got.wikipedia.org
10 in la.wiktionary.org
9 in zh.wikipedia.org
3 in so.wikipedia.org
1 in en.wikibooks.org
1 in ja.wikipedia.org
1 in nl.wikibooks.org
I've put the full list of pages here:
http://meta.wikimedia.org/wiki/User:Brion_VIBBER/Unicode_high_chars
Most of the en.wiktionary entries are individual letters in the Deseret
and Shavian alphabets (invented alphabets for English; historical
curiosities).
The Gothic alphabet is entirely in the high-character area, but it's a
long-dead language and not exactly an active wiki. Perhaps we should
just close it down...
Latin Wiktionary contains several Gothic terms...
The Chinese Wikipedia contains several apparently legitimate articles
(from what I can tell) using high characters; these might have to be
moved. The Japanese Wikipedia has one redirect with such a character.
The Somali Wikipedia contains three one-sentence stub pages pages using
the Osmanya script; Omniglot's article on it says this script is no
longer in use since adoption of the Latin alphabet in 1972.
English Wikibooks has a user account with a Gothic-script name, which
has edited a number of pages about the Gothic language and has a user page.
Dutch Wikibooks has one Gothic-titled redirect.
-- brion vibber (brion @ pobox.com)
I am trying to add references in such a way that
1) i have added a button in the toolbar that should open a form for user to
enter the citation info, and enter a no for this info in the article. When the
user saves the page the citations will be added at the end in the order that
they appear in the article.
2)Each of thses will be links with their own page and talk page.
3) I think the 2nd part will be similar to working of the category namespace.
Can anyone please tell me how i can open a pop up form when a butotn in the
toolbar is clicked. I am stuck at that right now. Also how do i add pages to
the cctaegory namespace automatically?
Thanks,
--
Amruta
I have a copy of cvs 1.6 working on postgresql and would like to submit
patches. The website says I should ask for CVS access, which I am comfortbale
with, though I am also comfortable submitting patches if there is
someone/place specific I should be sending them too. Thanks in advance.
--
Robert Treat
Build A Brighter Lamp :: Linux Apache *mediawiki* PostgreSQL
You seem to misunderstand. If it does not exist fully as of yet (although it seems that there was at one time some work on the feature, or else why the restrictions field in the cur table), I'm willing to write it and give the change back to the community. I was asking for advice on implementation in the rest of the email.
Gabe
-----Original Message-----
From: wikitech-l-bounces(a)wikimedia.org [mailto:wikitech-l-bounces@wikimedia.org] On Behalf Of Brion Vibber
Sent: Thursday, October 13, 2005 4:31 PM
To: Wikimedia developers
Subject: Re: [Wikitech-l] Adding page view restrictions
Sechan, Gabe wrote:
> I'm using MediaWiki to create a corporate wiki, and need to allow
> access to certain pages only to certain groups. I don't see full
> support for this in MediaWiki as of 1.4.11
There is no support for that in MediaWiki; I recommend you use some other product which suits your requirements.
-- brion vibber (brion @ pobox.com)
I've started a user help page on Special:CheckUser at
[[:meta:Help:CheckUser]]. I've tried to write it neutrally for any
MediaWiki installation, not just Wikimedia. Clarifications and
additions are welcomed! I've been particularly cautious in my
assertions concerning the privacy policy ...
- d.