I'm using the web API and db42.
API example:
http://en.wikipedia.org/w/api.php?action=query&prop=revisions&title…
MySQL example:
mysql> select rev_id, rev_sha1 from revision where rev_page = 15661504 and
rev_id <= 488033783 order by rev_timestamp desc limit 2;
+-----------+---------------------------------+
| rev_id | rev_sha1 |
+-----------+---------------------------------+
| 488033783 | i8x0e29kxs2t1o9f1h03ks7q59yyyrv |
| 485404713 | |
+-----------+---------------------------------+
2 rows in set (0.00 sec)
-Aaron
On Thu, May 10, 2012 at 4:03 PM, Diederik van Liere <dvanliere(a)wikimedia.org
wrote:
Which machine are you accessing?
D
On Thu, May 10, 2012 at 4:58 PM, Aaron Halfaker <aaron.halfaker(a)gmail.com>wrote;wrote:
Hey guys,
I'm trying to use the sha1 hashes of Wiki content for the first time
(woot! Props to D et al. for seeing it through) but I'm having some
trouble actually getting them out of the API/databases. It looks like the
checksums only go back to April 19th. Is this true of all pages? Is there
any plan to propagate the metric backwards?
-Aaron
_______________________________________________
Analytics mailing list
Analytics(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics
_______________________________________________
Analytics mailing list
Analytics(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics