I'm using the web API and db42.
API example: http://en.wikipedia.org/w/api.php?action=query&prop=revisions&titles...
MySQL example: mysql> select rev_id, rev_sha1 from revision where rev_page = 15661504 and rev_id <= 488033783 order by rev_timestamp desc limit 2; +-----------+---------------------------------+ | rev_id | rev_sha1 | +-----------+---------------------------------+ | 488033783 | i8x0e29kxs2t1o9f1h03ks7q59yyyrv | | 485404713 | | +-----------+---------------------------------+ 2 rows in set (0.00 sec)
-Aaron
On Thu, May 10, 2012 at 4:03 PM, Diederik van Liere <dvanliere@wikimedia.org
wrote:
Which machine are you accessing? D
On Thu, May 10, 2012 at 4:58 PM, Aaron Halfaker aaron.halfaker@gmail.comwrote:
Hey guys,
I'm trying to use the sha1 hashes of Wiki content for the first time (woot! Props to D et al. for seeing it through) but I'm having some trouble actually getting them out of the API/databases. It looks like the checksums only go back to April 19th. Is this true of all pages? Is there any plan to propagate the metric backwards?
-Aaron
Analytics mailing list Analytics@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/analytics
Analytics mailing list Analytics@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/analytics