Hi Aaron,
I'm using the web API and db42.MySQL example:mysql> select rev_id, rev_sha1 from revision where rev_page = 15661504 and rev_id <= 488033783 order by rev_timestamp desc limit 2;+-----------+---------------------------------+| rev_id | rev_sha1 |+-----------+---------------------------------+| 488033783 | i8x0e29kxs2t1o9f1h03ks7q59yyyrv || 485404713 | |+-----------+---------------------------------+2 rows in set (0.00 sec)-AaronOn Thu, May 10, 2012 at 4:03 PM, Diederik van Liere <dvanliere@wikimedia.org> wrote:
Which machine are you accessing?
DOn Thu, May 10, 2012 at 4:58 PM, Aaron Halfaker <aaron.halfaker@gmail.com> wrote:
_______________________________________________Hey guys,I'm trying to use the sha1 hashes of Wiki content for the first time (woot! Props to D et al. for seeing it through) but I'm having some trouble actually getting them out of the API/databases. It looks like the checksums only go back to April 19th. Is this true of all pages? Is there any plan to propagate the metric backwards?-Aaron
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics
_______________________________________________
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics
_______________________________________________
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics