I'm using the web API and db42. 

API example:  http://en.wikipedia.org/w/api.php?action=query&prop=revisions&titles=User:EpochFail&rvprop=ids|sha1&format=jsonfm&rvlimit=2&rvdir=older&rvstartid=488033783

MySQL example:
mysql> select rev_id, rev_sha1 from revision where rev_page = 15661504 and rev_id <= 488033783 order by rev_timestamp desc limit 2;
+-----------+---------------------------------+
| rev_id    | rev_sha1                        |
+-----------+---------------------------------+
| 488033783 | i8x0e29kxs2t1o9f1h03ks7q59yyyrv |
| 485404713 |                                 |
+-----------+---------------------------------+
2 rows in set (0.00 sec)

-Aaron

On Thu, May 10, 2012 at 4:03 PM, Diederik van Liere <dvanliere@wikimedia.org> wrote:
Which machine are you accessing?
D

On Thu, May 10, 2012 at 4:58 PM, Aaron Halfaker <aaron.halfaker@gmail.com> wrote:
Hey guys,

I'm trying to use the sha1 hashes of Wiki content for the first time (woot!  Props to D et al. for seeing it through) but I'm having some trouble actually getting them out of the API/databases.  It looks like the checksums only go back to April 19th.  Is this true of all pages?  Is there any plan to propagate the metric backwards?

-Aaron

_______________________________________________
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics



_______________________________________________
Analytics mailing list
Analytics@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/analytics