Mine is somewhat more complete in terms of queries, but I don't have editing support written yet.
Also, I think we are tackling it in slightly different ways in that you have the higher level helper functions for various API calls which I don't have, which mine expects the client code to do themselves (but i provide a generic "list" helper function in a similar way to your fetchWithContinuation.). I am constantly in two minds about how much to implement in the module itself, as I don't want a small api change to mean we need to update the module (I would prefer that the client code is instead changed)
Perhaps we can work together to make a single library. I have been planning to work on rewriting mine (it was originally proof-of-concept code, and significant reorganization is needed). My library is field-tested in VeblenBot and the WP 1.0 bot on enwiki.
Sure. Sounds like a good idea. Our bot was written to take over the handling of http://www.exotica.org.uk/wiki/UnExoticA which is currently using an older perl module. I've almost switched it over now to use my new module.
I'll be glad to dual-license my code in both GPL and the Perl license.
actually I was going to make mine GPL, I just haven't edited the license file etc. so I'm happy to go GPL also.
- The server may be temporarily unavailable, or the DNS may temporarily fail to resolve. This is true for any web-based service. The query has to be repeated if the server is temporarily unavailable.
I don't have any such stuff and I guess I should. Down to me making the bot primarily to replace some code we were using before, which we run on our own wiki (and on the server machine itself).
The squids may return an error message, which is marked by an HTTP 'x-squid-error' header. This means the query must be repeated.
You will quite routinely get a 0 byte response with a sucessful HTTP status code. This means the query must be repeated.
You need to use a maxlag= parameter, and detect when the maxlag is exceeded. If it is exceeded, the query must be repeated after the appropriate delay.
Rarely, you may get a truncated XML response, which causes the XML parser to throw an exception. This means the query must be repeated.
I assume these are all connected with Wikipedia as I never came across them on our installation. I will look at implementing this stuff though. I guess a config option with a retry count might be an idea.
You can see this error handling in the poorly named 'makeHTMLrequest' and 'makeXMLrequest' functions in my API.pm. These should be 'makeHTTPrequest' and 'makeXMLrequest', respectively.
Thanks. Will take a close look later (Since it is 5:47 am now and I really should not be awake).
Best Regards
Jools