Gordon @ IA was most friendly and helpful. archive-it is a subscription service for focused collections of sites; he had a different idea better suited to our work.
Gordon writes:
Now, given the importance of Wikipedia and editorial significant of things it outlinks-to, perhaps we could set up something specially focused on its content (and the de facto stream of newly-occurring outlinks), that would require no conscious effort by editors but greatly increase the odds that anything linked from Wikipedia would (a few months down the line) also be in our Archive. Is there (or could there be) a feed of all outlinks that IA could crawl almost nonstop?
That sounds excellent to me, if possible (and I think close to what emijrp had in mind!) What would it take to produce such a feed?
SJ
PS - An aside: IA's policies include taking down any links on request, so this would not be a foolproof archive, but a 99% one.
On Tue, Aug 24, 2010 at 9:13 PM, Samuel Klein meta.sj@gmail.com wrote:
I've asked Gordon Mohr @ IA about how to work with archive-it. I will cc: this thread on any response.
SJ
On Tue, Aug 24, 2010 at 8:56 PM, George Herbert george.herbert@gmail.com wrote:
On Tue, Aug 24, 2010 at 5:48 PM, Samuel Klein meta.sj@gmail.com wrote:
Here's the Archive's on-demand service:
That would be the most reliable way to set up the partnership emijrp proposes. And it's certainly a good idea. Figuring out how to make it work for almost all editors and make it spam-proof may be interesting.
SJ
On Tue, Aug 24, 2010 at 8:45 PM, Ray Saintonge saintonge@telus.net wrote:
David Gerard wrote:
On 24 August 2010 14:57, emijrp emijrp@gmail.com wrote:
I want to make a proposal about external links preservation. Many times, when you check an external link or a link reference, the website is dead or offline. This websites are important, because they are the sources for the facts showed in the articles. Internet Archive searches for interesting websites to save in their hard disks, so, we can send them our external links sql tables (all projects and languages of course). They improve their database and we always have a copy of the sources text to check when needed. I think that this can be a cool partnership.
+1
Are people who clean up dead links taking the time to check Internet Archive to se if the page in question is there?
Ec
foundation-l mailing list foundation-l@lists.wikimedia.org Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/foundation-l
-- Samuel Klein identi.ca:sj w:user:sj
foundation-l mailing list foundation-l@lists.wikimedia.org Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/foundation-l
I actually proposed some form of Wikimedia / IArchive link collaboration some years ago to a friend who worked there at the time; however, they left shortly afterwards.
I like SJ's particular idea. Who has current contacts with Brewster Kahle or someone else over there?
-- -george william herbert george.herbert@gmail.com
foundation-l mailing list foundation-l@lists.wikimedia.org Unsubscribe: https://lists.wikimedia.org/mailman/listinfo/foundation-l
-- Samuel Klein identi.ca:sj w:user:sj