Do we send a cc: to Wikimedia tech mailing list?
2010/8/28 Samuel Klein <meta.sj(a)gmail.com>
Gordon @ IA was most friendly and helpful. archive-it
is a
subscription service for focused collections of sites; he had a
different idea better suited to our work.
Gordon writes:
Now, given the importance of Wikipedia and
editorial significant of
things
it outlinks-to, perhaps we could set up something
specially focused on
its
content (and the de facto stream of
newly-occurring outlinks), that would
require no conscious effort by editors but greatly increase the odds that
anything linked from Wikipedia would (a few months down the line) also be
in our Archive. Is there (or could there be) a feed of all outlinks that
IA
could crawl almost nonstop?
That sounds excellent to me, if possible (and I think close to what
emijrp had in mind!) What would it take to produce such a feed?
SJ
PS - An aside: IA's policies include taking down any links on request,
so this would not be a foolproof archive, but a 99% one.
On Tue, Aug 24, 2010 at 9:13 PM, Samuel Klein <meta.sj(a)gmail.com> wrote:
I've asked Gordon Mohr @ IA about how to work
with archive-it. I will
cc: this thread on any response.
SJ
On Tue, Aug 24, 2010 at 8:56 PM, George Herbert
<george.herbert(a)gmail.com> wrote:
> On Tue, Aug 24, 2010 at 5:48 PM, Samuel Klein <meta.sj(a)gmail.com>
wrote:
>> Here's the Archive's on-demand
service:
>>
>>
http://archive-it.org
>>
>> That would be the most reliable way to set up the partnership emijrp
>> proposes. And it's certainly a good idea. Figuring out how to make
>> it work for almost all editors and make it spam-proof may be
>> interesting.
>>
>> SJ
>>
>>
>>
>> On Tue, Aug 24, 2010 at 8:45 PM, Ray Saintonge <saintonge(a)telus.net>
wrote:
>>> David Gerard wrote:
>>>> On 24 August 2010 14:57, emijrp <emijrp(a)gmail.com> wrote:
>>>>
>>>>> I want to make a proposal about external links preservation. Many
times,
>>>>> when you check an external
link or a link reference, the website is
dead or
>>>>> offline. This websites are
important, because they are the sources
for the
>>>>> facts showed in the articles.
Internet Archive searches for
interesting
>>>>> websites to save in their
hard disks, so, we can send them our
external
>>>>> links sql tables (all
projects and languages of course). They
improve their
>>>>> database and we always have a
copy of the sources text to check when
needed.
>>>>> I think that this can be a
cool partnership.
>>>>>
>>>> +1
>>>>
>>>>
>>> Are people who clean up dead links taking the time to check Internet
>>> Archive to se if the page in question is there?
>>>
>>>
>>> Ec
>>>
>>> _______________________________________________
>>> foundation-l mailing list
>>> foundation-l(a)lists.wikimedia.org
>>> Unsubscribe:
https://lists.wikimedia.org/mailman/listinfo/foundation-l
>
--
Samuel Klein identi.ca:sj w:user:sj
_______________________________________________
foundation-l mailing list
foundation-l(a)lists.wikimedia.org
Unsubscribe:
https://lists.wikimedia.org/mailman/listinfo/foundation-l
I actually proposed some form of Wikimedia / IArchive link
collaboration some years ago to a friend who worked there at the time;
however, they left shortly afterwards.
I like SJ's particular idea. Who has current contacts with Brewster
Kahle or someone else over there?
--
-george william herbert
george.herbert(a)gmail.com
_______________________________________________
foundation-l mailing list
foundation-l(a)lists.wikimedia.org
Unsubscribe:
https://lists.wikimedia.org/mailman/listinfo/foundation-l
--
Samuel Klein identi.ca:sj w:user:sj
--
Samuel Klein identi.ca:sj w:user:sj
_______________________________________________
foundation-l mailing list
foundation-l(a)lists.wikimedia.org
Unsubscribe:
https://lists.wikimedia.org/mailman/listinfo/foundation-l