In reply to:
Message: 9
Date: Fri, 12 Mar 2010 20:44:33 +0300
From: Dmitriy Sintsov <questpc(a)rambler.ru>
Subject: Re: [Wikitech-l] Uploads on small wikis
To: Wikimedia developers <wikitech-l(a)lists.wikimedia.org>
Message-ID: <596823262.1268415874.80897352.68859(a)mcgi68.rambler.ru>
Content-Type: text/plain; charset="us-ascii"; format="flowed"
* Aryeh Gregor <Simetrical+wikilist(a)gmail.com> [Fri, 12 Mar 2010
12:31:39 -0500]:
>> In principle, one could imagine hacking Squid to be smart enough to
>> cache contents and interface separately, and paste them together on
>> view about as quickly as it can serve plain requests now. But I don't
>> know how feasible that would be in practice.
>
>Perhaps an uploading from small wiki to commons via api while small wiki
>localizes client output with it's own language messages. Or, perhaps
>some parts of localization can be done with Javascript (which can read
>cookies, too), again not altering squid caching.
>Dmitriy
Stupid question, if having the ?uselang parameter can be done without
significant problems, wouldn't having javascript that just appeneds
?uselang= <whatevever its set to> to all internal links on a page
whenever the url for the current page has ?uselang=<something> also
not cause significant problems (well making uselang persist)?
(although then again, I'd imagine that would not be much different
then the use a cookie solution in terms of caching)
-bawolff
On Fri, Mar 12, 2010 at 1:12 PM, Michael Dale <mdale(a)wikimedia.org> wrote:
> Guillaume Paumier wrote:
>> Just FYI, we're working on both (crosswiki-upload and 1-click crosswiki
>> file move), but we're not quite there yet.
>>
> As mentioned on commons list a cross site upload tool is in early /
> alpha / experimental testing:
> http://lists.wikimedia.org/pipermail/commons-l/2010-March/005335.html
>
> To summarize from that post you can visit:
> http://en.wikipedia.org/w/index.php?title=Wikipedia:Sandbox&action=edit&wit…
>
I haven't seen this new withJS parameter anywhere in trunk or wmf-deployment,
only in js2-work, unless I'm being really dense today. When did this go live?
-Chad
Brian J Mingus wrote:
>
> On Wed, Mar 10, 2010 at 8:54 PM, Tomasz Finc <tfinc(a)wikimedia.org
> <mailto:tfinc@wikimedia.org>> wrote:
>
> Yup, that's the one. If you have a fast upload pipe then I'm more then
> happy to setup space for it. Otherwise it should be arriving in our
> snail mail after a couple of days.
>
> -tomasz
>
>
> Anyone may download the file from me here:
>
> http://grey.colorado.edu/enwiki-20080103-pages-meta-history.xml.7z
>
> The md5sum is:
>
> 20a201afc05a4e5f2f6c3b9b7afa225c enwiki-20080103-pages-meta-history.xml.7z
>
> The file size is:
>
> 18522193111 (~18 gigabytes)
>
> I'm sure you will find my pipe fat enough..;-)
>
>
> ------------------------------------------------------------------------
>
> _______________________________________________
> Xmldatadumps-admin-l mailing list
> Xmldatadumps-admin-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-admin-l
That seem way too tiny to be the real thing.
--tomasz
Newer snapshots super cede their old brethren. So a 20100130 already
includes all of the old content of 20080103 baring and format changes.
--tomasz
Kevin Webb wrote:
> Also, does the 20080103 dump combined with lastest 20100130 dump
> provide a complete edit history of Wikipedia? I'm unclear about
> whether the 20080103 dump was cumulative or if there was some other
> previous cut off point.
>
> Is it correct to assume that future dumps will begin post 2010-01-30?
>
> Thanks!
> kpw
>
> On Wed, Mar 10, 2010 at 10:55 PM, Kevin Webb <kpwebb(a)gmail.com> wrote:
>> It's in EC2 so I could get it to you in about 20 mins. Just hit me
>> with an email off-list with the desired destination...
>>
>> kpw
>>
>> On Wed, Mar 10, 2010 at 10:54 PM, Tomasz Finc <tfinc(a)wikimedia.org> wrote:
>>> Yup, that's the one. If you have a fast upload pipe then I'm more then happy
>>> to setup space for it. Otherwise it should be arriving in our snail mail
>>> after a couple of days.
>>>
>>> -tomasz
>>>
>>> Kevin Webb wrote:
>>>> Many thanks to everyone involved.
>>>>
>>>> Also, in case it's of use to anyone I have a copy of the
>>>> enwiki-20080103-pages-meta-history.xml dump in 7z form. Is that the
>>>> backup that's beeing referred to or is it in fact 20081003?
>>>>
>>>> kpw
>>>>
>>>> On Wed, Mar 10, 2010 at 10:20 PM, Tomasz Finc <tfinc(a)wikimedia.org> wrote:
>>>>> Thankfully due to an awesome volunteer we'll be able to get that 2008
>>>>> snapshot in our archive. I'll mail out when it shows up in our snail
>>>>> mail.
>>>>>
>>>>> --tomasz
>>>>>
>>>>> Erik Zachte wrote:
>>>>>> I'm thrilled. Big thanks to Tim and Tomasz for pulling this off.
>>>>>> For the record the 2008-10-03 dump existed for a short while only.
>>>>>> It evaporated before wikistats and many others could parse it,
>>>>>> so now we can finally catch up from 3.5 (!) years backlog.
>>>>>>
>>>>>> Erik Zachte
>>>>>>
>>>>>>> -----Original Message-----
>>>>>>> From: wikitech-l-bounces(a)lists.wikimedia.org [mailto:wikitech-l-
>>>>>>> bounces(a)lists.wikimedia.org] On Behalf Of Tomasz Finc
>>>>>>> Sent: Thursday, March 11, 2010 4:11
>>>>>>> To: Wikimedia developers; xmldatadumps-admin-l(a)lists.wikimedia.org;
>>>>>>> xmldatadumps(a)lists.wikimedia.org
>>>>>>> Subject: [Wikitech-l] 2010-03-11 01:10:08: enwiki Checksumming pages-
>>>>>>> meta-history.xml.bz2 :D
>>>>>>>
>>>>>>> New full history en wiki snapshot is hot off the presses!
>>>>>>>
>>>>>>> It's currently being checksummed which will take a while for 280GB+ of
>>>>>>> compressed data but for those brave souls willing to test please grab
>>>>>>> it
>>>>>>> from
>>>>>>>
>>>>>>> http://download.wikipedia.org/enwiki/20100130/enwiki-20100130-pages-
>>>>>>> meta-history.xml.bz2
>>>>>>>
>>>>>>> and give us feedback about its quality. This run took just over a month
>>>>>>> and gained a huge speed up after Tims work on re-compressing ES. If we
>>>>>>> see no hiccups with this data snapshot, I'll start mirroring it to
>>>>>>> other
>>>>>>> locations (internet archive, amazon public data sets, etc).
>>>>>>>
>>>>>>> For those not familiar, the last successful run that we've seen of this
>>>>>>> data goes all the way back to 2008-10-03. That's over 1.5 years of
>>>>>>> people waiting to get access to these data bits.
>>>>>>>
>>>>>>> I'm excited to say that we seem to have it :)
>>>>>>>
>>>>>>> --tomasz
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Wikitech-l mailing list
>>>>>>> Wikitech-l(a)lists.wikimedia.org
>>>>>>> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
>>>>>>
>>>>>> _______________________________________________
>>>>>> Xmldatadumps-admin-l mailing list
>>>>>> Xmldatadumps-admin-l(a)lists.wikimedia.org
>>>>>> https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-admin-l
>>>>> _______________________________________________
>>>>> Xmldatadumps-admin-l mailing list
>>>>> Xmldatadumps-admin-l(a)lists.wikimedia.org
>>>>> https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-admin-l
>>>>>
>>>
Yup, that's the one. If you have a fast upload pipe then I'm more then
happy to setup space for it. Otherwise it should be arriving in our
snail mail after a couple of days.
-tomasz
Kevin Webb wrote:
> Many thanks to everyone involved.
>
> Also, in case it's of use to anyone I have a copy of the
> enwiki-20080103-pages-meta-history.xml dump in 7z form. Is that the
> backup that's beeing referred to or is it in fact 20081003?
>
> kpw
>
> On Wed, Mar 10, 2010 at 10:20 PM, Tomasz Finc <tfinc(a)wikimedia.org> wrote:
>> Thankfully due to an awesome volunteer we'll be able to get that 2008
>> snapshot in our archive. I'll mail out when it shows up in our snail mail.
>>
>> --tomasz
>>
>> Erik Zachte wrote:
>>> I'm thrilled. Big thanks to Tim and Tomasz for pulling this off.
>>> For the record the 2008-10-03 dump existed for a short while only.
>>> It evaporated before wikistats and many others could parse it,
>>> so now we can finally catch up from 3.5 (!) years backlog.
>>>
>>> Erik Zachte
>>>
>>>> -----Original Message-----
>>>> From: wikitech-l-bounces(a)lists.wikimedia.org [mailto:wikitech-l-
>>>> bounces(a)lists.wikimedia.org] On Behalf Of Tomasz Finc
>>>> Sent: Thursday, March 11, 2010 4:11
>>>> To: Wikimedia developers; xmldatadumps-admin-l(a)lists.wikimedia.org;
>>>> xmldatadumps(a)lists.wikimedia.org
>>>> Subject: [Wikitech-l] 2010-03-11 01:10:08: enwiki Checksumming pages-
>>>> meta-history.xml.bz2 :D
>>>>
>>>> New full history en wiki snapshot is hot off the presses!
>>>>
>>>> It's currently being checksummed which will take a while for 280GB+ of
>>>> compressed data but for those brave souls willing to test please grab
>>>> it
>>>> from
>>>>
>>>> http://download.wikipedia.org/enwiki/20100130/enwiki-20100130-pages-
>>>> meta-history.xml.bz2
>>>>
>>>> and give us feedback about its quality. This run took just over a month
>>>> and gained a huge speed up after Tims work on re-compressing ES. If we
>>>> see no hiccups with this data snapshot, I'll start mirroring it to
>>>> other
>>>> locations (internet archive, amazon public data sets, etc).
>>>>
>>>> For those not familiar, the last successful run that we've seen of this
>>>> data goes all the way back to 2008-10-03. That's over 1.5 years of
>>>> people waiting to get access to these data bits.
>>>>
>>>> I'm excited to say that we seem to have it :)
>>>>
>>>> --tomasz
>>>>
>>>> _______________________________________________
>>>> Wikitech-l mailing list
>>>> Wikitech-l(a)lists.wikimedia.org
>>>> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
>>>
>>>
>>> _______________________________________________
>>> Xmldatadumps-admin-l mailing list
>>> Xmldatadumps-admin-l(a)lists.wikimedia.org
>>> https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-admin-l
>>
>> _______________________________________________
>> Xmldatadumps-admin-l mailing list
>> Xmldatadumps-admin-l(a)lists.wikimedia.org
>> https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-admin-l
>>
Does anybody have a bot or script that aligns the
categories for year of death across language
interwiki links, to make sure the death of a living
person is recorded on all languages? I think this
is something we need to run once or twice each year.
Maybe the interwiki bot could do it?
--
Lars Aronsson (lars(a)aronsson.se)
Aronsson Datateknik - http://aronsson.se
On Fri, Feb 12, 2010 at 2:36 PM, Siebrand Mazeland <s.mazeland(a)xs4all.nl> wrote:
> There are 36 FIXMEs left in Code Review at the moment (-18 compared to 2010-01-17). 14 are about commits made 2 or more months ago.
>
> Details: http://tr.im/NWF5 (mediawiki.org Code Review)
>
> Find below a table with open FIXME count per committer, 25 in total (-5 compared to last time).
>
> If you have not already addressed the Code Review comment(s), please do so. If the comments have been addressed, please set the issues status to 'new', or if you cannot do this, ask another user to do this for you.
>
> Thanks!
>
> Siebrand
>
> nimishg 4
> conrad 3
> mah 3
> reedy 2
> freakolowsky 2
> dale 2
> ialex 2
> ning 1
> ashley 1
> tomasz 1
> philip 1
> brion 1
> purodha 1
> happy-melon 1
> diana 1
> raymond 1
> jonwilliford 1
> shmichael 1
> tstarling 1
> tparscal 1
> vasilievvv 1
> adam 1
> werdna 1
> maxsem 1
> mglaser 1
>
>
> _______________________________________________
> Wikitech-l mailing list
> Wikitech-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
>
Currently, there are 24 FIXMEs on CodeReview, this is -12 compared
to last month. Below is the list of fixmes, by author:
adam 1
ashley 1
brion 1
catrope 1
dale 1
diana 1
happy-melon 1
ishimatsu 1
mah 1
nimishg 2
ning 1
pdhanda 2
philip 3
purodha 1
reedy 2
shmichael 1
tomasz 1
werdna 2
If you have not already addressed the Code Review comment(s), please
do so. If the comments have been addressed, please set the diff status
back to 'new', or if you cannot do this, ask another user to do this for you.
Also as a friendly reminder, please do not ever mark your own code as
OK or RESOLVED, even if you think it's fine. It's always great to review
other people's code, but let someone else review yours.
-Chad