There's an extension called LiveHTTPHeaders which allows the relevant request information to be captured and saved to a file.
http://livehttpheaders.mozdev.org/
Open it from the tools menu, then do the things that are slow while it's opened, then click "save all" and save it to a file. Don't post the file publically, since it will contain your login cookies, but you can send it to Krinkle as an email attachment, if he wants it.
-- Tim Starling
On 10/08/12 11:21, Eugene Zelenko wrote:
Hi, Krinkle!
I'm very sorry for beginner question, but how could get such log in Firefox 14? Is some extension available which could dump all pages with timestamps downloaded to view particular page? Or may be Firefox could do this itself?
Eugene.
On Thu, Aug 9, 2012 at 7:59 AM, Krinkle krinklemail@gmail.com wrote:
On Aug 9, 2012, at 4:49 PM, Eugene Zelenko wrote:
Hi!
I noticed that content from bits.wikimedia.org (including WikiEditor) is updated quite regularly - ~ every 20 minutes on Commons.
Such behavior is definitely creates problem for users with slow connections or with payed data traffic.
Are JavaScript/CCS are really updated so often?
Eugene.
Can you elaborate a bit? (urls, timestamps, http headers, ..)
-- Krinkle
Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l