[Mediawiki-l] maintinence script to append a new section to a page

Rich Evans revansx at yahoo.com
Wed Dec 24 02:07:33 UTC 2008


got it!

C:\>wget.exe -O file.txt "http://en.wikipedia.org/wiki/Wget"
--2008-12-23 21:04:59--  http://en.wikipedia.org/wiki/Wget
Resolving en.wikipedia.org... 208.80.152.2
Connecting to en.wikipedia.org|208.80.152.2|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 61182 (60K) [text/html]
Saving to: `file.txt'

100%[==============================================================================>] 61,182       107K/s   in 0.6s

2008-12-23 21:05:00 (107 KB/s) - `file.txt' saved [61182/61182]

C:\>

Thanks!
- Rich



----- Original Message ----
From: "jidanni at jidanni.org" <jidanni at jidanni.org>
To: mediawiki-l at lists.wikimedia.org
Sent: Tuesday, December 23, 2008 12:57:37 PM
Subject: Re: [Mediawiki-l] maintinence script to append a new section to a page

Never mind libwww-perl's GET, let's use the more well known wget:
$ wget -O file.txt 'http://en.wikipedia.org/w/index.php?title=Not_Dead_Yet&action=raw'
--2008-12-24 01:52:21--  http://en.wikipedia.org/w/index.php?title=Not_Dead_Yet&action=raw
Resolving en.wikipedia.org... 208.80.152.2
Connecting to en.wikipedia.org|208.80.152.2|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 1732 (1.7K) [text/x-wiki]
Saving to: `file.txt'

100%[======================================>] 1,732       --.-K/s   in 0.04s  

2008-12-24 01:52:22 (45.4 KB/s) - `file.txt' saved [1732/1732]

Anyway, your mission is to save URLs like the above into a file.

_______________________________________________
MediaWiki-l mailing list
MediaWiki-l at lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l



      



More information about the MediaWiki-l mailing list