got it!
C:\>wget.exe -O file.txt "http://en.wikipedia.org/wiki/Wget"
--2008-12-23 21:04:59--
http://en.wikipedia.org/wiki/Wget
Resolving
en.wikipedia.org... 208.80.152.2
Connecting to en.wikipedia.org|208.80.152.2|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 61182 (60K) [text/html]
Saving to: `file.txt'
100%[==============================================================================>]
61,182 107K/s in 0.6s
2008-12-23 21:05:00 (107 KB/s) - `file.txt' saved [61182/61182]
C:\>
Thanks!
- Rich
----- Original Message ----
From: "jidanni(a)jidanni.org" <jidanni(a)jidanni.org>
To: mediawiki-l(a)lists.wikimedia.org
Sent: Tuesday, December 23, 2008 12:57:37 PM
Subject: Re: [Mediawiki-l] maintinence script to append a new section to a page
Never mind libwww-perl's GET, let's use the more well known wget:
$ wget -O file.txt
'http://en.wikipedia.org/w/index.php?title=Not_Dead_Yet&action=raw'
--2008-12-24 01:52:21--
http://en.wikipedia.org/w/index.php?title=Not_Dead_Yet&action=raw
Resolving
en.wikipedia.org... 208.80.152.2
Connecting to en.wikipedia.org|208.80.152.2|:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 1732 (1.7K) [text/x-wiki]
Saving to: `file.txt'
100%[======================================>] 1,732 --.-K/s in 0.04s
2008-12-24 01:52:22 (45.4 KB/s) - `file.txt' saved [1732/1732]
Anyway, your mission is to save URLs like the above into a file.
_______________________________________________
MediaWiki-l mailing list
MediaWiki-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l