I know of two options:
1. (suggested by Brion Vibber) use the ImportDump maintenance program:
php maintenance/importDump.php big-old-file.xml
-or-
for x in *.xml
do
php maintenance/importDump.php $x
done
2. Use the Pywikipediabot Framework:
a. Install the pywikipediabot and configure it for your site.
b. Manipulate the data for your pages into one large file, preferrably named
dict.txt with {{start}} tags at the top of each page, {{stop}} tags at the
bottom, and the '''pagename''' as the first bolded item on each
page
(optionally enclosed in an HTML-style <!-- comment -->
c. Run the login.py script to log in.
d. Run the pagefromfile.py script. This will parse through the dict.txt file
quickly creating each page in the file on your wiki.
This should be a fairly quick process... complete in under an hour or two.
The pagefromfile.py script can occasionally bomb out. If that happens, look
at your wiki recentchanges log, note the last file uploaded, remove the data
in dict.txt from the top of the file to that page, and restart.
Good luck!
Tim
-----Original Message-----
From: wikitech-l-bounces(a)wikimedia.org
[mailto:wikitech-l-bounces@wikimedia.org] On Behalf Of Liz Kim
Sent: Wednesday, September 06, 2006 3:19 PM
To: wikitech-l(a)wikimedia.org
Subject: [Wikitech-l] directly write to database
Hello,
I have over 200 .html files that need to be turned into wiki. These are all
very simple text html files with maybe <b> tags and <p>.. very simple.
Instead of copying and pasting all these files one by one, is there a way to
access the database to write to it directly? Is there a way to easily
understand the db strucutre? I am thinking there could be a script that can
read the html code and i can easily parse out what I want to put into teh
edit window of wiki.. Thank you, Liz
_______________________________________________
Wikitech-l mailing list
Wikitech-l(a)wikimedia.org
http://mail.wikipedia.org/mailman/listinfo/wikitech-l