It tries to split the import / export to smaller batches. This will take a
while longer, but it works with bigger databases.
Best,
Simon
2015-05-15 18:03 GMT+02:00 John Foster <jfoster81747(a)gmail.com>om>:
I'm using the basic mysqldump for dumping and
mysql syntax for importing.
Moving from a Linux Mint LMDE server to an Ubuntu Enterprise server.
pretty vanilla stuff. The database is too big for phpmyadmin to do either &
it's just too slow. Also tried BigDump but that did not perform either.
Mostly figure I need to beef up mysql timing & tuning & have no experience
doing that.
On 05/14/2015 09:58 PM, John wrote:
are you using the dump/import script or a feature
of the database
software?
mysql or mariadb?
On Thu, May 14, 2015 at 10:52 PM, John Foster <jfoster81747(a)gmail.com>
wrote:
I have to move a mediawiki site that has a database of approx 4.1GB. It
is
somthing I have done before and I usually just use the command line
syntax
to do it. However I have tried 3 times to export it and the files are
always either incomplete due to server (Mysql) disconnect. I finally did
get one that seemed to complete OK but then I tried several times to
import
it into the new server and likewise got server disconnect errors. I am
aware that there are a multitude of possibilities, so what I'm asking
here
is for any tips on moving databases of this size or larger. I'm beginning
to wonder if Mysql is the way to go here.
Thanks!
--
John Foster
JW Foster & Associates
_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l
_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l
--
John Foster
JW Foster & Associates
_______________________________________________
MediaWiki-l mailing list
To unsubscribe, go to:
https://lists.wikimedia.org/mailman/listinfo/mediawiki-l