Hi all,
Import problem:
I tried to import (in several ways) the latest English dump, converted to proper SQL, but seemingly there's limit/setting somewhere halting the import when the text table reaches 4GB size (1995267 rows) in MyISAM (page and revision imported without problem though). What have I missed? A setting somewhere in mysql, table kind, no? --
Recommended DB tools?:
I also wonder if anyone has seen any very very well working db tools which shuffles data between databases efficiently. I need the following (unfortunately MySQL master/slave replication won't do (all) the tricks for me):
1. Both Local and Remote replication (via TCP/IP and/or Http/PHP "tunneling" for more restrictive hosts). 2. "Restore" database from sql-dump AND directly from another database, see #1 (i.e. Data Transfer and Data Synchronization, one way suffice). 3. Data Compression on remote connections.
In short, the tools should be able to do "all the tricks" regarding Backup/Restore & Transfer and Synchronization - with good speed over regular ~20Mb(/1) broadband internet connections. Perhaps one need to pay for such a tool but I would'nt pay much over $100.
What I have tried: 1. MySQL Adminstrator. I've played around with MySQL Adminstrator (Backup/Restore) but it takes tiiiiime to shuffle big tables. It took like 5 hours to upload the enwiki_page table. Unless I have missed some settings that's way too slow.
2. DumpTimer. http://www.dumptimer.com/index.php A lightweight and smart tool, opens stable connections to most any host out there (gzip-->unpacked by a serverside php-script), and uploads with fairly good pace, but only until ~200 Gb something is reached when it starts to slow down significantly. It's almost stalled at some 300 Gb. Dead end for big tables.
3. Navicat. http://www.navicat.com/ Looks really really (yes, really) good at first, it has all kinds of combinations possible (Backup/Restore, Transfer, Synchro, etc) but your joy is turned into desperation when it occurs to you that the thing first "Processes" the data, say 4,5 million rows... all the while memory peak is, well, increasing and increasing, and increasing... and when when passing 2 GB it decides to start shuffle the data also, but when the counters starts to spin its time for "out of memory". Sigh. (seems to be a good choice though if you don't have huuuge tables).
So, anyone, what's is the "ultimate tool" out there? Some tool which deals with both size and speed. I mean, can't be the only one having the need for a tool which does "all the stuff" you need to do with a bunch of remote databases...
Oh, and all features must be possible to schedule too, of course.
TIA,
// Rolf Lampa
Rolf Lampa wrote:
<snip>
- DumpTimer. ... uploads with fairly good pace, but only until
~200 *MB* something ... almost stalled at some 300 *MB*. Dead end for big tables.
MB, of course.
// Rolf Lampa
Rolf Lampa wrote:
Import problem:
<...> seemingly there's limit/setting somewhere halting the import when the text table reaches 4GB size (1995267 rows)
Found the problem: One of the extra (removable) disks had FAT32 which limits the file size to 4GB.
// Rolf Lampa
mediawiki-l@lists.wikimedia.org