The data should be complete at the time I pulled it. It will be relatively cheap to update, so we could set a cron job to check against sitematrix every night. See details below.
analytics-store.eqiad.wmnet [staging]> explain wiki_info;
+-----------------+----------------+------+-----+---------+-------+
| Field | Type | Null | Key | Default | Extra |
+-----------------+----------------+------+-----+---------+-------+
| wiki | varbinary(100) | NO | PRI | | |
| code | varbinary(100) | YES | | NULL | |
| sitename | varbinary(100) | YES | | NULL | |
| url | varbinary(255) | YES | | NULL | |
| lang_id | int(11) | YES | | NULL | |
| lang_code | varbinary(100) | YES | | NULL | |
| lang_name | varbinary(255) | YES | | NULL | |
| lang_local_name | varbinary(255) | YES | | NULL | |
+-----------------+----------------+------+-----+---------+-------+
8 rows in set (0.00 sec)
analytics-store.eqiad.wmnet [staging]> select * from wiki_info limit 3;
+--------------+------------+-----------+--------------------------+---------+-----------+-----------+-----------------+
| wiki | code | sitename | url | lang_id | lang_code | lang_name | lang_local_name |
+--------------+------------+-----------+--------------------------+---------+-----------+-----------+-----------------+
+--------------+------------+-----------+--------------------------+---------+-----------+-----------+-----------------+
3 rows in set (0.02 sec)