For wikistats I would like to aggregate edit counts per editor for all
wikis, so that we get a better count for how many people work on our
projects.
I checked several wiki pages on SUL, but am a bit unclear to where we are
now.
At the moment: can different Wikimedia wikis still have same names for
different users?
So that John or Jane on one wiki belongs to person x, and on another wiki to
person y?
My hunch is yes, as I believe total migration is still a work in progress,
but please confirm.
And if so, is there a way I can tell which user names on which wikis are
post SUL and unified?
Thanks, Erik Zachte
Does anyone on the wikitech mailing list happen to know whether it
would be possible for some of the larger wikipedia database downloads
(which are, say, 16GB or so in size) to be split into parts so that
they can be downloaded. For whatever reason, whenever I have
attempted to download the ~14GB files (say, from
http://static.wikipedia.org/downloads/2008-06/en/ ), I have found that
only 2GB (presumably, the first 2GB) of what I have sought to download
has actually been downloaded. Is there anyway around this? Could
anyone possibly suggest what possible reasons there might be for this
difficulty in downloading the material?
Thanks. .
And $wgExtNewTables and friends
-Chad
On Apr 14, 2009 5:56 PM, "Roan Kattouw" <roan.kattouw(a)gmail.com> wrote:
2009/4/14 Platonides <Platonides(a)gmail.com>:
> Prince Gerald Albert wrote: >> Is it a good practice to create my own
table to store my extension ...
Note that there's also the LoadExtensionSchemaUpdates hook [1], which
you can use to have your schema changes done as part of update.php .
Of course users still have to run update.php in that case, but they
have to do that on upgrade as well.
Roan Kattouw (Catrope)
[1] http://www.mediawiki.org/wiki/Manual:Hooks/LoadExtensionSchemaUpdates
_______________________________________________ Wikitech-l mailing list
Wikitech-l(a)lists.wikimedia....
I'm writing an extension for mediawiki, where I require creating tables
to store my data and access it back.
I tried searching several places to know how I can create a new table to
use it for my extension.
Can somebody guide me in doing so or point me to links where I can know
the details of it.
Is it a good practice to create my own table to store my extension
related data to it? Is there any standard practice for it?
Thanks
Prince Gerald.
DISCLAIMER:
-----------------------------------------------------------------------------------------------------------------------
The contents of this e-mail and any attachment(s) are confidential and intended for the named recipient(s) only.
It shall not attach any liability on the originator or HCL or its affiliates. Any views or opinions presented in
this email are solely those of the author and may not necessarily reflect the opinions of HCL or its affiliates.
Any form of reproduction, dissemination, copying, disclosure, modification, distribution and / or publication of
this message without the prior written consent of the author of this e-mail is strictly prohibited. If you have
received this email in error please delete it and notify the sender immediately. Before opening any mail and
attachments please check them for viruses and defect.
-----------------------------------------------------------------------------------------------------------------------
I don't edit very often anymore, but I saw this as it happened to be be
immediately above something else....
https://secure.wikimedia.org/wikipedia/en/wiki/Wikipedia:Templates_for_dele…
A large 68K template, listing all the congresscritters ever elected in NY,
organized by district number. Meta template usage. Lots of template calls.
All the information is already in a 100K+ master list article, or another
organized by district, or another series of district articles.
Since all the information is also in a category, it's pointless. I've
tried to suggest alternatives (such as an existing template that points to
each district article, where the same information is already displayed).
Mostly occurs on 1,900+ stubs, where the underlying edit is only a few K.
As displayed, the huge template overwhelms the tiny articles.
Nominated for deletion because of its size. Thought folks might like to
know about it -- previous discussions of template server thrashing.
Added today:
Rememberthedot -- general patches and awesomeness.
Aude -- OpenStreetMap, SlippyMap extension, and other geodata-related
goodies.
Welcome aboard!
I've asked a few other folks from my email archives for some account
details and sample patches. If you've been inquiring about commit access
and *haven't* heard from me today, please resend!
I'll see about setting up a clearer open request queue and giving a
couple more people access to add folks so we don't get so behind.
-- brion
I am wondering if anyone has some contributer "browser-client"
information handy or can point me to a some dataset that I could query
to get the following information:
1) What is wikipedia client browser usage percentage distribution ( I
recall that being published recently but I have misplaced the link)
2) What is the distribution of browsers for people that visit the
?action=edit pages (on en.wikipedia)
3) What is the distribution of browsers for people that submit post data
to /wiki/Special:Upload on commons. (ie upload images)
thanks,
michael
On thistle with DB=dewiki:
mysql> explain select * from recentchanges
left join tag_summary on ts_rc_id=rc_id
order by rc_timestamp desc limit 50\G
*************************** 1. row ***************************
table: recentchanges
type: ALL
possible_keys: NULL
key: NULL
key_len: NULL
ref: NULL
rows: 1179921
Extra: Using temporary; Using filesort
*************************** 2. row ***************************
table: tag_summary
type: ALL
possible_keys: ts_rc_id
key: NULL
key_len: NULL
ref: NULL
rows: 4
Extra:
2 rows in set (0.00 sec)
Whenever you do a join with a limit, MySQL gets the query plan wrong.
It scans the small table and filesorts the large table. You have to
use FORCE INDEX on the small table to suppress the scan. We've seen
this many times. It's very difficult to detect during code review and
frequently crashes the site.
Does anyone know a DBMS where joining with limits actually works?
Because I'm sick of this crap.
-- Tim Starling