OK, now that I've looked into the code some more (looking at 1.9.3). I see that I have a somewhat different problem.
Here's what's going on. I have some wikis where I regularly update page content from an external source using a script that generates xml like what the exporter generates. These get loaded via importdump.php and are marked in the page histories as revisions...which is working fine.
The problem is that I want to be able to see in recentchanges when a human has edited the pages. So, I want to have the script-generated pages marked as bot edits and let the human changes show through. I can mark the script-generated pages as being from a bot by flagging them in the recentchanges...but the human ones still aren't there because the limit on 5000 entries in recentchanges when rebuildrecentchanges.php runs. My scripts run daily and they typically affect more than 5000 pages at a time.
I'm concerned that just raising or eliminating the limit will make things unacceptably slow and make the table too large. I'm wondering about hacking the script to separate passes for bots and non-bots. Thoughts?
Jim
p.s. I decided to sent this just to mediawiki-l and not wikitech, since presumably the wikipedias don't have this problem.
On May 20, 2007, at 5:26 PM, Jim Hu wrote:
Yes, I know that.. I should have said recentchanges instead of revisions.
But they're not flagged in the recentchanges table. The problem is that they show up whether or not one uses hide bots in Special:Recentchanges. This is fixable by manually updating recentchanges with,
update recentchanges set rc_bot=1 where rc_user=<bot_user_id>;
but I was hoping that importDump would do this automatically based on recognizing the username. However, upon reflection, I believe that importDump doesn't do anything directly to recentchanges - I usually have to rebuild to get changes to show up. Thus, if there is a fix needed, it should be in rebuildrecentchanges.php or somewhere else.
JH
On May 20, 2007, at 3:29 PM, Aaron Schulz wrote:
Hmm, bot edits are only observed as "bot edits" in the recentchanges table. Edits by bots in the revision table are not actually flagged as bot edits.
<div><FONT color=#3333cc>-Aaron Schulz</FONT></div></html>
From: Jim Hu jimhu@tamu.edu Reply-To: Wikimedia developers wikitech-l@lists.wikimedia.org To: Wikimedia developers <wikitech- l@lists.wikimedia.org>,MediaWiki announcements and site admin listmediawiki-l@lists.wikimedia.org Subject: [Wikitech-l] importDump and setting rc_bot Date: Sun, 20 May 2007 13:22:07 -0400
As far as I can tell, importDump does not mark imported pages as coming from a bot, even when the user is a bot in the User table. Is that correct? Is there a way to indicate a bot revision in the xml, or do I need to do this in the db afterward? ===================================== Jim Hu Associate Professor Dept. of Biochemistry and Biophysics 2128 TAMU Texas A&M Univ. College Station, TX 77843-2128 979-862-4054
Wikitech-l mailing list Wikitech-l@lists.wikimedia.org http://lists.wikimedia.org/mailman/listinfo/wikitech-l
Make every IM count. Download Messenger and join the i’m Initiative now. It’s free. http://im.live.com/messenger/im/home/? source=TAGHM_MAY07
Wikitech-l mailing list Wikitech-l@lists.wikimedia.org http://lists.wikimedia.org/mailman/listinfo/wikitech-l
===================================== Jim Hu Associate Professor Dept. of Biochemistry and Biophysics 2128 TAMU Texas A&M Univ. College Station, TX 77843-2128 979-862-4054
Wikitech-l mailing list Wikitech-l@lists.wikimedia.org http://lists.wikimedia.org/mailman/listinfo/wikitech-l
===================================== Jim Hu Associate Professor Dept. of Biochemistry and Biophysics 2128 TAMU Texas A&M Univ. College Station, TX 77843-2128 979-862-4054