Hi there,
among magic words we have
{{PAGESINCATEGORY:categoryname}}
counting quit well.
What I am looking for is a sibling, doing the same for special
namespaces and beeing filtered to *non categorized* pages.
DPL can filter based on namespaces and categories (here even with _none_
!) - but results are lists of articles but not counts.
Does anybody know a ready to use extension or another way to get this
figure dynamically in WikiText?
I want to avoid waste programming...
Thx.
Greetings,
Uwe
Dear list,
do you provide more information about the pages-logging dump somewhere?
While parsing it we came across some questions that we are trying to
clarify:
* Why are there only ~40 million logs (at over 450 million revisions)?
Which logs does the pages-logging dump / the "public" logging table
(not) contain? (We double-checked the number of logs on the database,
using our Toolserver-Account (Logging table).)
* Can you give us more information about the TextElement that is defined
for <logitem> in the XML Schema?
Definition: <element name="text" type="mw:TextType"/>
In pages-logging it sometimes occurs as:
<text deleted="deleted" />
How is this being used?
Kind regards,
Katja Mueller
Happy New Year everybody -- we survived 2011 and have another wacky fun
year of MediaWiki goodness to look forward to in 2012!
As we roll over our calendars, let's not forget it's also getting towards
time to roll out another MediaWiki release. 1.19's seen a lot of
refactoring goodness, bug fixes, and improvements; we need to make sure we
can get those improvements out to the public.
I'd like us to plan for a code freeze on trunk -- at least a feature &
refactoring freeze -- starting within a few days to give us all a chance to
tidy up, catch up with code review, and prepare for deployments.
The combinations of review, testing, and actual deployments are important
parts of our quality control system; we know from past experience that long
waits between deployments have lead to poorly-tested code getting pushed
out long after they've fallen out of our collective memory, making bugs
newly discovered in them harder to find.
I believe we've got some general plans to hit deployment in February; if we
hit code slush this week or so that gives some time for everybody to catch
up on review, do more thorough testing and -- perhaps most importantly of
all -- make sure we have good documentation on what's changed and what
needs to be tested and tried out by real humans!
Note that I'm recommending a slush/freeze on trunk rather than simply
branching and doing review on the release branch because that's been hard
to manage in the past. I think we really need to concentrate on release
polish, especially making sure that we don't have unexpected regressions
and that people know what to expect has changed in the user interface and
in behavior. Even "small" features like the image rotation changes we
landed in 1.18 can have surprising consequences, and need to be clearly on
peoples' radar before they're irrevocable.
-- brion vibber (brion @ pobox.com / bvibber @ wikimedia.org)
Welcome Andrew! Happy to see you onboard the Wikimedia train :-)
--
Alolita Sharma
Features Engineering
WMF
----- Reply message -----
From: "Diederik van Liere" <dvanliere(a)gmail.com>
To: "Wikimedia developers" <wikitech-l(a)lists.wikimedia.org>
Cc: "Wikimedia developers" <wikitech-l(a)lists.wikimedia.org>, "Andrew Otto" <aotto(a)wikimedia.org>
Subject: [Wikitech-l] Welcome, Andrew Otto - Software Developer for Analytics
Date: Fri, Jan 6, 2012 10:28 am
Welcome andrew!
Super excited to have you joining us!
Diederik
Sent from my iPhone
On 2012-01-06, at 13:13, Sumana Harihareswara <sumanah(a)wikimedia.org> wrote:
> On 01/06/2012 01:08 PM, Rob Lanphier wrote:
>> We're really excited to have Andrew on board to help bring some
>> systems rigor to our data gathering process. Our current data mining
>> regime involves a few pieces of lightweight data gathering
>> infrastructure (e.g. udp2log), a combination of one-off special
>> purpose log crunching scripts, along with other scripts that started
>> their lives as one-off special purpose scripts, but have gradually
>> become core infrastructure. Most of these scripts have single
>> maintainers, and there is a lot of duplication of effort. In
>> addition, the systems have a nasty tendency to break at the least
>> opportune times. Andrew's background bringing sanity to insane
>> environments will be enormously helpful here.
>
> (See episode S10E07, "The Shadow Scripts,"
> https://blog.wikimedia.org/2011/10/31/data-analytics-at-wikimedia-foundatio…
> )*
>
>> Andrew is based out of Virginia, but is still traveling the world.
>> Right now, you'll find him in New York City. Please join me in
>> welcoming Andrew to the team!
>
> I congratulated him IN PERSON five minutes ago, because we're coworking
> today. There's another New Yorker now, yay!
>
> --
> Sumana Harihareswara
> Volunteer Development Coordinator
> Wikimedia Foundation
>
> * I am being silly and acting as though this blog entry were an episode
> of a science fiction TV show.
>
> _______________________________________________
> Wikitech-l mailing list
> Wikitech-l(a)lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikitech-l
_______________________________________________
Wikitech-l mailing list
Wikitech-l(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Hi everyone,
I'm pleased to announce Andrew Otto will be coming to Wikimedia
Foundation as a software developer in Platform Engineering, focused on
analytics. We've been hiring for this spot for quite some time, and
I'm happy we held out for Andrew.
Andrew comes to us from CouchSurfing, where he worked for the past
four years as one of the very early technical staff there, working in
various places throughout the world (Thailand, Alaska, and New York
are the ones I recall). His team scaled their systems from a few web
servers and one monolithic database, to a cluster of over 30 machines
handling almost 100 million page views per month. He was responsible
for introducing Puppet for system configuration at his last job, and
much of his work at CouchSurfing has been in reviewing code and
maintaining a consistent architecture for CouchSurfing.
We're really excited to have Andrew on board to help bring some
systems rigor to our data gathering process. Our current data mining
regime involves a few pieces of lightweight data gathering
infrastructure (e.g. udp2log), a combination of one-off special
purpose log crunching scripts, along with other scripts that started
their lives as one-off special purpose scripts, but have gradually
become core infrastructure. Most of these scripts have single
maintainers, and there is a lot of duplication of effort. In
addition, the systems have a nasty tendency to break at the least
opportune times. Andrew's background bringing sanity to insane
environments will be enormously helpful here.
Andrew has an email address and is technically starting the onboarding
process, but is still wrapping up at CouchSurfing. He'll be with us
part-time starting January 17, and ramping up to full-time starting in
April.
Andrew is based out of Virginia, but is still traveling the world.
Right now, you'll find him in New York City. Please join me in
welcoming Andrew to the team!
Rob
Hi,
one of our database servers, db22, had a disk failure a little while
ago, and while this failed disk was to be replaced another RAID
problem appeared.
This caused downtime of db22 and users started reporting problems at
around 7 pm:
19:09 < malafaya> so, what's wrong?
was even before:
19:19 <+nagios-wm> PROBLEM - Host db22 is DOWN: PING CRITICAL - Packet
loss = 100%
Since this affected CentralAuth, users kept getting error messages
like: [db22: s4] 10.0.6.32
Database ops immediately started moving a database slave to be the new
master, while the hardware issue on db22 is still being investigated.
The current effect is that commons is read-only. The expected downtime
was at 10 minutes when writing this.
--
--
Daniel Zahn <dzahn(a)wikimedia.org>
Dear all,
I hope your season's holidays are going well.
I'm writing to inform you that I have released on Github a first (0.1)
version of wikicapthca[1] a ReCAPTCHA-like program for Wiki*. The
thing is born from an initial observation by Alex brollo[2] about
which we have discussed much both in WMI mailing list and also on
Wikisource-l[3].
For starters the code there is a rewriting of Alex's scripts and
nothing more, and here's what the program does for now: 1) gets a
djvu, 2) extracts the text layer, 3) identifies non recognized words
and 4) produce a tiff image of them.
But I would like to write a "proof of concept" of the whole process
from getting OCR-ed djvu's from Commons to producing challenges,
serving them, collecting answer and then using those answers in some
useful way. That's said, you see there's still a long way to go.
Obviously in the long run we could use this system as a backup for the
current one, which has demonstrated some limitations[4], but I'm sure
there are many aspects of the problem which go beyond my knowledge
(I'm a physicist not a computer scientist, you know) so any help
and/or advice is welcome.
Thanks for your time.
Cristian
[1]https://github.com/CristianCantoro/wikicaptcha
[2]http://it.wikisource.org/wiki/Utente:Alex_brollo
[3]http://lists.wikimedia.org/pipermail/wikisource-l/2011-February/000939.ht…
[4]http://lists.wikimedia.org/pipermail/wikitech-l/2011-November/056078.html
Greetings!
Utilizing the model set by Wikimedia staff IRC office hours, MediaWiki.org's WikiProject Extensions is presenting our first MediaWiki Workshops for developers (volunteer and staff). Preparing extensions for MediaWiki 1.19 will be held on 13 January, 2012 at 19:00 UTC in IRC (#wikimedia-dev).
This IRC workshop will be an opportunity to find out about changes in MediaWiki 1.19 that may require revisions to extensions or skins. Also an opportunity to ask MediaWiki developers questions regarding extension development.
Everyone is invited to attend. Developers interested in serving as "extensions" or "MediaWiki 1.19" experts are encouraged to signup as participants at: http://www.mediawiki.org/wiki/Project:WikiProject_Extensions/MediaWiki_Work…
If there's continued interest / demand, MediaWiki Workshops will typically last one hour, and happen no more than twice a month to present trainings, hold discussions and collaborate on community or WikiProject Extensions projects. Facilitators host the session to introduce any presenters, determine the order of questions, and generally helps to keep things going. Time of day will vary in order to offer people in different parts of the world the opportunity to participate. Future topics will likely include MediaWiki.org documentation, ResourceLouder orientation and workshops similar to our inaugural chat to prepare for each MediaWiki release.
More information: http://www.mediawiki.org/wiki/Project:WikiProject_Extensions/MediaWiki_Work…MediaWiki.org's WikiProject SysAdmins will also likely host similar MediaWiki Workshops to help third-party wiki system administrators. Stay tuned for more information, and chime in at: http://www.mediawiki.org/wiki/Project:WikiProject_SysAdmins/Ideas
Look forward to seeing folks next Friday! Please feel free to forward this along to any interested folks.
-greg aka varnent
PS. Sorry if this is a duplicate - was having listserv problems earlier.
-------
Gregory Varnum
Lead Administrator, WikiQueer
Lead, Aequalitas Project
@GregVarnum
fb.com/GregVarnum