Dear All,
I need to integrate the number of Wikipedia pages containing specific word in my application. for example if I write car, I want to receive all page count that are in Wikipedia and containing that word in any position of the article content. I have tried to use special search and retrieve and raw wikitext article but it is not efficient. I will appreciate if you can forward any links or information leading to that.
Cheers,
Muhidin A.
Hi guys,
Hmmm, I am in digest mode in the mailing list and I am not sure on how
to respond efficiently to the digest email. Let's try.
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Thu, 07 Nov 2013 18:52:06 +0100
> From: Alvaro del Castillo <acs(a)bitergia.com>
> To: mediawiki-api(a)lists.wikimedia.org
> Subject: [Mediawiki-api] Working in a MediaWiki activity analyzer
> Message-ID: <1383846726.14529.135.camel@lenovix>
> Content-Type: text/plain; charset="UTF-8"
>
> Hi guys,
>
> We are working in an analyzer to gather the activity from MediaWiki
> sites using the cool MediaWiki API.
>
> Right now the idea is to get a list of wiki pages using:
>
> "action=query&list=allpages&aplimit=500"
>
> and then for each page get all revisions with:
>
> action=query&prop=revisions&titles=Main%20Page&rvlimit=500
>
> so we have all revisions activity (pretty similar to having all commits
> to source code) for a MediaWiki site.
>
> We are doing it as Open Source in:
>
> https://github.com/MetricsGrimoire/MediaWikiAnalysis
>
> Any comments are welcomed!
>
> After having all the data we plan to use:
>
> https://github.com/VizGrimoire/VizGrimoireR
>
> for data analysis (SQL+R) combination
>
> and for doing the viz:
>
> https://github.com/VizGrimoire/VizGrimoireJS
>
> Kudos to Sumana Harihareswara for pointing me to this list!
>
> Cheers
> --
> |\_____/| Alvaro del Castillo San Félix
> [o] [o] acs(a)bitergia.com - Chief Technical Officer (CTO)
> | V | http://www.bitergia.com
> | | "Bridging the gap between developers and stakeholders"
> -ooo-ooo-
>
>
>
>
> ------------------------------
>
> Message: 2
> Date: Thu, 7 Nov 2013 13:29:56 -0500
> From: "Brad Jorsch (Anomie)" <bjorsch(a)wikimedia.org>
> To: "MediaWiki API announcements & discussion"
> <mediawiki-api(a)lists.wikimedia.org>
> Subject: Re: [Mediawiki-api] Working in a MediaWiki activity analyzer
> Message-ID:
> <CAEepRSswUtAT7UXP7A8QAv6rXD34F6A0gG2O2d_OmAWeVbJN1Q(a)mail.gmail.com>
> Content-Type: text/plain; charset=ISO-8859-1
>
> If you're using WMF sites, you'll want to download database dumps
> instead. See http://dumps.wikimedia.org/.
No, we are not using WMF sites. It is for getting metrics for website
using MediaWiki as its engine.
>
> On Thu, Nov 7, 2013 at 12:52 PM, Alvaro del Castillo <acs(a)bitergia.com> wrote:
> > Hi guys,
> >
> > We are working in an analyzer to gather the activity from MediaWiki
> > sites using the cool MediaWiki API.
> >
> > Right now the idea is to get a list of wiki pages using:
> >
> > "action=query&list=allpages&aplimit=500"
> >
> > and then for each page get all revisions with:
> >
> > action=query&prop=revisions&titles=Main%20Page&rvlimit=500
> >
> > so we have all revisions activity (pretty similar to having all commits
> > to source code) for a MediaWiki site.
> >
> > We are doing it as Open Source in:
> >
> > https://github.com/MetricsGrimoire/MediaWikiAnalysis
> >
> > Any comments are welcomed!
> >
> > After having all the data we plan to use:
> >
> > https://github.com/VizGrimoire/VizGrimoireR
> >
> > for data analysis (SQL+R) combination
> >
> > and for doing the viz:
> >
> > https://github.com/VizGrimoire/VizGrimoireJS
> >
> > Kudos to Sumana Harihareswara for pointing me to this list!
> >
> > Cheers
> > --
> > |\_____/| Alvaro del Castillo San Félix
> > [o] [o] acs(a)bitergia.com - Chief Technical Officer (CTO)
> > | V | http://www.bitergia.com
> > | | "Bridging the gap between developers and stakeholders"
> > -ooo-ooo-
> >
> >
> > _______________________________________________
> > Mediawiki-api mailing list
> > Mediawiki-api(a)lists.wikimedia.org
> > https://lists.wikimedia.org/mailman/listinfo/mediawiki-api
>
>
>
> --
> Brad Jorsch (Anomie)
> Software Engineer
> Wikimedia Foundation
>
>
>
> ------------------------------
>
> Message: 3
> Date: Thu, 07 Nov 2013 22:40:14 +0100
> From: Platonides <platonides(a)gmail.com>
> To: MediaWiki API announcements & discussion
> <mediawiki-api(a)lists.wikimedia.org>
> Subject: Re: [Mediawiki-api] Working in a MediaWiki activity analyzer
> Message-ID: <527C08BE.5010607(a)gmail.com>
> Content-Type: text/plain; charset=UTF-8; format=flowed
>
> On 07/11/13 19:29, Brad Jorsch (Anomie) wrote:
> > If you're using WMF sites, you'll want to download database dumps
> > instead. See http://dumps.wikimedia.org/.
>
> Indeed. Moreover, for your usecase stub dumps may be enough for you.
>
>
>
> ------------------------------
>
> Message: 4
> Date: Thu, 7 Nov 2013 23:02:28 -0500
> From: Nathan Larson <nathanlarson3141(a)gmail.com>
> To: "MediaWiki API announcements & discussion"
> <mediawiki-api(a)lists.wikimedia.org>
> Subject: Re: [Mediawiki-api] Working in a MediaWiki activity analyzer
> Message-ID:
> <CAF-JeUwx-VA1Hs3wcNjo-GP7JsVECrdki-Bs5o91uVZYT+D1xQ(a)mail.gmail.com>
> Content-Type: text/plain; charset="iso-8859-1"
>
> On Thu, Nov 7, 2013 at 12:52 PM, Alvaro del Castillo <acs(a)bitergia.com>wrote:
>
> > Hi guys,
> >
> > We are working in an analyzer to gather the activity from MediaWiki
> > sites using the cool MediaWiki API.
> >
> >
> Sounds like an interesting project. I am working on something similar that
> will start with the most recent dump and do API queries to stay up to date.
> Perhaps we can collaborate? I created
> https://www.mediawiki.org/wiki/MediaWikiAnalysis as a starting point for
> further on-wiki discussion, if you're interested. Thanks.
We have no dumps at all so we must use the API. I have finished a first
version of the tool and you can see it in action in this sample:
http://bitergia.com/projects/redhat-rdo-dashboard/browser/mediawiki.html
It is showing the evolution in time of reviews (better editions?) and
authors (better editors?). Also the panel is showing some global data
about the WikiMedia site and the Top Authors (Editors).
Our idea is to continue with the development in order to viz the
MediaWiki website and its community evolution.
I am taking a look to:
https://www.mediawiki.org/wiki/MediaWikiAnalysis
Cool! I plan to continue working and I'll try to share the work in this
wiki page.
Next goals: show pages evolution, continue with a new query when the
limit is reached and do the work incremental so we can minimize the
effort for updating the info.
Guys, I plan to create such a viz also for mediawiki.org under the
project:
http://korma.wmflabs.org/browser/
Cheers!
--
|\_____/| Alvaro del Castillo San Félix
[o] [o] acs(a)bitergia.com - Chief Technical Officer (CTO)
| V | http://www.bitergia.com
| | "Bridging the gap between developers and stakeholders"
-ooo-ooo-
Hi guys,
We are working in an analyzer to gather the activity from MediaWiki
sites using the cool MediaWiki API.
Right now the idea is to get a list of wiki pages using:
"action=query&list=allpages&aplimit=500"
and then for each page get all revisions with:
action=query&prop=revisions&titles=Main%20Page&rvlimit=500
so we have all revisions activity (pretty similar to having all commits
to source code) for a MediaWiki site.
We are doing it as Open Source in:
https://github.com/MetricsGrimoire/MediaWikiAnalysis
Any comments are welcomed!
After having all the data we plan to use:
https://github.com/VizGrimoire/VizGrimoireR
for data analysis (SQL+R) combination
and for doing the viz:
https://github.com/VizGrimoire/VizGrimoireJS
Kudos to Sumana Harihareswara for pointing me to this list!
Cheers
--
|\_____/| Alvaro del Castillo San Félix
[o] [o] acs(a)bitergia.com - Chief Technical Officer (CTO)
| V | http://www.bitergia.com
| | "Bridging the gap between developers and stakeholders"
-ooo-ooo-