I am writing a Java program to extract the abstract of the wikipedia page
given the title of the wikipedia page. I have done some research and found
out that the abstract with be in rvsection=0
So for example if I want the abstract of 'Eiffel Tower" wiki page then I am
querying using the api in the following way.
and parse the XML data which we get and take the wikitext in the tag <rev
xml:space="preserve"> which represents the abstract of the wikipedia page.
But this wiki text also contains the infobox data which I do not need. I
would like to know if there is anyway in which I can remove the infobox data
and get only the wikitext related to the page's abstract Or if there is any
alternative method by which I can get the abstract of the page directly.
Looking forward to your help.
Thanks in Advance
Hi, in response to bug 54607 , we've changed the semantics of the
mobileformat parameter to action=parse
== Summary ==
Previously, it used to accept strings 'html' or 'wml', later just
'html' and modify the structure of output (see below). This was problematic
because you needed to retrieve the HTML from output in different ways,
depending on whether mobileformat is specified or not. Now,
mobileformat is a boolean parameter, that is if there's a 'mobileformat'
parameter in request, it will be treated as "the output should be
mobile-friendly", regardless of value. And the output structure will
be the same. For compatibility with older callers,
mobileformat=(html|wml) will be special-cased to return the older
structure at least for 6 month from now. These changes will start
being rolled out to the WMF sites starting from tomorrow, Tuesday
October 24th and this process will be complete by October 31st.
== Examples ==
=== Non-mobile parse ===
<parse title="..." displaytitle="...">
=== Parse that outputs mobile HTML, old style ===
<parse title="..." text="foo" displaytitle="...">
=== Parse that outputs mobile HTML, new style ===
Same as for non-mobile parses.
== FAQ ==
Q: I didn't use mobileformat before, does anything change for me?
Q: I use mobileformat=html, will my bot/tool be broken now?
A: No, you will have 6 months to switch to new style.
Q: I'm only planning to use mobileformat, what should I do?
A: Just use the new style.
Q: How did this format discrepancy appear in the first place?
A: To err is human.
Max Semenik ([[User:MaxSem]])
I am working on an application which requires the total wikipedia page counts that contain specific word or a pair of words. I have tried to to use advanced search to get these page counts and retrieve the the HTML page to extract that. Please see the attached snapshot is information pointed by the red arrow the actual total pages containing the name David Beckham ?. Any advice is highly appreciated.
Muhidin A. Mohamed,
School of Electrical, Electronic and Computer Engineering,
University of Birmingham,
is there any minimum number of votes for fixing a bug?
I am wondering which is - if there is - an average time for having a normal
bug fixed, and see if I may help.
How would if possible to help without installing mediawiki software?
(I am a bit skilled in python, not other languages)
My name is Kenrick, I'm an Indonesian Wikipedia administrator trying to
experiment things with MediaWiki API.
I am currently doing a project to list down the new articles using
MediaWiki API with these parameters:
action = query
list = recentchanges
rctype = new
rcshow = !redirect
But then this method also includes those articles which current revision is
a redirect page.
Is there any way to perfectly list down new articles like Special:Newpages
There are 2 move operations in recent change:
"move/move" and "move/move_redir".
And in "move/move" changes, there may be tag suppressedredirect="" which
means move the page without creating the redirect (
Here comes my question:
does "move/move" change without the suppressedredirect="" equal to
"move/move_redir", i.e. it's a move redirect operation?
I am writing a mobile client that can show wikipedia content. My
approach is to download the raw media-wiki markup instead of the
generated html. This allows me more control and I avoid using a html
The approach is quite successful except for when I encounter images.
In the markup I can see something like;
I use the API to fetch some metadata;
The piece of the puzzle I am still missing is how to find out the
actual download URL for any given image.
I've seen images start with;
But I don't really understand how to decide what url to prefix to my
Anyone can shed some light on this?
we've been developing a product to access the knowledge in the wikipedia in
other ways... and of course we make use of the mediawiki api.
We are using geodata extension to construct a nearby similar to the wiki,
but I'd like the possibility to retrieve articles beyond the 10000 limit
radius: this is the case when I drag the map.
As example, you can zoom out your map at "nations" scale, drag it around
the globe, and you'll see other articles which will be pinned on the map.
I've seen something similar within google maps and other mobile app, so I
read again the doc http://www.mediawiki.org/wiki/Extension:GeoData
I figured out this feature should be available from geohack, which is
already included, by the use of parameter "scale".
However, I didn't understand how to use it; also geohack is not mentioned
in the main doc
Should I pipe it with a generator? Any example for doing it ?
Once again, thank you so much for your help to all devs!
As part of the visualisation tool I'm building I'm fetching the parsed
revisions of an article. When the article is of a considerable size , eg
latest revisions of Barack Obama it takes 10+ seconds. As the tool is
interactive and it shows the edits made to an article as an animation the
time taken by the server does not bode well. (The requests are only read )
I'm currently not making parallel requests. What would a reasonable degree
of parallel requests. Are there other ways to get around this latency issue
https://meta.wikimedia.org/wiki/Grants:IEG/Replay_Edits talks about the
tool and the project.