I am writing a Java program to extract the abstract of the wikipedia page
given the title of the wikipedia page. I have done some research and found
out that the abstract with be in rvsection=0
So for example if I want the abstract of 'Eiffel Tower" wiki page then I am
querying using the api in the following way.
and parse the XML data which we get and take the wikitext in the tag <rev
xml:space="preserve"> which represents the abstract of the wikipedia page.
But this wiki text also contains the infobox data which I do not need. I
would like to know if there is anyway in which I can remove the infobox data
and get only the wikitext related to the page's abstract Or if there is any
alternative method by which I can get the abstract of the page directly.
Looking forward to your help.
Thanks in Advance
When list=allusers is used with auactiveusers, a property 'recenteditcount'
is returned in the result. In bug 67301 it was pointed out that this
property is including various other logged actions, and so should really be
named something like "recentactions".
Gerrit change 130093, merged today, adds the "recentactions" result
property. "recenteditcount" is also returned for backwards compatability,
but will be removed at some point during the MediaWiki 1.25 development
Any clients using this property should be updated to use the new property
name. The new property will be available on WMF wikis with 1.24wmf12, see
https://www.mediawiki.org/wiki/MediaWiki_1.24/Roadmap for the schedule.
Brad Jorsch (Anomie)
Mediawiki-api-announce mailing list
can you please tell me if there is way to get bread crumb navigation..
For example i need this for istanbul page:
Europe > Turkey > Marmara (region) > Istanbul Province > Istanbul
I already have all page HTMLs through parse action in API but they did not
have this part.I read documentation roughly could not see any option
Am I correct in assuming that if I wanted to track if a particular Wikipedia page has changed from the last time I looked at it would be the id within revision element?
how can I extract just clean plain text from a Wikipedia article?
Without wiki-stuff, without html, without pictures, without json.
Just clean text.
I can't seem to find this exact solution.
I had taken a back up of the Wikipedia data and have a range / set of
How do I find out the range of pageids that are currently in use in
Do pageids always remain constant?
Can a certain pageid cease to exist?
Please let me know