Hello,
I am writing a Java program to extract the abstract of the wikipedia page
given the title of the wikipedia page. I have done some research and found
out that the abstract with be in rvsection=0
So for example if I want the abstract of 'Eiffel Tower" wiki page then I am
querying using the api in the following way.
http://en.wikipedia.org/w/api.php?action=query&prop=revisions&titles=Eiffel…
and parse the XML data which we get and take the wikitext in the tag <rev
xml:space="preserve"> which represents the abstract of the wikipedia page.
But this wiki text also contains the infobox data which I do not need. I
would like to know if there is anyway in which I can remove the infobox data
and get only the wikitext related to the page's abstract Or if there is any
alternative method by which I can get the abstract of the page directly.
Looking forward to your help.
Thanks in Advance
Aditya Uppu
When list=allusers is used with auactiveusers, a property 'recenteditcount'
is returned in the result. In bug 67301[1] it was pointed out that this
property is including various other logged actions, and so should really be
named something like "recentactions".
Gerrit change 130093,[2] merged today, adds the "recentactions" result
property. "recenteditcount" is also returned for backwards compatability,
but will be removed at some point during the MediaWiki 1.25 development
cycle.
Any clients using this property should be updated to use the new property
name. The new property will be available on WMF wikis with 1.24wmf12, see
https://www.mediawiki.org/wiki/MediaWiki_1.24/Roadmap for the schedule.
[1]: https://bugzilla.wikimedia.org/show_bug.cgi?id=67301
[2]: https://gerrit.wikimedia.org/r/#/c/130093/
--
Brad Jorsch (Anomie)
Software Engineer
Wikimedia Foundation
_______________________________________________
Mediawiki-api-announce mailing list
Mediawiki-api-announce(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/mediawiki-api-announce
I've just deployed the ApiFeatureUsage extension to Beta Labs for testing.
If an API client (e.g. a bot or a user script) sets a unique User-Agent or
Api-User-Agent header, a summary of deprecated API features hit by that
client may be fetched from Special:ApiFeatureUsage[1] or from the API
itself.[2]
Please try it out, and report any issues in Phabricator using the
MediaWiki-extension-ApiFeatureUsage project. If things look good after
testing I'll be looking at getting it deployed to production in the next
month or so.
[1]: http://en.wikipedia.beta.wmflabs.org/wiki/Special:ApiFeatureUsage
[2]:
http://en.wikipedia.beta.wmflabs.org/w/api.php?action=help&modules=query+fe…
--
Brad Jorsch (Anomie)
Software Engineer
Wikimedia Foundation
_______________________________________________
Mediawiki-api-announce mailing list
Mediawiki-api-announce(a)lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/mediawiki-api-announce
Hi
can you please tell me if there is way to get bread crumb navigation..
For example i need this for istanbul page:
Europe > Turkey > Marmara (region) > Istanbul Province > Istanbul
I already have all page HTMLs through parse action in API but they did not
have this part.I read documentation roughly could not see any option
best
Oz
Hi,
Am I correct in assuming that if I wanted to track if a particular Wikipedia page has changed from the last time I looked at it would be the id within revision element?
<revision>
<id>631144794</id>
<parentid>381202555</parentid>
<timestamp>2014-10-26T04:50:23Z</timestamp>
Thanks
Hi,
When I submit the API http://en.wikipedia.org/w/api.php?format=json <http://en.wikipedia.org/w/api.php?format=json&action=query&prop=revisions&r…> &action=query&prop=revisions&rvprop=content|ids&pageids=7412236 I get the below with one of the properties as a *. Is it possible to get with a name like say content instead of the *. I am not able to de-serialize the response.
revisions
------{}
*={{about|the person|the biography|Steve Jobs (book)|the 2013 biographical film|Jobs (film)}}……..
contentformat=text/x-wiki
contentmodel=wikitext
parentid=638035784
revid=638103683
Thanks
Hello everyone,
how can I extract just clean plain text from a Wikipedia article?
Without wiki-stuff, without html, without pictures, without json.
Just clean text.
I can't seem to find this exact solution.
Best regards
Mikoto
Hi,
I had taken a back up of the Wikipedia data and have a range / set of
pageids.
How do I find out the range of pageids that are currently in use in
wikipedia?
Do pageids always remain constant?
Can a certain pageid cease to exist?
Please let me know
thanks