Hello,
I am writing a Java program to extract the abstract of the wikipedia page
given the title of the wikipedia page. I have done some research and found
out that the abstract with be in rvsection=0
So for example if I want the abstract of 'Eiffel Tower" wiki page then I am
querying using the api in the following way.
http://en.wikipedia.org/w/api.php?action=query&prop=revisions&titles=Eiffel…
and parse the XML data which we get and take the wikitext in the tag <rev
xml:space="preserve"> which represents the abstract of the wikipedia page.
But this wiki text also contains the infobox data which I do not need. I
would like to know if there is anyway in which I can remove the infobox data
and get only the wikitext related to the page's abstract Or if there is any
alternative method by which I can get the abstract of the page directly.
Looking forward to your help.
Thanks in Advance
Aditya Uppu
(sending to main tech lists, crossposted to Tech Blog[0], feel free to forward
anywhere else you'd like)
Greetings MediaWiki hackers!
I am pleased to announce the upcoming MediaWiki Hack-A-Ton in Washington, DC.
As you are all aware, every year in April our good friends at
Wikimedia Deutschland
host the annual "MediaWiki Developers Meetup" in Berlin. At that
event, the program
is focused on demonstrations, workshops and small group discussions. To
complement this, we're planning the DC meetup to be focused solely on hacking,
bugfixing and getting down and dirty with the code.
We're scheduling this for October 22nd-24th in Washington, DC. Some of
the details
haven't been ironed out yet, but will be announced over the coming
days as it is.
So clear your calendars, and keep your eyes on MediaWiki.org[1] and the mailing
lists for more information.
Some travel assistance may be available for those coming a long way. I've also
been told there will be swag of some sort for attendees :)
-Chad
[0] http://techblog.wikimedia.org/2010/09/hack-a-ton-dc/
[1] http://www.mediawiki.org/wiki/Hack-A-Ton_DC
Hi,
I'm looking for some API that could help me get the "tables" alone from
a page if it exists.
I've looked at : http://en.wikipedia.org/w/api.php
But nothing seems to be working on tables alone.
I'm new here, so if I am not looking at the right place, please point me
to appropriate links.
Thanks
Chandra Sekhar
This afternoon (UTC), the API got overloaded for a while; you may have
noticed slow responses or HTTP 504s. Tim tried to solve this by
disabling action=parse, which worked surprisingly well. This means
action=parse is disabled for now and will remain disabled until I
figure out what was going on at the time of the outage.
In the meantime, the API will act as though the action=parse module
doesn't exist, so you'll get errors along the lines of "unknown value
for parameter action: 'parse'".
I'll re-enable action=parse once I'm relatively confident it won't
cause the entire API to melt down again.
Roan Kattouw (Catrope)
Hi all,
You know what happened with API today?
Did it change some functions?
action=parse is not being supported anymore. So what is new API to get article content?
Best regards,
Sergio Gaona Bautista
________________________________
"El contenido de este documento y/o sus anexos son para uso exclusivo de su destinatario intencional y puede contener Informaci?n legalmente protegida por ser privilegiada o confidencial. Si usted no es el destinatario intencional de este documento por favor Inf?rmenos de inmediato y elimine el documento y sus anexos. Igualmente cualquier retenci?n, revisi?n no autorizada, distribuci?n, divulgaci?n, reenv?o, copia, impresi?n, reproducci?n, o uso indebido de este documento y/o sus anexos, est? estrictamente prohibida y sancionada legalmente. Agradecemos su atenci?n. UNE EPM Telecomunicaciones S.A."
--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
"The contents of this document and / or its attachments are for exclusive use of the intended recipient and may contain privileged or confidential information. If you are not the intended recipient of this document please inform us immediately and delete the document and its attachments. Likewise, any retention, unauthorized review, distribution, disclosure, forwarding, copying, printing, reproduction, or misuse of this document and / or its attachments is strictly prohibited and punishable by law. Thank you for your attention. UNE EPM Telecomunicaciones SA"
Hi,
Has there been any change in the API very recently (login request) ?
My tool, WikiCleaner [1], stopped working in the last hours : when trying to
login, a status 504 (Gateway timeout) is now returned.
Thanks for any help
Nico
[1] http://en.wikipedia.org/wiki/User:NicoV/Wikipedia_Cleaner/Documentation
On Thu, Sep 16, 2010 at 1:40 PM, Aude <aude.wiki(a)gmail.com> wrote:
> Very awesome that no travel is required! I'm not a hardcore mediawiki
> contributor but would like to be more involved, can work on mapping
> and data capabilites, or on aspects of core code (If you want help
> with something)
>
Of course all developers are welcome! We're going to be tracking
bugs for that weekend using a new keyword in Bugzilla called
"bugsmash."
(That's also a general invitation for people to start targeting some
items for the Hack-A-Ton)
> What help do you need from local DC wikipedians? Although we are not
> yet a chapter, we could perhaps help with getting a venue (though
> danese probably can handle it too) or other ways of helping organize.
>
I should've reached out to you guys a few weeks ago ;-) The venue
is almost settled--hopefully will have an information to share about
that this evening. I might need a hand the day before (I will be driving
up the 21st) to help me get things organized and ready for everyone's
arrival on the 22nd.
-Chad
Hi,
I'm looking for help using the MediaWiki API to get some page view counts.
I'm able to get page view counts on some sites, such as the wikia muppet site:
http://muppet.wikia.com//api.php?action=query&prop=info&titles=Beaker&inpro…
But that's not working on en.wikipedia.org:
http://en.wikipedia.org/w/api.php?action=query&titles=Health&prop=info&inpr…
That gives an error message that says: "Unrecognized value for
parameter 'inprop': views".
A few questions:
+ Does that error message happen because Wikipedia is configured with
$wgDisableCounters=true, or am I doing something wrong?
+ Is there some other API that I should look at for getting page counts?
+ Should I ping Henrik about maybe screen-scraping from his
stats.grok.se page traffic stats pages?
Thanks, I appreciate any help you can give me. I'm doing this work as
part of the Health Speaks effort to translate health information into
more languages, and we want to get a sense of how useful the newly
translated pages turn out to be.
Brian Skinner
Health Speaks, Google.orghttp://sitescontent.google.com/healthspeaks/