Thank you! I don't protect non-API wikis at all, they have a room in the museum, I only want to be sure that we can do everything with API.

2013/4/22 Alex S.H. Lin <alexsh.tw@gmail.com>
(sorry for poor english grammar )

 I think you want to distunguish the link page is exist or not in wikitext.
You don't need to analyze HTML tags now, just use
  • Page().linkedPages() to get all links Page() object in the page you want.
  • getall(site, pagelist) to load all links detail.
  • pagelist[x].get() to get wikitexts.
if page is not exist, you will get the exception NoPage.

There's use two API requests, and easy to solve it.


2013/4/22 Bináris <wikiposta@gmail.com>
Can we solve everything through API? I have a script that I wrote a few years ago, and I had to analyze HTML code to distunguish "red" and "blue" links on a page because there was no API function for that.

--
Bináris

_______________________________________________
Pywikipedia-l mailing list
Pywikipedia-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/pywikipedia-l




--
I always keep my spirit minority.

_______________________________________________
Pywikipedia-l mailing list
Pywikipedia-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/pywikipedia-l




--
Bináris