Hi,
I'm currently maintaining wikilint (cf.
<URI:http://toolserver.org/~timl/cgi-bin/wikilint>) that re-
views Wikipedia articles for common problems. At the moment,
it is a powerful, but ugly mess of regular expressions ga-
lore. Fixing bugs is a nightmare.
Ideally, a redesign would parse the source in a tree-like
structure and then work on that. So I went to CPAN and
[[mw:Alternative parsers]] and found out that:
a) there are lots of "release early, release once" "imple-
mentations" that do not anything useful and do not seem
to be in further development, and
b) for many people, "parser" seems to have the meaning
"converter".
So I'll probably have to start another try. As for wikilint
I do not have to be able to parse 100 % of all thinkable wi-
ki markup (if the article cannot be parsed, it probably is
broken anyway), I could go for a rather "lean" approach. For
the tree structure, I would opt for DOM to maximize code re-
usability with wiki markup in a separate namespace. If there
are no relevant fundaments to build on, I would prefer Perl,
ideally enhancing an existing CPAN module like
WWW::Wikipedia::Entry.
Any pointers to things that I overlooked? Thoughts on in-
terfaces & Co.? Volunteers? :-)
Tim
*blows dust off list*
90% documentation of wikitext is not hard. The last 10% has killed
every attempt.
[percentages not exact]
So. I suspect what will happen to get WYSIWYG editing is that
FCKeditor or similar will get a fair way into that 10%, then whatever
remains will be deprecated and bot-converted away from.
Then we translate FCKeditor into EBNF/ANTLR/WTFBOL and the world is a
utopia of third-party parsers galore and everything is rainbows and
unicorns!
The question that springs to my mind is: What is that last 10%? Has
anyone documented the pitfalls that killed the intrepid explorers of
the past?
- d.
Hi,
I am new to the list. I would like to parse wikitext files from
wikipedia and have some doubts:
1) what is the best parser that gets the best rendering for wikipedia
content ? And the fastest
parser ?
2) If I decide to write my parser, how can I handle rendering of
infoboxes/macros ? Maybe the idea
is to parse the wikitext but handle the macros/infoboxes/templates
with a separate library (as
I imagine there are lots of different templates and it might be silly
to try to parse/render each one
as this might be handled well for existing libraries)
thanks and regards
jose
Hi,
Some how i was able to install tiny mce as wysiwyg plugin for media wiki .
But i am facing one issue, I am not able to embed any flash files there.
The images are getting inserted properly, But incase of flash i do not why
the HTML tag < is getting converted to <
Does any body know how to fix this. Please let me know.
Thanks
pravams
Hello,
Apologies if this has been asked before. I've been hunting but not found anything yet on mediawiki, Wikipedia, or the discussion list archive.
Is it possible to parse or call the data from template A on page X, and spit parts of it back out using another template (say, template B) on page Y, and in a different namespace? For the sake of argument, assume there is only one template A iteration on page X.
For example,
Template A
{{bookdata
| image = no
| maplink = Image:Title_p82.gif
}}
Template B
{{mapdata
| maplink = [call from bookdata]
}}
It seems like there should be a way to do this, so if it means reading manuals on something I'd be happy to. At this point I'm just not sure which direction I should be going to read up on this.
Thanks for any advice.
Craig
I am a new member of the list. But it seems that this list was so quiet for
months.
any new progress of any plan of parsers?
I have tried Steve Bennett's antlr grammar, but was failed to generate
parser for it.
And I noticed also that no new progress for this antlr grammar.
Dose the MediaWiki tech team have any roadmaps for the parser?
Who dose push the things progress more?
Thanks for your reply.
Regards,
Mingli
hello,
We am doing knowledge Harvesting project,
where we are parsing wikipedia xml pages and
there it contain wiki markup .
so we are suppose to parse that wiki too,
we got stuck here, how to parse wiki markup.
do i need to write wiki parser of its already available.
do reply if any one know about this query.
Thanking you,
Hitesh Thakkar.
I saw we get a lot of troubles with wikitext. I wonder this idea is
possible: we will concentrate to build a wysiwyg editor which produces
xml-like content so that:
- people don't have to know any wiki language
- developer is able to forget wikitext and its complexity
To prove that it is possible with a friend of mine, I had write a xml-like
language for wiki (in Vietnamese). I can ask him to translate it into
English (because I'm not really good in English).