On 11/22/05, S. Woodside
<sbwoodside(a)yahoo.com> wrote:
Conjecture 1. That the distance between any two
wikipedia pages,
randomly chosen, as measured by wikilinks, is on average 6.
Conjecture 2. That wikipedia is sufficiently formal and complete
that
you could build a useful general purpose AI knowledge base using it.
Conjecture 3. That wikipedia has low information entropy.
Conjecture 4. That the development of a wikipedia article over time
occurs in a manner consistent to the biological evolution of a
species.
Conjecture 5. That the relationship between the amount of
material in
wikipedia and the number of article views is exponential.
Conjecture 6. That wikipedia is, on average, factually accurate.
Ohoh ! can I produce one?
Conjecture 7. The amount of wanky conjectures produced by bloggers is
bound only by the number of sites that will allow them to spam their
URLs.
Do I get a prize? ;)
On a more constructive note, If you're interested in AI taught using
wikipedia... you should take a look at the state of the art in
English
parsers (such as [[Link Grammar]]) and how they fair on Wikipedia.
Impressive that it works at all, but I don't think we need to worry
about a hostile takeover of the planet by a self-aware encyclopedia
any time soon.
_______________________________________________
Wikitech-l mailing list
Wikitech-l(a)wikimedia.org
http://mail.wikipedia.org/mailman/listinfo/wikitech-l
--
"Take away their language, destroy their souls." -- Joseph Stalin
_______________________________________________
Wikitech-l mailing list
Wikitech-l(a)wikimedia.org