Abdel Samad, Rawia, 21/01/2015 09:47:
I work for a consulting firm called Strategy&. We
have been engaged by
Facebook on behalf of
Internet.org to conduct a study on assessing the
state of connectivity globally. One key area of focus is the
availability of relevant online content. We are using a the availability
of encyclopedic knowledge in one’s primary language as a proxy for
relevant content. We define this as 100K+ Wikipedia articles in one’s
primary language.
Hello Rawia,
is there any update on this project? Have you contacted Google about
similar "content availability" and "content ingestion" activities they
conducted in the past, also related to machine translation
(
https://meta.wikimedia.org/wiki/Machine_translation )?
We are very interested in this sort of initiatives (see also
https://lists.wikimedia.org/pipermail/wiki-research-l/2015-March/004297.html
), but experience taught us that looking at the wrong things can have
terrible consequences.
Nemo
We have a few questions related to this analysis
prior
to publishing it:
·We are currently using the article count by language based on
Wikimedia’s foundation public link: Source:
http://meta.wikimedia.org/wiki/List_of_Wikipedias. Is this a reliable
source for article count – does it include stubs?
·Is it possible to get historic data for article count. It would be
great to monitor the evolution of the metric we have defined over time?
·What are the biggest drivers you’ve seen for step change in the number
of articles (e.g., number of active admins, machine translation, etc.)
·We had to map Wikipedia language codes to ISO 639-3 language codes in
Ethnologue (source we are using for primary language data). The 2
language code for a wikipedia language in the “List of Wikipedias”
sometimes matches but not always the ISO 639-1 code. Is there an easy
way to do the mapping?
Many Thanks,
Rawia