I am considering the task of converting the templates from the gene articles in Wikipedia (http://en.wikipedia.org/wiki/Portal:Gene_Wiki) to use/create wikidata assertions. This involves an extensive update of the template structure as well as the code for the bot that keeps them in sync with external public databases. (https://bitbucket.org/sulab/pygenewiki)
More specifically I'm thinking about working with a Google Summer of Code student on this project.
Given a time frame of now through August, would it make sense for us to pursue this objective directly in the context of wikidata (through the public API). Or would it be better for us to install our own version of the wikibase software (kept in sync with code updates) and develop the new gene wiki bot code locally with the aim of switching to the public API later? Or is it too early to consider this project?
I want to get involved and support wikidata with this important data, but I'm hesitant to ramp up development (especially with a student) in a moving target situation.
Any thoughts? thanks! -Ben
Hi Benjamin,
there is a similar task. I am not sure this fits here. Hungarian Wikipedia lists the redlist status of species of plants and animals (threatened, not threatened etc.). There is a request to update this by bot from http://www.iucnredlist.org/. Also biological property, also listed in templates and also originates from an external database. I told them to wait for Wikidata phase 2, which is now alive, but what next? Do you think these tasks are related?
2013/3/1 Benjamin Good ben.mcgee.good@gmail.com
I am considering the task of converting the templates from the gene articles in Wikipedia (http://en.wikipedia.org/wiki/Portal:Gene_Wiki) to use/create wikidata assertions. This involves an extensive update of the template structure as well as the code for the bot that keeps them in sync with external public databases. (https://bitbucket.org/sulab/pygenewiki)
More specifically I'm thinking about working with a Google Summer of Code student on this project.
Given a time frame of now through August, would it make sense for us to pursue this objective directly in the context of wikidata (through the public API). Or would it be better for us to install our own version of the wikibase software (kept in sync with code updates) and develop the new gene wiki bot code locally with the aim of switching to the public API later? Or is it too early to consider this project?
I want to get involved and support wikidata with this important data, but I'm hesitant to ramp up development (especially with a student) in a moving target situation.
Any thoughts? thanks! -Ben
Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
On Fri, Mar 1, 2013 at 12:12 AM, Bináris wikiposta@gmail.com wrote:
Hi Benjamin,
there is a similar task. I am not sure this fits here. Hungarian Wikipedia lists the redlist status of species of plants and animals (threatened, not threatened etc.). There is a request to update this by bot from http://www.iucnredlist.org/.
My bot imported a bunch of these over the past two days from English Wikipedia categories. You can see some of the logs at [1] and [2]. I don't think it is fully complete, but it is a huge start. Since my bot can import from any Wikipedia project if Hungarian Wikipedia has it sorted into categories as well, my bot can complement the enwp data with it.
Also biological property, also listed in templates and also originates from an external database. I told them to wait for Wikidata phase 2, which is now alive, but what next? Do you think these tasks are related?
I am planning on importing taxonomy data from enwp (and possibly other projects) hopefully next week. There's a quick page at [3] that an enwp user who is familiar with the taxonomy templates put together for me. If other languages also have similar templates it would be good to add to that list as well. I am also going to look into supplementing with information from Wikispecies if possible.
2013/3/1 Benjamin Good ben.mcgee.good@gmail.com
I am considering the task of converting the templates from the gene articles in Wikipedia (http://en.wikipedia.org/wiki/Portal:Gene_Wiki) to use/create wikidata assertions. This involves an extensive update of the template structure as well as the code for the bot that keeps them in sync with external public databases. (https://bitbucket.org/sulab/pygenewiki)
I am not very familiar with enwp's coverage of genes. Does this require updates on Wikidata side or Wikipedia side?
More specifically I'm thinking about working with a Google Summer of Code
student on this project.
Given a time frame of now through August, would it make sense for us to pursue this objective directly in the context of wikidata (through the public API). Or would it be better for us to install our own version of the wikibase software (kept in sync with code updates) and develop the new gene wiki bot code locally with the aim of switching to the public API later? Or is it too early to consider this project?
I would think that as soon as the data can go on to the main Wikidata database, the better.
I want to get involved and support wikidata with this important data, but I'm hesitant to ramp up development (especially with a student) in a moving target situation.
Any thoughts? thanks! -Ben
--
Bináris
[1] https://www.wikidata.org/wiki/User:Legobot/properties.js/Archive/2013/02/27 [2] https://www.wikidata.org/wiki/User:Legobot/properties.js/Archive/2013/02/28 [3] https://www.wikidata.org/wiki/User:Legobot/taxon
--Legoktm
Lego,
nice work, but I am not sure whether we speak abpout the same thing. What I would like is to update property values in Wikidata from iucnredlist.orgregularly, then insert into Wikipedias the always actual data from Wikidata.
Oh I see what you mean. You want to go database-->wikidata-->wikipedia, whereas I am going database-->wikipedia-->wikidata. I think it is a good idea :) -- Legoktm
On Fri, Mar 1, 2013 at 1:22 AM, Bináris wikiposta@gmail.com wrote:
Lego,
nice work, but I am not sure whether we speak abpout the same thing. What I would like is to update property values in Wikidata from iucnredlist.orgregularly, then insert into Wikipedias the always actual data from Wikidata.
-- Bináris _______________________________________________ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
There will soon be a mechanism where Wikipedia can display data from Wikidata directly, as it currently does with the language links. No need to bot-edit Wikipedia.
On Fri, Mar 1, 2013 at 7:22 AM, Bináris wikiposta@gmail.com wrote:
Lego,
nice work, but I am not sure whether we speak abpout the same thing. What I would like is to update property values in Wikidata from iucnredlist.orgregularly, then insert into Wikipedias the always actual data from Wikidata.
-- Bináris _______________________________________________ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
2013/3/1 Magnus Manske magnusmanske@googlemail.com
There will soon be a mechanism where Wikipedia can display data from Wikidata directly, as it currently does with the language links. No need to bot-edit Wikipedia.
Yes, yes, that's what I speak about. :-) As phase 2 is working now in Wikidata repo, this is the good time to speak about this.
Sorry to be the contrarian, but I'm not sure we should be talking about pulling data from Wikidata into Wikipedia until the devs announce that they are close to deploying it. It makes no sense to build an infrastructure now if the assumptions about functionality and API that you're basing the infrastructure on aren't officially locked in yet.
Sven
On Fri, Mar 1, 2013 at 5:12 AM, Bináris wikiposta@gmail.com wrote:
2013/3/1 Magnus Manske magnusmanske@googlemail.com
There will soon be a mechanism where Wikipedia can display data from Wikidata directly, as it currently does with the language links. No need to bot-edit Wikipedia.
Yes, yes, that's what I speak about. :-) As phase 2 is working now in Wikidata repo, this is the good time to speak about this.
-- Bináris _______________________________________________ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
Of course, I know what I speak about. Please let me think a few weeks in advance. Yes, we definitely have to speak about the purpose of the project. Tha basic question is of course how to transfer data from outer sources to Wikidata properties, this is the only important point in the whole thread.
2013/3/1 Sven Manguard svenmanguard@gmail.com
Sorry to be the contrarian, but I'm not sure we should be talking about pulling data from Wikidata into Wikipedia until the devs announce that they are close to deploying it. It makes no sense to build an infrastructure now if the assumptions about functionality and API that you're basing the infrastructure on aren't officially locked in yet.
Sven
On Fri, Mar 1, 2013 at 5:12 AM, Bináris wikiposta@gmail.com wrote:
2013/3/1 Magnus Manske magnusmanske@googlemail.com
There will soon be a mechanism where Wikipedia can display data from Wikidata directly, as it currently does with the language links. No need to bot-edit Wikipedia.
Yes, yes, that's what I speak about. :-) As phase 2 is working now in Wikidata repo, this is the good time to speak about this.
-- Bináris _______________________________________________ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
Inclusion syntax: http://meta.wikimedia.org/wiki/Wikidata/Notes/Inclusion_syntax
There will be a new code rollout on Wikidata March 6, which should bring support for more data types. Not sure when the Wikipedia client support will be installed, but can't be too far off.
On Fri, Mar 1, 2013 at 3:42 PM, Sven Manguard svenmanguard@gmail.comwrote:
Sorry to be the contrarian, but I'm not sure we should be talking about pulling data from Wikidata into Wikipedia until the devs announce that they are close to deploying it. It makes no sense to build an infrastructure now if the assumptions about functionality and API that you're basing the infrastructure on aren't officially locked in yet.
Sven
On Fri, Mar 1, 2013 at 5:12 AM, Bináris wikiposta@gmail.com wrote:
2013/3/1 Magnus Manske magnusmanske@googlemail.com
There will soon be a mechanism where Wikipedia can display data from Wikidata directly, as it currently does with the language links. No need to bot-edit Wikipedia.
Yes, yes, that's what I speak about. :-) As phase 2 is working now in Wikidata repo, this is the good time to speak about this.
-- Bináris _______________________________________________ Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l
Wikidata-l mailing list Wikidata-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-l