Heya folks,
I just wanted to let you know that the next Wikidata office hours will
be on soon. Denny and I will be around on IRC in #wikimedia-wikidata
to answer any question you might have and discuss. Logs will be
published afterwards.
English: July 18 at 22:00 UTC
(http://www.timeanddate.com/worldclock/fixedtime.html?hour=22&min=00&sec=0&d…)
German: July 23 at 16:30 UTC
(http://www.timeanddate.com/worldclock/fixedtime.html?hour=16&min=30&sec=0&d…)
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Community Communications for Wikidata
Wikimedia Deutschland e.V.
Obentrautstr. 72
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA1
Hello All!
I just wanted to quickly introduce DrTrigonBot to you all in case
you do not know it already... ;)
I think DrTrigonBot could easily be used to add a wide variety of
data from several pages to the database, as explained in [1] also.
[1]
http://meta.wikimedia.org/w/index.php?title=Wikidata/Data_collaborators&dif…
The bot is running since quite some time with success and updates
templates in dewiki, enwiki, nlwiki and others.
So I think it would be a good thing to make it compatible with
wikidata and use it there also. So what would I have to do in
order to achieve this and be of some use to you?
Thanks and Greetings
DrTrigon
-----BEGIN PGP SIGNATURE-----
Version: GnuPG v1.4.12 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org/
iEYEARECAAYFAk//APQACgkQAXWvBxzBrDDx9gCfdCIsGmlBrdfSi6qwnAA5Ef2X
QJQAn3mlS+siO/N/DntPEm/+8ZwdTipI
=CfCH
-----END PGP SIGNATURE-----
Heya folks :)
Here's what we've been up to during the past week.
http://meta.wikimedia.org/wiki/Wikidata/Status_updates/2012_07_13 has
the wiki version.
= Development =
* started implementing UI editing story
* revised upcoming layout update regarding phase II
* review and bugfixing
* start on Undo for items
* investigated all the items currently marked as “needs testing”,
added missing tests and moved them to done
* investigation of bug #38181, no fix yet
* reviewing API changes
See http://meta.wikimedia.org/wiki/Wikidata/Development/Current_sprint
for what we’re working on next.
= Discussions/Press =
* We discussed with WMF ops how Wikidata should be using the parser cache and
squid cache. We decided to bypass the parser cache for now and split the squid
cache by language.
* Started discussions on deployment plan
= Events =
see http://meta.wikimedia.org/wiki/Wikidata/Events
* Wikimania
* upcoming: IRC office hours
* upcoming: Wikidata intro at the New York Times
= other noteworthy stuff =
* We have a logo!
http://lists.wikimedia.org/pipermail/wikidata-l/2012-July/000832.html
Anything to add? Please share! :)
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Community Communications for Wikidata
Wikimedia Deutschland e.V.
Obentrautstr. 72
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
*********************************************************************
We apologize if you received multiple copies of this Call for Papers.
Please feel free to distribute it to those who might be interested.
*********************************************************************
This is an announcement of Semantic MediaWiki
Conference 'SMWCon Fall 2012'.
WHERE
Cologne (Köln), Germany,
WHEN
October 24-26 2012.
CONFERENCE WEBSITE:
http://semantic-mediawiki.org/wiki/SMWCon_Fall_2012
SMWCon is an event for everyone who is interested in collaborative
knowledge creation and semantic wikis. If you use SMW in your projects
or want to know how it can be applied to your needs or want to discuss
the future of the project, SMWCon is a right place to go.
As always, in the first day we will have some introductory tutorials for
those who want to learn more about Semantic MediaWiki. The second and
the third day of the conference will include talks from the developers
and users of SMW.
We also will have some talks about Wikidata project and Semantic
MediaWiki roadmap.
HOW TO PARTICIPATE
We have started to form a program of the conference.
To register or write the abstract of your talk you should edit
the page http://semantic-mediawiki.org/wiki/SMWCon_Fall_2012.
IMPORTANT DATES
September 24 - proposal on a wiki (the earlier the better)
October 24 - the conference itself
See you in Cologne!
Sincerely yours,
Yury Katkov, program chair
Heya folks,
If you are at Wikimania Denny, Jeroen, Daniel and I would love to talk to
you. Come and say hi. If you don't know us come to the Wikidata talks later
today (3:40pm) or the panel tomorrow (10:50am). (You should come to them
anyway ;))
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Community Communications for Wikidata
Wikimedia Deutschland e.V.
Obentrautstr. 72
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg unter
der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das Finanzamt für
Körperschaften I Berlin, Steuernummer 27/681/51985.
>From your last weekly digest:
> we will have a talk at the New York Times, open to the public, on Thursday, July 19th, at 7pm
I did some web searching and looked on
https://meta.wikimedia.org/wiki/Wikidata/Events and didn't see more
information on this -- could you give a few more details so I can
forward this to NYC acquaintances? I don't see it at
https://www.nytimes.com/marketing/timesopen/ -- is it part of TimesOpen?
Thanks!
--
Sumana Harihareswara
Engineering Community Manager
Wikimedia Foundation
Heya folks,
Here is the Wikidata summary of the week before 2012-07-06. The wiki
version is at http://meta.wikimedia.org/wiki/Wikidata/Status_updates/2012_07_01
I've been away for a week of vacation but it looks like the rest of
the team did well without me for a bit :P
= Development =
* Sprint 8 has finished and Sprint 9 has started, a two-week sprint
that will go on throughout Wikimania
* Most of the items linked are now shown with their label instead of
the cryptic ID
* Switching languages is getting better and better
* Search is working
* A doxygen documentation is online at http://wikidata-docs.wikimedia.de/
See http://meta.wikimedia.org/wiki/Wikidata/Development/Current_sprint
for what we’re working on next.
= Discussions/Press =
* The Wikidata logo submission has been closed as planned, and the
voting has started. So far we had more than 500 votes, and the voting
is open until next week. The winner of the contest will be announced
at Wikimania.
= Events =
see http://meta.wikimedia.org/wiki/Wikidata/Events
* Wikipedia Academy
* Wikipedia Meetup NYC
* upcoming: Wikimania with several talks
* On a short notice, we will have a talk at the New York Times, open
to the public, on Thursday, July 19th, at 7pm
Anything to add? Please share! :)
Cheers
Lydia
--
Lydia Pintscher - http://about.me/lydia.pintscher
Community Communications for Wikidata
Wikimedia Deutschland e.V.
Obentrautstr. 72
10963 Berlin
www.wikimedia.de
Wikimedia Deutschland - Gesellschaft zur Förderung Freien Wissens e. V.
Eingetragen im Vereinsregister des Amtsgerichts Berlin-Charlottenburg
unter der Nummer 23855 Nz. Als gemeinnützig anerkannt durch das
Finanzamt für Körperschaften I Berlin, Steuernummer 27/681/51985.
Hello
A few notes on the BBC's use of DBpedia which Dan thought might be of
interest to this list:
Not sure how familiar you are with bbc web stuff so a brief introduction
<skip-me>
We have a large and somewhat sprawling website with 2 main sections: news
article related stuff (including sports) and programme related stuff (tv and
radio). In between these sections are various other domain specific bits
(http://www.bbc.co.uk/music, http://www.bbc.co.uk/food,
http://www.bbc.co.uk/nature etc)
In the main we have actual content / data for news articles and programmes.
Most of the other bits of co.uk are really just different ways of cutting
this content / new aggregations. Because we don't have data for these
domains we borrow from elsewhere (mostly from the LOD cloud). So /music is
based on a backbone of musicbrainz data, /nature is based on numerous data
sources (open and not so open) all tied together with dbpedia identifiers...
In the main we don't really use dbpedia as a data source but rather as a
source of identifiers to triangulate with other data sources
So for example, we have 2 tools for "tagging" programmes with dbpedia
identifiers. Short clips are tagged with one tool using dbpedia information
resource uris, full episodes are tagged with another tool using dbpedia
non-information resource uris (< don't ask)
Taking /music as an example: because it's based on musicbrainz and because
musicbrainz includes wikipedia uris for artists we can easily derive dbpedia
uris (of whatever flavour) and query the programme systems for programmes
tagged with that artist's dbpedia uri
</skip-me>
=== some problems we've found when using dbpedia ===
1. it's not really intended for use for data extraction. The semantics of
extraction depend on the infobox data and this isn't always applied
correctly. So http://en.wikipedia.org/wiki/Fox_News_Channel and
http://en.wikipedia.org/wiki/Fox_News_Channel_controversies share the same
main infobox meaning dbpedia sees them both as tv channels
2. wikipedia tends to conflate many objects into a single item / page. Eg
http://en.wikipedia.org/wiki/Penny_Lane has composer details, duration
details and release information conflating composition with recording with
release
3. the data extraction is a bit flakey in parts. Mainly because it's been
done by a small team and it covers so many different domains.
4. wikipedia doesn't do redirects properly. So
http://en.wikipedia.org/wiki/Spring_watch and
http://en.wikipedia.org/wiki/Autumn_watch are based on the same data /
return the same content and are flagged as a redirect internally but they
don't actually 30x. This is confusing for editorial staff knowing which uri
to "tag" with
5. wikipedia uris are derived from the article title. If the article title
changes the uri changes. Dbpedia uris are derived from wikipedia uris so
they also change when wikipedia uris / titles change. This has caused us no
end of upsets. An example: bbc.co.uk/nature uses wiki|dbpedia uri slugs. So
http://en.wikipedia.org/wiki/Stoat on wikipedia is
http://www.bbc.co.uk/nature/life/Stoat on bbc.co.uk
Apparently people in the UK call stoats stoats and people in the US call
them ermine (or the other way round) which lead to an edit war on wikipedia
which caused the dbpedia uri to flip repeatedly and our aggregations to
break. We've had similar problems with music artists (can't quite remember
the details but seem to remember some arguments about how the "and" should
appear in Florence and the Machine
http://en.wikipedia.org/wiki/Florence_and_the_Machine
6. Titles do change often enough to cause us problems. Particularly names
for people
Nic (cced) has done some work on dbpedia lite (http://dbpedialite.org/)
which aims to provide stable identifiers for dbpedia concepts based on (I
think) wikipedia table row identifiers (which wikimedia do claim are
guaranteed)
7. wikipedia has a policy that aims toward one outbound link per infobox. So
for a person or organisation page eg they tend to settle on that person /
orgs's homepage and not their social media accounts or web presence(s)
elsewhere. Which makes dbpedia less useful as an identifier triangulation
point
=== end of problems (at least the one's I can remember) ===
So I think we'd be interested in wikidata for 2 (maybe 3) reasons:
1. as a source of data for domains where there's no established (open)
authority (eg the equivalent of musicbrainz for films)
2. as a better, more stable source of identifiers to triangulate to other
data sources
?3?. Possibly as a place to contribute of some of our data (eg we're
donating our classical music data to musicbrainz; there may be data we have
that would be useful to wikidata)
Have glanced quickly at the proposed wikidata uri scheme
(http://meta.wikimedia.org/wiki/Wikidata/Notes/URI_scheme#Proposal_for_Wikid
ata) and
<snip>
http://{site}.wikidata.org/item/{Title} is a semi-persistent convenience URI
for the item about the article Title on the selected site
Semi-persistent refers to the fact that Wikipedia titles can change over
time, although this happens rarely
</snip>
Not sure on the definition of infrequently but I know it's caused us
problems.
Wondering if the id in http://wikidata.org/id/Q{id} is the wikipedia row ID
(as used by dbpedialite)? Also wondering why there's a different set of URIs
for machine-readable access rather than just using content negotiation?
Cheers
Michael
http://www.bbc.co.uk/
This e-mail (and any attachments) is confidential and may contain personal views which are not the views of the BBC unless specifically stated.
If you have received it in error, please delete it from your system.
Do not use, copy or disclose the information in any way nor act in reliance on it and notify the sender immediately.
Please note that the BBC monitors e-mails sent or received.
Further communication will signify your consent to this.