Thank you so much Magnus. I will follow up if I get IDs and names.
I will take a deep look at this tool and the asset types, etc. It seems like some of these smaller datasets would be really good potential Editathon projects? It would be great to see almost all green checkmarks!
As far as databases, I was also hoping to work with the Rigler-Deutsch Index (https://en.wikipedia.org/wiki/Rigler-Deutsch_Index) which is a union catalog of 78 RPM records held in the U.S. There is an extraction project underway (http://searchworks.stanford.edu/view/km199fq2766) and the increased discover-ability of the music would be sort of great. It is an important database for music.
Thanks again,
- Erika
*Erika Herzog* Wikipedia *User:BrillLyle https://en.wikipedia.org/wiki/User:BrillLyle*
On Wed, Sep 7, 2016 at 4:39 AM, Magnus Manske magnusmanske@googlemail.com wrote:
One "typical" approach for a data set this type and size is Mix'n'match: https://tools.wmflabs.org/mix-n-match/
If you get a list of IDs and names, let me know.
On Tue, Sep 6, 2016 at 7:17 PM Brill Lyle wp.brilllyle@gmail.com wrote:
Hi Wikidatans,
After going past my 500th edit on Wikidata #Whee! I was hoping to dip my toe into doing something that would involve a larger scale project, like adding database information to Wikidata.
There's a database I use all the time that is excellent, rich, deep, and well-deployed -- at JewishGen.org
main search page: http://www.jewishgen.org/Communities/Search.asp example page: http://data.jewishgen.org/wconnect/wc.dll?jg~jgsys~ community~-524980
I started a Property proposal here:
https://www.wikidata.org/wiki/Wikidata:Property_proposal/ Place#JewishGen_Locality_ID_English
I have also contacted the folks over at JewishGen to ask if they might provide me with raw data, initially even just with the locality page IDs, then hopefully more rich / fuller data that's in the database.
I was wondering if this is
(a) the typical approach people use when importing data (b) if you have any advice / best practices to share (c) also, if I should try and do a wget to scrub for this data (if that's even possible)? do people do this to grab data?
This information, I envision being used as part of a unique identifier that could be built into infoboxes, and might also be a sort of templatized box even (although I don't hugely love the issue of restricted / redirected editing away from Wikipedia). But I would really like to see this information in a pathway to Wikipedia. I think it would improve a lot of these town pages, a lot of which are stubs.
Best -- and thanks in advance for any advice,
Erika
*Erika Herzog* Wikipedia *User:BrillLyle https://en.wikipedia.org/wiki/User:BrillLyle* _______________________________________________ Wikidata mailing list Wikidata@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata
Wikidata mailing list Wikidata@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata