-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
OpenRefine+Wikidata quick demo #5
Comments
Yes, your reconciliation tool looks great and would be a good thing to demo, play with and hack on. We haven't fully figured out how to organize the lightning talks, other than the generic slot for them in the program. We'll update that as things become more concrete. |
I just gave this a try. I downloaded the results of this SPARQL query
and fed them into OpenRefine, which resulted in 351 matching rows that I then converted to the format of the new author resolver, which gave this list, from where I then picked cases to look at in more detail, which resulted in ca. 1k replacements of P2093 statements with the corresponding P50 statements. That looks promising. Things that still need attention in this workflow:
Pinging @magnusmanske I also tried to tackle another problem by way of a similar pipeline:
This is where I got stuck. Can the tool be used at all in any way that would help with replacing those journal items in P921 statements with the corresponding items about the actual topics, as per the mapping here? |
@Daniel-Mietchen Thanks for giving it a spin! But your links to OpenRefine refer to your local instance of the tool, that we cannot access. Screenshots? Concerning your mapping, I have some ideas to make this work if the mapping is stored on Wikidata (as journal to topic statements). I'll add the relevant endpoint and make screenshots to explain how to use it. |
This was my first try with Open Refine, so I'm still trying to find my way around. Is there no way to make my OpenRefine projects open, perhaps even by default? Would be nice to have them synced with Zenodo or so for every "release". In this specific case, though, I don't think it matters too much (and screenshots wouldn't make much of a difference), since I simply took the outputs of both SPARQL queries (in csv format) and imported them into OpenRefine. |
Re lightning talks, we now have #13 to get them organized. |
+1 |
Very cool! I'm testing it out on this dataset https://figshare.com/articles/COAF_Jisc_and_RCUK_APC_data_2013-2015/3462620 |
So, we've done a lot of things on this topic:
So many thanks to all who got involved! |
I have been working on a tool that sounds quite relevant for the event:
https://tools.wmflabs.org/openrefine-wikidata/
It helps align datasets to Wikidata in OpenRefine, a super cool software to deal with messy data. If you are still looking for lightning talks during the event, I would be happy to give a quick demo of the tool. I'd love it if we could then play with the tool on some research data (and I'm sure some attendees will know of many interesting datasets).
The text was updated successfully, but these errors were encountered: