[09:16:42] Lydia_WMDE: has anyone worked out yet whether bots importing non-cc0 datasets is actually acceptable? [09:17:11] nikki: in general the answer is: not ok [09:17:48] what do we do about bots which have already imported non-cc0 stuff? [09:18:13] have a stern talk with the owner i guess :/ [09:18:29] I'm wondering because I just noticed https://www.wikidata.org/wiki/Wikidata:Requests_for_permissions/Bot/WDBot was approved to import a cc-by dataset, they think it's ok because they've included attribution in the reference [09:18:51] :( [09:18:56] not ok [09:26:35] would you mind leaving them a message? I'm not really sure what to say [09:34:18] ok [09:36:43] nikki: left a note on the talk page of the operator [09:36:57] thanks :) [09:54:48] Lydia_WMDE: do you happen to know some sparql geeks who would be willing to help us at MusicBrainz? [09:55:17] We'd love to have something that can be queried directly in combination with Wikidata even, but we don't have anyone who knows how to do anything like this [09:56:57] reosarevok: The best way is probably to setup a SPARQL endpoint and then use SPARQL federation to query both query.wikidata.org and MusicBrainz endpoint at the same time [09:57:19] If I remind correctly your data are stored in SQL tables [09:57:46] It is. We *did* have some sort of SPARQL endpoint once [09:57:51] And there's some code for that [09:57:56] nice [09:57:59] There's just nobody who can maintain it at the moment [09:58:13] Since the one contributor who did got busy with IRL things [09:58:19] I see [09:58:24] https://github.com/ontop/ontop is a fairly nice project [09:58:35] https://github.com/LinkedBrainz [09:59:15] it translates SPARQL queries to SQL, send it to the SQL database and then returns the results as SPARQL [09:59:23] it allows to avoid all synchronization issues [09:59:50] and uses R2RML mapping just like you have [09:59:52] https://github.com/LinkedBrainz/MusicBrainz-R2RML [10:01:06] I am using it for a test of a SPARQL endpoint for Wikipedia database: https://www.mediawiki.org/wiki/MW2SPARQL [10:02:02] Ontop is developped by an Italian university [10:04:56] reosarevok: there is also https://www.wikidata.org/wiki/Wikidata:Request_a_query for getting help with writing specific queries [10:05:44] Lydia_WMDE: I can write queries most of the time, it's just about having an actual endpoint up :) [10:05:55] gotcha [10:06:45] reosarevok: I'm quite buzy at the moment, if would want I could make you connected with the Ontop guys. They would probably be interested to help you. [10:07:05] Basically we had RDF and SPARQL before it was cool - and then it still wasn't cool, and nobody used it, and now when it's finally useful nobody is left to deal with it :D [10:07:15] If not, I could maybe take some time during my Christmas break [10:07:15] Tpt[m]: wouldn't hurt! [10:07:27] This is not something we're in a hurry with, anyway [10:07:38] cool [10:07:41] It's just all the cool kids are playing with the stuff now (finally) [10:07:43] :D [10:09:04] could you give me your email in DM? [10:09:15] it's the best way to reach Ontop people [10:09:25] the other way is for you to open a task on GitHub [15:13:49] addshore: Hello, the reason you had the failures last time was because you didn't have PageImages extension installed locally [15:13:57] Was able to reproduce the error [15:14:09] [= [15:14:24] All fixed now and PageImages extension is no longer a hard dependency if this is merged [15:14:37] https://gerrit.wikimedia.org/r/c/mediawiki/extensions/MobileFrontend/+/474472 [15:14:40] My env was clean :) [15:15:23] addshore: But thanks a lot for the heads up, that idea you gave also helped a lot in troubleshooting the problem [15:15:31] ;) [19:37:57] Hello [19:38:18] What's the tool for making queries to generate a list of article, already? [19:38:41] I need to get the list of exoplanets per constellation [19:38:49] for Wikipedia [19:38:55] (in French) [19:55:03] JN_Squire: Listeria? [19:55:33] (if you mean something like https://www.wikidata.org/wiki/Wikidata:WikiProject_sum_of_all_paintings/Collection/Tartu_City_Museum) [19:55:37] I just want the elements [19:55:43] Not a Wikipedia table [19:56:02] THe table is going to be too complet to be automated (notes) [19:56:10] complete* [19:56:35] Oh. Then just a SPARQL query should work? Or you could still use listeria to generate it, then just copy the table and remove the bot bits :) [19:56:44] (and extend it as desired) [19:56:57] There might be a second tool I just don't know about though :) [19:57:35] SPARQL! That's what I've forgotten! [19:58:20] Aaaand I have no idea about how to make the query I want (confirmed planet for a given constellation) [20:00:07] JN_Squire: what are the appropriate properties? [20:00:26] I can probably help but I know nothing about the topic, so you'd need to give me an example [20:10:21] Take for example https://www.wikidata.org/wiki/Q1198453 [20:10:44] Constellation is P59 [20:12:36] the attached constellation is Q9285 (Coma Berenice) [20:13:01] reosarevok: is it everything you need? [20:13:19] Probably :) Give me 5 min [20:13:24] Finishing a grant request [20:23:53] https://media.giphy.com/media/jSfihttps://media.giphy.com/media/jSfiX3lj42RDG/giphy.gif X3lj42RDG/giphy.gif [20:43:08] JN_Squire: http://tinyurl.com/y95ue7h9 that kind of thing? [20:44:52] Oh, you wanted French Wikipedia specifically, right? [20:44:55] Can also do that [20:45:07] reosarevok: nice! it's even better! I didn't expect to see so little planet's with the constellaltion label [20:45:13] Thank you! [20:45:35] So this is good enough? I can change a few things if needed :) [20:45:41] Is constellation labelling from ViziR doable by a bot? [20:46:22] Would that be http://vizier.u-strasbg.fr/vizier/VizieR/constellations.htx ? [20:46:48] Depends on licenses, I'd imagine [20:46:50] Is the data CC0? [20:47:58] Almost anything that has CC0 data can be put into Wikidata by a bot, the question is how hard it is :) [20:48:13] (anything with other licenses probably can't be put in) [21:09:44] reosarevok: Itsays this http://cds.u-strasbg.fr/vizier-org/licences_vizier.html [21:10:07] And the bot need to find each element and look for the right dataline [21:34:48] I'm going to sleo [21:34:51] Good night everyone