[08:34:08] is it possible to get the Q of a person, quering the person name in greek (assuming a label in greek exists in persons' wikidata page)? [09:26:58] multichill: hey, around? [09:27:29] addshore: hey, around [09:27:30] ? [09:32:21] hey Amir1 ! [09:32:44] you pinged me and I missed it then I pinged you then you missed it! [09:32:57] Hahaaa :P [09:33:06] I don't remember what it was about now :D [09:35:06] You were talking about wd-analyst [09:35:10] I'm working on it now [09:35:16] oh wait, it wait, Lydia said that you might be interested in the thing I am making to add references to Wikidata? :P [09:36:41] ohh yesss [09:37:28] are you trying to harvest references from wikipedia? [10:18:43] Amir1: yes! [10:18:52] bah, this conversation is so disjointed... [10:20:54] so, thoughts Amir1 ? :D [10:21:12] Lydia_WMDE: I'll do the 'summary' pages for each KPI today [10:21:22] I had ideas on how we can do this [10:21:34] but I abandoned them [10:21:55] Currently I am loading external links from articles, crawling the links for structured data and then if there is structured data using it to add reference [10:22:10] See https://www.wikidata.org/wiki/Special:Contributions/Addshore for my tests over the past 2 days [10:22:22] e.g. we can see what's the most used values birthdate so can easily add more [10:22:25] basically focusing on movies currently [10:22:35] I see [10:22:57] but expanding to other types where I can map schema.org to wikidata should be very easy [10:23:27] Can you share your source code with me? [10:23:35] maybe I can do some stuff with it [10:23:35] its on github :) [10:23:40] awesome [10:23:48] its a fair bit tider than it was yesterday morning, :) [10:23:58] my initial proof of concept was a mess to say the least.. [10:24:32] I mad it as part of this https://github.com/addwiki/addwiki the studd for the command is in https://github.com/addwiki/addwiki/tree/master/src/Commands/Wikimedia/WikidataReferencer [10:24:44] main class / command class is https://github.com/addwiki/addwiki/blob/master/src/Commands/Wikimedia/WikidataReferencer/WikidataReferencerCommand.php [10:26:28] I'll take a look at it [10:26:31] thanks [10:26:33] :) [10:26:49] it's probably going to go through another rapid stage of evolution this afternoon / evening [10:29:51] once you're done, tell me and I'll maybe make some PRs [10:29:56] :D [10:49:45] Lydia_WMDE: https://grafana.wikimedia.org/dashboard/db/wikidata-kpis [10:57:38] git add * [10:57:48] ha [11:09:00] * aude waves :) [11:10:26] Lydia_WMDE: http://tools.wmflabs.org/wd-analyst/index.php?p=P31&limit=24 [11:10:32] labels are supported now [11:24:53] Lydia_WMDE: please ping me when you're around [11:26:37] Amir1: hey [11:26:57] hey, [11:26:58] What do you think of the new look? [11:27:06] http://tools.wmflabs.org/wd-analyst/index.php?p=P31&limit=24 [11:27:06] addshore: yay! currently in kde office. will look into it later today [11:27:20] Amir1: looking [11:27:55] Amir1: i like it :) [11:27:58] yay for labels [11:28:15] \o/ [11:28:31] I hope make other things fixed [11:28:44] Specially I don't like the PHP code, it's so ugly and ad hoc [11:29:05] I might refactor it soon [11:29:50] :D [11:29:51] ok [11:41:30] I downloaded the wikidata json dump. Is there a program that gets a list of all page titles and redirects and their corresponding wikidata ids? [11:54:10] physikerwelt: not that I know of, but you could make one ;) [11:54:40] ahh yes Amir1 that php code looked interesting :P [11:55:13] it's the ugliest thing I've ever written [11:55:19] addshore: ok. That's what I'm doing right now... but I wanted to avoid duplicate work [11:55:21] Amir1: does all of that stuff come from dumps I guess? :) [11:55:32] yes [11:55:43] It goes a database [13:36:06] * aude wonder if anyone is at the office [13:36:13] so quiet [13:47:24] :P [13:47:36] * hoo is looking for Sandra [13:47:44] o_O [15:04:07] xD [16:20:04] i just dont know if it is odd to have promises of stuff everywhere rather than actual things [16:20:05] bah [16:20:26] need some way to typehint what thw unwrapping of a promise will return... [16:22:29] unless I have a bunch of Async method and a bunch of non Async methods.. [16:22:34] mhhmh [16:29:08] Hello? [16:29:59] How to create query to fetch data from Wikidata by a Program? [16:30:38] Someone here? [18:35:06] any translaters here? [18:35:56] why? [18:37:15] I've a question [18:37:35] I do some trnalsating sometimes, but am no expert [18:37:55] page for Wikidata Tours have "Start this tutorial" button in english in all languages [18:38:07] although its a template and translated to many [18:40:13] MGChecker see that? [18:40:46] HakanIST You mean on the https://www.wikidata.org/wiki/Wikidata:Tours/de page? The blue button? [18:41:02] At least in German for me its in German, too [18:41:10] Yes [18:41:29] button is in German for you? [18:41:40] Yes, but only with the /de link [18:42:05] If I switch the link to /en and have my user language to German, everything is English but the button [18:42:10] Wow, that looks weird [18:44:51] HakanIST I'll check the JS Code but I guess the button just uses the user language so it's always in the language of the user/the Wiki and doesn't change with the template. But that's just a wild guess, I'll look into it [18:45:26] frimelle you are right, I just switched my language [18:45:28] my bad [18:46:01] Don't worry. I guess you should still open a ticket, would be nice if the button and the text would be in the same language [18:47:37] yes I think it's necessary for anonymous [19:07:38] HakanIST exactly [19:07:52] If you open a ticket in phabricator, could you CC me? [19:10:00] sure thing [19:18:31] Tpt: ! you making this PR means I found this library again! :D https://github.com/ProjetPP/WikidataQueryApi/pull/4 [19:22:15] addshore... generting the list took a little while https://github.com/physikerwelt/WikidataListGenerator [19:23:01] oooh, physikerwelt if your doing it in java you should use the wikidata toolkit! [19:23:16] https://www.mediawiki.org/wiki/Wikidata_Toolkit [19:23:35] https://github.com/Wikidata/Wikidata-Toolkit [19:24:21] example implementation of some stuff https://github.com/wmde/wikidata-analysis/blob/master/java/analyzer/src/main/java/org/wikidata/analyzer/Processor/ReferenceProcessor.java [19:27:32] addshore: is there a method to fetch the wikidataid from the title from the dump [19:27:58] otherwise I would probably be blocked by wikidata;-) [19:28:12] wikidataid from the title from the dump ? [19:28:28] addshore: great! [19:28:37] I have a page title and want to get the wikidataid [19:29:23] addshore: but sending a request for each title to wikidata would probably be too much [19:29:55] Well if it is for Wikidata you can make the assumption itemIdPageTitle = itemId [19:30:09] for properties just strip the namespace "Property:" [19:30:39] I have the title of the english, russian or german wikipedia respectively [19:31:24] oh, the wikipedia title? [19:31:49] yes [19:32:14] yeh, I would write something using the wikidata toolkit! :) [19:33:21] I had a look at the wikidata toolkit before, but it was not clear to me if can process the whole dump offline [19:33:29] yup you can! [19:33:57] give me a sec, I'll see if I can bash something together for you [19:34:20] physikerwelt: you have enwiki page title right? [19:34:34] en de and ru [19:34:50] but we could start with en [19:44:43] addshore... I need to run [19:44:54] literally wait 45 more seconds [19:45:49] https://github.com/wmde/wikidata-analysis/tree/physikerwelt [19:46:01] https://github.com/wmde/wikidata-analysis/blob/physikerwelt/java/analyzer/src/main/java/org/wikidata/analyzer/Processor/PhysikerweltProcessor.java [19:46:41] you should be able to checkout that branch, build it and run it and it should give you a text file with ID "enwikititle" "dewikititle" "ruwikititle" on each line [19:47:15] ping me when you come to running it :) (if you already have the dump downloaded you can make it use that [19:47:34] if not you can run it on labs, or I can run it for you and dump the output somewhere [19:47:53] addshore... thank you very much [19:47:58] no worries ;) [19:48:10] my girfriend is waiting for me... but I'll try later [19:48:14] :D [19:48:19] Amir1 still around? [20:49:29] I want to construct a SPARQL query for all items that contain statements with multiple values for P2404 [20:49:39] How might I do that? [21:37:21] addshore are you still around? [21:37:27] yup [21:37:53] I checked out the physikerwelt branch and set up the analyzer using mvn [21:39:05] is there information how to run the program? [21:50:31] yes! [21:50:39] addshore ... I think I got it [21:50:43] so do you have a dump downloaded? [21:50:44] :D [21:50:46] however it downloads the file again now [21:51:07] so the data dir that you pass to it [21:51:14] put a dumpfiles directory in that [21:51:18] and put the dump file in that [21:51:23] yes but it did not find the dump [21:51:37] mabye because I unzipped it [21:51:38] hhhm, which format is your dump in? [21:51:44] yes, it needs to be the compressed one! [21:52:10] i downloaded the json file and I think bunzip2 delete the original [21:52:18] wahhaaa :p [21:52:22] rezip it? ;) [21:53:53] but the filename is ok wikidata-20151207-all.json [21:54:16] the program said Getting dump with date 20151214 [21:54:25] you can pass in a date too! [21:54:40] I think that will be faster comparing to zipping it again [21:55:49] do I have to get gz or bz2 [21:56:12] bz2 is a bit ugly... I never managed to read it in parallel ;-) [21:56:37] java -Xmx2g -jar ./wikidata-analyzer.jar Physikerwelt~/data 20151207 [21:56:42] java -Xmx2g -jar ./wikidata-analyzer.jar Physikerwelt ~/data 20151207 [21:57:34] ^^ ah that works [21:57:39] epic :) [22:00:14] physikerwelt: and it should output roughly how far through it is too [22:00:16] is that the expected output [22:00:23] https://gist.github.com/physikerwelt/35771b08007c9f1ba136 [22:00:28] and then at the end the data you want is in a file [22:00:40] Using online dump file so its downloading it again [22:01:45] format of local dumps needs to be wikidata-" + dumpDate + "-all.json.gz [22:02:47] ok.. thank you [22:03:13] I think I have to wait until the dump is downloaded... the conenction is quite slow at the moment [22:03:41] normally I get 20MB/s but currently it's only 2 [22:04:48] I'll let you know when I made progress... [22:07:22] physikerwelt: the dumps.wikimedia.org domain is limited I think [22:07:30] really, recompressing it will probably be faster! [23:23:53] if dumps.mw.o is to slow, use the mirror at http://dumps.wikimedia.your.org/ [23:24:38] hoo: I manged to download the dump file but not to read it with