[07:24:37] la la la [09:46:16] hoo: I made a patch after what we discussed last night [09:47:28] addshore: Ok... just told Thiemo and he also made a patch [09:47:36] but he removes it entirely from getentities [09:47:41] two approaches [09:48:25] ahh okay [09:48:28] link to his? [09:48:56] got it! [09:53:00] hoo: Thiemo_WMDE I think I would probably be happy with that approach too. [09:55:42] oh wait, but Thiemo_WMDE, you prefer the caching way? :) [11:57:11] dear colleagues, where is a good place to ask about lua + wikibase? [11:58:05] One question: how long it would take to make 100K items in Wikidata? [11:58:33] not me [11:58:36] people [11:58:52] i.e. I want growth speed of Wikidata [12:03:33] nvm. I got it. it's about a week [13:50:40] how would i get an interable of wikibase items from lua? [14:30:26] how would i create an interable over several wikidata items in lua? [14:30:40] Does anyone by any chance have a dump downloaded from january of this year? :d [14:30:42] :D [14:40:15] addshore: not one on labs [14:40:16] ? [14:40:29] nope, that are already gone [14:40:33] (a JSON one) ;) [14:43:32] addshore: think i do [14:43:53] aude: one from as close to 1st jan as possible would be great!!!! [14:44:57] jan 11 [14:45:10] it will take a few minutes to upload it from the office [14:46:17] also have one from march 2015 and one from june 2015 [14:46:30] awesome, the jan one would be amazing :) [14:46:53] will take ~12 minutes to upload [14:47:05] okay! [15:01:41] addshore: still says 9 minutes :( [15:01:49] wmde internetz is sloooooooooooow [15:01:53] haha, thats okay! [15:02:01] it'll probably take me just as long to download it [15:02:09] just give me a ping / link when it done :) [15:07:33] addshore: how urgent do you need this? [15:07:43] today would be epic [15:07:52] think i misread and it's 9 hours :( [15:07:59] aaahhhh xD [15:08:03] then tommorrow is fine :P [15:08:23] ok, maybe i can let this run tonight from my hostel :) [15:08:31] :D [15:08:35] maybe their internetz are faster [15:08:47] if not I can also give you the thing to run! [15:09:12] 111 kb/s [15:09:16] really not fast [15:11:10] aude: https://github.com/wikimedia/analytics-wmde-toolkit-analyzer-build [15:12:08] if uploading tonight isn't faster, then i cna run this [15:12:17] okay :) [15:12:26] just not sure how it all works and would like to make the dumps available anyway [15:12:58] java -Xmx2g -jar toolkit-analyzer.jar Metric ~/data 20160111 [15:13:05] ok [15:14:21] would be how you run it and you have to put the dump in ~/data/dumpfiles/json-20160111/20160111-all.json.gz [15:14:25] I think ;) [15:14:57] yeah, that's where i have it already [15:18:37] addshore: it tries to download the dump and then quits when it finds a dump already there [15:18:51] hmmm? [15:18:54] i will try to upload [15:18:59] okay! [15:19:21] java.nio.file.FileAlreadyExistsException [15:19:26] I should make it so you can just pass the path of a dump to run in too :P [15:19:31] oh, interesting [15:19:31] yeah [15:19:44] it should try this.dataDirectory + "/dumpfiles/json-" + dumpDate + "/" + dumpDate + "-all.json.gz" first [15:19:48] and i don't want anything bad to happen to my dump file, though i have backups in the US [15:20:21] i have the dumps for use with wikidata toolkit so should have them in an appropriate place [15:28:14] addshore: the dump will be on http://dumps.filbertkm.com/ (and want to setup https on my domain) [15:28:59] aude: ahh, well my code may have diverged from the toolkit stuff slightly [15:29:09] hmmm [15:29:09] as it all looks for local dumps on the cluster and labs [15:29:15] i see [15:29:19] you need to pass the data directory in when you run the script [15:29:54] and if the data directory is ~/data then the dump needs to be at ~/data/dumpfiles/json-" + dumpDate + "/" + dumpDate + "-all.json.gz" [15:30:31] i had stuff in ~/dumps/dumpfiles ... but tried to move them to ~/data/dumpfiles... and didn't work [15:31:21] hmm okay, I might need to double check that does still work then (it doesnt get run that often) as usually it just grabs the dumps from labs [15:32:10] what was the stacktrace with java.nio.file.FileAlreadyExistsException ? [15:32:12] i don't want it to somehow accidentally delete my file [15:32:22] until i upload a copy, at least [15:32:33] else we have to wait until thursday when i am back in the US [15:34:07] it wont delete it ;) [15:34:22] it never deletes anything :D [15:34:37] addshore: mmmeeeeerrrge https://gerrit.wikimedia.org/r/#/c/270020/28 [15:34:54] ooooh DanielK_WMDE_ awesomeee!! :D [15:36:14] DanielK_WMDE_: is this the kind of way you envisioned it being used? https://gerrit.wikimedia.org/r/#/c/281915/2/tests/phpunit/MediaWikiTestCase.php [15:44:30] addshore: yes, something like this. Perhaps overrideMwServices should just take a map map services as the first arguement. [15:45:13] yeh perhaps, it's also quite nice to be able to do them one by one [15:45:36] I'll keep work again on my patch before I merge yours to make sure it all feels right ;) [15:46:13] addshore: ok, but let's keep this going. I'm a bit blocked on this one [15:48:46] yeh, DanielK_WMDE_ I just need to figure out why the random special page test is failing, it almost seems like the service isn't being set back again after the test though :/ [16:07:59] addshore: it probably is, but the special page instance may still be around, and may have a reference to a stale service instance... [16:08:13] mhhhm, ill have a dig! [21:07:08] Stupid tool... adding the same claim that already exists :/ https://www.wikidata.org/w/index.php?title=Q3523054&diff=320048304&oldid=320048023