[07:11:26] hi there [07:11:29] everyone [11:09:07] Are the germans +1'ing https://youtrack.jetbrains.com/issue/WI-24808 ? :P [11:14:07] FYI, they don't care about votes [11:14:22] See also https://youtrack.jetbrains.com/issue/WI-473 [13:22:02] Here's an idea: is there any way to get API request data for Wikidata? While page views is one sort of metric, I feel like for Wikidata that API access is a more interesting piece of information. [13:34:21] harej: so you mean see what items have been accessed through the api? [13:34:29] yes, and how many times [13:35:02] interesting, I mean, it is defintly data that we have just not in a nice way [13:37:41] is it something that would be found on labs somewhere? [13:43:04] no [13:43:13] WMF have api logs and stats available... but not public [13:43:34] what would it take to make this information publicly available? obviously with the PII redacted [13:44:25] Probably a lot of effort [13:44:42] Depends what exactly you're warning [13:44:51] You could probably get analytics to write some reports for it [13:45:17] I mean, would it require someone with access to nonpublic information? [13:45:47] I guess [13:47:42] This information is important to me; I'm going to try to get funding to do this. If I find a person and pay them, could I get them through the process? [13:49:31] Depends on the WMF [13:49:48] But someone appropriately NDA'd should be able to get access [13:49:53] harej: I assume you already checked the Research committee suggestions? [13:49:53] Like I say, talk to analytics first [14:00:48] Nemo_bis: I hadn't thought of that; I wanted to ask if the report existed first [14:00:51] But sure, I can talk to analytics [15:58:44] Please remove gujarati language wikipedia of "Anwar Agewan" [15:59:12] hello [17:22:54] so I asked Analytics and they're kind of working on it? but i don't know if it's exactly the thing I want. [17:35:38] is Lydia_WMDE around? [17:38:28] harej: Lydia_WMDE has gone home for the day [17:39:09] you might know, in any case: is anyone currently working on metrics for access to wikidata's data by means other than visiting the website [17:39:25] i can get page view data for wikidata but i don't think that's very interesting compared to machines using wikidata [17:39:48] addshore was at some point poking at dump access data [17:39:53] but even that is incomplete [17:40:03] there are several mirrors for which we don't have access logs [17:40:11] yea, i was also going to suggest talking to addshore [17:40:20] addshore said he is not working on that precise thing [17:40:38] there is actualyl two things to consider: [17:40:45] access to the query endpoint [17:40:51] and access to the linked data interface [17:40:59] WDQS and Special:EntityData, respectively? [17:41:07] they are on different serves, using different urls and different infrastructure [17:41:19] harej: yes, exactly [17:41:24] For special:entitydata, see https://grafana.wikimedia.org/dashboard/db/wikidata-special-entitydata [17:41:33] and both of them may use the varnish caches for some things [17:41:38] but the blazegraph instances also access that [17:41:55] we have some stats for that, yea [17:42:12] so that graph gives me *all* the hits to entitydata; i'd like to see a breakdown by item [17:42:48] don't think we have that :/ [17:42:49] that's oprobably not possible for WDQS [17:43:10] WDQS does not capture which items are served in queries? [17:43:39] no. it doesn't know about items at all, actually [17:43:47] it only knows about triples and resource uris [17:44:04] it's an rdf triple store. it knows nothing of wikidata [17:44:10] can one infer items from triples and resource URIs, or are we out of luck there [17:44:25] you can, yes [17:44:40] but you would have to log the full query and full response [17:44:47] (i take it the triple is item/property/value?) [17:45:04] some triples are like that [17:45:31] but some are item/property/statement and some are statement/qualifier/value [17:45:40] and some are value/field/literal [17:45:42] etc [17:45:54] the rdf mapping is rather complex. [17:46:30] i can see how it would be expensive to get individual items out of that [17:46:40] so, let's disregard WDQS analytics for now. if i found you a volunteer and navigated them through the NDA process, could that person come up with entitydata access metrics on a per-item basis? [17:47:31] harej: that should be possible, yes. and analytics hopefully has the raw data for this. but i don't know the details [17:48:15] Yeah, we have the data in question in hive [17:48:30] just takes someone to crunch it [17:48:55] How much time should it take for an experienced engineer to produce such a dashboard? [17:51:37] not sure how that part currently works [17:51:46] so dunno how much work it would really be [17:51:54] but I presume not that much [17:58:12] harej: the api stuff is only very recently in hadoop / hive [18:11:12] * aude waves [18:11:47] hey aude :) [18:14:30] Just landed? [18:18:31] hoo|busy: last night [18:18:49] * aude assumes the train is happening today [18:19:01] I guess so [18:19:12] hope we can see a huge drop in UPDATEs on tendril [18:19:21] thanks for sorting out the deployment stuff [18:19:27] :) [18:21:36] :) [19:12:59] aude: https://tendril.wikimedia.org/host/view/db1018.eqiad.wmnet/3306 Look at that [19:13:04] "Query Write Traffic -- 24h / 5m" [19:16:27] \o/ [19:30:39] also https://tendril.wikimedia.org/host/view/db1052.eqiad.wmnet/3306 [20:15:17] Lydia_WMDE: Did you see the fun javascript hacks we did on Wikidata? ;-) [20:15:55] https://www.wikidata.org/w/index.php?title=Wikidata:Europeana_Art_History_Challenge/Content_Translation&withJS=MediaWiki:EuropeanaContentTranslation.js&itemid=Q2258948&destlang=nl [20:16:41] Javascript magic, feels like the Wiki Loves Monuments preparation [21:24:47] DanielK_WMDE__: I wonder if it is worth it to deploy https://gerrit.wikimedia.org/r/#/c/283555/ now, to understand https://phabricator.wikimedia.org/T132645 [21:29:12] REMINDER: about next week's data center migration. There will be NO EDITING for about 30 minutes on both Tuesday, 19 April at 14:00 UTC and Thursday, 21 April at 14:00 UTC. This will affect ALL of the wikis. You will be able to read the wikis and to log in, but you will not be able to edit. Read https://meta.wikimedia.org/wiki/Tech/Server_switch_2016 for more information. Please share this news. [21:29:13] If you find unexpected problems (e.g., can't read for more than a few minutes), then please report those in #wikimedia-tech, which will be monitored by members of the Operations team. #wikimedia-operations will be moderated until the switch is finished and everything is stable again.