[00:05:04] Stryn: that looks fun thanks [00:10:43] Stryn: I take that back [00:11:04] someone seems to have crawled through every possible ID on geonames [00:22:27] a simpler way to clear https://ceb.wikipedia.org/w/index.php?title=Espesyal:UnconnectedPages is by checking for duplicated wikidata geoname IDs [06:19:17] duplicate geoname should be a regular constraint [06:19:32] duplicate geonameID should be a regular constraint [07:32:34] * edoderoo gives stashbot some attention ;-) [15:58:35] Wikibase\DataModel\Deserializers\StatementDeserializer::deserialize [15:58:45] That one is called /a lot/, especially when dumping [15:58:55] I wonder whether we should inline its function calls [15:59:00] * hoo hides [15:59:55] Or at least pull the ifs from the private methods up [16:03:28] DanielK_WMDE: Have a second? [16:05:28] is it currently possible to query for items with duplicated geonames IDs? [16:06:06] there's a constraint report which should list them [16:06:15] although I haven't dared try to load the page for a long time [16:07:27] https://www.wikidata.org/wiki/Wikidata:Database_reports/Constraint_violations/P1566 see unique value [16:08:04] thanks nikki [16:08:04] should I have a lot of RAM to load that page? lol [16:08:38] query version (almost same number of results): http://tinyurl.com/hhj6joh [16:09:35] heh beat me to it [16:10:23] the page does load for me, but it's quite big and doesn't load particularly fast [16:10:40] I can't really say whether you need a lot of ram or not, I have plenty :/ [16:11:27] perfect thanks [16:18:49] well, it's exactly what I expected [16:19:15] ceb and sv pages unconnected with other human-created articles in other wikis [16:21:18] yeah, the bot has created so many articles that it's a nightmare to deal with. I was going to say keep on top of, but I don't think we ever have managed to get on top of it :P [16:24:44] I added it as a task suggestion on https://www.wikidata.org/wiki/Wikidata:Status_updates/Next#Monthly_Tasks [16:26:21] there's also duplicates where svwiki already has a human-created article, duplicates where the bot created two articles from the same geonames id, duplicates where the bot created multiple articles for places where we would normally have multiple p31 statements (like cebwiki now has 4 pages about berlin... so far no new wikidata items though) [16:27:55] lol [16:28:30] swedish wikipedians must hate that bot [16:33:31] Swedish Wikipedians seem quite flexible when it comes to bots ;) [16:34:43] :o I just found a pair which have nothing to do with the bot [16:36:15] me too, province and province capital with the same name [19:42:31] I'm importing some metadata from GRID into Wikidata via QuickStatements. I am adding statements such as "official website" (P856), and wonder if I should check before that the items don't already have a value for this property (as it is supposed to be single-valued). [19:43:13] here is an item that has been impacted by my QuickStatements (only for a trial run with 10 statements): https://www.wikidata.org/wiki/Q499510 [19:44:39] my statements add a correct website, but there was already an (incorrect, 404) website before