[13:35:53] hei. Would anyone be able to help me with a SPARQL query? I am trying to get Unit of the value of Property event interval (P2257) and failing to do it. I have very simple query example in https://w.wiki/Wnd [13:37:23] aaaac: https://w.wiki/Wnh [13:38:28] Lucas_WMDE: Thank you. I knew it would something simple [13:38:32] :) [13:38:33] no problem :) [13:44:32] Lucas_WMDE: the class wikibase:BestRank; is there simply to mark the statement with best rank for property, right? [13:45:08] it gives you the same effect as the wdt: prefix – only return the statement(s) with the best rank, and never deprecated statements [13:45:24] without this, switching from wdt: to p: might give you extra statements that you don’- want [13:45:29] *don’t [13:45:37] (I didn’t check if it makes a difference for this particular query) [13:45:50] Understood [13:46:07] that is actually quite useful. Thanks!! [21:51:47] aloha. anyone familiar with creating mix'n'match scrapers? :D [22:28:46] SothoTalKer_: I made one that works and one that doesn't, so based on that there's a 50:50 chance I can help :P [22:29:11] thats nice. explain it to me :D [22:29:38] what do you want to scrape? [22:31:09] This: http://data.rero.ch/02-A010048656 [22:33:20] do you have a list of ids you want to scrape then? [22:35:05] Anything from 02-A000000001 to 02-A030000000 [22:36:41] some exist, some don't [23:07:16] hm, 3 million is a lot of urls D: not sure if scraping that many pages would be good [23:07:56] but anyway, on https://mix-n-match.toolforge.org/mix-n-match/#/scraper/new you'd need to fill out the "catalog" section which should be fairly straightforward [23:08:34] under "levels" you would want to add a range level, I tried with start 56, end 60 and step 1 [23:09:13] then for the scraper I tried http://data.rero.ch/02-A0100486$1 as the url pattern - it replaces $1 with the numbers from the range [23:09:29] and I tried this for the regex entry: \s*100\s*.*?(.*?) [23:10:34] and in the resolve section, id: 02-A0100486$L1 ($L1 because you want to use the $1 from the previous section), name $1 (which is the bit captured inside the brackets in the regex entry) and url http://data.rero.ch/02-A0100486$L1 [23:10:46] and then when I click test scraper, it seems to extract something [23:37:16] too bad one cannot change a scraper after it was created [23:40:35] nikki: what about stuff that uses field 110 instead of 100? :> http://data.rero.ch/02-A000002990 [23:41:06] change 100 in the regex to 1[01]0? :D [23:41:06] El búfer 01 está vacío. [23:41:18] heheh confused the bot [23:41:30] 100 is for persons, 110 is for corporations :) [23:43:31] I probably also want subfield $d in addition to subfield $a, if given, for a better match. hmm :) [23:44:13] but yep, it's a good start :) [23:44:41] yeah I could only work with the example I had :D [23:45:38] too bad one cannot see the existing scrapers of mix'n'match catalogs [23:51:47] look at the time o_o good night [23:51:59] and tyvm# [23:52:06] you say that like you're not normally awake all night :P [23:52:10] you're welcome :D [23:53:02] i try to avoid that because i need my beauty sleep :x [23:53:37] maybe i'm gonna try to see the comet later, though :) [23:55:20] I saw Hale-Bopp in 1997 and it was great. so hopefully i can see a second one in the next night :D [23:55:37] ohh I should have a look too