[03:38:32] Hi [03:39:06] https://pastebin.com/uTQjvkyd is working on https://query.wikidata.org but not on the sparql endpoint with curl while yesterday this exact query worked [03:41:20] java.util.concurrent.ExecutionException: org.openrdf.query.MalformedQueryException: Encountered " "count" "COUNT "" at line 1, [05:01:10] Nevermind, I found the mistake [05:10:46] can i split an item? [05:11:09] and have all the reference to the item, now pointing to the two items [05:59:27] i did : ./loadData.sh -n wdq -d ./data/split , it reach 75 and serverkiled the proc. now i want to start form that file , i restart the server 9999 , from runBlazegraph.sh and "./loadData.sh -n wdq -d ./data/split -s 75 " [05:59:49] it start giving me loop error messages [05:59:58] Caused by: java.lang.RuntimeException: Could not load: url=file://./data/split/wikidump-000000075.ttl.gz, cause=java.net.UnknownHostException: . [06:00:09] with all rest of files [06:03:07] any advice [06:03:40] here is the error : https://pastebin.com/G1GBywXG [06:22:32] hmm it looks like it thinks . is the hostname. Try giving full path to -d [06:23:04] i did [06:23:08] same thing [06:23:34] can't be same thing, where it would take "./data/split" then? [06:24:12] what is the exact command line you are usng? [06:24:14] *using [06:24:17] Hold [06:24:24] lt seems working [06:24:31] okie [06:24:37] ./loadData.sh -n wdq -d /download/wiki/service-0.3.0/data/split -s 75 [06:26:18] Processing wikidump-000000075.ttl.gz [06:26:19] waiting ............. which mean it is working correct ? [06:26:29] yep this should work [06:26:33] yes this looks good [06:27:05] you'd probably need to wait for a while. If there's no exceptions then everything is good [06:27:12] Well , i like to ask , is it important to Processing all file to get human name and gender ? [06:27:38] i can get human names but gender labaile return as ID [06:27:54] well, the statemens are in no particular order, so if you want all humans you need all of them, or pre-process the dump (e.g. with something like wikidata toolkit) [06:28:06] I am getting "potential issues" with the Hoovers property (https://www.wikidata.org/wiki/Property:P5232) even when I input correct data; the URLs still work though. I believe the URL formatter regex is incorrect can someone please have a look at this. [06:28:20] you *can* make a TTL with just humans and just genders, but you'll need tool that can process dumps [06:29:07] hold let me show you please [06:30:58] lynx 'http://localhost:9999/bigdata/namespace/wdq/sparql?query=SELECT%20?given_name%20?given_nameLabel%20?fnameLabel%20?genderLabel%20?languageLabel%20WHERE%20{%20?given_name%20wdt:P31%20wd:Q5.%20OPTIONAL%20{?given_name%20wdt:P21%20?gender.}%20OPTIONAL%20{?given_name%20wdt:P735%20?fname.}%20OPTIONAL%20{?given_name%20wdt:P103%20?language.}%20SERVICE%20wikibase:label%20{%20bd:serviceParam%20w [06:30:58] ikibase:language%20"[AUTO_LANGUAGE],en".%20}%20}%20LIMIT%2010' [06:31:07] the output : [06:31:42] https://pastebin.com/WHwvpDhd [06:32:24] ok so what about it? [06:32:40] [06:32:40] Q4357927 [06:32:40] [06:32:46] Q4357927 [06:32:58] is this mean i have to wait all of them ? [06:33:04] to go to memory ? [06:33:18] ah, if you get Q in name then it means no label in that language probably [06:33:37] if you get the item in the result, it's probably already in the db, otherwise you wouldn't have it there [06:33:55] Nice , Well other one please , i am very happy , there is an expert to ask :) [06:34:15] all data on item are next to each other, so unless you've got it split between files, usually what you have is no label [06:34:59] no sure why though - Q4357927 does have en label on wikidata [06:35:04] as you can see i want all human names and gender , but i like to get all name translation to , how can i do that ... some times the name in other lang. [06:35:10] ahh wait [06:35:22] it's not item's label, it's name's label [06:35:32] then yes, it may mean Q4357927 is not loaded yet [06:35:48] the item you found has loaded, but the linked ones have not [06:36:18] so you'd have to wait to load all names and genders and so on, unfortunately [06:36:19] see my select [06:36:21] SELECT ?given_name ?given_nameLabel ?fnameLabel ?genderLabel ?languageLabel WHERE { ?given_name wdt:P31 wd:Q5. OPTIONAL {?given_name wdt:P21 ?gender.} OPTIONAL {?given_name wdt:P735 ?fname.} OPTIONAL {?given_name wdt:P103 ?language.} SERVICE wikibase:label { bd:serviceParam wikibase:language "[AUTO_LANGUAGE],en". } } LIMIT 10 [06:36:31] yeah I see it [06:37:01] Ok let me show you what i need ... from wiki [06:37:41] that's what I am saying - you'd have to wait until those entries load too. which unfortunately hard to predict since dumps are generated in split chunks so it's not exactly ordered inside... [06:38:33] i know but we have to add the select to name lang in other lang. [06:39:11] http://www.wikidata.org/entity/Q11672 [06:39:34] All entered languages [06:39:46] i need the name in all All entered languages [06:40:01] if it avalable [06:40:33] ah, ok, then do ?given_name rdfs:label ?name [06:40:59] instead of SERVICE [06:41:18] or wait, do you need given_nameLabel in all languages or fnameLabel? [06:41:39] that is what i dont know , i am learning , please can you add it to my select to see how it is [06:42:03] sure, but I need to know what you want to get as a result? [06:42:07] fnameLabel is the giving name which is only in one lang. [06:42:20] yes, sure [06:43:10] so i need the given_nameLabel in all languages [06:43:20] ah, ok [06:43:48] what about other labels? [06:44:51] which is https://i.imgur.com/7CKFP8y.png [06:45:25] sure, got that. but what about fnameLabel and genderLabel? do you need them? in which language? [06:45:50] gender in engelich [06:46:15] and fnameLabel in english since i will get the other translations [06:47:59] ok you can do something like this then: http://tinyurl.com/yalwp5vn [06:49:01] Oh hld let me study it :) [06:55:15] it is like magic [06:56:10] Well i am doing AI gender program , so i need all names as dataset , and then if you type any name in your mind even if it is not in the dataset [06:56:19] it will give you the gender :) [06:56:27] 95% of time [06:56:39] do you like to see a small demo :) [06:57:56] https://instaranker.com/panel/gender6.php check this out , once i have full dataset of names , WAW milions of data to learnrom [07:00:27] BTW can i group by same names ? like i have same name in 5 lang , "Pedro Aguirre Cerda" so if i have more than name "identecal" we chose the home lang. [07:00:50] the main name lang. [07:01:27] you probably can... yes, a lot of names will repeat [07:01:29] so in this case we will shrink the amount of same data [07:01:48] How , you are the master :) [07:01:53] basically if you've got a person names John Smith, then all latin script languages would have the same name [07:03:01] how can we say : give me the name if it is not the same of the given_nameLabel [07:03:27] got the idea ? [07:03:51] you could try something like this: http://tinyurl.com/y7ycgk4n [07:04:08] let me study it please [07:04:10] but you lose the language [07:04:28] or do post-processing with some script on the client end [07:05:00] maybe some sparql too can do, but can't think about something right now, too late for it here :) [07:08:22] LOL Thank so much i am laring , but why fnameLabel is empty , [07:09:31] https://www.wikidata.org/wiki/Q855 Joseph Stalin [07:12:31] (str(?given_nameLabel) as ?label) what is this mean :) [11:43:29] Translation admins: could someone please mark https://www.wikidata.org/wiki/Wikidata:Property_proposal/Proposal_preload for translation? [12:50:43] i am loading my file ./loadData.sh -n wdq -d /download/wiki/service-0.3.0/data/split -s , the wikidata.jnl and i am at wikidump-000000143.ttl.gz of wikidump-000000507.ttl.gz and the curent wikidata.jnl is 120GB , what is the expected final size :( [12:51:30] and is there any ready to download wikidata.jnl ? [14:17:40] DanielK_WMDE: I even just found an admin marking historical statements as deprecated :( [14:24:47] nikki: *sigh* [14:25:04] yeah :( [14:39:30] :( [18:22:55] Hi, ChAndrew :) [18:23:21] ChAndrew: Would you mark https://www.wikidata.org/wiki/Wikidata:Property_proposal/Proposal_preload for translation, please? O:) [18:26:06] If someone wants to win the race, they can too... [18:28:38] sjoerddebruin, would you? [18:29:36] don't forget to adjust the translate comments [18:29:39] multiple now [18:30:09] Hey :) [18:30:25] hi :) [18:30:44] I really don't want people to translate that several times [18:31:01] A single string would be enough [18:31:06] Is that possible? [18:31:31] i don't know, it doesn't proceed [18:32:43] So should I write Q1, Q2, Q3, etc., and value1, value2, value3, etc.? [18:33:53] I don't know, not that good in the translate extension [18:35:28] We could remove the repeated examples and just give one for example 1, what do you think? [18:35:53] I'm not really focused anymore, long day. [18:36:56] Time to rest then :) [18:37:19] "We could remove the repeated examples and just give one for example 1" → I'll do this [18:38:45] {{done}} [20:43:13] IS Thiemo around? I'm having some CI issues due to a phpunit test in WikibaseQuality [20:43:37] guessing not because of the time [20:44:43] have opened https://phabricator.wikimedia.org/T197038 if anyone knows that extension and can take a look! [20:51:55] So I know a little sparql and I'm a little surprised by the way wikidata sparql queries have ?Label terms [20:52:30] Its weird that you can have a query which collects for instance ?itemLabel but never mentions an ?itemLabel in the query [20:52:38] Can someone point me to relevant documentation? [20:55:17] those come from the label service [20:55:19] see https://www.mediawiki.org/wiki/Wikidata_Query_Service/User_Manual#Label_service [20:57:33] Thanks! [21:16:58] i am loading my file ./loadData.sh -n wdq -d /download/wiki/service-0.3.0/data/split -s , i am at wikidump-000000230.ttl.gz of wikidump-000000507.ttl.gz and the curent wikidata.jnl is 193G, what is the expected final size :( [21:17:22] and is there any ready to download new wikidata.jnl [21:22:11] any advice ? [23:10:14] beshoo: I don't think we have a ready wikidata jnl [23:11:06] The one we have at Wikimedia (for query.wikidata.org) currently has around 505G [23:11:10] * 550 [23:51:55] WAW , larger than my HDD [23:53:35] do think mine will reach tht number