[02:12:19] kaldari: So it sounds like the X-Frame-Options thing broke IE8 for you huh [02:12:29] yeah [02:12:38] kaldari: I'll send an e-mail to Chris CC you suggesting we soften the restriction to 'SAMEORIGIN' rather than 'DENY' [02:12:52] I guess we could just override $wgApiFrameOptions on Commons as a temp fix [02:12:57] Rather, I'll file a bug [02:13:11] Yes, but I'd rather not do that without approval from the security guy that wrote this code :) [02:13:19] yeah [02:14:12] Too bad WLM-US starts in a few hours [02:17:58] kaldari: Does only IE8 use an iframe, or does every browser submit into an iframe but is IE8 the only one choking on the DENY? What exactly happens here? [02:18:27] they all use an iframe, but only IE9 and lower seem to be choking [02:18:47] it gives a permission denied error [02:18:48] Do they submit a form inside an iframe, or submit a form with target="iframe", or what? [02:19:15] I believe it submits a form inside an iframe [02:21:38] Hmm, looks like in modern browsers it doesn't use an iframe? It doesn't seem to in Firefox [02:21:54] I suppose you could use FileAPI to get the contents of the local file, and something else to post a file upload? [02:23:39] changing $wgApiFrameOptions from 'DENY' to 'SAMEORIGIN' definitely fixes the problem [02:24:21] Yeah [02:24:28] We don't have Chris's phone number on the contact list unfortunately [02:24:40] So I doubt this is going to be fixed before Tuesday [02:25:11] Well that means we'll need to cancel the WLM-US banners until then [02:25:34] Could you selectively show them for supported browsers? [02:25:48] yeah, that's a good idea [02:26:48] I also strongly suspect that the reason IE9 works is not because it treats XFO differently, but because it has FileAPI & binary XHR and therefore doesn't use the iframe [02:27:01] I believe IE9 doesn't work [02:27:04] lemme see... [02:27:29] yeah, this breaks IE9 as well [02:28:02] Right [02:28:44] IE9 is about 10% of traffic BTW [02:29:07] and IE8 + IE9 is about 17% [02:34:04] if it isn't fixed before Tuesday, I know some people are going to be rather upset. There's a lot of media coverage of the event ramping up (besides CentralNotice), and WLM uploads only work through UploadWizard. [02:34:23] Right [02:34:55] Well if you can get a hold of Chris Steipp in a timely fashion and get him to agree, I can fix this for you [02:35:08] But I don't have his contact info [02:35:09] I'll see if I can track him down [02:35:30] You might be able to track him down via RobLa, he's Chris's boss [02:36:02] good idea [02:36:06] I'm about to finish writing the bug report (which CCs him), I'll give you the # when I'm done so you have a reference [02:36:25] for the future I'll suggest we add our security guys cell # to the contact list :P [02:36:30] Yeah srsly [02:37:25] My cell # is on there too, I suggest you grab that too while you're at it. I don't need Chris to deploy anything, I can handle that for you if you like (or you can do it yourself), call me any time [02:39:30] I'm going on a couple bike rides this weekend so I might be up to 2 (Saturday) or 5 (Monday) hours away from my computer [02:41:58] Rob says the security vulnerability this patches is relatively minor - he says we don't need to call Chris, he says it would be fine if we just change $wgApiFrameOptions to 'SAMEORIGIN' on Commons and fix UploadWizard later. [02:42:14] OK [02:42:18] Then I'll do it right now [02:44:17] thanks [02:44:41] are you doing a live hack or a commit? [02:44:48] A commit [02:44:51] Should be all set in 5 mins or so [02:44:57] yay [02:45:04] Want to finish the bug report first so I can reference the bug # [02:50:09] kaldari: All done, see https://bugzilla.wikimedia.org/show_bug.cgi?id=39877 and https://gerrit.wikimedia.org/r/#/c/22290/ [02:50:43] awesome, I'll test in IE9... [02:51:46] seems to be working now [02:52:28] thanks for the help! [02:52:40] 1000 WLM participants thank you :) [02:54:53] Yay [06:42:02] I need help, I'm using the python simplemediawiki api to query article data from wm [06:42:35] Does anyone know specific parameters to use to query the article text only? [06:43:03] I don't want wikitext or anything else. [06:43:22] just the article. Any help would be appreciated [07:02:26] Can I get some help! [07:03:20] I need some help making a proper media wiki api query! [07:03:25] LukeDev: There's no simple way to get "just the text". The API offers wikitext or HTML, your best bet is to grab one of those and try to strip it [07:03:38] (And shouting for help isn't gonna do much at midnight on a Friday night :) ) [07:03:56] forever alone. [07:06:18] Thanks Roan, I just couldn't find the request in the documentation and was surprised that something so essential was missing- I thought I was missing something. [07:06:38] I'll use an nat lang toolkit to process the html. [07:56:06] LukeDev: playing around with en.wikipedia.org/wiki/Special:ApiSandbox helps :) [07:56:26] and if you want text action=parse is what you should be looking for [09:14:17] YuviPanda! I met you at Wikimania [09:14:40] If you' [09:15:00] re still on I've been pulling text however it's a mess [17:59:07] we (turkish projects) seem to be experiencing a problem with dotted and dotless i's [17:59:30] could a dev assist me in developing a solution for this? [21:37:26] Could anyone help me understand the simline timeline? [21:51:44] I need to understand how events are created and tagged when people read create timelines. [21:57:41] LucianNovo: Can you link to some code? I don't know what you mean, but maybe I (or someone else) could help you understand it [21:59:36] Okay so the following timeline was created using json data fetched from wikipedia somehow:http://www.simile-widgets.org/timeline/examples/cubism/cubism.html [22:00:06] as you might see, it took json data: http://www.simile-widgets.org/timeline/examples/cubism/cubism.js [22:00:42] I want to understand how to fetch events [22:01:23] What api do I use for that? [22:01:27] LucianNovo: It looks like this person compiled their own JSON data, they didn't get it from WP itself....I think [22:01:36] really... [22:02:29] I don't see any requests to any domain but simile-widgets.org, so they aren't getting it from the client at least [22:02:49] They might be requesting it on the backend, but I can't really tell because they obfuscate their javascript [22:03:08] yea. I see thatn [22:04:09] Yeah, it's almost certainly a compilation that this person did [22:04:18] Sorry :) [22:06:09] for a timeline article like the following :http://en.wikipedia.org/wiki/Timeline_of_World_War_II_(1939) what mediaWiki query should I preform to get the headlines(ie:October 1939) followed by the paragraph formatted)content underneath? [22:06:32] Any advice? [22:07:32] nvm [22:10:23] @MarkTraceur thanks for the information. Could you tell me if there is anyway to access semantic data from wikipedia articles? [22:13:22] Like what if I wanted(my program) to determine what subject the article is about(an event in history, a person, etc), what would be the best way to access that data? [22:14:44] You'd have to analyse what metadata your given [22:14:49] Categories, infoboxes... [22:14:57] There's currently no abstraction of this sort of information at all [22:15:14] thanks [22:15:50] Are there any good methods I can run to extract this data in the (python)nltk library [22:15:53] ? [22:16:35] LucianNovo: tried dbpedia? [22:16:54] There isn't really any methods.. You can use the API to extract things like links, categories, images etc [22:17:13] @yuvipanda nope [22:17:45] @yuvipanda looking at it now [22:17:50] LucianNovo: http://dbpedia.org/About [22:17:51] yeah [22:19:25] this site seems to have link rot [22:19:39] but it has good data [22:19:45] (I was told) [22:21:42] trying to find a valuable querying method ...