[00:05:48] Reedy: you around? [00:07:27] perhaps someone else knows [00:07:39] I've been working on some tools related to new page patrolling [00:08:03] by searching through the logging table [00:08:24] and I've noticed that the log_params field of that table seems to have switched formats drastically, at some point in the last year [00:08:41] the new format looks like: a:3:{s:8:"4::curid";i:10091973;s:9:"5::previd";i:10091972;s:7:"6::auto";i:1;} [00:08:55] and the old format is just a number, like: 10074107 [00:10:00] where's the best place for someone to learn how to decode the information in both of those formats? I'm looking for information on how to determine if the patrol was automatic or manual [00:10:36] I assume that, with the new format, I'm just looking for "6::auto";i:1 or "6::auto";i:0 [00:10:45] but for the old format, I have no idea [00:11:32] looks like Reedy had some involvement with it at http://www.mediawiki.org/wiki/Special:Code/MediaWiki/112374#c31649 [00:11:46] The new format is just json [00:12:16] is it? [00:12:24] strangest json I've ever seen [00:12:24] yes [00:12:28] Why? [00:12:33] oh, sorry [00:12:39] it's serialised php isn't it? [00:12:55] might be, i'm not a php guy [00:13:24] the new format seems easy enough, but the old format is another story [00:14:08] http://p.defau.lt/?pdbVE_Wps2G7Fm4ojn0O0A [00:14:26] presumably the old ones are just the curid then [00:14:42] ahh, got it. that makes sense [00:14:51] so with the old format, there's no way to distinguish between autopatrols and manual patrols? [00:15:25] seemingly not [00:15:37] My comment on that revision shows to zeros on new lines [00:15:53] but that's no help to anyone [00:16:10] hmmm [00:17:04] yeah, looks like you're right, those are curid's [00:17:50] Though [00:17:50] v [00:17:52] list( $vals2['cur'], $vals2['prev'], $vals2['auto'] ) = $params; [00:18:04] log_params: 10074107 [00:18:05] 0 [00:18:05] 0 [00:18:07] i could swear that MW could distinguish between auto and manual patrols though.. the switch seems to have happened around feb/march 2012 [00:18:16] ohhhhhhhhhh [00:18:19] missed those [00:18:58] so the third number is auto [00:19:04] that's what I needed! [00:19:20] yeah, would look like that's the case [00:19:36] thanks for digging into that for me [00:19:39] appreciate it [00:36:57] susan is being stupid [00:37:00] the usual [00:37:47] domas: <3 [00:39:47] Susan: lollish, after you said that bug about edit conflicts doesn't interest you :p https://en.wikipedia.org/w/index.php?title=Wikipedia:Requests_for_comment/Article_feedback&curid=38126110&diff=534738832&oldid=534738771 [00:40:26] Nemo_bis: Which bug about edit conflicts? [00:40:30] I only saw one about undo. [00:41:04] susan: that was my point, that it is an anti-pattern [00:41:18] slapping cache on top of expensive data operations and expecting everything to work fine afterwards [00:41:42] Susan: it's the same. [00:41:55] Nemo_bis: I don't see how. [00:42:13] Susan: don't they all use diff3? [00:42:20] Anyway, to bed now [00:42:20] Who's they? [00:42:28] undo and edit conflicts [00:43:10] No. [00:43:37] domas: I guess we could add a field to store the counts. It's probably overkill, though. [00:44:05] is it? [00:44:16] domas: There are about 100 pages with over 50,000 revisions on en.wiki, it looks like. [00:44:23] Yes, WP:AN/I is an edge case. [00:44:37] You could just disable the counts for that page. :-) [00:44:55] I'm looking at . [00:44:56] so nice that I can cut and paste and don't have to write answers, e.g. [00:44:56] [16:36:56] susan is being stupid [00:44:57] [16:36:59] the usual [00:45:19] How expensive is a COUNT(*) on 25,000 rows? [00:45:57] multiple I/Os, megabyte of cache pressure, etc [00:46:11] * AaronSchulz waits for those SSDs [00:46:13] I guess if you overprovision hardware like crazy (or like wikipedia), then nobody cares [00:46:39] Google doesn't overprovision? [00:46:41] Twitter doesn't? [00:46:58] domas: https://gerrit.wikimedia.org/r/#/c/41196/ That one'll make you cry :P [00:47:13] The vast, vast majority of pages have very few revisions. [00:47:19] Probably somewhere around 99%. [00:47:25] Across Wikimedia wikis. [00:47:36] I'm not sure an extra DB field is worth it. [00:47:50] If it's saving a COUNT(*) on 5 rows. [00:48:21] google is exceptionally good at their resource management, I think [00:48:31] Susan: Actually why do you want to know the revision count and in which cases? [00:48:37] hoo: It's interesting information. [00:48:49] In which cases what? I don't follow the second question. [00:49:05] susan: well, how many have 1000 rows? [00:49:11] thats still multiple i/os and multiple pages in cache [00:49:26] how many have 100 rows? [00:49:26] ditto [00:49:33] now, multiple cumulative effect [00:49:41] domas: What would you say is a reasonable number to justify COUNT(*)? [00:49:47] I'm not sure that question is phrased well. [00:49:50] Susan: Why can't you use the TS DBs for that [00:49:53] COUNT(*) should never be done in web environments [00:49:58] or well... how often do you want to do this? [00:50:15] hoo: It's part of the info action in MediaWiki core. [00:50:17] domas: well if you know there is low upper bound it's fine [00:50:24] just like tiny filesorts [00:50:24] hoo: https://en.wikipedia.org/w/index.php?title=Sean_Combs&action=info [00:50:37] aaronschulz: meh, still pulls in more data than it needs [00:50:44] hello? [00:50:53] domas: :) [00:50:56] what is better, 10 counts that count 1000000 rows, or 1000000 counts that count 10 rows? [00:51:08] eric_lee: HI!II?!?! [00:51:23] doms [00:51:24] domas [00:51:26] domas: What is your recommendation? To store the counts in a DB field? [00:51:30] yup [00:51:32] Okay. [00:51:34] domas: on an ssd or hdd? :D [00:51:34] I'll file a bug. [00:51:35] not reco [00:51:37] but question [00:51:46] Susan: That'll probably never happen [00:51:47] * hoo hides [00:51:48] uh [00:51:53] what the counts are for indexes all in ram [00:52:02] hoo: What will never happen? Storing the counts? [00:52:03] domas: so I guess, "it depends" [00:52:10] Susan: Probably, yes [00:52:29] hoo: Well, I think (I know) people are interested in this information. [00:52:30] well, realistically the first one would suck for some other reasons [00:52:56] like tying row purging and all sorts of nonsense [00:53:00] Susan: And I know people actually working on nasty data analysis stuff [00:53:01] *tying up [00:53:01] and still that doesn't work the way they want it [00:53:15] (if someone got a spare computing cluster ping me) [00:53:33] hoo: I'm not sure what you're saying. [00:53:49] hoo: The action can be made more efficient, though I kind of hate stored counts. [00:53:54] Because they're invariably wrong. [00:54:38] they *are* always wrong, that's true :) [00:55:00] Doing that sounds like a nasty hack we should only do if REALLY needed desperatly [00:55:15] mysql> SELECT COUNT(*) FROM revision WHERE rev_page = 5149102; [00:55:16] +----------+ [00:55:16] | COUNT(*) | [00:55:16] +----------+ [00:55:16] | 214211 | [00:55:16] +----------+ [00:55:17] 1 row in set (15.94 sec) [00:55:23] It's not just you! http://upload.wikimedia.org looks down from here. [00:55:24] sorry for the flood, bah [00:55:24] ? [00:55:39] getting that kind of thing right and fast is about as easy as riding a unicycle through a wind tunnel while juggling handkerchiefs [00:56:04] My point should be: Creating that on demand in some rare cases IMO is fine [00:56:11] * was supposed to be [00:56:39] If we introduce a caching field for like everything, we're going to have a littly messy DB scheme [00:56:52] wodim: Seems fine to me. [00:57:34] yes, fine for some people... but not for me or for google appspot [00:57:34] hoo: Domas doesn't like COUNT(*) for Web environments. [00:57:40] so it's something in between i suppose :p [00:57:50] wodim: Do you know how to trace route? [00:57:58] yes i'm on it [00:58:03] Okay. [00:58:09] could use triggers [00:58:13] then the stored counts will be right [00:58:15] Susan: Well... we could of course get the whole data and count it in PHP, I'm not sure that's better though [00:58:36] TimStarling: MySQL triggers? Or you mean just += 1 to the count after every edit (or watch)? [00:58:39] Susan: oh, now it works fine, so... sorry for the noise [00:58:41] TimStarling: Do we really need a new DB field for that? If we do, I want some DB things myself [00:58:59] What DB things do you want? [00:59:00] Starting with some indexes [00:59:05] ah, yeah, triggers, a good way to anchor yourself to a particular rdbms [00:59:09] Heh. [00:59:20] they are quite powerful in pg, if you like the pg trigger language :) [00:59:30] hoo: you're saying we should just remove the feature? [00:59:30] I have a great deal of respect for Domas. But I'm still not convinced that COUNT(*)s are evil. [00:59:46] Susan: you are assuming he is literal [00:59:46] I am fine with removing the feature [01:00:00] domas always exaggerates, isn't that right domas? [01:00:14] TimStarling: You're fine with re-disabling the info action, you mean? [01:00:29] TimStarling: I'd like to see a focus on feature that look at the last X months of activity [01:00:29] no, just removing that particular data item from it [01:00:34] It took me something like seven years to get it re-enabled. [01:00:37] TimStarling: If it affects production, probably, yes [01:00:47] tends to be useful and have much better bounding performance [01:00:55] There are multiple COUNT(*)s in it currently. [01:01:07] Even I introduced one [01:01:23] Is there evidence that it's affecting the production cluster? [01:01:30] And I don't feel very guilty [01:01:38] Domas will fix that. [01:01:59] really, it could just not show some stuff if the estimated count is too high [01:02:02] some people will never understand that performance engineering is a combination of lots of 0.1%s [01:02:09] and it would work fine for 99.9% of pages [01:02:17] Heh. [01:02:22] we have had namespace selectors on RC and contributions pages for some years, despite domas complaining, and the cluster somehow survives [01:02:33] and the filesort for newbie contribs [01:02:33] except that from time to time it breaks shit [01:02:36] and we have to block IPs [01:02:37] and whatnot [01:02:40] domas: Well... the only difference over here is that we don't sum up those 0.1s but multiply 'em... [01:03:04] you multiply 1.01s [01:03:04] and usually scanning 50k rows or so doesn't take long [01:03:05] TimStarling: Right. That was mostly my point much earlier above. [01:03:06] not 0.1s [01:03:15] There are a few dozen ways to cause a large table scan currently. [01:03:18] and I wouldn't expect many pages to have more watchers than that [01:03:39] TimStarling: https://en.wikipedia.org/wiki/Wikipedia:Database_reports/Most-watched_pages [01:03:48] The highest is 76,000-ish. [01:03:54] And it's a steep drop-off. [01:03:55] when you need to read megabytes of data to serve a single digit [01:03:58] there's something wrong at that [01:03:58] domas: No... that's only a problem on rare cases (let's say 0.1) and that again is only a rare fraction of total page requests (let's again say 0.1), ... [01:04:02] but I guess wikimedia has moneys [01:04:11] TimStarling: does watchlist filesort? [01:04:18] domas: The money is much better spent on larger and more robust comments boxes? [01:04:49] AaronSchulz: yes, I think so [01:04:49] I mean, if you're going to raise $35 million, I think providing interesting information to readers and editors is better than most other uses of the money. [01:04:59] well the watch join would be LEFT, so nevermind [01:04:59] hoo: my point was there're lots of small things that are "not big problem" [01:05:10] sometimes it is easier to throw hardware at the problem and ignore outliers [01:05:11] if could be deferred till after the WHERE (god I'd hope that is what would happen) [01:05:16] as long as you don't monitor bad behavior [01:05:19] there is no bad behavior [01:05:21] TimStarling: I think Domas cares less about number of watchers than he does about number of revisions. [01:05:21] and everyone is happy [01:05:22] and cluster is up [01:05:27] * AaronSchulz still would not be comfortable with that [01:05:48] Can ?action=info be profiled via the Web UI? [01:05:48] I mentioned it is anti-pattern [01:05:56] It'd be interesting to see which queries it's running. [01:06:03] imagine that number being on every page, and it being beyond memcached [01:06:15] so it is very cheap when it is in the cache [01:06:19] Analytics is expensive; film at 11. ;-) [01:06:25] but when it is not in the cache, you end up with stampede and thousands of threads all counting same rows [01:06:27] tadaaaa [01:06:40] ddsh ... sudo service memcached restart [01:06:42] domas: Well, there are more nasty things which we yet do [01:06:49] and we will probably do in the future [01:06:53] "the domas performance test" [01:07:16] as someone mentioned, yeh, just buy lots of ssds [01:07:19] and it will be all fine [01:07:32] and fuck-all RAM for the innodb buffer [01:07:41] You could put the counts behind $wgMiserMode. [01:07:44] domas: So in you're opinion it's actually good to have those user_text fields and stuff? [01:07:56] But it feels pretty cruel to punish 700+ wikis because a few are popular. [01:09:16] domas does not care about cruelty to users, only servers [01:09:18] There are a few COUNT(*)s on recent changes. [01:09:25] recentchanges [01:09:26] why is enwiki punished with high cost-per-user of small wikis? [01:09:27] :) [01:09:43] Good night... I'm out, I don't really care about that counter anyway [01:09:56] just put the query in the watchlist group [01:09:57] ok, bye :) [01:10:13] anyway, you can do whatever you want [01:10:14] then it will be guaranteed to be in RAM, and then we are only talking about some tens of milliseconds of CPU time [01:10:21] I'm just telling what are good practices and what are the costs [01:10:38] I'm not sure storing counts for the recentchanges counts would make sense. [01:10:46] talking about overprovisioning hehe [01:10:55] For revision, probably makes sense. For watchlist, it definitely seems like overkill. [01:11:07] well [01:11:10] counts also deadlock easily [01:11:21] at least the way we do them [01:11:32] http://ganglia.wikimedia.org/latest/graph.php?h=db1043.eqiad.wmnet&m=cpu_report&r=day&s=by%20name&hc=4&mc=2&st=1359076267&g=cpu_report&z=medium&c=MySQL%20eqiad [01:11:42] you see that we have some spare CPU time on the watchlist server in eqiad [01:11:42] it needs a password [01:11:43] in any case, it's serializing unrelated thing [01:11:43] heh [01:11:43] For watchlist, after the top 1000 pages on en.wiki, we're talking about 650 rows or fewer, it looks like. [01:11:51] *things [01:12:02] domas: I imagine you're allowed to have the password. [01:12:05] I'm not. :-( [01:12:09] TimStarling: "some" [01:12:09] I had it somewhere [01:12:13] heh [01:12:27] timstarling: well, thats what I say, one can have whatever he wants as long as he overprovisions hardware [01:12:34] * domas is not relevant [01:12:42] irrelevant * [01:13:00] * domas was double-plus-relevant when we had no money to spend on hardware [01:13:10] :) [01:13:35] Of all the complaints I can list about the Wikimedia Foundation, I really can't add "buying too many servers" to the list. [01:13:49] I'm serious, instead of wfGetDB(DB_SLAVE)->select(...) it should be wfGetDB(DB_SLAVE, 'watchlist')->select(...) [01:14:26] I don't think anyone thought you were kidding. You should file a bug. :-) [01:14:26] someone here is serious? [01:14:30] Susan: *shrug*, I never asked for your list, nor I thought it was credible [01:14:52] you can wave your hands as much as you want [01:15:03] it is easy to get metrics on crap like that [01:15:14] I'm just saying that an argument that tries to make overprovisioning the enemy is kind of weak. [01:15:38] how much do you donate to WMF? [01:15:40] in cash? [01:15:47] Not much. [01:15:57] I haven't in years. [01:15:57] how much do your friends donate? [01:15:57] More than me! [01:16:22] how often do you have to do capacity management of any sorts? [01:16:27] I had a friend from work tell me he asked for a donation to Wikipedia for a Christmas gift. [01:16:38] I thought that was sweet. [01:16:46] Not often, thankfully. [01:16:48] I had friends, who were major donors, who stopped being major donors [01:16:59] I used to donate [01:17:01] I stopped too [01:17:02] heh [01:17:11] Sure. [01:17:13] world didn't stop [01:17:16] there's new generation of donors [01:17:21] Right. [01:17:25] and yet you donate a lot of your precious time in this discussion [01:17:27] :-) [01:17:28] And your time and expertise are much more valuable than cash. [01:17:50] I donate quite a lot. [01:17:53] you don't seem to care much about it anyway, ha ha [01:18:06] paravoid: yup [01:18:22] that's not true! people care enough to play along [01:18:30] I care. I just feel your argument is weak. [01:18:52] which one? [01:18:58] that one can solve problems by overprovisioning? [01:19:25] If you could say "X feature costs Y" and Y is understandable and relatable, that'd be a stronger argument, I think. [01:19:49] Instead, the argument seems to be "the problem can be solved with more servers." To which I say: that's a much better use of money than most other things WMF buys. [01:20:02] I'd take a few more servers over a few more staff any day. [01:20:49] i'd take a new datacenter not in tampa :) [01:20:52] And in the abstract, the Y cost doesn't mean anything to anyone but DB engineers. ;-) You could convert it to dollars. [01:20:54] wikipedia, where anyone can be HR/CFO/VP/CTO/etc at the same time [01:21:37] shutting down the datacentre in tampa would obviously be a bigger win for technical spending than removing the watchers count from action=info [01:22:00] well, there would be some dc to takes its place [01:22:17] *take [01:22:30] TimStarling: never said that is not the case [01:22:37] *shrug*, I didn't yet tell anywhere that anyone should do something [01:22:41] AaronSchulz: not really [01:22:41] There's a lot of waste in any non-profit. But overprovisioning servers doesn't feel like waste to me when a huge portion of the goal of the projects is to serve Web pages. [01:22:50] I said there're tradeoffs, indeed [01:22:52] TimStarling: is that not the plan? [01:22:56] the question is whether or not to have a spare datacentre [01:23:02] in case one fails [01:23:07] Susan: I guess in commercial world people have different perspective, because they have to buy their servers [01:23:19] Wikimedia steals them. [01:23:25] extorts [01:23:27] Heh. [01:23:30] would be correct word [01:23:58] TimStarling: there's a question about that? [01:24:14] I thought pmtpa was being taken offline completely. [01:24:15] maybe a question by tim ;) [01:24:27] I thought we were all agreeing that we need a secondary site [01:24:47] A secondary site or a secondary data center? [01:25:12] and that pmtpa has a great deal of technical debt accumulated from being evolved all these years, so it might make sense to build a new one somewhere [01:25:27] and west coast/SF was one of the proposals [01:25:29] I believe the plan is to build a caching site in SF. [01:25:35] But not a data center, per se. [01:26:09] no, the short-term plan is to build a caching center in SF (ulsfo), and possibly build a new DC somewhere in SF (but in ulsfo) [01:26:14] but not* [01:26:38] Oh, I hadn't heard anything about a new DC. [01:26:46] because it's not decided yet [01:26:54] it's just an idea that's been floating. [01:27:02] Sometimes there are discussions before decisions. [01:27:11] Or there used to be... [01:27:13] yes, and we haven't had that discussion yet. [01:27:51] What happened to yaseo and the knams? Are they both still around? [01:28:17] Guess yaseo is gone. [01:28:28] knams still seems to be around. [01:29:24] there are still knsq servers [01:29:31] it's esams now [01:29:37] evoswitch not kennisnet [01:30:06] Should http://wikitech.wikimedia.org/view/Kennisnet_cluster be marked historical then? [01:30:08] we still have some networking equipment in kennisnet but most of the amsterdam stuff is in the evoswitch DC [01:30:14] yea, we have knsq and amssq [01:30:37] sq? [01:30:40] squid [01:30:49] Ah. [01:30:55] Susan: no, it just shows a couple of bits of networking equipment [01:31:01] that's what is there [01:31:16] http://wikitech.wikimedia.org/view/Amsterdam_cluster has a note about how it's allegedly "Enqueued for move to the new wikitech". [01:31:32] I don't think the new wikitech idea ever went anywhere. [01:31:51] http://wikitech.wikimedia.org/view/Special:WhatLinksHere/Template:MoveToNewWiki [01:31:55] Hm. [01:36:07] Krinkle|detached: solution for your mailman question. login as list admin and go to "Edit the public HTML pages and text files", then edit the HTML of your "Subscribe results page" and remove or deactivate the form [03:51:32] TimStarling, Susan: don't you think we should just be a bit ruthless and delete 'historical' pages from Wikitech? Wikitech is less useful to the extent that it is a museum. [03:52:15] I don't think the old pages are in the way. [03:53:26] ori-l: should we delete articles from wikipedia just because they are historical? [03:54:13] p858snake|brb: the primary thing wikitech and wikipedia have in common is the software that powers them; I don't see how that isn't a straw man argument. [03:54:13] That's not really comparable. [03:55:05] I think it's somewhat comparable to having a medical manual that contains historic treatments for various conditions alongside contemporary, evidence-based ones. [03:55:40] The 'historical' template doesn't affect what comes up for search autocompletion or results, or what appears in category lists, or the appearance of wikilinks to the article [03:56:30] So there are a lot of views in which historic content makes an appearance without being clearly marked as such, and that diminishes from wikitech's usefulness as a quick diagnostic manual, and also makes it more difficult to assess the adequacy of coverage of various topics. [03:56:57] Do you have an example? [03:57:16] Sure, hang on a moment. [03:57:59] https://en.wikipedia.org/wiki/Wikipedia_talk:Miscellany_for_deletion/Wikipedia:Esperanza [03:58:16] "Messedrockerfy" is an interesting concept. [03:58:23] Not sure if you're familiar. It's kind of obscure these days. [03:59:38] Well, let's start with the simple and obvious: [03:59:41] http://wikitech.wikimedia.org/view/Main_Page [03:59:46] Which ones are stale? [03:59:53] Can you tell? [04:00:10] well of course historical content won't appear "historical" unless someone spends the time to mark them as such, its kinda a biased argument to make [04:00:57] ori-l: Right. That index just needs to be updated. [04:01:15] I don't have any issue with removing categories or de-linking irrelevant material. [04:01:49] If you start deleting content, I doubt you'll ever stop. The whole site is pretty bad. [04:01:51] right, but the preservationist attitude counsels treating content as precious [04:03:26] Krinkle doesn't seem to think so. [04:03:34] http://wikitech.wikimedia.org/history/Collected_Status [04:03:50] Let me put it another way: instead of deleting pages, why not fix them? :-) [04:04:51] If you delete the pages, you'll have red links that you'll still need to remove. [04:05:00] Why not just remove the links and leave the pages? [04:06:12] I've turned into a bit of a wiki-archaeologist, so it won't be easy to convince me to delete some of this stuff. Sometimes I reference it. [04:09:21] Sorry to be rude, my son woke up, so I have to go afk for a bit. [04:47:31] * Jasper_Deng would think http://en.wikipedia.org/wiki/Wikipedia:Village_pump_%28technical%29#content_not_refreshing_from_Hungary.2C_using_ipv6. 's premise is invalid [05:21:24] TimStarling: So duh tells me that the combination of a lack of interwiki transclusion and the support for global images (from Commons) has lead users to create global user pages created out of images: https://commons.wikimedia.org/wiki/File:Snowolf_gluserpage.png [05:21:30] I thought you might appreciate that. [05:22:41] It's a smart idea imo [05:34:04] smart, for those who want the hassle of creating a new image for each edit to the userpage [05:38:38] I don't understand the linkability. Do you use an imagemap with it, then? [05:40:05] rschen did [05:40:12] Snowolf used |link= [05:40:34] https://de.wikipedia.org/w/index.php?title=Benutzer:Snowolf&action=edit [07:54:16] Susan: I don't use an image map because to update it I'd have to change the image map on all wikis; I merely have a link to meta and I update the image on commons [07:54:29] Jasper_Deng_away: easier to update one image than 732 pages :) [07:54:53] Late at night, I hear the mice. [07:57:10] Have you tried restarting? [07:57:38] Not yet. [07:57:43] I'm at 14 days, I think. [07:57:48] 16 days. [12:38:41] can someone finally approve this https://gerrit.wikimedia.org/r/#/c/43843/ [12:38:49] I am waiting like a week for that or more [12:39:17] mutante, paravoid? [12:39:47] I am really tired of these 290% of ram free [12:39:50] hey [12:39:57] mutante has relocated to SF since months ago [12:40:00] hi [12:40:07] ok... but ur here! [12:40:09] so I wouldn't expect him to answer at this hour of the day :) [12:41:03] I don't understand your change [12:41:15] the previous check produces non-sense numbers [12:41:23] OK: 302% free memory [12:41:38] http://nagios.wmflabs.org/nagios3 [12:43:02] the names are crap [12:43:07] indeed [12:43:10] "full" [12:43:22] $full is the how much of ram is full [12:43:39] $free is how much is free, $free + $full is total ram [12:44:13] s/full/used/ [12:44:26] incidentally how "free" calls it [12:44:28] feel free to change it... [12:44:59] this check already exist and is used all over labs, I am not changing the names, I am fixing math [13:05:30] done [16:48:53] could someone take a look at the special pages jobs? [16:48:59] they havent been running for 5 days [16:49:07] andre__: ^^ [16:49:15] malafaya: is this across all of the WMF sites? [16:49:17] AaronSchulz: ^^ [16:49:19] jobqueue problem [16:49:21] filed already. [16:50:09] however hidden in https://bugzilla.wikimedia.org/show_bug.cgi?id=15434#c54 [16:50:10] andre__: this sounds like it might be Immediate [16:50:49] after finishing my current stuff I need to check if this is the case also on other pages than pt.wikt, if it is, yes. [16:51:26] okay, thank you andre__ [16:52:27] i would dare saying it across all wikis [16:52:35] for some it's 6 days [16:52:39] malafaya, a concrete link in that bug report would be welcome. [16:52:57] andre__, you mean in my comment? [16:53:16] It's not immediate [16:53:21] malafaya: yes. Make me click only once, imagine that I'm lazy :) [16:53:37] Though, I think that's the wrong bug to comment on [16:53:58] Done :) [16:54:14] The ones run regularily aren't "disabled special pages" [16:54:37] * sumanah leaves this to the professionals :-) [16:54:52] The question is, I guess, was it the datacentre migration that broke them? [16:56:53] malafaya: That's not the correct bug [16:59:06] Reedy, which one is it? That one is still active and recent [16:59:20] Open a new one [16:59:26] That bug is about disabled special pages [16:59:30] which are currently NEVER run [16:59:34] I'll open a new one. [16:59:35] maybe you mean that as the underlying cause is probably different a ... [17:00:05] bug 44348 [17:00:05] Those disabled special pages aren't supposed to be rununing [17:00:16] the request is to make them run (for most wikis) [17:01:12] you're right. i had it in my CC list and thought it was the right one [17:01:21] but i see andre__ already created a new one :) [17:01:23] anyway, it's now split into bug 44348. [17:01:24] thanks, andre__1 [17:01:31] happens, let's try to track down the new one. [17:01:33] np [17:02:50] Should be easy enough to confirm.. [17:04:04] I thought we logged that.. [17:05:20] /home/wikipedia/logs/norotate/updateSpecialPages.log [17:05:22] They didn't move [17:06:24] andre__: malafaya It's fine on any small wikis [17:06:40] On "non small" wikis [17:06:50] It's dying at Mostcategories on frwiki [17:08:03] The question now, is why [17:08:40] seems so [17:08:43] https://so.wiktionary.org/wiki/Special:WantedCategories is up to date [17:08:51] not https://fr.wikipedia.org/wiki/Sp%C3%A9cial:Cat%C3%A9gories_demand%C3%A9es [17:09:13] not the wikidata ones.... [17:09:34] http://www.wikidata.org/wiki/Special:WithoutInterwiki which should die! [17:09:48] aude: Alter InitialiseSettings.php to disable it ;) [17:10:02] how? [17:10:18] we can disable the updating but i only see a hook for actually eliminating special pages [17:10:26] <^demon> aude: Maybe Wikibase should disable WithoutInterwiki on wikis that have the extension installed? [17:10:32] <^demon> You could use the special page init list hook. [17:10:38] aude: wgDisableQueryPageUpdate [17:10:45] ^demon: i have a patch for that to submit [17:10:54] Reedy: that doesn't eliminate them [17:11:04] Stops them running though? [17:11:10] they still appear in the list and "without interwiki" is very confusing [17:11:17] Reedy: it does [17:11:58] * aude looks at the bug :) [17:12:11] http://p.defau.lt/?yHxRTU2FWJcTN8GwkG_w2A [17:12:24] Why do we run the update-special-pages on small wikis twice? :/ [17:12:45] * Reedy adds to his fixme list [17:13:02] <^demon> Heh, run on all, and run on small :p [17:13:10] heh [17:13:15] <^demon> Now, if it was run big and run small, ok. [17:13:17] <^demon> :p [17:14:00] <^demon> That's the only usage of small.dblist in puppet. [17:14:05] <^demon> Maybe we should just get rid of it. [17:14:49] That cronjob? [17:14:53] <^demon> Yeah. [17:14:54] That's what I thinking [17:15:47] <^demon> ensure => absent and git rm the shell script. [17:21:07] what about running "all" but not fr.wiki now? :) [17:22:58] malafaya: It doesn't fix the problem though [17:23:11] I've got to copy and modify the script. Create a dblist without frwiki... ;) [17:23:22] Currently I'm running only frwiki, to see if any errors spew [17:23:40] Actually [17:23:47] I wonder if frwiki has gotten "too big" [17:23:59] enwiki has Mostcategories disabled [17:24:08] hmm, possibly [17:24:19] i thought all wikis had the same special pages running [17:24:33] so it's possible that it got as big as enwiki [17:24:43] i mean, passed that threshold [17:24:44] Mostly, not completely though [17:24:45] http://p.defau.lt/?a99Z1OaRhq7dcsN2fO8yTg [17:25:06] Which is a hint I need to make the dblists available as groups (already got a TODO bug for that) [17:26:03] frwiki does have ~25% more pages than dewiki [17:26:21] eh, just delete everything beginning with a -- then we'll both get more people editing and reduce the size [17:35:33] Reedy, did it finish for frwiki? [17:35:51] No [17:35:56] I gave up [17:45:35] :( [17:46:41] Reedy: so bugs should be filed about that special page not being able to be updated? [17:46:49] Nemo_bis: Depends what the bug is [17:47:02] malafaya was right that they were broken [17:47:09] Reedy: it would be a blocker of the "re-enable everywhere" [17:47:50] https://gerrit.wikimedia.org/r/#/c/33713/ runs one page at a time, so fr.wiki failing wouldn't affect the other wikis. Still, it wouldn't work completely. [17:48:12] We're not talking about disabled special pages [17:48:49] Reedy: I know but that one also runs the 6 which are disabled only on en.wiki (and now fr.wiki), IIRC? [17:49:29] 6 are disabled everywhere [17:49:34] Oh, not, I left those for later. [17:49:44] 12 on enwiki and frwiki [17:49:48] Yes, that patch is only about those 6 (for now). [17:50:23] But how to investigate how intensive a special page is and whether it will complete or not? [17:50:58] run it and see [17:51:03] probably the best wayt [17:51:08] Is there something useful about the just disabled 6 on updateSpecialPages.log [17:51:32] I guess the way that the script runs when frwiki took too long and died (we had this with FR not long ago), it stops the rest of the script executing [17:51:47] Hi guys, What broke? [17:51:48] malafaya: ptwiki has been run manually. should be up to date now [17:53:12] Reeady, sorry, it was ptWIKT :) [17:53:16] *Reedy [17:53:23] * Reedy glares at malafaya [17:53:44] Ok but why does it die at Mostcategories :/ [17:53:50] that's what i linked to in the bug [17:54:05] Nemo_bis: it takes too long [17:54:15] I suspect it gets killed as a long running process [17:55:04] Reedy: where does this happen? [17:55:13] What do you mean where? [17:55:13] Poor lonely pages, they will be even more lonely now [17:55:19] What kills them [17:55:43] or it OOMs [17:56:01] # Disable all the query pages that take more than about 15 minutes to update [17:56:08] yes but that was ages ago [17:56:13] malafaya: also done [17:56:26] Reedy, I confirm it's updated. Thanks [17:56:30] is it possible to see if it OOMs? [17:56:44] Need to see if it's in the error long somewhere [17:57:14] flock -n /var/lock/update-special-pages /usr/local/bin/update-special-pages > /home/wikipedia/logs/norotate/updateSpecialPages.log 2>&1 [17:57:26] The errors should've been written to the log file too [18:00:11] * Reedy runs it in a screen session this time [18:04:34] Too bad http://stats.wikimedia.org/EN/TablesDatabaseLinks.htm has not been updated in ages. :[ We know that fr.wiki has 50 % more categories than de.wiki, though. [18:04:51] Weird [18:05:01] It completed in less than 5 minutes this time [18:05:15] Mostcategories got 5000 rows in 4m 6.28s [18:06:57] So, where are they supposed to be running from [18:07:38] <^demon> hume? [18:08:07] I wasn't sure if they'd been moved to a similar host in eqiad [18:13:08] <^demon> Oh, unsure. [18:13:11] Maybe it dies only if there's lag and when run all together the special page updates lag the DB? [18:14:01] It seems there's only hume with the cronjob entries, so i presume it's only on hume [18:31:56] How can job queue on mediawiki.org be so high? :/ [18:32:37] Not being run? [18:32:46] Commons very unhappy [18:32:58] en.wiki mildly unhappy (it got used at it I guess) [18:32:59] reedy@fenari:/home/wikipedia/common$ mwscript showJobs.php mediawikiwiki [18:33:00] 233 [18:33:34] https://gdash.wikimedia.org/dashboards/jobq/ is acting weird btw [18:34:29] It took a day off a week ago and is very low for last two days [18:39:15] Nemo_bis: The job runners are running, somewhat slowly though [18:39:28] I say that [18:39:37] And then a huge spam [18:41:48] They do look pretty lazy though [18:41:49] Current Load Avg (15, 5, 1m): [18:41:49] 5%, 5%, 5% [18:44:06] apergos: Fancy beating some lazy jobrunners? ;) [18:44:13] Aawon fled just on time [18:44:17] *Aaron [18:44:18] :-D [18:44:27] tomorrow, (or later tonight) I gotta get going actually [18:47:21] DarTar: ping [18:47:49] hey binasher [18:47:49] coming down [19:00:38] ever heard of "rietveld"? - "Code Review, hosted on Google App Engine" [19:00:56] <^demon> Yep, it was the predecessor to Gerrit. [19:01:02] ah, that makes sense then [19:01:10] i saw some reference to it in a Bugzilla extension [19:01:20] <^demon> Gerrit forked from Rietveld over differences in opinion over acls. [19:01:31] aha, thx [19:02:01] <^demon> The fork was *some time ago* and has since been totally rewritten, so it's not really compatible on a technical level anymore. [19:03:37] Bugzilla "MoreBugUrl" extension: [19:03:42]
  • An issue in a Rietveld installation.
  • [19:03:46]
  • A ticket in an RT installation.
  • [19:04:14] https://gerrit.wikimedia.org/r/#/c/44393/2/bugzilla-4.2/extensions/MoreBugUrl/lib/Rietveld.pm [19:04:33] package Bugzilla::Extension::MoreBugUrl::Rietveld; [19:12:43] i thought that couldn't be done without hacking core [19:12:44] interesting [19:13:42] !log Job queue not being run somewhat: no FuzzyBot on [[mw:Communication]] for a hour, en.wiki and Commons unhappy. [19:13:52] Logged the message, Master [19:15:38] 2013-01-25 19:15:16 mw1012 commonswiki: enotifNotify User_talk:Sreejithk2000 editor=MiszaBot editorID=163969 timestamp=20130125191516 summary=Robot: Archiving 2 threads (older than 21d) to [[User talk:Sreejithk2000/Archive 8]]. minorEdit=1 oldid=88800527 watchers=Array pageStatus=changed t=4 good [19:17:07] Commons is over 800k [19:17:12] Enwiki is around 110k [19:18:13] Nemo_bis: Like I said before, it seems the job runners are being pretty lazy [19:18:20] very low overall cpu usage [19:18:30] Total 2762011 [19:18:52] 2.7M [19:18:54] not so bad [19:19:36] notpeter: About? [19:19:43] sup [19:19:54] eqiad job runners seem to be being pretty lazy [19:20:23] it's cool. market forces will get them going again [19:20:29] 5% load average, but there's a backlog of 2.7 milllion jobs [19:20:49] that is kinda alot of jobs, it's true [19:21:09] watching runjobs there is splurges of a load being done [19:22:27] mw1001 periodically shows numerous php processes [19:22:30] yeah [19:22:32] looking at it now [19:25:59] thanks [19:28:05] yeah, man, they're just really lazy [19:28:09] what the fuck [19:28:18] what graphs are you looking at? [19:31:44] http://ganglia.wikimedia.org/latest/?r=hour&cs=&ce=&s=by+name&c=Jobrunners%2520eqiad&tab=m&vn= [19:31:53] Current Load Avg (15, 5, 1m): [19:31:53] 5%, 5%, 5% [19:31:53] Avg Utilization (last hour): [19:31:53] 5% [19:32:11] The week graph looks somewhat off on https://gdash.wikimedia.org/dashboards/jobq/ [19:38:25] skype? https://www.google.com/search?q=skype+credit+card+fraud [19:40:10] mutante: is skype as buggy as MSN? [19:40:18] Oh, right, it's M$ now, makes sense. [19:40:51] heh,yea [19:41:10] http://etherpad.wikimedia.org/WikimediaTelepresence [19:41:54] http://venturebeat.com/2011/06/28/microsoft-scores-patent-for-web-based-spying-technology/ [19:42:21] basically, interphone reloaded? [19:47:27] I wonder if we should add info about "Fatal exception of type MWException" somewhere on Meta or mw.o [19:51:12] Done: https://meta.wikimedia.org/wiki/Fatal_exception_of_type_MWException [19:51:23] mw.o has its manual page already [19:51:51] "Yes, the error message you're seeing is useless"? hah, ok [20:19:50] mutante: why didn't http://it.planet.wikimedia.org/ pick up the last item of http://xmau.com/notiziole/arch/wikipedia.xml ? it looks correct in puppet [21:37:59] Nemo_bis: re.taking a look [21:43:21] Nemo_bis: INFO:planet.runner:Updating feed http://xmau.com/notiziole/arch/wikipedia.xml [21:44:38] Nemo_bis: updating it.planet and it did.. the cron jobs currently just run once a day [21:44:41] mutante: now it appeared, together with a WLM entry [21:44:50] yea, when was it added? [21:44:59] the other entry was yesterday [21:45:08] feeds have been the same for months [21:45:57] ok, there is a cron and a logfile for each planet [21:46:11] and last time it ran it was INFO:planet.runner:Feed http://xmau.com/notiziole/arch/wikipedia.xml unchanged [21:46:24] runs at 0 0 [21:46:42] UTC [21:47:21] ok [21:48:02] mutante: and thanks for the switch of course, it was about time to get rid of the old one [21:48:24] mutante: what do you think about adding planet KDE to the sidebar? [21:48:40] oh yes, it was, wikitech said it was still managed by Brion and Erik:) [21:49:20] heh [21:49:28] Nemo_bis: i was wondering if people liked that i left Planet Apache/Debian/GNOME etc in the sidebar.. if that is ok, then adding KDE is fine too [21:49:54] mutante: nobody complained so I guess it's ok :p [21:50:04] I don't know how they were selected though [21:50:21] just edit ./templates/planet/index.html.tmpl.erb :) [21:50:30] operations/puppet repo [21:50:47] they were a default from old planet [21:50:53] from planetplanet [21:51:08] ah [21:51:12] i just kept it (the Planetarium part) [22:37:34] gn8 folks [23:21:51] is there a reason the site notice on enwiki wont say hidden? [23:25:22] Someone broke it earlier [23:25:30] and screwed up site css [23:49:06] Is that notice really going to stay up forever? [23:49:22] It's displaying for anons... [23:50:01] sigh [23:50:06] trust me we are trying to fix it [23:50:11] also i have 0 control over site notice [23:50:19] (technically with root access i do, but 0 control that won't get me fired)