[00:05:21] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [00:33:21] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 00:33:11 UTC 2013 [00:33:21] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [01:05:52] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [01:32:42] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 01:32:41 UTC 2013 [01:32:52] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [02:05:27] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [02:05:50] !log LocalisationUpdate completed (1.22wmf12) at Sun Aug 18 02:05:50 UTC 2013 [02:06:03] Logged the message, Master [02:11:07] PROBLEM - Puppet freshness on cp1063 is CRITICAL: No successful Puppet run in the last 10 hours [02:11:07] PROBLEM - Puppet freshness on terbium is CRITICAL: No successful Puppet run in the last 10 hours [02:13:27] !log LocalisationUpdate completed (1.22wmf13) at Sun Aug 18 02:13:27 UTC 2013 [02:13:38] Logged the message, Master [02:23:36] !log LocalisationUpdate ResourceLoader cache refresh completed at Sun Aug 18 02:23:36 UTC 2013 [02:23:48] Logged the message, Master [02:32:47] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 02:32:41 UTC 2013 [02:33:27] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [02:40:01] (03PS1) 10TTO: Add namespace aliases (shortcuts) for dewikivoyage [operations/mediawiki-config] - 10https://gerrit.wikimedia.org/r/79550 [02:40:28] (03CR) 10TTO: "Just pinging this, since it has fallen off the bottom of the dashboard." [operations/mediawiki-config] - 10https://gerrit.wikimedia.org/r/76277 (owner: 10TTO) [03:06:16] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [03:35:06] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 03:35:02 UTC 2013 [03:35:16] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [04:07:19] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [04:35:49] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 04:35:45 UTC 2013 [04:46:04] PROBLEM - Puppet freshness on zirconium is CRITICAL: No successful Puppet run in the last 10 hours [05:06:18] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [05:33:18] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 05:33:13 UTC 2013 [05:34:18] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [05:55:06] PROBLEM - search indices - check lucene status page on search27 is CRITICAL: HTTP CRITICAL: HTTP/1.1 200 OK - pattern found - 327 bytes in 0.054 second response time [06:06:08] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [06:32:48] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 06:32:38 UTC 2013 [06:33:08] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [06:53:02] Hi, db1026 is still super lagged. https://www.wikidata.org/w/api.php?action=query&meta=siteinfo&siprop=dbrepllag&sishowalldb= [07:00:26] PROBLEM - search indices - check lucene status page on search1009 is CRITICAL: HTTP CRITICAL: HTTP/1.1 200 OK - pattern found - 345 bytes in 0.002 second response time [07:05:48] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [07:07:58] legoktm, looking [07:08:12] MaxSem: thanks. it's been this way for over 24 hours now :/ [07:08:28] [Time] => 140255 [07:09:30] ? [07:09:38] in seconds [07:09:45] o_0 [07:10:08] PROBLEM - Puppet freshness on mw1126 is CRITICAL: No successful Puppet run in the last 10 hours [07:10:49] yeah :/ [07:11:12] Sleeping for 8.0 seconds, 2013-08-16 23:26:12 <-- that was when it started (pacific time) [07:15:49] dafuq [07:15:59] SELECT /* ApiQueryAllUsers::execute */ ipb_deleted,COUNT(*) AS recentedits,user_name [07:16:25] even this tiny part of the query is already batshit wrong [07:18:47] !log Killed a few hung queries on db1026 that were causing massive lag [07:18:59] Logged the message, Master [07:19:02] Lag is decreasing. [07:19:31] MaxSem: Did you grab the bad queries? [07:19:37] ^^^ [07:19:48] I mean the full queries. [07:19:53] It'd be good to have a bug report. [07:20:04] SHOW PROCESSLIST doesn't show full queries [07:20:15] show full processlist does. ;-) [07:20:25] Was it just ApiQueryAllUsers? [07:20:29] yup [07:20:31] K. [07:20:36] I'll file a bug against that. [07:20:38] yay thanks MaxSem! [07:21:03] "show full processlist\G" [07:22:27] hey, my window has only 5000 lines of scrollback fror \G! [07:22:58] :-) [07:23:09] Mine has as many lines as there is memory, I think. [07:23:26] @lag all [07:23:41] Blergh, SAL needs rotating. [07:23:51] It's at 400,000+ bytes. [07:26:15] MaxSem, legoktm: https://bugzilla.wikimedia.org/show_bug.cgi?id=52979 [07:26:21] MaxSem: Thanks for killing those queries. :-) [07:26:40] :) [07:31:16] hahaha [07:31:22] auactiveusers [07:31:56] https://git.wikimedia.org/blob/mediawiki%2Fcore.git/98c7138c3b079beef79ad34711bb2fece0f5d809/includes%2Fapi%2FApiQueryAllUsers.php [07:32:15] The COUNT(*) is easy to spot. [07:32:48] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 07:32:42 UTC 2013 [07:33:23] Not totally sure why it was hanging. [07:33:45] Hmmm. [07:33:48] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [07:33:48] PROBLEM - SSH on pdf1 is CRITICAL: CRITICAL - Socket timeout after 10 seconds [07:33:56] :DDD [07:34:39] RECOVERY - SSH on pdf1 is OK: SSH OK - OpenSSH_4.7p1 Debian-8ubuntu3 (protocol 2.0) [07:35:07] Something about Wikidata's unusually large recentchanges table, I suppose. [07:36:05] Hmmm. [07:37:04] 6154667 <-- count(*) from recentchanges on wikidatawiki_p [07:37:16] hehehe [07:37:17] 5933693 <-- count(*) from recentchanges on enwiki_p [07:37:38] They're comparable. [07:37:48] You'd think other wikis would be lagging/hanging. [07:38:06] Perhaps more scanning on wikidata due to fewer active users or something. [07:38:17] Query optimization is hard. Let's go shopping. [07:39:05] wikidata gets more edits / sec than en wp, has for months now [07:39:39] * apergos adds self to the bug [07:39:55] Right. Faster rate, but overall lower number per month, surprisingly. [07:40:11] Oh, never mind. [07:40:19] Higher number per month. Comparison is hard. [07:40:19] yes, nm [07:40:26] heh it's also sunday morning (here) [07:40:30] so yer excused [07:40:36] :-) [07:40:43] Yeah, I'm not fully awake. [07:40:56] Just enough to file a bug. ;-) [07:41:10] speaking of which I should start thinking about that cmelette with beet greens I was gonna have this morning [07:41:22] omelette [07:41:40] cmelette with cmjohnson [07:41:52] but I already got a director's shair fixed and a hammock hung this morning (the hammock had a lot of yaks first) [07:42:22] I thought "director's shair" was a meal for a second, heh. [07:42:37] no it was another bizarre typo [07:42:48] mehhh, ishmael is broken [07:42:51] an off by one given where the s and the c are on the keyboard [07:43:06] what's wrong with it? [07:43:07] Call me broken. [07:43:27] I'm hilarious. [07:43:34] I don't see sklow queries [07:44:04] Isn't there supposed to be a query killer for slow queries? [07:44:14] not yet [07:44:17] Think of the Domases. [07:44:32] I'm sure there was one at some point. [07:44:35] I thought it was set up [07:44:42] I think it breaks or something. [07:44:49] not active atm [07:44:54] * Elsie nods. [07:45:11] killing select queries should be safe [07:45:33] it was going to kill anything running as wikiuser that took longer than 5 mins I think [07:45:39] I even remember seeing actual code [07:46:06] MaxSem: If there's lag, perhaps. [07:46:40] It'd also be good to track which queries are hanging and file bugs. [07:46:46] Rather than just killing. [07:47:40] oh, those theoreticians:P [07:54:31] Ryan_Lane: Yo. Is the server admin log supposed to be moved to the archive? [07:54:42] It seems Daniel moved it rather than copying and pasting last time. [07:56:15] Hmmm, seems Antoine did it before him with Archive 20. [07:57:38] > 02:32 logmsgbot: LocalisationUpdate completed (1.22wmf12) at Sat Jul 27 02:31:17 UTC 2013 [07:57:59] Is there any system in which that timestamp makes sense? [08:00:06] rt claims we already have that pt-kill crn job. so somthing's up, I'll point to that bug and see if someone knows what's going on with that [08:00:41] ah what user were those queries running as? [08:00:53] MaxSem: any chance you remember? [08:02:05] LocalisationUpdate has update.php. [08:02:14] But Wikimedia must have a wrapper of some sort. [08:05:20] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [08:08:31] 45 KiB/s from Gerrit. [08:08:33] Woo. [08:11:30] apergos, some anon and some non-bot user [08:11:47] er I mean what mysql user... wikiuser? wikiadmin? [08:11:51] MaxSem: [08:11:55] wikiuser [08:11:59] hmm [08:12:02] as it was coming from API [08:12:39] and one more thing, any chance you remember how long the query had been running (more than 15 mins, more than an hour, ...?) [08:12:59] [Time] => 140255 [08:13:25] :-) [08:13:26] ok thanks, that needs to go on an rt ticket [08:13:51] Or that lovely Bugzilla bug. [08:14:15] Cloning operations/puppet is painfully slow. [08:14:28] By the time it finishes, I'll have forgotten what I was going to do with it. [08:14:32] uh well it needs to go into the rt ticket about pt-kill [08:15:25] RT is a plague. [08:15:43] that's how it goes [08:15:51] Indeed. [08:16:21] oh I see, sean took the ticket but it hasn't yet been implemented. meh [08:16:26] grrr [08:16:27] 3393485913 | wikiuser | 10.64.32.67:47517 | wikidatawiki | Query | 93 | Copying to tmp table | SELECT /* Wikibase\TermSqlIndex::getMatchingIDs */ DISTINCT term_entity_id,term_weight FROM `wb_te [08:17:00] I can hear Domas crying on the inside. [08:17:09] notme [08:17:10] PROBLEM - Puppet freshness on lvs1004 is CRITICAL: No successful Puppet run in the last 10 hours [08:17:10] PROBLEM - Puppet freshness on lvs1005 is CRITICAL: No successful Puppet run in the last 10 hours [08:17:10] PROBLEM - Puppet freshness on lvs1006 is CRITICAL: No successful Puppet run in the last 10 hours [08:17:10] PROBLEM - Puppet freshness on virt1 is CRITICAL: No successful Puppet run in the last 10 hours [08:17:10] PROBLEM - Puppet freshness on virt3 is CRITICAL: No successful Puppet run in the last 10 hours [08:17:11] PROBLEM - Puppet freshness on virt4 is CRITICAL: No successful Puppet run in the last 10 hours [08:17:50] full query: SELECT /* Wikibase\TermSqlIndex::getMatchingIDs */ DISTINCT term_entity_id,term_weight FROM `wb_terms` WHERE (term_language='zh-cn' AND term_search_key LIKE 'te%' AND term_type='label' AND term_entity_type='item') OR (term_language='zh-cn' AND term_search_key LIKE 'te%' AND term_type='alias' AND term_entity_type='item') LIMIT 5000 [08:18:04] DAFUUUUUQ??????!1 [08:18:42] Hmmmm. So it's `date`'s output that we're seeing in !log. [08:18:55] I wonder if that's worth a bug. [08:19:42] rows: 539106 [08:20:25] if I were Domas, I would've already been trashing WIkidata from servers:P [08:20:40] // Domas optimized [08:20:46] I got jokes. [08:20:54] Where do Puppet bugs go in Bugzilla? [08:21:25] Wikimedia --> General/Unknown it is. [08:22:17] puppet bugs should really go to RT [08:22:21] using intersect, using where, using temporary [08:22:48] RT is a plague. [08:22:50] it's not like explain wouldn't warn folks [08:23:03] This is also a really stupid bug I'm filing. [08:24:00] aha, it's https://bugzilla.wikimedia.org/show_bug.cgi?id=45529 [08:24:39] well that just got more urgent too didn't it [08:24:42] *sigh* [08:26:05] okoff to make omelette right now this second [08:26:56] https://bugzilla.wikimedia.org/show_bug.cgi?id=52986 [08:32:50] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 08:32:43 UTC 2013 [08:33:20] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [09:07:15] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [09:24:53] hm, what's going on? [09:25:00] going through backlog [09:26:27] paravoid: wd has crappy db queries by the looks [09:26:56] yeah [09:32:45] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 09:32:39 UTC 2013 [09:33:15] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [09:37:55] PROBLEM - RAID on searchidx1001 is CRITICAL: CHECK_NRPE: Socket timeout after 10 seconds. [09:38:55] RECOVERY - RAID on searchidx1001 is OK: OK: State is Optimal, checked 4 logical device(s) [10:05:33] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [10:13:37] !og resuming slave db52 after bug 52853 investigation [10:14:04] !log resuming slave db52 after bug 52853 investigation [10:14:15] Logged the message, Master [10:33:13] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 10:33:09 UTC 2013 [10:33:33] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [10:42:51] PROBLEM - RAID on searchidx1001 is CRITICAL: CHECK_NRPE: Socket timeout after 10 seconds. [10:43:51] RECOVERY - RAID on searchidx1001 is OK: OK: State is Optimal, checked 4 logical device(s) [10:46:51] PROBLEM - RAID on searchidx1001 is CRITICAL: CHECK_NRPE: Socket timeout after 10 seconds. [10:48:51] RECOVERY - RAID on searchidx1001 is OK: OK: State is Optimal, checked 4 logical device(s) [11:05:32] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [11:32:42] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 11:32:35 UTC 2013 [11:33:32] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [11:34:20] (03PS1) 10Mark Bergsma: Fix mobile redirect [operations/puppet] - 10https://gerrit.wikimedia.org/r/79563 [11:43:24] !log Moved eqiad text-varnish traffic back to Squid temporarily [11:43:37] Logged the message, Master [11:45:44] (03CR) 10Mark Bergsma: [C: 032] Fix mobile redirect [operations/puppet] - 10https://gerrit.wikimedia.org/r/79563 (owner: 10Mark Bergsma) [11:59:05] gaahh [11:59:13] it's the http:// [11:59:21] in req.host [11:59:24] how did I not see this [11:59:51] aaan I just saw you fixed this [12:00:11] :) [12:00:30] grmbl [12:06:23] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [12:11:23] PROBLEM - Puppet freshness on terbium is CRITICAL: No successful Puppet run in the last 10 hours [12:11:24] PROBLEM - Puppet freshness on cp1063 is CRITICAL: No successful Puppet run in the last 10 hours [12:26:12] !log authdns-update: add AAAA for sodium [12:26:23] Logged the message, Master [12:32:44] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 12:32:38 UTC 2013 [12:33:23] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [12:37:45] (03PS1) 10Mark Bergsma: Add nagios group for misc caches [operations/puppet] - 10https://gerrit.wikimedia.org/r/79567 [12:50:51] PROBLEM - RAID on searchidx1001 is CRITICAL: CHECK_NRPE: Socket timeout after 10 seconds. [12:50:57] (03CR) 10Mark Bergsma: [C: 032] Add nagios group for misc caches [operations/puppet] - 10https://gerrit.wikimedia.org/r/79567 (owner: 10Mark Bergsma) [12:51:51] RECOVERY - RAID on searchidx1001 is OK: OK: State is Optimal, checked 4 logical device(s) [13:05:42] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [13:15:30] (03PS1) 10Mark Bergsma: Add role::cache::ssl::wikimedia, factor out common resources [operations/puppet] - 10https://gerrit.wikimedia.org/r/79570 [13:16:36] (03CR) 10Mark Bergsma: [C: 032] Add role::cache::ssl::wikimedia, factor out common resources [operations/puppet] - 10https://gerrit.wikimedia.org/r/79570 (owner: 10Mark Bergsma) [13:20:03] (03PS1) 10Mark Bergsma: Switch the role::cache::misc cluster to role::cache::ssl::wikimedia [operations/puppet] - 10https://gerrit.wikimedia.org/r/79572 [13:21:04] (03CR) 10Mark Bergsma: [C: 032] Switch the role::cache::misc cluster to role::cache::ssl::wikimedia [operations/puppet] - 10https://gerrit.wikimedia.org/r/79572 (owner: 10Mark Bergsma) [13:28:33] PROBLEM - Puppetmaster HTTPS on stafford is CRITICAL: CRITICAL - Socket timeout after 10 seconds [13:29:32] RECOVERY - Puppetmaster HTTPS on stafford is OK: HTTP OK: Status line output matched 400 - 336 bytes in 6.048 second response time [13:33:32] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 13:33:27 UTC 2013 [13:33:42] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [13:42:53] PROBLEM - Host mw31 is DOWN: PING CRITICAL - Packet loss = 100% [13:43:43] RECOVERY - Host mw31 is UP: PING OK - Packet loss = 0%, RTA = 26.60 ms [13:57:34] PROBLEM - Puppetmaster HTTPS on stafford is CRITICAL: CRITICAL - Socket timeout after 10 seconds [13:58:23] RECOVERY - Puppetmaster HTTPS on stafford is OK: HTTP OK: Status line output matched 400 - 336 bytes in 0.129 second response time [14:06:16] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [14:22:40] (03PS1) 10Mark Bergsma: Serve a /robots.txt if the backend doesn't [operations/puppet] - 10https://gerrit.wikimedia.org/r/79574 [14:27:23] (03Abandoned) 10Mark Bergsma: Serve a /robots.txt if the backend doesn't [operations/puppet] - 10https://gerrit.wikimedia.org/r/79574 (owner: 10Mark Bergsma) [14:34:56] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 14:34:49 UTC 2013 [14:40:06] (03PS1) 10Mark Bergsma: Add Icinga monitoring (critical) for the misc-web cluster [operations/puppet] - 10https://gerrit.wikimedia.org/r/79576 [14:42:02] (03CR) 10Mark Bergsma: [C: 032] Add Icinga monitoring (critical) for the misc-web cluster [operations/puppet] - 10https://gerrit.wikimedia.org/r/79576 (owner: 10Mark Bergsma) [14:46:13] PROBLEM - Puppet freshness on zirconium is CRITICAL: No successful Puppet run in the last 10 hours [14:48:53] PROBLEM - RAID on searchidx1001 is CRITICAL: CHECK_NRPE: Socket timeout after 10 seconds. [14:48:53] PROBLEM - DPKG on searchidx1001 is CRITICAL: CHECK_NRPE: Socket timeout after 10 seconds. [14:49:44] RECOVERY - DPKG on searchidx1001 is OK: All packages OK [14:49:53] RECOVERY - RAID on searchidx1001 is OK: OK: State is Optimal, checked 4 logical device(s) [15:00:00] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [15:02:41] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 15:02:36 UTC 2013 [15:03:00] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [15:11:20] PROBLEM - SSH on amslvs1 is CRITICAL: Server answer: [15:17:20] RECOVERY - SSH on amslvs1 is OK: SSH OK - OpenSSH_5.9p1 Debian-5ubuntu1.1 (protocol 2.0) [15:20:30] PROBLEM - RAID on searchidx1001 is CRITICAL: CHECK_NRPE: Socket timeout after 10 seconds. [15:22:40] PROBLEM - Puppetmaster HTTPS on stafford is CRITICAL: CRITICAL - Socket timeout after 10 seconds [15:23:30] RECOVERY - Puppetmaster HTTPS on stafford is OK: HTTP OK: Status line output matched 400 - 336 bytes in 0.123 second response time [15:26:20] RECOVERY - RAID on searchidx1001 is OK: OK: State is Optimal, checked 4 logical device(s) [15:43:05] PROBLEM - RAID on searchidx1001 is CRITICAL: CHECK_NRPE: Socket timeout after 10 seconds. [15:43:55] RECOVERY - RAID on searchidx1001 is OK: OK: State is Optimal, checked 4 logical device(s) [16:10:16] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [16:20:03] (03CR) 10MZMcBride: "(1 comment)" [operations/puppet] - 10https://gerrit.wikimedia.org/r/78944 (owner: 10QChris) [16:32:47] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 16:32:41 UTC 2013 [16:33:16] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [17:06:02] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [17:10:52] PROBLEM - Puppet freshness on mw1126 is CRITICAL: No successful Puppet run in the last 10 hours [17:32:42] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 17:32:41 UTC 2013 [17:33:02] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [18:06:14] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [18:18:04] PROBLEM - Puppet freshness on lvs1004 is CRITICAL: No successful Puppet run in the last 10 hours [18:18:04] PROBLEM - Puppet freshness on lvs1005 is CRITICAL: No successful Puppet run in the last 10 hours [18:18:04] PROBLEM - Puppet freshness on lvs1006 is CRITICAL: No successful Puppet run in the last 10 hours [18:18:04] PROBLEM - Puppet freshness on virt1 is CRITICAL: No successful Puppet run in the last 10 hours [18:18:04] PROBLEM - Puppet freshness on virt3 is CRITICAL: No successful Puppet run in the last 10 hours [18:18:05] PROBLEM - Puppet freshness on virt4 is CRITICAL: No successful Puppet run in the last 10 hours [18:33:24] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 18:33:15 UTC 2013 [18:34:14] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [18:54:29] !log rotated server admin log wikitech page; testing logging [18:54:34] Logged the message, Master [18:54:47] All better. [19:06:25] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [19:32:45] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 19:32:38 UTC 2013 [19:33:25] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [19:43:59] (03CR) 10QChris: "(1 comment)" [operations/puppet] - 10https://gerrit.wikimedia.org/r/78944 (owner: 10QChris) [20:00:07] PROBLEM - RAID on searchidx1001 is CRITICAL: CHECK_NRPE: Socket timeout after 10 seconds. [20:02:07] RECOVERY - RAID on searchidx1001 is OK: OK: State is Optimal, checked 4 logical device(s) [20:06:24] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [20:32:44] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 20:32:40 UTC 2013 [20:33:24] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [21:02:18] (03CR) 10Edenhill: [C: 031] "Looks good!" [operations/software/varnish/varnishkafka] (debian) - 10https://gerrit.wikimedia.org/r/78782 (owner: 10Faidon) [21:04:10] (03CR) 10Edenhill: "(3 comments)" [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/78160 (owner: 10Edenhill) [21:07:50] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [21:32:40] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 21:32:38 UTC 2013 [21:32:50] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [22:06:07] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [22:11:27] PROBLEM - Puppet freshness on cp1063 is CRITICAL: No successful Puppet run in the last 10 hours [22:11:27] PROBLEM - Puppet freshness on terbium is CRITICAL: No successful Puppet run in the last 10 hours [22:24:55] (03Abandoned) 10Edenhill: Link with -lrt [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/78161 (owner: 10Edenhill) [22:25:21] (03Abandoned) 10Edenhill: Added JSON formatter, field name identifers and type casting option. [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/78160 (owner: 10Edenhill) [22:25:33] (03Abandoned) 10Edenhill: Added support for escaping troublesome characters in tag content. [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/78157 (owner: 10Edenhill) [22:32:47] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 22:32:42 UTC 2013 [22:33:07] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [22:40:16] (03PS1) 10Edenhill: Added support for escaping troublesome characters in tag content. [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/79745 [22:40:17] (03PS1) 10Edenhill: Added JSON formatter, field name identifers and type casting option. [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/79746 [22:40:18] (03PS1) 10Edenhill: Added 'output = null' for testing purposes. [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/79747 [22:40:19] (03PS1) 10Edenhill: When reading offline VSL files (-r ..) make a copy of each matched tags data since the data is volatile. [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/79748 [22:40:20] (03PS1) 10Edenhill: Handle "Var: Val" with empty " Val"s. [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/79749 [22:40:21] (03PS1) 10Edenhill: Indent fix and clarified comment. [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/79750 [22:40:22] (03PS1) 10Edenhill: Dont redeclare 'len' [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/79751 [22:40:23] (03PS1) 10Edenhill: Decrease default log.level to 6 (info) [operations/software/varnish/varnishkafka] - 10https://gerrit.wikimedia.org/r/79752 [23:06:14] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours [23:33:44] RECOVERY - Puppet freshness on mexia is OK: puppet ran at Sun Aug 18 23:33:38 UTC 2013 [23:34:14] PROBLEM - Puppet freshness on mexia is CRITICAL: No successful Puppet run in the last 10 hours