[00:05:50] PROBLEM - Puppet freshness on copper is CRITICAL: Puppet has not run in the last 10 hours [00:05:50] PROBLEM - Puppet freshness on zinc is CRITICAL: Puppet has not run in the last 10 hours [01:15:33] New patchset: Lcarr; "Adding in sun JDK to contint servers" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1591 [01:15:44] New review: gerrit2; "Change did not pass lint check. You will need to send an amended patchset for this (see: https://lab..." [operations/puppet] (production); V: -1 - https://gerrit.wikimedia.org/r/1591 [01:15:50] can I get a what, what? (maybe maplebed ? ) https://gerrit.wikimedia.org/r/1591 [01:18:51] New patchset: Lcarr; "Adding in sun JDK to contint servers" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1591 [01:19:03] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1591 [01:20:36] fyi it's for rt.wikimedia.org/Ticket/Display.html?id=2147 [01:30:32] New patchset: Lcarr; "Adding in sun JDK to contint servers" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1591 [01:37:08] New patchset: Lcarr; "Adding in sun JDK to contint servers" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1591 [01:37:18] New review: gerrit2; "Change did not pass lint check. You will need to send an amended patchset for this (see: https://lab..." [operations/puppet] (production); V: -1 - https://gerrit.wikimedia.org/r/1591 [01:39:38] New patchset: Lcarr; "Adding in sun JDK to contint servers" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1591 [01:39:49] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1591 [01:40:41] LeslieCarr: we need libswt-gtk-3.5-java too I think [01:41:15] http://rt.wikimedia.org/Ticket/Display.html?id=2147 [01:42:08] New patchset: Lcarr; "Adding in sun JDK to contint servers" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1591 [01:42:19] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1591 [01:44:39] New review: Bhartshorne; "(no comment)" [operations/puppet] (production); V: 0 C: 1; - https://gerrit.wikimedia.org/r/1591 [01:44:46] LeslieCarr: we might need ia32-libs as well [01:45:02] New review: Lcarr; "(no comment)" [operations/puppet] (production); V: 0 C: 2; - https://gerrit.wikimedia.org/r/1591 [01:45:02] Change merged: Lcarr; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1591 [01:45:44] LeslieCarr: we also need to export ANDROID_SWT= [01:46:05] preilly: just make sure to put it all in the ticket :) [01:57:48] preilly: ping? [01:58:24] 15 02:39:02 < jeremyb> someone have a public link to info about last night's mobile deploy? [01:58:24] he was here just a second ago. [01:58:27] 15 02:39:21 < jeremyb> specifically the breakage with desktop logins going to m. [01:59:26] jeremyb: ? [01:59:36] preilly: see my paste ^^ ? [01:59:45] jeremyb: from two night ago? [01:59:48] (2 nights ago now) [01:59:49] yes [02:00:09] jeremyb: it was just an issue the a redirect hook firing on non mobile logins [02:00:25] jeremyb: it was fixed by adding a check for the mobile only header [02:01:08] preilly: well, was there a mail in pipermail or a !log or a bug (or even RT) or some other thing i'm not thinking of that mentioned either the breakage or the fact that it had been fixed? [02:01:33] jeremyb: there was an email [02:01:49] jeremyb: I'm leaving the office now [02:01:52] preilly: well, there was the internal one, but one i can link to for the public? [02:01:56] sure [03:11:23] PROBLEM - Puppet freshness on es1002 is CRITICAL: Puppet has not run in the last 10 hours [03:12:14] PROBLEM - SSH on maerlant is CRITICAL: CRITICAL - Socket timeout after 10 seconds [03:30:27] PROBLEM - Puppet freshness on bast1001 is CRITICAL: Puppet has not run in the last 10 hours [03:32:26] PROBLEM - Puppet freshness on fenari is CRITICAL: Puppet has not run in the last 10 hours [04:44:04] RECOVERY - SSH on maerlant is OK: SSH OK - OpenSSH_5.3p1 Debian-3ubuntu7 (protocol 2.0) [04:54:51] PROBLEM - Recursive DNS on 91.198.174.6 is CRITICAL: CRITICAL - Plugin timed out while executing system call [06:37:55] PROBLEM - Puppet freshness on snapshot4 is CRITICAL: Puppet has not run in the last 10 hours [07:33:37] PROBLEM - Puppet freshness on ssl3002 is CRITICAL: Puppet has not run in the last 10 hours [07:33:37] PROBLEM - Puppet freshness on ssl2 is CRITICAL: Puppet has not run in the last 10 hours [07:33:37] PROBLEM - Puppet freshness on ssl3004 is CRITICAL: Puppet has not run in the last 10 hours [08:02:14] sigh, spence come on [08:02:30] have you notived the ganglia graphs for the past day? [08:02:47] i am looking at top right now [08:03:01] have you noticed the ganglia graphs for the cluser as a whole for the last day? [08:03:09] *cluster [08:03:13] oh [08:03:55] trying to open it, getting refused right now [08:04:01] yeah look slike it [08:04:09] well much of the last half of the day is missing data [08:04:26] when I checked a few minutes ago it was back to graphing, so I couldn't fix anything [08:06:35] would you think "ganglia_parser" is part of the problem? [08:07:52] no idea, can't get on the box [08:09:17] it's rsyncing stuff: /usr/bin/rsync -a /mnt/ganglia_tmp/rrds.pmtpa/ /var/lib/ganglia/rrds.pmtpa [08:09:37] and quite a bit of /usr/bin/python -tt /usr/sbin/ganglia_parser [08:09:57] that rsync needs to happen [08:10:01] once an hour or something doesn't it? [08:10:13] otherwise if ganglia restarts we lose the data [08:10:23] are there many of those? [08:10:44] yea, just wondering what makes it so slow [08:11:33] i'm checking, but it takes patience ... [08:12:35] 20 procs with string "ganglia" [08:13:27] 17 of them ganglia_parser, user nagios runs it [08:18:58] !log spence slow again, side-note: tried to use "sar" to investigate but "Please check if data collecting is enabled in /etc/default/sysstat" (want to?) [08:19:08] Logged the message, Master [08:23:46] load's dropping some but [08:23:59] still a ton of swap used [08:24:15] notifications.cgi and history.cgi use a lot of memory [08:24:23] yes indeed [08:24:27] makes me want to restart nagios again [08:24:30] k? [08:24:34] yes, please do [08:24:42] I would stop it [08:24:49] leave it stopped for 5 min, watch swap [08:24:53] restart [08:26:03] !log spence - almost unusable, Nagios notifications.cgi and history.cgi use a lot of memory, stopping Nagios, watching swap [08:26:12] Logged the message, Master [08:26:39] Warning - nagios did not exit in a timely manner [08:26:47] that's life [08:27:19] finally I am starting to see swap decrease, if slowly [08:29:33] killed 2 more procs, notifications.cgi [08:29:47] oh it was still going? heh [08:29:48] hrmm, let me also kill history.cgi [08:29:57] thank you [08:30:18] load avg going down [08:30:33] there were a couple of "noficiations", you got em both? [08:30:40] now "rsync" is the top CPU user [08:30:56] shouldn't take too long to finish [08:30:58] yes, got all notifications.cgi [08:31:22] no "cgi" at all anymore [08:32:47] ganglia_parser, run by user nagios, still doing stuff and didnt touch it [08:33:06] mmm [08:33:39] fine [08:37:32] so what I don't know is whether these other jobs owned by nagios ought to be shot before it gets restarted [08:37:34] mutante: hi :) so spence is dieting out because Nagios eats all memory? :-) [08:37:49] I don't know about the nsca stuff either [08:37:53] yep [08:38:09] so we're forcing it to take a little break and give back all the swap it ate [08:38:19] i wouldnt blame nsca, it has been running, didnt kill it, not using a lot [08:38:24] you probably have to many checks by now [08:38:26] and spence is much faster now [08:38:30] no, I don't blame those processes [08:38:53] I just wonder if they will be "left over " processes and therefore a problem when nagios is restarted [08:39:22] or you already restarted it? [08:39:26] or.... puppet restarted it? [08:39:37] NSCA is a demon which let you receive "passive checks" [08:39:44] hashar: hi, yeah, notifications.cgi [08:39:51] I see [08:39:58] apergos: 5 minutes over.. starting again now..hm [08:39:58] you can get a script feeding nagios with check monitoring [08:40:01] ok [08:40:10] for example giving it radius accounting or SNMP traps [08:40:17] well I see some puppet freshness checks in there already [08:40:34] I think puppet send SNMP traps [08:40:41] ok [08:40:47] they are then probably send back in NSCA for nagios processing [08:40:54] (all of that is pure supposition from me) [08:41:00] heh [08:41:03] (btw: nagios sucks) [08:41:05] ;-D [08:41:28] !log spence - had killed additional notifications.cgi and history.cgi procs, waited 5 minutes, load went down a lot, restarting nagios [08:41:37] Logged the message, Master [08:41:55] hashar: NSCA receives UDP packages [08:42:05] if we see it climb a huge amount again (shortly) this is going to prolly mean restarting the host [08:42:07] hashar: and they include the check results [08:44:17] the daemon just sits there and waits for incoming packages which include host names and service check results forwards them to Nagios, which compares that to existing host and service definitions. and if it never receives any result it also turns it into a CRIT. [08:44:18] RECOVERY - Recursive DNS on 91.198.174.6 is OK: DNS OK: 0.228 seconds response time. www.wikipedia.org returns 91.198.174.225 [08:44:48] (yea, it's encrypted) [08:46:10] you can use "send_nsca" on a monitored host to send the results of any plugin over to the nsca host [08:47:19] I have been using NRPE for that [08:47:31] it's "the other way around" [08:47:45] but we did use NSCA to receive informations from other software [08:47:57] NRPE = Nagios master executes stuff on other hosts , NSCA = hosts send results to the master, which just needs to wait for them [08:47:58] well actually we had a trap and syslog listener [08:48:14] thats why they are "passive" checks [08:48:14] then two perl daemons reading them and formatting those messages to then send then to NSCA [08:48:34] got you [08:48:40] :-)) [08:48:48] yep, the last step is send_nsca (its in puppet;) [08:49:13] then eventually everything was dropped in favor of nagios doing only active checks and sending alarms to a collector [08:49:25] all sys logs / accounting etc were also send to a collector [08:49:40] then we had NetCool on top of that to display nice alarms :D [08:49:59] I love monitoring, that is all funny [08:49:59] some people setup 2 layers of Nagios, one master which only receives passive results, but it receives that from multiple other Nagios (sub-)masters, who do active checks on hosts [08:50:39] that is a good way to scale [08:50:40] to scale up to more than a couple hundred hosts [08:50:44] * mutante nods [08:50:52] I think that is what the team did after I left [08:51:26] several "slaves" doing checks. A controller to generates their configuration and every checks result send to two master servers [08:51:35] not sure how they manage all of that though [08:52:05] anyway, two freelancers I have met there told me to have a look at http://www.opennms.org/ [08:52:23] hrmm, it almost feels like there should be a warning "dont use history.cgi" [08:52:23] seems to be way easier to setup and it could discovers services as I understood it [08:52:48] when you make nagios look through history, that seems to be reaallly hard work [08:53:04] by default it looks up for all hosts / all services [08:53:13] plus it uses flat files iirc [08:53:14] maybe we are keeping too much of it? blind guess [08:53:42] so it probably need to load the whole history file in memory then process it and display all the nice HMTL [08:53:49] yea [08:54:20] if it really matter to anyone, they should be export to a SQL database of some sort [08:54:25] I can't remember the plugin name though [08:54:45] and then disable all the histogram / history cgi [08:55:05] check out "Icinga" [08:55:44] https://www.icinga.org/nagios/ [08:56:32] oh well, but Nagios itself also is listed as offering MySQL support [08:56:55] just remembered that Icinga has it for sure https://www.icinga.org/nagios/feature-comparison/ [08:57:20] I have not followed nagios for the last 3 years or so :D [08:57:28] so I am a bit outdated with all theses forks [08:57:40] Icinga probably worth a look [08:57:43] also noteworthy keywords on that page: LDAP, IRC [08:58:18] looks good for now [08:58:20] as I understand it icinga is a community fork. Much like X.org forked from xf86 (or something similar) [08:58:34] I dunno if nagios is doing what it needs to do but ganglia is happy, load is happy, mem use is happy [08:58:45] apergos: cool:) [08:59:10] apergos: well, i tried to use history.cgi, saw it use a lot of CPU, killed it one more time, and not hitting that button anymore for now:) [08:59:14] * apergos just used the hell out of adblock on a site [08:59:38] mutante: makesure to disable history.cgi [08:59:41] maybe you should have to give explicit permission before a site can send you swf stuff [08:59:51] the best way is probably to deny it with an Apache directive [09:00:17] hashar: but i want it to work .. and not even sure if it really is the culprit yet [09:00:45] well it is probably better to have something disabled for now than having it crash the monitoring box :-D [09:00:52] hashar: then we cant check anymore when Nagios alarms have started and stopped after an event [09:01:08] we couldn't check anyhow :-P [09:01:12] yea [09:01:23] ok [09:05:08] PROBLEM - MySQL slave status on es1004 is CRITICAL: CRITICAL: Slave running: expected Yes, got No [09:07:08] PROBLEM - Puppet freshness on aluminium is CRITICAL: Puppet has not run in the last 10 hours [09:07:08] PROBLEM - Puppet freshness on db1001 is CRITICAL: Puppet has not run in the last 10 hours [09:07:08] PROBLEM - Puppet freshness on db1003 is CRITICAL: Puppet has not run in the last 10 hours [09:07:08] PROBLEM - Puppet freshness on db1005 is CRITICAL: Puppet has not run in the last 10 hours [09:07:08] PROBLEM - Puppet freshness on db1002 is CRITICAL: Puppet has not run in the last 10 hours [09:09:30] mutante: on another subject. TestSwarm is running on labs now :-D [09:09:32] https://testswarm.wmflabs.org/testswarm/ [09:09:48] pushing straight to test branch without review made it easier [09:09:54] err faster [09:11:27] hey, nice to hear it is running! yay [09:12:18] test branches ftw [09:12:28] I have been running the dumps out of one now for over a year :-D [09:15:27] ideally I would need a dedicated branch [09:15:33] and have my VM puppet daemon run from it [09:15:44] IIRC it is a work in progress [09:21:42] I should look at your stuff next time I have some spare time [09:21:49] also I should get a labs project set up [09:21:56] I have an account but no project [09:22:11] (I was one of the first test accounts, you couldn't create projects then) [09:29:15] i can setup security groups and instances, but i'm still not sure where you configure which group is used on which instance [09:29:27] next step i need on my labs vm is php5-cli [09:29:31] bbiaw, short break [09:32:05] well I am going back to bed :-) have a good day! [09:35:10] sleep well [09:35:28] I'm going to get groceries while it's sunny (it was pouring buckets earlier and is supposed to again, for the weekend) [09:42:48] PROBLEM - Puppet freshness on db1025 is CRITICAL: Puppet has not run in the last 10 hours [09:42:48] PROBLEM - Puppet freshness on db18 is CRITICAL: Puppet has not run in the last 10 hours [09:42:48] PROBLEM - Puppet freshness on db25 is CRITICAL: Puppet has not run in the last 10 hours [09:42:48] PROBLEM - Puppet freshness on mw1004 is CRITICAL: Puppet has not run in the last 10 hours [09:42:48] PROBLEM - Puppet freshness on erzurumi is CRITICAL: Puppet has not run in the last 10 hours [09:42:48] PROBLEM - Puppet freshness on mw1010 is CRITICAL: Puppet has not run in the last 10 hours [09:42:48] PROBLEM - Puppet freshness on mw1030 is CRITICAL: Puppet has not run in the last 10 hours [09:42:49] PROBLEM - Puppet freshness on mw1045 is CRITICAL: Puppet has not run in the last 10 hours [09:42:49] PROBLEM - Puppet freshness on mw1055 is CRITICAL: Puppet has not run in the last 10 hours [09:42:50] PROBLEM - Puppet freshness on mw1100 is CRITICAL: Puppet has not run in the last 10 hours [09:42:50] PROBLEM - Puppet freshness on mw5 is CRITICAL: Puppet has not run in the last 10 hours [09:42:51] PROBLEM - Puppet freshness on nfs1 is CRITICAL: Puppet has not run in the last 10 hours [09:42:51] PROBLEM - Puppet freshness on search15 is CRITICAL: Puppet has not run in the last 10 hours [09:42:52] PROBLEM - Puppet freshness on searchidx2 is CRITICAL: Puppet has not run in the last 10 hours [09:42:52] PROBLEM - Puppet freshness on srv269 is CRITICAL: Puppet has not run in the last 10 hours [09:42:53] PROBLEM - Puppet freshness on srv276 is CRITICAL: Puppet has not run in the last 10 hours [09:54:47] PROBLEM - Puppet freshness on mw1068 is CRITICAL: Puppet has not run in the last 10 hours [09:54:47] PROBLEM - Puppet freshness on mw1072 is CRITICAL: Puppet has not run in the last 10 hours [09:54:47] PROBLEM - Puppet freshness on mw1076 is CRITICAL: Puppet has not run in the last 10 hours [09:54:47] PROBLEM - Puppet freshness on mw1138 is CRITICAL: Puppet has not run in the last 10 hours [09:54:47] PROBLEM - Puppet freshness on mw36 is CRITICAL: Puppet has not run in the last 10 hours [09:54:48] PROBLEM - Puppet freshness on mw54 is CRITICAL: Puppet has not run in the last 10 hours [09:54:48] PROBLEM - Puppet freshness on search9 is CRITICAL: Puppet has not run in the last 10 hours [09:54:49] PROBLEM - Puppet freshness on srv211 is CRITICAL: Puppet has not run in the last 10 hours [09:54:49] PROBLEM - Puppet freshness on srv238 is CRITICAL: Puppet has not run in the last 10 hours [09:54:50] PROBLEM - Puppet freshness on virt2 is CRITICAL: Puppet has not run in the last 10 hours [09:58:58] PROBLEM - Puppet freshness on db1026 is CRITICAL: Puppet has not run in the last 10 hours [09:58:58] PROBLEM - Puppet freshness on db1012 is CRITICAL: Puppet has not run in the last 10 hours [09:58:58] PROBLEM - Puppet freshness on mw1025 is CRITICAL: Puppet has not run in the last 10 hours [09:58:58] PROBLEM - Puppet freshness on db1033 is CRITICAL: Puppet has not run in the last 10 hours [09:58:58] PROBLEM - Puppet freshness on mw1046 is CRITICAL: Puppet has not run in the last 10 hours [09:58:59] PROBLEM - Puppet freshness on mw11 is CRITICAL: Puppet has not run in the last 10 hours [09:58:59] PROBLEM - Puppet freshness on mw1113 is CRITICAL: Puppet has not run in the last 10 hours [09:59:00] PROBLEM - Puppet freshness on mw1127 is CRITICAL: Puppet has not run in the last 10 hours [09:59:00] PROBLEM - Puppet freshness on srv236 is CRITICAL: Puppet has not run in the last 10 hours [09:59:01] PROBLEM - Puppet freshness on mw1149 is CRITICAL: Puppet has not run in the last 10 hours [09:59:01] PROBLEM - Puppet freshness on mw42 is CRITICAL: Puppet has not run in the last 10 hours [09:59:02] PROBLEM - Puppet freshness on mw1133 is CRITICAL: Puppet has not run in the last 10 hours [09:59:02] PROBLEM - Puppet freshness on srv263 is CRITICAL: Puppet has not run in the last 10 hours [09:59:03] PROBLEM - Puppet freshness on srv256 is CRITICAL: Puppet has not run in the last 10 hours [10:00:57] PROBLEM - Puppet freshness on mw1064 is CRITICAL: Puppet has not run in the last 10 hours [10:00:57] PROBLEM - Puppet freshness on mw1151 is CRITICAL: Puppet has not run in the last 10 hours [10:00:57] PROBLEM - Puppet freshness on mw32 is CRITICAL: Puppet has not run in the last 10 hours [10:00:57] PROBLEM - Puppet freshness on sockpuppet is CRITICAL: Puppet has not run in the last 10 hours [10:04:27] PROBLEM - Puppet freshness on db1004 is CRITICAL: Puppet has not run in the last 10 hours [10:04:27] PROBLEM - Puppet freshness on db1031 is CRITICAL: Puppet has not run in the last 10 hours [10:04:27] PROBLEM - Puppet freshness on db10 is CRITICAL: Puppet has not run in the last 10 hours [10:04:27] PROBLEM - Puppet freshness on db1039 is CRITICAL: Puppet has not run in the last 10 hours [10:04:27] PROBLEM - Puppet freshness on db44 is CRITICAL: Puppet has not run in the last 10 hours [10:04:28] PROBLEM - Puppet freshness on db33 is CRITICAL: Puppet has not run in the last 10 hours [10:04:28] PROBLEM - Puppet freshness on grosley is CRITICAL: Puppet has not run in the last 10 hours [10:21:38] RECOVERY - MySQL slave status on es1004 is OK: OK: [10:30:27] PROBLEM - Puppet freshness on maerlant is CRITICAL: Puppet has not run in the last 10 hours [10:43:20] New patchset: Dzahn; "tweak nagios to relieve spence - external command_check_interval - event_broker_options" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1592 [10:47:29] New review: Dzahn; "external command check_interval was set to "-1" which is "as often as possible". (this would influen..." [operations/puppet] (production); V: 1 C: 1; - https://gerrit.wikimedia.org/r/1592 [10:49:21] New review: Dzahn; "external command check_interval was set to "-1" which is "as often as possible". (this would influen..." [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1592 [10:49:21] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1592 [10:56:22] aha! "test/mediawiki/core" stuff in gerrit! [11:21:13] New patchset: Dzahn; "add techblog apache_site configs - rebased - move to /misc/blogs.pp" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1433 [11:21:28] PROBLEM - Puppet freshness on copper is CRITICAL: Puppet has not run in the last 10 hours [11:21:28] PROBLEM - Puppet freshness on zinc is CRITICAL: Puppet has not run in the last 10 hours [11:22:24] New patchset: Dzahn; "add techblog apache_site configs - rebased - move to /misc/blogs.pp" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1433 [11:23:21] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1433 [11:23:22] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1433 [11:28:57] New patchset: Dzahn; "move blog.wm to blogs.pp, together with techblog" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1593 [11:29:58] New patchset: Dzahn; "move blog.wm to blogs.pp, together with techblog" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1593 [11:32:06] New patchset: Dzahn; "move blog.wm to blogs.pp, together with techblog - remove duplicate apache_site in techblog after merging configs" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1593 [11:32:42] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1593 [11:32:42] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1593 [11:34:46] New patchset: Dzahn; "fix apache_site name of techblog" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1594 [11:35:05] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1594 [11:35:06] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1594 [11:39:28] PROBLEM - Puppet freshness on spence is CRITICAL: Puppet has not run in the last 10 hours [11:42:54] New patchset: Dzahn; "add ocs.wikimania2009 apache_site configs - and basic puppet class - move to wikimanias.pp (wikimania special sites)" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1439 [11:45:20] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1439 [11:45:21] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1439 [11:54:48] New patchset: Dzahn; "add wm09schols, wm10schols, wm10reg apache_site configs & basic puppet classes, rebase, move to wikimanias.pp" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1435 [12:07:36] New patchset: Dzahn; "wm09schols, wm10schols, wm10reg apache_sites,rebase, move to wikimanias.pp, nicer class name structure, fix admin email" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1435 [12:09:09] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1435 [12:09:09] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1435 [12:14:13] New patchset: Dzahn; "consistent ServerAdmin email, somebody said software@ -> noc@, really for all though?" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1595 [12:15:27] New patchset: Dzahn; "consistent ServerAdmin email, somebody said software@ -> noc@, really for all though?" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1595 [12:28:10] New patchset: Mark Bergsma; "First step in cleaning up site.pp: folding node entries" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1596 [12:34:36] New review: Mark Bergsma; "(no comment)" [operations/puppet] (production); V: 0 C: 1; - https://gerrit.wikimedia.org/r/1596 [12:39:29] New patchset: Dzahn; "merge contacts.wm into a single apache config as well, move from misc-servers to outreach.pp, fix tabs, admin email" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1597 [12:41:04] New review: Mark Bergsma; "(no comment)" [operations/puppet] (production); V: 0 C: 2; - https://gerrit.wikimedia.org/r/1595 [12:41:04] Change merged: Mark Bergsma; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1595 [12:41:18] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1597 [12:41:19] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1597 [12:43:23] New review: Mark Bergsma; "Please fix the class name in the system_role definition" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1597 [12:44:38] New review: Mark Bergsma; "Can you rename misc/wikimanis.pp to misc/wikimania.pp, consistent with the class naming?" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1435 [12:47:47] New patchset: Dzahn; "sure, fix class name in system_role and rename wikimanias.pp to wikimania.pp" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1598 [12:50:45] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1598 [12:50:45] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1598 [12:51:08] New review: Mark Bergsma; "Perhaps I'm missing something, but how is this supposed to work then? Is SSL enabled on that vhost, ..." [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1575 [12:56:13] mark, it seems we might be having upload speed issues again [12:56:38] oh? [12:56:52] There's a bug in BZ saying https is slow, but secure.wm.o is ok [12:57:08] Uploading from Hume was getting 40 KB/s [12:57:14] New patchset: Dzahn; "add outreachcivi apache_site configs, merge into one file" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1432 [12:57:18] only https? [12:57:26] I was using HTTP [12:58:05] https://bugzilla.wikimedia.org/show_bug.cgi?id=31607 "Uploads at https://commons are slow/low-speed (0-20-40 KB/s), but fine with secure.wikimedia.org" [12:58:17] ok i'll look [12:58:39] New patchset: Dzahn; "add outreachcivi apache_site configs, merge into one file, fix tabs" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1432 [12:59:16] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1432 [12:59:17] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1432 [12:59:32] Thanks [13:01:47] hmmmz indeed [13:01:48] generic-receive-offload: on [13:03:04] It's not the same issue as last time, is it? [13:03:08] it is [13:03:11] but I wonder why [13:03:15] since we do set it to off [13:03:16] still it was on [13:03:30] hmm on most servers it's off [13:04:09] New patchset: Dzahn; "add outreachcivi to outreach.pp" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1599 [13:04:47] !log Found lvs5 and lvs6 with offload-gro enabled, even though it's set disabled in /etc/network/interfaces... corrected [13:04:52] weird [13:04:55] Logged the message, Master [13:04:58] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1599 [13:04:58] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1599 [13:12:34] mark, that looks a lot better, the bot is uploading a 3 meg file in under a minute now [13:12:34] Thanks [13:12:44] That's from TS, not hume [13:13:01] ok [13:13:05] but this is worrysome ;) [13:13:11] Puppet? [13:13:22] yeah but it sucks a bit [13:14:00] apergos, ^ [13:15:07] New review: Mark Bergsma; "(no comment)" [operations/puppet] (production); V: 0 C: 2; - https://gerrit.wikimedia.org/r/1596 [13:15:08] Change merged: Mark Bergsma; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1596 [13:15:37] New patchset: Dzahn; "add url.wm apache_site config, fix tabs, email (do we still use url.wm?)" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1440 [13:21:33] how is it from hume now? [13:23:03] Not sure, could stop Rotatebot and try it [13:23:31] might as well see if it's any better [13:29:13] New patchset: Mark Bergsma; "Enforce interface gro offloading in Puppet" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1600 [13:29:22] Feck [13:29:27] ? [13:29:31] 1 second to upload 4 meg :D [13:29:41] get to it :-D [13:29:52] what's today's backlog? [13:29:56] https://commons.wikimedia.org/w/index.php?limit=50&tagFilter=&title=Special%3AContributions&contribs=user&target=Reedy+RotateBot&namespace=&tagfilter=&year=&month=-1 [13:29:57] (how many files) [13:29:58] That's insane [13:30:01] 550 or so [13:30:27] ok. could you sleep 2 secs between each or so... give the scalers a little teeny breather? it will still finish fast [13:30:35] New patchset: Mark Bergsma; "Enforce interface gro offloading in Puppet" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1600 [13:30:41] Yeah [13:30:43] Well... [13:30:51] https://gerrit.wikimedia.org/r/#change,1600 [13:30:53] how's that? [13:30:54] On TS, it runs every 5 minutes via cron, if it's not already running [13:31:08] huh [13:31:43] ok well however you work it out, just so we don't pound on it too continuously [13:32:30] Reedy: however - we apparently cannot tag fast enough. I already placed notices to start mass tagging a day ago on the relevant sections. [13:32:32] hah [13:32:41] uh huh [13:33:05] New patchset: Dzahn; "add secure.wm apache site_config + basic puppet class, add proxy and proxy_http modules, fix tabs, move to secure.pp instead" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1436 [13:33:53] So it's whether we keep it on cluster, or let rotatebot do it, as that's significantly faster too [13:33:56] noone wants to review my commits :'-( [13:35:31] it was loading [13:35:38] heh, there is an extra whitespace in generic-definitions :) [13:35:46] took a while, sorry, my machine was swapping and when it does that new ff tabs take *forever* to open [13:36:10] well if you want to clear out the backlog fast running it from hume is th way to go [13:36:25] once every 5 mins seems fine to me [13:36:44] hah ff didn't even load the page [13:36:46] * apergos tries again [13:39:01] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 0 C: 1; - https://gerrit.wikimedia.org/r/1600 [13:40:33] New patchset: Mark Bergsma; "Enforce interface gro offloading in Puppet" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1600 [13:41:19] New review: Mark Bergsma; "(no comment)" [operations/puppet] (production); V: 0 C: 2; - https://gerrit.wikimedia.org/r/1600 [13:41:19] Change merged: Mark Bergsma; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1600 [13:43:51] New patchset: Mark Bergsma; "Fix syntax error" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1601 [13:44:02] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1601 [13:44:03] New review: Mark Bergsma; "(no comment)" [operations/puppet] (production); V: 0 C: 2; - https://gerrit.wikimedia.org/r/1601 [13:44:03] Change merged: Mark Bergsma; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1601 [13:44:37] why is lvs1003 still down? [13:44:47] when are the changes in /files/etc/network/interfaces applied? only at network startup right? [13:44:51] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1436 [13:44:57] ie typically at boot time ? [13:45:03] theoretically always when the network interface is brought up [13:45:38] so if the change went out and the hosts haven't had the interface cycled, they'd keep the old setting [13:47:43] New patchset: Dzahn; "add secure.wm apache site_config + basic puppet class, add proxy and proxy_http modules, fix tabs, move to secure.pp instead" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1436 [13:47:54] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1436 [13:48:26] New patchset: Dzahn; "add secure.wm apache site_config + basic puppet class, add proxy and proxy_http modules, fix tabs, move to secure.pp instead" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1436 [13:48:38] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1436 [13:48:50] New patchset: Mark Bergsma; "Remove superfluous ethtool -k parameter" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1602 [13:49:02] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1602 [13:49:07] New review: Mark Bergsma; "(no comment)" [operations/puppet] (production); V: 0 C: 2; - https://gerrit.wikimedia.org/r/1602 [13:49:08] Change merged: Mark Bergsma; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1602 [13:52:05] New patchset: Mark Bergsma; "Escape the dollar sign" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1603 [13:52:16] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1603 [13:52:21] New review: Mark Bergsma; "(no comment)" [operations/puppet] (production); V: 0 C: 2; - https://gerrit.wikimedia.org/r/1603 [13:52:21] Change merged: Mark Bergsma; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1603 [13:53:01] 4 mistakes, noone spotted them ;( [13:53:25] New review: Dzahn; "want a generic (tool?) class that ensures the ethtool package or add it to base?" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1602 [13:54:51] no [13:54:57] I hate those package classes [13:55:02] want to replace them by something else [13:55:54] oh, ok, got the impression i was supposed to use them [13:56:04] currently you are [13:56:07] since we have nothing better yet [13:56:26] yep, ok [13:56:59] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1436 [13:56:59] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1436 [13:58:45] sorry, I'm actually looking at the values on the hosts now [13:59:01] you have nothing better to do? ;) [13:59:02] k, almost got those all done (the ones that had been sitting there and all wanted to change misc-servers before) [13:59:24] I could have ignored what as written here and remain with as little knowledge as I had before you wrote anything [13:59:30] or I can take the opportunity to learn from it [14:00:36] what was/is url.wikimedia.org was actually for [14:01:10] dont know if we need it, it still has a vhost on singer, but no document root or data [14:01:21] then get rid of it [14:01:28] which hosts is it important to have this set on? [14:01:35] and, what's an alternative to ethtools? [14:01:48] go read my old private-l posts about upload slowness [14:02:17] july or september [14:04:02] I have it open [14:04:30] I still don't know beyond the vs hosts whether we want it off everywhere or whether on most hosts it doesn't matter [14:04:33] *lvs [14:04:34] Change abandoned: Dzahn; "nobody seems to miss it - or should speak up" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1440 [14:05:34] it doesn't help that I don't know what these are: 802.1Q tagged interface [14:05:42] * apergos goes to look it up [14:05:47] only on lvs servers [14:06:22] there are a lot of hosts in ALL (dsh) that must not exist any more btw... [14:06:30] New patchset: Mark Bergsma; "Fold interface_setting into interface_offload" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1604 [14:06:40] New review: gerrit2; "Change did not pass lint check. You will need to send an amended patchset for this (see: https://lab..." [operations/puppet] (production); V: -1 - https://gerrit.wikimedia.org/r/1604 [14:06:43] dsh usage does not exist anymore [14:06:51] really [14:07:01] * apergos gives up and goes back to coding [14:07:25] oh, it made me think of the RT we have that is like "generate dsh hosts files from puppet data" [14:07:35] New patchset: Mark Bergsma; "Fold interface_setting into interface_offload" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1604 [14:07:46] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1604 [14:07:50] ok, dsh ALL usage does not exist anymore [14:08:09] what do we use for the "all" list? [14:08:15] nothing [14:08:24] created a few semi-automatic, like srv and sq [14:08:36] I use puppet :P [14:08:46] New review: Mark Bergsma; "(no comment)" [operations/puppet] (production); V: 0 C: 2; - https://gerrit.wikimedia.org/r/1604 [14:08:46] Change merged: Mark Bergsma; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1604 [14:08:58] if you want a run a command live across all hosts and check the output, what do you use? we used to have this nice tool :-P [14:09:12] I rarely do [14:09:20] when I do, I just compile a fresh list of all hosts from e.g. puppet data [14:09:27] from its database or so [14:09:29] I see [14:09:37] yeah, autogenerated dsh node lists would be nice [14:09:44] it would [14:09:46] did from nagios hostlist, to check kernel versions [14:09:48] but the only really useful one is mediawiki_installation [14:09:51] the rest is so out of date [14:09:57] as they are now, apparently do [14:09:58] *so [14:10:29] I think we'll use mcollective to generate that [14:10:45] ok I have a few tabs open about 802.1Q tagged interfaces, I'll read em later [14:13:38] also the lwn article about generic-receive-offload [14:15:15] New patchset: Dzahn; "fully puppetize check_job_queue with custom check interval" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1473 [14:15:26] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1473 [14:16:06] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1473 [14:16:07] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1473 [14:16:59] apergos: ^ that makes sure the job queue check is not executed every minute anymore [14:19:59] New review: Demon; "(no comment)" [test/mediawiki/core] (master); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1574 [14:19:59] Change merged: Demon; [test/mediawiki/core] (master) - https://gerrit.wikimedia.org/r/1574 [14:20:19] yay [14:22:52] why do you need to escape the $? [14:23:38] seems like the single quotes should do it [14:23:40] ? who [14:23:43] mark [14:23:53] puppet interprets it [14:24:05] and $2 is not a puppet variable [14:24:16] nice [14:26:06] grr [14:26:10] who broke puppet on many systems [14:27:17] PROBLEM - Puppet freshness on es1002 is CRITICAL: Puppet has not run in the last 10 hours [14:33:48] mark: looks like i did, by adding nagios-plugins-basic to base :/ [14:34:04] mark: because the package is already defined in nrpe.pp .. i see that on sodium [14:34:21] can you work on fixing the issues? [14:34:24] look at puppet dashboard [14:34:38] yes [14:40:52] New patchset: Dzahn; "do not include nagios-plugins-basic here as it conflicts with class "nrpe" - real fix should be to have "nrpe" and "nrpe-light" or something, then" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1605 [14:42:30] New review: Dzahn; "(no comment)" [operations/puppet] (production); V: 1 C: 2; - https://gerrit.wikimedia.org/r/1605 [14:42:31] Change merged: Dzahn; [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1605 [14:44:08] mark: works for me again (sodium) [14:44:48] RECOVERY - Puppet freshness on sodium is OK: puppet ran at Fri Dec 16 14:43:50 UTC 2011 [14:44:48] RECOVERY - Puppet freshness on srv227 is OK: puppet ran at Fri Dec 16 14:43:52 UTC 2011 [14:44:48] RECOVERY - Puppet freshness on srv237 is OK: puppet ran at Fri Dec 16 14:43:53 UTC 2011 [14:44:48] RECOVERY - Puppet freshness on mw38 is OK: puppet ran at Fri Dec 16 14:43:57 UTC 2011 [14:44:48] RECOVERY - Puppet freshness on mw25 is OK: puppet ran at Fri Dec 16 14:43:58 UTC 2011 [14:44:48] RECOVERY - Puppet freshness on srv215 is OK: puppet ran at Fri Dec 16 14:44:02 UTC 2011 [14:44:48] RECOVERY - Puppet freshness on db1002 is OK: puppet ran at Fri Dec 16 14:44:02 UTC 2011 [14:44:49] RECOVERY - Puppet freshness on srv279 is OK: puppet ran at Fri Dec 16 14:44:04 UTC 2011 [14:44:49] RECOVERY - Puppet freshness on mw1088 is OK: puppet ran at Fri Dec 16 14:44:11 UTC 2011 [14:44:50] RECOVERY - Puppet freshness on mw32 is OK: puppet ran at Fri Dec 16 14:44:14 UTC 2011 [14:44:50] RECOVERY - Puppet freshness on locke is OK: puppet ran at Fri Dec 16 14:44:15 UTC 2011 [14:44:51] RECOVERY - Puppet freshness on mw3 is OK: puppet ran at Fri Dec 16 14:44:17 UTC 2011 [14:44:51] RECOVERY - Puppet freshness on mw21 is OK: puppet ran at Fri Dec 16 14:44:30 UTC 2011 [14:44:52] RECOVERY - Puppet freshness on db49 is OK: puppet ran at Fri Dec 16 14:44:31 UTC 2011 [14:44:52] RECOVERY - Puppet freshness on srv297 is OK: puppet ran at Fri Dec 16 14:44:38 UTC 2011 [14:44:53] RECOVERY - Puppet freshness on mw4 is OK: puppet ran at Fri Dec 16 14:44:38 UTC 2011 [14:44:53] RECOVERY - Puppet freshness on grosley is OK: puppet ran at Fri Dec 16 14:44:40 UTC 2011 [14:44:54] RECOVERY - Puppet freshness on mw47 is OK: puppet ran at Fri Dec 16 14:44:44 UTC 2011 [14:44:57] it hit all servers including "nrpe" [14:45:05] please check dashboard daily ;) [14:45:21] ok [14:45:47] RECOVERY - Puppet freshness on srv187 is OK: puppet ran at Fri Dec 16 14:44:48 UTC 2011 [14:45:48] RECOVERY - Puppet freshness on srv189 is OK: puppet ran at Fri Dec 16 14:44:55 UTC 2011 [14:45:48] RECOVERY - Puppet freshness on db47 is OK: puppet ran at Fri Dec 16 14:45:01 UTC 2011 [14:45:48] RECOVERY - Puppet freshness on srv298 is OK: puppet ran at Fri Dec 16 14:45:12 UTC 2011 [14:45:48] RECOVERY - Puppet freshness on mw1006 is OK: puppet ran at Fri Dec 16 14:45:16 UTC 2011 [14:45:48] RECOVERY - Puppet freshness on db27 is OK: puppet ran at Fri Dec 16 14:45:19 UTC 2011 [14:45:48] RECOVERY - Puppet freshness on es1004 is OK: puppet ran at Fri Dec 16 14:45:25 UTC 2011 [14:45:49] RECOVERY - Puppet freshness on db1022 is OK: puppet ran at Fri Dec 16 14:45:26 UTC 2011 [14:45:49] RECOVERY - Puppet freshness on db1044 is OK: puppet ran at Fri Dec 16 14:45:27 UTC 2011 [14:45:50] RECOVERY - Puppet freshness on snapshot2 is OK: puppet ran at Fri Dec 16 14:45:27 UTC 2011 [14:45:50] RECOVERY - Puppet freshness on mw30 is OK: puppet ran at Fri Dec 16 14:45:41 UTC 2011 [14:46:18] PROBLEM - Puppet freshness on bast1001 is CRITICAL: Puppet has not run in the last 10 hours [14:46:28] RECOVERY - Puppet freshness on srv272 is OK: puppet ran at Fri Dec 16 14:45:49 UTC 2011 [14:46:28] RECOVERY - Puppet freshness on mw1043 is OK: puppet ran at Fri Dec 16 14:45:55 UTC 2011 [14:46:28] RECOVERY - Puppet freshness on mw14 is OK: puppet ran at Fri Dec 16 14:45:55 UTC 2011 [14:46:28] RECOVERY - Puppet freshness on mw1085 is OK: puppet ran at Fri Dec 16 14:45:58 UTC 2011 [14:46:28] RECOVERY - Puppet freshness on srv256 is OK: puppet ran at Fri Dec 16 14:45:59 UTC 2011 [14:46:28] RECOVERY - Puppet freshness on db21 is OK: puppet ran at Fri Dec 16 14:46:06 UTC 2011 [14:46:28] RECOVERY - Puppet freshness on mw5 is OK: puppet ran at Fri Dec 16 14:46:08 UTC 2011 [14:46:29] RECOVERY - Puppet freshness on mw35 is OK: puppet ran at Fri Dec 16 14:46:08 UTC 2011 [14:46:38] RECOVERY - Puppet freshness on search7 is OK: puppet ran at Fri Dec 16 14:46:17 UTC 2011 [14:46:38] RECOVERY - Puppet freshness on mw6 is OK: puppet ran at Fri Dec 16 14:46:20 UTC 2011 [14:46:48] RECOVERY - Puppet freshness on mw1 is OK: puppet ran at Fri Dec 16 14:46:27 UTC 2011 [14:46:48] RECOVERY - Puppet freshness on srv241 is OK: puppet ran at Fri Dec 16 14:46:27 UTC 2011 [14:46:48] RECOVERY - Puppet freshness on srv207 is OK: puppet ran at Fri Dec 16 14:46:33 UTC 2011 [14:46:48] RECOVERY - Puppet freshness on mw43 is OK: puppet ran at Fri Dec 16 14:46:38 UTC 2011 [14:46:48] RECOVERY - Puppet freshness on srv273 is OK: puppet ran at Fri Dec 16 14:46:43 UTC 2011 [14:46:48] RECOVERY - Puppet freshness on db35 is OK: puppet ran at Fri Dec 16 14:46:43 UTC 2011 [14:46:48] RECOVERY - Puppet freshness on mw1052 is OK: puppet ran at Fri Dec 16 14:46:46 UTC 2011 [14:46:49] RECOVERY - Puppet freshness on db1039 is OK: puppet ran at Fri Dec 16 14:46:46 UTC 2011 [14:47:48] RECOVERY - Puppet freshness on mw1093 is OK: puppet ran at Fri Dec 16 14:46:48 UTC 2011 [14:47:48] RECOVERY - Puppet freshness on mw1121 is OK: puppet ran at Fri Dec 16 14:46:55 UTC 2011 [14:47:48] RECOVERY - Puppet freshness on mw1072 is OK: puppet ran at Fri Dec 16 14:46:56 UTC 2011 [14:47:48] RECOVERY - Puppet freshness on srv244 is OK: puppet ran at Fri Dec 16 14:47:18 UTC 2011 [14:47:48] RECOVERY - Puppet freshness on mw19 is OK: puppet ran at Fri Dec 16 14:47:22 UTC 2011 [14:47:48] RECOVERY - Puppet freshness on db1048 is OK: puppet ran at Fri Dec 16 14:47:27 UTC 2011 [14:47:48] RECOVERY - Puppet freshness on srv221 is OK: puppet ran at Fri Dec 16 14:47:27 UTC 2011 [14:47:49] RECOVERY - Puppet freshness on srv286 is OK: puppet ran at Fri Dec 16 14:47:35 UTC 2011 [14:47:49] RECOVERY - Puppet freshness on mw1131 is OK: puppet ran at Fri Dec 16 14:47:39 UTC 2011 [14:47:50] RECOVERY - Puppet freshness on gilman is OK: puppet ran at Fri Dec 16 14:47:39 UTC 2011 [14:47:50] RECOVERY - Puppet freshness on db1030 is OK: puppet ran at Fri Dec 16 14:47:47 UTC 2011 [14:48:18] PROBLEM - Puppet freshness on fenari is CRITICAL: Puppet has not run in the last 10 hours [14:48:23] !log puppet was broken on all servers including "nrpe" due to package conflict with nagios-plugins-basic i added to base, revert+fix [14:48:27] RECOVERY - Puppet freshness on mw1063 is OK: puppet ran at Fri Dec 16 14:47:49 UTC 2011 [14:48:27] RECOVERY - Puppet freshness on srv236 is OK: puppet ran at Fri Dec 16 14:47:51 UTC 2011 [14:48:27] RECOVERY - Puppet freshness on mw9 is OK: puppet ran at Fri Dec 16 14:47:56 UTC 2011 [14:48:27] RECOVERY - Puppet freshness on snapshot1 is OK: puppet ran at Fri Dec 16 14:47:56 UTC 2011 [14:48:28] RECOVERY - Puppet freshness on srv220 is OK: puppet ran at Fri Dec 16 14:47:59 UTC 2011 [14:48:28] RECOVERY - Puppet freshness on srv288 is OK: puppet ran at Fri Dec 16 14:48:04 UTC 2011 [14:48:28] RECOVERY - Puppet freshness on snapshot3 is OK: puppet ran at Fri Dec 16 14:48:10 UTC 2011 [14:48:29] RECOVERY - Puppet freshness on mw55 is OK: puppet ran at Fri Dec 16 14:48:12 UTC 2011 [14:48:32] Logged the message, Master [14:48:37] RECOVERY - Puppet freshness on db1038 is OK: puppet ran at Fri Dec 16 14:48:18 UTC 2011 [14:48:37] RECOVERY - Puppet freshness on search19 is OK: puppet ran at Fri Dec 16 14:48:18 UTC 2011 [14:48:37] RECOVERY - Puppet freshness on srv199 is OK: puppet ran at Fri Dec 16 14:48:18 UTC 2011 [14:48:37] RECOVERY - Puppet freshness on srv282 is OK: puppet ran at Fri Dec 16 14:48:22 UTC 2011 [14:48:38] RECOVERY - Puppet freshness on mw42 is OK: puppet ran at Fri Dec 16 14:48:25 UTC 2011 [14:48:48] RECOVERY - Puppet freshness on mw1078 is OK: puppet ran at Fri Dec 16 14:48:32 UTC 2011 [14:48:48] RECOVERY - Puppet freshness on search18 is OK: puppet ran at Fri Dec 16 14:48:34 UTC 2011 [14:48:48] RECOVERY - Puppet freshness on srv251 is OK: puppet ran at Fri Dec 16 14:48:35 UTC 2011 [14:48:48] RECOVERY - Puppet freshness on srv213 is OK: puppet ran at Fri Dec 16 14:48:40 UTC 2011 [14:48:48] RECOVERY - Puppet freshness on mw1044 is OK: puppet ran at Fri Dec 16 14:48:43 UTC 2011 [14:49:48] RECOVERY - Puppet freshness on db1007 is OK: puppet ran at Fri Dec 16 14:48:50 UTC 2011 [14:49:48] RECOVERY - Puppet freshness on srv285 is OK: puppet ran at Fri Dec 16 14:48:50 UTC 2011 [14:49:48] RECOVERY - Puppet freshness on virt2 is OK: puppet ran at Fri Dec 16 14:49:02 UTC 2011 [14:49:48] RECOVERY - Puppet freshness on mw1146 is OK: puppet ran at Fri Dec 16 14:49:08 UTC 2011 [14:49:48] RECOVERY - Puppet freshness on db1008 is OK: puppet ran at Fri Dec 16 14:49:22 UTC 2011 [14:49:48] RECOVERY - Puppet freshness on mw56 is OK: puppet ran at Fri Dec 16 14:49:26 UTC 2011 [14:49:48] RECOVERY - Puppet freshness on mw1154 is OK: puppet ran at Fri Dec 16 14:49:27 UTC 2011 [14:49:49] RECOVERY - Puppet freshness on srv295 is OK: puppet ran at Fri Dec 16 14:49:43 UTC 2011 [14:49:49] RECOVERY - Puppet freshness on mw68 is OK: puppet ran at Fri Dec 16 14:49:45 UTC 2011 [14:50:48] RECOVERY - Puppet freshness on db1001 is OK: puppet ran at Fri Dec 16 14:49:48 UTC 2011 [14:50:48] RECOVERY - Puppet freshness on mw1124 is OK: puppet ran at Fri Dec 16 14:49:50 UTC 2011 [14:50:48] RECOVERY - Puppet freshness on db1016 is OK: puppet ran at Fri Dec 16 14:49:51 UTC 2011 [14:50:48] RECOVERY - Puppet freshness on mw70 is OK: puppet ran at Fri Dec 16 14:49:54 UTC 2011 [14:50:48] RECOVERY - Puppet freshness on mw1034 is OK: puppet ran at Fri Dec 16 14:50:03 UTC 2011 [14:50:48] RECOVERY - Puppet freshness on mw61 is OK: puppet ran at Fri Dec 16 14:50:08 UTC 2011 [14:50:48] RECOVERY - Puppet freshness on srv226 is OK: puppet ran at Fri Dec 16 14:50:13 UTC 2011 [14:50:49] RECOVERY - Puppet freshness on mw1029 is OK: puppet ran at Fri Dec 16 14:50:33 UTC 2011 [14:50:49] RECOVERY - Puppet freshness on db18 is OK: puppet ran at Fri Dec 16 14:50:44 UTC 2011 [14:50:50] RECOVERY - Puppet freshness on db46 is OK: puppet ran at Fri Dec 16 14:50:45 UTC 2011 [14:51:48] RECOVERY - Puppet freshness on db1020 is OK: puppet ran at Fri Dec 16 14:50:47 UTC 2011 [14:51:48] RECOVERY - Puppet freshness on virt3 is OK: puppet ran at Fri Dec 16 14:50:49 UTC 2011 [14:51:48] RECOVERY - Puppet freshness on mw1116 is OK: puppet ran at Fri Dec 16 14:50:49 UTC 2011 [14:51:48] RECOVERY - Puppet freshness on srv269 is OK: puppet ran at Fri Dec 16 14:50:50 UTC 2011 [14:51:48] RECOVERY - Puppet freshness on search14 is OK: puppet ran at Fri Dec 16 14:50:54 UTC 2011 [14:53:48] RECOVERY - Puppet freshness on srv292 is OK: puppet ran at Fri Dec 16 14:52:48 UTC 2011 [14:53:48] RECOVERY - Puppet freshness on db36 is OK: puppet ran at Fri Dec 16 14:52:51 UTC 2011 [14:53:48] RECOVERY - Puppet freshness on db48 is OK: puppet ran at Fri Dec 16 14:52:52 UTC 2011 [14:53:48] RECOVERY - Puppet freshness on srv287 is OK: puppet ran at Fri Dec 16 14:52:53 UTC 2011 [14:53:48] RECOVERY - Puppet freshness on mw1003 is OK: puppet ran at Fri Dec 16 14:52:59 UTC 2011 [14:53:48] RECOVERY - Puppet freshness on mw1100 is OK: puppet ran at Fri Dec 16 14:53:16 UTC 2011 [14:53:48] RECOVERY - Puppet freshness on mw1130 is OK: puppet ran at Fri Dec 16 14:53:19 UTC 2011 [14:53:49] RECOVERY - Puppet freshness on db34 is OK: puppet ran at Fri Dec 16 14:53:21 UTC 2011 [14:53:49] RECOVERY - Puppet freshness on mw1105 is OK: puppet ran at Fri Dec 16 14:53:24 UTC 2011 [14:53:50] RECOVERY - Puppet freshness on mw1064 is OK: puppet ran at Fri Dec 16 14:53:28 UTC 2011 [14:53:50] RECOVERY - Puppet freshness on srv214 is OK: puppet ran at Fri Dec 16 14:53:29 UTC 2011 [14:53:51] RECOVERY - Puppet freshness on mw1028 is OK: puppet ran at Fri Dec 16 14:53:33 UTC 2011 [14:53:51] RECOVERY - Puppet freshness on db17 is OK: puppet ran at Fri Dec 16 14:53:41 UTC 2011 [14:53:52] RECOVERY - Puppet freshness on mw1014 is OK: puppet ran at Fri Dec 16 14:53:43 UTC 2011 [14:53:52] RECOVERY - Puppet freshness on mw15 is OK: puppet ran at Fri Dec 16 14:53:44 UTC 2011 [14:53:53] RECOVERY - Puppet freshness on db37 is OK: puppet ran at Fri Dec 16 14:53:45 UTC 2011 [14:54:48] RECOVERY - Puppet freshness on mw1007 is OK: puppet ran at Fri Dec 16 14:53:52 UTC 2011 [14:54:48] RECOVERY - Puppet freshness on mw1117 is OK: puppet ran at Fri Dec 16 14:53:56 UTC 2011 [14:54:48] RECOVERY - Puppet freshness on mw1032 is OK: puppet ran at Fri Dec 16 14:54:00 UTC 2011 [14:54:48] RECOVERY - Puppet freshness on es3 is OK: puppet ran at Fri Dec 16 14:54:03 UTC 2011 [14:54:48] RECOVERY - Puppet freshness on srv201 is OK: puppet ran at Fri Dec 16 14:54:05 UTC 2011 [14:54:48] RECOVERY - Puppet freshness on srv265 is OK: puppet ran at Fri Dec 16 14:54:08 UTC 2011 [14:54:48] RECOVERY - Puppet freshness on mw1015 is OK: puppet ran at Fri Dec 16 14:54:13 UTC 2011 [14:54:49] RECOVERY - Puppet freshness on mw1082 is OK: puppet ran at Fri Dec 16 14:54:18 UTC 2011 [14:54:49] RECOVERY - Puppet freshness on mw62 is OK: puppet ran at Fri Dec 16 14:54:21 UTC 2011 [14:54:50] RECOVERY - Puppet freshness on mw1035 is OK: puppet ran at Fri Dec 16 14:54:23 UTC 2011 [14:54:50] RECOVERY - Puppet freshness on mw1048 is OK: puppet ran at Fri Dec 16 14:54:24 UTC 2011 [14:54:51] RECOVERY - Puppet freshness on mw1040 is OK: puppet ran at Fri Dec 16 14:54:34 UTC 2011 [14:54:51] RECOVERY - Puppet freshness on mw1091 is OK: puppet ran at Fri Dec 16 14:54:39 UTC 2011 [14:54:52] RECOVERY - Puppet freshness on srv263 is OK: puppet ran at Fri Dec 16 14:54:40 UTC 2011 [14:54:52] RECOVERY - Puppet freshness on mw1125 is OK: puppet ran at Fri Dec 16 14:54:40 UTC 2011 [14:54:53] RECOVERY - Puppet freshness on srv254 is OK: puppet ran at Fri Dec 16 14:54:41 UTC 2011 [14:55:48] RECOVERY - Puppet freshness on mw29 is OK: puppet ran at Fri Dec 16 14:54:49 UTC 2011 [14:55:48] RECOVERY - Puppet freshness on srv219 is OK: puppet ran at Fri Dec 16 14:54:50 UTC 2011 [14:55:48] RECOVERY - Puppet freshness on mw39 is OK: puppet ran at Fri Dec 16 14:54:52 UTC 2011 [14:55:48] RECOVERY - Puppet freshness on srv196 is OK: puppet ran at Fri Dec 16 14:54:52 UTC 2011 [14:55:48] RECOVERY - Puppet freshness on db9 is OK: puppet ran at Fri Dec 16 14:54:55 UTC 2011 [14:55:48] RECOVERY - Puppet freshness on ms5 is OK: puppet ran at Fri Dec 16 14:55:02 UTC 2011 [14:55:48] RECOVERY - Puppet freshness on search4 is OK: puppet ran at Fri Dec 16 14:55:04 UTC 2011 [14:55:49] RECOVERY - Puppet freshness on srv293 is OK: puppet ran at Fri Dec 16 14:55:07 UTC 2011 [14:55:49] RECOVERY - Puppet freshness on mw1051 is OK: puppet ran at Fri Dec 16 14:55:15 UTC 2011 [14:55:50] RECOVERY - Puppet freshness on db1043 is OK: puppet ran at Fri Dec 16 14:55:19 UTC 2011 [14:55:50] RECOVERY - Puppet freshness on db1012 is OK: puppet ran at Fri Dec 16 14:55:24 UTC 2011 [14:55:51] RECOVERY - Puppet freshness on db42 is OK: puppet ran at Fri Dec 16 14:55:31 UTC 2011 [14:55:51] RECOVERY - Puppet freshness on mw1030 is OK: puppet ran at Fri Dec 16 14:55:39 UTC 2011 [14:56:48] RECOVERY - Puppet freshness on nfs2 is OK: puppet ran at Fri Dec 16 14:55:55 UTC 2011 [14:56:48] RECOVERY - Puppet freshness on srv255 is OK: puppet ran at Fri Dec 16 14:56:07 UTC 2011 [14:56:48] RECOVERY - Puppet freshness on mw1159 is OK: puppet ran at Fri Dec 16 14:56:09 UTC 2011 [14:56:48] RECOVERY - Puppet freshness on search9 is OK: puppet ran at Fri Dec 16 14:56:10 UTC 2011 [14:56:48] RECOVERY - Puppet freshness on mw1008 is OK: puppet ran at Fri Dec 16 14:56:10 UTC 2011 [14:56:48] RECOVERY - Puppet freshness on db16 is OK: puppet ran at Fri Dec 16 14:56:16 UTC 2011 [14:56:48] RECOVERY - Puppet freshness on mw1089 is OK: puppet ran at Fri Dec 16 14:56:17 UTC 2011 [14:56:49] RECOVERY - Puppet freshness on db1026 is OK: puppet ran at Fri Dec 16 14:56:24 UTC 2011 [14:56:49] RECOVERY - Puppet freshness on db1046 is OK: puppet ran at Fri Dec 16 14:56:30 UTC 2011 [14:56:50] RECOVERY - Puppet freshness on db1024 is OK: puppet ran at Fri Dec 16 14:56:34 UTC 2011 [14:56:50] RECOVERY - Puppet freshness on mw1141 is OK: puppet ran at Fri Dec 16 14:56:34 UTC 2011 [14:56:51] RECOVERY - Puppet freshness on aluminium is OK: puppet ran at Fri Dec 16 14:56:38 UTC 2011 [14:56:51] RECOVERY - Puppet freshness on mw1115 is OK: puppet ran at Fri Dec 16 14:56:40 UTC 2011 [14:56:52] RECOVERY - Puppet freshness on mw1140 is OK: puppet ran at Fri Dec 16 14:56:41 UTC 2011 [14:57:48] RECOVERY - Puppet freshness on srv195 is OK: puppet ran at Fri Dec 16 14:56:50 UTC 2011 [14:57:48] RECOVERY - Puppet freshness on mw1090 is OK: puppet ran at Fri Dec 16 14:56:58 UTC 2011 [14:57:48] RECOVERY - Puppet freshness on mw58 is OK: puppet ran at Fri Dec 16 14:57:00 UTC 2011 [14:57:48] RECOVERY - Puppet freshness on mw1060 is OK: puppet ran at Fri Dec 16 14:57:01 UTC 2011 [14:57:48] RECOVERY - Puppet freshness on db1017 is OK: puppet ran at Fri Dec 16 14:57:02 UTC 2011 [14:57:48] RECOVERY - Puppet freshness on mw1142 is OK: puppet ran at Fri Dec 16 14:57:11 UTC 2011 [14:57:48] RECOVERY - Puppet freshness on db1003 is OK: puppet ran at Fri Dec 16 14:57:11 UTC 2011 [14:57:49] RECOVERY - Puppet freshness on mw1001 is OK: puppet ran at Fri Dec 16 14:57:15 UTC 2011 [14:57:49] RECOVERY - Puppet freshness on mw1065 is OK: puppet ran at Fri Dec 16 14:57:15 UTC 2011 [14:57:50] RECOVERY - Puppet freshness on storage3 is OK: puppet ran at Fri Dec 16 14:57:22 UTC 2011 [14:57:50] RECOVERY - Puppet freshness on mw1024 is OK: puppet ran at Fri Dec 16 14:57:26 UTC 2011 [14:57:51] RECOVERY - Puppet freshness on mw63 is OK: puppet ran at Fri Dec 16 14:57:26 UTC 2011 [14:57:51] RECOVERY - Puppet freshness on mw1144 is OK: puppet ran at Fri Dec 16 14:57:27 UTC 2011 [14:57:52] RECOVERY - Puppet freshness on mw1004 is OK: puppet ran at Fri Dec 16 14:57:32 UTC 2011 [14:57:52] RECOVERY - Puppet freshness on srv240 is OK: puppet ran at Fri Dec 16 14:57:34 UTC 2011 [14:57:53] RECOVERY - Puppet freshness on srv280 is OK: puppet ran at Fri Dec 16 14:57:35 UTC 2011 [14:57:53] RECOVERY - Puppet freshness on mw45 is OK: puppet ran at Fri Dec 16 14:57:35 UTC 2011 [14:57:54] RECOVERY - Puppet freshness on db30 is OK: puppet ran at Fri Dec 16 14:57:40 UTC 2011 [14:57:54] RECOVERY - Puppet freshness on srv270 is OK: puppet ran at Fri Dec 16 14:57:46 UTC 2011 [14:58:48] RECOVERY - Puppet freshness on mw57 is OK: puppet ran at Fri Dec 16 14:57:47 UTC 2011 [14:58:48] RECOVERY - Puppet freshness on mw1155 is OK: puppet ran at Fri Dec 16 14:57:48 UTC 2011 [14:58:48] RECOVERY - Puppet freshness on db1009 is OK: puppet ran at Fri Dec 16 14:57:48 UTC 2011 [14:58:48] RECOVERY - Puppet freshness on srv262 is OK: puppet ran at Fri Dec 16 14:57:49 UTC 2011 [14:58:48] RECOVERY - Puppet freshness on mw1037 is OK: puppet ran at Fri Dec 16 14:57:52 UTC 2011 [14:58:48] RECOVERY - Puppet freshness on mw1104 is OK: puppet ran at Fri Dec 16 14:57:54 UTC 2011 [14:58:48] RECOVERY - Puppet freshness on srv198 is OK: puppet ran at Fri Dec 16 14:57:54 UTC 2011 [14:58:49] RECOVERY - Puppet freshness on mw1020 is OK: puppet ran at Fri Dec 16 14:57:57 UTC 2011 [14:58:49] RECOVERY - Puppet freshness on mw1057 is OK: puppet ran at Fri Dec 16 14:58:02 UTC 2011 [14:58:50] RECOVERY - Puppet freshness on mw1127 is OK: puppet ran at Fri Dec 16 14:58:08 UTC 2011 [14:58:50] RECOVERY - Puppet freshness on search3 is OK: puppet ran at Fri Dec 16 14:58:10 UTC 2011 [14:58:51] RECOVERY - Puppet freshness on srv228 is OK: puppet ran at Fri Dec 16 14:58:19 UTC 2011 [14:58:51] RECOVERY - Puppet freshness on mw1041 is OK: puppet ran at Fri Dec 16 14:58:19 UTC 2011 [14:58:52] RECOVERY - Puppet freshness on mw1126 is OK: puppet ran at Fri Dec 16 14:58:20 UTC 2011 [14:58:52] RECOVERY - Puppet freshness on mw27 is OK: puppet ran at Fri Dec 16 14:58:23 UTC 2011 [14:58:53] RECOVERY - Puppet freshness on db29 is OK: puppet ran at Fri Dec 16 14:58:38 UTC 2011 [14:58:53] RECOVERY - Puppet freshness on ms1004 is OK: puppet ran at Fri Dec 16 14:58:43 UTC 2011 [14:59:48] RECOVERY - Puppet freshness on mw1077 is OK: puppet ran at Fri Dec 16 14:59:13 UTC 2011 [14:59:48] RECOVERY - Puppet freshness on srv247 is OK: puppet ran at Fri Dec 16 14:59:27 UTC 2011 [14:59:48] RECOVERY - Puppet freshness on mw34 is OK: puppet ran at Fri Dec 16 14:59:28 UTC 2011 [14:59:48] RECOVERY - Puppet freshness on db1035 is OK: puppet ran at Fri Dec 16 14:59:32 UTC 2011 [14:59:48] RECOVERY - Puppet freshness on db40 is OK: puppet ran at Fri Dec 16 14:59:34 UTC 2011 [14:59:48] RECOVERY - Puppet freshness on mw1149 is OK: puppet ran at Fri Dec 16 14:59:34 UTC 2011 [14:59:48] RECOVERY - Puppet freshness on mw1076 is OK: puppet ran at Fri Dec 16 14:59:34 UTC 2011 [14:59:49] RECOVERY - Puppet freshness on mw1069 is OK: puppet ran at Fri Dec 16 14:59:40 UTC 2011 [14:59:49] RECOVERY - Puppet freshness on mw1156 is OK: puppet ran at Fri Dec 16 14:59:42 UTC 2011 [14:59:50] RECOVERY - Puppet freshness on mw10 is OK: puppet ran at Fri Dec 16 14:59:44 UTC 2011 [15:00:48] RECOVERY - Puppet freshness on mw18 is OK: puppet ran at Fri Dec 16 14:59:52 UTC 2011 [15:00:48] RECOVERY - Puppet freshness on db1042 is OK: puppet ran at Fri Dec 16 15:00:04 UTC 2011 [15:00:48] RECOVERY - Puppet freshness on srv271 is OK: puppet ran at Fri Dec 16 15:00:10 UTC 2011 [15:00:48] RECOVERY - Puppet freshness on es4 is OK: puppet ran at Fri Dec 16 15:00:11 UTC 2011 [15:00:48] RECOVERY - Puppet freshness on sockpuppet is OK: puppet ran at Fri Dec 16 15:00:13 UTC 2011 [15:00:48] RECOVERY - Puppet freshness on mw1129 is OK: puppet ran at Fri Dec 16 15:00:14 UTC 2011 [15:00:48] RECOVERY - Puppet freshness on mw8 is OK: puppet ran at Fri Dec 16 15:00:15 UTC 2011 [15:00:49] RECOVERY - Puppet freshness on mw33 is OK: puppet ran at Fri Dec 16 15:00:22 UTC 2011 [15:00:49] RECOVERY - Puppet freshness on mw28 is OK: puppet ran at Fri Dec 16 15:00:44 UTC 2011 [15:00:50] RECOVERY - Puppet freshness on mw52 is OK: puppet ran at Fri Dec 16 15:00:46 UTC 2011 [15:01:48] RECOVERY - Puppet freshness on nfs1 is OK: puppet ran at Fri Dec 16 15:00:47 UTC 2011 [15:01:48] RECOVERY - Puppet freshness on mw1002 is OK: puppet ran at Fri Dec 16 15:00:50 UTC 2011 [15:01:48] RECOVERY - Puppet freshness on es2 is OK: puppet ran at Fri Dec 16 15:00:52 UTC 2011 [15:01:48] RECOVERY - Puppet freshness on srv203 is OK: puppet ran at Fri Dec 16 15:00:57 UTC 2011 [15:01:48] RECOVERY - Puppet freshness on mw1068 is OK: puppet ran at Fri Dec 16 15:00:58 UTC 2011 [15:01:48] RECOVERY - Puppet freshness on mw1114 is OK: puppet ran at Fri Dec 16 15:01:01 UTC 2011 [15:01:48] RECOVERY - Puppet freshness on mw1123 is OK: puppet ran at Fri Dec 16 15:01:05 UTC 2011 [15:01:49] RECOVERY - Puppet freshness on mw11 is OK: puppet ran at Fri Dec 16 15:01:08 UTC 2011 [15:01:49] RECOVERY - Puppet freshness on srv258 is OK: puppet ran at Fri Dec 16 15:01:15 UTC 2011 [15:01:50] RECOVERY - Puppet freshness on srv276 is OK: puppet ran at Fri Dec 16 15:01:17 UTC 2011 [15:01:50] RECOVERY - Puppet freshness on mw1083 is OK: puppet ran at Fri Dec 16 15:01:18 UTC 2011 [15:01:51] RECOVERY - Puppet freshness on mw1135 is OK: puppet ran at Fri Dec 16 15:01:23 UTC 2011 [15:01:51] RECOVERY - Puppet freshness on srv277 is OK: puppet ran at Fri Dec 16 15:01:29 UTC 2011 [15:01:52] RECOVERY - Puppet freshness on db33 is OK: puppet ran at Fri Dec 16 15:01:32 UTC 2011 [15:01:52] RECOVERY - Puppet freshness on mw1073 is OK: puppet ran at Fri Dec 16 15:01:38 UTC 2011 [15:01:53] RECOVERY - Puppet freshness on db1014 is OK: puppet ran at Fri Dec 16 15:01:40 UTC 2011 [15:01:53] RECOVERY - Puppet freshness on mw1148 is OK: puppet ran at Fri Dec 16 15:01:42 UTC 2011 [15:02:48] RECOVERY - Puppet freshness on mw1097 is OK: puppet ran at Fri Dec 16 15:01:51 UTC 2011 [15:02:48] RECOVERY - Puppet freshness on mw1118 is OK: puppet ran at Fri Dec 16 15:01:56 UTC 2011 [15:02:48] RECOVERY - Puppet freshness on mw73 is OK: puppet ran at Fri Dec 16 15:02:01 UTC 2011 [15:02:48] RECOVERY - Puppet freshness on mw48 is OK: puppet ran at Fri Dec 16 15:02:02 UTC 2011 [15:02:48] RECOVERY - Puppet freshness on search20 is OK: puppet ran at Fri Dec 16 15:02:03 UTC 2011 [15:02:48] RECOVERY - Puppet freshness on mw17 is OK: puppet ran at Fri Dec 16 15:02:09 UTC 2011 [15:02:48] RECOVERY - Puppet freshness on mw1009 is OK: puppet ran at Fri Dec 16 15:02:10 UTC 2011 [15:02:49] RECOVERY - Puppet freshness on mw41 is OK: puppet ran at Fri Dec 16 15:02:11 UTC 2011 [15:02:49] RECOVERY - Puppet freshness on mw1098 is OK: puppet ran at Fri Dec 16 15:02:13 UTC 2011 [15:02:50] RECOVERY - Puppet freshness on mw67 is OK: puppet ran at Fri Dec 16 15:02:23 UTC 2011 [15:02:50] RECOVERY - Puppet freshness on srv192 is OK: puppet ran at Fri Dec 16 15:02:26 UTC 2011 [15:02:51] RECOVERY - Puppet freshness on db1025 is OK: puppet ran at Fri Dec 16 15:02:27 UTC 2011 [15:02:51] RECOVERY - Puppet freshness on mw66 is OK: puppet ran at Fri Dec 16 15:02:35 UTC 2011 [15:02:52] RECOVERY - Puppet freshness on mw1106 is OK: puppet ran at Fri Dec 16 15:02:39 UTC 2011 [15:03:48] RECOVERY - Puppet freshness on db23 is OK: puppet ran at Fri Dec 16 15:02:51 UTC 2011 [15:03:48] RECOVERY - Puppet freshness on srv267 is OK: puppet ran at Fri Dec 16 15:02:55 UTC 2011 [15:03:48] RECOVERY - Puppet freshness on mw36 is OK: puppet ran at Fri Dec 16 15:03:08 UTC 2011 [15:03:48] RECOVERY - Puppet freshness on srv224 is OK: puppet ran at Fri Dec 16 15:03:10 UTC 2011 [15:03:48] RECOVERY - Puppet freshness on srv239 is OK: puppet ran at Fri Dec 16 15:03:14 UTC 2011 [15:03:48] RECOVERY - Puppet freshness on srv204 is OK: puppet ran at Fri Dec 16 15:03:17 UTC 2011 [15:03:48] RECOVERY - Puppet freshness on srv249 is OK: puppet ran at Fri Dec 16 15:03:17 UTC 2011 [15:03:49] RECOVERY - Puppet freshness on srv278 is OK: puppet ran at Fri Dec 16 15:03:17 UTC 2011 [15:03:49] RECOVERY - Puppet freshness on srv294 is OK: puppet ran at Fri Dec 16 15:03:22 UTC 2011 [15:03:50] RECOVERY - Puppet freshness on mw1096 is OK: puppet ran at Fri Dec 16 15:03:23 UTC 2011 [15:03:50] RECOVERY - Puppet freshness on db1031 is OK: puppet ran at Fri Dec 16 15:03:24 UTC 2011 [15:03:51] RECOVERY - Puppet freshness on mw1119 is OK: puppet ran at Fri Dec 16 15:03:37 UTC 2011 [15:03:51] RECOVERY - Puppet freshness on mw1134 is OK: puppet ran at Fri Dec 16 15:03:37 UTC 2011 [15:03:52] RECOVERY - Puppet freshness on mw1023 is OK: puppet ran at Fri Dec 16 15:03:38 UTC 2011 [15:03:52] RECOVERY - Puppet freshness on srv191 is OK: puppet ran at Fri Dec 16 15:03:39 UTC 2011 [15:03:53] RECOVERY - Puppet freshness on db39 is OK: puppet ran at Fri Dec 16 15:03:41 UTC 2011 [15:03:53] RECOVERY - Puppet freshness on db10 is OK: puppet ran at Fri Dec 16 15:03:42 UTC 2011 [15:03:54] RECOVERY - Puppet freshness on mw1084 is OK: puppet ran at Fri Dec 16 15:03:43 UTC 2011 [15:03:54] RECOVERY - Puppet freshness on search15 is OK: puppet ran at Fri Dec 16 15:03:45 UTC 2011 [15:04:48] RECOVERY - Puppet freshness on db1013 is OK: puppet ran at Fri Dec 16 15:03:50 UTC 2011 [15:04:48] RECOVERY - Puppet freshness on srv225 is OK: puppet ran at Fri Dec 16 15:03:50 UTC 2011 [15:04:48] RECOVERY - Puppet freshness on hume is OK: puppet ran at Fri Dec 16 15:03:53 UTC 2011 [15:04:48] RECOVERY - Puppet freshness on mw49 is OK: puppet ran at Fri Dec 16 15:03:54 UTC 2011 [15:04:48] RECOVERY - Puppet freshness on mw1111 is OK: puppet ran at Fri Dec 16 15:04:07 UTC 2011 [15:04:48] RECOVERY - Puppet freshness on mw1153 is OK: puppet ran at Fri Dec 16 15:04:15 UTC 2011 [15:04:48] RECOVERY - Puppet freshness on es1 is OK: puppet ran at Fri Dec 16 15:04:17 UTC 2011 [15:04:49] RECOVERY - Puppet freshness on search6 is OK: puppet ran at Fri Dec 16 15:04:25 UTC 2011 [15:04:49] RECOVERY - Puppet freshness on mw1132 is OK: puppet ran at Fri Dec 16 15:04:26 UTC 2011 [15:04:50] RECOVERY - Puppet freshness on srv243 is OK: puppet ran at Fri Dec 16 15:04:28 UTC 2011 [15:04:50] RECOVERY - Puppet freshness on virt4 is OK: puppet ran at Fri Dec 16 15:04:31 UTC 2011 [15:04:51] RECOVERY - Puppet freshness on db44 is OK: puppet ran at Fri Dec 16 15:04:31 UTC 2011 [15:04:51] RECOVERY - Puppet freshness on search16 is OK: puppet ran at Fri Dec 16 15:04:33 UTC 2011 [15:04:52] RECOVERY - Puppet freshness on srv202 is OK: puppet ran at Fri Dec 16 15:04:38 UTC 2011 [15:04:52] RECOVERY - Puppet freshness on mw12 is OK: puppet ran at Fri Dec 16 15:04:45 UTC 2011 [15:04:53] RECOVERY - Puppet freshness on mw1011 is OK: puppet ran at Fri Dec 16 15:04:47 UTC 2011 [15:05:48] RECOVERY - Puppet freshness on srv257 is OK: puppet ran at Fri Dec 16 15:04:48 UTC 2011 [15:05:48] RECOVERY - Puppet freshness on mw1160 is OK: puppet ran at Fri Dec 16 15:04:53 UTC 2011 [15:05:48] RECOVERY - Puppet freshness on db12 is OK: puppet ran at Fri Dec 16 15:04:55 UTC 2011 [15:05:48] RECOVERY - Puppet freshness on mw1075 is OK: puppet ran at Fri Dec 16 15:04:57 UTC 2011 [15:05:48] RECOVERY - Puppet freshness on mw1143 is OK: puppet ran at Fri Dec 16 15:04:59 UTC 2011 [15:05:48] RECOVERY - Puppet freshness on mw1062 is OK: puppet ran at Fri Dec 16 15:05:04 UTC 2011 [15:05:48] RECOVERY - Puppet freshness on srv248 is OK: puppet ran at Fri Dec 16 15:05:04 UTC 2011 [15:05:49] RECOVERY - Puppet freshness on mw1101 is OK: puppet ran at Fri Dec 16 15:05:12 UTC 2011 [15:05:49] RECOVERY - Puppet freshness on mw20 is OK: puppet ran at Fri Dec 16 15:05:16 UTC 2011 [15:05:50] RECOVERY - Puppet freshness on mw1059 is OK: puppet ran at Fri Dec 16 15:05:22 UTC 2011 [15:05:50] RECOVERY - Puppet freshness on searchidx2 is OK: puppet ran at Fri Dec 16 15:05:30 UTC 2011 [15:05:51] RECOVERY - Puppet freshness on mw59 is OK: puppet ran at Fri Dec 16 15:05:46 UTC 2011 [15:05:54] yay [15:06:48] RECOVERY - Puppet freshness on db24 is OK: puppet ran at Fri Dec 16 15:05:49 UTC 2011 [15:06:48] RECOVERY - Puppet freshness on stafford is OK: puppet ran at Fri Dec 16 15:05:50 UTC 2011 [15:06:48] RECOVERY - Puppet freshness on db1019 is OK: puppet ran at Fri Dec 16 15:05:50 UTC 2011 [15:06:48] RECOVERY - Puppet freshness on srv238 is OK: puppet ran at Fri Dec 16 15:05:55 UTC 2011 [15:06:48] RECOVERY - Puppet freshness on srv233 is OK: puppet ran at Fri Dec 16 15:06:01 UTC 2011 [15:06:48] RECOVERY - Puppet freshness on mw44 is OK: puppet ran at Fri Dec 16 15:06:03 UTC 2011 [15:06:48] RECOVERY - Puppet freshness on professor is OK: puppet ran at Fri Dec 16 15:06:10 UTC 2011 [15:06:49] RECOVERY - Puppet freshness on srv299 is OK: puppet ran at Fri Dec 16 15:06:11 UTC 2011 [15:06:49] RECOVERY - Puppet freshness on srv188 is OK: puppet ran at Fri Dec 16 15:06:16 UTC 2011 [15:06:50] RECOVERY - Puppet freshness on mw13 is OK: puppet ran at Fri Dec 16 15:06:22 UTC 2011 [15:06:50] RECOVERY - Puppet freshness on mw72 is OK: puppet ran at Fri Dec 16 15:06:23 UTC 2011 [15:06:51] RECOVERY - Puppet freshness on db1010 is OK: puppet ran at Fri Dec 16 15:06:26 UTC 2011 [15:06:51] RECOVERY - Puppet freshness on mw1058 is OK: puppet ran at Fri Dec 16 15:06:29 UTC 2011 [15:06:52] RECOVERY - Puppet freshness on mw1158 is OK: puppet ran at Fri Dec 16 15:06:30 UTC 2011 [15:06:52] RECOVERY - Puppet freshness on srv290 is OK: puppet ran at Fri Dec 16 15:06:34 UTC 2011 [15:06:53] RECOVERY - Puppet freshness on srv268 is OK: puppet ran at Fri Dec 16 15:06:37 UTC 2011 [15:06:53] RECOVERY - Puppet freshness on srv234 is OK: puppet ran at Fri Dec 16 15:06:38 UTC 2011 [15:06:54] RECOVERY - Puppet freshness on mw1122 is OK: puppet ran at Fri Dec 16 15:06:41 UTC 2011 [15:07:48] RECOVERY - Puppet freshness on mw1128 is OK: puppet ran at Fri Dec 16 15:06:51 UTC 2011 [15:07:48] RECOVERY - Puppet freshness on db1004 is OK: puppet ran at Fri Dec 16 15:06:56 UTC 2011 [15:07:48] RECOVERY - Puppet freshness on srv274 is OK: puppet ran at Fri Dec 16 15:07:03 UTC 2011 [15:07:48] RECOVERY - Puppet freshness on mw1087 is OK: puppet ran at Fri Dec 16 15:07:05 UTC 2011 [15:07:48] RECOVERY - Puppet freshness on mw1013 is OK: puppet ran at Fri Dec 16 15:07:12 UTC 2011 [15:07:48] RECOVERY - Puppet freshness on mw1151 is OK: puppet ran at Fri Dec 16 15:07:18 UTC 2011 [15:07:48] RECOVERY - Puppet freshness on mw64 is OK: puppet ran at Fri Dec 16 15:07:20 UTC 2011 [15:07:49] RECOVERY - Puppet freshness on db1033 is OK: puppet ran at Fri Dec 16 15:07:26 UTC 2011 [15:07:49] RECOVERY - Puppet freshness on mw1147 is OK: puppet ran at Fri Dec 16 15:07:26 UTC 2011 [15:07:50] RECOVERY - Puppet freshness on mw1017 is OK: puppet ran at Fri Dec 16 15:07:36 UTC 2011 [15:07:50] RECOVERY - Puppet freshness on srv264 is OK: puppet ran at Fri Dec 16 15:07:39 UTC 2011 [15:07:51] RECOVERY - Puppet freshness on mw51 is OK: puppet ran at Fri Dec 16 15:07:46 UTC 2011 [15:07:51] RECOVERY - Puppet freshness on srv205 is OK: puppet ran at Fri Dec 16 15:07:46 UTC 2011 [15:08:48] RECOVERY - Puppet freshness on mw1074 is OK: puppet ran at Fri Dec 16 15:07:54 UTC 2011 [15:08:48] RECOVERY - Puppet freshness on srv209 is OK: puppet ran at Fri Dec 16 15:07:57 UTC 2011 [15:08:48] RECOVERY - Puppet freshness on mw2 is OK: puppet ran at Fri Dec 16 15:08:00 UTC 2011 [15:08:48] RECOVERY - Puppet freshness on mw1016 is OK: puppet ran at Fri Dec 16 15:08:00 UTC 2011 [15:08:48] RECOVERY - Puppet freshness on mw40 is OK: puppet ran at Fri Dec 16 15:08:13 UTC 2011 [15:08:48] RECOVERY - Puppet freshness on search12 is OK: puppet ran at Fri Dec 16 15:08:15 UTC 2011 [15:08:48] RECOVERY - Puppet freshness on srv291 is OK: puppet ran at Fri Dec 16 15:08:18 UTC 2011 [15:08:49] RECOVERY - Puppet freshness on db1040 is OK: puppet ran at Fri Dec 16 15:08:20 UTC 2011 [15:08:49] RECOVERY - Puppet freshness on mw1086 is OK: puppet ran at Fri Dec 16 15:08:20 UTC 2011 [15:08:50] RECOVERY - Puppet freshness on mw53 is OK: puppet ran at Fri Dec 16 15:08:21 UTC 2011 [15:08:50] RECOVERY - Puppet freshness on mw16 is OK: puppet ran at Fri Dec 16 15:08:22 UTC 2011 [15:08:51] RECOVERY - Puppet freshness on db1027 is OK: puppet ran at Fri Dec 16 15:08:26 UTC 2011 [15:08:51] RECOVERY - Puppet freshness on mw1136 is OK: puppet ran at Fri Dec 16 15:08:27 UTC 2011 [15:08:52] RECOVERY - Puppet freshness on emery is OK: puppet ran at Fri Dec 16 15:08:28 UTC 2011 [15:08:52] RECOVERY - Puppet freshness on db1028 is OK: puppet ran at Fri Dec 16 15:08:34 UTC 2011 [15:08:53] RECOVERY - Puppet freshness on db5 is OK: puppet ran at Fri Dec 16 15:08:37 UTC 2011 [15:08:53] RECOVERY - Puppet freshness on mw1145 is OK: puppet ran at Fri Dec 16 15:08:38 UTC 2011 [15:08:54] RECOVERY - Puppet freshness on db13 is OK: puppet ran at Fri Dec 16 15:08:39 UTC 2011 [15:09:48] RECOVERY - Puppet freshness on mw65 is OK: puppet ran at Fri Dec 16 15:08:48 UTC 2011 [15:09:48] RECOVERY - Puppet freshness on mw1070 is OK: puppet ran at Fri Dec 16 15:08:50 UTC 2011 [15:09:48] RECOVERY - Puppet freshness on mw1019 is OK: puppet ran at Fri Dec 16 15:08:52 UTC 2011 [15:09:48] RECOVERY - Puppet freshness on mw69 is OK: puppet ran at Fri Dec 16 15:08:53 UTC 2011 [15:09:48] RECOVERY - Puppet freshness on mw1067 is OK: puppet ran at Fri Dec 16 15:09:00 UTC 2011 [15:09:48] RECOVERY - Puppet freshness on srv283 is OK: puppet ran at Fri Dec 16 15:09:02 UTC 2011 [15:09:48] RECOVERY - Puppet freshness on mw1095 is OK: puppet ran at Fri Dec 16 15:09:04 UTC 2011 [15:09:49] RECOVERY - Puppet freshness on ms1002 is OK: puppet ran at Fri Dec 16 15:09:09 UTC 2011 [15:09:49] RECOVERY - Puppet freshness on mw1102 is OK: puppet ran at Fri Dec 16 15:09:14 UTC 2011 [15:09:50] RECOVERY - Puppet freshness on db1021 is OK: puppet ran at Fri Dec 16 15:09:18 UTC 2011 [15:09:50] RECOVERY - Puppet freshness on mw37 is OK: puppet ran at Fri Dec 16 15:09:19 UTC 2011 [15:09:51] RECOVERY - Puppet freshness on srv260 is OK: puppet ran at Fri Dec 16 15:09:22 UTC 2011 [15:09:51] RECOVERY - Puppet freshness on srv235 is OK: puppet ran at Fri Dec 16 15:09:29 UTC 2011 [15:09:52] RECOVERY - Puppet freshness on search13 is OK: puppet ran at Fri Dec 16 15:09:30 UTC 2011 [15:09:52] RECOVERY - Puppet freshness on srv252 is OK: puppet ran at Fri Dec 16 15:09:31 UTC 2011 [15:09:53] RECOVERY - Puppet freshness on srv190 is OK: puppet ran at Fri Dec 16 15:09:32 UTC 2011 [15:09:53] RECOVERY - Puppet freshness on mw24 is OK: puppet ran at Fri Dec 16 15:09:37 UTC 2011 [15:09:54] RECOVERY - Puppet freshness on search1 is OK: puppet ran at Fri Dec 16 15:09:38 UTC 2011 [15:09:54] RECOVERY - Puppet freshness on mw1112 is OK: puppet ran at Fri Dec 16 15:09:39 UTC 2011 [15:09:55] RECOVERY - Puppet freshness on mw1050 is OK: puppet ran at Fri Dec 16 15:09:39 UTC 2011 [15:09:55] RECOVERY - Puppet freshness on es1001 is OK: puppet ran at Fri Dec 16 15:09:41 UTC 2011 [15:10:48] RECOVERY - Puppet freshness on srv212 is OK: puppet ran at Fri Dec 16 15:09:47 UTC 2011 [15:10:48] RECOVERY - Puppet freshness on mw1047 is OK: puppet ran at Fri Dec 16 15:09:55 UTC 2011 [15:10:48] RECOVERY - Puppet freshness on db8 is OK: puppet ran at Fri Dec 16 15:09:56 UTC 2011 [15:10:48] RECOVERY - Puppet freshness on db32 is OK: puppet ran at Fri Dec 16 15:10:01 UTC 2011 [15:10:48] RECOVERY - Puppet freshness on srv275 is OK: puppet ran at Fri Dec 16 15:10:03 UTC 2011 [15:10:48] RECOVERY - Puppet freshness on mw1005 is OK: puppet ran at Fri Dec 16 15:10:04 UTC 2011 [15:10:48] RECOVERY - Puppet freshness on db11 is OK: puppet ran at Fri Dec 16 15:10:14 UTC 2011 [15:10:49] RECOVERY - Puppet freshness on srv229 is OK: puppet ran at Fri Dec 16 15:10:16 UTC 2011 [15:10:49] RECOVERY - Puppet freshness on mw1031 is OK: puppet ran at Fri Dec 16 15:10:17 UTC 2011 [15:10:50] RECOVERY - Puppet freshness on mw1025 is OK: puppet ran at Fri Dec 16 15:10:25 UTC 2011 [15:10:50] RECOVERY - Puppet freshness on mw7 is OK: puppet ran at Fri Dec 16 15:10:29 UTC 2011 [15:10:51] RECOVERY - Puppet freshness on srv194 is OK: puppet ran at Fri Dec 16 15:10:29 UTC 2011 [15:10:51] RECOVERY - Puppet freshness on db1006 is OK: puppet ran at Fri Dec 16 15:10:44 UTC 2011 [15:10:58] RECOVERY - Puppet freshness on mw1152 is OK: puppet ran at Fri Dec 16 15:10:47 UTC 2011 [15:11:08] RECOVERY - Puppet freshness on db1045 is OK: puppet ran at Fri Dec 16 15:10:48 UTC 2011 [15:11:08] RECOVERY - Puppet freshness on mw1138 is OK: puppet ran at Fri Dec 16 15:10:49 UTC 2011 [15:11:08] RECOVERY - Puppet freshness on mw1049 is OK: puppet ran at Fri Dec 16 15:10:53 UTC 2011 [15:11:08] RECOVERY - Puppet freshness on srv218 is OK: puppet ran at Fri Dec 16 15:10:55 UTC 2011 [15:11:18] RECOVERY - Puppet freshness on mw1055 is OK: puppet ran at Fri Dec 16 15:11:04 UTC 2011 [15:11:18] RECOVERY - Puppet freshness on db1041 is OK: puppet ran at Fri Dec 16 15:11:05 UTC 2011 [15:11:28] RECOVERY - Puppet freshness on mw1036 is OK: puppet ran at Fri Dec 16 15:11:08 UTC 2011 [15:11:28] RECOVERY - Puppet freshness on srv245 is OK: puppet ran at Fri Dec 16 15:11:09 UTC 2011 [15:11:28] RECOVERY - Puppet freshness on srv289 is OK: puppet ran at Fri Dec 16 15:11:10 UTC 2011 [15:11:28] RECOVERY - Puppet freshness on srv197 is OK: puppet ran at Fri Dec 16 15:11:15 UTC 2011 [15:11:28] RECOVERY - Puppet freshness on db1029 is OK: puppet ran at Fri Dec 16 15:11:15 UTC 2011 [15:11:28] RECOVERY - Puppet freshness on es1003 is OK: puppet ran at Fri Dec 16 15:11:16 UTC 2011 [15:11:38] RECOVERY - Puppet freshness on srv223 is OK: puppet ran at Fri Dec 16 15:11:18 UTC 2011 [15:11:38] RECOVERY - Puppet freshness on db26 is OK: puppet ran at Fri Dec 16 15:11:19 UTC 2011 [15:11:38] RECOVERY - Puppet freshness on search11 is OK: puppet ran at Fri Dec 16 15:11:20 UTC 2011 [15:11:38] RECOVERY - Puppet freshness on srv208 is OK: puppet ran at Fri Dec 16 15:11:23 UTC 2011 [15:11:48] RECOVERY - Puppet freshness on mw1053 is OK: puppet ran at Fri Dec 16 15:11:33 UTC 2011 [15:11:48] RECOVERY - Puppet freshness on mw1157 is OK: puppet ran at Fri Dec 16 15:11:42 UTC 2011 [15:11:48] RECOVERY - Puppet freshness on db41 is OK: puppet ran at Fri Dec 16 15:11:46 UTC 2011 [15:12:18] RECOVERY - Puppet freshness on db1034 is OK: puppet ran at Fri Dec 16 15:11:47 UTC 2011 [15:12:18] RECOVERY - Puppet freshness on mw1042 is OK: puppet ran at Fri Dec 16 15:11:49 UTC 2011 [15:12:18] RECOVERY - Puppet freshness on db1018 is OK: puppet ran at Fri Dec 16 15:11:51 UTC 2011 [15:12:18] RECOVERY - Puppet freshness on search8 is OK: puppet ran at Fri Dec 16 15:11:57 UTC 2011 [15:12:18] RECOVERY - Puppet freshness on db1015 is OK: puppet ran at Fri Dec 16 15:12:04 UTC 2011 [15:12:28] RECOVERY - Puppet freshness on srv296 is OK: puppet ran at Fri Dec 16 15:12:07 UTC 2011 [15:12:28] RECOVERY - Puppet freshness on srv210 is OK: puppet ran at Fri Dec 16 15:12:09 UTC 2011 [15:12:28] RECOVERY - Puppet freshness on mw31 is OK: puppet ran at Fri Dec 16 15:12:15 UTC 2011 [15:12:28] RECOVERY - Puppet freshness on mw1022 is OK: puppet ran at Fri Dec 16 15:12:16 UTC 2011 [15:12:38] RECOVERY - Puppet freshness on mw1137 is OK: puppet ran at Fri Dec 16 15:12:23 UTC 2011 [15:12:38] RECOVERY - Puppet freshness on mw1080 is OK: puppet ran at Fri Dec 16 15:12:25 UTC 2011 [15:12:48] RECOVERY - Puppet freshness on mw1109 is OK: puppet ran at Fri Dec 16 15:12:27 UTC 2011 [15:12:48] RECOVERY - Puppet freshness on mw1150 is OK: puppet ran at Fri Dec 16 15:12:31 UTC 2011 [15:12:48] RECOVERY - Puppet freshness on srv301 is OK: puppet ran at Fri Dec 16 15:12:32 UTC 2011 [15:12:48] RECOVERY - Puppet freshness on mw1133 is OK: puppet ran at Fri Dec 16 15:12:41 UTC 2011 [15:12:48] RECOVERY - Puppet freshness on mw74 is OK: puppet ran at Fri Dec 16 15:12:45 UTC 2011 [15:13:48] RECOVERY - Puppet freshness on srv216 is OK: puppet ran at Fri Dec 16 15:12:53 UTC 2011 [15:13:48] RECOVERY - Puppet freshness on mw1071 is OK: puppet ran at Fri Dec 16 15:12:54 UTC 2011 [15:13:48] RECOVERY - Puppet freshness on srv211 is OK: puppet ran at Fri Dec 16 15:12:55 UTC 2011 [15:13:48] RECOVERY - Puppet freshness on mw1079 is OK: puppet ran at Fri Dec 16 15:12:55 UTC 2011 [15:13:48] RECOVERY - Puppet freshness on search10 is OK: puppet ran at Fri Dec 16 15:12:58 UTC 2011 [15:13:48] RECOVERY - Puppet freshness on srv300 is OK: puppet ran at Fri Dec 16 15:12:58 UTC 2011 [15:13:48] RECOVERY - Puppet freshness on search2 is OK: puppet ran at Fri Dec 16 15:13:00 UTC 2011 [15:13:49] RECOVERY - Puppet freshness on srv259 is OK: puppet ran at Fri Dec 16 15:13:01 UTC 2011 [15:13:49] RECOVERY - Puppet freshness on mw1045 is OK: puppet ran at Fri Dec 16 15:13:07 UTC 2011 [15:13:50] RECOVERY - Puppet freshness on mw1110 is OK: puppet ran at Fri Dec 16 15:13:12 UTC 2011 [15:13:50] RECOVERY - Puppet freshness on mw1012 is OK: puppet ran at Fri Dec 16 15:13:13 UTC 2011 [15:13:51] RECOVERY - Puppet freshness on srv246 is OK: puppet ran at Fri Dec 16 15:13:15 UTC 2011 [15:13:51] RECOVERY - Puppet freshness on db22 is OK: puppet ran at Fri Dec 16 15:13:21 UTC 2011 [15:13:52] RECOVERY - Puppet freshness on mw71 is OK: puppet ran at Fri Dec 16 15:13:29 UTC 2011 [15:13:52] RECOVERY - Puppet freshness on mw54 is OK: puppet ran at Fri Dec 16 15:13:39 UTC 2011 [15:17:08] PROBLEM - SSH on maerlant is CRITICAL: Server answer: [15:21:24] !log reinstalling dataset1 [15:21:32] Logged the message, RobH [15:22:07] cmjohnson1: are you on site right now? [15:22:23] not sure dataset1 will take my command to boot from network, it seems to take my f12 entry as delete =P [15:22:27] robh: no not at the moment...waiting for the repair guy [15:22:34] ahh, no worries [15:22:48] I should be in after lunch time [15:23:01] i can cheat and change the boot order to network first in bios [15:23:05] then change it back when install is done [15:23:19] its what i have done before, its just more annoying =] [15:23:38] np [15:23:42] how's the car? [15:23:51] good as new [15:23:54] got it back yesterday [15:24:12] was only cosmetic damage, no frame damage [15:24:25] they basically pulled off my entire rear bumper assembly panels and replaced them all [15:24:33] that is good...frame damage is death to a vehicle [15:24:38] my car is 2007 though, so they actually had the parts and paint locally [15:25:08] that is good! no more walky for u =] [15:25:17] get you back in the DC! [15:25:37] yay [15:25:58] I'm actually completely functional today too which means I will pound on in when it's done... er... if it installs [15:26:17] congrats onthe car [15:28:58] RobH: can you fix lvs1003 at some point? [15:29:07] not sure what its current status is... just that it has been broken and/or not up for a long time [15:33:06] sure, i will take a look at it after dataset1 [15:33:36] well, trying dataset1, it wont take my damned key entry to do anything but enter bios, so killed all the boot options and hopefully will do network [15:33:47] if not it will wait for cmjohnson1 to be on site this afternoon, no problem [15:34:31] cursed thing [15:35:02] robh: should I bring the sledge hammer? maybe we could fly apergos in to have first whack! [15:35:23] the day dataset1 is slated to really die, I am flying down there with a rifle. [15:35:52] next ops meeting [15:35:53] !log lvs1003 unresponsive to serial console, rebooting [15:35:57] heheh [15:36:01] Logged the message, RobH [15:36:04] we can have it on site.... [15:36:12] cmjohnson1: yea, i need to you console dataset1 later when on site with me [15:36:23] so you can do the F12 to boot from network, it wont take my entry over serial for anything but bios [15:36:32] bios can be set to boot cdrom or raid, network isnt a boot order option [15:36:40] network can only be selected from the POST entry screen [15:36:41] sucks. [15:37:19] * RobH is also hacking at making es1002 work with the r610 sas controller [15:37:22] no joy so far =[ [15:37:33] lvs1003 is booting back up [15:38:31] mark: seems lvs1003 lost its primary boot disk, its attempting to netbot [15:38:52] seeing if i can boot off its secondary and swap it later [15:40:26] looks like mutante made a ticket for this awhile ago, stealing it [15:40:59] wow... ok, no disks found, so not software raid on these i guess, or the entire thing is borked [15:41:17] mark: I am going to update the ticket, if I do not go to the DC today, it will be on Monday, is that ok? [15:41:26] the sas topology in the controller bios sees no drives at all [15:42:22] !log lvs1003 disk dead per RT 1549, will torubleshoot on site later today or Monday [15:42:31] Logged the message, RobH [15:43:04] it will prolly be monday, since yesterday proved to me that I need to adjust my datacenter time to leave at 730 and tonight i have plans at 720 =P [15:44:30] !log es1002 being actively worked on for hdd controller testing [15:44:38] Logged the message, RobH [15:44:57] wtf [15:44:58] console: Serial Device 2 is currently in use [15:45:10] that isnt supposed to happen on new drac6 [15:45:11] wtf [15:49:02] yay for supposed to [15:49:58] had to reset the drac, on drac6 [15:50:00] today is a sad day. [15:54:16] awww [15:55:39] Change abandoned: Dzahn; "fixed meanwhile, already running for a while" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1453 [15:56:59] huh, seems that the lvs100X servers are all single disk [15:57:19] mark: would we want to install a second drive for raid1 mirroring to prevent what happened now with lvs1003? [15:57:34] or are we find with single disk and redundancy is that we have 4 lvs servers? [15:58:02] I figure a mirror that we can simply reboot to sdb in the event of drive failure may be preferred for vital servers [15:59:02] they are the r610s, so they are hot swap [15:59:15] but we would have to order the standard drives, all i have for spares in eqiad are ssd [15:59:37] though we do have plenty of spare 2.5" disks in tampa [15:59:44] so could always have cmjohnson1 ship up 4 of those [16:00:02] a lot of them are old backups for the sun database servers, but they are a ton of those, so we can easily spare some [16:00:53] do they have 2 drives? [16:01:05] lvs1003 is not urgent [16:01:08] it's been broken for months [16:01:12] but that's why I want something to happen ;) [16:01:16] seems everyone forgot about it [16:01:26] New review: Dzahn; "yes, generic::webserver::php5': ssl => 'true', but true about the certificate, lemme add that" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1575 [16:01:35] mark: yea, i didnt notice it, i can fix no problem [16:01:42] but since we have the spare disks, wouldnt software raid be ideal? [16:02:09] we would need to reinstall the other lvs servers if we add is all [16:02:16] but since they are puppetized and usch that shouldnt be hard right? [16:02:30] no not hard at all [16:02:46] cool, i will drop a ticket for chris to ship them out then (cmjohnson1 that cool with you?) [16:03:10] why don't they have two disks then? [16:03:12] we forgot them? [16:03:16] yep. [16:03:55] i logged into the others, they only have single disk [16:04:32] i will ask chris to send me some spare drives for these, and to have some on hand. [16:09:23] robh: I will get those out to you [16:09:41] printing the fedex label now [16:10:32] * RobH is locked out of portal.equinix.com [16:10:49] seems it needs IE for the security confirmation screens, cuz it shows blank page info for me in FF, safari, and chrome [16:10:50] =[ [16:10:51] oh yeah, do I need to do anything for that? [16:10:57] that link expired [16:11:09] I figured you were on it [16:11:10] so if you didnt click and set a password right away (I did not) you need to ask glenda to send you another link [16:11:13] come on. not ff *or* chrome? [16:11:15] i can ask her to send to you two as well [16:11:20] i bet ct didnt do it yet either =] [16:11:29] mark: i will ask her to resend to you as well [16:11:35] tnx [16:11:36] New patchset: Dzahn; "planet - use star.wmf ssl cert, move to own file, remove hard-coded IP,.." [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1606 [16:11:48] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1606 [16:13:15] heh [16:13:16] (Note: This link will expire on December 8, 2011 11:40:45 PM GMT) [16:13:45] they never took old portal offline, so i forgot about this and used old portal over and over, heh [16:14:20] i haven't been in there in months ;) [16:15:12] all my work in there is incoming and outgoing shipment requests [16:15:13] robh: do you want that ground, 2day or next day...i have it set for 2day [16:15:22] grrrr! [16:15:32] cmjohnson1: how much is price difference between overnight and two day [16:15:41] wait 1 [16:15:49] two day is fine [16:15:53] this is not in a hurry at all [16:16:00] ok, what mark said is good then [16:16:02] it can take weeks :P [16:16:26] its on my radar now, i want it fixed before i go on holiday break ;] [16:16:31] two day is fine though [16:16:53] i'd prefer it fixed AFTER my holiday break :P [16:16:54] cmjohnson1: sorry to make ya look it up =] [16:16:57] the lvs servers are already redundant [16:17:02] but who knows what you break by reinstalling ;) [16:17:12] i would only reinstall the lvs1003 [16:17:12] I'm going skiing for a week on the 30th btw [16:17:16] ok [16:17:17] the others would get the drives in and wait for you ;] [16:17:30] 30th till the 8th [16:17:31] just installing drives wont affect them, so they can sit unused in servers [16:17:42] nice [16:17:47] i need to learn to ski this year [16:17:49] or snowboard [16:18:05] (prolly snowboard, i wanna be cool ;) [16:18:06] heh [16:18:18] (i figured ;) [16:18:24] most in our group are snowboarding [16:18:25] plus i have been on a cheap snowboard [16:18:27] never been on skis [16:18:42] cheap ebing not a real binding, velcro strapping only [16:18:50] no real boots or proper setup [16:19:05] plus was on tiny hills that folks were sledding on, i wanna get the real experience this year [16:19:28] my sister and brother in law do both [16:19:32] never done a snowboard though I would like to [16:19:37] i used to skateboard [16:19:38] done skis a few ties, it's fun [16:19:43] *times [16:19:58] Well, before I was lacking the climate, and the elevation [16:19:59] robh: 2day 46.34 and overnight 104.14....this is at 15lbs [16:20:01] now i have both within two hours [16:20:10] <^demon> I'd love to go again, but I don't think I'll have the time/money this season. [16:20:10] yes and this is a good time of year [16:20:12] cmjohnson1: yikes! two day [16:20:29] ^demon: i let you know when I am going, I am going to pull my buddy justin with me [16:20:32] but more folks is more fun [16:21:01] k [16:21:07] cmjohnson1: if it wasnt holidays, i woudl say ground, but i worry about delivery dates since its not promised like on two day [16:21:48] <^demon> RobH: I'm going on a cruise in March, so trying to not do too much traveling between now and then :) [16:21:50] <^demon> Saving $$$ [16:21:59] nice, where to and what line? [16:22:04] I agree...that is why I went w/ 2day...it is only a little more and you will get it by Tuesday [16:22:13] cmjohnson1: awesome thank you =] [16:23:16] hmm never done one of those (cruise)... guess I would not want to be on a cruise ship. but on a schooner doing some sort of trip with lots of stop would be cool [16:23:16] s [16:23:35] <^demon> RobH: Carnival, eastern caribbean (eg: San Juan, St. Thomas, St. Lucia, Barbados, etc) [16:23:39] i have done a couple cruises [16:23:44] carnival is the party line =] [16:23:45] good call [16:23:54] st thomas is pretty much just stunning [16:24:06] apergos: I want to do a sailing cruise next [16:24:13] they take a larger sailboat, and you get to cruise and crew [16:24:16] it looks awesome. [16:24:37] do they take people without crew experience? [16:25:00] they have the pateince to teach some specific tasks I'm in [16:25:33] yea, they do it all [16:25:56] there seems to be varying degrees, some ships are mostly you can relax, and learn when you want [16:26:07] others are more encouraging to have folks stand watches with crew and really work the boat [16:26:15] <^demon> I'm doing a sailing thing in St. Maarten. [16:26:18] but i get the impression the entire point is to teach folks about sailing [16:26:32] ^demon: i ahve done a similar cruise on another line in that area [16:26:34] sweet [16:26:40] how i miss warm weather and teal water [16:26:43] although I hate to ask how expensive something like that is [16:27:20] its about the same as a normal cruise ship [16:27:29] so anywhere from 500 to 5k, depending on how long you are on and where you go [16:27:35] usually the airfare and such are packaged to get you there [16:28:09] since its a lot of smaller craft, there are tons more companies [16:28:12] robh: I will ping you when I get to DC. [16:28:27] cmjohnson1: awesome, thanks! then we can get dataset1 squared away. my fault it waited this long [16:28:44] NBD [16:28:50] but dont feel you need to rush on my account, i will be around all day =] [16:29:05] okay...thx [16:29:16] em 5k is way outa my league [16:29:28] yea me too [16:29:37] most the ones i have seen that i would be interested in are much closer to 1k [16:29:44] but its still too much money right now for me =[ [16:29:56] i just crew on other folks boats, thats free. [16:30:03] though now its too cold, i miss sailing =[ [16:30:03] that's a very sweet deal [16:30:20] yea i will be renting a boat next wikimania for folks who wanna go sailing =] [16:30:22] you gotta find a winter hobby. [16:30:28] get thee to a ski loft! [16:30:28] <^demon> http://is.gd/VNJcVf - this is the exact cruise I'll be on [16:30:50] that is a nice schedule [16:30:57] its good to be on the boat a week, lets you relax. [16:31:04] plus only one sea day, thats nice [16:31:13] being at sea is neat but really not interesting [16:31:39] i have not been to st kitt [16:31:42] i have heard good things [16:31:53] (my dad used to put doctors on cruise ships for continuing education courses) [16:31:58] <^demon> That's why we decided to leave from San Juan. Leaving from Miami/Ft. Lauderdale you lose a day. [16:32:02] its how i went on the ones i went on, he made me come help him work =P [16:32:08] good call [16:32:50] that looks like a lot of fun [16:33:06] i would suggest you not get too drunk and miss the boat, as I once almost did [16:33:11] <^demon> First cruise :) [16:33:23] i was the last one up the gangway and they were going to leave me [16:34:05] you get a card that is for your room, also acts as your passport to go on and off ship [16:34:32] though you will walk thorugh a metal detector going on and off ship [16:34:37] and they wont let you bring your own booze [16:34:42] unless you somehow hide it properly [16:34:51] vodka to the water bottle, etc... [16:34:57] heh heh [16:35:04] so the food is free, its drinking where they make money [16:35:21] ahhh [16:35:31] free food (well you pay for it in the cost of the cruise) [16:35:33] but I get ya [16:35:35] yea [16:35:40] but even soda they charge [16:35:44] water and ice tea are free [16:35:45] wtf [16:35:50] not that I want soda [16:35:51] soda is free if you buy the stamp at cruise start [16:35:53] but juice at least [16:36:01] they have like a 30 dollar sicker they slap on your id card [16:36:06] and if you have it and show it, you get soda free [16:36:14] juice water tea coffee are free [16:36:19] oh yay [16:36:30] and usually have nice coffee and expresso for folks [16:36:36] but i need diet coke to live. [16:42:53] New patchset: Dzahn; "blogs - consistent class names" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1607 [16:43:05] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1607 [17:46:43] mark: you around? [17:46:50] yes [17:47:03] did you get a chance to look at the swift stuff? [17:47:13] I sent you a mail reply [17:47:17] basically it's a puppet bug [17:47:26] so I guess we should get rid of the inheritance for now [17:47:29] which is not too horrid [17:47:38] but if you want to convert to hashes again, I guess we can [17:47:53] oh. "It seems like there is currently no way to inherit from a parametrized class" lame. [17:48:08] maybe you can tell me if this would work: [17:48:27] we can remove the default settings in swift-cluster::base [17:48:32] have a class to set the config variables that does a case switch on cluster name, then sets all the individual variables. [17:48:32] and copy them to the two descendents [17:48:34] then it should work [17:48:50] then call out to that class from the template. [17:49:37] that's basically what we do, right? [17:49:42] except we don't do a case switch [17:49:45] but have two classes [17:49:58] why not copy the default values to the children, it's clear enough... [17:51:01] that's a smaller changeset. I'll try that first. [17:51:10] there's no reason why that wouldn't work [17:53:14] mark: my other question for you - RobH got an answer from dell - the r510 only supports those two controllers. [17:53:18] PROBLEM - Puppet freshness on snapshot4 is CRITICAL: Puppet has not run in the last 10 hours [17:53:38] they said only the raid controllers wont work, i already fired back a question about sas controller [17:53:44] since i asked both, they only answered one so far [17:54:22] though, when i was in the datacenter yesterday [17:54:33] looking online, the sas controller port on the backplane seems standard to me [17:54:39] same thing in dells, in suns, so on [17:54:51] I wonder, though, if it's just that the partman recipe is the wrong one, since it was written for all disks under one raid10. [17:54:53] so we may be able to add a third party controller just fine [17:55:09] hrmm, it is failing on initing disks, not even seeing them to partition [17:55:12] if the partman recipe is wrong, just partition manually [17:55:15] so i dont think its hitting the partman at all [17:55:19] then probably not no [17:55:27] what is the error message? [17:55:37] hrmph. [17:55:48] lemme pull it up [17:56:12] no error, gotes to blank screen after activating sata or something, resettin gto get the error [17:56:19] in the installer, on detect disks [17:56:27] now, since its serial, i cannot just hop onto another console to see =/ [17:56:38] you can't go back and into shell? [17:56:55] nope, it literally locks on black screen [17:57:00] once you say to activate the sata devices [17:57:07] i can get shell before it tries that, but not after [17:57:22] after it is just unresponsive [17:57:24] strange [17:57:31] i think its just not supported is my fear [17:57:34] sounds like the sas controller doesn't really support it [17:57:38] like that card cannot work in that confdiguration [17:57:40] yea [17:57:49] so we need a sata controller than can run 6 drives [17:57:51] let's buy a supermicro ds1 ;) [17:57:52] per cable [17:58:15] so i am hoping dell will tell us about a sas controller they have that will work [17:58:18] but still waiting to hear back [17:58:46] maplebed: can you read what kind of controller/chipset runs the disks on the suns that do this fine? [17:58:47] you can look around for retail SAS controllers as well [17:59:11] heh, figured would be easier to start from chipset and move from there [17:59:14] for retail stuff [17:59:31] * maplebed knows nothing about sun hardware. [17:59:46] ms1 uses sata_mv [17:59:48] so that's marvell [17:59:52] i can check then no worries, oh cool [17:59:56] ok, thats a decent starting point [18:00:14] so the each-disk-as-raid0... [18:00:16] ms3 and up use lsi [18:00:21] I got it to boot with 5 drives, just not 12. [18:00:27] I feel like that's worth trying again. [18:00:40] es1002? [18:00:41] I would try LSI first [18:00:54] heh, first google result [18:00:55] http://www.lsi.com/channel/products/storagecomponents/Pages/LSISAS9211-8i.aspx [18:01:16] so i will find a few for us to review [18:01:24] then i can use company card to order one for testing [18:01:45] if this works out, we would just order the 510s with no raid, no controller [18:01:59] and of course, there's the famous backblaze storage pod blog post [18:02:34] http://blog.backblaze.com/2011/07/20/petabytes-on-a-budget-v2-0revealing-more-secrets/ [18:02:36] the one where they have specs, but you build? [18:02:41] yes [18:02:51] yea, neat but then we are in the boat of custom hardware [18:03:04] which is not nice when you have the midsize footprint we do [18:03:06] imho [18:03:14] my point is [18:03:21] you can take the SATA controllers from that [18:03:21] oh, that controller you mean? [18:03:23] yea [18:03:56] Syba PCI Express SATA II 4-Port RAID Controller Card SY-PEX40008 [18:04:27] looks like that powers 4 drives per channel though, but will look [18:04:33] yeah [18:05:14] http://www.sybausa.com/productInfo.php?iid=537 [18:05:20] hello [18:05:39] the raid is fake raid, like the sas controller [18:05:40] s [18:05:48] heya hashar [18:06:08] probably going to make mark scream again :D [18:08:32] i just shot the sybausa tech support a question, telling them what we are trying to drive and how [18:08:45] and asking for their input on what controller they have, referencing that blog post, that may work [18:08:59] the detail pages dont seem to say now many devices it can drive off a single channel [18:09:32] though they have similar controllers only two channel, which is what we would actually need [18:09:48] http://www.sybausa.com/productInfo.php?iid=536¤tPage=0 [18:10:25] it isn't one drive per channel? [18:11:26] nope [18:11:37] not if i am reading this right [18:11:45] they are using the one controller, and running 4 backplanes off it [18:11:52] which means 4 drives per backplane, per channel [18:12:14] or i am just confused [18:13:25] maplebed: so in the dell r610 [18:13:28] like the sq servers [18:13:39] it has a SAS controller with two cables, that plug into the backplane [18:13:46] everything i see refers to that controller as two channel [18:13:52] but it runs up to 6 drives [18:13:58] so its running 3 per channel there [18:14:21] but this is all kind of what i think is right, based off how it has been, i have no textbook authoritative answer =] [citation needed] [18:14:51] and i love this. http://store.xkcd.com/xkcd/#CitationNeededSticker [18:15:11] The sybausa cards - it looks like the throughput isn't as high as the LSI - LSI says 6G/port and the sybausa 3. [18:15:33] though they won't be raided, I do expect swift to want to talk to many drives in parallel. [18:16:35] good point [18:17:00] I am putting a ticket to sales with LSI with the exact same questions I sent sybausa =] [18:19:05] I am also looking at the stuff online, but its not enough [18:19:12] and digging in old forum posts is always suspect [18:19:21] so i would like them to confirm stuff before we buy it anyhow [18:23:50] 14 [18:23:54] wtg wmbot. [18:23:56] cmjohnson1: if i recall, you have a ton of them [18:24:15] wm-bot: if you dont want it to be negative, dont suck so much. [18:24:38] cmjohnson1: I recall like 30, if you dont have that many, i need at minimum 6 [18:24:45] so if not a ton, lets do like 8 [18:24:48] that gives me two spares [18:25:03] but lets make sure you have a good 8 to 10 on hand when done [18:25:07] at minimum [18:25:10] if you dont, i can order them isntead [18:25:12] instead even [18:27:31] eww. [18:27:34] ok, i dont like that at all [18:27:39] well, open? [18:27:49] the problem was before andrew i had everything labeled and bagged [18:27:59] he removed all those drives from the damned bags and pout them in plastic drawers [18:28:02] sigh [18:28:14] so, yea, there are the sun drives that arent in bags [18:28:20] those 'should' be fine [18:28:30] so send me 8 of the opened ones [18:28:37] that are with the sun label on them [18:28:44] i think they are fine, and I will certainly find out [18:29:16] heh [18:29:20] if you can find one large static bag, you can just put them all together in there [18:29:30] cmjohnson1: yep, but not until after you get back from vacation [18:29:37] hopefully i am ordering a few servers today [18:29:57] fyi: ryan is looking for petan to fix wmbot [18:30:05] oh, i pinged him, heh [18:30:27] crap [18:30:31] I dropped it from the wrong channel [18:30:37] @commands [18:30:44] @drop [18:30:54] hehe [18:30:55] hahahahaha [18:31:01] i lol'd [18:31:21] well, I've fucked this up properly [18:31:22] Ryan_Lane: you bring brion my hat? [18:31:35] I was supposed to? [18:31:39] yesterday? [18:31:46] I'll bring it to work today, when I go in [18:31:51] no owrries [18:31:59] i havent chatted with him about it [18:32:13] i just know he wont mind sending it to me, since i am sending them east coast stuffs [18:32:30] cmjohnson1: i will connect to it right now, go ahead and crash cart it up [18:33:32] ok, connectiong now [18:33:56] ok, can you reboot it and enter the normal bios for me please? [18:34:10] i need to revert my changes there, then we will reboot and do a network boot option [18:37:13] cmjohnson1: crap, didnt mean to switch to that channel [18:37:15] sorry about that [18:37:23] lets work in here, this is all public [18:37:59] heh, on serial its even easier to miss [18:38:06] since it updates more slowly than console [18:38:09] yep [18:38:22] you can restart, there is no way for us to cleanly shutdown the installed os [18:38:25] since we are locked out of it [18:38:37] and we are going to replace it anyhow =] [18:39:00] got it that time, i saw it update [18:39:13] of course, now i need the mac, and we have to do it again in a minute [18:39:56] i am going to read the mac off the dhcp server log [18:40:10] so it will prolly timeout on the netboot before i can snag it, update dhcp, and restart dhcp [18:40:45] cmjohnson1: crap, it went past too fast hrmm [18:41:20] yea, and it took it [18:41:31] its there, just hard to see, none of its bolded [18:41:41] when you reboot, its on left side, 1/3rd way down [18:41:45] under memory tests [18:41:49] now, i think the mac is 78:2b:cb:0a:11:7a [18:42:24] but that mac doesnt show on generic registry sites for anyone, but that sounds normal [18:42:31] well, can you drop into bios and see if it tells you the mac [18:42:36] you can do it faster than i can on serial [18:42:45] see if it is 78:2b:cb:0a:11:7a, and if it isnt, take down the mac for me pls [18:43:56] it blends in [18:44:07] ok, i am going to boot the bot. [21:33:31] LeslieCarr: was about to subject something similar :-) [21:33:38] subject/suggest [21:34:55] Oh ffs [21:36:02] heh [21:36:02] I guess it's broken [21:36:10] we'll need petan to look at it [21:36:18] @commands [21:36:24] @add #wikimedia-labs [21:36:36] @trusted [21:36:43] @add #wikimedia-labs [21:36:52] it's actually already in there [21:36:56] I edited the config file [21:45:05] can some one possibly review my changes to production in gerrit? [21:45:11] they should be straightforward [21:46:18] PROBLEM - Puppet freshness on maerlant is CRITICAL: Puppet has not run in the last 10 hours [21:47:08] New patchset: Bhartshorne; "setting real auth tokens for pmtpa test cluster" [operations/puppet] (production) - https://gerrit.wikimedia.org/r/1613 [21:47:19] New review: gerrit2; "Lint check passed." [operations/puppet] (production); V: 1 - https://gerrit.wikimedia.org/r/1613 [21:51:17] bugzilla is down? [21:52:14] hrm... and the cause isn't a enwiki blackout since enwiki is up [21:52:37] did someone just decide to take bugzilla down to protest SOPA? [21:52:42] hexmode: its up for me. [21:52:47] yeah, that'll get there attention [21:53:01] oh gerrit got me [21:53:01] RobH: maybe I should kill my cookies [21:53:14] hexmode: rt is down too it seems [21:53:53] bleh, closed wrong window [21:54:27] and then client crashes... [21:54:31] hexmode: i take that back [21:54:33] hashar: robh: killed all my wiki cookies and still bugzilla times out [21:54:35] it loads the main page, i didnt try to login [21:54:38] heh [21:54:47] kaulen is up at least [21:54:55] but I can't access the site itself [21:54:59] lookin [21:55:22] although it works when unlogged [21:55:35] RobH: message says "Gateway Time-out -- The gateway did not receive a timely response from the upstream server or application. --- Apache Server at bugzilla.wikimedia.org Port 443" [21:55:49] so what happened with ds1? is it installed and read yfor action? [21:55:56] and when I say "wiki cookies" I mean wikimedia.org [21:56:07] apergos: its installed, hasnt had puppet run [21:56:14] so first login has to be with the new install key on sockpuppet [21:56:15] if it can help, I have the same issue with RT [21:56:17] but its ready otherwise [21:56:23] ok [21:56:36] the main page using http://rt.wm.org/ works but the httpS does not [21:57:02] hashar: rt is another server, but yea, its not pulling up fo rme [21:57:27] they are probably unrelated [21:57:36] or the backend for both is borked [21:57:39] unless you have a nice loadbalancer/ssl offloaded in front of both of them [21:57:46] both live on same database server [21:57:49] could they use the same db? [21:57:52] but it doesnt seem bad [21:57:53] lol [21:58:51] its on db9, but db9 is responsive [21:58:58] bleh, there is a lot of shit on the same server as bugzilla [21:59:26] !log kicking apache for bz related issues on kaulen [21:59:34] Logged the message, RobH [22:00:23] ohhh [22:00:26] db9 disk full [22:00:34] !log db9 disk full, related services are messing up, fixing [22:00:44] Logged the message, RobH [22:02:20] Ryan_Lane: you about? [22:02:34] It seems like db9 is pretty damned full, wanted second eyes on it [22:03:23] yeah [22:03:26] purge some binlogs? [22:03:32] it can bascailly hold enough binlogs for like 4 days now [22:03:47] not that many to purge, but dumped half [22:03:47] asher has purged otrs db [22:03:50] and now its 98% [22:03:54] but not reclaimed that space yet [22:04:27] well, thats no good. [22:04:37] so how many binlogs can i really afford to kill off really? [22:04:40] all but past 24 hours? [22:04:45] probably [22:04:46] (I'm doing the puppet stuff on ds1... in theory anyways) [22:04:55] apergos: wont affect it [22:04:59] this is purely space on db9 [22:05:03] and backend issues for services [22:05:10] ok, removing all but past 24 hours of binlogs. [22:05:10] I mean, the solution is to figure out how to reclaim the space of the purged db [22:05:13] no no, I'm just saying, so ds1 is off yer plate [22:05:22] but I don't want to go delling data without being damn sure [22:05:22] notpeter: yea, but that sounds hinky for someone to do in a hurry who has not done it [22:05:28] yes [22:06:23] !log db9 space reclaimed back to 94% full, related services should start recovering [22:06:31] Logged the message, RobH [22:06:41] hexmod discovered our outage [22:06:45] hexmode even [22:07:17] so anyone know if i need to do anything else, stuff isnt back yet [22:07:20] hey guys, it looks like blog.wikimedia.org is down for realz [22:07:24] like kick mysql in some way? [22:07:28] LeslieCarr: all db9 related services are [22:07:34] oh it's db9 [22:07:34] ok [22:07:35] db9 ran out of disk [22:07:44] i removed all but recent binlogs, it has 8% free now [22:07:55] anythihn else and we ditch binlogs less than 24 hours old. [22:08:01] robh: is there an alert needed for db9 diskspace? So *I* don't have to notice it ;) [22:08:16] hexmode: its on nagios, and someone should have taken care of it before now [22:08:23] ah [22:08:26] but it was an active issue that asher is relocationg services off to reclaim [22:08:32] so i think the rest of us just assumed he was handling it [22:08:53] i think with so much nagios borkedness recently (and puppet borkedness) it's been getting a bit ignored [22:08:53] but he also was quite clear in a recent email that he is on vacation today, so ooops ;P [22:09:15] heh [22:09:25] of course, i have no idea what to do next [22:09:29] FAIL on FAIL [22:09:30] i removed the binlogs, usually that recovers service [22:09:33] but its still not working. [22:09:42] Do I need to call Asher on his day off? [22:09:54] hrmm, domas_ you abouts? [22:10:13] Ryan_Lane: any ideas? [22:10:28] maplebed: ^ [22:10:32] is it out of space? [22:10:37] I'm gonna need to kill that bot again [22:10:50] I deleted all but the past 24 hours of binlogs [22:10:54] its at 94% full [22:11:02] how did you delete them? [22:11:12] i just deleted them. [22:11:17] ? [22:11:21] * maplebed reads backscroll [22:11:21] which was how i was told to do it a long time ago [22:11:22] you didn't use the mysql command? [22:11:25] nope. [22:11:38] so i fucked up huh? [22:11:43] I dunno [22:11:48] I always follow the docs :) [22:11:48] no, you didn't. [22:11:53] i have done it this way in the past to free up space without issues [22:11:56] ok [22:12:03] it's ok to delete them by hand, sorta. [22:12:07] there is a wikitech page I think [22:12:21] IIRC if you delete them by hand it disables the auto-deletion stuff, but that's the only effect. [23:28:50] @trustadd .*@wikimedia/RobH admin [23:28:50] Successfuly added .*@wikimedia/RobH [23:28:55] @trustadd .*@wikimedia/Ryan-lane admin [23:28:55] Successfuly added .*@wikimedia/Ryan-lane [23:29:17] !g 10 [23:29:22] hm... [23:29:24] !gerrit [23:29:24] https://gerrit.wikimedia.org/ [23:29:32] !r 10 [23:29:37] !change 10 [23:29:37] https://gerrit.wikimedia.org/r/10 [23:29:41] weird [23:29:59] @trusted [23:29:59] I trust: petan|w.*wikimedia/Petrb, .*@wikimedia/.*, .*@mediawiki/.*, .*@mediawiki/Catrope, .*@wikimedia/RobH, .*@wikimedia/Ryan-lane, [23:30:21] if there was problem just type [23:30:22] @part [23:30:37] @trusted [23:30:37] I trust: petan|w.*wikimedia/Petrb, .*@wikimedia/.*, .*@mediawiki/.*, .*@mediawiki/Catrope, .*@wikimedia/RobH, .*@wikimedia/Ryan-lane, petan!.*@wikimedia/Petrb, [23:30:52] coolness, thanks! [23:30:53] .@drop remove even db so that's probably something you don't want to do [23:31:15] !g alias change [23:31:15] Alias is already existing! [23:31:18] im out for the night folks, cyas [23:31:20] !g test [23:31:25] !g unalias [23:31:25] Alias removed [23:31:30] !g alias change [23:31:30] Successfully created [23:31:33] !g 10 [23:31:33] I am giving up on ds1 [23:31:39] meh [23:31:41] I copied over the cert req to [23:32:03] /var/lib/puppet/server/ssl/certificate_requests/ and set perms and owners to puppet and 640 like everythin else in there [23:32:05] it won't see it [23:32:23] after a ton of time on google, still nothing so I'm going ot bed and it will sit for the weekend :-( [23:42:52] !r 67 [23:42:52] on is http://www.mediawiki.org/wiki/Special:Code/MediaWiki/67 [23:49:43] !r del [23:49:43] Unable to find the specified key in db [23:49:59] @search is [23:49:59] Results: revision, alert, [23:50:05] !revision del [23:50:05] Successfully removed revision [23:50:41] !revision is https://www.mediawiki.org/wiki/Special:Code/MediaWiki/$1 [23:50:41] Key was added! [23:50:46] !r 10 [23:50:46] https://www.mediawiki.org/wiki/Special:Code/MediaWiki/10 [23:51:30] fixed :) [23:52:41] it was throwin error because of ipv 6 hostnames it should be ok now