-#
-# Copyright (c) 2004 The Trustees of Princeton University (Trustees).
-#
-# Faiyaz Ahmed <faiyaza@cs.princeton.edu>
-#
-# $Id: policy.py,v 1.17 2007/08/29 17:26:50 soltesz Exp $
-#
-# Policy Engine.
+#!/usr/bin/python
+
+# This script is used to manipulate the operational state of nodes in
+# different node groups. These are basically set operations on nodes via the
+# PLC api.
+#
+# Take the ng name as an argument....
+# optionally,
+# * get a list of nodes in the given nodegroup.
+# * set some or all in the set to rins.
+# * restart them all.
+# * do something else to them all.
+#
-#from monitor import *
-from threading import *
+import os
import time
-import logging
-import mailer
-import emailTxt
-import pickle
-import Queue
-import plc
+import traceback
import sys
-import os
-import reboot
-import database
-import string
-from www.printbadnodes import cmpCategoryVal
-from config import config
-#print "policy"
-config = config()
-
-DAT="./monitor.dat"
-
-logger = logging.getLogger("monitor")
-
-# Time to enforce policy
-POLSLEEP = 7200
-
-# Where to email the summary
-SUMTO = "soltesz@cs.princeton.edu"
-TECHEMAIL="tech-%s@sites.planet-lab.org"
-PIEMAIL="pi-%s@sites.planet-lab.org"
-SLICEMAIL="%s@slices.planet-lab.org"
-PLCEMAIL="support@planet-lab.org"
-
-#Thresholds (DAYS)
-SPERMIN = 60
-SPERHOUR = 60*60
-SPERDAY = 86400
-PITHRESH = 7 * SPERDAY
-SLICETHRESH = 7 * SPERDAY
-# Days before attempting rins again
-RINSTHRESH = 5 * SPERDAY
-
-# Days before calling the node dead.
-DEADTHRESH = 30 * SPERDAY
-# Minimum number of nodes up before squeezing
-MINUP = 2
-
-TECH=1
-PI=2
-USER=4
-ADMIN=8
-
-# IF:
-# no SSH, down.
-# bad disk, down
-# DNS, kinda down (sick)
-# clock, kinda down (sick)
-# Full disk, going to be down
-
-# Actions:
-# Email
-# suspend slice creation
-# kill slices
-def array_to_priority_map(array):
- """ Create a mapping where each entry of array is given a priority equal
- to its position in the array. This is useful for subsequent use in the
- cmpMap() function."""
- map = {}
- count = 0
- for i in array:
- map[i] = count
- count += 1
- return map
-
-def getdebug():
- return config.debug
-
-def print_stats(key, stats):
- if key in stats: print "%20s : %d" % (key, stats[key])
-
-def get_ticket_id(record):
- if 'ticket_id' in record and record['ticket_id'] is not "" and record['ticket_id'] is not None:
- return record['ticket_id']
- elif 'found_rt_ticket' in record and \
- record['found_rt_ticket'] is not "" and \
- record['found_rt_ticket'] is not None:
- return record['found_rt_ticket']
- else:
- return None
-
-class Merge(Thread):
- def __init__(self, l_merge, toRT):
- self.toRT = toRT
- self.merge_list = l_merge
- # the hostname to loginbase mapping
- self.plcdb_hn2lb = database.dbLoad("plcdb_hn2lb")
-
- # Previous actions taken on nodes.
- self.act_all = database.if_cached_else(1, "act_all", lambda : {})
- self.findbad = database.if_cached_else(1, "findbad", lambda : {})
-
- self.cache_all = database.if_cached_else(1, "act_all", lambda : {})
- self.sickdb = {}
- self.mergedb = {}
- Thread.__init__(self)
-
- def run(self):
- # populate sickdb
- self.accumSickSites()
- # read data from findbad and act_all
- self.mergeActionsAndBadDB()
- # pass node_records to RT
- self.sendToRT()
-
- def accumSickSites(self):
- """
- Take all nodes, from l_diagnose, look them up in the act_all database,
- and insert them into sickdb[] as:
-
- sickdb[loginbase][nodename] = fb_record
- """
- # look at all problems reported by findbad
- l_nodes = self.findbad['nodes'].keys()
- count = 0
- for nodename in l_nodes:
- if nodename not in self.merge_list:
- continue # skip this node, since it's not wanted
-
- count += 1
- loginbase = self.plcdb_hn2lb[nodename]
- values = self.findbad['nodes'][nodename]['values']
-
- fb_record = {}
- fb_record['nodename'] = nodename
- try:
- fb_record['category'] = values['category']
- except:
- print values
- print nodename
- print self.findbad['nodes'][nodename]
- count -= 1
- continue
- fb_record['state'] = values['state']
- fb_record['comonstats'] = values['comonstats']
- fb_record['plcnode'] = values['plcnode']
- fb_record['kernel'] = self.getKernel(values['kernel'])
- fb_record['stage'] = "findbad"
- fb_record['message'] = None
- fb_record['bootcd'] = values['bootcd']
- fb_record['args'] = None
- fb_record['info'] = None
- fb_record['time'] = time.time()
- fb_record['date_created'] = time.time()
-
- if loginbase not in self.sickdb:
- self.sickdb[loginbase] = {}
+from optparse import OptionParser
- self.sickdb[loginbase][nodename] = fb_record
+from monitor import config
+from monitor import parser as parsermodule
+from monitor.common import *
+from monitor.model import *
+from monitor.wrapper import plc
+from monitor.wrapper import plccache
+from monitor.database.info.model import *
+from monitor.database.info.interface import *
- print "Found %d nodes" % count
+from nodequery import verify,query_to_dict,node_select
- def getKernel(self, unamestr):
- s = unamestr.split()
- if len(s) > 2:
- return s[2]
- else:
- return ""
-
- def mergeActionsAndBadDB(self):
- """
- - Look at the sick node_records as reported in findbad,
- - Then look at the node_records in act_all.
-
- There are four cases:
- 1) Problem in findbad, no problem in act_all
- this ok, b/c it just means it's a new problem
- 2) Problem in findbad, problem in act_all
- -Did the problem get better or worse?
- -If Same, or Worse, then continue looking for open tickets.
- -If Better, or No problem, then "back-off" penalties.
- This judgement may need to wait until 'Diagnose()'
-
- 3) No problem in findbad, problem in act_all
- The the node is operational again according to Findbad()
-
- 4) No problem in findbad, no problem in act_all
- There won't be a record in either db, so there's no code.
- """
-
- sorted_sites = self.sickdb.keys()
- sorted_sites.sort()
- # look at all problems reported by findbad
- for loginbase in sorted_sites:
- d_fb_nodes = self.sickdb[loginbase]
- sorted_nodes = d_fb_nodes.keys()
- sorted_nodes.sort()
- for nodename in sorted_nodes:
- fb_record = self.sickdb[loginbase][nodename]
- x = fb_record
- if loginbase not in self.mergedb:
- self.mergedb[loginbase] = {}
-
- # take the info either from act_all or fb-record.
- # if node not in act_all
- # then take it from fbrecord, obviously.
- # else node in act_all
- # if act_all == 0 length (no previous records)
- # then take it from fbrecord.
- # else
- # take it from act_all.
- #
-
- # We must compare findbad state with act_all state
- if nodename not in self.act_all:
- # 1) ok, b/c it's a new problem. set ticket_id to null
- self.mergedb[loginbase][nodename] = {}
- self.mergedb[loginbase][nodename].update(x)
- self.mergedb[loginbase][nodename]['ticket_id'] = ""
- self.mergedb[loginbase][nodename]['prev_category'] = "NORECORD"
- else:
- if len(self.act_all[nodename]) == 0:
- self.mergedb[loginbase][nodename] = {}
- self.mergedb[loginbase][nodename].update(x)
- self.mergedb[loginbase][nodename]['ticket_id'] = ""
- self.mergedb[loginbase][nodename]['prev_category'] = "NORECORD"
- else:
- y = self.act_all[nodename][0]
- y['prev_category'] = y['category']
-
- self.mergedb[loginbase][nodename] = {}
- self.mergedb[loginbase][nodename].update(y)
- self.mergedb[loginbase][nodename]['comonstats'] = x['comonstats']
- self.mergedb[loginbase][nodename]['category'] = x['category']
- self.mergedb[loginbase][nodename]['state'] = x['state']
- self.mergedb[loginbase][nodename]['kernel']=x['kernel']
- self.mergedb[loginbase][nodename]['bootcd']=x['bootcd']
- self.mergedb[loginbase][nodename]['plcnode']=x['plcnode']
- ticket = get_ticket_id(self.mergedb[loginbase][nodename])
- self.mergedb[loginbase][nodename]['rt'] = mailer.getTicketStatus(ticket)
-
- # delete the entry from cache_all to keep it out of case 3)
- del self.cache_all[nodename]
-
- # 3) nodes that remin in cache_all were not identified by findbad.
- # Do we keep them or not?
- # NOTE: i think that since the categories are performed before this
- # step now, and by a monitor-controlled agent.
-
- # TODO: This does not work correctly. Do we need this?
- #for hn in self.cache_all.keys():
- # y = self.act_all[hn][0]
- # if 'monitor' in y['bucket']:
- # loginbase = self.plcdb_hn2lb[hn]
- # if loginbase not in self.sickdb:
- # self.sickdb[loginbase] = {}
- # self.sickdb[loginbase][hn] = y
- # else:
- # del self.cache_all[hn]
+api = plc.getAuthAPI()
- print "len of cache_all: %d" % len(self.cache_all.keys())
- return
+def logic():
- def sendToRT(self):
- sorted_sites = self.mergedb.keys()
- sorted_sites.sort()
- # look at all problems reported by merge
- for loginbase in sorted_sites:
- d_merge_nodes = self.mergedb[loginbase]
- for nodename in d_merge_nodes.keys():
- record = self.mergedb[loginbase][nodename]
- self.toRT.put(record)
-
- # send signal to stop reading
- self.toRT.put(None)
- return
-
-class Diagnose(Thread):
- def __init__(self, fromRT):
- self.fromRT = fromRT
- self.plcdb_hn2lb = database.dbLoad("plcdb_hn2lb")
- self.findbad = database.if_cached_else(1, "findbad", lambda : {})
-
- self.diagnose_in = {}
- self.diagnose_out = {}
- Thread.__init__(self)
-
-
- def run(self):
- self.accumSickSites()
-
- print "Accumulated %d sick sites" % len(self.diagnose_in.keys())
- logger.debug("Accumulated %d sick sites" % len(self.diagnose_in.keys()))
+ plc.nodeBootState(host, 'reinstall')
+ node_end_record(host)
+def main(hostnames, sitenames):
+ # commands:
+ i = 1
+ node_count = 1
+ site_count = 1
+ #print "hosts: %s" % hostnames
+ for i,host in enumerate(hostnames):
try:
- stats = self.diagnoseAll()
- except Exception, err:
- print "----------------"
- import traceback
- print traceback.print_exc()
- print err
- #if config.policysavedb:
- sys.exit(1)
-
- print_stats("sites_observed", stats)
- print_stats("sites_diagnosed", stats)
- print_stats("nodes_diagnosed", stats)
-
- if config.policysavedb:
- print "Saving Databases... diagnose_out"
- database.dbDump("diagnose_out", self.diagnose_out)
-
- def accumSickSites(self):
- """
- Take all nodes, from l_diagnose, look them up in the diagnose_out database,
- and insert them into diagnose_in[] as:
-
- diagnose_in[loginbase] = [diag_node1, diag_node2, ...]
- """
- while 1:
- node_record = self.fromRT.get(block = True)
- if node_record == None:
- break;
-
- nodename = node_record['nodename']
- loginbase = self.plcdb_hn2lb[nodename]
-
- if loginbase not in self.diagnose_in:
- self.diagnose_in[loginbase] = {}
-
- self.diagnose_in[loginbase][nodename] = node_record
-
- return
-
- def diagnoseAll(self):
- i_sites_observed = 0
- i_sites_diagnosed = 0
- i_nodes_diagnosed = 0
- i_nodes_actedon = 0
- i_sites_emailed = 0
- l_allsites = []
-
- sorted_sites = self.diagnose_in.keys()
- sorted_sites.sort()
- self.diagnose_out= {}
- for loginbase in sorted_sites:
- l_allsites += [loginbase]
-
- d_diag_nodes = self.diagnose_in[loginbase]
- d_act_records = self.__diagnoseSite(loginbase, d_diag_nodes)
- # store records in diagnose_out, for saving later.
- self.diagnose_out.update(d_act_records)
-
- if len(d_act_records[loginbase]['nodes'].keys()) > 0:
- i_nodes_diagnosed += (len(d_act_records[loginbase]['nodes'].keys()))
- i_sites_diagnosed += 1
- i_sites_observed += 1
-
- return {'sites_observed': i_sites_observed,
- 'sites_diagnosed': i_sites_diagnosed,
- 'nodes_diagnosed': i_nodes_diagnosed,
- 'allsites':l_allsites}
-
- pass
-
- def getDaysDown(cls, diag_record):
- daysdown = -1
- last_contact = diag_record['plcnode']['last_contact']
- date_created = diag_record['plcnode']['date_created']
-
- if diag_record['comonstats']['uptime'] != "null" and diag_record['comonstats']['uptime'] != "-1":
- daysdown = - int(float(diag_record['comonstats']['uptime'])) // (60*60*24)
- elif last_contact is None:
- if date_created is not None:
- now = time.time()
- diff = now - date_created
- daysdown = diff // (60*60*24)
- else:
- daysdown = -1
- else:
- now = time.time()
- diff = now - last_contact
- daysdown = diff // (60*60*24)
- return daysdown
- getDaysDown = classmethod(getDaysDown)
-
- def getStrDaysDown(cls, diag_record):
- daysdown = "unknown"
- last_contact = diag_record['plcnode']['last_contact']
- date_created = diag_record['plcnode']['date_created']
-
- if diag_record['comonstats']['uptime'] != "null" and \
- diag_record['comonstats']['uptime'] != "-1":
- daysdown = int(float(diag_record['comonstats']['uptime'])) // (60*60*24)
- daysdown = "%d days up" % daysdown
-
- elif last_contact is None:
- if date_created is not None:
- now = time.time()
- diff = now - date_created
- daysdown = diff // (60*60*24)
- daysdown = "Never contacted PLC, created %s days ago" % daysdown
- else:
- daysdown = "Never contacted PLC"
- else:
- now = time.time()
- diff = now - last_contact
- daysdown = diff // (60*60*24)
- daysdown = "%s days down" % daysdown
- return daysdown
- getStrDaysDown = classmethod(getStrDaysDown)
- #def getStrDaysDown(cls, diag_record):
- # daysdown = cls.getDaysDown(diag_record)
- # if daysdown > -1:
- # return "%d days down"%daysdown
- # elif daysdown == -1:
- # return "Has never contacted PLC"
- # else:
- # return "%d days up"% -daysdown
- #getStrDaysDown = classmethod(getStrDaysDown)
-
- def __getCDVersion(self, diag_record, nodename):
- cdversion = ""
- #print "Getting kernel for: %s" % diag_record['nodename']
- cdversion = diag_record['kernel']
- return cdversion
-
- def __diagnoseSite(self, loginbase, d_diag_nodes):
- """
- d_diag_nodes are diagnose_in entries.
- """
- d_diag_site = {loginbase : { 'config' :
- {'squeeze': False,
- 'email': False
- },
- 'nodes': {}
- }
- }
- sorted_nodes = d_diag_nodes.keys()
- sorted_nodes.sort()
- for nodename in sorted_nodes:
- node_record = d_diag_nodes[nodename]
- diag_record = self.__diagnoseNode(loginbase, node_record)
-
- if diag_record != None:
- d_diag_site[loginbase]['nodes'][nodename] = diag_record
-
- # NOTE: improvement means, we need to act/squeeze and email.
- #print "DIAG_RECORD", diag_record
- if 'monitor-end-record' in diag_record['stage'] or \
- 'nmreset' in diag_record['stage']:
- # print "resetting loginbase!"
- d_diag_site[loginbase]['config']['squeeze'] = True
- d_diag_site[loginbase]['config']['email'] = True
- #else:
- # print "NO IMPROVEMENT!!!!"
- else:
- pass # there is nothing to do for this node.
-
- # NOTE: these settings can be overridden by command line arguments,
- # or the state of a record, i.e. if already in RT's Support Queue.
- nodes_up = self.getUpAtSite(loginbase, d_diag_site)
- if nodes_up < MINUP:
- d_diag_site[loginbase]['config']['squeeze'] = True
-
- max_slices = self.getMaxSlices(loginbase)
- num_nodes = self.getNumNodes(loginbase)
- # NOTE: when max_slices == 0, this is either a new site (the old way)
- # or an old disabled site from previous monitor (before site['enabled'])
- if nodes_up < num_nodes and max_slices != 0:
- d_diag_site[loginbase]['config']['email'] = True
-
- if len(d_diag_site[loginbase]['nodes'].keys()) > 0:
- print "SITE: %20s : %d nodes up, at most" % (loginbase, nodes_up)
-
- return d_diag_site
-
- def diagRecordByCategory(self, node_record):
- nodename = node_record['nodename']
- category = node_record['category']
- state = node_record['state']
- loginbase = self.plcdb_hn2lb[nodename]
- diag_record = None
-
- if "ERROR" in category: # i.e. "DOWN"
- diag_record = {}
- diag_record.update(node_record)
- daysdown = self.getDaysDown(diag_record)
- if daysdown < 7:
- format = "DIAG: %20s : %-40s Down only %s days NOTHING DONE"
- print format % (loginbase, nodename, daysdown)
- return None
-
- s_daysdown = self.getStrDaysDown(diag_record)
- diag_record['message'] = emailTxt.mailtxt.newdown
- diag_record['args'] = {'nodename': nodename}
- diag_record['info'] = (nodename, s_daysdown, "")
-
- if 'reboot_node_failed' in node_record:
- # there was a previous attempt to use the PCU.
- if node_record['reboot_node_failed'] == False:
- # then the last attempt apparently, succeeded.
- # But, the category is still 'ERROR'. Therefore, the
- # PCU-to-Node mapping is broken.
- #print "Setting message for ERROR node to PCU2NodeMapping: %s" % nodename
- diag_record['message'] = emailTxt.mailtxt.pcutonodemapping
- diag_record['email_pcu'] = True
-
- if 'ticket_id' in diag_record:
- if diag_record['ticket_id'] == "":
- if 'found_rt_ticket' in diag_record:
- ticket_id = diag_record['found_rt_ticket']
- else:
- ticket_id = "None"
- else:
- ticket_id = diag_record['ticket_id']
- else:
- ticket_id = "None"
-
- diag_record['log'] = "DOWN: %20s : %-40s == %20s %s" % \
- (loginbase, nodename, diag_record['info'][1:], ticket_id)
-
- elif "OLDBOOTCD" in category:
- # V2 boot cds as determined by findbad
- s_daysdown = self.getStrDaysDown(node_record)
- s_cdversion = self.__getCDVersion(node_record, nodename)
- diag_record = {}
- diag_record.update(node_record)
- #if "2.4" in diag_record['kernel'] or "v2" in diag_record['bootcd']:
- diag_record['message'] = emailTxt.mailtxt.newbootcd
- diag_record['args'] = {'nodename': nodename}
- diag_record['info'] = (nodename, s_daysdown, s_cdversion)
- if diag_record['ticket_id'] == "":
- diag_record['log'] = "BTCD: %20s : %-40s == %20s %20s %s" % \
- (loginbase, nodename, diag_record['kernel'],
- diag_record['bootcd'], diag_record['found_rt_ticket'])
- else:
- diag_record['log'] = "BTCD: %20s : %-40s == %20s %20s %s" % \
- (loginbase, nodename, diag_record['kernel'],
- diag_record['bootcd'], diag_record['ticket_id'])
-
- elif "PROD" in category:
- if "DEBUG" in state:
- # Not sure what to do with these yet. Probably need to
- # reboot, and email.
- print "DEBG: %20s : %-40s NOTHING DONE" % (loginbase, nodename)
- return None
- elif "BOOT" in state:
- # no action needed.
- # TODO: remove penalties, if any are applied.
- now = time.time()
- last_contact = node_record['plcnode']['last_contact']
- if last_contact == None:
- time_diff = 0
- else:
- time_diff = now - last_contact;
-
- if 'improvement' in node_record['stage']:
- # then we need to pass this on to 'action'
- diag_record = {}
- diag_record.update(node_record)
- diag_record['message'] = emailTxt.mailtxt.newthankyou
- diag_record['args'] = {'nodename': nodename}
- diag_record['info'] = (nodename, node_record['prev_category'],
- node_record['category'])
- if 'email_pcu' in diag_record:
- if diag_record['email_pcu']:
- # previously, the pcu failed to reboot, so send
- # email. Now, reset these values to try the reboot
- # again.
- diag_record['email_pcu'] = False
- del diag_record['reboot_node_failed']
-
- if diag_record['ticket_id'] == "":
- diag_record['log'] = "IMPR: %20s : %-40s == %20s %20s %s %s" % \
- (loginbase, nodename, diag_record['stage'],
- state, category, diag_record['found_rt_ticket'])
- else:
- diag_record['log'] = "IMPR: %20s : %-40s == %20s %20s %s %s" % \
- (loginbase, nodename, diag_record['stage'],
- state, category, diag_record['ticket_id'])
- return diag_record
- #elif time_diff >= 6*SPERHOUR:
- # # heartbeat is older than 30 min.
- # # then reset NM.
- # #print "Possible NM problem!! %s - %s = %s" % (now, last_contact, time_diff)
- # diag_record = {}
- # diag_record.update(node_record)
- # diag_record['message'] = emailTxt.mailtxt.NMReset
- # diag_record['args'] = {'nodename': nodename}
- # diag_record['stage'] = "nmreset"
- # diag_record['info'] = (nodename,
- # node_record['prev_category'],
- # node_record['category'])
- # if diag_record['ticket_id'] == "":
- # diag_record['log'] = "NM : %20s : %-40s == %20s %20s %s %s" % \
- # (loginbase, nodename, diag_record['stage'],
- # state, category, diag_record['found_rt_ticket'])
- # else:
- # diag_record['log'] = "NM : %20s : %-40s == %20s" % \
- # (loginbase, nodename, diag_record['stage'])
-#
-# return diag_record
- else:
- return None
- else:
- # unknown
- pass
- elif "ALPHA" in category:
- pass
- elif "clock_drift" in category:
- pass
- elif "dns" in category:
- pass
- elif "filerw" in category:
- pass
- else:
- print "Unknown category!!!! %s" % category
- sys.exit(1)
-
- return diag_record
-
- def __diagnoseNode(self, loginbase, node_record):
- # TODO: change the format of the hostname in this
- # record to something more natural.
- nodename = node_record['nodename']
- category = node_record['category']
- prev_category = node_record['prev_category']
- state = node_record['state']
- #if 'prev_category' in node_record:
- # prev_category = node_record['prev_category']
- #else:
- # prev_category = "ERROR"
- if node_record['prev_category'] != "NORECORD":
-
- val = cmpCategoryVal(category, prev_category)
- print "%s went from %s -> %s" % (nodename, prev_category, category)
- if val == 1:
- # improved
- if node_record['ticket_id'] == "" or node_record['ticket_id'] == None:
- print "closing record with no ticket: ", node_record['nodename']
- node_record['action'] = ['close_rt']
- node_record['message'] = None
- node_record['stage'] = 'monitor-end-record'
- return node_record
- else:
- node_record['stage'] = 'improvement'
-
- #if 'monitor-end-record' in node_record['stage']:
- # # just ignore it if it's already ended.
- # # otherwise, the status should be worse, and we won't get
- # # here.
- # print "monitor-end-record: ignoring ", node_record['nodename']
- # return None
-#
-# #return None
- elif val == -1:
- # current category is worse than previous, carry on
- pass
- else:
- #values are equal, carry on.
- #print "why are we here?"
- pass
-
- if 'rt' in node_record and 'Status' in node_record['rt']:
- if node_record['stage'] == 'ticket_waitforever':
- if 'resolved' in node_record['rt']['Status']:
- print "ending waitforever record for: ", node_record['nodename']
- node_record['action'] = ['noop']
- node_record['message'] = None
- node_record['stage'] = 'monitor-end-record'
- print "oldlog: %s" % node_record['log'],
- print "%15s" % node_record['action']
- return node_record
- if 'new' in node_record['rt']['Status'] and \
- 'Queue' in node_record['rt'] and \
- 'Monitor' in node_record['rt']['Queue']:
-
- print "RESETTING stage to findbad"
- node_record['stage'] = 'findbad'
-
- #### COMPARE category and prev_category
- # if not_equal
- # then assign a stage based on relative priorities
- # else equal
- # then check category for stats.
- diag_record = self.diagRecordByCategory(node_record)
- if diag_record == None:
- #print "diag_record == None"
- return None
-
- #### found_RT_ticket
- # TODO: need to record time found, and maybe add a stage for acting on it...
- # NOTE: after found, if the support ticket is resolved, the block is
- # not removed. How to remove the block on this?
- if 'found_rt_ticket' in diag_record and \
- diag_record['found_rt_ticket'] is not None:
- if diag_record['stage'] is not 'improvement':
- diag_record['stage'] = 'ticket_waitforever'
+ lb = plccache.plcdb_hn2lb[host]
+ except:
+ print "unknown host in plcdb_hn2lb %s" % host
+ email_exception(host)
+ continue
+
+ nodeblack = BlacklistRecord.get_by(hostname=host)
+
+ if nodeblack and not nodeblack.expired():
+ print "skipping %s due to blacklist. will expire %s" % (host, nodeblack.willExpire() )
+ continue
+
+ sitehist = SiteInterface.get_or_make(loginbase=lb)
+
+ recent_actions = sitehist.getRecentActions(hostname=host)
+
+ nodehist = HistoryNodeRecord.findby_or_create(hostname=host)
+
+ print "%s %s %s" % (i, nodehist.hostname, nodehist.status)
+ if nodehist.status == 'good' and \
+ changed_lessthan(nodehist.last_changed, 1.0) and \
+ found_within(recent_actions, 'down_notice', 7.0) and \
+ not found_within(recent_actions, 'online_notice', 0.5):
+ # NOTE: chronicly flapping nodes will not get 'online' notices
+ # since, they are never up long enough to be 'good'.
+ # NOTE: searching for down_notice proves that the node has
+ # gone through a 'down' state first, rather than just
+ # flapping through: good, offline, online, ...
+ #
+ # NOTE: there is a narrow window in which this command must be
+ # evaluated, otherwise the notice will not go out.
+ # this is not ideal.
+ sitehist.sendMessage('online_notice', hostname=host, viart=False, saveact=True)
+ print "send message for host %s online" % host
+
+
+ # if a node is offline and doesn't have a PCU, remind the user that they should have one.
+ #if not nodehist.haspcu and nodehist.status in ['offline', 'down'] and \
+ # changed_greaterthan(nodehist.last_changed,1.0) and \
+ # not found_within(recent_actions, 'pcumissing_notice', 7.0):
+ #
+ # sitehist.sendMessage('pcumissing_notice', hostname=host)
+ # print "send message for host %s pcumissing_notice" % host
+
+ # if it is offline and HAS a PCU, then try to use it.
+ if nodehist.haspcu and nodehist.status in ['offline', 'down'] and \
+ changed_greaterthan(nodehist.last_changed,1.0) and \
+ not found_between(recent_actions, 'try_reboot', 3.5, 1):
+
+ sitehist.attemptReboot(host)
+ print "send message for host %s try_reboot" % host
+
+ # NOTE: non-intuitive is that found_between(try_reboot, 3.5, 1)
+ # will be false for a day after the above condition is satisfied
+ if nodehist.haspcu and nodehist.status in ['offline', 'down'] and \
+ changed_greaterthan(nodehist.last_changed,1.5) and \
+ found_between(recent_actions, 'try_reboot', 3.5, 1) and \
+ not found_within(recent_actions, 'pcufailed_notice', 3.5):
- current_time = time.time()
- # take off four days, for the delay that database caused.
- # TODO: generalize delays at PLC, and prevent enforcement when there
- # have been no emails.
- # NOTE: 7*SPERDAY exists to offset the 'bad week'
- #delta = current_time - diag_record['time'] - 7*SPERDAY
- delta = current_time - diag_record['time']
-
- message = diag_record['message']
- act_record = {}
- act_record.update(diag_record)
-
- #### DIAGNOSE STAGES
- if 'findbad' in diag_record['stage']:
- # The node is bad, and there's no previous record of it.
- act_record['email'] = TECH
- act_record['action'] = ['noop']
- act_record['message'] = message[0]
- act_record['stage'] = 'stage_actinoneweek'
-
- elif 'nmreset' in diag_record['stage']:
- act_record['email'] = ADMIN
- act_record['action'] = ['reset_nodemanager']
- act_record['message'] = message[0]
- act_record['stage'] = 'nmreset'
- return None
-
- elif 'reboot_node' in diag_record['stage']:
- act_record['email'] = TECH
- act_record['action'] = ['noop']
- act_record['message'] = message[0]
- act_record['stage'] = 'stage_actinoneweek'
-
- elif 'improvement' in diag_record['stage']:
- # - backoff previous squeeze actions (slice suspend, nocreate)
- # TODO: add a backoff_squeeze section... Needs to runthrough
- print "backing off of %s" % nodename
- act_record['action'] = ['close_rt']
- act_record['message'] = message[0]
- act_record['stage'] = 'monitor-end-record'
-
- elif 'actinoneweek' in diag_record['stage']:
- if delta >= 7 * SPERDAY:
- act_record['email'] = TECH | PI
- act_record['stage'] = 'stage_actintwoweeks'
- act_record['message'] = message[1]
- act_record['action'] = ['nocreate' ]
- act_record['time'] = current_time # reset clock for waitforever
- elif delta >= 3* SPERDAY and not 'second-mail-at-oneweek' in act_record:
- act_record['email'] = TECH
- act_record['message'] = message[0]
- act_record['action'] = ['sendmailagain-waitforoneweekaction' ]
- act_record['second-mail-at-oneweek'] = True
- else:
- act_record['message'] = None
- act_record['action'] = ['waitforoneweekaction' ]
- print "ignoring this record for: %s" % act_record['nodename']
- return None # don't send if there's no action
-
- elif 'actintwoweeks' in diag_record['stage']:
- if delta >= 7 * SPERDAY:
- act_record['email'] = TECH | PI | USER
- act_record['stage'] = 'stage_waitforever'
- act_record['message'] = message[2]
- act_record['action'] = ['suspendslices']
- act_record['time'] = current_time # reset clock for waitforever
- elif delta >= 3* SPERDAY and not 'second-mail-at-twoweeks' in act_record:
- act_record['email'] = TECH | PI
- act_record['message'] = message[1]
- act_record['action'] = ['sendmailagain-waitfortwoweeksaction' ]
- act_record['second-mail-at-twoweeks'] = True
- else:
- act_record['message'] = None
- act_record['action'] = ['waitfortwoweeksaction']
- return None # don't send if there's no action
-
- elif 'ticket_waitforever' in diag_record['stage']:
- act_record['email'] = TECH
- if 'first-found' not in act_record:
- act_record['first-found'] = True
- act_record['log'] += " firstfound"
- act_record['action'] = ['ticket_waitforever']
- act_record['message'] = message[0]
- act_record['time'] = current_time
- else:
- if delta >= 7*SPERDAY:
- act_record['action'] = ['ticket_waitforever']
- if 'rt' in act_record and 'Status' in act_record['rt'] and \
- act_record['rt']['Status'] == 'new':
- act_record['message'] = message[0]
- else:
- act_record['message'] = None
-
- act_record['time'] = current_time # reset clock
- else:
- act_record['action'] = ['ticket_waitforever']
- act_record['message'] = None
- return None
-
- elif 'waitforever' in diag_record['stage']:
- # more than 3 days since last action
- # TODO: send only on weekdays.
- # NOTE: expects that 'time' has been reset before entering waitforever stage
- if delta >= 3*SPERDAY:
- act_record['action'] = ['email-againwaitforever']
- act_record['message'] = message[2]
- act_record['time'] = current_time # reset clock
- else:
- act_record['action'] = ['waitforever']
- act_record['message'] = None
- return None # don't send if there's no action
-
- else:
- # There is no action to be taken, possibly b/c the stage has
- # already been performed, but diagnose picked it up again.
- # two cases,
- # 1. stage is unknown, or
- # 2. delta is not big enough to bump it to the next stage.
- # TODO: figure out which. for now assume 2.
- print "UNKNOWN stage for %s; nothing done" % nodename
- act_record['action'] = ['unknown']
- act_record['message'] = message[0]
-
- act_record['email'] = TECH
- act_record['action'] = ['noop']
- act_record['message'] = message[0]
- act_record['stage'] = 'stage_actinoneweek'
- act_record['time'] = current_time # reset clock
- #print "Exiting..."
- #return None
- #sys.exit(1)
-
- print "%s" % act_record['log'],
- print "%15s" % act_record['action']
- return act_record
-
- def getMaxSlices(self, loginbase):
- # if sickdb has a loginbase, then it will have at least one node.
- site_stats = None
-
- for nodename in self.diagnose_in[loginbase].keys():
- if nodename in self.findbad['nodes']:
- site_stats = self.findbad['nodes'][nodename]['values']['plcsite']
- break
-
- if site_stats == None:
- raise Exception, "loginbase with no nodes in findbad"
- else:
- return site_stats['max_slices']
-
- def getNumNodes(self, loginbase):
- # if sickdb has a loginbase, then it will have at least one node.
- site_stats = None
-
- for nodename in self.diagnose_in[loginbase].keys():
- if nodename in self.findbad['nodes']:
- site_stats = self.findbad['nodes'][nodename]['values']['plcsite']
- break
-
- if site_stats == None:
- raise Exception, "loginbase with no nodes in findbad"
+ # send pcu failure message
+ #act = ActionRecord(**kwargs)
+ sitehist.sendMessage('pcufailed_notice', hostname=host)
+ print "send message for host %s PCU Failure" % host
+
+ if nodehist.status == 'monitordebug' and \
+ changed_greaterthan(nodehist.last_changed, 1) and \
+ not found_between(recent_actions, 'bootmanager_restore', 0.5, 0):
+ # send down node notice
+ # delay 0.5 days before retrying...
+
+ print "send message for host %s bootmanager_restore" % host
+ sitehist.runBootManager(host)
+ # sitehist.sendMessage('retry_bootman', hostname=host)
+
+ if nodehist.status == 'down' and \
+ changed_greaterthan(nodehist.last_changed, 2) and \
+ not found_within(recent_actions, 'down_notice', 3.5):
+ # send down node notice
+
+ sitehist.sendMessage('down_notice', hostname=host)
+ print "send message for host %s down" % host
+
+ node_count = node_count + 1
+ print "time: ", time.strftime('%Y-%m-%d %H:%M:%S')
+ sys.stdout.flush()
+ session.flush()
+
+ for i,site in enumerate(sitenames):
+ sitehist = SiteInterface.get_or_make(loginbase=site)
+ siteblack = BlacklistRecord.get_by(loginbase=site)
+ skip_due_to_blacklist=False
+
+ if siteblack and not siteblack.expired():
+ print "skipping %s due to blacklist. will expire %s" % (site, siteblack.willExpire() )
+ skip_due_to_blacklist=True
+ continue
+
+ # TODO: make query only return records within a certin time range,
+ # i.e. greater than 0.5 days ago. or 5 days, etc.
+ recent_actions = sitehist.getRecentActions(loginbase=site)
+
+ print "%s %s %s" % (i, sitehist.db.loginbase, sitehist.db.status)
+
+ # determine if there are penalties within the last 30 days?
+ # if so, add a 'pause_penalty' action.
+ if sitehist.db.message_id != 0 and sitehist.db.message_status == 'open' and \
+ sitehist.db.penalty_level > 0 and not found_within(recent_actions, 'pause_penalty', 30):
+ # pause escalation
+ print "Pausing penalties for %s" % site
+ sitehist.pausePenalty()
else:
- if 'num_nodes' in site_stats:
- return site_stats['num_nodes']
- else:
- return 0
- """
- Returns number of up nodes as the total number *NOT* in act_all with a
- stage other than 'steady-state' .
- """
- def getUpAtSite(self, loginbase, d_diag_site):
- # TODO: THIS DOESN"T WORK!!! it misses all the 'debug' state nodes
- # that aren't recorded yet.
+ if sitehist.db.status == 'down':
+ if not found_within(recent_actions, 'pause_penalty', 30) and \
+ not found_within(recent_actions, 'increase_penalty', 7) and \
+ changed_greaterthan(sitehist.db.last_changed, 7):
- numnodes = self.getNumNodes(loginbase)
- # NOTE: assume nodes we have no record of are ok. (too conservative)
- # TODO: make the 'up' value more representative
- up = numnodes
- for nodename in d_diag_site[loginbase]['nodes'].keys():
+ # TODO: catch errors
+ sitehist.increasePenalty()
+ sitehist.applyPenalty()
+ sitehist.sendMessage('increase_penalty')
- rec = d_diag_site[loginbase]['nodes'][nodename]
- if rec['stage'] != 'monitor-end-record':
- up -= 1
- else:
- pass # the node is assumed to be up.
+ print "send message for site %s penalty increase" % site
- #if up != numnodes:
- # print "ERROR: %s total nodes up and down != %d" % (loginbase, numnodes)
+ if sitehist.db.status == 'good':
+ # clear penalty
+ # NOTE: because 'all clear' should have an indefinite status, we
+ # have a boolean value rather than a 'recent action'
+ if sitehist.db.penalty_applied:
+ # send message that penalties are cleared.
- return up
+ sitehist.clearPenalty()
+ sitehist.applyPenalty()
+ sitehist.sendMessage('clear_penalty')
+ sitehist.closeTicket()
+ print "send message for site %s penalty cleared" % site
-class SiteAction:
- def __init__(self, parameter_names=['hostname', 'ticket_id']):
- self.parameter_names = parameter_names
- def checkParam(self, args):
- for param in self.parameter_names:
- if param not in args:
- raise Exception("Parameter %s not provided in args"%param)
- def run(self, args):
- self.checkParam(args)
- return self._run(args)
- def _run(self, args):
- pass
-class SuspendAction(SiteAction):
- def _run(self, args):
- return plc.suspendSlices(args['hostname'])
+ site_count = site_count + 1
-class RemoveSliceCreation(SiteAction):
- def _run(self, args):
- return plc.removeSliceCreation(args['hostname'])
+ print "time: ", time.strftime('%Y-%m-%d %H:%M:%S')
+ sys.stdout.flush()
+ session.flush()
-class BackoffActions(SiteAction):
- def _run(self, args):
- plc.enableSlices(args['hostname'])
- plc.enableSliceCreation(args['hostname'])
- return True
-
-# TODO: create class for each action below,
-# allow for lists of actions to be performed...
-
-def close_rt_backoff(args):
- if 'ticket_id' in args and (args['ticket_id'] != "" and args['ticket_id'] != None):
- mailer.closeTicketViaRT(args['ticket_id'],
- "Ticket CLOSED automatically by SiteAssist.")
- plc.enableSlices(args['hostname'])
- plc.enableSliceCreation(args['hostname'])
- return
-
-def reboot_node(args):
- host = args['hostname']
- return reboot.reboot_policy(host, True, config.debug)
-
-def reset_nodemanager(args):
- os.system("ssh root@%s /sbin/service nm restart" % nodename)
+ session.flush()
return
-class Action(Thread):
- def __init__(self, l_action):
- self.l_action = l_action
-
- # the hostname to loginbase mapping
- self.plcdb_hn2lb = database.dbLoad("plcdb_hn2lb")
-
- # Actions to take.
- self.diagnose_db = database.if_cached_else(1, "diagnose_out", lambda : {})
- # Actions taken.
- self.act_all = database.if_cached_else(1, "act_all", lambda : {})
-
- # A dict of actions to specific functions. PICKLE doesnt' like lambdas.
- self.actions = {}
- self.actions['suspendslices'] = lambda args: plc.suspendSlices(args['hostname'])
- self.actions['nocreate'] = lambda args: plc.removeSliceCreation(args['hostname'])
- self.actions['close_rt'] = lambda args: close_rt_backoff(args)
- self.actions['rins'] = lambda args: plc.nodeBootState(args['hostname'], "rins")
- self.actions['noop'] = lambda args: args
- self.actions['reboot_node'] = lambda args: reboot_node(args)
- self.actions['reset_nodemanager'] = lambda args: args # reset_nodemanager(args)
-
- self.actions['ticket_waitforever'] = lambda args: args
- self.actions['waitforever'] = lambda args: args
- self.actions['unknown'] = lambda args: args
- self.actions['waitforoneweekaction'] = lambda args: args
- self.actions['waitfortwoweeksaction'] = lambda args: args
- self.actions['sendmailagain-waitforoneweekaction'] = lambda args: args
- self.actions['sendmailagain-waitfortwoweeksaction'] = lambda args: args
- self.actions['email-againwaitforever'] = lambda args: args
- self.actions['email-againticket_waitforever'] = lambda args: args
-
-
- self.sickdb = {}
- Thread.__init__(self)
-
- def run(self):
- self.accumSites()
- print "Accumulated %d sick sites" % len(self.sickdb.keys())
- logger.debug("Accumulated %d sick sites" % len(self.sickdb.keys()))
-
- try:
- stats = self.analyseSites()
- except Exception, err:
- print "----------------"
- import traceback
- print traceback.print_exc()
- print err
- if config.policysavedb:
- print "Saving Databases... act_all"
- database.dbDump("act_all", self.act_all)
- sys.exit(1)
-
- print_stats("sites_observed", stats)
- print_stats("sites_diagnosed", stats)
- print_stats("nodes_diagnosed", stats)
- print_stats("sites_emailed", stats)
- print_stats("nodes_actedon", stats)
- print string.join(stats['allsites'], ",")
-
- if config.policysavedb:
- print "Saving Databases... act_all"
- #database.dbDump("policy.eventlog", self.eventlog)
- # TODO: remove 'diagnose_out',
- # or at least the entries that were acted on.
- database.dbDump("act_all", self.act_all)
-
- def accumSites(self):
- """
- Take all nodes, from l_action, look them up in the diagnose_db database,
- and insert them into sickdb[] as:
-
- This way only the given l_action nodes will be acted on regardless
- of how many from diagnose_db are available.
-
- sickdb[loginbase][nodename] = diag_record
- """
- # TODO: what if l_action == None ?
- for nodename in self.l_action:
-
- loginbase = self.plcdb_hn2lb[nodename]
-
- if loginbase in self.diagnose_db and \
- nodename in self.diagnose_db[loginbase]['nodes']:
-
- diag_record = self.diagnose_db[loginbase]['nodes'][nodename]
-
- if loginbase not in self.sickdb:
- self.sickdb[loginbase] = {'nodes' : {}}
-
- # NOTE: don't copy all node records, since not all will be in l_action
- self.sickdb[loginbase]['nodes'][nodename] = diag_record
- # NOTE: but, we want to get the loginbase config settings,
- # this is the easiest way.
- self.sickdb[loginbase]['config'] = self.diagnose_db[loginbase]['config']
- #else:
- #print "%s not in diagnose_db!!" % loginbase
- return
-
- def __emailSite(self, loginbase, roles, message, args):
- """
- loginbase is the unique site abbreviation, prepended to slice names.
- roles contains TECH, PI, USER roles, and derive email aliases.
- record contains {'message': [<subj>,<body>], 'args': {...}}
- """
- ticket_id = 0
- args.update({'loginbase':loginbase})
-
- if not config.mail and not config.debug and config.bcc:
- roles = ADMIN
- if config.mail and config.debug:
- roles = ADMIN
-
- # build targets
- contacts = []
- if ADMIN & roles:
- contacts += [config.email]
- if TECH & roles:
- contacts += [TECHEMAIL % loginbase]
- if PI & roles:
- contacts += [PIEMAIL % loginbase]
- if USER & roles:
- slices = plc.slices(loginbase)
- if len(slices) >= 1:
- for slice in slices:
- contacts += [SLICEMAIL % slice]
- print "SLIC: %20s : %d slices" % (loginbase, len(slices))
- else:
- print "SLIC: %20s : 0 slices" % loginbase
-
- try:
- subject = message[0] % args
- body = message[1] % args
- if ADMIN & roles:
- # send only to admin
- if 'ticket_id' in args:
- subj = "Re: [PL #%s] %s" % (args['ticket_id'], subject)
- else:
- subj = "Re: [PL noticket] %s" % subject
- mailer.email(subj, body, contacts)
- ticket_id = args['ticket_id']
- else:
- ticket_id = mailer.emailViaRT(subject, body, contacts, args['ticket_id'])
- except Exception, err:
- print "exception on message:"
- import traceback
- print traceback.print_exc()
- print message
-
- return ticket_id
-
-
- def _format_diaginfo(self, diag_node):
- info = diag_node['info']
- if diag_node['stage'] == 'monitor-end-record':
- hlist = " %s went from '%s' to '%s'\n" % (info[0], info[1], info[2])
- else:
- hlist = " %s %s - %s\n" % (info[0], info[2], info[1]) #(node,ver,daysdn)
- return hlist
-
-
- def get_email_args(self, act_recordlist, loginbase=None):
-
- email_args = {}
- email_args['hostname_list'] = ""
-
- for act_record in act_recordlist:
- email_args['hostname_list'] += act_record['msg_format']
- email_args['hostname'] = act_record['nodename']
- if 'plcnode' in act_record and \
- 'pcu_ids' in act_record['plcnode'] and \
- len(act_record['plcnode']['pcu_ids']) > 0:
- print "setting 'pcu_id' for email_args %s"%email_args['hostname']
- email_args['pcu_id'] = act_record['plcnode']['pcu_ids'][0]
- else:
- email_args['pcu_id'] = "-1"
-
- if 'ticket_id' in act_record:
- if act_record['ticket_id'] == 0 or act_record['ticket_id'] == '0':
- print "Enter the ticket_id for %s @ %s" % (loginbase, act_record['nodename'])
- sys.stdout.flush()
- line = sys.stdin.readline()
- try:
- ticket_id = int(line)
- except:
- print "could not get ticket_id from stdin..."
- os._exit(1)
- else:
- ticket_id = act_record['ticket_id']
-
- email_args['ticket_id'] = ticket_id
-
- return email_args
-
- def get_unique_issues(self, act_recordlist):
- # NOTE: only send one email per site, per problem...
- unique_issues = {}
- for act_record in act_recordlist:
- act_key = act_record['action'][0]
- if act_key not in unique_issues:
- unique_issues[act_key] = []
-
- unique_issues[act_key] += [act_record]
-
- return unique_issues
-
-
- def __actOnSite(self, loginbase, site_record):
- i_nodes_actedon = 0
- i_nodes_emailed = 0
-
- act_recordlist = []
-
- for nodename in site_record['nodes'].keys():
- diag_record = site_record['nodes'][nodename]
- act_record = self.__actOnNode(diag_record)
- #print "nodename: %s %s" % (nodename, act_record)
- if act_record is not None:
- act_recordlist += [act_record]
-
- unique_issues = self.get_unique_issues(act_recordlist)
-
- for issue in unique_issues.keys():
- print "\tworking on issue: %s" % issue
- issue_record_list = unique_issues[issue]
- email_args = self.get_email_args(issue_record_list, loginbase)
-
- # for each record.
- for act_record in issue_record_list:
- # if there's a pcu record and email config is set
- if 'email_pcu' in act_record:
- if act_record['message'] != None and act_record['email_pcu'] and site_record['config']['email']:
- # and 'reboot_node' in act_record['stage']:
-
- email_args['hostname'] = act_record['nodename']
- ticket_id = self.__emailSite(loginbase,
- act_record['email'],
- emailTxt.mailtxt.pcudown[0],
- email_args)
- if ticket_id == 0:
- # error.
- print "got a ticket_id == 0!!!! %s" % act_record['nodename']
- os._exit(1)
- pass
- email_args['ticket_id'] = ticket_id
-
-
- act_record = issue_record_list[0]
- # send message before squeezing
- print "\t\tconfig.email: %s and %s" % (act_record['message'] != None,
- site_record['config']['email'])
- if act_record['message'] != None and site_record['config']['email']:
- ticket_id = self.__emailSite(loginbase, act_record['email'],
- act_record['message'], email_args)
-
- if ticket_id == 0:
- # error.
- print "ticket_id == 0 for %s %s" % (loginbase, act_record['nodename'])
- os._exit(1)
- pass
-
- # Add ticket_id to ALL nodenames
- for act_record in issue_record_list:
- nodename = act_record['nodename']
- # update node record with RT ticket_id
- if nodename in self.act_all:
- self.act_all[nodename][0]['ticket_id'] = "%s" % ticket_id
- # if the ticket was previously resolved, reset it to new.
- if 'rt' in act_record and \
- 'Status' in act_record['rt'] and \
- act_record['rt']['Status'] == 'resolved':
- mailer.setTicketStatus(ticket_id, "new")
- status = mailer.getTicketStatus(ticket_id)
- self.act_all[nodename][0]['rt'] = status
- if config.mail: i_nodes_emailed += 1
-
- print "\t\tconfig.squeeze: %s and %s" % (config.squeeze,
- site_record['config']['squeeze'])
- if config.squeeze and site_record['config']['squeeze']:
- for act_key in act_record['action']:
- self.actions[act_key](email_args)
- i_nodes_actedon += 1
-
- if config.policysavedb:
- print "Saving Databases... act_all, diagnose_out"
- database.dbDump("act_all", self.act_all)
- # remove site record from diagnose_out, it's in act_all as done.
- del self.diagnose_db[loginbase]
- database.dbDump("diagnose_out", self.diagnose_db)
-
- print "sleeping for 1 sec"
- time.sleep(1)
- #print "Hit enter to continue..."
- #sys.stdout.flush()
- #line = sys.stdin.readline()
-
- return (i_nodes_actedon, i_nodes_emailed)
-
- def __actOnNode(self, diag_record):
- nodename = diag_record['nodename']
- message = diag_record['message']
-
- act_record = {}
- act_record.update(diag_record)
- act_record['nodename'] = nodename
- act_record['msg_format'] = self._format_diaginfo(diag_record)
- print "act_record['stage'] == %s " % act_record['stage']
-
- # avoid end records, and nmreset records
- # reboot_node_failed, is set below, so don't reboot repeatedly.
-
- if 'monitor-end-record' not in act_record['stage'] and \
- 'nmreset' not in act_record['stage'] and \
- 'reboot_node_failed' not in act_record:
-
- if "DOWN" in act_record['log'] and \
- 'pcu_ids' in act_record['plcnode'] and \
- len(act_record['plcnode']['pcu_ids']) > 0:
-
- print "%s" % act_record['log'],
- print "%15s" % (['reboot_node'],)
- # Set node to re-install
- plc.nodeBootState(act_record['nodename'], "rins")
- try:
- ret = reboot_node({'hostname': act_record['nodename']})
- except Exception, exc:
- print "exception on reboot_node:"
- import traceback
- print traceback.print_exc()
- ret = False
-
- if ret: # and ( 'reboot_node_failed' not in act_record or act_record['reboot_node_failed'] == False):
- # Reboot Succeeded
- print "reboot succeeded for %s" % act_record['nodename']
- act_record2 = {}
- act_record2.update(act_record)
- act_record2['action'] = ['reboot_node']
- act_record2['stage'] = "reboot_node"
- act_record2['reboot_node_failed'] = False
- act_record2['email_pcu'] = False
-
- if nodename not in self.act_all:
- self.act_all[nodename] = []
- print "inserting 'reboot_node' record into act_all"
- self.act_all[nodename].insert(0,act_record2)
-
- # return None to avoid further action
- print "Taking no further action"
- return None
- else:
- print "reboot failed for %s" % act_record['nodename']
- # set email_pcu to also send pcu notice for this record.
- act_record['reboot_node_failed'] = True
- act_record['email_pcu'] = True
-
- print "%s" % act_record['log'],
- print "%15s" % act_record['action']
-
- if act_record['stage'] is not 'monitor-end-record' and \
- act_record['stage'] is not 'nmreset':
- if nodename not in self.act_all:
- self.act_all[nodename] = []
-
- self.act_all[nodename].insert(0,act_record)
- else:
- print "Not recording %s in act_all" % nodename
-
- return act_record
-
- def analyseSites(self):
- i_sites_observed = 0
- i_sites_diagnosed = 0
- i_nodes_diagnosed = 0
- i_nodes_actedon = 0
- i_sites_emailed = 0
- l_allsites = []
-
- sorted_sites = self.sickdb.keys()
- sorted_sites.sort()
- for loginbase in sorted_sites:
- site_record = self.sickdb[loginbase]
- print "sites: %s" % loginbase
-
- i_nodes_diagnosed += len(site_record.keys())
- i_sites_diagnosed += 1
-
- (na,ne) = self.__actOnSite(loginbase, site_record)
-
- i_sites_observed += 1
- i_nodes_actedon += na
- i_sites_emailed += ne
-
- l_allsites += [loginbase]
-
- return {'sites_observed': i_sites_observed,
- 'sites_diagnosed': i_sites_diagnosed,
- 'nodes_diagnosed': i_nodes_diagnosed,
- 'sites_emailed': i_sites_emailed,
- 'nodes_actedon': i_nodes_actedon,
- 'allsites':l_allsites}
-
- def print_stats(self, key, stats):
- print "%20s : %d" % (key, stats[key])
-
-
-
- #"""
- #Prints, logs, and emails status of up nodes, down nodes, and buckets.
- #"""
- #def status(self):
- # sub = "Monitor Summary"
- # msg = "\nThe following nodes were acted upon: \n\n"
- # for (node, (type, date)) in self.emailed.items():
- # # Print only things acted on today.
- # if (time.gmtime(time.time())[2] == time.gmtime(date)[2]):
- # msg +="%s\t(%s)\t%s\n" %(node, type, time.ctime(date))
- # msg +="\n\nThe following sites have been 'squeezed':\n\n"
- # for (loginbase, (date, type)) in self.squeezed.items():
- # # Print only things acted on today.
- # if (time.gmtime(time.time())[2] == time.gmtime(date)[2]):
- # msg +="%s\t(%s)\t%s\n" %(loginbase, type, time.ctime(date))
- # mailer.email(sub, msg, [SUMTO])
- # logger.info(msg)
- # return
-
- #"""
- #Store/Load state of emails. When, where, what.
- #"""
- #def emailedStore(self, action):
- # try:
- # if action == "LOAD":
- # f = open(DAT, "r+")
- # logger.info("POLICY: Found and reading " + DAT)
- # self.emailed.update(pickle.load(f))
- # if action == "WRITE":
- # f = open(DAT, "w")
- # #logger.debug("Writing " + DAT)
- # pickle.dump(self.emailed, f)
- # f.close()
- # except Exception, err:
- # logger.info("POLICY: Problem with DAT, %s" %err)
-
-
-#class Policy(Thread):
-def main():
- print "policy.py is a module, not a script for running directly."
+if __name__ == "__main__":
+ parser = parsermodule.getParser(['nodesets'])
+ parser.set_defaults( timewait=0,
+ skip=0,
+ rins=False,
+ reboot=False,
+ findbad=False,
+ force=False,
+ nosetup=False,
+ verbose=False,
+ quiet=False,)
+
+ parser.add_option("", "--stopselect", dest="stopselect", metavar="",
+ help="The select string that must evaluate to true for the node to be considered 'done'")
+ parser.add_option("", "--findbad", dest="findbad", action="store_true",
+ help="Re-run findbad on the nodes we're going to check before acting.")
+ parser.add_option("", "--force", dest="force", action="store_true",
+ help="Force action regardless of previous actions/logs.")
+ parser.add_option("", "--rins", dest="rins", action="store_true",
+ help="Set the boot_state to 'rins' for all nodes.")
+ parser.add_option("", "--reboot", dest="reboot", action="store_true",
+ help="Actively try to reboot the nodes, keeping a log of actions.")
+
+ parser.add_option("", "--verbose", dest="verbose", action="store_true",
+ help="Extra debug output messages.")
+ parser.add_option("", "--nosetup", dest="nosetup", action="store_true",
+ help="Do not perform the orginary setup phase.")
+ parser.add_option("", "--skip", dest="skip",
+ help="Number of machines to skip on the input queue.")
+ parser.add_option("", "--timewait", dest="timewait",
+ help="Minutes to wait between iterations of 10 nodes.")
+
+ parser = parsermodule.getParser(['defaults'], parser)
+ config = parsermodule.parse_args(parser)
+
+ fbquery = HistoryNodeRecord.query.all()
+ hostnames = [ n.hostname for n in fbquery ]
+
+ fbquery = HistorySiteRecord.query.all()
+ sitenames = [ s.loginbase for s in fbquery ]
+
+ if config.site:
+ # TODO: replace with calls to local db. the api fails so often that
+ # these calls should be regarded as unreliable.
+ l_nodes = plccache.GetNodesBySite(config.site)
+ filter_hostnames = [ n['hostname'] for n in l_nodes ]
+
+ hostnames = filter(lambda x: x in filter_hostnames, hostnames)
+ sitenames = [config.site]
+
+ if config.node:
+ hostnames = [ config.node ]
+ sitenames = [ plccache.plcdb_hn2lb[config.node] ]
-if __name__ == '__main__':
- import os
- import plc
try:
- main()
+ main(hostnames, sitenames)
+ session.flush()
except KeyboardInterrupt:
- print "Killed. Exitting."
- logger.info('Monitor Killed')
- os._exit(0)
+ print "Killed by interrupt"
+ session.flush()
+ sys.exit(0)
+ except:
+ email_exception()
+ print traceback.print_exc();
+ print "fail all..."