X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=policy.py;h=78232375ee593c21319dcaff10e11837bf90e757;hb=refs%2Fheads%2F1.0;hp=f6629fb8e2b0cbd0a5d444dfdefb2d2eb29d9f64;hpb=dbfe6652a8654ec633c4ffb9e86e9a61a5516dc6;p=monitor.git diff --git a/policy.py b/policy.py index f6629fb..7823237 100644 --- a/policy.py +++ b/policy.py @@ -3,7 +3,7 @@ # # Faiyaz Ahmed # -# $Id: policy.py,v 1.7 2007/01/11 21:39:07 faiyaza Exp $ +# $Id: policy.py,v 1.17 2007/08/29 17:26:50 soltesz Exp $ # # Policy Engine. @@ -16,7 +16,12 @@ import emailTxt import pickle import Queue import plc +import sys +import os import reboot +import database +import string +from unified_model import cmpCategoryVal import config DAT="./monitor.dat" @@ -27,22 +32,31 @@ logger = logging.getLogger("monitor") POLSLEEP = 7200 # Where to email the summary -SUMTO = "faiyaza@cs.princeton.edu" +SUMTO = "soltesz@cs.princeton.edu" TECHEMAIL="tech-%s@sites.planet-lab.org" PIEMAIL="pi-%s@sites.planet-lab.org" SLICEMAIL="%s@slices.planet-lab.org" PLCEMAIL="support@planet-lab.org" #Thresholds (DAYS) +SPERMIN = 60 +SPERHOUR = 60*60 SPERDAY = 86400 -PITHRESH = 2 * SPERDAY -SLICETHRESH = 5 * SPERDAY +PITHRESH = 7 * SPERDAY +SLICETHRESH = 7 * SPERDAY # Days before attempting rins again RINSTHRESH = 5 * SPERDAY +# Days before calling the node dead. +DEADTHRESH = 30 * SPERDAY # Minimum number of nodes up before squeezing MINUP = 2 +TECH=1 +PI=2 +USER=4 +ADMIN=8 + # IF: # no SSH, down. # bad disk, down @@ -54,242 +68,1312 @@ MINUP = 2 # Email # suspend slice creation # kill slices -class Policy(Thread): - def __init__(self, comonthread, sickNoTicket, emailed): - self.cmn = comonthread - # host - > (time of email, type of email) - self.emailed = emailed - # all sick nodes w/o tickets - self.sickNoTicket = sickNoTicket - # Sitess we've Squeezed. - self.squeezed = {} +def array_to_priority_map(array): + """ Create a mapping where each entry of array is given a priority equal + to its position in the array. This is useful for subsequent use in the + cmpMap() function.""" + map = {} + count = 0 + for i in array: + map[i] = count + count += 1 + return map + +def getdebug(): + return config.debug + +def print_stats(key, stats): + if key in stats: print "%20s : %d" % (key, stats[key]) + + +class Merge(Thread): + def __init__(self, l_merge, toRT): + self.toRT = toRT + self.merge_list = l_merge + # the hostname to loginbase mapping + self.plcdb_hn2lb = database.dbLoad("plcdb_hn2lb") + + # Previous actions taken on nodes. + self.act_all = database.if_cached_else(1, "act_all", lambda : {}) + self.findbad = database.if_cached_else(1, "findbad", lambda : {}) + + self.cache_all = database.if_cached_else(1, "act_all", lambda : {}) + self.sickdb = {} + self.mergedb = {} Thread.__init__(self) - - - ''' - What to do when node is in dbg (as reported by CoMon). - ''' - def __actOnDebug(self, node): - # Check to see if we've done this before - if (node in self.emailed.keys()): - if (self.emailed[node][0] == "dbg"): - delta = time.time() - self.emailed[node][1] - if (delta <= RINSTHRESH ): - # Don't mess with node if under Thresh. - # Return, move on. - logger.info("POLICY: %s in dbg, but acted on %s days ago" % (node, delta // SPERDAY)) - return - logger.info("POLICY: Node in dbg - " + node) - plc.nodeBootState(node, "rins") - # If it has a PCU - return reboot.reboot(node) - - ''' - What to do when node is in dbg (as reported by CoMon). - ''' - def __actOnFilerw(self, node): - target = [PLCEMAIL] - logger.info("POLICY: Emailing PLC for " + node) - tmp = emailTxt.mailtxt.filerw - sbj = tmp[0] % {'hostname': node} - msg = tmp[1] % {'hostname': node} - mailer.email(sbj, msg, target) - self.emailed[node] = ("filerw", time.time()) - - - ''' - Acts on sick nodes. - ''' - def actOnSick(self): - # Get list of nodes in debug from PLC - #dbgNodes = NodesDebug() - global TECHEMAIL, PIEMAIL - # Grab a node from the queue (pushed by rt thread). - node = self.sickNoTicket.get(block = True) - # Get the login base - loginbase = plc.siteId(node) + + def run(self): + # populate sickdb + self.accumSickSites() + # read data from findbad and act_all + self.mergeActionsAndBadDB() + # pass node_records to RT + self.sendToRT() + + def accumSickSites(self): + """ + Take all nodes, from l_diagnose, look them up in the act_all database, + and insert them into sickdb[] as: + + sickdb[loginbase][nodename] = fb_record + """ + # look at all problems reported by findbad + l_nodes = self.findbad['nodes'].keys() + count = 0 + for nodename in l_nodes: + if nodename not in self.merge_list: + continue # skip this node, since it's not wanted + + count += 1 + loginbase = self.plcdb_hn2lb[nodename] + values = self.findbad['nodes'][nodename]['values'] + + fb_record = {} + fb_record['nodename'] = nodename + try: + fb_record['category'] = values['category'] + except: + print values + print nodename + print self.findbad['nodes'][nodename] + count -= 1 + continue + fb_record['state'] = values['state'] + fb_record['comonstats'] = values['comonstats'] + fb_record['plcnode'] = values['plcnode'] + fb_record['kernel'] = self.getKernel(values['kernel']) + fb_record['stage'] = "findbad" + fb_record['message'] = None + fb_record['bootcd'] = values['bootcd'] + fb_record['args'] = None + fb_record['info'] = None + fb_record['time'] = time.time() + fb_record['date_created'] = time.time() + + if loginbase not in self.sickdb: + self.sickdb[loginbase] = {} + + self.sickdb[loginbase][nodename] = fb_record + + print "Found %d nodes" % count + + def getKernel(self, unamestr): + s = unamestr.split() + if len(s) > 2: + return s[2] + else: + return "" + + def mergeActionsAndBadDB(self): + """ + - Look at the sick node_records as reported in findbad, + - Then look at the node_records in act_all. + + There are four cases: + 1) Problem in findbad, no problem in act_all + this ok, b/c it just means it's a new problem + 2) Problem in findbad, problem in act_all + -Did the problem get better or worse? + -If Same, or Worse, then continue looking for open tickets. + -If Better, or No problem, then "back-off" penalties. + This judgement may need to wait until 'Diagnose()' + + 3) No problem in findbad, problem in act_all + The the node is operational again according to Findbad() + + 4) No problem in findbad, no problem in act_all + There won't be a record in either db, so there's no code. + """ + + sorted_sites = self.sickdb.keys() + sorted_sites.sort() + # look at all problems reported by findbad + for loginbase in sorted_sites: + d_fb_nodes = self.sickdb[loginbase] + sorted_nodes = d_fb_nodes.keys() + sorted_nodes.sort() + for nodename in sorted_nodes: + fb_record = self.sickdb[loginbase][nodename] + x = fb_record + if loginbase not in self.mergedb: + self.mergedb[loginbase] = {} + + # take the info either from act_all or fb-record. + # if node not in act_all + # then take it from fbrecord, obviously. + # else node in act_all + # if act_all == 0 length (no previous records) + # then take it from fbrecord. + # else + # take it from act_all. + # + + # We must compare findbad state with act_all state + if nodename not in self.act_all: + # 1) ok, b/c it's a new problem. set ticket_id to null + self.mergedb[loginbase][nodename] = {} + self.mergedb[loginbase][nodename].update(x) + self.mergedb[loginbase][nodename]['ticket_id'] = "" + self.mergedb[loginbase][nodename]['prev_category'] = "NORECORD" + else: + if len(self.act_all[nodename]) == 0: + self.mergedb[loginbase][nodename] = {} + self.mergedb[loginbase][nodename].update(x) + self.mergedb[loginbase][nodename]['ticket_id'] = "" + self.mergedb[loginbase][nodename]['prev_category'] = "NORECORD" + else: + y = self.act_all[nodename][0] + y['prev_category'] = y['category'] + + self.mergedb[loginbase][nodename] = {} + self.mergedb[loginbase][nodename].update(y) + self.mergedb[loginbase][nodename]['comonstats'] = x['comonstats'] + self.mergedb[loginbase][nodename]['category'] = x['category'] + self.mergedb[loginbase][nodename]['state'] = x['state'] + self.mergedb[loginbase][nodename]['kernel']=x['kernel'] + self.mergedb[loginbase][nodename]['bootcd']=x['bootcd'] + self.mergedb[loginbase][nodename]['plcnode']=x['plcnode'] + ticket = get_ticket_id(self.mergedb[loginbase][nodename]) + self.mergedb[loginbase][nodename]['rt'] = mailer.getTicketStatus(ticket) + + # delete the entry from cache_all to keep it out of case 3) + del self.cache_all[nodename] + + # 3) nodes that remin in cache_all were not identified by findbad. + # Do we keep them or not? + # NOTE: i think that since the categories are performed before this + # step now, and by a monitor-controlled agent. + + # TODO: This does not work correctly. Do we need this? + #for hn in self.cache_all.keys(): + # y = self.act_all[hn][0] + # if 'monitor' in y['bucket']: + # loginbase = self.plcdb_hn2lb[hn] + # if loginbase not in self.sickdb: + # self.sickdb[loginbase] = {} + # self.sickdb[loginbase][hn] = y + # else: + # del self.cache_all[hn] + + print "len of cache_all: %d" % len(self.cache_all.keys()) + return + + def sendToRT(self): + sorted_sites = self.mergedb.keys() + sorted_sites.sort() + # look at all problems reported by merge + for loginbase in sorted_sites: + d_merge_nodes = self.mergedb[loginbase] + for nodename in d_merge_nodes.keys(): + record = self.mergedb[loginbase][nodename] + self.toRT.put(record) + + # send signal to stop reading + self.toRT.put(None) + return + +class Diagnose(Thread): + def __init__(self, fromRT): + self.fromRT = fromRT + self.plcdb_hn2lb = database.dbLoad("plcdb_hn2lb") + self.findbad = database.if_cached_else(1, "findbad", lambda : {}) + + self.diagnose_in = {} + self.diagnose_out = {} + Thread.__init__(self) + + + def run(self): + self.accumSickSites() + + print "Accumulated %d sick sites" % len(self.diagnose_in.keys()) + logger.debug("Accumulated %d sick sites" % len(self.diagnose_in.keys())) + + try: + stats = self.diagnoseAll() + except Exception, err: + print "----------------" + import traceback + print traceback.print_exc() + from nodecommon import email_exception + email_exception() + print err + #if config.policysavedb: + sys.exit(1) + + print_stats("sites_observed", stats) + print_stats("sites_diagnosed", stats) + print_stats("nodes_diagnosed", stats) + + if config.policysavedb: + print "Saving Databases... diagnose_out" + database.dbDump("diagnose_out", self.diagnose_out) + + def accumSickSites(self): + """ + Take all nodes, from l_diagnose, look them up in the diagnose_out database, + and insert them into diagnose_in[] as: + + diagnose_in[loginbase] = [diag_node1, diag_node2, ...] + """ + while 1: + node_record = self.fromRT.get(block = True) + if node_record == None: + break; + + nodename = node_record['nodename'] + loginbase = self.plcdb_hn2lb[nodename] + + if loginbase not in self.diagnose_in: + self.diagnose_in[loginbase] = {} + + self.diagnose_in[loginbase][nodename] = node_record + + return + + def diagnoseAll(self): + i_sites_observed = 0 + i_sites_diagnosed = 0 + i_nodes_diagnosed = 0 + i_nodes_actedon = 0 + i_sites_emailed = 0 + l_allsites = [] + + sorted_sites = self.diagnose_in.keys() + sorted_sites.sort() + self.diagnose_out= {} + for loginbase in sorted_sites: + l_allsites += [loginbase] + + d_diag_nodes = self.diagnose_in[loginbase] + d_act_records = self.__diagnoseSite(loginbase, d_diag_nodes) + # store records in diagnose_out, for saving later. + self.diagnose_out.update(d_act_records) + + if len(d_act_records[loginbase]['nodes'].keys()) > 0: + i_nodes_diagnosed += (len(d_act_records[loginbase]['nodes'].keys())) + i_sites_diagnosed += 1 + i_sites_observed += 1 + + return {'sites_observed': i_sites_observed, + 'sites_diagnosed': i_sites_diagnosed, + 'nodes_diagnosed': i_nodes_diagnosed, + 'allsites':l_allsites} + + pass - # Princeton Backdoor - if loginbase == "princeton": return - - # Send appropriate message for node if in appropriate bucket. - # If we know where to send a message - if not loginbase: - logger.info("POLICY: loginbase for %s not found" %node) - # And we didn't email already. + def getDaysDown(cls, diag_record): + daysdown = -1 + last_contact = diag_record['plcnode']['last_contact'] + date_created = diag_record['plcnode']['date_created'] + + if diag_record['comonstats']['uptime'] != "null" and diag_record['comonstats']['uptime'] != "-1": + daysdown = - int(float(diag_record['comonstats']['uptime'])) // (60*60*24) + elif last_contact is None: + if date_created is not None: + now = time.time() + diff = now - date_created + daysdown = diff // (60*60*24) + else: + daysdown = -1 else: - # If first email, send to Tech - target = [TECHEMAIL % loginbase] + now = time.time() + diff = now - last_contact + daysdown = diff // (60*60*24) + return daysdown + getDaysDown = classmethod(getDaysDown) + + def getStrDaysDown(cls, diag_record): + daysdown = "unknown" + last_contact = diag_record['plcnode']['last_contact'] + date_created = diag_record['plcnode']['date_created'] + + if diag_record['comonstats']['uptime'] != "null" and \ + diag_record['comonstats']['uptime'] != "-1": + daysdown = int(float(diag_record['comonstats']['uptime'])) // (60*60*24) + daysdown = "%d days up" % daysdown + + elif last_contact is None: + if date_created is not None: + now = time.time() + diff = now - date_created + daysdown = diff // (60*60*24) + daysdown = "Never contacted PLC, created %s days ago" % daysdown + else: + daysdown = "Never contacted PLC" + else: + now = time.time() + diff = now - last_contact + daysdown = diff // (60*60*24) + daysdown = "%s days down" % daysdown + return daysdown + getStrDaysDown = classmethod(getStrDaysDown) + #def getStrDaysDown(cls, diag_record): + # daysdown = cls.getDaysDown(diag_record) + # if daysdown > -1: + # return "%d days down"%daysdown + # elif daysdown == -1: + # return "Has never contacted PLC" + # else: + # return "%d days up"% -daysdown + #getStrDaysDown = classmethod(getStrDaysDown) + + def __getCDVersion(self, diag_record, nodename): + cdversion = "" + #print "Getting kernel for: %s" % diag_record['nodename'] + cdversion = diag_record['kernel'] + return cdversion + + def __diagnoseSite(self, loginbase, d_diag_nodes): + """ + d_diag_nodes are diagnose_in entries. + """ + d_diag_site = {loginbase : { 'config' : + {'squeeze': False, + 'email': False + }, + 'nodes': {} + } + } + sorted_nodes = d_diag_nodes.keys() + sorted_nodes.sort() + for nodename in sorted_nodes: + node_record = d_diag_nodes[nodename] + diag_record = self.__diagnoseNode(loginbase, node_record) + + if diag_record != None: + d_diag_site[loginbase]['nodes'][nodename] = diag_record + + # NOTE: improvement means, we need to act/squeeze and email. + #print "DIAG_RECORD", diag_record + if 'monitor-end-record' in diag_record['stage'] or \ + 'nmreset' in diag_record['stage']: + # print "resetting loginbase!" + d_diag_site[loginbase]['config']['squeeze'] = True + d_diag_site[loginbase]['config']['email'] = True + #else: + # print "NO IMPROVEMENT!!!!" + else: + pass # there is nothing to do for this node. + + # NOTE: these settings can be overridden by command line arguments, + # or the state of a record, i.e. if already in RT's Support Queue. + nodes_up = self.getUpAtSite(loginbase, d_diag_site) + if nodes_up < MINUP: + d_diag_site[loginbase]['config']['squeeze'] = True + + max_slices = self.getMaxSlices(loginbase) + num_nodes = self.getNumNodes(loginbase) + # NOTE: when max_slices == 0, this is either a new site (the old way) + # or an old disabled site from previous monitor (before site['enabled']) + if nodes_up < num_nodes and max_slices != 0: + d_diag_site[loginbase]['config']['email'] = True + + if len(d_diag_site[loginbase]['nodes'].keys()) > 0: + print "SITE: %20s : %d nodes up, at most" % (loginbase, nodes_up) + + return d_diag_site + + def diagRecordByCategory(self, node_record): + nodename = node_record['nodename'] + category = node_record['category'] + state = node_record['state'] + loginbase = self.plcdb_hn2lb[nodename] + diag_record = None + + if "ERROR" in category: # i.e. "DOWN" + diag_record = {} + diag_record.update(node_record) + daysdown = self.getDaysDown(diag_record) + if daysdown < 7: + format = "DIAG: %20s : %-40s Down only %s days NOTHING DONE" + print format % (loginbase, nodename, daysdown) + return None + + s_daysdown = self.getStrDaysDown(diag_record) + diag_record['message'] = emailTxt.mailtxt.newdown + diag_record['args'] = {'nodename': nodename} + diag_record['info'] = (nodename, s_daysdown, "") + + if 'reboot_node_failed' in node_record: + # there was a previous attempt to use the PCU. + if node_record['reboot_node_failed'] == False: + # then the last attempt apparently, succeeded. + # But, the category is still 'ERROR'. Therefore, the + # PCU-to-Node mapping is broken. + #print "Setting message for ERROR node to PCU2NodeMapping: %s" % nodename + diag_record['message'] = emailTxt.mailtxt.pcutonodemapping + diag_record['email_pcu'] = True + + if 'ticket_id' in diag_record: + if diag_record['ticket_id'] == "": + if 'found_rt_ticket' in diag_record: + ticket_id = diag_record['found_rt_ticket'] + else: + ticket_id = "None" + else: + ticket_id = diag_record['ticket_id'] + else: + ticket_id = "None" + + diag_record['log'] = "DOWN: %20s : %-40s == %20s %s" % \ + (loginbase, nodename, diag_record['info'][1:], ticket_id) + + elif "OLDBOOTCD" in category: + # V2 boot cds as determined by findbad + s_daysdown = self.getStrDaysDown(node_record) + s_cdversion = self.__getCDVersion(node_record, nodename) + diag_record = {} + diag_record.update(node_record) + #if "2.4" in diag_record['kernel'] or "v2" in diag_record['bootcd']: + diag_record['message'] = emailTxt.mailtxt.newbootcd + diag_record['args'] = {'nodename': nodename} + diag_record['info'] = (nodename, s_daysdown, s_cdversion) + if diag_record['ticket_id'] == "": + diag_record['log'] = "BTCD: %20s : %-40s == %20s %20s %s" % \ + (loginbase, nodename, diag_record['kernel'], + diag_record['bootcd'], diag_record['found_rt_ticket']) + else: + diag_record['log'] = "BTCD: %20s : %-40s == %20s %20s %s" % \ + (loginbase, nodename, diag_record['kernel'], + diag_record['bootcd'], diag_record['ticket_id']) + + elif "PROD" in category: + if "DEBUG" in state: + # Not sure what to do with these yet. Probably need to + # reboot, and email. + print "DEBG: %20s : %-40s NOTHING DONE" % (loginbase, nodename) + return None + elif "BOOT" in state: + # no action needed. + # TODO: remove penalties, if any are applied. + now = time.time() + last_contact = node_record['plcnode']['last_contact'] + if last_contact == None: + time_diff = 0 + else: + time_diff = now - last_contact; + + if 'improvement' in node_record['stage']: + # then we need to pass this on to 'action' + diag_record = {} + diag_record.update(node_record) + diag_record['message'] = emailTxt.mailtxt.newthankyou + diag_record['args'] = {'nodename': nodename} + diag_record['info'] = (nodename, node_record['prev_category'], + node_record['category']) + if 'email_pcu' in diag_record: + if diag_record['email_pcu']: + # previously, the pcu failed to reboot, so send + # email. Now, reset these values to try the reboot + # again. + diag_record['email_pcu'] = False + del diag_record['reboot_node_failed'] + + if diag_record['ticket_id'] == "": + diag_record['log'] = "IMPR: %20s : %-40s == %20s %20s %s %s" % \ + (loginbase, nodename, diag_record['stage'], + state, category, diag_record['found_rt_ticket']) + else: + diag_record['log'] = "IMPR: %20s : %-40s == %20s %20s %s %s" % \ + (loginbase, nodename, diag_record['stage'], + state, category, diag_record['ticket_id']) + return diag_record + #elif time_diff >= 6*SPERHOUR: + # # heartbeat is older than 30 min. + # # then reset NM. + # #print "Possible NM problem!! %s - %s = %s" % (now, last_contact, time_diff) + # diag_record = {} + # diag_record.update(node_record) + # diag_record['message'] = emailTxt.mailtxt.NMReset + # diag_record['args'] = {'nodename': nodename} + # diag_record['stage'] = "nmreset" + # diag_record['info'] = (nodename, + # node_record['prev_category'], + # node_record['category']) + # if diag_record['ticket_id'] == "": + # diag_record['log'] = "NM : %20s : %-40s == %20s %20s %s %s" % \ + # (loginbase, nodename, diag_record['stage'], + # state, category, diag_record['found_rt_ticket']) + # else: + # diag_record['log'] = "NM : %20s : %-40s == %20s" % \ + # (loginbase, nodename, diag_record['stage']) +# +# return diag_record + else: + return None + else: + # unknown + pass + elif "ALPHA" in category: + pass + elif "clock_drift" in category: + pass + elif "dns" in category: + pass + elif "filerw" in category: + pass + else: + print "Unknown category!!!! %s" % category + sys.exit(1) + + return diag_record + + def __diagnoseNode(self, loginbase, node_record): + # TODO: change the format of the hostname in this + # record to something more natural. + nodename = node_record['nodename'] + category = node_record['category'] + prev_category = node_record['prev_category'] + state = node_record['state'] + #if 'prev_category' in node_record: + # prev_category = node_record['prev_category'] + #else: + # prev_category = "ERROR" + if node_record['prev_category'] != "NORECORD": + + val = cmpCategoryVal(category, prev_category) + print "%s went from %s -> %s" % (nodename, prev_category, category) + if val == 1: + # improved + if node_record['ticket_id'] == "" or node_record['ticket_id'] == None: + print "closing record with no ticket: ", node_record['nodename'] + node_record['action'] = ['close_rt'] + node_record['message'] = None + node_record['stage'] = 'monitor-end-record' + return node_record + else: + node_record['stage'] = 'improvement' + + #if 'monitor-end-record' in node_record['stage']: + # # just ignore it if it's already ended. + # # otherwise, the status should be worse, and we won't get + # # here. + # print "monitor-end-record: ignoring ", node_record['nodename'] + # return None +# +# #return None + elif val == -1: + # current category is worse than previous, carry on + pass + else: + #values are equal, carry on. + #print "why are we here?" + pass + + if 'rt' in node_record and 'Status' in node_record['rt']: + if node_record['stage'] == 'ticket_waitforever': + if 'resolved' in node_record['rt']['Status']: + print "ending waitforever record for: ", node_record['nodename'] + node_record['action'] = ['noop'] + node_record['message'] = None + node_record['stage'] = 'monitor-end-record' + print "oldlog: %s" % node_record['log'], + print "%15s" % node_record['action'] + return node_record + if 'new' in node_record['rt']['Status'] and \ + 'Queue' in node_record['rt'] and \ + 'Monitor' in node_record['rt']['Queue']: + + print "RESETTING stage to findbad" + node_record['stage'] = 'findbad' - # If disk is foobarred, PLC should check it. - if (node in self.cmn.filerw) and \ - (node not in self.emailed.keys()): - self.__actOnFilerw(node) - return - - # If in dbg, set to rins, then reboot. Inform PLC. - if (node in self.cmn.dbg): - # If reboot failure via PCU, POD and send email - # if contacted PCU, return - if self.__actOnDebug(node): return - - if (node in self.emailed.keys()) and \ - (node not in self.cmn.filerw) and \ - (node not in self.cmn.clock_drift): - # If we emailed before, how long ago? - delta = time.time() - self.emailed[node][1] - if delta < SPERDAY: - logger.info("POLICY: already acted on %s today." % node) - return - - logger.info("POLICY: acted %s on %s days ago" % (node, - delta // SPERDAY)) + #### COMPARE category and prev_category + # if not_equal + # then assign a stage based on relative priorities + # else equal + # then check category for stats. + diag_record = self.diagRecordByCategory(node_record) + if diag_record == None: + #print "diag_record == None" + return None + + #### found_RT_ticket + # TODO: need to record time found, and maybe add a stage for acting on it... + # NOTE: after found, if the support ticket is resolved, the block is + # not removed. How to remove the block on this? + if 'found_rt_ticket' in diag_record and \ + diag_record['found_rt_ticket'] is not None: + if diag_record['stage'] is not 'improvement': + diag_record['stage'] = 'ticket_waitforever' + + current_time = time.time() + # take off four days, for the delay that database caused. + # TODO: generalize delays at PLC, and prevent enforcement when there + # have been no emails. + # NOTE: 7*SPERDAY exists to offset the 'bad week' + #delta = current_time - diag_record['time'] - 7*SPERDAY + delta = current_time - diag_record['time'] + + message = diag_record['message'] + act_record = {} + act_record.update(diag_record) + + #### DIAGNOSE STAGES + if 'findbad' in diag_record['stage']: + # The node is bad, and there's no previous record of it. + act_record['email'] = TECH + act_record['action'] = ['noop'] + act_record['message'] = message[0] + act_record['stage'] = 'stage_actinoneweek' + + elif 'nmreset' in diag_record['stage']: + act_record['email'] = ADMIN + act_record['action'] = ['reset_nodemanager'] + act_record['message'] = message[0] + act_record['stage'] = 'nmreset' + return None + + elif 'reboot_node' in diag_record['stage']: + act_record['email'] = TECH + act_record['action'] = ['noop'] + act_record['message'] = message[0] + act_record['stage'] = 'stage_actinoneweek' - # If no luck with tech, email PI - if (delta >= 1): - target.append(PIEMAIL % loginbase) - - # If more than PI thresh, but less than slicethresh - if (delta >= PITHRESH) and (delta < SLICETHRESH): - #remove slice creation if enough nodes arent up - if not self.enoughUp(loginbase): - slices = plc.slices(loginbase) - if len(slices) >= 1: - for slice in slices: - target.append(SLICEMAIL % slice) - logger.info("POLICY: Removing slice creation from %s" % loginbase) - tmp = emailTxt.mailtxt.removedSliceCreation - sbj = tmp[0] - msg = tmp[1] % {'loginbase': loginbase} - plc.removeSliceCreation(node) - mailer.email(sbj, msg, target) - self.squeezed[loginbase] = (time.time(), "creation") - self.emailed[node] = ("creation", time.time()) - return - - # If more than PI thresh and slicethresh - if (delta >= PITHRESH) and (delta > SLICETHRESH): - target.append(PIEMAIL % loginbase) - # Email slices at site. - slices = plc.slices(loginbase) - if len(slices) >= 1: - for slice in slices: - target.append(SLICEMAIL % slice) - # If not enough up, freeze slices and email everyone. - if not self.enoughUp(loginbase): - logger.info("POLICY: Suspending %s slices." % loginbase) - tmp = emailTxt.mailtxt.suspendSlices - sbj = tmp[0] - msg = tmp[1] % {'loginbase': loginbase} - plc.suspendSlices(node) - self.squeezed[loginbase] = (time.time(), "freeze") - mailer.email(sbj, msg, target) - self.emailed[node] = ("freeze", time.time()) - return - - # Find the bucket the node is in and send appropriate email - # to approriate list of people. - for bkt in self.cmn.comonbkts.keys(): - if (node in getattr(self.cmn, bkt)): - # Send predefined message for that bucket. - logger.info("POLICY: Emailing (%s) %s - %s"\ - %(bkt, node, target)) - tmp = getattr(emailTxt.mailtxt, bkt) - sbj = tmp[0] % {'hostname': node} - msg = tmp[1] % {'hostname': node} - mailer.email(sbj, msg, target) - self.emailed[node] = (bkt , time.time()) - return - - - ''' - Prints, logs, and emails status of up nodes, down nodes, and buckets. - ''' - def status(self): - sub = "Monitor Summary" - msg = "\nThe following nodes were acted upon: \n\n" - for (node, (type, date)) in self.emailed.items(): - # Print only things acted on today. - if (time.gmtime(time.time())[2] == time.gmtime(date)[2]): - msg +="%s\t(%s)\t%s\n" %(node, type, time.ctime(date)) - msg +="\n\nThe following sites have been 'squeezed':\n\n" - for (loginbase, (date, type)) in self.squeezed.items(): - # Print only things acted on today. - if (time.gmtime(time.time())[2] == time.gmtime(date)[2]): - msg +="%s\t(%s)\t%s\n" %(loginbase, type, time.ctime(date)) - mailer.email(sub, msg, [SUMTO]) - logger.info(msg) - return - - ''' - Store/Load state of emails. When, where, what. - ''' - def emailedStore(self, action): + elif 'improvement' in diag_record['stage']: + # - backoff previous squeeze actions (slice suspend, nocreate) + # TODO: add a backoff_squeeze section... Needs to runthrough + print "backing off of %s" % nodename + act_record['action'] = ['close_rt'] + act_record['message'] = message[0] + act_record['stage'] = 'monitor-end-record' + + elif 'actinoneweek' in diag_record['stage']: + if delta >= 7 * SPERDAY: + act_record['email'] = TECH | PI + act_record['stage'] = 'stage_actintwoweeks' + act_record['message'] = message[1] + act_record['action'] = ['nocreate' ] + act_record['time'] = current_time # reset clock for waitforever + elif delta >= 3* SPERDAY and not 'second-mail-at-oneweek' in act_record: + act_record['email'] = TECH + act_record['message'] = message[0] + act_record['action'] = ['sendmailagain-waitforoneweekaction' ] + act_record['second-mail-at-oneweek'] = True + else: + act_record['message'] = None + act_record['action'] = ['waitforoneweekaction' ] + print "ignoring this record for: %s" % act_record['nodename'] + return None # don't send if there's no action + + elif 'actintwoweeks' in diag_record['stage']: + if delta >= 7 * SPERDAY: + act_record['email'] = TECH | PI | USER + act_record['stage'] = 'stage_waitforever' + act_record['message'] = message[2] + act_record['action'] = ['suspendslices'] + act_record['time'] = current_time # reset clock for waitforever + elif delta >= 3* SPERDAY and not 'second-mail-at-twoweeks' in act_record: + act_record['email'] = TECH | PI + act_record['message'] = message[1] + act_record['action'] = ['sendmailagain-waitfortwoweeksaction' ] + act_record['second-mail-at-twoweeks'] = True + else: + act_record['message'] = None + act_record['action'] = ['waitfortwoweeksaction'] + return None # don't send if there's no action + + elif 'ticket_waitforever' in diag_record['stage']: + act_record['email'] = TECH + if 'first-found' not in act_record: + act_record['first-found'] = True + act_record['log'] += " firstfound" + act_record['action'] = ['ticket_waitforever'] + act_record['message'] = message[0] + act_record['time'] = current_time + else: + if delta >= 7*SPERDAY: + act_record['action'] = ['ticket_waitforever'] + if 'rt' in act_record and 'Status' in act_record['rt'] and \ + act_record['rt']['Status'] == 'new': + act_record['message'] = message[0] + else: + act_record['message'] = None + + act_record['time'] = current_time # reset clock + else: + act_record['action'] = ['ticket_waitforever'] + act_record['message'] = None + return None + + elif 'waitforever' in diag_record['stage']: + # more than 3 days since last action + # TODO: send only on weekdays. + # NOTE: expects that 'time' has been reset before entering waitforever stage + if delta >= 3*SPERDAY: + act_record['action'] = ['email-againwaitforever'] + act_record['message'] = message[2] + act_record['time'] = current_time # reset clock + else: + act_record['action'] = ['waitforever'] + act_record['message'] = None + return None # don't send if there's no action + + else: + # There is no action to be taken, possibly b/c the stage has + # already been performed, but diagnose picked it up again. + # two cases, + # 1. stage is unknown, or + # 2. delta is not big enough to bump it to the next stage. + # TODO: figure out which. for now assume 2. + print "UNKNOWN stage for %s; nothing done" % nodename + act_record['action'] = ['unknown'] + act_record['message'] = message[0] + + act_record['email'] = TECH + act_record['action'] = ['noop'] + act_record['message'] = message[0] + act_record['stage'] = 'stage_actinoneweek' + act_record['time'] = current_time # reset clock + #print "Exiting..." + #return None + #sys.exit(1) + + print "%s" % act_record['log'], + print "%15s" % act_record['action'] + return act_record + + def getMaxSlices(self, loginbase): + # if sickdb has a loginbase, then it will have at least one node. + site_stats = None + + for nodename in self.diagnose_in[loginbase].keys(): + if nodename in self.findbad['nodes']: + site_stats = self.findbad['nodes'][nodename]['values']['plcsite'] + break + + if site_stats == None: + raise Exception, "loginbase with no nodes in findbad" + else: + return site_stats['max_slices'] + + def getNumNodes(self, loginbase): + # if sickdb has a loginbase, then it will have at least one node. + site_stats = None + + for nodename in self.diagnose_in[loginbase].keys(): + if nodename in self.findbad['nodes']: + site_stats = self.findbad['nodes'][nodename]['values']['plcsite'] + break + + if site_stats == None: + raise Exception, "loginbase with no nodes in findbad" + else: + if 'num_nodes' in site_stats: + return site_stats['num_nodes'] + else: + return 0 + + """ + Returns number of up nodes as the total number *NOT* in act_all with a + stage other than 'steady-state' . + """ + def getUpAtSite(self, loginbase, d_diag_site): + # TODO: THIS DOESN"T WORK!!! it misses all the 'debug' state nodes + # that aren't recorded yet. + + numnodes = self.getNumNodes(loginbase) + # NOTE: assume nodes we have no record of are ok. (too conservative) + # TODO: make the 'up' value more representative + up = numnodes + for nodename in d_diag_site[loginbase]['nodes'].keys(): + + rec = d_diag_site[loginbase]['nodes'][nodename] + if rec['stage'] != 'monitor-end-record': + up -= 1 + else: + pass # the node is assumed to be up. + + #if up != numnodes: + # print "ERROR: %s total nodes up and down != %d" % (loginbase, numnodes) + + return up + + +class SiteAction: + def __init__(self, parameter_names=['hostname', 'ticket_id']): + self.parameter_names = parameter_names + def checkParam(self, args): + for param in self.parameter_names: + if param not in args: + raise Exception("Parameter %s not provided in args"%param) + def run(self, args): + self.checkParam(args) + return self._run(args) + def _run(self, args): + pass + +class SuspendAction(SiteAction): + def _run(self, args): + return plc.suspendSlices(args['hostname']) + +class RemoveSliceCreation(SiteAction): + def _run(self, args): + return plc.removeSliceCreation(args['hostname']) + +class BackoffActions(SiteAction): + def _run(self, args): + plc.enableSlices(args['hostname']) + plc.enableSliceCreation(args['hostname']) + return True + +# TODO: create class for each action below, +# allow for lists of actions to be performed... + + + +def reset_nodemanager(args): + os.system("ssh root@%s /sbin/service nm restart" % nodename) + return + +class Action(Thread): + def __init__(self, l_action): + self.l_action = l_action + + # the hostname to loginbase mapping + self.plcdb_hn2lb = database.dbLoad("plcdb_hn2lb") + + # Actions to take. + self.diagnose_db = database.if_cached_else(1, "diagnose_out", lambda : {}) + # Actions taken. + self.act_all = database.if_cached_else(1, "act_all", lambda : {}) + + # A dict of actions to specific functions. PICKLE doesnt' like lambdas. + self.actions = {} + self.actions['suspendslices'] = lambda args: plc.suspendSlices(args['hostname']) + self.actions['nocreate'] = lambda args: plc.removeSliceCreation(args['hostname']) + self.actions['close_rt'] = lambda args: close_rt_backoff(args) + self.actions['rins'] = lambda args: plc.nodeBootState(args['hostname'], "reinstall") + self.actions['noop'] = lambda args: args + self.actions['reboot_node'] = lambda args: reboot_node(args) + self.actions['reset_nodemanager'] = lambda args: args # reset_nodemanager(args) + + self.actions['ticket_waitforever'] = lambda args: args + self.actions['waitforever'] = lambda args: args + self.actions['unknown'] = lambda args: args + self.actions['waitforoneweekaction'] = lambda args: args + self.actions['waitfortwoweeksaction'] = lambda args: args + self.actions['sendmailagain-waitforoneweekaction'] = lambda args: args + self.actions['sendmailagain-waitfortwoweeksaction'] = lambda args: args + self.actions['email-againwaitforever'] = lambda args: args + self.actions['email-againticket_waitforever'] = lambda args: args + + + self.sickdb = {} + Thread.__init__(self) + + def run(self): + self.accumSites() + print "Accumulated %d sick sites" % len(self.sickdb.keys()) + logger.debug("Accumulated %d sick sites" % len(self.sickdb.keys())) + try: - if action == "LOAD": - f = open(DAT, "r+") - logger.info("POLICY: Found and reading " + DAT) - self.emailed.update(pickle.load(f)) - if action == "WRITE": - f = open(DAT, "w") - #logger.debug("Writing " + DAT) - pickle.dump(self.emailed, f) - f.close() + stats = self.analyseSites() except Exception, err: - logger.info("POLICY: Problem with DAT, %s" %err) - - ''' - Returns True if more than MINUP nodes are up at a site. - ''' - def enoughUp(self, loginbase): - allsitenodes = plc.getSiteNodes(loginbase) - if len(allsitenodes) == 0: - logger.info("Node not in db") - return - - numnodes = len(allsitenodes) - sicknodes = [] - # Get all sick nodes from comon - for bucket in self.cmn.comonbkts.keys(): - for host in getattr(self.cmn, bucket): - sicknodes.append(host) - # Diff. - for node in allsitenodes: - if node in sicknodes: - numnodes -= 1 - - if numnodes < MINUP: - logger.info(\ -"POLICY: site with %s has nodes %s up." %(loginbase, numnodes)) - return False - else: - return True + print "----------------" + import traceback + print traceback.print_exc() + print err + if config.policysavedb: + print "Saving Databases... act_all" + database.dbDump("act_all", self.act_all) + sys.exit(1) + + print_stats("sites_observed", stats) + print_stats("sites_diagnosed", stats) + print_stats("nodes_diagnosed", stats) + print_stats("sites_emailed", stats) + print_stats("nodes_actedon", stats) + print string.join(stats['allsites'], ",") + + if config.policysavedb: + print "Saving Databases... act_all" + #database.dbDump("policy.eventlog", self.eventlog) + # TODO: remove 'diagnose_out', + # or at least the entries that were acted on. + database.dbDump("act_all", self.act_all) + + def accumSites(self): + """ + Take all nodes, from l_action, look them up in the diagnose_db database, + and insert them into sickdb[] as: + + This way only the given l_action nodes will be acted on regardless + of how many from diagnose_db are available. + + sickdb[loginbase][nodename] = diag_record + """ + # TODO: what if l_action == None ? + for nodename in self.l_action: + + loginbase = self.plcdb_hn2lb[nodename] + + if loginbase in self.diagnose_db and \ + nodename in self.diagnose_db[loginbase]['nodes']: + + diag_record = self.diagnose_db[loginbase]['nodes'][nodename] + + if loginbase not in self.sickdb: + self.sickdb[loginbase] = {'nodes' : {}} + + # NOTE: don't copy all node records, since not all will be in l_action + self.sickdb[loginbase]['nodes'][nodename] = diag_record + # NOTE: but, we want to get the loginbase config settings, + # this is the easiest way. + self.sickdb[loginbase]['config'] = self.diagnose_db[loginbase]['config'] + #else: + #print "%s not in diagnose_db!!" % loginbase + return + + def __emailSite(self, loginbase, roles, message, args): + """ + loginbase is the unique site abbreviation, prepended to slice names. + roles contains TECH, PI, USER roles, and derive email aliases. + record contains {'message': [,], 'args': {...}} + """ + ticket_id = 0 + args.update({'loginbase':loginbase}) + + if not config.mail and not config.debug and config.bcc: + roles = ADMIN + if config.mail and config.debug: + roles = ADMIN + + # build targets + contacts = [] + if ADMIN & roles: + contacts += [config.email] + if TECH & roles: + contacts += [TECHEMAIL % loginbase] + if PI & roles: + contacts += [PIEMAIL % loginbase] + if USER & roles: + slices = plc.slices(loginbase) + if len(slices) >= 1: + for slice in slices: + contacts += [SLICEMAIL % slice] + print "SLIC: %20s : %d slices" % (loginbase, len(slices)) + else: + print "SLIC: %20s : 0 slices" % loginbase + + try: + subject = message[0] % args + body = message[1] % args + if ADMIN & roles: + # send only to admin + if 'ticket_id' in args: + subj = "Re: [PL #%s] %s" % (args['ticket_id'], subject) + else: + subj = "Re: [PL noticket] %s" % subject + mailer.email(subj, body, contacts) + ticket_id = args['ticket_id'] + else: + ticket_id = mailer.emailViaRT(subject, body, contacts, args['ticket_id']) + except Exception, err: + print "exception on message:" + import traceback + print traceback.print_exc() + print message + + return ticket_id + + + def _format_diaginfo(self, diag_node): + info = diag_node['info'] + if diag_node['stage'] == 'monitor-end-record': + hlist = " %s went from '%s' to '%s'\n" % (info[0], info[1], info[2]) + else: + hlist = " %s %s - %s\n" % (info[0], info[2], info[1]) #(node,ver,daysdn) + return hlist + + + def get_email_args(self, act_recordlist, loginbase=None): + + email_args = {} + email_args['hostname_list'] = "" + + for act_record in act_recordlist: + email_args['hostname_list'] += act_record['msg_format'] + email_args['hostname'] = act_record['nodename'] + if 'plcnode' in act_record and \ + 'pcu_ids' in act_record['plcnode'] and \ + len(act_record['plcnode']['pcu_ids']) > 0: + print "setting 'pcu_id' for email_args %s"%email_args['hostname'] + email_args['pcu_id'] = act_record['plcnode']['pcu_ids'][0] + else: + email_args['pcu_id'] = "-1" + + if 'ticket_id' in act_record: + if act_record['ticket_id'] == 0 or act_record['ticket_id'] == '0': + print "Enter the ticket_id for %s @ %s" % (loginbase, act_record['nodename']) + sys.stdout.flush() + line = sys.stdin.readline() + try: + ticket_id = int(line) + except: + print "could not get ticket_id from stdin..." + os._exit(1) + else: + ticket_id = act_record['ticket_id'] + + email_args['ticket_id'] = ticket_id + + return email_args + + def get_unique_issues(self, act_recordlist): + # NOTE: only send one email per site, per problem... + unique_issues = {} + for act_record in act_recordlist: + act_key = act_record['action'][0] + if act_key not in unique_issues: + unique_issues[act_key] = [] + + unique_issues[act_key] += [act_record] + return unique_issues + + + def __actOnSite(self, loginbase, site_record): + i_nodes_actedon = 0 + i_nodes_emailed = 0 + + act_recordlist = [] + + for nodename in site_record['nodes'].keys(): + diag_record = site_record['nodes'][nodename] + act_record = self.__actOnNode(diag_record) + #print "nodename: %s %s" % (nodename, act_record) + if act_record is not None: + act_recordlist += [act_record] + + unique_issues = self.get_unique_issues(act_recordlist) + + for issue in unique_issues.keys(): + print "\tworking on issue: %s" % issue + issue_record_list = unique_issues[issue] + email_args = self.get_email_args(issue_record_list, loginbase) + + # for each record. + for act_record in issue_record_list: + # if there's a pcu record and email config is set + if 'email_pcu' in act_record: + if act_record['message'] != None and act_record['email_pcu'] and site_record['config']['email']: + # and 'reboot_node' in act_record['stage']: + + email_args['hostname'] = act_record['nodename'] + ticket_id = self.__emailSite(loginbase, + act_record['email'], + emailTxt.mailtxt.pcudown[0], + email_args) + if ticket_id == 0: + # error. + print "got a ticket_id == 0!!!! %s" % act_record['nodename'] + os._exit(1) + pass + email_args['ticket_id'] = ticket_id + + + act_record = issue_record_list[0] + # send message before squeezing + print "\t\tconfig.email: %s and %s" % (act_record['message'] != None, + site_record['config']['email']) + if act_record['message'] != None and site_record['config']['email']: + ticket_id = self.__emailSite(loginbase, act_record['email'], + act_record['message'], email_args) + + if ticket_id == 0: + # error. + print "ticket_id == 0 for %s %s" % (loginbase, act_record['nodename']) + os._exit(1) + pass + + # Add ticket_id to ALL nodenames + for act_record in issue_record_list: + nodename = act_record['nodename'] + # update node record with RT ticket_id + if nodename in self.act_all: + self.act_all[nodename][0]['ticket_id'] = "%s" % ticket_id + # if the ticket was previously resolved, reset it to new. + if 'rt' in act_record and \ + 'Status' in act_record['rt'] and \ + act_record['rt']['Status'] == 'resolved': + mailer.setTicketStatus(ticket_id, "new") + status = mailer.getTicketStatus(ticket_id) + self.act_all[nodename][0]['rt'] = status + if config.mail: i_nodes_emailed += 1 + + print "\t\tconfig.squeeze: %s and %s" % (config.squeeze, + site_record['config']['squeeze']) + if config.squeeze and site_record['config']['squeeze']: + for act_key in act_record['action']: + self.actions[act_key](email_args) + i_nodes_actedon += 1 + if config.policysavedb: + print "Saving Databases... act_all, diagnose_out" + database.dbDump("act_all", self.act_all) + # remove site record from diagnose_out, it's in act_all as done. + del self.diagnose_db[loginbase] + database.dbDump("diagnose_out", self.diagnose_db) + print "sleeping for 1 sec" + time.sleep(1) + #print "Hit enter to continue..." + #sys.stdout.flush() + #line = sys.stdin.readline() - def run(self): - while 1: - self.actOnSick() - self.emailedStore("WRITE") + return (i_nodes_actedon, i_nodes_emailed) + + def __actOnNode(self, diag_record): + nodename = diag_record['nodename'] + message = diag_record['message'] + + act_record = {} + act_record.update(diag_record) + act_record['nodename'] = nodename + act_record['msg_format'] = self._format_diaginfo(diag_record) + print "act_record['stage'] == %s " % act_record['stage'] + + # avoid end records, and nmreset records + # reboot_node_failed, is set below, so don't reboot repeatedly. + if 'monitor-end-record' not in act_record['stage'] and \ + 'nmreset' not in act_record['stage'] and \ + 'reboot_node_failed' not in act_record: + + if "DOWN" in act_record['log'] and \ + 'pcu_ids' in act_record['plcnode'] and \ + len(act_record['plcnode']['pcu_ids']) > 0: + + print "%s" % act_record['log'], + print "%15s" % (['reboot_node'],) + # Set node to re-install + plc.nodeBootState(act_record['nodename'], "reinstall") + try: + ret = reboot_node({'hostname': act_record['nodename']}) + except Exception, exc: + print "exception on reboot_node:" + import traceback + print traceback.print_exc() + ret = False + + if ret: # and ( 'reboot_node_failed' not in act_record or act_record['reboot_node_failed'] == False): + # Reboot Succeeded + print "reboot succeeded for %s" % act_record['nodename'] + act_record2 = {} + act_record2.update(act_record) + act_record2['action'] = ['reboot_node'] + act_record2['stage'] = "reboot_node" + act_record2['reboot_node_failed'] = False + act_record2['email_pcu'] = False + + if nodename not in self.act_all: + self.act_all[nodename] = [] + print "inserting 'reboot_node' record into act_all" + self.act_all[nodename].insert(0,act_record2) + + # return None to avoid further action + print "Taking no further action" + return None + else: + print "reboot failed for %s" % act_record['nodename'] + # set email_pcu to also send pcu notice for this record. + act_record['reboot_node_failed'] = True + act_record['email_pcu'] = True + + print "%s" % act_record['log'], + print "%15s" % act_record['action'] + + if act_record['stage'] is not 'monitor-end-record' and \ + act_record['stage'] is not 'nmreset': + if nodename not in self.act_all: + self.act_all[nodename] = [] + + self.act_all[nodename].insert(0,act_record) + else: + print "Not recording %s in act_all" % nodename + + return act_record + + def analyseSites(self): + i_sites_observed = 0 + i_sites_diagnosed = 0 + i_nodes_diagnosed = 0 + i_nodes_actedon = 0 + i_sites_emailed = 0 + l_allsites = [] + + sorted_sites = self.sickdb.keys() + sorted_sites.sort() + for loginbase in sorted_sites: + site_record = self.sickdb[loginbase] + print "sites: %s" % loginbase + + i_nodes_diagnosed += len(site_record.keys()) + i_sites_diagnosed += 1 + + (na,ne) = self.__actOnSite(loginbase, site_record) + + i_sites_observed += 1 + i_nodes_actedon += na + i_sites_emailed += ne + + l_allsites += [loginbase] + + return {'sites_observed': i_sites_observed, + 'sites_diagnosed': i_sites_diagnosed, + 'nodes_diagnosed': i_nodes_diagnosed, + 'sites_emailed': i_sites_emailed, + 'nodes_actedon': i_nodes_actedon, + 'allsites':l_allsites} + + def print_stats(self, key, stats): + print "%20s : %d" % (key, stats[key]) + + + + #""" + #Prints, logs, and emails status of up nodes, down nodes, and buckets. + #""" + #def status(self): + # sub = "Monitor Summary" + # msg = "\nThe following nodes were acted upon: \n\n" + # for (node, (type, date)) in self.emailed.items(): + # # Print only things acted on today. + # if (time.gmtime(time.time())[2] == time.gmtime(date)[2]): + # msg +="%s\t(%s)\t%s\n" %(node, type, time.ctime(date)) + # msg +="\n\nThe following sites have been 'squeezed':\n\n" + # for (loginbase, (date, type)) in self.squeezed.items(): + # # Print only things acted on today. + # if (time.gmtime(time.time())[2] == time.gmtime(date)[2]): + # msg +="%s\t(%s)\t%s\n" %(loginbase, type, time.ctime(date)) + # mailer.email(sub, msg, [SUMTO]) + # logger.info(msg) + # return + + #""" + #Store/Load state of emails. When, where, what. + #""" + #def emailedStore(self, action): + # try: + # if action == "LOAD": + # f = open(DAT, "r+") + # logger.info("POLICY: Found and reading " + DAT) + # self.emailed.update(pickle.load(f)) + # if action == "WRITE": + # f = open(DAT, "w") + # #logger.debug("Writing " + DAT) + # pickle.dump(self.emailed, f) + # f.close() + # except Exception, err: + # logger.info("POLICY: Problem with DAT, %s" %err) + + +#class Policy(Thread): def main(): - logger.setLevel(logging.DEBUG) - ch = logging.StreamHandler() - ch.setLevel(logging.DEBUG) - formatter = logging.Formatter('%(message)s') - ch.setFormatter(formatter) - logger.addHandler(ch) - - #print NodesDebug() - #tmp = Queue.Queue() - #a = Policy(None, tmp) - #a.emailedStore("LOAD") - #print a.emailed - - print plc.slices(plc.siteId("alice.cs.princeton.edu")) - os._exit(0) + print "policy.py is a module, not a script for running directly." + if __name__ == '__main__': import os import plc