X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=sfa%2Fsenslab%2Fslabdriver.py;h=e7c95d7746f7eec207245e0598f953dcf12f6936;hb=9c93ec5573a58c63a6f572434913562b35e61f90;hp=c11e1c6d6b4d710a844cc6a4d6af0bba5d51bc2e;hpb=b3b654d4ab6b3615470b19cf9511d3bfd03cd50a;p=sfa.git diff --git a/sfa/senslab/slabdriver.py b/sfa/senslab/slabdriver.py index c11e1c6d..e7c95d77 100644 --- a/sfa/senslab/slabdriver.py +++ b/sfa/senslab/slabdriver.py @@ -1,50 +1,40 @@ -import sys import subprocess -import datetime -from time import gmtime, strftime -from sfa.util.faults import MissingSfaInfo , SliverDoesNotExist +from datetime import datetime +from dateutil import tz +from time import strftime, gmtime + +from sfa.util.faults import SliverDoesNotExist, UnknownSfaType from sfa.util.sfalogging import logger -from sfa.util.defaultdict import defaultdict -from sfa.storage.record import Record from sfa.storage.alchemy import dbsession -from sfa.storage.model import RegRecord +from sfa.storage.model import RegRecord, RegUser, RegSlice +from sfa.trust.credential import Credential -from sfa.trust.certificate import * -from sfa.trust.credential import * -from sfa.trust.gid import GID from sfa.managers.driver import Driver from sfa.rspecs.version_manager import VersionManager from sfa.rspecs.rspec import RSpec -from sfa.util.xrn import hrn_to_urn, urn_to_sliver_id -from sfa.util.plxrn import slicename_to_hrn, hostname_to_hrn, hrn_to_pl_slicename +from sfa.util.xrn import hrn_to_urn, urn_to_sliver_id, get_leaf + ## thierry: everything that is API-related (i.e. handling incoming requests) # is taken care of # SlabDriver should be really only about talking to the senslab testbed -## thierry : please avoid wildcard imports :) + from sfa.senslab.OARrestapi import OARrestapi from sfa.senslab.LDAPapi import LDAPapi -from sfa.senslab.parsing import parse_filter -from sfa.senslab.slabpostgres import SlabDB, slab_dbsession,SliceSenslab -from sfa.senslab.slabaggregate import SlabAggregate +from sfa.senslab.slabpostgres import SlabDB, slab_dbsession, SliceSenslab, JobSenslab +from sfa.senslab.slabaggregate import SlabAggregate, slab_xrn_to_hostname, slab_xrn_object from sfa.senslab.slabslices import SlabSlices -def list_to_dict(recs, key): - """ - convert a list of dictionaries into a dictionary keyed on the - specified dictionary key - """ - # print>>sys.stderr, " \r\n \t\t 1list_to_dict : rec %s \r\n \t\t list_to_dict key %s" %(recs,key) - keys = [rec[key] for rec in recs] - #print>>sys.stderr, " \r\n \t\t list_to_dict : rec %s \r\n \t\t list_to_dict keys %s" %(recs,keys) - return dict(zip(keys, recs)) + + + # thierry : note # this inheritance scheme is so that the driver object can receive @@ -54,68 +44,81 @@ class SlabDriver(Driver): def __init__(self, config): Driver.__init__ (self, config) - self.config=config + self.config = config self.hrn = config.SFA_INTERFACE_HRN - - self.root_auth = config.SFA_REGISTRY_ROOT_AUTH - - - print >>sys.stderr, "\r\n_____________ SFA SENSLAB DRIVER \r\n" - # thierry - just to not break the rest of this code + self.root_auth = config.SFA_REGISTRY_ROOT_AUTH - #self.oar = OARapi() self.oar = OARrestapi() - self.ldap = LDAPapi() - #self.users = SenslabImportUsers() + self.ldap = LDAPapi() self.time_format = "%Y-%m-%d %H:%M:%S" - self.db = SlabDB(config) - #self.logger=sfa_logger() - self.cache=None + self.db = SlabDB(config,debug = True) + self.cache = None - - def sliver_status(self,slice_urn,slice_hrn): - # receive a status request for slice named urn/hrn urn:publicid:IDN+senslab+nturro_slice hrn senslab.nturro_slice - # shall return a structure as described in - # http://groups.geni.net/geni/wiki/GAPI_AM_API_V2#SliverStatus - # NT : not sure if we should implement this or not, but used by sface. + + def sliver_status(self, slice_urn, slice_hrn): + """Receive a status request for slice named urn/hrn + urn:publicid:IDN+senslab+nturro_slice hrn senslab.nturro_slice + shall return a structure as described in + http://groups.geni.net/geni/wiki/GAPI_AM_API_V2#SliverStatus + NT : not sure if we should implement this or not, but used by sface. - - sl = self.GetSlices(slice_filter= slice_hrn, filter_type = 'slice_hrn') - if len(slices) is 0: + """ + + #First get the slice with the slice hrn + sl = self.GetSlices(slice_filter = slice_hrn, \ + slice_filter_type = 'slice_hrn') + if len(sl) is 0: raise SliverDoesNotExist("%s slice_hrn" % (slice_hrn)) - #sl = slices[0] - print >>sys.stderr, "\r\n \r\n_____________ Sliver status urn %s hrn %s slices %s \r\n " %(slice_urn,slice_hrn,slices) + + top_level_status = 'unknown' + nodes_in_slice = sl['node_ids'] + recuser = dbsession.query(RegRecord).filter_by(record_id = \ + sl['record_id_user']).first() + sl.update({'user':recuser.hrn}) + if len(nodes_in_slice) is 0: + raise SliverDoesNotExist("No slivers allocated ") + else: + top_level_status = 'ready' + + logger.debug("Slabdriver - sliver_status Sliver status urn %s hrn %s sl\ + %s \r\n " %(slice_urn, slice_hrn, sl)) + if sl['oar_job_id'] is not -1: - - # report about the local nodes only - nodes = self.GetNodes({'hostname':sl['node_ids']}, - ['node_id', 'hostname','site_login_base','boot_state']) - if len(nodes) is 0: - raise SliverDoesNotExist("No slivers allocated ") - - - site_logins = [node['site_login_base'] for node in nodes] - + #A job is running on Senslab for this slice + # report about the local nodes that are in the slice only + + nodes_all = self.GetNodes({'hostname':nodes_in_slice}, + ['node_id', 'hostname','site','boot_state']) + nodeall_byhostname = dict([(n['hostname'], n) for n in nodes_all]) + + result = {} - top_level_status = 'unknown' - if nodes: - top_level_status = 'ready' result['geni_urn'] = slice_urn - result['slab_login'] = sl['job_user'] + result['pl_login'] = sl['user'] #For compatibility + - timestamp = float(sl['startTime']) + float(sl['walltime']) - result['slab_expires'] = strftime(self.time_format, gmtime(float(timestamp))) + timestamp = float(sl['startTime']) + float(sl['walltime']) + result['pl_expires'] = strftime(self.time_format, \ + gmtime(float(timestamp))) + #result['slab_expires'] = strftime(self.time_format,\ + #gmtime(float(timestamp))) resources = [] - for node in nodes: + for node in nodeall_byhostname: res = {} - res['slab_hostname'] = node['hostname'] - res['slab_boot_state'] = node['boot_state'] + #res['slab_hostname'] = node['hostname'] + #res['slab_boot_state'] = node['boot_state'] - sliver_id = urn_to_sliver_id(slice_urn, sl['record_id_slice'], node['node_id']) - res['geni_urn'] = sliver_id - if node['boot_state'] == 'Alive': + res['pl_hostname'] = nodeall_byhostname[node]['hostname'] + res['pl_boot_state'] = nodeall_byhostname[node]['boot_state'] + res['pl_last_contact'] = strftime(self.time_format, \ + gmtime(float(timestamp))) + sliver_id = urn_to_sliver_id(slice_urn, sl['record_id_slice'], \ + nodeall_byhostname[node]['node_id']) + res['geni_urn'] = sliver_id + if nodeall_byhostname[node]['boot_state'] == 'Alive': + res['geni_status'] = 'ready' else: res['geni_status'] = 'failed' @@ -127,106 +130,247 @@ class SlabDriver(Driver): result['geni_status'] = top_level_status result['geni_resources'] = resources - print >>sys.stderr, "\r\n \r\n_____________ Sliver status resources %s res %s \r\n " %(resources,res) + logger.debug("SLABDRIVER \tsliver_statusresources %s res %s "\ + %(resources,res)) return result - def create_sliver (self, slice_urn, slice_hrn, creds, rspec_string, users, options): - aggregate = SlabAggregate(self) + def synchronize_oar_and_slice_table(self, slice_hrn = None): + #Get list of leases + oar_leases_list = self.GetReservedNodes() + + logger.debug("SLABDRIVER \tsynchronize_oar_and_slice_table \r\n \r\n : oar_leases_list %s\r\n" %( oar_leases_list)) + #Get list of slices/leases . multiple entry per user depending on number of jobs + #At this point we don't have the slice_hrn so that's why + #we are calling Getslices, which holds a field with slice_hrn + + if slice_hrn : + sfa_slices_list = self.GetSlices(slice_filter = slice_hrn, slice_filter_type = 'slice_hrn') + self.synchronize_oar_and_slice_table_for_slice_hrn(slice_hrn, oar_leases_list, sfa_slices_list) + else : + sfa_slices_list = self.GetSlices() + + sfa_slices_dict_by_slice_hrn = {} + for sfa_slice in sfa_slices_list: + if sfa_slice['slice_hrn'] not in sfa_slices_dict_by_slice_hrn: + sfa_slices_dict_by_slice_hrn[sfa_slice['slice_hrn']] = [] + sfa_slices_dict_by_slice_hrn[sfa_slice['slice_hrn']].append(sfa_slice) + else : + sfa_slices_dict_by_slice_hrn[sfa_slice['slice_hrn']].append(sfa_slice) + + for slice_hrn in sfa_slices_dict_by_slice_hrn: + list_slices_sfa = sfa_slices_dict_by_slice_hrn[slice_hrn] + if slice_hrn =='senslab2.avakian_slice': + logger.debug("SLABDRIVER \tsynchronize_oar_and_slice_table slice_hrn %s list_slices_sfa %s\r\n \r\n" %( slice_hrn,list_slices_sfa)) + self.synchronize_oar_and_slice_table_for_slice_hrn(slice_hrn, oar_leases_list, list_slices_sfa) + + return + + + def synchronize_oar_and_slice_table_for_slice_hrn(self,slice_hrn, oar_leases_list, sfa_slices_list): + + #Get list of slices/leases . multiple entry per user depending on number of jobs + #sfa_slices_list = self.GetSlices(slice_filter = slice_hrn, slice_filter_type = 'slice_hrn') + sfa_slices_dict = {} + oar_leases_dict = {} + login = slice_hrn.split(".")[1].split("_")[0] + + #Create dictionnaries based on the tuple user login/ job id + #for the leases list and the slices list + + for sl in sfa_slices_list: + if sl['oar_job_id'] != [] : + for oar_jobid in sl['oar_job_id']: + if (login, oar_jobid) not in sfa_slices_dict: + sfa_slices_dict[(login,oar_jobid)] = sl + + for lease in oar_leases_list: + if (lease['user'], lease['lease_id']) not in oar_leases_dict: + oar_leases_dict[(lease['user'], lease['lease_id'])] = lease + + #Find missing entries in the sfa slices list dict by comparing + #the keys in both dictionnaries + #Add the missing entries in the slice sneslab table + for lease in oar_leases_dict : + logger.debug(" =============SLABDRIVER \t\t\ synchronize_oar_and_slice_table_for_slice_hrn oar_leases_list %s \r\n \t\t\t SFA_SLICES_DICT %s \r\n \r\n LOGIN %s \r\n " %( oar_leases_list,sfa_slices_dict,login)) + if lease not in sfa_slices_dict and login == lease[0]: + + #if lease in GetReservedNodes not in GetSlices update the db + #First get the list of nodes hostnames for this job + oar_reserved_nodes_listdict = oar_leases_dict[lease]['reserved_nodes'] + oar_reserved_nodes_list = [] + for node_dict in oar_reserved_nodes_listdict: + oar_reserved_nodes_list.append(node_dict['hostname']) + #And update the db with slice hrn, job id and node list + self.db.add_job(slice_hrn, lease[1], oar_reserved_nodes_list) + + for lease in sfa_slices_dict: + #Job is now terminated or in Error, either way ot is not going to run again + #Remove it from the db + if lease not in oar_leases_dict: + self.db.delete_job( slice_hrn, lease[1]) + + return + + def create_sliver (self, slice_urn, slice_hrn, creds, rspec_string, \ + users, options): + aggregate = SlabAggregate(self) + slices = SlabSlices(self) peer = slices.get_peer(slice_hrn) sfa_peer = slices.get_sfa_peer(slice_hrn) - slice_record=None - - + slice_record = None + if not isinstance(creds, list): creds = [creds] - - + if users: slice_record = users[0].get('slice_record', {}) # parse rspec rspec = RSpec(rspec_string) - requested_attributes = rspec.version.get_slice_attributes() + logger.debug("SLABDRIVER.PY \tcreate_sliver \trspec.version %s " \ + %(rspec.version)) - # ensure site record exists - #site = slices.verify_site(slice_hrn, slice_record, peer, sfa_peer, options=options) + self.synchronize_oar_and_slice_table(slice_hrn) + # ensure site record exists? # ensure slice record exists - slice = slices.verify_slice(slice_hrn, slice_record, peer, sfa_peer, options=options) + sfa_slice = slices.verify_slice(slice_hrn, slice_record, peer, \ + sfa_peer, options=options) + requested_attributes = rspec.version.get_slice_attributes() + + logger.debug("SLABDRIVER.PY create_sliver slice %s " %(sfa_slice)) + # ensure person records exists - persons = slices.verify_persons(slice_hrn, slice, users, peer, sfa_peer, options=options) - # ensure slice attributes exists - #slices.verify_slice_attributes(slice, requested_attributes, options=options) + persons = slices.verify_persons(slice_hrn, sfa_slice, users, peer, \ + sfa_peer, options=options) - # add/remove slice from nodes - requested_slivers = [node.get('component_name') for node in rspec.version.get_nodes_with_slivers()] - nodes = slices.verify_slice_nodes(slice, requested_slivers, peer) - - - - # handle MyPLC peer association. - # only used by plc and ple. - #slices.handle_peer(site, slice, persons, peer) + + + # add/remove slice from nodes + + requested_slivers = [node.get('component_name') \ + for node in rspec.version.get_nodes_with_slivers()] + l = [ node for node in rspec.version.get_nodes_with_slivers() ] + logger.debug("SLADRIVER \tcreate_sliver requested_slivers \ + requested_slivers %s listnodes %s" %(requested_slivers,l)) + + nodes = slices.verify_slice_nodes(sfa_slice, requested_slivers, peer) + + # add/remove leases + requested_lease_list = [] + kept_leases = [] + for lease in rspec.version.get_leases(): + single_requested_lease = {} + logger.debug("SLABDRIVER.PY \tcreate_sliver lease %s " %(lease)) + if not lease.get('lease_id'): + single_requested_lease['hostname'] = \ + slab_xrn_to_hostname(lease.get('component_id').strip()) + single_requested_lease['start_time'] = lease.get('start_time') + single_requested_lease['duration'] = lease.get('duration') + else: + kept_leases.append(int(lease['lease_id'])) + if single_requested_lease.get('hostname'): + requested_lease_list.append(single_requested_lease) + + #dCreate dict of leases by start_time, regrouping nodes reserved at the same + #time, for the same amount of time = one job on OAR + requested_job_dict = {} + for lease in requested_lease_list: + + #In case it is an asap experiment start_time is empty + if lease['start_time'] == '': + lease['start_time'] = '0' + + if lease['start_time'] not in requested_job_dict: + if isinstance(lease['hostname'], str): + lease['hostname'] = [lease['hostname']] + + requested_job_dict[lease['start_time']] = lease + + else : + job_lease = requested_job_dict[lease['start_time']] + if lease['duration'] == job_lease['duration'] : + job_lease['hostname'].append(lease['hostname']) + + + + + logger.debug("SLABDRIVER.PY \tcreate_sliver requested_job_dict %s " %(requested_job_dict)) + + leases = slices.verify_slice_leases(sfa_slice, \ + requested_job_dict, kept_leases, peer) return aggregate.get_rspec(slice_xrn=slice_urn, version=rspec.version) def delete_sliver (self, slice_urn, slice_hrn, creds, options): - slices = self.GetSlices(slice_filter= slice_hrn, filter_type = 'slice_hrn') - if not slices: + sfa_slice = self.GetSlices(slice_filter = slice_hrn, \ + slice_filter_type = 'slice_hrn') + logger.debug("SLABDRIVER.PY delete_sliver slice %s" %(sfa_slice)) + if not sfa_slice: return 1 - slice = slices[0] - + + slices = SlabSlices(self) # determine if this is a peer slice - # xxx I wonder if this would not need to use PlSlices.get_peer instead - # in which case plc.peers could be deprecated as this here - # is the only/last call to this last method in plc.peers - peer = peers.get_peer(self, slice_hrn) + + peer = slices.get_peer(slice_hrn) + logger.debug("SLABDRIVER.PY delete_sliver peer %s" %(peer)) try: if peer: - self.UnBindObjectFromPeer('slice', slice['slice_id'], peer) - self.DeleteSliceFromNodes(slice_hrn, slice['node_ids']) + self.UnBindObjectFromPeer('slice', \ + sfa_slice['record_id_slice'], peer) + self.DeleteSliceFromNodes(sfa_slice) finally: if peer: - self.BindObjectToPeer('slice', slice['slice_id'], peer, slice['peer_slice_id']) + self.BindObjectToPeer('slice', sfa_slice['record_id_slice'], \ + peer, sfa_slice['peer_slice_id']) return 1 - - + def AddSlice(self, slice_record): + slab_slice = SliceSenslab( slice_hrn = slice_record['slice_hrn'], \ + record_id_slice= slice_record['record_id_slice'] , \ + record_id_user= slice_record['record_id_user'], \ + peer_authority = slice_record['peer_authority']) + logger.debug("SLABDRIVER.PY \tAddSlice slice_record %s slab_slice %s" \ + %(slice_record,slab_slice)) + slab_dbsession.add(slab_slice) + slab_dbsession.commit() + return + # first 2 args are None in case of resource discovery def list_resources (self, slice_urn, slice_hrn, creds, options): #cached_requested = options.get('cached', True) version_manager = VersionManager() # get the rspec's return format from options - rspec_version = version_manager.get_version(options.get('geni_rspec_version')) + rspec_version = \ + version_manager.get_version(options.get('geni_rspec_version')) version_string = "rspec_%s" % (rspec_version) #panos adding the info option to the caching key (can be improved) if options.get('info'): - version_string = version_string + "_"+options.get('info', 'default') + version_string = version_string + "_" + \ + options.get('info', 'default') # look in cache first #if cached_requested and self.cache and not slice_hrn: #rspec = self.cache.get(version_string) #if rspec: - #logger.debug("SlabDriver.ListResources: returning cached advertisement") + #logger.debug("SlabDriver.ListResources: \ + #returning cached advertisement") #return rspec #panos: passing user-defined options - #print "manager options = ",options aggregate = SlabAggregate(self) origin_hrn = Credential(string=creds[0]).get_gid_caller().get_hrn() - print>>sys.stderr, " \r\n \r\n \t SLABDRIVER get_rspec origin_hrn %s" %(origin_hrn) options.update({'origin_hrn':origin_hrn}) - print>>sys.stderr, " \r\n \r\n \t SLABDRIVER get_rspec options %s" %(options) - rspec = aggregate.get_rspec(slice_xrn=slice_urn, version=rspec_version, - options=options) - + rspec = aggregate.get_rspec(slice_xrn=slice_urn, \ + version=rspec_version, options=options) + # cache the result #if self.cache and not slice_hrn: #logger.debug("Slab.ListResources: stores advertisement in cache") @@ -244,11 +388,15 @@ class SlabDriver(Driver): #return slices # get data from db - print>>sys.stderr, " \r\n \t\t SLABDRIVER.PY list_slices" - slices = self.GetSlices() - slice_hrns = [slicename_to_hrn(self.hrn, slice['slice_hrn']) for slice in slices] - slice_urns = [hrn_to_urn(slice_hrn, 'slice') for slice_hrn in slice_hrns] - + + slices = self.GetSlices() + logger.debug("SLABDRIVER.PY \tlist_slices hrn %s \r\n \r\n" %(slices)) + slice_hrns = [slab_slice['slice_hrn'] for slab_slice in slices] + #slice_hrns = [slicename_to_hrn(self.hrn, slab_slice['slice_hrn']) \ + #for slab_slice in slices] + slice_urns = [hrn_to_urn(slice_hrn, 'slice') \ + for slice_hrn in slice_hrns] + # cache the result #if self.cache: #logger.debug ("SlabDriver.list_slices stores value in cache") @@ -258,30 +406,27 @@ class SlabDriver(Driver): #No site or node register supported def register (self, sfa_record, hrn, pub_key): - type = sfa_record['type'] - slab_record = self.sfa_fields_to_slab_fields(type, hrn, sfa_record) - - #if type == 'authority': - #sites = self.shell.GetSites([slab_record['login_base']]) - #if not sites: - #pointer = self.shell.AddSite(slab_record) - #else: - #pointer = sites[0]['site_id'] - - if type == 'slice': - acceptable_fields=['url', 'instantiation', 'name', 'description'] + record_type = sfa_record['type'] + slab_record = self.sfa_fields_to_slab_fields(record_type, hrn, \ + sfa_record) + + + if record_type == 'slice': + acceptable_fields = ['url', 'instantiation', 'name', 'description'] for key in slab_record.keys(): if key not in acceptable_fields: slab_record.pop(key) - print>>sys.stderr, " \r\n \t\t SLABDRIVER.PY register" - slices = self.GetSlices(slice_filter =slab_record['hrn'], filter_type = 'slice_hrn') + logger.debug("SLABDRIVER.PY register") + slices = self.GetSlices(slice_filter =slab_record['hrn'], \ + slice_filter_type = 'slice_hrn') if not slices: - pointer = self.AddSlice(slab_record) + pointer = self.AddSlice(slab_record) else: - pointer = slices[0]['slice_id'] + pointer = slices[0]['slice_id'] - elif type == 'user': - persons = self.GetPersons([sfa_record['hrn']]) + elif record_type == 'user': + persons = self.GetPersons([sfa_record]) + #persons = self.GetPersons([sfa_record['hrn']]) if not persons: pointer = self.AddPerson(dict(sfa_record)) #add in LDAP @@ -290,50 +435,56 @@ class SlabDriver(Driver): #Does this make sense to senslab ? #if 'enabled' in sfa_record and sfa_record['enabled']: - #self.UpdatePerson(pointer, {'enabled': sfa_record['enabled']}) + #self.UpdatePerson(pointer, \ + #{'enabled': sfa_record['enabled']}) - # add this person to the site only if she is being added for the first - # time by sfa and doesont already exist in plc + #TODO register Change this AddPersonToSite stuff 05/07/2012 SA + # add this person to the site only if + # she is being added for the first + # time by sfa and doesnt already exist in plc if not persons or not persons[0]['site_ids']: login_base = get_leaf(sfa_record['authority']) self.AddPersonToSite(pointer, login_base) # What roles should this user have? + #TODO : DElete this AddRoleToPerson 04/07/2012 SA + #Function prototype is : + #AddRoleToPerson(self, auth, role_id_or_name, person_id_or_email) + #what's the pointer doing here? self.AddRoleToPerson('user', pointer) # Add the user's key if pub_key: - self.AddPersonKey(pointer, {'key_type' : 'ssh', 'key' : pub_key}) + self.AddPersonKey(pointer, {'key_type' : 'ssh', \ + 'key' : pub_key}) #No node adding outside OAR - #elif type == 'node': - #login_base = hrn_to_slab_login_base(sfa_record['authority']) - #nodes = self.GetNodes([slab_record['hostname']]) - #if not nodes: - #pointer = self.AddNode(login_base, slab_record) - #else: - #pointer = nodes[0]['node_id'] - + return pointer #No site or node record update allowed def update (self, old_sfa_record, new_sfa_record, hrn, new_key): pointer = old_sfa_record['pointer'] - type = old_sfa_record['type'] + old_sfa_record_type = old_sfa_record['type'] # new_key implemented for users only - if new_key and type not in [ 'user' ]: - raise UnknownSfaType(type) + if new_key and old_sfa_record_type not in [ 'user' ]: + raise UnknownSfaType(old_sfa_record_type) #if (type == "authority"): #self.shell.UpdateSite(pointer, new_sfa_record) - if type == "slice": - slab_record=self.sfa_fields_to_slab_fields(type, hrn, new_sfa_record) + if old_sfa_record_type == "slice": + slab_record = self.sfa_fields_to_slab_fields(old_sfa_record_type, \ + hrn, new_sfa_record) if 'name' in slab_record: slab_record.pop('name') + #Prototype should be UpdateSlice(self, + #auth, slice_id_or_name, slice_fields) + #Senslab cannot update slice since slice = job + #so we must delete and create another job self.UpdateSlice(pointer, slab_record) - elif type == "user": + elif old_sfa_record_type == "user": update_fields = {} all_fields = new_sfa_record for key in all_fields.keys(): @@ -358,29 +509,31 @@ class SlabDriver(Driver): else: key_exists = True if not key_exists: - self.AddPersonKey(pointer, {'key_type': 'ssh', 'key': new_key}) - - #elif type == "node": - #self.UpdateNode(pointer, new_sfa_record) + self.AddPersonKey(pointer, {'key_type': 'ssh', \ + 'key': new_key}) + return True def remove (self, sfa_record): - type=sfa_record['type'] - hrn=sfa_record['hrn'] - record_id= sfa_record['record_id'] - if type == 'user': - username = hrn.split(".")[len(hrn.split(".")) -1] - #get user in ldap - persons = self.GetPersons(username) - # only delete this person if he has site ids. if he doesnt, it probably means - # he was just removed from a site, not actually deleted - if persons and persons[0]['site_ids']: - self.DeletePerson(username) - elif type == 'slice': - if self.GetSlices(slice_filter = hrn, filter_type = 'slice_hrn'): - self.DeleteSlice(hrn) + sfa_record_type = sfa_record['type'] + hrn = sfa_record['hrn'] + record_id = sfa_record['record_id'] + if sfa_record_type == 'user': + + #get user from senslab ldap + person = self.GetPersons(sfa_record) + #No registering at a given site in Senslab. + #Once registered to the LDAP, all senslab sites are + #accesible. + if person : + #Mark account as disabled in ldap + self.DeletePerson(sfa_record) + elif sfa_record_type == 'slice': + if self.GetSlices(slice_filter = hrn, \ + slice_filter_type = 'slice_hrn'): + self.DeleteSlice(sfa_record_type) #elif type == 'authority': #if self.GetSites(pointer): @@ -388,202 +541,423 @@ class SlabDriver(Driver): return True - def GetPeers (self,auth = None, peer_filter=None, return_fields=None): + + + #TODO clean GetPeers. 05/07/12SA + def GetPeers (self, auth = None, peer_filter=None, return_fields_list=None): existing_records = {} - existing_hrns_by_types= {} - print >>sys.stderr, "\r\n \r\n SLABDRIVER GetPeers auth = %s, peer_filter %s, return_field %s " %(auth , peer_filter, return_fields) + existing_hrns_by_types = {} + logger.debug("SLABDRIVER \tGetPeers auth = %s, peer_filter %s, \ + return_field %s " %(auth , peer_filter, return_fields_list)) all_records = dbsession.query(RegRecord).filter(RegRecord.type.like('%authority%')).all() for record in all_records: - existing_records[record.hrn] = record + existing_records[(record.hrn, record.type)] = record if record.type not in existing_hrns_by_types: existing_hrns_by_types[record.type] = [record.hrn] - print >>sys.stderr, "\r\n \r\n SLABDRIVER GetPeers \t NOT IN existing_hrns_by_types %s " %( existing_hrns_by_types) + logger.debug("SLABDRIVER \tGetPeer\t NOT IN \ + existing_hrns_by_types %s " %( existing_hrns_by_types)) else: - print >>sys.stderr, "\r\n \r\n SLABDRIVER GetPeers \t INNN type %s hrn %s " %( record.type,record.hrn ) - existing_hrns_by_types.update({record.type:(existing_hrns_by_types[record.type].append(record.hrn))}) + logger.debug("SLABDRIVER \tGetPeer\t \INNN type %s hrn %s " \ + %(record.type,record.hrn)) + existing_hrns_by_types[record.type].append(record.hrn) + - print >>sys.stderr, "\r\n \r\n SLABDRIVER GetPeers existing_hrns_by_types %s " %( existing_hrns_by_types) - records_list= [] + logger.debug("SLABDRIVER \tGetPeer\texisting_hrns_by_types %s "\ + %( existing_hrns_by_types)) + records_list = [] - try: - for hrn in existing_hrns_by_types['authority+sa']: - records_list.append(existing_records[hrn]) - print >>sys.stderr, "\r\n \r\n SLABDRIVER GetPeers records_list %s " %(records_list) - - except: - pass + try: + if peer_filter: + records_list.append(existing_records[(peer_filter,'authority')]) + else : + for hrn in existing_hrns_by_types['authority']: + records_list.append(existing_records[(hrn,'authority')]) + + logger.debug("SLABDRIVER \tGetPeer \trecords_list %s " \ + %(records_list)) - if not peer_filter and not return_fields: + except: + pass + + return_records = records_list + if not peer_filter and not return_fields_list: return records_list - return_records = parse_filter(records_list,peer_filter, 'peers', return_fields) - + + + logger.debug("SLABDRIVER \tGetPeer return_records %s " \ + %(return_records)) return return_records - - def GetPersons(self, person_filter=None, return_fields=None): - - person_list = self.ldap.ldapFind({'authority': self.root_auth }) - - #check = False - #if person_filter and isinstance(person_filter, dict): - #for k in person_filter.keys(): - #if k in person_list[0].keys(): - #check = True - - return_person_list = parse_filter(person_list,person_filter ,'persons', return_fields) - if return_person_list: - print>>sys.stderr, " \r\n GetPersons person_filter %s return_fields %s " %(person_filter,return_fields) - return return_person_list + #TODO : Handling OR request in make_ldap_filters_from_records + #instead of the for loop + #over the records' list + def GetPersons(self, person_filter=None, return_fields_list=None): + """ + person_filter should be a list of dictionnaries when not set to None. + Returns a list of users whose accounts are enabled found in ldap. + + """ + logger.debug("SLABDRIVER \tGetPersons person_filter %s" \ + %(person_filter)) + person_list = [] + if person_filter and isinstance(person_filter, list): + #If we are looking for a list of users (list of dict records) + #Usually the list contains only one user record + for searched_attributes in person_filter: + + #Get only enabled user accounts in senslab LDAP : + #add a filter for make_ldap_filters_from_record + person = self.ldap.LdapFindUser(searched_attributes, \ + is_user_enabled=True) + person_list.append(person) + + else: + #Get only enabled user accounts in senslab LDAP : + #add a filter for make_ldap_filters_from_record + person_list = self.ldap.LdapFindUser(is_user_enabled=True) + + return person_list def GetTimezone(self): - time = self.oar.parser.SendRequest("GET_timezone") - return time + server_timestamp, server_tz = self.oar.parser.\ + SendRequest("GET_timezone") + return server_timestamp, server_tz - def DeleteJobs(self, job_id, username): - if not job_id: + def DeleteJobs(self, job_id, slice_hrn): + if not job_id or job_id is -1: return + username = slice_hrn.split(".")[-1].rstrip("_slice") reqdict = {} reqdict['method'] = "delete" reqdict['strval'] = str(job_id) - answer = self.oar.POSTRequestToOARRestAPI('DELETE_jobs_id',reqdict,username) - print>>sys.stderr, "\r\n \r\n jobid DeleteJobs %s " %(answer) + + self.db.delete_job(slice_hrn, job_id) + answer = self.oar.POSTRequestToOARRestAPI('DELETE_jobs_id', \ + reqdict,username) + logger.debug("SLABDRIVER \tDeleteJobs jobid %s \r\n answer %s username %s" \ + %(job_id,answer, username)) + return answer + + - - def GetJobs(self,job_id= None, resources=True,return_fields=None, username = None): - #job_resources=['reserved_resources', 'assigned_resources','job_id', 'job_uri', 'assigned_nodes',\ - #'api_timestamp'] + ##TODO : Unused GetJobsId ? SA 05/07/12 + #def GetJobsId(self, job_id, username = None ): + #""" + #Details about a specific job. + #Includes details about submission time, jot type, state, events, + #owner, assigned ressources, walltime etc... + + #""" + #req = "GET_jobs_id" + #node_list_k = 'assigned_network_address' + ##Get job info from OAR + #job_info = self.oar.parser.SendRequest(req, job_id, username) + + #logger.debug("SLABDRIVER \t GetJobsId %s " %(job_info)) + #try: + #if job_info['state'] == 'Terminated': + #logger.debug("SLABDRIVER \t GetJobsId job %s TERMINATED"\ + #%(job_id)) + #return None + #if job_info['state'] == 'Error': + #logger.debug("SLABDRIVER \t GetJobsId ERROR message %s "\ + #%(job_info)) + #return None + + #except KeyError: + #logger.error("SLABDRIVER \tGetJobsId KeyError") + #return None + + #parsed_job_info = self.get_info_on_reserved_nodes(job_info, \ + #node_list_k) + ##Replaces the previous entry + ##"assigned_network_address" / "reserved_resources" + ##with "node_ids" + #job_info.update({'node_ids':parsed_job_info[node_list_k]}) + #del job_info[node_list_k] + #logger.debug(" \r\nSLABDRIVER \t GetJobsId job_info %s " %(job_info)) + #return job_info + + + def GetJobsResources(self, job_id, username = None): + #job_resources=['reserved_resources', 'assigned_resources',\ + #'job_id', 'job_uri', 'assigned_nodes',\ + #'api_timestamp'] #assigned_res = ['resource_id', 'resource_uri'] #assigned_n = ['node', 'node_uri'] - - - if job_id and resources is False: - req = "GET_jobs_id" - node_list_k = 'assigned_network_address' - - if job_id and resources : - req = "GET_jobs_id_resources" - node_list_k = 'reserved_resources' - + req = "GET_jobs_id_resources" + node_list_k = 'reserved_resources' - #Get job info from OAR - job_info = self.oar.parser.SendRequest(req, job_id, username) - print>>sys.stderr, "\r\n \r\n \t\t GetJobs %s " %(job_info) - - if 'state' in job_info : - if job_info['state'] == 'Terminated': - print>>sys.stderr, "\r\n \r\n \t\t GetJobs TERMINELEBOUSIN " - return None - if job_info['state'] == 'Error': - print>>sys.stderr, "\r\n \r\n \t\t GetJobs ERROR message %s " %(job_info) - return None - - #Get a dict of nodes . Key :hostname of the node - node_list = self.GetNodes() - node_hostname_list = [] - for node in node_list: - node_hostname_list.append(node['hostname']) - node_dict = dict(zip(node_hostname_list,node_list)) - - #print>>sys.stderr, "\r\n \r\n \r\n \r\n \r\n \t\t GetJobs GetNODES %s " %(node_list) - try : - - #for n in job_info[node_list]: - #n = str(self.root_auth) + str(n) + #Get job resources list from OAR + node_id_list = self.oar.parser.SendRequest(req, job_id, username) + logger.debug("SLABDRIVER \t GetJobsResources %s " %(node_id_list)) + + hostname_list = \ + self.__get_hostnames_from_oar_node_ids(node_id_list) + + #parsed_job_info = self.get_info_on_reserved_nodes(job_info, \ + #node_list_k) + #Replaces the previous entry "assigned_network_address" / + #"reserved_resources" + #with "node_ids" + job_info = {'node_ids': hostname_list} + + return job_info - liste =job_info[node_list_k] - print>>sys.stderr, "\r\n \r\n \t\t GetJobs resources job_info liste%s" %(liste) - for k in range(len(liste)): - job_info[node_list_k][k] = node_dict[job_info[node_list_k][k]]['hostname'] - - print>>sys.stderr, "\r\n \r\n \t\t YYYYYYYYYYYYGetJobs resources job_info %s" %(job_info) - job_info.update({'node_ids':job_info[node_list_k]}) - del job_info[node_list_k] - return job_info + def get_info_on_reserved_nodes(self, job_info, node_list_name): + #Get the list of the testbed nodes records and make a + #dictionnary keyed on the hostname out of it + node_list_dict = self.GetNodes() + #node_hostname_list = [] + node_hostname_list = [node['hostname'] for node in node_list_dict] + #for node in node_list_dict: + #node_hostname_list.append(node['hostname']) + node_dict = dict(zip(node_hostname_list, node_list_dict)) + try : + reserved_node_hostname_list = [] + for index in range(len(job_info[node_list_name])): + #job_info[node_list_name][k] = + reserved_node_hostname_list[index] = \ + node_dict[job_info[node_list_name][index]]['hostname'] + + logger.debug("SLABDRIVER \t get_info_on_reserved_nodes \ + reserved_node_hostname_list %s" \ + %(reserved_node_hostname_list)) except KeyError: - print>>sys.stderr, "\r\n \r\n \t\t GetJobs KEYERROR " + logger.error("SLABDRIVER \t get_info_on_reserved_nodes KEYERROR " ) - + return reserved_node_hostname_list - - + def GetNodesCurrentlyInUse(self): + """Returns a list of all the nodes already involved in an oar job""" + return self.oar.parser.SendRequest("GET_running_jobs") + + def __get_hostnames_from_oar_node_ids(self, resource_id_list ): + full_nodes_dict_list = self.GetNodes() + #Put the full node list into a dictionary keyed by oar node id + oar_id_node_dict = {} + for node in full_nodes_dict_list: + oar_id_node_dict[node['oar_id']] = node + + logger.debug("SLABDRIVER \t __get_hostnames_from_oar_node_ids\ + oar_id_node_dict %s" %(oar_id_node_dict)) + hostname_list = [] + hostname_dict_list = [] + for resource_id in resource_id_list: + #Because jobs requested "asap" do not have defined resources + if resource_id is not "Undefined": + hostname_dict_list.append({'hostname' : \ + oar_id_node_dict[resource_id]['hostname'], + 'site_id' : oar_id_node_dict[resource_id]['site']}) + + #hostname_list.append(oar_id_node_dict[resource_id]['hostname']) + return hostname_dict_list + + def GetReservedNodes(self): + #Get the nodes in use and the reserved nodes + reservation_dict_list = \ + self.oar.parser.SendRequest("GET_reserved_nodes") + + + for resa in reservation_dict_list: + logger.debug ("GetReservedNodes resa %s"%(resa)) + #dict list of hostnames and their site + resa['reserved_nodes'] = \ + self.__get_hostnames_from_oar_node_ids(resa['resource_ids']) + + #del resa['resource_ids'] + return reservation_dict_list - def GetNodes(self,node_filter= None, return_fields=None): - - node_dict =self.oar.parser.SendRequest("GET_resources_full") - print>>sys.stderr, "\r\n \r\n \t\t SLABDRIVER.PY GetNodes " + def GetNodes(self, node_filter_dict = None, return_fields_list = None): + """ + node_filter_dict : dictionnary of lists + + """ + node_dict_by_id = self.oar.parser.SendRequest("GET_resources_full") + node_dict_list = node_dict_by_id.values() + + #No filtering needed return the list directly + if not (node_filter_dict or return_fields_list): + return node_dict_list + return_node_list = [] - if not (node_filter or return_fields): - return_node_list = node_dict.values() - return return_node_list - - return_node_list= parse_filter(node_dict.values(),node_filter ,'node', return_fields) + if node_filter_dict: + for filter_key in node_filter_dict: + try: + #Filter the node_dict_list by each value contained in the + #list node_filter_dict[filter_key] + for value in node_filter_dict[filter_key]: + for node in node_dict_list: + if node[filter_key] == value: + if return_fields_list : + tmp = {} + for k in return_fields_list: + tmp[k] = node[k] + return_node_list.append(tmp) + else: + return_node_list.append(node) + except KeyError: + logger.log_exc("GetNodes KeyError") + return + + return return_node_list - def GetSites(self, site_filter = None, return_fields=None): - site_dict =self.oar.parser.SendRequest("GET_sites") - print>>sys.stderr, "\r\n \r\n \t\t SLABDRIVER.PY GetSites " + def GetSites(self, site_filter_name_list = None, return_fields_list = None): + site_dict = self.oar.parser.SendRequest("GET_sites") + #site_dict : dict where the key is the sit ename return_site_list = [] - if not ( site_filter or return_fields): - return_site_list = site_dict.values() - return return_site_list - - return_site_list = parse_filter(site_dict.values(), site_filter,'site', return_fields) - return return_site_list + if not ( site_filter_name_list or return_fields_list): + return_site_list = site_dict.values() + return return_site_list + for site_filter_name in site_filter_name_list: + if site_filter_name in site_dict: + if return_fields_list: + for field in return_fields_list: + tmp = {} + try: + tmp[field] = site_dict[site_filter_name][field] + except KeyError: + logger.error("GetSites KeyError %s "%(field)) + return None + return_site_list.append(tmp) + else: + return_site_list.append( site_dict[site_filter_name]) + - def GetSlices(self,slice_filter = None, filter_type = None, return_fields=None): + return return_site_list + #warning return_fields_list paramr emoved (Not used) + def GetSlices(self, slice_filter = None, slice_filter_type = None): + #def GetSlices(self, slice_filter = None, slice_filter_type = None, \ + #return_fields_list = None): + """ Get the slice records from the slab db. + Returns a slice ditc if slice_filter and slice_filter_type + are specified. + Returns a list of slice dictionnaries if there are no filters + specified. + + """ + login = None return_slice_list = [] slicerec = {} - ftypes = ['slice_hrn', 'record_id_user'] - if filter_type and filter_type in ftypes: - if filter_type == 'slice_hrn': - slicerec = slab_dbsession.query(SliceSenslab).filter_by(slice_hrn = slice_filter).first() - if filter_type == 'record_id_user': + slicerec_dict = {} + authorized_filter_types_list = ['slice_hrn', 'record_id_user'] + + if slice_filter_type in authorized_filter_types_list: + #Get list of slices based on the slice hrn + if slice_filter_type == 'slice_hrn': + + login = slice_filter.split(".")[1].split("_")[0] + + #DO NOT USE RegSlice - reg_researchers to get the hrn of the user + #otherwise will mess up the RegRecord in Resolve, don't know + #why - SA 08/08/2012 + + #Only one entry for one user = one slice in slice_senslab table + slicerec = slab_dbsession.query(SliceSenslab).filter_by(slice_hrn = slice_filter).first() + + #Get slice based on user id + if slice_filter_type == 'record_id_user': slicerec = slab_dbsession.query(SliceSenslab).filter_by(record_id_user = slice_filter).first() - if slicerec: - rec = slicerec.dumpquerytodict() - login = slicerec.slice_hrn.split(".")[1].split("_")[0] - print >>sys.stderr, " \r\n \r\n \tSLABDRIVER.PY slicerec GetSlices %s " %(slicerec) - if slicerec.oar_job_id is not -1: - rslt = self.GetJobs( slicerec.oar_job_id, resources=False, username = login ) - print >>sys.stderr, " \r\n \r\n \tSLABDRIVER.PY GetSlices GetJobs %s " %(rslt) - if rslt : - rec.update(rslt) - rec.update({'hrn':str(rec['slice_hrn'])}) - #If GetJobs is empty, this means the job is now in the 'Terminated' state - #Update the slice record - else : - self.db.update_job(slice_filter, job_id = '-1') - rec['oar_job_id'] = '-1' - rec.update({'hrn':str(rec['slice_hrn'])}) - - print >>sys.stderr, " \r\n \r\n \tSLABDRIVER.PY GetSlices rec %s" %(rec) - return rec + if slicerec is None: + return [] + + #slicerec_dictlist = [] + slicerec_dict = slicerec.dump_sqlalchemyobj_to_dict() + if login is None : + login = slicerec_dict['slice_hrn'].split(".")[1].split("_")[0] + + #for record in slicerec: + #slicerec_dictlist.append(record.dump_sqlalchemyobj_to_dict()) + #if login is None : + #login = slicerec_dictlist[0]['slice_hrn'].split(".")[1].split("_")[0] + + #One slice can have multiple jobs + sqljob_list = slab_dbsession.query(JobSenslab).filter_by( slice_hrn=slicerec_dict['slice_hrn']).all() + job_list = [] + for job in sqljob_list: + job_list.append(job.dump_sqlalchemyobj_to_dict()) - else: - return_slice_list = slab_dbsession.query(SliceSenslab).all() - - print >>sys.stderr, " \r\n \r\n \tSLABDRIVER.PY GetSlices slices %s slice_filter %s " %(return_slice_list,slice_filter) - - #if return_fields: - #return_slice_list = parse_filter(sliceslist, slice_filter,'slice', return_fields) + logger.debug("\r\n SLABDRIVER \tGetSlices login %s \ + slice record %s" \ + %(login, slicerec_dict)) + + #Several jobs for one slice + slicerec_dict['oar_job_id'] = [] + for job in job_list : + #if slicerec_dict['oar_job_id'] is not -1: + #Check with OAR the status of the job if a job id is in + #the slice record + + rslt = self.GetJobsResources(job['oar_job_id'], \ + username = login) + logger.debug("SLABDRIVER.PY \tGetSlices rslt fromn GetJobsResources %s"\ + %(rslt)) + if rslt : + slicerec_dict['oar_job_id'].append(job['oar_job_id']) + slicerec_dict.update(rslt) + slicerec_dict.update({'hrn':\ + str(slicerec_dict['slice_hrn'])}) + #If GetJobsResources is empty, this means the job is + #now in the 'Terminated' state + #Update the slice record + else : + self.db.delete_job(slice_filter, job['oar_job_id']) + slicerec_dict.\ + update({'hrn':str(slicerec_dict['slice_hrn'])}) + try: + slicerec_dict['node_ids'] = job['node_list'] + except KeyError: + pass + + logger.debug("SLABDRIVER.PY \tGetSlices RETURN slicerec_dict %s"\ + %(slicerec_dict)) + + return [slicerec_dict] + + + else: + slice_list = slab_dbsession.query(SliceSenslab).all() + sqljob_list = slab_dbsession.query(JobSenslab).all() + + job_list = [] + for job in sqljob_list: + job_list.append(job.dump_sqlalchemyobj_to_dict()) + + return_slice_list = [] + for record in slice_list: + return_slice_list.append(record.dump_sqlalchemyobj_to_dict()) + + for slicerec_dict in return_slice_list: + slicerec_dict['oar_job_id'] = [] + for job in job_list: + if slicerec_dict['slice_hrn'] in job: + slicerec_dict['oar_job_id'].append(job['oar_job_id']) + + logger.debug("SLABDRIVER.PY \tGetSlices RETURN slices %s \ + slice_filter %s " %(return_slice_list, slice_filter)) - + #if return_fields_list: + #return_slice_list = parse_filter(sliceslist, \ + #slice_filter,'slice', return_fields_list) + return return_slice_list - + + - def testbed_name (self): return "senslab2" + def testbed_name (self): return self.hrn # 'geni_request_rspec_versions' and 'geni_ad_rspec_versions' are mandatory def aggregate_version (self): @@ -615,7 +989,7 @@ class SlabDriver(Driver): # @param sfa_fields dictionary of SFA fields # @param slab_fields dictionary of PLC fields (output) - def sfa_fields_to_slab_fields(self, type, hrn, record): + def sfa_fields_to_slab_fields(self, sfa_type, hrn, record): def convert_ints(tmpdict, int_fields): for field in int_fields: @@ -626,18 +1000,21 @@ class SlabDriver(Driver): #for field in record: # slab_record[field] = record[field] - if type == "slice": + if sfa_type == "slice": #instantion used in get_slivers ? if not "instantiation" in slab_record: slab_record["instantiation"] = "senslab-instantiated" - slab_record["hrn"] = hrn_to_pl_slicename(hrn) - print >>sys.stderr, "\r\n \r\n \t SLABDRIVER.PY sfa_fields_to_slab_fields slab_record %s hrn_to_pl_slicename(hrn) hrn %s " %(slab_record['hrn'], hrn) - if "url" in record: - slab_record["url"] = record["url"] - if "description" in record: - slab_record["description"] = record["description"] - if "expires" in record: - slab_record["expires"] = int(record["expires"]) + #slab_record["hrn"] = hrn_to_pl_slicename(hrn) + #Unused hrn_to_pl_slicename because Slab's hrn already in the appropriate form SA 23/07/12 + slab_record["hrn"] = hrn + logger.debug("SLABDRIVER.PY sfa_fields_to_slab_fields \ + slab_record %s " %(slab_record['hrn'])) + if "url" in record: + slab_record["url"] = record["url"] + if "description" in record: + slab_record["description"] = record["description"] + if "expires" in record: + slab_record["expires"] = int(record["expires"]) #nodes added by OAR only and then imported to SFA #elif type == "node": @@ -666,285 +1043,395 @@ class SlabDriver(Driver): return slab_record - - - - def AddSliceToNodes(self, slice_name, added_nodes, slice_user=None): - - site_list = [] - nodeid_list =[] - resource = "" - reqdict = {} - reqdict['property'] ="network_address in (" - for node in added_nodes: - #Get the ID of the node : remove the root auth and put the site in a separate list - s=node.split(".") - # NT: it's not clear for me if the nodenames will have the senslab prefix - # so lets take the last part only, for now. - lastpart=s[-1] - #if s[0] == self.root_auth : - # Again here it's not clear if nodes will be prefixed with _, lets split and tanke the last part for now. - s=lastpart.split("_") - nodeid=s[-1] - reqdict['property'] += "'"+ nodeid +"', " - nodeid_list.append(nodeid) - #site_list.append( l[0] ) - reqdict['property'] = reqdict['property'][0: len( reqdict['property'])-2] +")" - reqdict['resource'] ="network_address="+ str(len(nodeid_list)) - reqdict['resource']+= ",walltime=" + str(00) + ":" + str(12) + ":" + str(20) #+2 min 20 - reqdict['script_path'] = "/bin/sleep 620" #+20 sec - reqdict['type'] = "deploy" - reqdict['directory']= "" - reqdict['name']= "TestSandrine" - timestamp = self.GetTimezone() - print>>sys.stderr, "\r\n \r\n AddSliceToNodes slice_name %s added_nodes %s username %s reqdict %s " %(slice_name,added_nodes,slice_user, reqdict) - readable_time = strftime(self.time_format, gmtime(float(timestamp))) - print >>sys.stderr," \r\n \r\n \t\t\t\t AVANT ParseTimezone readable_time %s timestanp %s " %(readable_time, timestamp ) - timestamp = timestamp+ 3620 #Add 3 min to server time - readable_time = strftime(self.time_format, gmtime(float(timestamp))) - - print >>sys.stderr," \r\n \r\n \t\t\t\tAPRES ParseTimezone readable_time %s timestanp %s " %(readable_time , timestamp) - reqdict['reservation'] = readable_time - - # first step : start the OAR job - print>>sys.stderr, "\r\n \r\n AddSliceToNodes reqdict %s \r\n site_list %s" %(reqdict,site_list) - #OAR = OARrestapi() - answer = self.oar.POSTRequestToOARRestAPI('POST_job',reqdict,slice_user) - print>>sys.stderr, "\r\n \r\n AddSliceToNodes jobid %s " %(answer) - #self.db.update('slice',['oar_job_id'], [answer['id']], 'slice_hrn', slice_name) - + - self.db.update_job( slice_name, job_id = answer['id'] ) - jobid=answer['id'] - print>>sys.stderr, "\r\n \r\n AddSliceToNodes jobid %s added_nodes %s slice_user %s" %(jobid,added_nodes,slice_user) - # second step : configure the experiment - # we need to store the nodes in a yaml (well...) file like this : - # [1,56,23,14,45,75] with name /tmp/sfa.json - f=open('/tmp/sfa/'+str(jobid)+'.json','w') - f.write('[') - f.write(str(added_nodes[0].strip('node'))) - for node in added_nodes[1:len(added_nodes)] : - f.write(','+node.strip('node')) - f.write(']') - f.close() - - # third step : call the senslab-experiment wrapper - #command= "java -jar target/sfa-1.0-jar-with-dependencies.jar "+str(jobid)+" "+slice_user - javacmdline="/usr/bin/java" - jarname="/opt/senslabexperimentwrapper/sfa-1.0-jar-with-dependencies.jar" - #ret=subprocess.check_output(["/usr/bin/java", "-jar", ", str(jobid), slice_user]) - output = subprocess.Popen([javacmdline, "-jar", jarname, str(jobid), slice_user],stdout=subprocess.PIPE).communicate()[0] - - print>>sys.stderr, "\r\n \r\n AddSliceToNodes wrapper returns %s " %(output) - return + + def __transforms_timestamp_into_date(self, xp_utc_timestamp = None): + """ Transforms unix timestamp into valid OAR date format """ + + #Used in case of a scheduled experiment (not immediate) + #To run an XP immediately, don't specify date and time in RSpec + #They will be set to None. + if xp_utc_timestamp: + #transform the xp_utc_timestamp into server readable time + xp_server_readable_date = datetime.fromtimestamp(int(\ + xp_utc_timestamp)).strftime(self.time_format) + + return xp_server_readable_date + + else: + return None + + + + + def LaunchExperimentOnOAR(self, added_nodes, slice_name, \ + lease_start_time, lease_duration, slice_user=None): + lease_dict = {} + lease_dict['lease_start_time'] = lease_start_time + lease_dict['lease_duration'] = lease_duration + lease_dict['added_nodes'] = added_nodes + lease_dict['slice_name'] = slice_name + lease_dict['slice_user'] = slice_user + lease_dict['grain'] = self.GetLeaseGranularity() + lease_dict['time_format'] = self.time_format + + def __create_job_structure_request_for_OAR(lease_dict): + """ Creates the structure needed for a correct POST on OAR. + Makes the timestamp transformation into the appropriate format. + Sends the POST request to create the job with the resources in + added_nodes. + + """ + + site_list = [] + nodeid_list = [] + resource = "" + reqdict = {} + + + reqdict['workdir'] = '/tmp' + reqdict['resource'] = "{network_address in (" + + for node in lease_dict['added_nodes']: + logger.debug("\r\n \r\n OARrestapi \t __create_job_structure_request_for_OAR \ + node %s" %(node)) + + # Get the ID of the node + nodeid = node + reqdict['resource'] += "'" + nodeid + "', " + nodeid_list.append(nodeid) + + custom_length = len(reqdict['resource'])- 2 + reqdict['resource'] = reqdict['resource'][0:custom_length] + \ + ")}/nodes=" + str(len(nodeid_list)) + def __process_walltime(duration): + """ Calculates the walltime in seconds from the duration in H:M:S + specified in the RSpec. + + """ + if duration: + # Fixing the walltime by adding a few delays. + # First put the walltime in seconds oarAdditionalDelay = 20; + # additional delay for /bin/sleep command to + # take in account prologue and epilogue scripts execution + # int walltimeAdditionalDelay = 120; additional delay + desired_walltime = duration + total_walltime = desired_walltime + 140#+2 min 20 + sleep_walltime = desired_walltime + 20 #+20 sec + walltime = [] + #Put the walltime back in str form + #First get the hours + walltime.append(str(total_walltime / 3600)) + total_walltime = total_walltime - 3600 * int(walltime[0]) + #Get the remaining minutes + walltime.append(str(total_walltime / 60)) + total_walltime = total_walltime - 60 * int(walltime[1]) + #Get the seconds + walltime.append(str(total_walltime)) + + else: + logger.log_exc(" __process_walltime duration null") + + return walltime, sleep_walltime + + walltime, sleep_walltime = \ + __process_walltime(int(lease_dict['lease_duration'])*lease_dict['grain']) + + + reqdict['resource'] += ",walltime=" + str(walltime[0]) + \ + ":" + str(walltime[1]) + ":" + str(walltime[2]) + reqdict['script_path'] = "/bin/sleep " + str(sleep_walltime) + + #In case of a scheduled experiment (not immediate) + #To run an XP immediately, don't specify date and time in RSpec + #They will be set to None. + if lease_dict['lease_start_time'] is not '0': + #Readable time accepted by OAR + start_time = datetime.fromtimestamp(int(lease_dict['lease_start_time'])).\ + strftime(lease_dict['time_format']) + reqdict['reservation'] = start_time + #If there is not start time, Immediate XP. No need to add special + # OAR parameters + + + reqdict['type'] = "deploy" + reqdict['directory'] = "" + reqdict['name'] = "SFA_" + lease_dict['slice_user'] + + return reqdict + + + #Create the request for OAR + reqdict = __create_job_structure_request_for_OAR(lease_dict) + # first step : start the OAR job and update the job + logger.debug("SLABDRIVER.PY \tLaunchExperimentOnOAR reqdict %s\ + \r\n " %(reqdict)) + + answer = self.oar.POSTRequestToOARRestAPI('POST_job', \ + reqdict, slice_user) + logger.debug("SLABDRIVER \tLaunchExperimentOnOAR jobid %s " %(answer)) + try: + jobid = answer['id'] + except KeyError: + logger.log_exc("SLABDRIVER \tLaunchExperimentOnOAR \ + Impossible to create job %s " %(answer)) + return + + + def __configure_experiment(jobid, added_nodes): + # second step : configure the experiment + # we need to store the nodes in a yaml (well...) file like this : + # [1,56,23,14,45,75] with name /tmp/sfa.json + job_file = open('/tmp/sfa/'+ str(jobid) + '.json', 'w') + job_file.write('[') + job_file.write(str(added_nodes[0].strip('node'))) + for node in added_nodes[1:len(added_nodes)] : + job_file.write(', '+ node.strip('node')) + job_file.write(']') + job_file.close() + return + + def __launch_senslab_experiment(jobid): + # third step : call the senslab-experiment wrapper + #command= "java -jar target/sfa-1.0-jar-with-dependencies.jar + # "+str(jobid)+" "+slice_user + javacmdline = "/usr/bin/java" + jarname = \ + "/opt/senslabexperimentwrapper/sfa-1.0-jar-with-dependencies.jar" + #ret=subprocess.check_output(["/usr/bin/java", "-jar", ", \ + #str(jobid), slice_user]) + output = subprocess.Popen([javacmdline, "-jar", jarname, str(jobid), \ + slice_user],stdout=subprocess.PIPE).communicate()[0] + + logger.debug("SLABDRIVER \t __configure_experiment wrapper returns%s " \ + %(output)) + return - def DeleteSliceFromNodes(self, slice_name, deleted_nodes): + + if jobid : + logger.debug("SLABDRIVER \tLaunchExperimentOnOAR jobid %s \ + added_nodes %s slice_user %s" %(jobid, added_nodes, slice_user)) + self.db.add_job( slice_name, jobid, added_nodes) + + __configure_experiment(jobid, added_nodes) + __launch_senslab_experiment(jobid) + + return + + def AddLeases(self, hostname_list, slice_record, lease_start_time, lease_duration): + logger.debug("SLABDRIVER \r\n \r\n \t AddLeases hostname_list %s \ + slice_record %s lease_start_time %s lease_duration %s "\ + %( hostname_list, slice_record , lease_start_time, \ + lease_duration)) + + tmp = slice_record['PI'][0].split(".") + username = tmp[(len(tmp)-1)] + self.LaunchExperimentOnOAR(hostname_list, slice_record['name'], lease_start_time, lease_duration, username) + start_time = datetime.fromtimestamp(int(lease_start_time)).strftime(self.time_format) + logger.debug("SLABDRIVER \t AddLeases hostname_list start_time %s " %(start_time)) + + return + + + #Delete the jobs and updates the job id in the senslab table + #to set it to -1 + #Does not clear the node list + def DeleteSliceFromNodes(self, slice_record): + # Get user information + + self.DeleteJobs(slice_record['oar_job_id'], slice_record['hrn']) + return - - def fill_record_sfa_info(self, records): - - def startswith(prefix, values): - return [value for value in values if value.startswith(prefix)] - - # get person ids - person_ids = [] - site_ids = [] - for record in records: - person_ids.extend(record.get("person_ids", [])) - site_ids.extend(record.get("site_ids", [])) - if 'site_id' in record: - site_ids.append(record['site_id']) - - #print>>sys.stderr, "\r\n \r\n _fill_record_sfa_info ___person_ids %s \r\n \t\t site_ids %s " %(person_ids, site_ids) - - # get all pis from the sites we've encountered - # and store them in a dictionary keyed on site_id - site_pis = {} - if site_ids: - pi_filter = {'|roles': ['pi'], '|site_ids': site_ids} - pi_list = self.GetPersons( pi_filter, ['person_id', 'site_ids']) - #print>>sys.stderr, "\r\n \r\n _fill_record_sfa_info ___ GetPersons ['person_id', 'site_ids'] pi_ilist %s" %(pi_list) - - for pi in pi_list: - # we will need the pi's hrns also - person_ids.append(pi['person_id']) - - # we also need to keep track of the sites these pis - # belong to - for site_id in pi['site_ids']: - if site_id in site_pis: - site_pis[site_id].append(pi) - else: - site_pis[site_id] = [pi] - - # get sfa records for all records associated with these records. - # we'll replace pl ids (person_ids) with hrns from the sfa records - # we obtain + def GetLeaseGranularity(self): + """ Returns the granularity of Senslab testbed. + Defined in seconds. """ - # get the sfa records - #table = SfaTable() - existing_records = {} - all_records = dbsession.query(RegRecord).all() - for record in all_records: - existing_records[(record.type,record.pointer)] = record + grain = 60 + return grain + + def GetLeases(self, lease_filter_dict=None, return_fields_list=None): + unfiltered_reservation_list = self.GetReservedNodes() + + ##Synchronize slice_table of sfa senslab db + #self.synchronize_oar_and_slice_table(unfiltered_reservation_list) + + reservation_list = [] + #Find the slice associated with this user senslab ldap uid + logger.debug(" SLABDRIVER.PY \tGetLeases ") + #Create user dict first to avoir looking several times for + #the same user in LDAP SA 27/07/12 + resa_user_dict = {} + for resa in unfiltered_reservation_list: + logger.debug("SLABDRIVER \tGetLeases USER %s"\ + %(resa['user'])) + if resa['user'] not in resa_user_dict: + logger.debug("SLABDRIVER \tGetLeases userNOTIN ") + ldap_info = self.ldap.LdapSearch('(uid='+resa['user']+')') + ldap_info = ldap_info[0][1] + user = dbsession.query(RegUser).filter_by(email = \ + ldap_info['mail'][0]).first() + #Separated in case user not in database : record_id not defined SA 17/07//12 + query_slice_info = slab_dbsession.query(SliceSenslab).filter_by(record_id_user = user.record_id) + if query_slice_info: + slice_info = query_slice_info.first() + else: + slice_info = None + + resa_user_dict[resa['user']] = {} + resa_user_dict[resa['user']]['ldap_info'] = user + resa_user_dict[resa['user']]['slice_info'] = slice_info + + logger.debug("SLABDRIVER \tGetLeases resa_user_dict %s"\ + %(resa_user_dict)) + for resa in unfiltered_reservation_list: - print >>sys.stderr, " \r\r\n SLABDRIVER fill_record_sfa_info existing_records %s " %(existing_records) - person_list, persons = [], {} - #person_list = table.find({'type': 'user', 'pointer': person_ids}) - try: - for p_id in person_ids: - person_list.append( existing_records.get(('user',p_id))) - except KeyError: - print >>sys.stderr, " \r\r\n SLABDRIVER fill_record_sfa_info ERRRRRRRRRROR" - - # create a hrns keyed on the sfa record's pointer. - # Its possible for multiple records to have the same pointer so - # the dict's value will be a list of hrns. - persons = defaultdict(list) - for person in person_list: - persons[person['pointer']].append(person) - - # get the pl records - slab_person_list, slab_persons = [], {} - slab_person_list = self.GetPersons(person_ids, ['person_id', 'roles']) - slab_persons = list_to_dict(slab_person_list, 'person_id') - #print>>sys.stderr, "\r\n \r\n _fill_record_sfa_info ___ _list %s \r\n \t\t SenslabUsers.GetPersons ['person_id', 'roles'] slab_persons %s \r\n records %s" %(slab_person_list, slab_persons,records) - # fill sfa info - - for record in records: - # skip records with no pl info (top level authorities) - #Sandrine 24 oct 11 2 lines - #if record['pointer'] == -1: - #continue - sfa_info = {} - type = record['type'] - if (type == "slice"): - # all slice users are researchers - #record['geni_urn'] = hrn_to_urn(record['hrn'], 'slice') ? besoin ou pas ? - record['PI'] = [] - record['researcher'] = [] - for person_id in record.get('person_ids', []): - #Sandrine 24 oct 11 line - #for person_id in record['person_ids']: - hrns = [person['hrn'] for person in persons[person_id]] - record['researcher'].extend(hrns) - - # pis at the slice's site - slab_pis = site_pis[record['site_id']] - pi_ids = [pi['person_id'] for pi in slab_pis] - for person_id in pi_ids: - hrns = [person['hrn'] for person in persons[person_id]] - record['PI'].extend(hrns) - record['geni_urn'] = hrn_to_urn(record['hrn'], 'slice') - record['geni_creator'] = record['PI'] - - elif (type == "authority"): - record['PI'] = [] - record['operator'] = [] - record['owner'] = [] - for pointer in record['person_ids']: - if pointer not in persons or pointer not in slab_persons: - # this means there is not sfa or pl record for this user - continue - hrns = [person['hrn'] for person in persons[pointer]] - roles = slab_persons[pointer]['roles'] - if 'pi' in roles: - record['PI'].extend(hrns) - if 'tech' in roles: - record['operator'].extend(hrns) - if 'admin' in roles: - record['owner'].extend(hrns) - # xxx TODO: OrganizationName - elif (type == "node"): - sfa_info['dns'] = record.get("hostname", "") - # xxx TODO: URI, LatLong, IP, DNS - - elif (type == "user"): - sfa_info['email'] = record.get("email", "") - sfa_info['geni_urn'] = hrn_to_urn(record['hrn'], 'user') - sfa_info['geni_certificate'] = record['gid'] - # xxx TODO: PostalAddress, Phone - - #print>>sys.stderr, "\r\n \r\rn \t\t \t <<<<<<<<<<<<<<<<<<<<<<<< fill_record_sfa_info sfa_info %s \r\n record %s : "%(sfa_info,record) - record.update(sfa_info) + #ldap_info = self.ldap.LdapSearch('(uid='+resa['user']+')') + #ldap_info = ldap_info[0][1] + + #user = dbsession.query(RegUser).filter_by(email = \ + #ldap_info['mail'][0]).first() + ##Separated in case user not in database : record_id not defined SA 17/07//12 + #query_slice_info = slab_dbsession.query(SliceSenslab).filter_by(record_id_user = user.record_id) + #if query_slice_info: + #slice_info = query_slice_info.first() + #Put the slice_urn + resa['slice_hrn'] = resa_user_dict[resa['user']]['slice_info'].slice_hrn + resa['slice_id'] = hrn_to_urn(resa['slice_hrn'], 'slice') + #Put the slice_urn + #resa['slice_id'] = hrn_to_urn(slice_info.slice_hrn, 'slice') + resa['component_id_list'] = [] + #Transform the hostnames into urns (component ids) + for node in resa['reserved_nodes']: + #resa['component_id_list'].append(hostname_to_urn(self.hrn, \ + #self.root_auth, node['hostname'])) + slab_xrn = slab_xrn_object(self.root_auth, node['hostname']) + resa['component_id_list'].append(slab_xrn.urn) + + #Filter the reservation list if necessary + #Returns all the leases associated with a given slice + if lease_filter_dict: + logger.debug("SLABDRIVER \tGetLeases lease_filter_dict %s"\ + %(lease_filter_dict)) + for resa in unfiltered_reservation_list: + if lease_filter_dict['name'] == resa['slice_hrn']: + reservation_list.append(resa) + else: + reservation_list = unfiltered_reservation_list + + logger.debug(" SLABDRIVER.PY \tGetLeases reservation_list %s"\ + %(reservation_list)) + return reservation_list def augment_records_with_testbed_info (self, sfa_records): return self.fill_record_info (sfa_records) - def fill_record_info(self, records): + def fill_record_info(self, record_list): """ Given a SFA record, fill in the senslab specific and SFA specific fields in the record. """ - print >>sys.stderr, "\r\n \t\t SLABDRIVER.PY fill_record_info 000000000 fill_record_info %s " %(records) - if not isinstance(records, list): - records = [records] - - parkour = records + logger.debug("SLABDRIVER \tfill_record_info records %s " %(record_list)) + if not isinstance(record_list, list): + record_list = [record_list] + try: - for record in parkour: - + for record in record_list: + #If the record is a SFA slice record, then add information + #about the user of this slice. This kind of + #information is in the Senslab's DB. if str(record['type']) == 'slice': - print >>sys.stderr, "\r\n \t\t SLABDRIVER.PY fill_record_info \t \t record %s" %(record) - #sfatable = SfaTable() + #Get slab slice record. + recslice_list = self.GetSlices(slice_filter = \ + str(record['hrn']),\ + slice_filter_type = 'slice_hrn') - #existing_records_by_id = {} - #all_records = dbsession.query(RegRecord).all() - #for rec in all_records: - #existing_records_by_id[rec.record_id] = rec - #print >>sys.stderr, "\r\n \t\t SLABDRIVER.PY fill_record_info \t\t existing_records_by_id %s" %(existing_records_by_id[record['record_id']]) - - #recslice = self.db.find('slice',{'slice_hrn':str(record['hrn'])}) - #recslice = slab_dbsession.query(SliceSenslab).filter_by(slice_hrn = str(record['hrn'])).first() - recslice = self.GetSlices(slice_filter = str(record['hrn']), filter_type = 'slice_hrn') - print >>sys.stderr, "\r\n \t\t SLABDRIVER.PY fill_record_info \t\t HOY HOY reclise %s" %(recslice) - #if isinstance(recslice,list) and len(recslice) == 1: - #recslice = recslice[0] - - recuser = dbsession.query(RegRecord).filter_by(record_id = recslice['record_id_user']).first() - #existing_records_by_id[recslice['record_id_user']] - print >>sys.stderr, "\r\n \t\t SLABDRIVER.PY fill_record_info \t\t recuser %s" %(recuser) - - + recuser = dbsession.query(RegRecord).filter_by(record_id = \ + recslice_list[0]['record_id_user']).first() + logger.debug("SLABDRIVER \tfill_record_info TYPE SLICE RECUSER %s " %(recuser)) record.update({'PI':[recuser.hrn], - 'researcher': [recuser.hrn], - 'name':record['hrn'], - 'oar_job_id':recslice['oar_job_id'], - 'node_ids': [], - 'person_ids':[recslice['record_id_user']]}) - - elif str(record['type']) == 'user': - print >>sys.stderr, "\r\n \t\t SLABDRIVER.PY fill_record_info USEEEEEEEEEERDESU!" - - rec = self.GetSlices(slice_filter = record['record_id'], filter_type = 'record_id_user') - #Append record in records list, therfore fetches user and slice info again(one more loop) + 'researcher': [recuser.hrn], + 'name':record['hrn'], + 'oar_job_id':[rec['oar_job_id'] for rec in recslice_list], + 'node_ids': [], + 'person_ids':[recslice_list[0]['record_id_user']], + 'geni_urn':'', #For client_helper.py compatibility + 'keys':'', #For client_helper.py compatibility + 'key_ids':''}) #For client_helper.py compatibility + + #for rec in recslice_list: + #record['oar_job_id'].append(rec['oar_job_id']) + logger.debug( "SLABDRIVER.PY \t fill_record_info SLICE \ + recslice_list %s \r\n \t RECORD %s \r\n \r\n" %(recslice_list,record)) + if str(record['type']) == 'user': + #The record is a SFA user record. + #Get the information about his slice from Senslab's DB + #and add it to the user record. + recslice_list = self.GetSlices(\ + slice_filter = record['record_id'],\ + slice_filter_type = 'record_id_user') + + logger.debug( "SLABDRIVER.PY \t fill_record_info TYPE USER \ + recslice_list %s \r\n \t RECORD %s \r\n" %(recslice_list , record)) + #Append slice record in records list, + #therefore fetches user and slice info again(one more loop) #Will update PIs and researcher for the slice - - rec.update({'type':'slice','hrn':rec['slice_hrn']}) - records.append(rec) - print >>sys.stderr, "\r\n \t\t SLABDRIVER.PY fill_record_info ADDING SLIC EINFO rec %s" %(rec) + recuser = dbsession.query(RegRecord).filter_by(record_id = \ + recslice_list[0]['record_id_user']).first() + logger.debug( "SLABDRIVER.PY \t fill_record_info USER \ + recuser %s \r\n \r\n" %(recuser)) + recslice = {} + recslice = recslice_list[0] + recslice.update({'PI':[recuser.hrn], + 'researcher': [recuser.hrn], + 'name':record['hrn'], + 'node_ids': [], + 'oar_job_id': [rec['oar_job_id'] for rec in recslice_list], + 'person_ids':[recslice_list[0]['record_id_user']]}) + recslice.update({'type':'slice', \ + 'hrn':recslice_list[0]['slice_hrn']}) + #for rec in recslice_list: + #recslice['oar_job_id'].append(rec['oar_job_id']) + + #GetPersons takes [] as filters + #user_slab = self.GetPersons([{'hrn':recuser.hrn}]) + user_slab = self.GetPersons([record]) + - print >>sys.stderr, "\r\n \t\t SLABDRIVER.PY fill_record_info OKrecords %s" %(records) - except TypeError: - print >>sys.stderr, "\r\n \t\t SLABDRIVER fill_record_info EXCEPTION RECORDS : %s" %(records) - return + record.update(user_slab[0]) + #For client_helper.py compatibility + record.update( { 'geni_urn':'', + 'keys':'', + 'key_ids':'' }) + record_list.append(recslice) + + logger.debug("SLABDRIVER.PY \tfill_record_info ADDING SLICE\ + INFO TO USER records %s" %(record_list)) + logger.debug("SLABDRIVER.PY \tfill_record_info END \ + #record %s \r\n \r\n " %(record)) + + except TypeError, error: + logger.log_exc("SLABDRIVER \t fill_record_info EXCEPTION %s"\ + %(error)) + #logger.debug("SLABDRIVER.PY \t fill_record_info ENDENDEND ") + + return #self.fill_record_slab_info(records) - ##print >>sys.stderr, "\r\n \t\t after fill_record_slab_info %s" %(records) - #self.fill_record_sfa_info(records) - #print >>sys.stderr, "\r\n \t\t after fill_record_sfa_info" - + + - - #def update_membership_list(self, oldRecord, record, listName, addFunc, delFunc): + #TODO Update membership? update_membership_list SA 05/07/12 + #def update_membership_list(self, oldRecord, record, listName, addFunc, \ + #delFunc): ## get a list of the HRNs tht are members of the old and new records #if oldRecord: #oldList = oldRecord.get(listName, []) @@ -985,7 +1472,7 @@ class SlabDriver(Driver): #delFunc(self.plauth, personId, containerId) #def update_membership(self, oldRecord, record): - #print >>sys.stderr, " \r\n \r\n ***SLABDRIVER.PY update_membership record ", record + #if record.type == "slice": #self.update_membership_list(oldRecord, record, 'researcher', #self.users.AddPersonToSlice, @@ -997,3 +1484,191 @@ class SlabDriver(Driver): ### thierry # I don't think you plan on running a component manager at this point # let me clean up the mess of ComponentAPI that is deprecated anyways + + +#TODO FUNCTIONS SECTION 04/07/2012 SA + + #TODO : Is UnBindObjectFromPeer still necessary ? Currently does nothing + #04/07/2012 SA + def UnBindObjectFromPeer(self, auth, object_type, object_id, shortname): + """ This method is a hopefully temporary hack to let the sfa correctly + detach the objects it creates from a remote peer object. This is + needed so that the sfa federation link can work in parallel with + RefreshPeer, as RefreshPeer depends on remote objects being correctly + marked. + Parameters: + auth : struct, API authentication structure + AuthMethod : string, Authentication method to use + object_type : string, Object type, among 'site','person','slice', + 'node','key' + object_id : int, object_id + shortname : string, peer shortname + FROM PLC DOC + + """ + logger.warning("SLABDRIVER \tUnBindObjectFromPeer EMPTY-\ + DO NOTHING \r\n ") + return + + #TODO Is BindObjectToPeer still necessary ? Currently does nothing + #04/07/2012 SA + def BindObjectToPeer(self, auth, object_type, object_id, shortname=None, \ + remote_object_id=None): + """This method is a hopefully temporary hack to let the sfa correctly + attach the objects it creates to a remote peer object. This is needed + so that the sfa federation link can work in parallel with RefreshPeer, + as RefreshPeer depends on remote objects being correctly marked. + Parameters: + shortname : string, peer shortname + remote_object_id : int, remote object_id, set to 0 if unknown + FROM PLC API DOC + + """ + logger.warning("SLABDRIVER \tBindObjectToPeer EMPTY - DO NOTHING \r\n ") + return + + #TODO UpdateSlice 04/07/2012 SA + #Funciton should delete and create another job since oin senslab slice=job + def UpdateSlice(self, auth, slice_id_or_name, slice_fields=None): + """Updates the parameters of an existing slice with the values in + slice_fields. + Users may only update slices of which they are members. + PIs may update any of the slices at their sites, or any slices of + which they are members. Admins may update any slice. + Only PIs and admins may update max_nodes. Slices cannot be renewed + (by updating the expires parameter) more than 8 weeks into the future. + Returns 1 if successful, faults otherwise. + FROM PLC API DOC + + """ + logger.warning("SLABDRIVER UpdateSlice EMPTY - DO NOTHING \r\n ") + return + + #TODO UpdatePerson 04/07/2012 SA + def UpdatePerson(self, auth, person_id_or_email, person_fields=None): + """Updates a person. Only the fields specified in person_fields + are updated, all other fields are left untouched. + Users and techs can only update themselves. PIs can only update + themselves and other non-PIs at their sites. + Returns 1 if successful, faults otherwise. + FROM PLC API DOC + + """ + logger.warning("SLABDRIVER UpdatePerson EMPTY - DO NOTHING \r\n ") + return + + #TODO GetKeys 04/07/2012 SA + def GetKeys(self, auth, key_filter=None, return_fields=None): + """Returns an array of structs containing details about keys. + If key_filter is specified and is an array of key identifiers, + or a struct of key attributes, only keys matching the filter + will be returned. If return_fields is specified, only the + specified details will be returned. + + Admin may query all keys. Non-admins may only query their own keys. + FROM PLC API DOC + + """ + logger.warning("SLABDRIVER GetKeys EMPTY - DO NOTHING \r\n ") + return + + #TODO DeleteKey 04/07/2012 SA + def DeleteKey(self, auth, key_id): + """ Deletes a key. + Non-admins may only delete their own keys. + Returns 1 if successful, faults otherwise. + FROM PLC API DOC + + """ + logger.warning("SLABDRIVER DeleteKey EMPTY - DO NOTHING \r\n ") + return + + + #TODO : Check rights to delete person + def DeletePerson(self, auth, person_record): + """ Disable an existing account in senslab LDAP. + Users and techs can only delete themselves. PIs can only + delete themselves and other non-PIs at their sites. + ins can delete anyone. + Returns 1 if successful, faults otherwise. + FROM PLC API DOC + + """ + #Disable user account in senslab LDAP + ret = self.ldap.LdapMarkUserAsDeleted(person_record) + logger.warning("SLABDRIVER DeletePerson %s " %(person_record)) + return ret + + #TODO Check DeleteSlice, check rights 05/07/2012 SA + def DeleteSlice(self, auth, slice_record): + """ Deletes the specified slice. + Senslab : Kill the job associated with the slice if there is one + using DeleteSliceFromNodes. + Updates the slice record in slab db to remove the slice nodes. + + Users may only delete slices of which they are members. PIs may + delete any of the slices at their sites, or any slices of which + they are members. Admins may delete any slice. + Returns 1 if successful, faults otherwise. + FROM PLC API DOC + + """ + self.DeleteSliceFromNodes(slice_record) + logger.warning("SLABDRIVER DeleteSlice %s "%(slice_record)) + return + + #TODO AddPerson 04/07/2012 SA + def AddPerson(self, auth, person_fields=None): + """Adds a new account. Any fields specified in person_fields are used, + otherwise defaults are used. + Accounts are disabled by default. To enable an account, + use UpdatePerson(). + Returns the new person_id (> 0) if successful, faults otherwise. + FROM PLC API DOC + + """ + logger.warning("SLABDRIVER AddPerson EMPTY - DO NOTHING \r\n ") + return + + #TODO AddPersonToSite 04/07/2012 SA + def AddPersonToSite (self, auth, person_id_or_email, \ + site_id_or_login_base=None): + """ Adds the specified person to the specified site. If the person is + already a member of the site, no errors are returned. Does not change + the person's primary site. + Returns 1 if successful, faults otherwise. + FROM PLC API DOC + + """ + logger.warning("SLABDRIVER AddPersonToSite EMPTY - DO NOTHING \r\n ") + return + + #TODO AddRoleToPerson : Not sure if needed in senslab 04/07/2012 SA + def AddRoleToPerson(self, auth, role_id_or_name, person_id_or_email): + """Grants the specified role to the person. + PIs can only grant the tech and user roles to users and techs at their + sites. Admins can grant any role to any user. + Returns 1 if successful, faults otherwise. + FROM PLC API DOC + + """ + logger.warning("SLABDRIVER AddRoleToPerson EMPTY - DO NOTHING \r\n ") + return + + #TODO AddPersonKey 04/07/2012 SA + def AddPersonKey(self, auth, person_id_or_email, key_fields=None): + """Adds a new key to the specified account. + Non-admins can only modify their own keys. + Returns the new key_id (> 0) if successful, faults otherwise. + FROM PLC API DOC + + """ + logger.warning("SLABDRIVER AddPersonKey EMPTY - DO NOTHING \r\n ") + return + + def DeleteLeases(self, leases_id_list, slice_hrn ): + for job_id in leases_id_list: + self.DeleteJobs(job_id, slice_hrn) + + logger.debug("SLABDRIVER DeleteLeases leases_id_list %s slice_hrn %s \r\n " %(leases_id_list, slice_hrn)) + return