import sys
+import subprocess
+import datetime
+from time import gmtime, strftime
-from sfa.util.faults import MissingSfaInfo
+from sfa.util.faults import MissingSfaInfo , SliverDoesNotExist
+#from sfa.util.sfatime import datetime_to_string
from sfa.util.sfalogging import logger
from sfa.storage.table import SfaTable
from sfa.util.defaultdict import defaultdict
from sfa.rspecs.version_manager import VersionManager
from sfa.rspecs.rspec import RSpec
-from sfa.util.xrn import hrn_to_urn
+from sfa.util.xrn import hrn_to_urn, urn_to_sliver_id
from sfa.util.plxrn import slicename_to_hrn, hostname_to_hrn, hrn_to_pl_slicename, hrn_to_pl_login_base
## thierry: everything that is API-related (i.e. handling incoming requests)
self.cache=None
+ def sliver_status(self,slice_urn,slice_hrn):
+ # receive a status request for slice named urn/hrn urn:publicid:IDN+senslab+nturro_slice hrn senslab.nturro_slice
+ # shall return a structure as described in
+ # http://groups.geni.net/geni/wiki/GAPI_AM_API_V2#SliverStatus
+ # NT : not sure if we should implement this or not, but used by sface.
+ slices = self.GetSlices([slice_hrn])
+ if len(slices) is 0:
+ raise SliverDoesNotExist("%s slice_hrn" % (slice_hrn))
+ sl = slices[0]
+ print >>sys.stderr, "\r\n \r\n_____________ Sliver status urn %s hrn %s slices %s \r\n " %(slice_urn,slice_hrn,slices)
+ if sl['oar_job_id'] is not -1:
+
+ # report about the local nodes only
+ nodes = self.GetNodes({'hostname':sl['node_ids']},
+ ['node_id', 'hostname','site_login_base','boot_state'])
+ if len(nodes) is 0:
+ raise SliverDoesNotExist("No slivers allocated ")
+
+
+ site_logins = [node['site_login_base'] for node in nodes]
+
+ result = {}
+ top_level_status = 'unknown'
+ if nodes:
+ top_level_status = 'ready'
+ result['geni_urn'] = slice_urn
+ result['pl_login'] = sl['job_user']
+ timestamp = float(sl['startTime']) + float(sl['walltime'])
+ result['pl_expires'] = strftime(self.time_format, gmtime(float(timestamp)))
+
+ resources = []
+ for node in nodes:
+ res = {}
+ res['pl_hostname'] = node['hostname']
+ res['pl_boot_state'] = node['boot_state']
+
+ sliver_id = urn_to_sliver_id(slice_urn, sl['record_id_slice'], node['node_id'])
+ res['geni_urn'] = sliver_id
+ if node['boot_state'] == 'Alive':
+ res['geni_status'] = 'ready'
+ else:
+ res['geni_status'] = 'failed'
+ top_level_status = 'failed'
+
+ res['geni_error'] = ''
+
+ resources.append(res)
+
+ result['geni_status'] = top_level_status
+ result['geni_resources'] = resources
+ print >>sys.stderr, "\r\n \r\n_____________ Sliver status resources %s res %s \r\n " %(resources,res)
+ return result
+
+
def create_sliver (self, slice_urn, slice_hrn, creds, rspec_string, users, options):
aggregate = SlabAggregate(self)
#aggregate = SlabAggregate(self)
#panos: passing user-defined options
#print "manager options = ",options
aggregate = SlabAggregate(self)
+ origin_hrn = Credential(string=creds[0]).get_gid_caller().get_hrn()
+ print>>sys.stderr, " \r\n \r\n \t SLABDRIVER get_rspec origin_hrn %s" %(origin_hrn)
+ options.update({'origin_hrn':origin_hrn})
+ print>>sys.stderr, " \r\n \r\n \t SLABDRIVER get_rspec options %s" %(options)
rspec = aggregate.get_rspec(slice_xrn=slice_urn, version=rspec_version,
options=options)
return_person_list = parse_filter(person_list,person_filter ,'persons', return_fields)
if return_person_list:
- print>>sys.stderr, " \r\n GetPersons person_filter %s return_fields %s return_person_list %s " %(person_filter,return_fields,return_person_list)
+ print>>sys.stderr, " \r\n GetPersons person_filter %s return_fields %s " %(person_filter,return_fields)
return return_person_list
+ def GetTimezone(self):
+ time = self.oar.parser.SendRequest("GET_timezone")
+ return time
+
- def GetJobs(self,job_id= None, resources=True,return_fields=None, details = None):
+ def DeleteJobs(self, job_id, username):
+ if not job_id:
+ return
+ reqdict = {}
+ reqdict['method'] = "delete"
+ reqdict['strval'] = str(job_id)
+ answer = self.oar.POSTRequestToOARRestAPI('DELETE_jobs_id',reqdict,username)
+ print>>sys.stderr, "\r\n \r\n jobid DeleteJobs %s " %(answer)
+
+
+ def GetJobs(self,job_id= None, resources=True,return_fields=None, username = None):
#job_resources=['reserved_resources', 'assigned_resources','job_id', 'job_uri', 'assigned_nodes',\
#'api_timestamp']
#assigned_res = ['resource_id', 'resource_uri']
#assigned_n = ['node', 'node_uri']
-
+
if job_id and resources is False:
- job_info = self.oar.parser.SendRequest("GET_jobs_id", job_id)
- print>>sys.stderr, "\r\n \r\n \t\t GetJobs resources is False job_info %s" %(job_info)
+ req = "GET_jobs_id"
+ node_list_k = 'assigned_network_address'
+
+ if job_id and resources :
+ req = "GET_jobs_id_resources"
+ node_list_k = 'reserved_resources'
+
+
+
+ #Get job info from OAR
+ job_info = self.oar.parser.SendRequest(req, job_id, username)
+ print>>sys.stderr, "\r\n \r\n \t\t GetJobs %s " %(job_info)
+
+ if 'state' in job_info :
+ if job_info['state'] == 'Terminated':
+ print>>sys.stderr, "\r\n \r\n \t\t GetJobs TERMINELEBOUSIN "
+ return None
+ if job_info['state'] == 'Error':
+ print>>sys.stderr, "\r\n \r\n \t\t GetJobs ERROR message %s " %(job_info)
+ return None
+
+ #Get a dict of nodes . Key :hostname of the node
+ node_list = self.GetNodes()
+ node_hostname_list = []
+ for node in node_list:
+ node_hostname_list.append(node['hostname'])
+ node_dict = dict(zip(node_hostname_list,node_list))
+
+ #print>>sys.stderr, "\r\n \r\n \r\n \r\n \r\n \t\t GetJobs GetNODES %s " %(node_list)
+ try :
+
+ #for n in job_info[node_list]:
+ #n = str(self.root_auth) + str(n)
- if job_id and resources :
- job_info = self.oar.parser.SendRequest("GET_jobs_id_resources", job_id)
- print>>sys.stderr, "\r\n \r\n \t\t GetJobs job_info %s" %(job_info)
+ liste =job_info[node_list_k]
+ print>>sys.stderr, "\r\n \r\n \t\t GetJobs resources job_info liste%s" %(liste)
+ for k in range(len(liste)):
+ job_info[node_list_k][k] = node_dict[job_info[node_list_k][k]]['hostname']
- if job_info['state'] == 'Terminated':
- print>>sys.stderr, "\r\n \r\n \t\t GetJobs TERMINELEBOUSIN "
- return None
- else:
+ print>>sys.stderr, "\r\n \r\n \t\t YYYYYYYYYYYYGetJobs resources job_info %s" %(job_info)
+ job_info.update({'node_ids':job_info[node_list_k]})
+ del job_info[node_list_k]
return job_info
-
+
+ except KeyError:
+ print>>sys.stderr, "\r\n \r\n \t\t GetJobs KEYERROR "
+
+
+
+
def GetNodes(self,node_filter= None, return_fields=None):
node_dict =self.oar.parser.SendRequest("GET_resources_full")
-
+ print>>sys.stderr, "\r\n \r\n \t\t SLABDRIVER.PY GetNodes "
return_node_list = []
if not (node_filter or return_fields):
return_node_list = node_dict.values()
return_node_list= parse_filter(node_dict.values(),node_filter ,'node', return_fields)
return return_node_list
- #def GetSites(self, auth, site_filter = None, return_fields=None):
- #self.oar.parser.SendRequest("GET_resources_full")
- #site_dict = self.oar.parser.GetSitesFromOARParse()
- #return_site_list = []
- #site = site_dict.values()[0]
- #if not (site_filter or return_fields):
- #return_site_list = site_dict.values()
- #return return_site_list
-
- #return_site_list = parse_filter(site_dict.values(),site_filter ,'site', return_fields)
- #return return_site_list
+
def GetSlices(self,slice_filter = None, return_fields=None):
-
sliceslist = self.db.find('slice',columns = ['oar_job_id', 'slice_hrn', 'record_id_slice','record_id_user'], record_filter=slice_filter)
print >>sys.stderr, " \r\n \r\n \tSLABDRIVER.PY GetSlices slices %s slice_filter %s " %(sliceslist,slice_filter)
return_slice_list = parse_filter(sliceslist, slice_filter,'slice', return_fields)
-
+
if return_slice_list:
for sl in return_slice_list:
+ login = sl['slice_hrn'].split(".")[1].split("_")[0]
+ print >>sys.stderr, " \r\n \r\n \tSLABDRIVER.PY GetSlices sl %s " %(sl)
if sl['oar_job_id'] is not -1:
- rslt = self.GetJobs( sl['oar_job_id'],resources=False)
-
+ rslt = self.GetJobs( sl['oar_job_id'],resources=False, username = login )
+ print >>sys.stderr, " \r\n \r\n \tSLABDRIVER.PY GetSlices GetJobs %s " %(rslt)
if rslt :
sl.update(rslt)
sl.update({'hrn':str(sl['slice_hrn'])})
sl['oar_job_id'] = '-1'
sl.update({'hrn':str(sl['slice_hrn'])})
self.db.update_senslab_slice(sl)
-
-
-
+
print >>sys.stderr, " \r\n \r\n \tSLABDRIVER.PY GetSlices return_slice_list %s" %(return_slice_list)
return return_slice_list
+
-
-
-
-
- return return_slice_list
def testbed_name (self): return "senslab2"
def AddSliceToNodes(self, slice_name, added_nodes, slice_user=None):
- print>>sys.stderr, "\r\n \r\n AddSliceToNodes slice_name %s added_nodes %s username %s" %(slice_name,added_nodes,slice_user )
+
site_list = []
nodeid_list =[]
resource = ""
reqdict['property'] ="network_address in ("
for node in added_nodes:
#Get the ID of the node : remove the root auth and put the site in a separate list
- tmp = node.strip(self.root_auth+".")
- l = tmp.split("_")
-
- nodeid= (l[len(l)-1])
+ s=node.split(".")
+ # NT: it's not clear for me if the nodenames will have the senslab prefix
+ # so lets take the last part only, for now.
+ lastpart=s[-1]
+ #if s[0] == self.root_auth :
+ # Again here it's not clear if nodes will be prefixed with <site>_, lets split and tanke the last part for now.
+ s=lastpart.split("_")
+ nodeid=s[-1]
reqdict['property'] += "'"+ nodeid +"', "
nodeid_list.append(nodeid)
- site_list.append( l[0] )
-
+ #site_list.append( l[0] )
reqdict['property'] = reqdict['property'][0: len( reqdict['property'])-2] +")"
reqdict['resource'] ="network_address="+ str(len(nodeid_list))
- reqdict['resource']+= ",walltime=" + str(00) + ":" + str(05) + ":" + str(00)
- reqdict['script_path'] = "/bin/sleep 320"
- #reqdict['type'] = "deploy"
+ reqdict['resource']+= ",walltime=" + str(00) + ":" + str(12) + ":" + str(20) #+2 min 20
+ reqdict['script_path'] = "/bin/sleep 620" #+20 sec
+ reqdict['type'] = "deploy"
+ reqdict['directory']= ""
+ reqdict['name']= "TestSandrine"
+ timestamp = self.GetTimezone()
+ print>>sys.stderr, "\r\n \r\n AddSliceToNodes slice_name %s added_nodes %s username %s reqdict %s " %(slice_name,added_nodes,slice_user, reqdict)
+ readable_time = strftime(self.time_format, gmtime(float(timestamp)))
+ print >>sys.stderr," \r\n \r\n \t\t\t\t AVANT ParseTimezone readable_time %s timestanp %s " %(readable_time, timestamp )
+ timestamp = timestamp+ 3620 #Add 3 min to server time
+ readable_time = strftime(self.time_format, gmtime(float(timestamp)))
+
+ print >>sys.stderr," \r\n \r\n \t\t\t\tAPRES ParseTimezone readable_time %s timestanp %s " %(readable_time , timestamp)
+ reqdict['reservation'] = readable_time
+
+ # first step : start the OAR job
print>>sys.stderr, "\r\n \r\n AddSliceToNodes reqdict %s \r\n site_list %s" %(reqdict,site_list)
- OAR = OARrestapi()
- answer = OAR.POSTRequestToOARRestAPI('POST_job',reqdict,slice_user)
+ #OAR = OARrestapi()
+ answer = self.oar.POSTRequestToOARRestAPI('POST_job',reqdict,slice_user)
print>>sys.stderr, "\r\n \r\n AddSliceToNodes jobid %s " %(answer)
self.db.update('slice',['oar_job_id'], [answer['id']], 'slice_hrn', slice_name)
+
+ jobid=answer['id']
+ print>>sys.stderr, "\r\n \r\n AddSliceToNodes jobid %s added_nodes %s slice_user %s" %(jobid,added_nodes,slice_user)
+ # second step : configure the experiment
+ # we need to store the nodes in a yaml (well...) file like this :
+ # [1,56,23,14,45,75] with name /tmp/sfa<jobid>.json
+ f=open('/tmp/sfa/'+str(jobid)+'.json','w')
+ f.write('[')
+ f.write(str(added_nodes[0].strip('node')))
+ for node in added_nodes[1:len(added_nodes)] :
+ f.write(','+node.strip('node'))
+ f.write(']')
+ f.close()
+
+ # third step : call the senslab-experiment wrapper
+ #command= "java -jar target/sfa-1.0-jar-with-dependencies.jar "+str(jobid)+" "+slice_user
+ javacmdline="/usr/bin/java"
+ jarname="/opt/senslabexperimentwrapper/sfa-1.0-jar-with-dependencies.jar"
+ #ret=subprocess.check_output(["/usr/bin/java", "-jar", ", str(jobid), slice_user])
+ output = subprocess.Popen([javacmdline, "-jar", jarname, str(jobid), slice_user],stdout=subprocess.PIPE).communicate()[0]
+
+ print>>sys.stderr, "\r\n \r\n AddSliceToNodes wrapper returns %s " %(output)
return