X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=sfa%2Fmanagers%2Faggregate_manager_pl.py;h=75fdc8b68ee859e3b60ee7e22a5aed83d2d912d5;hb=613eea54a3317606aa6541b5e150f02638bb947b;hp=0c70ddaeb9ec76cf4f860754e13c7282bd68f7ba;hpb=ba2aaa438f939a4b5c697052e37b1c3218901319;p=sfa.git diff --git a/sfa/managers/aggregate_manager_pl.py b/sfa/managers/aggregate_manager_pl.py index 0c70ddae..75fdc8b6 100644 --- a/sfa/managers/aggregate_manager_pl.py +++ b/sfa/managers/aggregate_manager_pl.py @@ -4,11 +4,10 @@ import traceback import sys import re from types import StringTypes -from dateutil.parser import parse from sfa.util.faults import * -from sfa.util.xrn import get_authority, hrn_to_urn, urn_to_hrn -from sfa.util.plxrn import slicename_to_hrn, hrn_to_pl_slicename +from sfa.util.xrn import get_authority, hrn_to_urn, urn_to_hrn, Xrn +from sfa.util.plxrn import slicename_to_hrn, hrn_to_pl_slicename, hostname_to_urn from sfa.util.rspec import * from sfa.util.specdict import * from sfa.util.record import SfaRecord @@ -20,8 +19,23 @@ from sfa.trust.credential import Credential import sfa.plc.peers as peers from sfa.plc.network import * from sfa.plc.api import SfaAPI +from sfa.plc.aggregate import Aggregate from sfa.plc.slices import * - +from sfa.util.version import version_core +from sfa.rspecs.rspec_version import RSpecVersion +from sfa.rspecs.rspec_parser import parse_rspec +from sfa.util.sfatime import utcparse +from sfa.util.callids import Callids + +def GetVersion(api): + xrn=Xrn(api.hrn) + return version_core({'interface':'aggregate', + 'testbed':'myplc', + 'hrn':xrn.get_hrn(), + 'input_rspec' : ['PG 2', 'SFA 1'], + 'output_rspec' : ["SFA 1"], + 'ad_rspec' : ["PG 2", "SFA 1"], + }) def __get_registry_objects(slice_xrn, creds, users): """ @@ -42,7 +56,6 @@ def __get_registry_objects(slice_xrn, creds, users): slicename = hrn_to_pl_slicename(hrn) login_base = slicename.split('_')[0] reg_objects = {} - site = {} site['site_id'] = 0 site['name'] = 'geni.%s' % login_base @@ -58,7 +71,12 @@ def __get_registry_objects(slice_xrn, creds, users): reg_objects['site'] = site slice = {} - slice['expires'] = int(time.mktime(Credential(string=creds[0]).get_expiration().timetuple())) + + extime = Credential(string=creds[0]).get_expiration() + # If the expiration time is > 60 days from now, set the expiration time to 60 days from now + if extime > datetime.datetime.utcnow() + datetime.timedelta(days=60): + extime = datetime.datetime.utcnow() + datetime.timedelta(days=60) + slice['expires'] = int(time.mktime(extime.timetuple())) slice['hrn'] = hrn slice['name'] = hrn_to_pl_slicename(hrn) slice['url'] = hrn @@ -83,14 +101,10 @@ def __get_hostnames(nodes): hostnames.append(node.hostname) return hostnames -def get_version(): - version = {} - version['geni_api'] = 1 - version['sfa'] = 1 - return version +def SliverStatus(api, slice_xrn, creds, call_id): + if Callids().already_handled(call_id): return {} -def slice_status(api, slice_xrn, creds): - hrn, type = urn_to_hrn(slice_xrn) + (hrn, type) = urn_to_hrn(slice_xrn) # find out where this slice is currently running api.logger.info(hrn) slicename = hrn_to_pl_slicename(hrn) @@ -100,99 +114,113 @@ def slice_status(api, slice_xrn, creds): raise Exception("Slice %s not found (used %s as slicename internally)" % slice_xrn, slicename) slice = slices[0] - nodes = api.plshell.GetNodes(api.plauth, slice['node_ids'], - ['hostname', 'boot_state', 'last_contact']) - api.logger.info(slice) - api.logger.info(nodes) - + # report about the local nodes only + nodes = api.plshell.GetNodes(api.plauth, {'node_id':slice['node_ids'],'peer_id':None}, + ['hostname', 'site_id', 'boot_state', 'last_contact']) + site_ids = [node['site_id'] for node in nodes] + sites = api.plshell.GetSites(api.plauth, site_ids, ['site_id', 'login_base']) + sites_dict = dict ( [ (site['site_id'],site['login_base'] ) for site in sites ] ) + result = {} - result['geni_urn'] = slice_xrn - result['geni_status'] = 'unknown' + top_level_status = 'unknown' + if nodes: + top_level_status = 'ready' + result['geni_urn'] = Xrn(slice_xrn, 'slice').get_urn() result['pl_login'] = slice['name'] - result['pl_expires'] = slice['expires'] + result['pl_expires'] = datetime.datetime.fromtimestamp(slice['expires']).ctime() resources = [] - for node in nodes: res = {} res['pl_hostname'] = node['hostname'] res['pl_boot_state'] = node['boot_state'] res['pl_last_contact'] = node['last_contact'] - res['geni_urn'] = '' - res['geni_status'] = 'unknown' + if node['last_contact'] is not None: + res['pl_last_contact'] = datetime.datetime.fromtimestamp(node['last_contact']).ctime() + res['geni_urn'] = hostname_to_urn(api.hrn, sites_dict[node['site_id']], node['hostname']) + if node['boot_state'] == 'boot': + res['geni_status'] = 'ready' + else: + res['geni_status'] = 'failed' + top_level_staus = 'failed' + res['geni_error'] = '' resources.append(res) + result['geni_status'] = top_level_status result['geni_resources'] = resources + # XX remove me + #api.logger.info(result) + # XX remove me return result -def create_slice(api, slice_xrn, creds, rspec, users): +def CreateSliver(api, slice_xrn, creds, rspec_string, users, call_id): """ Create the sliver[s] (slice) at this aggregate. Verify HRN and initialize the slice record in PLC if necessary. """ + if Callids().already_handled(call_id): return "" reg_objects = __get_registry_objects(slice_xrn, creds, users) - hrn, type = urn_to_hrn(slice_xrn) + (hrn, type) = urn_to_hrn(slice_xrn) peer = None + aggregate = Aggregate(api) slices = Slices(api) peer = slices.get_peer(hrn) sfa_peer = slices.get_sfa_peer(hrn) registry = api.registries[api.hrn] credential = api.getCredential() - site_id, remote_site_id = slices.verify_site(registry, credential, hrn, - peer, sfa_peer, reg_objects) + (site_id, remote_site_id) = slices.verify_site(registry, credential, hrn, + peer, sfa_peer, reg_objects) - slice_record = slices.verify_slice(registry, credential, hrn, site_id, - remote_site_id, peer, sfa_peer, reg_objects) + slice = slices.verify_slice(registry, credential, hrn, site_id, + remote_site_id, peer, sfa_peer, reg_objects) - network = Network(api) - - slice = network.get_slice(api, hrn) - slice.peer_id = slice_record['peer_slice_id'] - current = __get_hostnames(slice.get_nodes()) - - network.addRSpec(rspec, api.config.SFA_AGGREGATE_RSPEC_SCHEMA) - request = __get_hostnames(network.nodesWithSlivers()) - + nodes = api.plshell.GetNodes(api.plauth, slice['node_ids'], ['hostname']) + current_slivers = [node['hostname'] for node in nodes] + rspec = parse_rspec(rspec_string) + requested_slivers = [str(host) for host in rspec.get_nodes_with_slivers()] # remove nodes not in rspec - deleted_nodes = list(set(current).difference(request)) + deleted_nodes = list(set(current_slivers).difference(requested_slivers)) # add nodes from rspec - added_nodes = list(set(request).difference(current)) + added_nodes = list(set(requested_slivers).difference(current_slivers)) try: if peer: - api.plshell.UnBindObjectFromPeer(api.plauth, 'slice', slice.id, peer) + api.plshell.UnBindObjectFromPeer(api.plauth, 'slice', slice['slice_id'], peer) - api.plshell.AddSliceToNodes(api.plauth, slice.name, added_nodes) - api.plshell.DeleteSliceFromNodes(api.plauth, slice.name, deleted_nodes) + api.plshell.AddSliceToNodes(api.plauth, slice['name'], added_nodes) + api.plshell.DeleteSliceFromNodes(api.plauth, slice['name'], deleted_nodes) - network.updateSliceTags() + # TODO: update slice tags + #network.updateSliceTags() finally: if peer: api.plshell.BindObjectToPeer(api.plauth, 'slice', slice.id, peer, slice.peer_id) - # print network.toxml() - - return True + return aggregate.get_rspec(slice_xrn=slice_xrn) -def renew_slice(api, xrn, creds, expiration_time): - hrn, type = urn_to_hrn(xrn) +def RenewSliver(api, xrn, creds, expiration_time, call_id): + if Callids().already_handled(call_id): return True + (hrn, type) = urn_to_hrn(xrn) slicename = hrn_to_pl_slicename(hrn) slices = api.plshell.GetSlices(api.plauth, {'name': slicename}, ['slice_id']) if not slices: raise RecordNotFound(hrn) slice = slices[0] - requested_time = parse(expiration_time) + requested_time = utcparse(expiration_time) record = {'expires': int(time.mktime(requested_time.timetuple()))} - api.plshell.UpdateSlice(api.plauth, slice['slice_id'], record) - return 1 + try: + api.plshell.UpdateSlice(api.plauth, slice['slice_id'], record) + return True + except: + return False def start_slice(api, xrn, creds): hrn, type = urn_to_hrn(xrn) @@ -227,8 +255,9 @@ def reset_slice(api, xrn): # XX not implemented at this interface return 1 -def delete_slice(api, xrn, creds): - hrn, type = urn_to_hrn(xrn) +def DeleteSliver(api, xrn, creds, call_id): + if Callids().already_handled(call_id): return "" + (hrn, type) = urn_to_hrn(xrn) slicename = hrn_to_pl_slicename(hrn) slices = api.plshell.GetSlices(api.plauth, {'name': slicename}) if not slices: @@ -246,9 +275,13 @@ def delete_slice(api, xrn, creds): api.plshell.BindObjectToPeer(api.plauth, 'slice', slice['slice_id'], peer, slice['peer_slice_id']) return 1 -def get_slices(api, creds): +# xxx Thierry : caching at the aggregate level sounds wrong... +caching=True +#caching=False +def ListSlices(api, creds, call_id): + if Callids().already_handled(call_id): return [] # look in cache first - if api.cache: + if caching and api.cache: slices = api.cache.get('slices') if slices: return slices @@ -259,32 +292,40 @@ def get_slices(api, creds): slice_urns = [hrn_to_urn(slice_hrn, 'slice') for slice_hrn in slice_hrns] # cache the result - if api.cache: + if caching and api.cache: api.cache.add('slices', slice_urns) return slice_urns -def get_rspec(api, creds, options): +def ListResources(api, creds, options,call_id): + if Callids().already_handled(call_id): return "" # get slice's hrn from options xrn = options.get('geni_slice_urn', '') - hrn, type = urn_to_hrn(xrn) + (hrn, type) = urn_to_hrn(xrn) + # get the rspec's return format from options + rspec_version = RSpecVersion(options.get('rspec_version', 'SFA 1')) + version_string = "rspec_%s_%s" % (rspec_version.format, rspec_version.version) + # look in cache first - if api.cache and not xrn: - rspec = api.cache.get('nodes') + if caching and api.cache and not xrn: + rspec = api.cache.get(version_string) if rspec: + api.logger.info("aggregate.ListResources: returning cached value for hrn %s"%hrn) return rspec - network = Network(api) - if (hrn): - if network.get_slice(api, hrn): - network.addSlice() + aggregate = Aggregate(api) - rspec = network.toxml() + if xrn: + # get this rspec for the specified slice + rspec = aggregate.get_rspec(slice_xrn=hrn, version=rspec_version) + else: + # generate rspec in both pg and sfa formats + rspec = aggregate.get_rspec(version=rspec_version) # cache the result - if api.cache and not xrn: - api.cache.add('nodes', rspec) + if caching and api.cache and not xrn: + api.cache.add(version_string, rspec) return rspec @@ -303,7 +344,7 @@ def get_ticket(api, xrn, creds, rspec, users): credential = api.getCredential() records = registry.Resolve(xrn, credential) - # similar to create_slice, we must verify that the required records exist + # similar to CreateSliver, we must verify that the required records exist # at this aggregate before we can issue a ticket site_id, remote_site_id = slices.verify_site(registry, credential, slice_hrn, peer, sfa_peer, reg_objects) @@ -352,15 +393,15 @@ def get_ticket(api, xrn, creds, rspec, users): def main(): api = SfaAPI() """ - rspec = get_rspec(api, "plc.princeton.sapan", None) - #rspec = get_rspec(api, "plc.princeton.coblitz", None) - #rspec = get_rspec(api, "plc.pl.sirius", None) + rspec = ListResources(api, "plc.princeton.sapan", None, 'pl_test_sapan') + #rspec = ListResources(api, "plc.princeton.coblitz", None, 'pl_test_coblitz') + #rspec = ListResources(api, "plc.pl.sirius", None, 'pl_test_sirius') print rspec """ f = open(sys.argv[1]) xml = f.read() f.close() - create_slice(api, "plc.princeton.sapan", xml) + CreateSliver(api, "plc.princeton.sapan", xml, 'CreateSliver_sapan') if __name__ == "__main__": main()