X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=sfa%2Fopenstack%2Fosaggregate.py;h=0a87704ef33dad35aaf81b2d331cb3050218d78d;hb=1505f3a42ea06ef30959e9e54fa7298eb55d83ee;hp=cb40a081b4e5ab4f35c5ad866e351262d006d229;hpb=dbe9e9bbf064bf4a21b1d60e725f6d690127c548;p=sfa.git diff --git a/sfa/openstack/osaggregate.py b/sfa/openstack/osaggregate.py index cb40a081..0a87704e 100644 --- a/sfa/openstack/osaggregate.py +++ b/sfa/openstack/osaggregate.py @@ -3,11 +3,12 @@ import os import socket import base64 import string -import random +import random +import time from collections import defaultdict from nova.exception import ImageNotFound from nova.api.ec2.cloud import CloudController -from sfa.util.faults import SfaAPIError, SliverDoesNotExist +from sfa.util.faults import SliverDoesNotExist from sfa.util.sfatime import utcparse, datetime_to_string, datetime_to_epoch from sfa.rspecs.rspec import RSpec from sfa.rspecs.elements.hardware_type import HardwareType @@ -17,13 +18,16 @@ from sfa.rspecs.elements.login import Login from sfa.rspecs.elements.disk_image import DiskImage from sfa.rspecs.elements.services import Services from sfa.rspecs.elements.interface import Interface +from sfa.rspecs.elements.fw_rule import FWRule from sfa.util.xrn import Xrn -from sfa.planetlab.plxrn import PlXrn +from sfa.planetlab.plxrn import PlXrn from sfa.openstack.osxrn import OSXrn, hrn_to_os_slicename from sfa.rspecs.version_manager import VersionManager from sfa.openstack.security_group import SecurityGroup +from sfa.client.multiclient import MultiClient from sfa.util.sfalogging import logger + def pubkeys_to_user_data(pubkeys): user_data = "#!/bin/bash\n\n" for pubkey in pubkeys: @@ -34,27 +38,16 @@ def pubkeys_to_user_data(pubkeys): user_data += "\n" return user_data -def instance_to_sliver(instance, xrn=None): - sliver_urn = None - if slice_xrn: - sliver_xrn = Xrn(xrn=slice_xrn, type='slice', id=instance.id).get_urn() - - sliver = Sliver({'slice_id': sliver_urn, - 'name': instance.name, - 'type': instance.name, - 'cpus': str(instance.vcpus), - 'memory': str(instance.ram), - 'storage': str(instance.disk)}) - return sliver def image_to_rspec_disk_image(image): img = DiskImage() img['name'] = image['name'] img['description'] = image['name'] img['os'] = image['name'] - img['version'] = image['name'] + img['version'] = image['name'] return img - + + class OSAggregate: def __init__(self, driver): @@ -68,38 +61,53 @@ class OSAggregate: zones = [zone.name for zone in zones] return zones - - def list_resources(self, version=None, options={}): + def list_resources(self, version=None, options=None): + if options is None: + options = {} version_manager = VersionManager() version = version_manager.get_version(version) - rspec_version = version_manager._get_version(version.type, version.version, 'ad') + rspec_version = version_manager._get_version( + version.type, version.version, 'ad') rspec = RSpec(version=version, user_options=options) nodes = self.get_aggregate_nodes() rspec.version.add_nodes(nodes) return rspec.toxml() - def describe(self, urns, version=None, options={}): + def describe(self, urns, version=None, options=None): + if options is None: + options = {} # update nova connection tenant_name = OSXrn(xrn=urns[0], type='slice').get_tenant_name() self.driver.shell.nova_manager.connect(tenant=tenant_name) instances = self.get_instances(urns) - if len(instances) == 0: - raise SliverDoesNotExist("You have not allocated any slivers here") + # lookup the sliver allocations + sliver_ids = [sliver['sliver_id'] for sliver in slivers] + constraint = SliverAllocation.sliver_id.in_(sliver_ids) + sliver_allocations = self.driver.api.dbsession().query( + SliverAllocation).filter(constraint) + sliver_allocation_dict = {} + for sliver_allocation in sliver_allocations: + sliver_allocation_dict[ + sliver_allocation.sliver_id] = sliver_allocation geni_slivers = [] rspec_nodes = [] for instance in instances: rspec_nodes.append(self.instance_to_rspec_node(instance)) - geni_slivers.append(self.instance_to_geni_sliver(instance)) + geni_sliver = self.instance_to_geni_sliver( + instance, sliver_sllocation_dict) + geni_slivers.append(geni_sliver) version_manager = VersionManager() version = version_manager.get_version(version) - rspec_version = version_manager._get_version(version.type, version.version, 'manifest') - rspec = RSpec(version=version, user_options=options) + rspec_version = version_manager._get_version( + version.type, version.version, 'manifest') + rspec = RSpec(version=rspec_version, user_options=options) + rspec.xml.set('expires', datetime_to_string(utcparse(time.time()))) rspec.version.add_nodes(rspec_nodes) result = {'geni_urn': Xrn(urns[0]).get_urn(), - 'geni_rspec': rspec.toxml(), + 'geni_rspec': rspec.toxml(), 'geni_slivers': geni_slivers} - + return result def get_instances(self, urns): @@ -108,19 +116,20 @@ class OSAggregate: ids = set() for urn in urns: xrn = OSXrn(xrn=urn) - names.add(xrn.get_slice_name()) - if xrn.id: - ids.add(xrn.id) + if xrn.type == 'slice': + names.add(xrn.get_slice_name()) + elif xrn.type == 'sliver': + ids.add(xrn.leaf) # look up instances instances = [] - for name in name: - servers = self.driver.shell.nova_manager.servers.findall(name=name) - instances.extend(servers) - - # filter on id + filter = {} + if names: + filter['name'] = names if ids: - instances = [server in servers if server.id in ids] + filter['id'] = ids + servers = self.driver.shell.nova_manager.servers.findall(**filter) + instances.extend(servers) return instances @@ -135,86 +144,127 @@ class OSAggregate: rspec_node = Node() rspec_node['component_id'] = node_xrn.urn rspec_node['component_name'] = node_xrn.name - rspec_node['component_manager_id'] = Xrn(self.driver.hrn, 'authority+cm').get_urn() + rspec_node['component_manager_id'] = Xrn( + self.driver.hrn, 'authority+cm').get_urn() + rspec_node['sliver_id'] = OSXrn( + name=instance.name, type='slice', id=instance.id).get_urn() if instance.metadata.get('client_id'): rspec_node['client_id'] = instance.metadata.get('client_id') - flavor = self.driver.shell.nova_manager.flavors.find(id=instance.flavor['id']) - rspec_node['slivers'] = [self.instance_to_sliver(flavor)] - image = self.driver.shell.image_manager.get_images(id=instance.image['id']) + + # get sliver details + flavor = self.driver.shell.nova_manager.flavors.find( + id=instance.flavor['id']) + sliver = self.instance_to_sliver(flavor) + # get firewall rules + fw_rules = [] + group_name = instance.metadata.get('security_groups') + if group_name: + group = self.driver.shell.nova_manager.security_groups.find( + name=group_name) + for rule in group.rules: + port_range = "%s:%s" % (rule['from_port'], rule['to_port']) + fw_rule = FWRule({'protocol': rule['ip_protocol'], + 'port_range': port_range, + 'cidr_ip': rule['ip_range']['cidr']}) + fw_rules.append(fw_rule) + sliver['fw_rules'] = fw_rules + rspec_node['slivers'] = [sliver] + + # get disk image + image = self.driver.shell.image_manager.get_images( + id=instance.image['id']) if isinstance(image, list) and len(image) > 0: image = image[0] disk_image = image_to_rspec_disk_image(image) sliver['disk_image'] = [disk_image] - # build interfaces + # get interfaces rspec_node['services'] = [] rspec_node['interfaces'] = [] addresses = instance.addresses - # HACK: public ips are stored in the list of private, but - # this seems wrong. Assume pub ip is the last in the list of - # private ips until openstack bug is fixed. + # HACK: public ips are stored in the list of private, but + # this seems wrong. Assume pub ip is the last in the list of + # private ips until openstack bug is fixed. if addresses.get('private'): login = Login({'authentication': 'ssh-keys', 'hostname': addresses.get('private')[-1]['addr'], - 'port':'22', 'username': 'root'}) + 'port': '22', 'username': 'root'}) service = Services({'login': login}) - rspec_node['services'].append(service) - + rspec_node['services'].append(service) + for private_ip in addresses.get('private', []): - if_xrn = PlXrn(auth=self.driver.hrn, - interface='node%s:eth0' % (instance.hostId)) - interface = Interface({'component_id': if_xrn.urn}) - interface['ips'] = [{'address': private_ip['addr'], + if_xrn = PlXrn(auth=self.driver.hrn, + interface='node%s' % (instance.hostId)) + if_client_id = Xrn(if_xrn.urn, type='interface', + id="eth%s" % if_index).urn + if_sliver_id = Xrn( + rspec_node['sliver_id'], type='slice', id="eth%s" % if_index).urn + interface = Interface({'component_id': if_xrn.urn, + 'client_id': if_client_id, + 'sliver_id': if_sliver_id}) + interface['ips'] = [{'address': private_ip['addr'], #'netmask': private_ip['network'], 'type': private_ip['version']}] - rspec_node['interfaces'].append(interface) - + rspec_node['interfaces'].append(interface) + # slivers always provide the ssh service for public_ip in addresses.get('public', []): - login = Login({'authentication': 'ssh-keys', - 'hostname': public_ip['addr'], - 'port':'22', 'username': 'root'}) + login = Login({'authentication': 'ssh-keys', + 'hostname': public_ip['addr'], + 'port': '22', 'username': 'root'}) service = Services({'login': login}) rspec_node['services'].append(service) return rspec_node - def instance_to_sliver(self, instance, xrn=None): if xrn: - xrn = Xrn(xrn=slice_xrn, type='slice', id=instance.id).get_urn() + sliver_hrn = '%s.%s' % (self.driver.hrn, instance.id) + sliver_id = Xrn(sliver_hrn, type='sliver').urn - sliver = Sliver({'sliver_id': xrn.get_urn(), + sliver = Sliver({'sliver_id': sliver_id, 'name': instance.name, 'type': instance.name, 'cpus': str(instance.vcpus), 'memory': str(instance.ram), 'storage': str(instance.disk)}) - return sliver - - def instance_to_geni_sliver(self, instance): - op_status = "geni_unknown" - state = instance.state.lower() - if state == 'active': - op_status = 'geni_ready' - elif state == 'building': - op_status = 'geni_configuring' - elif state == 'failed': - op_status =' geni_failed' - - urn = OSXrn(name=instance.name, type='slice', id=instance.id).get_urn() + return sliver + + def instance_to_geni_sliver(self, instance, sliver_allocations=None): + if sliver_allocations is None: + sliver_allocations = {} + sliver_hrn = '%s.%s' % (self.driver.hrn, instance.id) + sliver_id = Xrn(sliver_hrn, type='sliver').urn + + # set sliver allocation and operational status + sliver_allocation = sliver_allocations[sliver_id] + if sliver_allocation: + allocation_status = sliver_allocation.allocation_state + if allocation_status == 'geni_allocated': + op_status = 'geni_pending_allocation' + elif allocation_status == 'geni_provisioned': + state = instance.state.lower() + if state == 'active': + op_status = 'geni_ready' + elif state == 'building': + op_status = 'geni_notready' + elif state == 'failed': + op_status = ' geni_failed' + else: + op_status = 'geni_unknown' + else: + allocation_status = 'geni_unallocated' # required fields - geni_sliver = {'geni_sliver_urn': urn, + geni_sliver = {'geni_sliver_urn': sliver_id, 'geni_expires': None, - 'geni_allocation_status': 'geni_provisioned', + 'geni_allocation_status': allocation_status, 'geni_operational_status': op_status, 'geni_error': None, 'plos_created_at': datetime_to_string(utcparse(instance.created)), 'plos_sliver_type': self.shell.nova_manager.flavors.find(id=instance.flavor['id']).name, - } - + } return geni_sliver - + def get_aggregate_nodes(self): zones = self.get_availability_zones() # available sliver/instance/vm types @@ -223,44 +273,50 @@ class OSAggregate: instances = instances.values() # available images images = self.driver.shell.image_manager.get_images_detailed() - disk_images = [image_to_rspec_disk_image(img) for img in images if img['container_format'] in ['ami', 'ovf']] + disk_images = [image_to_rspec_disk_image(img) for img in images if img[ + 'container_format'] in ['ami', 'ovf']] rspec_nodes = [] for zone in zones: rspec_node = Node() xrn = OSXrn(zone, type='node') rspec_node['component_id'] = xrn.urn rspec_node['component_name'] = xrn.name - rspec_node['component_manager_id'] = Xrn(self.driver.hrn, 'authority+cm').get_urn() + rspec_node['component_manager_id'] = Xrn( + self.driver.hrn, 'authority+cm').get_urn() rspec_node['exclusive'] = 'false' rspec_node['hardware_types'] = [HardwareType({'name': 'plos-pc'}), - HardwareType({'name': 'pc'})] + HardwareType({'name': 'pc'})] slivers = [] for instance in instances: sliver = self.instance_to_sliver(instance) sliver['disk_image'] = disk_images slivers.append(sliver) - + rspec_node['available'] = 'true' rspec_node['slivers'] = slivers - rspec_nodes.append(rspec_node) + rspec_nodes.append(rspec_node) - return rspec_nodes + return rspec_nodes def create_tenant(self, tenant_name): - tenants = self.driver.shell.auth_manager.tenants.findall(name=tenant_name) + tenants = self.driver.shell.auth_manager.tenants.findall( + name=tenant_name) if not tenants: - self.driver.shell.auth_manager.tenants.create(tenant_name, tenant_name) - tenant = self.driver.shell.auth_manager.tenants.find(name=tenant_name) + self.driver.shell.auth_manager.tenants.create( + tenant_name, tenant_name) + tenant = self.driver.shell.auth_manager.tenants.find( + name=tenant_name) else: tenant = tenants[0] return tenant - + def create_instance_key(self, slice_hrn, user): slice_name = Xrn(slice_hrn).leaf user_name = Xrn(user['urn']).leaf key_name = "%s_%s" % (slice_name, user_name) pubkey = user['keys'][0] key_found = False - existing_keys = self.driver.shell.nova_manager.keypairs.findall(name=key_name) + existing_keys = self.driver.shell.nova_manager.keypairs.findall( + name=key_name) for existing_key in existing_keys: if existing_key.public_key != pubkey: self.driver.shell.nova_manager.keypairs.delete(existing_key) @@ -269,119 +325,140 @@ class OSAggregate: if not key_found: self.driver.shell.nova_manager.keypairs.create(key_name, pubkey) - return key_name - + return key_name - def create_security_group(self, slicename, fw_rules=[]): + def create_security_group(self, slicename, fw_rules=None): + if fw_rules is None: + fw_rules = [] # use default group by default - group_name = 'default' + group_name = 'default' if isinstance(fw_rules, list) and fw_rules: # Each sliver get's its own security group. # Keep security group names unique by appending some random # characters on end. - random_name = "".join([random.choice(string.letters+string.digits) - for i in xrange(6)]) - group_name = slicename + random_name + random_name = "".join([random.choice(string.letters + string.digits) + for i in xrange(6)]) + group_name = slicename + random_name security_group = SecurityGroup(self.driver) security_group.create_security_group(group_name) for rule in fw_rules: - security_group.add_rule_to_group(group_name, - protocol = rule.get('protocol'), - cidr_ip = rule.get('cidr_ip'), - port_range = rule.get('port_range'), - icmp_type_code = rule.get('icmp_type_code')) + security_group.add_rule_to_group(group_name, + protocol=rule.get('protocol'), + cidr_ip=rule.get('cidr_ip'), + port_range=rule.get( + 'port_range'), + icmp_type_code=rule.get('icmp_type_code')) + # Open ICMP by default + security_group.add_rule_to_group(group_name, + protocol="icmp", + cidr_ip="0.0.0.0/0", + icmp_type_code="-1:-1") return group_name def add_rule_to_security_group(self, group_name, **kwds): security_group = SecurityGroup(self.driver) - security_group.add_rule_to_group(group_name=group_name, - protocol=kwds.get('protocol'), - cidr_ip =kwds.get('cidr_ip'), - icmp_type_code = kwds.get('icmp_type_code')) - - + security_group.add_rule_to_group(group_name=group_name, + protocol=kwds.get('protocol'), + cidr_ip=kwds.get('cidr_ip'), + icmp_type_code=kwds.get('icmp_type_code')) def run_instances(self, instance_name, tenant_name, rspec, key_name, pubkeys): - #logger.debug('Reserving an instance: image: %s, flavor: ' \ + # logger.debug('Reserving an instance: image: %s, flavor: ' \ # '%s, key: %s, name: %s' % \ # (image_id, flavor_id, key_name, slicename)) # make sure a tenant exists for this slice - tenant = self.create_tenant(tenant_name) + tenant = self.create_tenant(tenant_name) # add the sfa admin user to this tenant and update our nova client connection # to use these credentials for the rest of this session. This emsures that the instances # we create will be assigned to the correct tenant. - sfa_admin_user = self.driver.shell.auth_manager.users.find(name=self.driver.shell.auth_manager.opts['OS_USERNAME']) + sfa_admin_user = self.driver.shell.auth_manager.users.find( + name=self.driver.shell.auth_manager.opts['OS_USERNAME']) user_role = self.driver.shell.auth_manager.roles.find(name='user') admin_role = self.driver.shell.auth_manager.roles.find(name='admin') - self.driver.shell.auth_manager.roles.add_user_role(sfa_admin_user, admin_role, tenant) - self.driver.shell.auth_manager.roles.add_user_role(sfa_admin_user, user_role, tenant) - self.driver.shell.nova_manager.connect(tenant=tenant.name) + self.driver.shell.auth_manager.roles.add_user_role( + sfa_admin_user, admin_role, tenant) + self.driver.shell.auth_manager.roles.add_user_role( + sfa_admin_user, user_role, tenant) + self.driver.shell.nova_manager.connect(tenant=tenant.name) authorized_keys = "\n".join(pubkeys) files = {'/root/.ssh/authorized_keys': authorized_keys} rspec = RSpec(rspec) requested_instances = defaultdict(list) + # iterate over clouds/zones/nodes + slivers = [] for node in rspec.version.get_nodes_with_slivers(): instances = node.get('slivers', []) if not instances: continue for instance in instances: - try: + try: metadata = {} - flavor_id = self.driver.shell.nova_manager.flavors.find(name=instance['name']) + flavor_id = self.driver.shell.nova_manager.flavors.find(name=instance[ + 'name']) image = instance.get('disk_image') if image and isinstance(image, list): image = image[0] - image_id = self.driver.shell.nova_manager.images.find(name=image['name']) + else: + raise InvalidRSpec( + "Must specify a disk_image for each VM") + image_id = self.driver.shell.nova_manager.images.find(name=image[ + 'name']) fw_rules = instance.get('fw_rules', []) - group_name = self.create_security_group(instance_name, fw_rules) + group_name = self.create_security_group( + instance_name, fw_rules) metadata['security_groups'] = group_name if node.get('component_id'): metadata['component_id'] = node['component_id'] if node.get('client_id'): - metadata['client_id'] = node['client_id'] - self.driver.shell.nova_manager.servers.create(flavor=flavor_id, - image=image_id, - key_name = key_name, - security_groups = [group_name], - files=files, - meta=metadata, - name=instance_name) - except Exception, err: - logger.log_exc(err) - - - - def delete_instance(self, tenant_name, instance_name, id=None): - - def _delete_security_group(instance): - security_group = instance.metadata.get('security_groups', '') + metadata['client_id'] = node['client_id'] + server = self.driver.shell.nova_manager.servers.create( + flavor=flavor_id, + image=image_id, + key_name=key_name, + security_groups=[group_name], + files=files, + meta=metadata, + name=instance_name) + slivers.append(server) + except Exception as err: + logger.log_exc(err) + + return slivers + + def delete_instance(self, instance): + + def _delete_security_group(inst): + security_group = inst.metadata.get('security_groups', '') if security_group: manager = SecurityGroup(self.driver) - timeout = 10.0 # wait a maximum of 10 seconds before forcing the security group delete + timeout = 10.0 # wait a maximum of 10 seconds before forcing the security group delete start_time = time.time() instance_deleted = False while instance_deleted == False and (time.time() - start_time) < timeout: - inst = self.driver.shell.nova_manager.servers.findall(id=instance.id) - if not inst: + tmp_inst = self.driver.shell.nova_manager.servers.findall( + id=inst.id) + if not tmp_inst: instance_deleted = True + time.sleep(.5) manager.delete_security_group(security_group) - thread_manager = ThreadManager() - self.driver.shell.nova_manager.connect(tenant=tenant_name) - args = {'name': instance_name} - if id: - args['id'] = id + multiclient = MultiClient() + tenant = self.driver.shell.auth_manager.tenants.find( + id=instance.tenant_id) + self.driver.shell.nova_manager.connect(tenant=tenant.name) + args = {'name': instance.name, + 'id': instance.id} instances = self.driver.shell.nova_manager.servers.findall(**args) security_group_manager = SecurityGroup(self.driver) for instance in instances: # destroy instance self.driver.shell.nova_manager.servers.delete(instance) # deleate this instance's security groups - thread_manager.run(_delete_security_group, instance) + multiclient.run(_delete_security_group, instance) return 1 def stop_instances(self, instance_name, tenant_name, id=None): @@ -407,7 +484,7 @@ class OSAggregate: def restart_instances(self, instacne_name, tenant_name, id=None): self.stop_instances(instance_name, tenant_name, id) self.start_instances(instance_name, tenant_name, id) - return 1 + return 1 def update_instances(self, project_name): pass