-
-import os
-import socket
-import base64
-import string
-import random
-import time
-from collections import defaultdict
-from nova.exception import ImageNotFound
-from nova.api.ec2.cloud import CloudController
-from sfa.util.faults import SliverDoesNotExist
-from sfa.util.sfatime import utcparse, datetime_to_string, datetime_to_epoch
-from sfa.rspecs.rspec import RSpec
-from sfa.rspecs.elements.hardware_type import HardwareType
-from sfa.rspecs.elements.node import Node
-from sfa.rspecs.elements.sliver import Sliver
-from sfa.rspecs.elements.login import Login
-from sfa.rspecs.elements.disk_image import DiskImage
-from sfa.rspecs.elements.services import Services
-from sfa.rspecs.elements.interface import Interface
-from sfa.rspecs.elements.fw_rule import FWRule
-from sfa.util.xrn import Xrn
-from sfa.planetlab.plxrn import PlXrn
-from sfa.openstack.osxrn import OSXrn, hrn_to_os_slicename
-from sfa.rspecs.version_manager import VersionManager
-from sfa.openstack.security_group import SecurityGroup
-from sfa.client.multiclient import MultiClient
-from sfa.util.sfalogging import logger
-
-
-def pubkeys_to_user_data(pubkeys):
- user_data = "#!/bin/bash\n\n"
- for pubkey in pubkeys:
- pubkey = pubkey.replace('\n', '')
- user_data += "echo %s >> /root/.ssh/authorized_keys" % pubkey
- user_data += "\n"
- user_data += "echo >> /root/.ssh/authorized_keys"
- user_data += "\n"
- return user_data
-
-
-def image_to_rspec_disk_image(image):
- img = DiskImage()
- img['name'] = image['name']
- img['description'] = image['name']
- img['os'] = image['name']
- img['version'] = image['name']
- return img
-
-
-class OSAggregate:
-
- def __init__(self, driver):
- self.driver = driver
-
- def get_availability_zones(self):
- zones = self.driver.shell.nova_manager.dns_domains.domains()
- if not zones:
- zones = ['cloud']
- else:
- zones = [zone.name for zone in zones]
- return zones
-
- def list_resources(self, version=None, options=None):
- if options is None:
- options = {}
- version_manager = VersionManager()
- version = version_manager.get_version(version)
- rspec_version = version_manager._get_version(
- version.type, version.version, 'ad')
- rspec = RSpec(version=version, user_options=options)
- nodes = self.get_aggregate_nodes()
- rspec.version.add_nodes(nodes)
- return rspec.toxml()
-
- def describe(self, urns, version=None, options=None):
- if options is None:
- options = {}
- # update nova connection
- tenant_name = OSXrn(xrn=urns[0], type='slice').get_tenant_name()
- self.driver.shell.nova_manager.connect(tenant=tenant_name)
- instances = self.get_instances(urns)
- # lookup the sliver allocations
- sliver_ids = [sliver['sliver_id'] for sliver in slivers]
- constraint = SliverAllocation.sliver_id.in_(sliver_ids)
- sliver_allocations = self.driver.api.dbsession().query(
- SliverAllocation).filter(constraint)
- sliver_allocation_dict = {}
- for sliver_allocation in sliver_allocations:
- sliver_allocation_dict[
- sliver_allocation.sliver_id] = sliver_allocation
-
- geni_slivers = []
- rspec_nodes = []
- for instance in instances:
- rspec_nodes.append(self.instance_to_rspec_node(instance))
- geni_sliver = self.instance_to_geni_sliver(
- instance, sliver_sllocation_dict)
- geni_slivers.append(geni_sliver)
- version_manager = VersionManager()
- version = version_manager.get_version(version)
- rspec_version = version_manager._get_version(
- version.type, version.version, 'manifest')
- rspec = RSpec(version=rspec_version, user_options=options)
- rspec.xml.set('expires', datetime_to_string(utcparse(time.time())))
- rspec.version.add_nodes(rspec_nodes)
- result = {'geni_urn': Xrn(urns[0]).get_urn(),
- 'geni_rspec': rspec.toxml(),
- 'geni_slivers': geni_slivers}
-
- return result
-
- def get_instances(self, urns):
- # parse slice names and sliver ids
- names = set()
- ids = set()
- for urn in urns:
- xrn = OSXrn(xrn=urn)
- if xrn.type == 'slice':
- names.add(xrn.get_slice_name())
- elif xrn.type == 'sliver':
- ids.add(xrn.leaf)
-
- # look up instances
- instances = []
- filter = {}
- if names:
- filter['name'] = names
- if ids:
- filter['id'] = ids
- servers = self.driver.shell.nova_manager.servers.findall(**filter)
- instances.extend(servers)
-
- return instances
-
- def instance_to_rspec_node(self, instance):
- # determine node urn
- node_xrn = instance.metadata.get('component_id')
- if not node_xrn:
- node_xrn = OSXrn('cloud', type='node')
- else:
- node_xrn = OSXrn(xrn=node_xrn, type='node')
-
- rspec_node = Node()
- rspec_node['component_id'] = node_xrn.urn
- rspec_node['component_name'] = node_xrn.name
- rspec_node['component_manager_id'] = Xrn(
- self.driver.hrn, 'authority+cm').get_urn()
- rspec_node['sliver_id'] = OSXrn(
- name=instance.name, type='slice', id=instance.id).get_urn()
- if instance.metadata.get('client_id'):
- rspec_node['client_id'] = instance.metadata.get('client_id')
-
- # get sliver details
- flavor = self.driver.shell.nova_manager.flavors.find(
- id=instance.flavor['id'])
- sliver = self.instance_to_sliver(flavor)
- # get firewall rules
- fw_rules = []
- group_name = instance.metadata.get('security_groups')
- if group_name:
- group = self.driver.shell.nova_manager.security_groups.find(
- name=group_name)
- for rule in group.rules:
- port_range = "%s:%s" % (rule['from_port'], rule['to_port'])
- fw_rule = FWRule({'protocol': rule['ip_protocol'],
- 'port_range': port_range,
- 'cidr_ip': rule['ip_range']['cidr']})
- fw_rules.append(fw_rule)
- sliver['fw_rules'] = fw_rules
- rspec_node['slivers'] = [sliver]
-
- # get disk image
- image = self.driver.shell.image_manager.get_images(
- id=instance.image['id'])
- if isinstance(image, list) and len(image) > 0:
- image = image[0]
- disk_image = image_to_rspec_disk_image(image)
- sliver['disk_image'] = [disk_image]
-
- # get interfaces
- rspec_node['services'] = []
- rspec_node['interfaces'] = []
- addresses = instance.addresses
- # HACK: public ips are stored in the list of private, but
- # this seems wrong. Assume pub ip is the last in the list of
- # private ips until openstack bug is fixed.
- if addresses.get('private'):
- login = Login({'authentication': 'ssh-keys',
- 'hostname': addresses.get('private')[-1]['addr'],
- 'port': '22', 'username': 'root'})
- service = Services({'login': login})
- rspec_node['services'].append(service)
-
- for private_ip in addresses.get('private', []):
- if_xrn = PlXrn(auth=self.driver.hrn,
- interface='node%s' % (instance.hostId))
- if_client_id = Xrn(if_xrn.urn, type='interface',
- id="eth%s" % if_index).urn
- if_sliver_id = Xrn(
- rspec_node['sliver_id'], type='slice', id="eth%s" % if_index).urn
- interface = Interface({'component_id': if_xrn.urn,
- 'client_id': if_client_id,
- 'sliver_id': if_sliver_id})
- interface['ips'] = [{'address': private_ip['addr'],
- #'netmask': private_ip['network'],
- 'type': private_ip['version']}]
- rspec_node['interfaces'].append(interface)
-
- # slivers always provide the ssh service
- for public_ip in addresses.get('public', []):
- login = Login({'authentication': 'ssh-keys',
- 'hostname': public_ip['addr'],
- 'port': '22', 'username': 'root'})
- service = Services({'login': login})
- rspec_node['services'].append(service)
- return rspec_node
-
- def instance_to_sliver(self, instance, xrn=None):
- if xrn:
- sliver_hrn = '%s.%s' % (self.driver.hrn, instance.id)
- sliver_id = Xrn(sliver_hrn, type='sliver').urn
-
- sliver = Sliver({'sliver_id': sliver_id,
- 'name': instance.name,
- 'type': instance.name,
- 'cpus': str(instance.vcpus),
- 'memory': str(instance.ram),
- 'storage': str(instance.disk)})
- return sliver
-
- def instance_to_geni_sliver(self, instance, sliver_allocations=None):
- if sliver_allocations is None:
- sliver_allocations = {}
- sliver_hrn = '%s.%s' % (self.driver.hrn, instance.id)
- sliver_id = Xrn(sliver_hrn, type='sliver').urn
-
- # set sliver allocation and operational status
- sliver_allocation = sliver_allocations[sliver_id]
- if sliver_allocation:
- allocation_status = sliver_allocation.allocation_state
- if allocation_status == 'geni_allocated':
- op_status = 'geni_pending_allocation'
- elif allocation_status == 'geni_provisioned':
- state = instance.state.lower()
- if state == 'active':
- op_status = 'geni_ready'
- elif state == 'building':
- op_status = 'geni_notready'
- elif state == 'failed':
- op_status = ' geni_failed'
- else:
- op_status = 'geni_unknown'
- else:
- allocation_status = 'geni_unallocated'
- # required fields
- geni_sliver = {'geni_sliver_urn': sliver_id,
- 'geni_expires': None,
- 'geni_allocation_status': allocation_status,
- 'geni_operational_status': op_status,
- 'geni_error': None,
- 'plos_created_at': datetime_to_string(utcparse(instance.created)),
- 'plos_sliver_type': self.shell.nova_manager.flavors.find(id=instance.flavor['id']).name,
- }
-
- return geni_sliver
-
- def get_aggregate_nodes(self):
- zones = self.get_availability_zones()
- # available sliver/instance/vm types
- instances = self.driver.shell.nova_manager.flavors.list()
- if isinstance(instances, dict):
- instances = instances.values()
- # available images
- images = self.driver.shell.image_manager.get_images_detailed()
- disk_images = [image_to_rspec_disk_image(img) for img in images if img[
- 'container_format'] in ['ami', 'ovf']]
- rspec_nodes = []
- for zone in zones:
- rspec_node = Node()
- xrn = OSXrn(zone, type='node')
- rspec_node['component_id'] = xrn.urn
- rspec_node['component_name'] = xrn.name
- rspec_node['component_manager_id'] = Xrn(
- self.driver.hrn, 'authority+cm').get_urn()
- rspec_node['exclusive'] = 'false'
- rspec_node['hardware_types'] = [HardwareType({'name': 'plos-pc'}),
- HardwareType({'name': 'pc'})]
- slivers = []
- for instance in instances:
- sliver = self.instance_to_sliver(instance)
- sliver['disk_image'] = disk_images
- slivers.append(sliver)
- rspec_node['available'] = 'true'
- rspec_node['slivers'] = slivers
- rspec_nodes.append(rspec_node)
-
- return rspec_nodes
-
- def create_tenant(self, tenant_name):
- tenants = self.driver.shell.auth_manager.tenants.findall(
- name=tenant_name)
- if not tenants:
- self.driver.shell.auth_manager.tenants.create(
- tenant_name, tenant_name)
- tenant = self.driver.shell.auth_manager.tenants.find(
- name=tenant_name)
- else:
- tenant = tenants[0]
- return tenant
-
- def create_instance_key(self, slice_hrn, user):
- slice_name = Xrn(slice_hrn).leaf
- user_name = Xrn(user['urn']).leaf
- key_name = "%s_%s" % (slice_name, user_name)
- pubkey = user['keys'][0]
- key_found = False
- existing_keys = self.driver.shell.nova_manager.keypairs.findall(
- name=key_name)
- for existing_key in existing_keys:
- if existing_key.public_key != pubkey:
- self.driver.shell.nova_manager.keypairs.delete(existing_key)
- elif existing_key.public_key == pubkey:
- key_found = True
-
- if not key_found:
- self.driver.shell.nova_manager.keypairs.create(key_name, pubkey)
- return key_name
-
- def create_security_group(self, slicename, fw_rules=None):
- if fw_rules is None:
- fw_rules = []
- # use default group by default
- group_name = 'default'
- if isinstance(fw_rules, list) and fw_rules:
- # Each sliver get's its own security group.
- # Keep security group names unique by appending some random
- # characters on end.
- random_name = "".join([random.choice(string.letters + string.digits)
- for i in xrange(6)])
- group_name = slicename + random_name
- security_group = SecurityGroup(self.driver)
- security_group.create_security_group(group_name)
- for rule in fw_rules:
- security_group.add_rule_to_group(group_name,
- protocol=rule.get('protocol'),
- cidr_ip=rule.get('cidr_ip'),
- port_range=rule.get(
- 'port_range'),
- icmp_type_code=rule.get('icmp_type_code'))
- # Open ICMP by default
- security_group.add_rule_to_group(group_name,
- protocol="icmp",
- cidr_ip="0.0.0.0/0",
- icmp_type_code="-1:-1")
- return group_name
-
- def add_rule_to_security_group(self, group_name, **kwds):
- security_group = SecurityGroup(self.driver)
- security_group.add_rule_to_group(group_name=group_name,
- protocol=kwds.get('protocol'),
- cidr_ip=kwds.get('cidr_ip'),
- icmp_type_code=kwds.get('icmp_type_code'))
-
- def run_instances(self, instance_name, tenant_name, rspec, key_name, pubkeys):
- # logger.debug('Reserving an instance: image: %s, flavor: ' \
- # '%s, key: %s, name: %s' % \
- # (image_id, flavor_id, key_name, slicename))
-
- # make sure a tenant exists for this slice
- tenant = self.create_tenant(tenant_name)
-
- # add the sfa admin user to this tenant and update our nova client connection
- # to use these credentials for the rest of this session. This emsures that the instances
- # we create will be assigned to the correct tenant.
- sfa_admin_user = self.driver.shell.auth_manager.users.find(
- name=self.driver.shell.auth_manager.opts['OS_USERNAME'])
- user_role = self.driver.shell.auth_manager.roles.find(name='user')
- admin_role = self.driver.shell.auth_manager.roles.find(name='admin')
- self.driver.shell.auth_manager.roles.add_user_role(
- sfa_admin_user, admin_role, tenant)
- self.driver.shell.auth_manager.roles.add_user_role(
- sfa_admin_user, user_role, tenant)
- self.driver.shell.nova_manager.connect(tenant=tenant.name)
-
- authorized_keys = "\n".join(pubkeys)
- files = {'/root/.ssh/authorized_keys': authorized_keys}
- rspec = RSpec(rspec)
- requested_instances = defaultdict(list)
-
- # iterate over clouds/zones/nodes
- slivers = []
- for node in rspec.version.get_nodes_with_slivers():
- instances = node.get('slivers', [])
- if not instances:
- continue
- for instance in instances:
- try:
- metadata = {}
- flavor_id = self.driver.shell.nova_manager.flavors.find(name=instance[
- 'name'])
- image = instance.get('disk_image')
- if image and isinstance(image, list):
- image = image[0]
- else:
- raise InvalidRSpec(
- "Must specify a disk_image for each VM")
- image_id = self.driver.shell.nova_manager.images.find(name=image[
- 'name'])
- fw_rules = instance.get('fw_rules', [])
- group_name = self.create_security_group(
- instance_name, fw_rules)
- metadata['security_groups'] = group_name
- if node.get('component_id'):
- metadata['component_id'] = node['component_id']
- if node.get('client_id'):
- metadata['client_id'] = node['client_id']
- server = self.driver.shell.nova_manager.servers.create(
- flavor=flavor_id,
- image=image_id,
- key_name=key_name,
- security_groups=[group_name],
- files=files,
- meta=metadata,
- name=instance_name)
- slivers.append(server)
- except Exception as err:
- logger.log_exc(err)
-
- return slivers
-
- def delete_instance(self, instance):
-
- def _delete_security_group(inst):
- security_group = inst.metadata.get('security_groups', '')
- if security_group:
- manager = SecurityGroup(self.driver)
- timeout = 10.0 # wait a maximum of 10 seconds before forcing the security group delete
- start_time = time.time()
- instance_deleted = False
- while instance_deleted == False and (time.time() - start_time) < timeout:
- tmp_inst = self.driver.shell.nova_manager.servers.findall(
- id=inst.id)
- if not tmp_inst:
- instance_deleted = True
- time.sleep(.5)
- manager.delete_security_group(security_group)
-
- multiclient = MultiClient()
- tenant = self.driver.shell.auth_manager.tenants.find(
- id=instance.tenant_id)
- self.driver.shell.nova_manager.connect(tenant=tenant.name)
- args = {'name': instance.name,
- 'id': instance.id}
- instances = self.driver.shell.nova_manager.servers.findall(**args)
- security_group_manager = SecurityGroup(self.driver)
- for instance in instances:
- # destroy instance
- self.driver.shell.nova_manager.servers.delete(instance)
- # deleate this instance's security groups
- multiclient.run(_delete_security_group, instance)
- return 1
-
- def stop_instances(self, instance_name, tenant_name, id=None):
- self.driver.shell.nova_manager.connect(tenant=tenant_name)
- args = {'name': instance_name}
- if id:
- args['id'] = id
- instances = self.driver.shell.nova_manager.servers.findall(**args)
- for instance in instances:
- self.driver.shell.nova_manager.servers.pause(instance)
- return 1
-
- def start_instances(self, instance_name, tenant_name, id=None):
- self.driver.shell.nova_manager.connect(tenant=tenant_name)
- args = {'name': instance_name}
- if id:
- args['id'] = id
- instances = self.driver.shell.nova_manager.servers.findall(**args)
- for instance in instances:
- self.driver.shell.nova_manager.servers.resume(instance)
- return 1
-
- def restart_instances(self, instacne_name, tenant_name, id=None):
- self.stop_instances(instance_name, tenant_name, id)
- self.start_instances(instance_name, tenant_name, id)
- return 1
-
- def update_instances(self, project_name):
- pass