Merge Master in geni-v3 conflict resolution
[sfa.git] / sfa / openstack / osaggregate.py
1
2 import os
3 import socket
4 import base64
5 import string
6 import random
7 import time    
8 from collections import defaultdict
9 from nova.exception import ImageNotFound
10 from nova.api.ec2.cloud import CloudController
11 from sfa.util.faults import SfaAPIError, SliverDoesNotExist
12 from sfa.util.sfatime import utcparse, datetime_to_string, datetime_to_epoch
13 from sfa.rspecs.rspec import RSpec
14 from sfa.rspecs.elements.hardware_type import HardwareType
15 from sfa.rspecs.elements.node import Node
16 from sfa.rspecs.elements.sliver import Sliver
17 from sfa.rspecs.elements.login import Login
18 from sfa.rspecs.elements.disk_image import DiskImage
19 from sfa.rspecs.elements.services import Services
20 from sfa.rspecs.elements.interface import Interface
21 from sfa.rspecs.elements.fw_rule import FWRule
22 from sfa.util.xrn import Xrn
23 from sfa.planetlab.plxrn import PlXrn 
24 from sfa.openstack.osxrn import OSXrn, hrn_to_os_slicename
25 from sfa.rspecs.version_manager import VersionManager
26 from sfa.openstack.security_group import SecurityGroup
27 from sfa.server.threadmanager import ThreadManager
28 from sfa.util.sfalogging import logger
29
30 def pubkeys_to_user_data(pubkeys):
31     user_data = "#!/bin/bash\n\n"
32     for pubkey in pubkeys:
33         pubkey = pubkey.replace('\n', '')
34         user_data += "echo %s >> /root/.ssh/authorized_keys" % pubkey
35         user_data += "\n"
36         user_data += "echo >> /root/.ssh/authorized_keys"
37         user_data += "\n"
38     return user_data
39
40 def image_to_rspec_disk_image(image):
41     img = DiskImage()
42     img['name'] = image['name']
43     img['description'] = image['name']
44     img['os'] = image['name']
45     img['version'] = image['name']    
46     return img
47     
48 class OSAggregate:
49
50     def __init__(self, driver):
51         self.driver = driver
52
53     def get_availability_zones(self):
54         zones = self.driver.shell.nova_manager.dns_domains.domains()
55         if not zones:
56             zones = ['cloud']
57         else:
58             zones = [zone.name for zone in zones]
59         return zones
60
61     def list_resources(self, version=None, options={}):
62         version_manager = VersionManager()
63         version = version_manager.get_version(version)
64         rspec_version = version_manager._get_version(version.type, version.version, 'ad')
65         rspec = RSpec(version=version, user_options=options)
66         nodes = self.get_aggregate_nodes()
67         rspec.version.add_nodes(nodes)
68         return rspec.toxml()
69
70     def describe(self, urns, version=None, options={}):
71         # update nova connection
72         tenant_name = OSXrn(xrn=urns[0], type='slice').get_tenant_name()
73         self.driver.shell.nova_manager.connect(tenant=tenant_name)
74         instances = self.get_instances(urns)
75         # lookup the sliver allocations
76         sliver_ids = [sliver['sliver_id'] for sliver in slivers]
77         constraint = SliverAllocation.sliver_id.in_(sliver_ids)
78         sliver_allocations = dbsession.query(SliverAllocation).filter(constraint)
79         sliver_allocation_dict = {}
80         for sliver_allocation in sliver_allocations:
81             sliver_allocation_dict[sliver_allocation.sliver_id] = sliver_allocation
82
83         geni_slivers = []
84         rspec_nodes = []
85         for instance in instances:
86             rspec_nodes.append(self.instance_to_rspec_node(instance))
87             geni_sliver = self.instance_to_geni_sliver(instance, sliver_sllocation_dict)
88             geni_slivers.append(geni_sliver)
89         version_manager = VersionManager()
90         version = version_manager.get_version(version)
91         rspec_version = version_manager._get_version(version.type, version.version, 'manifest')
92         rspec = RSpec(version=rspec_version, user_options=options)
93         rspec.xml.set('expires',  datetime_to_string(utcparse(time.time())))
94         rspec.version.add_nodes(rspec_nodes)
95         result = {'geni_urn': Xrn(urns[0]).get_urn(),
96                   'geni_rspec': rspec.toxml(), 
97                   'geni_slivers': geni_slivers}
98         
99         return result
100
101     def get_instances(self, urns):
102         # parse slice names and sliver ids
103         names = set()
104         ids = set()
105         for urn in urns:
106             xrn = OSXrn(xrn=urn)
107             if xrn.type == 'slice':
108                 names.add(xrn.get_slice_name())
109             elif xrn.type == 'sliver':
110                 ids.add(xrn.leaf)
111
112         # look up instances
113         instances = []
114         filter = {}
115         if names:
116             filter['name'] = names
117         if ids:
118             filter['id'] = ids   
119         servers = self.driver.shell.nova_manager.servers.findall(**filter)
120         instances.extend(servers)
121
122         return instances
123
124     def instance_to_rspec_node(self, instance):
125         # determine node urn
126         node_xrn = instance.metadata.get('component_id')
127         if not node_xrn:
128             node_xrn = OSXrn('cloud', type='node')
129         else:
130             node_xrn = OSXrn(xrn=node_xrn, type='node')
131
132         rspec_node = Node()
133         rspec_node['component_id'] = node_xrn.urn
134         rspec_node['component_name'] = node_xrn.name
135         rspec_node['component_manager_id'] = Xrn(self.driver.hrn, 'authority+cm').get_urn()
136         rspec_node['sliver_id'] = OSXrn(name=instance.name, type='slice', id=instance.id).get_urn() 
137         if instance.metadata.get('client_id'):
138             rspec_node['client_id'] = instance.metadata.get('client_id')
139
140         # get sliver details
141         flavor = self.driver.shell.nova_manager.flavors.find(id=instance.flavor['id'])
142         sliver = self.instance_to_sliver(flavor)
143         # get firewall rules
144         fw_rules = []
145         group_name = instance.metadata.get('security_groups')
146         if group_name:
147             group = self.driver.shell.nova_manager.security_groups.find(name=group_name)
148             for rule in group.rules:
149                 port_range ="%s:%s" % (rule['from_port'], rule['to_port'])
150                 fw_rule = FWRule({'protocol': rule['ip_protocol'],
151                                   'port_range': port_range,
152                                   'cidr_ip': rule['ip_range']['cidr']})
153                 fw_rules.append(fw_rule)
154         sliver['fw_rules'] = fw_rules 
155         rspec_node['slivers'] = [sliver]
156
157         # get disk image
158         image = self.driver.shell.image_manager.get_images(id=instance.image['id'])
159         if isinstance(image, list) and len(image) > 0:
160             image = image[0]
161         disk_image = image_to_rspec_disk_image(image)
162         sliver['disk_image'] = [disk_image]
163
164         # get interfaces            
165         rspec_node['services'] = []
166         rspec_node['interfaces'] = []
167         addresses = instance.addresses
168         # HACK: public ips are stored in the list of private, but 
169         # this seems wrong. Assume pub ip is the last in the list of 
170         # private ips until openstack bug is fixed.      
171         if addresses.get('private'):
172             login = Login({'authentication': 'ssh-keys',
173                            'hostname': addresses.get('private')[-1]['addr'],
174                            'port':'22', 'username': 'root'})
175             service = Services({'login': login})
176             rspec_node['services'].append(service)    
177         
178         for private_ip in addresses.get('private', []):
179             if_xrn = PlXrn(auth=self.driver.hrn, 
180                            interface='node%s' % (instance.hostId)) 
181             if_client_id = Xrn(if_xrn.urn, type='interface', id="eth%s" %if_index).urn
182             if_sliver_id = Xrn(rspec_node['sliver_id'], type='slice', id="eth%s" %if_index).urn
183             interface = Interface({'component_id': if_xrn.urn,
184                                    'client_id': if_client_id,
185                                    'sliver_id': if_sliver_id})
186             interface['ips'] =  [{'address': private_ip['addr'],
187                                  #'netmask': private_ip['network'],
188                                  'type': private_ip['version']}]
189             rspec_node['interfaces'].append(interface) 
190         
191         # slivers always provide the ssh service
192         for public_ip in addresses.get('public', []):
193             login = Login({'authentication': 'ssh-keys', 
194                            'hostname': public_ip['addr'], 
195                            'port':'22', 'username': 'root'})
196             service = Services({'login': login})
197             rspec_node['services'].append(service)
198         return rspec_node
199
200
201     def instance_to_sliver(self, instance, xrn=None):
202         if xrn:
203             sliver_hrn = '%s.%s' % (self.driver.hrn, instance.id)
204             sliver_id = Xrn(sliver_hrn, type='sliver').urn
205
206         sliver = Sliver({'sliver_id': sliver_id,
207                          'name': instance.name,
208                          'type': instance.name,
209                          'cpus': str(instance.vcpus),
210                          'memory': str(instance.ram),
211                          'storage':  str(instance.disk)})
212         return sliver   
213
214     def instance_to_geni_sliver(self, instance, sliver_allocations = {}):
215         sliver_hrn = '%s.%s' % (self.driver.hrn, instance.id)
216         sliver_id = Xrn(sliver_hrn, type='sliver').urn
217  
218         # set sliver allocation and operational status
219         sliver_allocation = sliver_allocations[sliver_id]
220         if sliver_allocation:
221             allocation_status = sliver_allocation.allocation_state
222             if allocation_status == 'geni_allocated':
223                 op_status =  'geni_pending_allocation'
224             elif allocation_status == 'geni_provisioned':
225                 state = instance.state.lower()
226                 if state == 'active':
227                     op_status = 'geni_ready'
228                 elif state == 'building':
229                     op_status = 'geni_notready'
230                 elif state == 'failed':
231                     op_status =' geni_failed'
232                 else:
233                     op_status = 'geni_unknown'
234             else:
235                 allocation_status = 'geni_unallocated'    
236         # required fields
237         geni_sliver = {'geni_sliver_urn': sliver_id, 
238                        'geni_expires': None,
239                        'geni_allocation_status': allocation_status,
240                        'geni_operational_status': op_status,
241                        'geni_error': None,
242                        'plos_created_at': datetime_to_string(utcparse(instance.created)),
243                        'plos_sliver_type': self.shell.nova_manager.flavors.find(id=instance.flavor['id']).name,
244                         }
245
246         return geni_sliver
247                         
248     def get_aggregate_nodes(self):
249         zones = self.get_availability_zones()
250         # available sliver/instance/vm types
251         instances = self.driver.shell.nova_manager.flavors.list()
252         if isinstance(instances, dict):
253             instances = instances.values()
254         # available images
255         images = self.driver.shell.image_manager.get_images_detailed()
256         disk_images  = [image_to_rspec_disk_image(img) for img in images if img['container_format'] in ['ami', 'ovf']]
257         rspec_nodes = []
258         for zone in zones:
259             rspec_node = Node()
260             xrn = OSXrn(zone, type='node')
261             rspec_node['component_id'] = xrn.urn
262             rspec_node['component_name'] = xrn.name
263             rspec_node['component_manager_id'] = Xrn(self.driver.hrn, 'authority+cm').get_urn()
264             rspec_node['exclusive'] = 'false'
265             rspec_node['hardware_types'] = [HardwareType({'name': 'plos-pc'}),
266                                                 HardwareType({'name': 'pc'})]
267             slivers = []
268             for instance in instances:
269                 sliver = self.instance_to_sliver(instance)
270                 sliver['disk_image'] = disk_images
271                 slivers.append(sliver)
272             rspec_node['available'] = 'true'
273             rspec_node['slivers'] = slivers
274             rspec_nodes.append(rspec_node) 
275
276         return rspec_nodes 
277
278     def create_tenant(self, tenant_name):
279         tenants = self.driver.shell.auth_manager.tenants.findall(name=tenant_name)
280         if not tenants:
281             self.driver.shell.auth_manager.tenants.create(tenant_name, tenant_name)
282             tenant = self.driver.shell.auth_manager.tenants.find(name=tenant_name)
283         else:
284             tenant = tenants[0]
285         return tenant
286             
287     def create_instance_key(self, slice_hrn, user):
288         slice_name = Xrn(slice_hrn).leaf
289         user_name = Xrn(user['urn']).leaf
290         key_name = "%s_%s" % (slice_name, user_name)
291         pubkey = user['keys'][0]
292         key_found = False
293         existing_keys = self.driver.shell.nova_manager.keypairs.findall(name=key_name)
294         for existing_key in existing_keys:
295             if existing_key.public_key != pubkey:
296                 self.driver.shell.nova_manager.keypairs.delete(existing_key)
297             elif existing_key.public_key == pubkey:
298                 key_found = True
299
300         if not key_found:
301             self.driver.shell.nova_manager.keypairs.create(key_name, pubkey)
302         return key_name       
303         
304
305     def create_security_group(self, slicename, fw_rules=[]):
306         # use default group by default
307         group_name = 'default' 
308         if isinstance(fw_rules, list) and fw_rules:
309             # Each sliver get's its own security group.
310             # Keep security group names unique by appending some random
311             # characters on end.
312             random_name = "".join([random.choice(string.letters+string.digits)
313                                            for i in xrange(6)])
314             group_name = slicename + random_name 
315             security_group = SecurityGroup(self.driver)
316             security_group.create_security_group(group_name)
317             for rule in fw_rules:
318                 security_group.add_rule_to_group(group_name, 
319                                              protocol = rule.get('protocol'), 
320                                              cidr_ip = rule.get('cidr_ip'), 
321                                              port_range = rule.get('port_range'), 
322                                              icmp_type_code = rule.get('icmp_type_code'))
323             # Open ICMP by default
324             security_group.add_rule_to_group(group_name,
325                                              protocol = "icmp",
326                                              cidr_ip = "0.0.0.0/0",
327                                              icmp_type_code = "-1:-1")
328         return group_name
329
330     def add_rule_to_security_group(self, group_name, **kwds):
331         security_group = SecurityGroup(self.driver)
332         security_group.add_rule_to_group(group_name=group_name, 
333                                          protocol=kwds.get('protocol'), 
334                                          cidr_ip =kwds.get('cidr_ip'), 
335                                          icmp_type_code = kwds.get('icmp_type_code'))
336
337  
338
339     def run_instances(self, instance_name, tenant_name, rspec, key_name, pubkeys):
340         #logger.debug('Reserving an instance: image: %s, flavor: ' \
341         #            '%s, key: %s, name: %s' % \
342         #            (image_id, flavor_id, key_name, slicename))
343
344         # make sure a tenant exists for this slice
345         tenant = self.create_tenant(tenant_name)  
346
347         # add the sfa admin user to this tenant and update our nova client connection
348         # to use these credentials for the rest of this session. This emsures that the instances
349         # we create will be assigned to the correct tenant.
350         sfa_admin_user = self.driver.shell.auth_manager.users.find(name=self.driver.shell.auth_manager.opts['OS_USERNAME'])
351         user_role = self.driver.shell.auth_manager.roles.find(name='user')
352         admin_role = self.driver.shell.auth_manager.roles.find(name='admin')
353         self.driver.shell.auth_manager.roles.add_user_role(sfa_admin_user, admin_role, tenant)
354         self.driver.shell.auth_manager.roles.add_user_role(sfa_admin_user, user_role, tenant)
355         self.driver.shell.nova_manager.connect(tenant=tenant.name)  
356
357         authorized_keys = "\n".join(pubkeys)
358         files = {'/root/.ssh/authorized_keys': authorized_keys}
359         rspec = RSpec(rspec)
360         requested_instances = defaultdict(list)
361         
362         # iterate over clouds/zones/nodes
363         slivers = []
364         for node in rspec.version.get_nodes_with_slivers():
365             instances = node.get('slivers', [])
366             if not instances:
367                 continue
368             for instance in instances:
369                 try: 
370                     metadata = {}
371                     flavor_id = self.driver.shell.nova_manager.flavors.find(name=instance['name'])
372                     image = instance.get('disk_image')
373                     if image and isinstance(image, list):
374                         image = image[0]
375                     else:
376                         raise InvalidRSpec("Must specify a disk_image for each VM")
377                     image_id = self.driver.shell.nova_manager.images.find(name=image['name'])
378                     fw_rules = instance.get('fw_rules', [])
379                     group_name = self.create_security_group(instance_name, fw_rules)
380                     metadata['security_groups'] = group_name
381                     if node.get('component_id'):
382                         metadata['component_id'] = node['component_id']
383                     if node.get('client_id'):
384                         metadata['client_id'] = node['client_id'] 
385                     server = self.driver.shell.nova_manager.servers.create(
386                                                             flavor=flavor_id,
387                                                             image=image_id,
388                                                             key_name = key_name,
389                                                             security_groups = [group_name],
390                                                             files=files,
391                                                             meta=metadata, 
392                                                             name=instance_name)
393                     slivers.append(server)
394                 except Exception, err:    
395                     logger.log_exc(err)                                
396                            
397         return slivers        
398
399     def delete_instance(self, instance):
400     
401         def _delete_security_group(inst):
402             security_group = inst.metadata.get('security_groups', '')
403             if security_group:
404                 manager = SecurityGroup(self.driver)
405                 timeout = 10.0 # wait a maximum of 10 seconds before forcing the security group delete
406                 start_time = time.time()
407                 instance_deleted = False
408                 while instance_deleted == False and (time.time() - start_time) < timeout:
409                     tmp_inst = self.driver.shell.nova_manager.servers.findall(id=inst.id)
410                     if not tmp_inst:
411                         instance_deleted = True
412                     time.sleep(.5)
413                 manager.delete_security_group(security_group)
414
415         thread_manager = ThreadManager()
416         tenant = self.driver.shell.auth_manager.tenants.find(id=instance.tenant_id)  
417         self.driver.shell.nova_manager.connect(tenant=tenant.name)
418         args = {'name': instance.name,
419                 'id': instance.id}
420         instances = self.driver.shell.nova_manager.servers.findall(**args)
421         security_group_manager = SecurityGroup(self.driver)
422         for instance in instances:
423             # destroy instance
424             self.driver.shell.nova_manager.servers.delete(instance)
425             # deleate this instance's security groups
426             thread_manager.run(_delete_security_group, instance)
427         return 1
428
429     def stop_instances(self, instance_name, tenant_name, id=None):
430         self.driver.shell.nova_manager.connect(tenant=tenant_name)
431         args = {'name': instance_name}
432         if id:
433             args['id'] = id
434         instances = self.driver.shell.nova_manager.servers.findall(**args)
435         for instance in instances:
436             self.driver.shell.nova_manager.servers.pause(instance)
437         return 1
438
439     def start_instances(self, instance_name, tenant_name, id=None):
440         self.driver.shell.nova_manager.connect(tenant=tenant_name)
441         args = {'name': instance_name}
442         if id:
443             args['id'] = id
444         instances = self.driver.shell.nova_manager.servers.findall(**args)
445         for instance in instances:
446             self.driver.shell.nova_manager.servers.resume(instance)
447         return 1
448
449     def restart_instances(self, instacne_name, tenant_name, id=None):
450         self.stop_instances(instance_name, tenant_name, id)
451         self.start_instances(instance_name, tenant_name, id)
452         return 1 
453
454     def update_instances(self, project_name):
455         pass