in delete_instances(), wait until instance is destroyed until attempting to delete...
[sfa.git] / sfa / openstack / osaggregate.py
1
2 import os
3 import socket
4 import base64
5 import string
6 import random    
7 from collections import defaultdict
8 from nova.exception import ImageNotFound
9 from nova.api.ec2.cloud import CloudController
10 from sfa.util.faults import SfaAPIError
11 from sfa.rspecs.rspec import RSpec
12 from sfa.rspecs.elements.hardware_type import HardwareType
13 from sfa.rspecs.elements.node import Node
14 from sfa.rspecs.elements.sliver import Sliver
15 from sfa.rspecs.elements.login import Login
16 from sfa.rspecs.elements.disk_image import DiskImage
17 from sfa.rspecs.elements.services import Services
18 from sfa.rspecs.elements.interface import Interface
19 from sfa.util.xrn import Xrn
20 from sfa.planetlab.plxrn import PlXrn 
21 from sfa.openstack.osxrn import OSXrn, hrn_to_os_slicename
22 from sfa.rspecs.version_manager import VersionManager
23 from sfa.openstack.security_group import SecurityGroup
24 from sfa.server.threadmanager import ThreadManager
25 from sfa.util.sfalogging import logger
26
27 def pubkeys_to_user_data(pubkeys):
28     user_data = "#!/bin/bash\n\n"
29     for pubkey in pubkeys:
30         pubkey = pubkey.replace('\n', '')
31         user_data += "echo %s >> /root/.ssh/authorized_keys" % pubkey
32         user_data += "\n"
33         user_data += "echo >> /root/.ssh/authorized_keys"
34         user_data += "\n"
35     return user_data
36
37 def instance_to_sliver(instance, slice_xrn=None):
38     sliver_id = None
39     if slice_xrn:
40         xrn = Xrn(slice_xrn, 'slice')
41         sliver_id = xrn.get_sliver_id(instance.project_id, instance.hostname, instance.id)
42
43     sliver = Sliver({'slice_id': sliver_id,
44                      'name': instance.name,
45                      'type': instance.name,
46                      'cpus': str(instance.vcpus),
47                      'memory': str(instance.ram),
48                      'storage':  str(instance.disk)})
49     return sliver
50
51 def image_to_rspec_disk_image(image):
52     img = DiskImage()
53     img['name'] = image['name']
54     img['description'] = image['name']
55     img['os'] = image['name']
56     img['version'] = image['name']    
57     return img
58     
59 class OSAggregate:
60
61     def __init__(self, driver):
62         self.driver = driver
63
64     def get_rspec(self, slice_xrn=None, version=None, options={}):
65         version_manager = VersionManager()
66         version = version_manager.get_version(version)
67         if not slice_xrn:
68             rspec_version = version_manager._get_version(version.type, version.version, 'ad')
69             nodes = self.get_aggregate_nodes()
70         else:
71             rspec_version = version_manager._get_version(version.type, version.version, 'manifest')
72             nodes = self.get_slice_nodes(slice_xrn)
73         rspec = RSpec(version=rspec_version, user_options=options)
74         rspec.version.add_nodes(nodes)
75         return rspec.toxml()
76
77     def get_availability_zones(self):
78         # essex release
79         zones = self.driver.shell.nova_manager.dns_domains.domains()
80
81         if not zones:
82             zones = ['cloud']
83         else:
84             zones = [zone.name for zone in zones]
85         return zones
86
87     def get_slice_nodes(self, slice_xrn):
88         # update nova connection
89         tenant_name = OSXrn(xrn=slice_xrn, type='slice').get_tenant_name()
90         self.driver.shell.nova_manager.connect(tenant=tenant_name)    
91         
92         zones = self.get_availability_zones()
93         name = hrn_to_os_slicename(slice_xrn)
94         instances = self.driver.shell.nova_manager.servers.findall(name=name)
95         node_dict = {}
96         for instance in instances:
97             # determine node urn
98             node_xrn = instance.metadata.get('component_id')
99             if not node_xrn:
100                 node_xrn = OSXrn('cloud', type='node')
101             else:
102                 node_xrn = OSXrn(xrn=node_xrn, type='node')
103
104             if not node_xrn.urn in node_dict:
105                 rspec_node = Node()
106                 rspec_node['component_id'] = node_xrn.urn
107                 rspec_node['component_name'] = node_xrn.name
108                 rspec_node['component_manager_id'] = Xrn(self.driver.hrn, 'authority+cm').get_urn()
109                 rspec_node['slivers'] = []
110                 node_dict[node_xrn.urn] = rspec_node
111             else:
112                 rspec_node = node_dict[node_xrn.urn]
113
114             if instance.metadata.get('client_id'):
115                 rspec_node['client_id'] = instance.metadata.get('client_id')
116             
117             flavor = self.driver.shell.nova_manager.flavors.find(id=instance.flavor['id'])
118             sliver = instance_to_sliver(flavor)
119             rspec_node['slivers'].append(sliver)
120             image = self.driver.shell.image_manager.get_images(id=instance.image['id'])
121             if isinstance(image, list) and len(image) > 0:
122                 image = image[0]
123             disk_image = image_to_rspec_disk_image(image)
124             sliver['disk_image'] = [disk_image]
125
126             # build interfaces            
127             rspec_node['services'] = []
128             rspec_node['interfaces'] = []
129             addresses = instance.addresses
130             # HACK: public ips are stored in the list of private, but 
131             # this seems wrong. Assume pub ip is the last in the list of 
132             # private ips until openstack bug is fixed.      
133             if addresses.get('private'):
134                 login = Login({'authentication': 'ssh-keys',
135                                'hostname': addresses.get('private')[-1]['addr'],
136                                'port':'22', 'username': 'root'})
137                 service = Services({'login': login})
138                 rspec_node['services'].append(service)    
139             
140             for private_ip in addresses.get('private', []):
141                 if_xrn = PlXrn(auth=self.driver.hrn, 
142                                interface='node%s:eth0' % (instance.hostId)) 
143                 interface = Interface({'component_id': if_xrn.urn})
144                 interface['ips'] =  [{'address': private_ip['addr'],
145                                      #'netmask': private_ip['network'],
146                                      'type': private_ip['version']}]
147                 rspec_node['interfaces'].append(interface) 
148             
149             # slivers always provide the ssh service
150             for public_ip in addresses.get('public', []):
151                 login = Login({'authentication': 'ssh-keys', 
152                                'hostname': public_ip['addr'], 
153                                'port':'22', 'username': 'root'})
154                 service = Services({'login': login})
155                 rspec_node['services'].append(service)
156             rspec_nodes.append(rspec_node)
157         return node_dict.values()
158
159     def get_aggregate_nodes(self):
160         zones = self.get_availability_zones()
161         # available sliver/instance/vm types
162         instances = self.driver.shell.nova_manager.flavors.list()
163         if isinstance(instances, dict):
164             instances = instances.values()
165         # available images
166         images = self.driver.shell.image_manager.get_images_detailed()
167         disk_images  = [image_to_rspec_disk_image(img) for img in images if img['container_format'] in ['ami', 'ovf']]
168         rspec_nodes = []
169         for zone in zones:
170             rspec_node = Node()
171             xrn = OSXrn(zone, type='node')
172             rspec_node['component_id'] = xrn.urn
173             rspec_node['component_name'] = xrn.name
174             rspec_node['component_manager_id'] = Xrn(self.driver.hrn, 'authority+cm').get_urn()
175             rspec_node['exclusive'] = 'false'
176             rspec_node['hardware_types'] = [HardwareType({'name': 'plos-pc'}),
177                                                 HardwareType({'name': 'pc'})]
178             slivers = []
179             for instance in instances:
180                 sliver = instance_to_sliver(instance)
181                 sliver['disk_image'] = disk_images
182                 slivers.append(sliver)
183         
184             rspec_node['slivers'] = slivers
185             rspec_nodes.append(rspec_node) 
186
187         return rspec_nodes 
188
189
190     def create_tenant(self, tenant_name):
191         tenants = self.driver.shell.auth_manager.tenants.findall(name=tenant_name)
192         if not tenants:
193             self.driver.shell.auth_manager.tenants.create(tenant_name, tenant_name)
194             tenant = self.driver.shell.auth_manager.tenants.find(name=tenant_name)
195         else:
196             tenant = tenants[0]
197         return tenant
198             
199
200     def create_instance_key(self, slice_hrn, user):
201         slice_name = Xrn(slice_hrn).leaf
202         user_name = Xrn(user['urn']).leaf
203         key_name = "%s_%s" % (slice_name, user_name)
204         pubkey = user['keys'][0]
205         key_found = False
206         existing_keys = self.driver.shell.nova_manager.keypairs.findall(name=key_name)
207         for existing_key in existing_keys:
208             if existing_key.public_key != pubkey:
209                 self.driver.shell.nova_manager.keypairs.delete(existing_key)
210             elif existing_key.public_key == pubkey:
211                 key_found = True
212
213         if not key_found:
214             self.driver.shell.nova_manager.keypairs.create(key_name, pubkey)
215         return key_name       
216         
217
218     def create_security_group(self, slicename, fw_rules=[]):
219         # use default group by default
220         group_name = 'default' 
221         if isinstance(fw_rules, list) and fw_rules:
222             # Each sliver get's its own security group.
223             # Keep security group names unique by appending some random
224             # characters on end.
225             random_name = "".join([random.choice(string.letters+string.digits)
226                                            for i in xrange(6)])
227             group_name = slicename + random_name 
228             security_group = SecurityGroup(self.driver)
229             security_group.create_security_group(group_name)
230             for rule in fw_rules:
231                 security_group.add_rule_to_group(group_name, 
232                                              protocol = rule.get('protocol'), 
233                                              cidr_ip = rule.get('cidr_ip'), 
234                                              port_range = rule.get('port_range'), 
235                                              icmp_type_code = rule.get('icmp_type_code'))
236         return group_name
237
238     def add_rule_to_security_group(self, group_name, **kwds):
239         security_group = SecurityGroup(self.driver)
240         security_group.add_rule_to_group(group_name=group_name, 
241                                          protocol=kwds.get('protocol'), 
242                                          cidr_ip =kwds.get('cidr_ip'), 
243                                          icmp_type_code = kwds.get('icmp_type_code'))
244
245  
246
247     def run_instances(self, instance_name, tenant_name, rspec, key_name, pubkeys):
248         #logger.debug('Reserving an instance: image: %s, flavor: ' \
249         #            '%s, key: %s, name: %s' % \
250         #            (image_id, flavor_id, key_name, slicename))
251
252         # make sure a tenant exists for this slice
253         tenant = self.create_tenant(tenant_name)  
254
255         # add the sfa admin user to this tenant and update our nova client connection
256         # to use these credentials for the rest of this session. This emsures that the instances
257         # we create will be assigned to the correct tenant.
258         sfa_admin_user = self.driver.shell.auth_manager.users.find(name=self.driver.shell.auth_manager.opts['OS_USERNAME'])
259         user_role = self.driver.shell.auth_manager.roles.find(name='user')
260         admin_role = self.driver.shell.auth_manager.roles.find(name='admin')
261         self.driver.shell.auth_manager.roles.add_user_role(sfa_admin_user, admin_role, tenant)
262         self.driver.shell.auth_manager.roles.add_user_role(sfa_admin_user, user_role, tenant)
263         self.driver.shell.nova_manager.connect(tenant=tenant.name)  
264
265         authorized_keys = "\n".join(pubkeys)
266         files = {'/root/.ssh/authorized_keys': authorized_keys}
267         rspec = RSpec(rspec)
268         requested_instances = defaultdict(list)
269         # iterate over clouds/zones/nodes
270         for node in rspec.version.get_nodes_with_slivers():
271             instances = node.get('slivers', [])
272             if not instances:
273                 continue
274             for instance in instances:
275                 try: 
276                     metadata = {}
277                     flavor_id = self.driver.shell.nova_manager.flavors.find(name=instance['name'])
278                     image = instance.get('disk_image')
279                     if image and isinstance(image, list):
280                         image = image[0]
281                     image_id = self.driver.shell.nova_manager.images.find(name=image['name'])
282                     fw_rules = instance.get('fw_rules', [])
283                     group_name = self.create_security_group(instance_name, fw_rules)
284                     metadata['security_groups'] = group_name
285                     if node.get('component_id'):
286                         metadata['component_id'] = node['component_id']
287                     if node.get('client_id'):
288                         metadata['client_id'] = node['client_id']
289                     self.driver.shell.nova_manager.servers.create(flavor=flavor_id,
290                                                             image=image_id,
291                                                             key_name = key_name,
292                                                             security_groups = [group_name],
293                                                             files=files,
294                                                             meta=metadata, 
295                                                             name=instance_name)
296                 except Exception, err:    
297                     logger.log_exc(err)                                
298                            
299
300
301     def delete_instances(self, instance_name, tenant_name):
302
303         def _delete_security_group(instance):
304             security_group = instance.metadata.get('security_groups', '')
305             if security_group:
306                 manager = SecurityGroup(self.driver)
307                 timeout = 10.0 # wait a maximum of 10 seconds before forcing the security group delete
308                 start_time = time.time()
309                 instance_deleted = False
310                 while instance_deleted == False and (time.time() - start_time) < timeout:
311                     inst = self.driver.shell.nova_manager.servers.findall(id=instance.id)
312                     if not inst:
313                         instance_deleted = True
314                 manager.delete_security_group(security_group)
315
316         thread_manager = ThreadManager()
317         self.driver.shell.nova_manager.connect(tenant=tenant_name)
318         instances = self.driver.shell.nova_manager.servers.findall(name=instance_name)
319         for instance in instances:
320             # destroy instance
321             self.driver.shell.nova_manager.servers.delete(instance)
322             # deleate this instance's security groups
323             thread_manager.run(_delete_security_group, instance)
324         return 1
325
326
327     def stop_instances(self, instance_name, tenant_name):
328         self.driver.shell.nova_manager.connect(tenant=tenant_name)
329         instances = self.driver.shell.nova_manager.servers.findall(name=instance_name)
330         for instance in instances:
331             self.driver.shell.nova_manager.servers.pause(instance)
332         return 1
333
334     def update_instances(self, project_name):
335         pass