+#
+
"""VServer slivers.
There are a couple of tricky things going on here. First, the kernel
"""
import errno
+import traceback
import os, os.path
+import sys
import time
-import commands
+from threading import BoundedSemaphore
+import subprocess
+# the util-vserver-pl module
import vserver
-import accounts
import logger
import tools
-from threading import BoundedSemaphore
-
-globalsem = BoundedSemaphore()
+from account import Account
+from initscript import Initscript
# special constant that tells vserver to keep its existing settings
KEEP_LIMIT = vserver.VC_LIM_KEEP
# populate the sliver/vserver specific default allocations table,
# which is used to look for slice attributes
DEFAULT_ALLOCATION = {}
-for rlimit in vserver.RLIMITS.keys():
+for rlimit in list(vserver.RLIMITS.keys()):
rlim = rlimit.lower()
- DEFAULT_ALLOCATION["%s_min"%rlim]=KEEP_LIMIT
- DEFAULT_ALLOCATION["%s_soft"%rlim]=KEEP_LIMIT
- DEFAULT_ALLOCATION["%s_hard"%rlim]=KEEP_LIMIT
+ DEFAULT_ALLOCATION["{}_min".format(rlim)] = KEEP_LIMIT
+ DEFAULT_ALLOCATION["{}_soft".format(rlim)] = KEEP_LIMIT
+ DEFAULT_ALLOCATION["{}_hard".format(rlim)] = KEEP_LIMIT
-class Sliver_VS(accounts.Account, vserver.VServer):
+class Sliver_VS(vserver.VServer, Account, Initscript):
"""This class wraps vserver.VServer to make its interface closer to what we need."""
SHELL = '/bin/vsh'
TYPE = 'sliver.VServer'
- _init_disk_info_sem = globalsem
+ _init_disk_info_sem = BoundedSemaphore()
def __init__(self, rec):
- logger.verbose ('initing Sliver_VS with name=%s'%rec['name'])
+ name = rec['name']
+ logger.verbose ('sliver_vs: {} init'.format(name))
try:
- vserver.VServer.__init__(self, rec['name'],logfile='/var/log/nm')
- except Exception, err:
+ logger.log("sliver_vs: {}: first chance...".format(name))
+ vserver.VServer.__init__(self, name, logfile='/var/log/nodemanager')
+ Account.__init__ (self, name)
+ Initscript.__init__ (self, name)
+ except Exception as err:
if not isinstance(err, vserver.NoSuchVServer):
# Probably a bad vserver or vserver configuration file
- logger.log_exc(self.name)
- logger.log('%s: recreating bad vserver' % rec['name'])
- self.destroy(rec['name'])
- self.create(rec['name'], rec['vref'])
- vserver.VServer.__init__(self, rec['name'],logfile='/var/log/nm')
+ logger.log_exc("sliver_vs:__init__ (first chance)", name=name)
+ logger.log('sliver_vs: {}: recreating bad vserver'.format(name))
+ self.destroy(name)
+ self.create(name, rec)
+ vserver.VServer.__init__(self, name, logfile='/var/log/nodemanager')
+ Account.__init__ (self, name)
+ Initscript.__init__ (self, name)
- self.keys = ''
self.rspec = {}
- self.initscript = ''
self.slice_id = rec['slice_id']
self.disk_usage_initialized = False
- self.initscriptchanged = False
+ self.enabled = True
+ # xxx this almost certainly is wrong...
self.configure(rec)
- _root_context_arch=None
- @staticmethod
- def root_context_arch():
- if not Sliver_VS._root_context_arch:
- Sliver_VS._root_context_arch=commands.getoutput("uname -i")
- return Sliver_VS._root_context_arch
-
@staticmethod
- def personality (arch):
- personality="linux32"
- if arch.find("64")>=0:
- personality="linux64"
- return personality
-
- @staticmethod
- def create(name, vref = None):
- logger.verbose('Sliver_VS:create - name=%s'%name)
+ def create(name, rec = None):
+ logger.verbose('sliver_vs: {}: create'.format(name))
+ vref = rec['vref']
if vref is None:
- vref='default'
+ # added by caglar
+ # band-aid for short period as old API doesn't have GetSliceFamily function
+ vref = "planetlab-f8-i386"
+ logger.log("sliver_vs: {}: ERROR - no vref attached, using hard-wired default {}"
+ .format(name, vref))
+
+ # used to look in /etc/planetlab/family,
+ # now relies on the 'GetSliceFamily' extra attribute in GetSlivers()
+ # which for legacy is still exposed here as the 'vref' key
+
+ # check the template exists -- there's probably a better way..
+ if not os.path.isdir ("/vservers/.vref/{}".format(vref)):
+ logger.log ("sliver_vs: {}: ERROR Could not create sliver - vreference image {} not found"
+ .format(name, vref))
+ return
+
+ # compute guest personality
try:
- ### locating the right slicefamily
- # this is a first draft, and more a proof of concept thing
- # the idea is to parse vref for dash-separated wishes,
- # and to project these against the defaults
- # so e.g. if the default slice family (as found in /etc/planetlab/slicefamily)
- # is planetlab-f8-i386, then here is what we get
- # vref=x86_64 -> vuseradd -t planetlab-f8-x86_64
- # vref=centos5 -> vuseradd -t planetlab-centos5-i386
- # vref=centos5-onelab -> vuseradd -t onelab-centos5-i386
- # vref=planetflow -> vuseradd -t planetflow-f8-i386
- # vref=x86_64-planetflow -> vuseradd -t planetflow-f8-x86_64
-
- # default
- default=file("/etc/planetlab/slicefamily").read().strip()
- (pldistro,fcdistro,arch) = default.split("-")
-
- known_archs = [ 'i386', 'x86_64' ]
- known_fcdistros = [ 'f8', 'f9', 'centos5' ]
- # from the slice attribute: cut dashes and try to figure the meaning
- slice_wishes = vref.split("-")
- for wish in slice_wishes:
- if wish in known_archs:
- arch=wish
- elif wish in known_fcdistros:
- fcdistro=wish
- else:
- pldistro=wish
-
- # rejoin the parts
- refname="-".join( (pldistro,fcdistro,arch) )
-
- # check the template exists -- there's probably a better way..
- if os.path.isdir ("/vservers/.vref/%s"% vref): refname = vref
-
- if not os.path.isdir ("/vservers/.vref/%s"% refname):
- logger.verbose("%s (%s) : vref %s not found, using default %s"%(
- name,vref,refname,default))
- refname=default
- # reset so arch is right
- (pldistro,fcdistro,arch) = default.split("-")
- # could check again, but as we have /etc/slicefamily
- # there's probably no /vservers/.vref/default
-
- except IOError:
- # have not found slicefamily
- logger.verbose("%s (%s): legacy node - using fallback vrefname 'default'"%(name,vref))
- # for legacy nodes
- refname="default"
- arch="i386"
+ (x, y, arch) = vref.split('-')
+ # mh, this of course applies when 'vref' is e.g. 'netflow'
+ # and that's not quite right
except:
- import traceback
- logger.log("%s (%s) : unexpected error follows - using 'default'"%(name,vref))
- logger.log(traceback.format_exc())
- refname="default"
- arch="i386"
-
- logger.log_call('/usr/sbin/vuseradd', '-t', refname, name)
+ arch = 'i386'
+
+ def personality (arch):
+ return "linux64" if arch.find("64") >= 0 else "linux32"
+
+ command = []
+ # be verbose
+ command += ['/bin/bash', '-x', ]
+ command += ['/usr/sbin/vuseradd', ]
+ if 'attributes' in rec and 'isolate_loopback' in rec['attributes'] and rec['attributes']['isolate_loopback'] == '1':
+ command += [ "-i", ]
+ # the vsliver imge to use
+ command += [ '-t', vref, ]
+ # slice name
+ command += [ name, ]
+ logger.log_call(command, timeout=15*60)
# export slicename to the slice in /etc/slicename
- file('/vservers/%s/etc/slicename' % name, 'w').write(name)
+ with open('/vservers/{}/etc/slicename'.format(name), 'w') as slicenamefile:
+ slicenamefile.write(name)
+ with open('/vservers/{}/etc/slicefamily'.format(name), 'w') as slicefamilyfile:
+ slicefamilyfile.write(vref)
# set personality: only if needed (if arch's differ)
- if Sliver_VS.root_context_arch() != arch:
- file('/etc/vservers/%s/personality' % name, 'w').write(Sliver_VS.personality(arch))
- logger.log('%s: set personality to %s'%(name,Sliver_VS.personality(arch)))
+ if tools.root_context_arch() != arch:
+ with open('/etc/vservers/{}/personality'.format(name), 'w') as personalityfile:
+ personalityfile.write(personality(arch)+"\n")
+ logger.log('sliver_vs: {}: set personality to {}'.format(name, personality(arch)))
@staticmethod
- def destroy(name): logger.log_call('/usr/sbin/vuserdel', name)
+ def destroy(name):
+ # need to umount before we trash, otherwise we end up with sequels in
+ # /vservers/slicename/ (namely in home/ )
+ # also because this is a static method we cannot check for 'omf_control'
+ # but it is no big deal as umount_ssh_dir checks before it umounts..
+ Account.umount_ssh_dir(name)
+ logger.log("sliver_vs: destroying {}".format(name))
+ logger.log_call(['/bin/bash', '-x', '/usr/sbin/vuserdel', name, ])
+
def configure(self, rec):
+ # in case we update nodemanager..
+ self.install_and_enable_vinit()
+
new_rspec = rec['_rspec']
if new_rspec != self.rspec:
self.rspec = new_rspec
self.set_resources()
- new_initscript = rec['initscript']
- if new_initscript != self.initscript:
- self.initscript = new_initscript
- self.initscriptchanged = True
-
- accounts.Account.configure(self, rec) # install ssh keys
+ # do the configure part from Initscript
+ # i.e. install slice initscript if defined
+ Initscript.configure(self, rec)
+ # install ssh keys
+ Account.configure(self, rec)
+ # remember configure() always gets called *before* start()
+ # in particular the slice initscript
+ # is expected to be in place already at this point
def start(self, delay=0):
- if self.rspec['enabled'] > 0:
- logger.log('%s: starting in %d seconds' % (self.name, delay))
- time.sleep(delay)
- child_pid = os.fork()
- if child_pid == 0:
- if self.initscriptchanged:
- logger.log('%s: installing initscript' % self.name)
- def install_initscript():
- flags = os.O_WRONLY | os.O_CREAT | os.O_TRUNC
- fd = os.open('/etc/rc.vinit', flags, 0755)
- os.write(fd, new_initscript)
- os.close(fd)
- try:
- self.chroot_call(install_initscript)
- self.initscriptchanged = False
- except: logger.log_exc(self.name)
- # VServer.start calls fork() internally,
- # so just close the nonstandard fds and fork once to avoid creating zombies
- tools.close_nonstandard_fds()
- vserver.VServer.start(self)
- os._exit(0)
- else: os.waitpid(child_pid, 0)
- else: logger.log('%s: not starting, is not enabled' % self.name)
+ if self.rspec['enabled'] <= 0:
+ logger.log('sliver_vs: not starting {}, is not enabled'.format(self.name))
+ return
+ logger.log('sliver_vs: {}: starting in {} seconds'.format(self.name, delay))
+ time.sleep(delay)
+ # the generic /etc/init.d/vinit script is permanently refreshed, and enabled
+ self.install_and_enable_vinit()
+ # expose .ssh for omf_friendly slivers
+ if 'omf_control' in self.rspec['tags']:
+ Account.mount_ssh_dir(self.name)
+ child_pid = os.fork()
+ if child_pid == 0:
+ # VServer.start calls fork() internally,
+ # so just close the nonstandard fds and fork once to avoid creating zombies
+ tools.close_nonstandard_fds()
+ vserver.VServer.start(self)
+ os._exit(0)
+ else:
+ os.waitpid(child_pid, 0)
def stop(self):
- logger.log('%s: stopping' % self.name)
+ logger.log('sliver_vs: {}: stopping'.format(self.name))
vserver.VServer.stop(self)
+ def is_running(self):
+ return vserver.VServer.is_running(self)
+
+ # this one seems to belong in Initscript at first sight,
+ # but actually depends on the underlying vm techno
+ # so let's keep it here
+ def rerun_slice_vinit(self):
+ command = "/usr/sbin/vserver {} exec /etc/rc.d/init.d/vinit restart"\
+ .format(self.name)
+ logger.log("vsliver_vs: {}: Rerunning slice initscript: {}"
+ .format(self.name, command))
+ subprocess.call(command + "&", stdin=open('/dev/null', 'r'),
+ stdout=open('/dev/null', 'w'), stderr=subprocess.STDOUT, shell=True)
+
def set_resources(self):
disk_max = self.rspec['disk_max']
- logger.log('%s: setting max disk usage to %d KiB' % (self.name, disk_max))
+ logger.log('sliver_vs: {}: setting max disk usage to {} KiB'
+ .format(self.name, disk_max))
try: # if the sliver is over quota, .set_disk_limit will throw an exception
if not self.disk_usage_initialized:
self.vm_running = False
Sliver_VS._init_disk_info_sem.acquire()
- logger.log('%s: computing disk usage: beginning' % self.name)
+ logger.log('sliver_vs: {}: computing disk usage: beginning'.format(self.name))
+ # init_disk_info is inherited from VServer
try: self.init_disk_info()
finally: Sliver_VS._init_disk_info_sem.release()
- logger.log('%s: computing disk usage: ended' % self.name)
+ logger.log('sliver_vs: {}: computing disk usage: ended'.format(self.name))
self.disk_usage_initialized = True
vserver.VServer.set_disklimit(self, max(disk_max, self.disk_blocks))
except:
- logger.log('%s: failed to set max disk usage' % self.name)
- logger.log_exc(self.name)
+ logger.log_exc('sliver_vs: failed to set max disk usage', name=self.name)
# get/set the min/soft/hard values for all of the vserver
# related RLIMITS. Note that vserver currently only
# implements support for hard limits.
- for limit in vserver.RLIMITS.keys():
+ for limit in list(vserver.RLIMITS.keys()):
type = limit.lower()
- minimum = self.rspec['%s_min'%type]
- soft = self.rspec['%s_soft'%type]
- hard = self.rspec['%s_hard'%type]
+ minimum = self.rspec['{}_min'.format(type)]
+ soft = self.rspec['{}_soft'.format(type)]
+ hard = self.rspec['{}_hard'.format(type)]
update = self.set_rlimit(limit, hard, soft, minimum)
if update:
- logger.log('%s: setting rlimit %s to (%d, %d, %d)'
- % (self.name, type, hard, soft, minimum))
+ logger.log('sliver_vs: {}: setting rlimit {} to ({}, {}, {})'
+ .format(self.name, type, hard, soft, minimum))
self.set_capabilities_config(self.rspec['capabilities'])
if self.rspec['capabilities']:
- logger.log('%s: setting capabilities to %s' % (self.name, self.rspec['capabilities']))
+ logger.log('sliver_vs: {}: setting capabilities to {}'
+ .format(self.name, self.rspec['capabilities']))
cpu_pct = self.rspec['cpu_pct']
cpu_share = self.rspec['cpu_share']
+ count = 1
+ for key in list(self.rspec.keys()):
+ if key.find('sysctl.') == 0:
+ sysctl = key.split('.')
+ try:
+ # /etc/vservers/<guest>/sysctl/<id>/
+ dirname = "/etc/vservers/{}/sysctl/{}".format(self.name, count)
+ try:
+ os.makedirs(dirname, 0o755)
+ except:
+ pass
+ with open("{}/setting".format(dirname), "w") as setting:
+ setting.write("{}\n".format(key.lstrip("sysctl.")))
+ with open("{}/value".format(dirname), "w") as value:
+ value.write("{}\n".format(self.rspec[key]))
+ count += 1
+
+ logger.log("sliver_vs: {}: writing {}={}"
+ .format(self.name, key, self.rspec[key]))
+ except IOError as e:
+ logger.log("sliver_vs: {}: could not set {}={}"
+ .format(self.name, key, self.rspec[key]))
+ logger.log("sliver_vs: {}: error = {}".format(self.name, e))
+
+
if self.rspec['enabled'] > 0:
if cpu_pct > 0:
- logger.log('%s: setting cpu reservation to %d%%' % (self.name, cpu_pct))
+ logger.log('sliver_vs: {}: setting cpu reservation to {}%'
+ .format(self.name, cpu_pct))
else:
cpu_pct = 0
if cpu_share > 0:
- logger.log('%s: setting cpu share to %d' % (self.name, cpu_share))
+ logger.log('sliver_vs: {}: setting cpu share to {}'
+ .format(self.name, cpu_share))
else:
cpu_share = 0
self.set_sched_config(cpu_pct, cpu_share)
# if IP address isn't set (even to 0.0.0.0), sliver won't be able to use network
if self.rspec['ip_addresses'] != '0.0.0.0':
- logger.log('%s: setting IP address(es) to %s' % \
- (self.name, self.rspec['ip_addresses']))
- self.set_ipaddresses_config(self.rspec['ip_addresses'])
-
- if self.is_running():
- logger.log("%s: Setting name to %s" % (self.name, self.slice_id),2)
- self.setname(self.slice_id)
-
+ logger.log('sliver_vs: {}: setting IP address(es) to {}'
+ .format(self.name, self.rspec['ip_addresses']))
+ add_loopback = True
+ if 'isolate_loopback' in self.rspec['tags']:
+ add_loopback = self.rspec['tags']['isolate_loopback'] != "1"
+ self.set_ipaddresses_config(self.rspec['ip_addresses'], add_loopback)
+
+ #logger.log("sliver_vs: {}: Setting name to {}".format(self.name, self.slice_id))
+ #self.setname(self.slice_id)
+ #logger.log("sliver_vs: {}: Storing slice id of {} for PlanetFlow".format(self.name, self.slice_id))
+ try:
+ vserver_config_path = '/etc/vservers/{}'.format(self.name)
+ if not os.path.exists (vserver_config_path):
+ os.makedirs (vserver_config_path)
+ with open('{}/slice_id'.format(vserver_config_path), 'w') as sliceidfile:
+ sliceidfile.write("{}\n".format(self.slice_id))
+ logger.log("sliver_vs: Recorded slice id {} for slice {}"
+ .format(self.slice_id, self.name))
+ except IOError as e:
+ logger.log("sliver_vs: Could not record slice_id for slice {}. Error: {}"
+ .format(self.name, str(e)))
+ except Exception as e:
+ logger.log_exc("sliver_vs: Error recording slice id: {}".format(e), name=self.name)
+
+
+ if self.enabled == False:
+ self.enabled = True
+ self.start()
+
if False: # Does not work properly yet.
if self.have_limits_changed():
- logger.log('%s: limits have changed --- restarting' % self.name)
+ logger.log('sliver_vs: {}: limits have changed --- restarting'.format(self.name))
stopcount = 10
while self.is_running() and stopcount > 0:
self.stop()
self.start()
else: # tell vsh to disable remote login by setting CPULIMIT to 0
- logger.log('%s: disabling remote login' % self.name)
+ logger.log('sliver_vs: {}: disabling remote login'.format(self.name))
self.set_sched_config(0, 0)
+ self.enabled = False
self.stop()