X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=system%2FTestSlice.py;h=0043d4c5a0913efe6beca828499304380ff49c39;hb=8980ce2087e5dc37a815e6fca78d881bdb22521c;hp=1e53cac36b85f76947222918767bb0434eabccfd;hpb=88ce17543e3f87197e7d6fd792ea56d99b7d67c3;p=tests.git diff --git a/system/TestSlice.py b/system/TestSlice.py index 1e53cac..0043d4c 100644 --- a/system/TestSlice.py +++ b/system/TestSlice.py @@ -8,11 +8,11 @@ import time from TestKey import TestKey from TestUser import TestUser -from TestNode import TestNode +from TestNode import TestNode, CompleterTaskNodeSsh from TestSsh import TestSsh from Completer import Completer, CompleterTask -class CompleterTaskSshSlice (CompleterTask): +class CompleterTaskSliceSsh (CompleterTask): def __init__ (self, test_plc, hostname, slicename, private_key,command, expected, dry_run): self.test_plc=test_plc @@ -30,8 +30,11 @@ class CompleterTaskSshSlice (CompleterTask): if self.dry_run: return True if self.expected: return retcod==0 else: return retcod!=0 - def failure_message (self): - return "Could not ssh into slice %s @ %s"%(self.slicename,self.hostname) + def failure_epilogue (self): + if self.expected: + print "Could not ssh into sliver %s@%s"%(self.slicename,self.hostname) + else: + print "Could still ssh into sliver%s@%s (that was expected to be down)"%(self.slicename,self.hostname) class TestSlice: @@ -110,7 +113,7 @@ class TestSlice: hostnames=[] for nodename in self.slice_spec['nodenames']: node_spec=self.test_site.locate_node(nodename) - test_node=TestNode(self,self.test_site,node_spec) + test_node=TestNode(self.test_plc,self.test_site,node_spec) hostnames += [test_node.name()] utils.header("Adding %r in %s"%(hostnames,slice_name)) self.test_plc.apiserver.AddSliceToNodes(auth, slice_name, hostnames) @@ -140,23 +143,24 @@ class TestSlice: key_names += user_spec['key_names'] return self.test_plc.locate_private_key_from_key_names (key_names) - # trying to reach the slice through ssh - expected to answer - def ssh_slice (self, options, *args, **kwds): - "tries to ssh-enter the slice with the user key, to ensure slice creation" - return self.do_ssh_slice(options, expected=True, *args, **kwds) + # for TestPlc.slice_mapper__tasks + # i.e. returns a list of CompleterTasks that are merged into the same Completer run + # to avoid waiting for as many slices as the Plc has + # also the __doc__ lines are used for the TestPlc methods, e.g. just 'ssh_slice' + def ssh_slice__tasks (self, options, *args, **kwds): + "tries to ssh-enter the slice with the user key, to check for slice creation" + return self.ssh_tasks(options, expected=True, *args, **kwds) # when we expect the slice is not reachable - def ssh_slice_off (self, options, *args, **kwds): + def ssh_slice_off__tasks (self, options, *args, **kwds): "tries to ssh-enter the slice with the user key, expecting it to be unreachable" - return self.do_ssh_slice(options, expected=False, *args, **kwds) + return self.ssh_tasks(options, expected=False, *args, **kwds) - def do_ssh_slice(self,options,expected=True, - timeout_minutes=20,silent_minutes=10,period_seconds=15,command=None): - "tries to enter a slice" - - timeout = timedelta(minutes=timeout_minutes) - graceout = timedelta(minutes=silent_minutes) - period = timedelta(seconds=period_seconds) + def ssh_tasks(self,options, expected=True, command=None): +# timeout_minutes=20,silent_minutes=10,period_seconds=15): +# timeout = timedelta(minutes=timeout_minutes) +# graceout = timedelta(minutes=silent_minutes) +# period = timedelta(seconds=period_seconds) if not command: command="echo hostname ; hostname; echo id; id; echo uname -a ; uname -a" # locate a key @@ -175,9 +179,10 @@ class TestSlice: dry_run = getattr(options,'dry_run',False) for nodename in self.slice_spec['nodenames']: (site_spec,node_spec) = self.test_plc.locate_node(nodename) - tasks.append( CompleterTaskSshSlice(self.test_plc,node_spec['node_fields']['hostname'], + tasks.append( CompleterTaskSliceSsh(self.test_plc,node_spec['node_fields']['hostname'], slicename,private_key,command,expected,dry_run)) - return Completer (tasks).run (timeout, graceout, period) + return tasks +# return Completer (tasks).run (timeout, graceout, period) def ssh_slice_basics (self, options, *args, **kwds): "the slice is expected to be UP and we just check a few simple sanity commands, including 'ps' to check for /proc" @@ -186,6 +191,7 @@ class TestSlice: if not self.do_ssh_slice_once(options,expected=False, command='false'): overall=False if not self.do_ssh_slice_once(options,expected=False, command='someimprobablecommandname'): overall=False if not self.do_ssh_slice_once(options,expected=True, command='ps'): overall=False + if not self.do_ssh_slice_once(options,expected=False, command='ls /vservers'): overall=False return overall # pick just one nodename and runs the ssh command once @@ -215,3 +221,29 @@ class TestSlice: else: success = retcod!=0 if not success: utils.header ("WRONG RESULT for %s"%msg) return success + + # for TestPlc.slice_mapper__tasks + # check that /vservers/<> is present/deleted + def slice_fs_present__tasks (self, options): + "checks that /vservers/ exists on the filesystem" + return self.check_rootfs_tasks(options,expected=True) + def slice_fs_deleted__tasks (self, options): + "checks that /vservers/ has been properly wiped off" + return self.check_rootfs_tasks (options,expected=False) + + def check_rootfs_tasks (self, options, expected): + # use constant admin key + local_key = "keys/key_admin.rsa" + node_infos = self.test_plc.all_node_infos() + rootfs="/vservers/%s"%self.name() + class CompleterTaskRootfs (CompleterTaskNodeSsh): + def __init__ (self, nodename, qemuname): + CompleterTaskNodeSsh.__init__(self,nodename, qemuname, local_key, expected=expected, + command="ls -d %s"%rootfs) + def failure_epilogue (self): + if expected: + print "Could not stat %s - was expected to be present"%rootfs + else: + print "Sliver rootfs %s still present - this is unexpected"%rootfs + utils.system(self.test_ssh.actual_command("ls -l %s; du -hs %s"%(rootfs,rootfs),dry_run=self.dry_run)) + return [ CompleterTaskRootfs (nodename, qemuname) for (nodename,qemuname) in node_infos ]