From 79b4fa2971f7f6fca8dba0d0875a94556b26764f Mon Sep 17 00:00:00 2001 From: Thierry Parmentelat Date: Fri, 24 Nov 2006 12:06:00 +0000 Subject: [PATCH] Persons gets cached and connected to Keys (not tested until GetSlivers yet) reference output of TestPeers under cvs yet todo: check GetSlivers Optimize the set of calls each peer sends other peers, the current scheme tends to be slow because of too many small requests being issued consider writing a dedicated Method at some point --- PLC/Cache.py | 39 +++-- PLC/Persons.py | 18 ++- TestPeers.py | 257 ++++++++++++++++++++----------- TestPeers.ref | 405 +++++++++++++++++++++++++++++++++++++++++++++++++ planetlab4.sql | 7 +- 5 files changed, 622 insertions(+), 104 deletions(-) create mode 100644 TestPeers.ref diff --git a/PLC/Cache.py b/PLC/Cache.py index fb0fce5..ff9f180 100644 --- a/PLC/Cache.py +++ b/PLC/Cache.py @@ -146,7 +146,7 @@ class Cache: # (*) or naming conflicts local_objects = table_class (self.api) ### index upon class_key for future searches - verbose ('local objects:',local_objects) + #verbose ('local objects:',local_objects) verbose ('class_key',class_key) local_objects_index = local_objects.dict(class_key) verbose ('update_table',classname,local_objects_index.keys()) @@ -164,11 +164,13 @@ class Cache: # scan the peer's local objects for alien_object in alien_object_list: + object_name = alien_object[class_key] + ### ignore, e.g. system-wide slices if lambda_ignore(alien_object): + verbose('Ignoring',object_name) continue - object_name = alien_object[class_key] verbose ('update_table - Considering',object_name) # create or update @@ -176,6 +178,7 @@ class Cache: ### We know about this object already local_object = local_objects_index[object_name] if local_object ['peer_id'] is None: + ### xxx send e-mail print 'We are in trouble here' print 'The %s object named %s is natively defined twice'%(classname,object_name) print 'Once on this PLC and once on peer %d'%peer_id @@ -271,22 +274,38 @@ class Cache: def refresh_peer (self): - peer_local_keys = self.peer_server.GetKeys(self.auth,{'peer_id':None}) - peer_local_nodes = self.peer_server.GetNodes(self.auth,None,None,'local') - # xxx would ideally get our own nodes only, - # requires to know remote peer's peer_id for ourselves, mmhh - peer_foreign_nodes = self.peer_server.GetNodes(self.auth,None,None,'foreign') peer_local_slices = self.peer_server.GetSlices(self.auth,{'peer_id':None}) + # refresh keys + peer_local_keys = self.peer_server.GetKeys(self.auth,{'peer_id':None}) nb_new_keys = self.update_table('Key', peer_local_keys) + # refresh nodes + peer_local_nodes = self.peer_server.GetNodes(self.auth,{'peer_id':None}) nb_new_nodes = self.update_table('Node', peer_local_nodes) - # rough and temporary - nb_new_slices = self.refresh_slices(peer_local_slices,peer_local_nodes+peer_foreign_nodes) - + # refresh persons + peer_local_persons = self.peer_server.GetPersons(self.auth,{'peer_id':None}) + # xxx ideally get our own persons only + # requires to know remote peer's peer_id for ourselves, mmhh + peer_all_keys = peer_local_keys + self.peer_server.GetKeys(self.auth,{'~peer_id':None}) + nb_new_persons = self.update_table ('Person', peer_local_persons, + { 'Key': peer_all_keys} ) + + # refresh slices + def is_system_slice (slice): + return slice['creator_person_id'] == 1 + # xxx would ideally get our own nodes only, + peer_all_nodes = peer_local_nodes+self.peer_server.GetNodes(self.auth,{'~peer_id':None}) + + nb_new_slices = self.update_table ('Slice', peer_local_slices, + {'Node':peer_all_nodes}, + is_system_slice) + + return {'plcname':self.api.config.PLC_NAME, 'new_keys':nb_new_keys, 'new_nodes':nb_new_nodes, + 'new_persons':nb_new_persons, 'new_slices':nb_new_slices} diff --git a/PLC/Persons.py b/PLC/Persons.py index 305c1ec..b519302 100644 --- a/PLC/Persons.py +++ b/PLC/Persons.py @@ -4,7 +4,7 @@ # Mark Huang # Copyright (C) 2006 The Trustees of Princeton University # -# $Id: Persons.py,v 1.18 2006/11/09 03:07:42 mlhuang Exp $ +# $Id: Persons.py,v 1.19 2006/11/09 19:43:55 mlhuang Exp $ # from types import StringTypes @@ -50,8 +50,24 @@ class Person(Row): 'site_ids': Parameter([int], "List of site identifiers"), 'key_ids': Parameter([int], "List of key identifiers"), 'slice_ids': Parameter([int], "List of slice identifiers"), + 'peer_id': Parameter(int, "Peer at which this slice was created", nullok = True), } + # for Cache + class_key = 'email' + foreign_fields = ['first_name', 'last_name', 'title', 'email', 'phone', 'url', + 'bio', 'enabled', 'password', 'last_updated', 'date_created'] + #foreign_xrefs = { 'Node' : { 'field' : 'node_ids' , + # 'table': 'slice_node' } } + foreign_xrefs = { + 'Key' : { 'field' : 'key_ids', 'table' : 'person_key' } +# 'key_ids': Parameter([int], "List of key identifiers"), +# 'role_ids': Parameter([int], "List of role identifiers"), +# 'roles': Parameter([str], "List of roles"), +# 'site_ids': Parameter([int], "List of site identifiers"), +# 'slice_ids': Parameter([int], "List of slice identifiers"), +} + def validate_email(self, email): """ Validate email address. Stolen from Mailman. diff --git a/TestPeers.py b/TestPeers.py index ae59068..8ea5e58 100755 --- a/TestPeers.py +++ b/TestPeers.py @@ -50,51 +50,63 @@ import os plc1={ 'plcname':'plc1 in federation', 'hostname':'lurch.cs.princeton.edu', 'url-format':'https://%s:443/PLCAPI/', - 'builtin_admin_id':'root@localhost.localdomain', - 'builtin_admin_password':'root', - 'peer_admin_name':'plc1@planet-lab.org', - 'peer_admin_password':'peer', - 'peer_admin_key':'ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAIEAqK1lKNf61lAGYAXzG6xKnFQkfv3ViG0GP2Krp1zD7d/93IkmsXVEjLfGhEJQjRjzRc9/gFdATP703cDzp4Ag2eR2wdQz0e6SXOBd2sLuW3LqTwor1XMmp5f0QCOg5OSKXwozE3Tlt0+ewBNvAE8HWwZFjou5CFnrFMVZPjqfhpU= thierry.parmentelat@sophia.inria.fr', + 'builtin-admin-id':'root@plc1.org', + 'builtin-admin-password':'root', + 'peer-admin-name':'plc1@planet-lab.org', + 'peer-admin-password':'peer', 'node-format':'n1%02d.plc1.org', 'plainname' : 'one', 'slice-format' : 's1%02d', + 'person-format' : 'user1-%d@plc1.org', + 'key-format':'ssh-rsa 1111111111111111 user%d-key%d', + 'person-password' : 'password1', } plc2={ 'plcname':'plc2 in federation', 'hostname':'planetlab-devbox.inria.fr', 'url-format':'https://%s:443/PLCAPI/', - 'builtin_admin_id':'root@localhost.localdomain', - 'builtin_admin_password':'root', - 'peer_admin_name':'plc2@planet-lab.org', - 'peer_admin_password':'peer', - 'peer_admin_key':'ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAIEAlsX+X+sgN/rsNizPhsXMdHzArxVdwN1KJMG4vTY1m0KQMFJSilaX6xlMIirKhFmNDrVkrOPT2On59K4fDFjGgDq9gMMfdaAfEhxDdy2x1k/H/vJRQE/CqkoRZE8mVAt/cgMsOHTLJTDxBm/0RedaBlcXBpCwqBi3n05sGkS2BjM= VTHD-tester-ssh-v2', + 'builtin-admin-id':'root@plc2.org', + 'builtin-admin-password':'root', + 'peer-admin-name':'plc2@planet-lab.org', + 'peer-admin-password':'peer', 'node-format':'n2%02d.plc2.org', 'plainname' : 'two', 'slice-format' : 's2%02d', + 'person-format' : 'user2-%d@plc2.org', + 'key-format':'ssh-rsa 2222222222222222 user%d-key%d', + 'person-password' : 'password2', } #################### # set initial conditions -def define_test (nodes,slices): - global number_nodes, number_slices +def define_test (keys,persons,nodes,slices): + global number_keys, number_persons, number_nodes, number_slices + number_keys=keys + number_persons=persons number_nodes=nodes number_slices=slices def fast(): - define_test(1,1) + define_test(1,1,1,1) -define_test (nodes=5,slices=3) +define_test (keys=4,persons=2,nodes=5,slices=3) # predefined stuff +# number of 'system' persons +# builtin maint, local root, 2 persons for the peering +system_persons = 4 +# among that, 1 gets refreshed - other ones have conflicting names +system_persons_cross = 1 + system_slices_ids = (1,2) def system_slices (): return len(system_slices_ids) +def total_slices (): + return number_slices+system_slices() + # temporary - the myplc I use doesnt know about 'system' yet def system_slivers (): # return len(system_slices_ids) return 0 - -def total_slices (): - return number_slices+system_slices() def total_slivers (): return number_slices+system_slivers() @@ -114,6 +126,12 @@ def node_name (i,n): def slice_name (i,n): return plc[i]['plainname']+'_'+plc[i]['slice-format']%n +def person_name (i,n): + return plc[i]['person-format']%n + +def key_name (i,n,k): + return plc[i]['key-format']%(n,k) + # to have indexes start at 1 def myrange (n): return range (1,n+1,1) @@ -135,39 +153,58 @@ def test00_init (args=[1,2]): plc[i]['url']=url s[i]=xmlrpclib.ServerProxy(url,allow_none=True) print 'initializing s[%d]'%i,url - aa[i]={'Username':plc[i]['builtin_admin_id'], + aa[i]={'Username':plc[i]['builtin-admin-id'], 'AuthMethod':'password', - 'AuthString':plc[i]['builtin_admin_password'], + 'AuthString':plc[i]['builtin-admin-password'], 'Role':'admin'} print 'initialized aa[%d]'%i, aa[i] - a[i]={'Username':plc[i]['peer_admin_name'], + a[i]={'Username':plc[i]['peer-admin-name'], 'AuthMethod':'password', - 'AuthString':plc[i]['peer_admin_password'], + 'AuthString':plc[i]['peer-admin-password'], 'Role':'admin'} print 'initialized a[%d]'%i, a[i] def test00_print (args=[1,2]): - global plc,s,a,aa for i in args: print 's[%d]'%i,s[i] print 'aa[%d]'%i, aa[i] print 'a[%d]'%i, a[i] -def check_nodes (en,ef,args=[1,2]): - global plc,s,a +def check_nodes (el,ef,args=[1,2]): for i in args: # use a single request and sort afterwards for efficiency # could have used GetNodes's scope as well all_nodes = s[i].GetNodes(a[i]) n = len ([ x for x in all_nodes if x['peer_id'] is None]) f = len ([ x for x in all_nodes if x['peer_id'] is not None]) - print '%02d: Checking nodes: got %d local (e=%d) & %d foreign (e=%d)'%(i,n,en,f,ef) - assert n==en + print '%02d: Checking nodes: got %d local (e=%d) & %d foreign (e=%d)'%(i,n,el,f,ef) + assert n==el + assert f==ef + +def check_keys (el,ef,args=[1,2]): + for i in args: + # use a single request and sort afterwards for efficiency + # could have used GetKeys's scope as well + all_keys = s[i].GetKeys(a[i]) + n = len ([ x for x in all_keys if x['peer_id'] is None]) + f = len ([ x for x in all_keys if x['peer_id'] is not None]) + print '%02d: Checking keys: got %d local (e=%d) & %d foreign (e=%d)'%(i,n,el,f,ef) + assert n==el + assert f==ef + +def check_persons (el,ef,args=[1,2]): + for i in args: + # use a single request and sort afterwards for efficiency + # could have used GetPersons's scope as well + all_persons = s[i].GetPersons(a[i]) + n = len ([ x for x in all_persons if x['peer_id'] is None]) + f = len ([ x for x in all_persons if x['peer_id'] is not None]) + print '%02d: Checking persons: got %d local (e=%d) & %d foreign (e=%d)'%(i,n,el,f,ef) + assert n==el assert f==ef # expected : local slices, foreign slices def check_slices (els,efs,args=[1,2]): - global plc,s,a for i in args: ls=len(s[i].GetSlices(a[i],{'peer_id':None})) fs=len(s[i].GetSlices(a[i],{'~peer_id':None})) @@ -193,7 +230,6 @@ def check_slice_nodes (expected_nodes, is_local_slice, args=[1,2]): check_slice_nodes_n (ns,expected_nodes, is_local_slice, args) def check_slice_nodes_n (ns,expected_nodes, is_local_slice, args=[1,2]): - global plc,s,a for i in args: peer=peer_index(i) if is_local_slice: @@ -225,7 +261,6 @@ def check_conf_files (args=[1,2]): check_conf_files_n (nn,args) def check_conf_files_n (nn,args=[1,2]): - global plc,s,a for i in args: nodename=node_name(i,nn) ndict= s[i].GetSlivers(a[i],[nodename])[0] @@ -250,7 +285,6 @@ def check_slivers_1 (esn,args=[1,2]): check_slivers_n (1,esn,args) def check_slivers_n (nn,esn,args=[1,2]): - global plc,s,a for i in args: nodename=node_name(i,nn) ndict= s[i].GetSlivers(a[i],[nodename])[0] @@ -266,74 +300,46 @@ def check_slivers_n (nn,esn,args=[1,2]): #################### def test00_admin_person (args=[1,2]): - global plc,s,a + global plc for i in args: - email = plc[i]['peer_admin_name'] + email = plc[i]['peer-admin-name'] try: p=s[i].GetPersons(a[i],[email])[0] - plc[i]['peer_admin_id']=p['person_id'] + plc[i]['peer-admin-id']=p['person_id'] except: person_id=s[i].AddPerson(aa[i],{'first_name':'Local', 'last_name':'PeerPoint', 'role_ids':[10], 'email':email, - 'password':plc[i]['peer_admin_password']}) - print '%02d: created peer admin account %d, %s - %s'%(i, - person_id,plc[i]['peer_admin_name'], - plc[i]['peer_admin_password']) - plc[i]['peer_admin_id']=person_id - s[i].AddPersonKey(aa[i],email,{'key_type':'ssh', - 'key':plc[i]['peer_admin_key']}) - print '%02d: added key to peer admin '%i + 'password':plc[i]['peer-admin-password']}) + print '%02d:== created peer admin account %d, %s - %s'%(i, + person_id,plc[i]['peer-admin-name'], + plc[i]['peer-admin-password']) + plc[i]['peer-admin-id']=person_id def test00_admin_enable (args=[1,2]): - global plc,s,a for i in args: - s[i].AdmSetPersonEnabled(aa[i],plc[i]['peer_admin_id'],True) - s[i].AddRoleToPerson(aa[i],'admin',plc[i]['peer_admin_id']) - print '%02d: enabled+admin on account %d:%s'%(i,plc[i]['peer_admin_id'],plc[i]['peer_admin_name']) - -#################### -def test01_site (args=[1,2]): - global plc,s,a - for i in args: - peer=peer_index(i) - ### create a site (required for creating a slice) - sitename=site_name(i) - abbrev_name="abbr"+str(i) - login_base=plc[i]['plainname'] - ### should be enough - needs to check we can add/del slices - max_slices = number_slices - try: - s[i].GetSites(a[i],{'login_base':login_base})[0] - except: - site_id=s[i].AddSite (a[i], {'name':plc_name(i), - 'abbreviated_name': abbrev_name, - 'login_base': login_base, - 'is_public': True, - 'url': 'http://%s.com/'%abbrev_name, - 'max_slices':max_slices}) - ### max_slices does not seem taken into account at that stage - s[i].UpdateSite(a[i],site_id,{'max_slices':max_slices}) - print '%02d: Created site %d with max_slices=%d'%(i,site_id,max_slices) - plc[i]['site_id']=site_id + s[i].AdmSetPersonEnabled(aa[i],plc[i]['peer-admin-id'],True) + s[i].AddRoleToPerson(aa[i],'admin',plc[i]['peer-admin-id']) + print '%02d:== enabled+admin on account %d:%s'%(i,plc[i]['peer-admin-id'],plc[i]['peer-admin-name']) def test01_peer_person (args=[1,2]): - global plc,s,a + global plc for i in args: peer=peer_index(i) - email=plc[peer]['peer_admin_name'] + email=plc[peer]['peer-admin-name'] try: p=s[i].GetPersons(a[i],[email])[0] plc[i]['peer_person_id']=p['person_id'] except: person_id = s[i].AddPerson (a[i], {'first_name':'Peering(plain passwd)', 'last_name':plc_name(peer), 'role_ids':[3000], - 'email':email,'password':plc[peer]['peer_admin_password']}) + 'email':email,'password':plc[peer]['peer-admin-password']}) print '%02d:Created person %d as the peer person'%(i,person_id) plc[i]['peer_person_id']=person_id +#################### def test01_peer (args=[1,2]): - global plc,s,a + global plc for i in args: peer=peer_index(i) peername = plc_name(peer) @@ -348,13 +354,36 @@ def test01_peer (args=[1,2]): print "PLEASE manually set password for person_id=%d in DB%d"%(plc[i]['peer_person_id'],i) def test01_peer_passwd (args=[1,2]): - global plc,s,a for i in args: # using an ad-hoc local command for now - never could get quotes to reach sql.... print "Attempting to set passwd for person_id=%d in DB%d"%(plc[i]['peer_person_id'],i), retcod=os.system("ssh root@%s new_plc_api/person-password.sh %d"%(plc[i]['hostname'],plc[i]['peer_person_id'])) print '-> system returns',retcod +def test01_site (args=[1,2]): + global plc + for i in args: + peer=peer_index(i) + ### create a site (required for creating a slice) + sitename=site_name(i) + abbrev_name="abbr"+str(i) + login_base=plc[i]['plainname'] + ### should be enough - needs to check we can add/del slices + max_slices = number_slices + try: + s[i].GetSites(a[i],{'login_base':login_base})[0] + except: + site_id=s[i].AddSite (a[i], {'name':plc_name(i), + 'abbreviated_name': abbrev_name, + 'login_base': login_base, + 'is_public': True, + 'url': 'http://%s.com/'%abbrev_name, + 'max_slices':max_slices}) + ### max_slices does not seem taken into account at that stage + s[i].UpdateSite(a[i],site_id,{'max_slices':max_slices}) + print '%02d:== Created site %d with max_slices=%d'%(i,site_id,max_slices) + plc[i]['site_id']=site_id + ############################## # this one gets cached def get_peer_id (i): @@ -367,10 +396,9 @@ def get_peer_id (i): return peer_id def test01_refresh (message,args=[1,2]): - global plc,s,a print '=== refresh',message for i in args: - print '%02d: Refreshing peer'%(i), + print '%02d:== Refreshing peer'%(i), retcod=s[i].RefreshPeer(a[i],get_peer_id(i)) print 'got ',retcod @@ -381,12 +409,11 @@ def get_local_node_id(i,nodename): # clean all local nodes - foreign nodes are not supposed to be cleaned up manually def clean_all_nodes (args=[1,2]): - global plc,s,a for i in args: - print '%02d: Cleaning all nodes'%i + print '%02d:== Cleaning all nodes'%i loc_nodes = s[i].GetNodes(a[i],None,None,'local') for node in loc_nodes: - print '%02d: > Cleaning node %d'%(i,node['node_id']) + print '%02d:==== Cleaning node %d'%(i,node['node_id']) s[i].DeleteNode(a[i],node['node_id']) def test02_node (args=[1,2]): @@ -394,36 +421,67 @@ def test02_node (args=[1,2]): test02_node_n (nn,args) def test02_node_n (nn,args=[1,2]): - global plc,s,a for i in args: nodename = node_name(i,nn) try: get_local_node_id(i,nodename) except: n=s[i].AddNode(a[i],1,{'hostname': nodename}) - print '%02d: Added node %d %s'%(i,n,node_name(i,i)) + print '%02d:== Added node %d %s'%(i,n,node_name(i,i)) def test02_delnode (args=[1,2]): for nn in myrange(number_nodes): test02_delnode_n (nn,args) def test02_delnode_n (nn,args=[1,2]): - global plc,s,a for i in args: nodename = node_name(i,nn) node_id = get_local_node_id (i,nodename) retcod=s[i].DeleteNode(a[i],nodename) - print '%02d: Deleted node %d, returns %s'%(i,node_id,retcod) + print '%02d:== Deleted node %d, returns %s'%(i,node_id,retcod) + +#################### +def test05_person (args=[1,2]): + for np in myrange(number_persons): + test05_person_n (np,True,args) + +def test05_del_person (args=[1,2]): + for np in myrange(number_persons): + test05_person_n (np,False,args) + +def test05_person_n (np,add_if_true,args=[1,2]): + test05_person_n_ks (np, myrange(number_keys),add_if_true,args) + +def test05_person_n_ks (np,nks,add_if_true,args=[1,2]): + for i in args: + email = person_name(i,np) + try: + person_id=s[i].GetPersons(a[i],[email])[0]['person_id'] + if not add_if_true: + s[i].DeletePerson(a[i],person_id) + print "%02d:== deleted person_id %d"%(i,person_id) + except: + if add_if_true: + password = plc[i]['person-password'] + person_id=s[i].AddPerson(a[i],{'first_name':'Your average', + 'last_name':'User%d'%np, + 'role_ids':[30], + 'email':email, + 'password': password }) + print '%02d:== created user account %d, %s - %s'%(i, person_id,email,password) + for nk in nks: + key=key_name(i,np,nk) + s[i].AddPersonKey(aa[i],email,{'key_type':'ssh', 'key':key}) + print '%02d:== added key %s to person %s'%(i,key,email) #################### def clean_all_slices (args=[1,2]): - global plc,s,a for i in args: - print '%02d: Cleaning all slices'%i + print '%02d:== Cleaning all slices'%i for slice in s[i].GetSlices(a[i],{'peer_id':None}): slice_id = slice['slice_id'] if slice_id not in system_slices_ids: - print '%02d: > Cleaning slice %d'%(i,slice_id) + print '%02d:==== Cleaning slice %d'%(i,slice_id) s[i].DeleteSlice(a[i],slice_id) def get_local_slice_id (i,name): @@ -434,7 +492,6 @@ def test03_slice (args=[1,2]): test03_slice_n (n,args) def test03_slice_n (ns,args=[1,2]): - global plc,s,a for i in args: peer=peer_index(i) plcname=plc_name(i) @@ -449,7 +506,7 @@ def test03_slice_n (ns,args=[1,2]): 'max_nodes':max_nodes, 'instanciation':'plc-instantiated', }) - print '%02d: created slice %d - max nodes=%d'%(i,slice_id,max_nodes) + print '%02d:== created slice %d - max nodes=%d'%(i,slice_id,max_nodes) def test04_node_slice (is_local, add_if_true, args=[1,2]): @@ -460,7 +517,6 @@ def test04_node_slice_ns (ns,is_local, add_if_true, args=[1,2]): test04_node_slice_nl_n (myrange(number_nodes),ns,is_local, add_if_true, args) def test04_node_slice_nl_n (nnl,ns,is_local, add_if_true, args=[1,2]): - global plc,s,a for i in args: peer=peer_index(i) slice_id = get_local_slice_id (i,slice_name (i,ns)) @@ -477,7 +533,7 @@ def test04_node_slice_nl_n (nnl,ns,is_local, add_if_true, args=[1,2]): else: s[i].DeleteSliceFromNodes (a[i], slice_id,hostnames) message="deleted" - print '%02d: %s in slice %d %s '%(i,message,slice_id,nodetype), + print '%02d:== %s in slice %d %s '%(i,message,slice_id,nodetype), print hostnames def test04_slice_add_lnode (args=[1,2]): @@ -617,8 +673,26 @@ def test_all_slices (): test_all_addslices () test_all_delslices () +def test_all_persons (): + test05_del_person() + check_keys(0,0) + check_persons(system_persons,0) + test01_refresh ('before persons&keys creation') + check_keys(0,0) + check_persons(system_persons,system_persons_cross) + message ("Creating persons&keys - 1 extra del/add cycle for unique indexes") + test05_person () + test05_del_person([2]) + test05_person([2]) + check_keys(number_persons*number_keys,0) + check_persons(system_persons+number_persons,system_persons_cross) + test01_refresh ('after persons&keys creation') + check_keys(number_persons*number_keys,number_persons*number_keys) + check_persons(system_persons+number_persons,system_persons_cross+number_persons) + def test_all (): test_all_init () + test_all_persons () test_all_nodes () test_all_slices () @@ -633,6 +707,7 @@ def populate (): def test_now (): test_all_init() + test_all_persons () # clean_all_nodes() # clean_all_slices() # populate() diff --git a/TestPeers.ref b/TestPeers.ref new file mode 100644 index 0000000..d3a69a7 --- /dev/null +++ b/TestPeers.ref @@ -0,0 +1,405 @@ +5 nodes & 3 slices +==================== ('INIT',) +initializing s[1] https://lurch.cs.princeton.edu:443/PLCAPI/ +initialized aa[1] {'Username': 'root@plc1.org', 'AuthMethod': 'password', 'Role': 'admin', 'AuthString': 'root'} +initialized a[1] {'Username': 'plc1@planet-lab.org', 'AuthMethod': 'password', 'Role': 'admin', 'AuthString': 'peer'} +initializing s[2] https://planetlab-devbox.inria.fr:443/PLCAPI/ +initialized aa[2] {'Username': 'root@plc2.org', 'AuthMethod': 'password', 'Role': 'admin', 'AuthString': 'root'} +initialized a[2] {'Username': 'plc2@planet-lab.org', 'AuthMethod': 'password', 'Role': 'admin', 'AuthString': 'peer'} +s[1] +aa[1] {'Username': 'root@plc1.org', 'AuthMethod': 'password', 'Role': 'admin', 'AuthString': 'root'} +a[1] {'Username': 'plc1@planet-lab.org', 'AuthMethod': 'password', 'Role': 'admin', 'AuthString': 'peer'} +s[2] +aa[2] {'Username': 'root@plc2.org', 'AuthMethod': 'password', 'Role': 'admin', 'AuthString': 'root'} +a[2] {'Username': 'plc2@planet-lab.org', 'AuthMethod': 'password', 'Role': 'admin', 'AuthString': 'peer'} +01:== created peer admin account 3, plc1@planet-lab.org - peer +02:== created peer admin account 3, plc2@planet-lab.org - peer +01:== enabled+admin on account 3:plc1@planet-lab.org +02:== enabled+admin on account 3:plc2@planet-lab.org +01:Created person 4 as the peer person +02:Created person 4 as the peer person +01:Created peer 1 +PLEASE manually set password for person_id=4 in DB1 +02:Created peer 1 +PLEASE manually set password for person_id=4 in DB2 +Attempting to set passwd for person_id=4 in DB1UPDATE 1 + -> system returns 0 +Attempting to set passwd for person_id=4 in DB2UPDATE 1 + -> system returns 0 +01:== Created site 2 with max_slices=3 +02:== Created site 2 with max_slices=3 +01: Checking keys: got 0 local (e=0) & 0 foreign (e=0) +02: Checking keys: got 0 local (e=0) & 0 foreign (e=0) +01: Checking persons: got 4 local (e=4) & 0 foreign (e=0) +02: Checking persons: got 4 local (e=4) & 0 foreign (e=0) +=== refresh before persons&keys creation +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 1, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 1, 'new_nodes': 0} +01: Checking keys: got 0 local (e=0) & 0 foreign (e=0) +02: Checking keys: got 0 local (e=0) & 0 foreign (e=0) +01: Checking persons: got 4 local (e=4) & 1 foreign (e=1) +02: Checking persons: got 4 local (e=4) & 1 foreign (e=1) +==================== ('Creating persons&keys - 1 extra del/add cycle for unique indexes',) +01:== created user account 6, user1-1@plc1.org - password1 +01:== added key ssh-rsa 1111111111111111 user1-key1 to person user1-1@plc1.org +01:== added key ssh-rsa 1111111111111111 user1-key2 to person user1-1@plc1.org +01:== added key ssh-rsa 1111111111111111 user1-key3 to person user1-1@plc1.org +01:== added key ssh-rsa 1111111111111111 user1-key4 to person user1-1@plc1.org +02:== created user account 6, user2-1@plc2.org - password2 +02:== added key ssh-rsa 2222222222222222 user1-key1 to person user2-1@plc2.org +02:== added key ssh-rsa 2222222222222222 user1-key2 to person user2-1@plc2.org +02:== added key ssh-rsa 2222222222222222 user1-key3 to person user2-1@plc2.org +02:== added key ssh-rsa 2222222222222222 user1-key4 to person user2-1@plc2.org +01:== created user account 7, user1-2@plc1.org - password1 +01:== added key ssh-rsa 1111111111111111 user2-key1 to person user1-2@plc1.org +01:== added key ssh-rsa 1111111111111111 user2-key2 to person user1-2@plc1.org +01:== added key ssh-rsa 1111111111111111 user2-key3 to person user1-2@plc1.org +01:== added key ssh-rsa 1111111111111111 user2-key4 to person user1-2@plc1.org +02:== created user account 7, user2-2@plc2.org - password2 +02:== added key ssh-rsa 2222222222222222 user2-key1 to person user2-2@plc2.org +02:== added key ssh-rsa 2222222222222222 user2-key2 to person user2-2@plc2.org +02:== added key ssh-rsa 2222222222222222 user2-key3 to person user2-2@plc2.org +02:== added key ssh-rsa 2222222222222222 user2-key4 to person user2-2@plc2.org +02:== deleted person_id 6 +02:== deleted person_id 7 +02:== created user account 8, user2-1@plc2.org - password2 +02:== added key ssh-rsa 2222222222222222 user1-key1 to person user2-1@plc2.org +02:== added key ssh-rsa 2222222222222222 user1-key2 to person user2-1@plc2.org +02:== added key ssh-rsa 2222222222222222 user1-key3 to person user2-1@plc2.org +02:== added key ssh-rsa 2222222222222222 user1-key4 to person user2-1@plc2.org +02:== created user account 9, user2-2@plc2.org - password2 +02:== added key ssh-rsa 2222222222222222 user2-key1 to person user2-2@plc2.org +02:== added key ssh-rsa 2222222222222222 user2-key2 to person user2-2@plc2.org +02:== added key ssh-rsa 2222222222222222 user2-key3 to person user2-2@plc2.org +02:== added key ssh-rsa 2222222222222222 user2-key4 to person user2-2@plc2.org +01: Checking keys: got 8 local (e=8) & 0 foreign (e=0) +02: Checking keys: got 8 local (e=8) & 0 foreign (e=0) +01: Checking persons: got 6 local (e=6) & 1 foreign (e=1) +02: Checking persons: got 6 local (e=6) & 1 foreign (e=1) +=== refresh after persons&keys creation +01:== Refreshing peer got {'new_keys': 8, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 2, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 8, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 2, 'new_nodes': 0} +01: Checking keys: got 8 local (e=8) & 8 foreign (e=8) +02: Checking keys: got 8 local (e=8) & 8 foreign (e=8) +01: Checking persons: got 6 local (e=6) & 3 foreign (e=3) +02: Checking persons: got 6 local (e=6) & 3 foreign (e=3) +==================== ('RESETTING NODES',) +01:== Cleaning all nodes +02:== Cleaning all nodes +=== refresh cleaned nodes +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +01: Checking nodes: got 0 local (e=0) & 0 foreign (e=0) +02: Checking nodes: got 0 local (e=0) & 0 foreign (e=0) +==================== ('CREATING NODES',) +01:== Added node 1 n101.plc1.org +02:== Added node 1 n202.plc2.org +01:== Added node 2 n101.plc1.org +02:== Added node 2 n202.plc2.org +01:== Added node 3 n101.plc1.org +02:== Added node 3 n202.plc2.org +01:== Added node 4 n101.plc1.org +02:== Added node 4 n202.plc2.org +01:== Added node 5 n101.plc1.org +02:== Added node 5 n202.plc2.org +01: Checking nodes: got 5 local (e=5) & 0 foreign (e=0) +02: Checking nodes: got 5 local (e=5) & 0 foreign (e=0) +=== refresh after node creation +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 5} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 5} +01: Checking nodes: got 5 local (e=5) & 5 foreign (e=5) +02: Checking nodes: got 5 local (e=5) & 5 foreign (e=5) +==================== ('2 extra del/add cycles on plc2 for different indexes',) +02:== Deleted node 1, returns 1 +02:== Deleted node 2, returns 1 +02:== Deleted node 3, returns 1 +02:== Deleted node 4, returns 1 +02:== Deleted node 5, returns 1 +02:== Added node 11 n202.plc2.org +02:== Added node 12 n202.plc2.org +02:== Added node 13 n202.plc2.org +02:== Added node 14 n202.plc2.org +02:== Added node 15 n202.plc2.org +02:== Deleted node 11, returns 1 +02:== Deleted node 12, returns 1 +02:== Deleted node 13, returns 1 +02:== Deleted node 14, returns 1 +02:== Deleted node 15, returns 1 +02:== Added node 16 n202.plc2.org +02:== Added node 17 n202.plc2.org +02:== Added node 18 n202.plc2.org +02:== Added node 19 n202.plc2.org +02:== Added node 20 n202.plc2.org +02:== Deleted node 16, returns 1 +02:== Deleted node 17, returns 1 +02:== Deleted node 18, returns 1 +02:== Deleted node 19, returns 1 +02:== Deleted node 20, returns 1 +02: Checking nodes: got 0 local (e=0) & 5 foreign (e=5) +=== refresh after deletion on plc2 +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': -5} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +01: Checking nodes: got 5 local (e=5) & 0 foreign (e=0) +02: Checking nodes: got 0 local (e=0) & 5 foreign (e=5) +==================== ('ADD on plc2 for different indexes',) +02:== Added node 21 n202.plc2.org +02:== Added node 22 n202.plc2.org +02:== Added node 23 n202.plc2.org +02:== Added node 24 n202.plc2.org +02:== Added node 25 n202.plc2.org +01: Checking nodes: got 5 local (e=5) & 0 foreign (e=0) +02: Checking nodes: got 5 local (e=5) & 5 foreign (e=5) +=== refresh after re-creation on plc2 +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 5} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +01: Checking nodes: got 5 local (e=5) & 5 foreign (e=5) +02: Checking nodes: got 5 local (e=5) & 5 foreign (e=5) +==================== ('RESETTING SLICES TEST',) +01:== Cleaning all nodes +01:==== Cleaning node 1 +01:==== Cleaning node 2 +01:==== Cleaning node 3 +01:==== Cleaning node 4 +01:==== Cleaning node 5 +02:== Cleaning all nodes +02:==== Cleaning node 21 +02:==== Cleaning node 22 +02:==== Cleaning node 23 +02:==== Cleaning node 24 +02:==== Cleaning node 25 +01:== Added node 16 n101.plc1.org +02:== Added node 26 n202.plc2.org +01:== Added node 17 n101.plc1.org +02:== Added node 27 n202.plc2.org +01:== Added node 18 n101.plc1.org +02:== Added node 28 n202.plc2.org +01:== Added node 19 n101.plc1.org +02:== Added node 29 n202.plc2.org +01:== Added node 20 n101.plc1.org +02:== Added node 30 n202.plc2.org +01:== Cleaning all slices +01:==== Cleaning slice 3 +02:== Cleaning all slices +02:==== Cleaning slice 3 +=== refresh After slices init +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +==================== ('CREATING SLICES on plc1',) +01:== created slice 4 - max nodes=5 +01:== created slice 5 - max nodes=5 +01:== created slice 6 - max nodes=5 +01: Checking slices: got 5 local (e=5) & 0 foreign (e=0) +02: Checking slices: got 2 local (e=2) & 0 foreign (e=0) +=== refresh after slice created on plc1 +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 3, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +01: Checking slices: got 5 local (e=5) & 0 foreign (e=0) +02: Checking slices: got 2 local (e=2) & 3 foreign (e=3) +01: local slice one_s101 (e=0) on nodes [] +01: local slice one_s102 (e=0) on nodes [] +01: local slice one_s103 (e=0) on nodes [] +02: foreign slice one_s101 (e=0) on nodes [] +02: foreign slice one_s102 (e=0) on nodes [] +02: foreign slice one_s103 (e=0) on nodes [] +==================== ('ADDING LOCAL NODES IN SLICES',) +01:== added in slice 4 local ['n101.plc1.org', 'n102.plc1.org', 'n103.plc1.org', 'n104.plc1.org', 'n105.plc1.org'] +01:== added in slice 5 local ['n101.plc1.org', 'n102.plc1.org', 'n103.plc1.org', 'n104.plc1.org', 'n105.plc1.org'] +01:== added in slice 6 local ['n101.plc1.org', 'n102.plc1.org', 'n103.plc1.org', 'n104.plc1.org', 'n105.plc1.org'] +01: local slice one_s101 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: local slice one_s102 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: local slice one_s103 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s101 (e=0) on nodes [] +02: foreign slice one_s102 (e=0) on nodes [] +02: foreign slice one_s103 (e=0) on nodes [] +=== refresh After local nodes were added on plc1 +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +01: local slice one_s101 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: local slice one_s102 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: local slice one_s103 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s101 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s102 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s103 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +==================== ('ADDING FOREIGN NODES IN SLICES',) +01:== added in slice 4 foreign ['n201.plc2.org', 'n202.plc2.org', 'n203.plc2.org', 'n204.plc2.org', 'n205.plc2.org'] +01:== added in slice 5 foreign ['n201.plc2.org', 'n202.plc2.org', 'n203.plc2.org', 'n204.plc2.org', 'n205.plc2.org'] +01:== added in slice 6 foreign ['n201.plc2.org', 'n202.plc2.org', 'n203.plc2.org', 'n204.plc2.org', 'n205.plc2.org'] +01: local slice one_s101 (e=10) on nodes [16, 17, 18, 19, 20, 11, 12, 13, 14, 15] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +[FOR:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +01: local slice one_s102 (e=10) on nodes [16, 17, 18, 19, 20, 11, 12, 13, 14, 15] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +[FOR:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +01: local slice one_s103 (e=10) on nodes [16, 17, 18, 19, 20, 11, 12, 13, 14, 15] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +[FOR:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +02: foreign slice one_s101 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s102 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s103 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +=== refresh After foreign nodes were added in plc1 +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +01: local slice one_s101 (e=10) on nodes [16, 17, 18, 19, 20, 11, 12, 13, 14, 15] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +[FOR:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +01: local slice one_s102 (e=10) on nodes [16, 17, 18, 19, 20, 11, 12, 13, 14, 15] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +[FOR:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +01: local slice one_s103 (e=10) on nodes [16, 17, 18, 19, 20, 11, 12, 13, 14, 15] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +[FOR:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +02: foreign slice one_s101 (e=10) on nodes [8, 9, 10, 6, 7, 26, 27, 28, 29, 30] +[LOC:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s102 (e=10) on nodes [8, 9, 10, 6, 7, 26, 27, 28, 29, 30] +[LOC:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s103 (e=10) on nodes [8, 9, 10, 6, 7, 26, 27, 28, 29, 30] +[LOC:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: 3 slivers (exp. 3) in GetSlivers for node n101.plc1.org +>>slivername = one_s101 +{ 'attributes': [], + 'expires': normalized, + 'instantiation': 'plc-instantiated', + 'keys': [], + 'name': 'one_s101', + 'slice_id': 4} +>>slivername = one_s102 +{ 'attributes': [], + 'expires': normalized, + 'instantiation': 'plc-instantiated', + 'keys': [], + 'name': 'one_s102', + 'slice_id': 5} +>>slivername = one_s103 +{ 'attributes': [], + 'expires': normalized, + 'instantiation': 'plc-instantiated', + 'keys': [], + 'name': 'one_s103', + 'slice_id': 6} +02: 3 slivers (exp. 3) in GetSlivers for node n201.plc2.org +>>slivername = one_s101 +{ 'attributes': [], + 'expires': normalized, + 'instantiation': 'plc-instantiated', + 'keys': [], + 'name': 'one_s101', + 'slice_id': 4} +>>slivername = one_s102 +{ 'attributes': [], + 'expires': normalized, + 'instantiation': 'plc-instantiated', + 'keys': [], + 'name': 'one_s102', + 'slice_id': 5} +>>slivername = one_s103 +{ 'attributes': [], + 'expires': normalized, + 'instantiation': 'plc-instantiated', + 'keys': [], + 'name': 'one_s103', + 'slice_id': 6} +==================== ('DELETING FOREIGN NODES FROM SLICES',) +01:== deleted in slice 4 foreign ['n201.plc2.org', 'n202.plc2.org', 'n203.plc2.org', 'n204.plc2.org', 'n205.plc2.org'] +01:== deleted in slice 5 foreign ['n201.plc2.org', 'n202.plc2.org', 'n203.plc2.org', 'n204.plc2.org', 'n205.plc2.org'] +01:== deleted in slice 6 foreign ['n201.plc2.org', 'n202.plc2.org', 'n203.plc2.org', 'n204.plc2.org', 'n205.plc2.org'] +01: local slice one_s101 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: local slice one_s102 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: local slice one_s103 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s101 (e=10) on nodes [8, 9, 10, 6, 7, 26, 27, 28, 29, 30] +[LOC:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s102 (e=10) on nodes [8, 9, 10, 6, 7, 26, 27, 28, 29, 30] +[LOC:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s103 (e=10) on nodes [8, 9, 10, 6, 7, 26, 27, 28, 29, 30] +[LOC:5] : n201.plc2.org n202.plc2.org n203.plc2.org n204.plc2.org n205.plc2.org +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: 3 slivers (exp. 3) in GetSlivers for node n101.plc1.org +>>slivername = one_s101 +{ 'attributes': [], + 'expires': normalized, + 'instantiation': 'plc-instantiated', + 'keys': [], + 'name': 'one_s101', + 'slice_id': 4} +>>slivername = one_s102 +{ 'attributes': [], + 'expires': normalized, + 'instantiation': 'plc-instantiated', + 'keys': [], + 'name': 'one_s102', + 'slice_id': 5} +>>slivername = one_s103 +{ 'attributes': [], + 'expires': normalized, + 'instantiation': 'plc-instantiated', + 'keys': [], + 'name': 'one_s103', + 'slice_id': 6} +=== refresh After foreign nodes were removed on plc1 +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +01: local slice one_s101 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: local slice one_s102 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +01: local slice one_s103 (e=5) on nodes [16, 17, 18, 19, 20] +[LOC:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s101 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s102 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s103 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +==================== ('DELETING LOCAL NODES FROM SLICES',) +01:== deleted in slice 4 local ['n101.plc1.org', 'n102.plc1.org', 'n103.plc1.org', 'n104.plc1.org', 'n105.plc1.org'] +01:== deleted in slice 5 local ['n101.plc1.org', 'n102.plc1.org', 'n103.plc1.org', 'n104.plc1.org', 'n105.plc1.org'] +01:== deleted in slice 6 local ['n101.plc1.org', 'n102.plc1.org', 'n103.plc1.org', 'n104.plc1.org', 'n105.plc1.org'] +01: local slice one_s101 (e=0) on nodes [] +01: local slice one_s102 (e=0) on nodes [] +01: local slice one_s103 (e=0) on nodes [] +02: foreign slice one_s101 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s102 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +02: foreign slice one_s103 (e=5) on nodes [8, 9, 10, 6, 7] +[FOR:5] : n101.plc1.org n102.plc1.org n103.plc1.org n104.plc1.org n105.plc1.org +=== refresh After local nodes were removed on plc1 +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +01: local slice one_s101 (e=0) on nodes [] +01: local slice one_s102 (e=0) on nodes [] +01: local slice one_s103 (e=0) on nodes [] +02: foreign slice one_s101 (e=0) on nodes [] +02: foreign slice one_s102 (e=0) on nodes [] +02: foreign slice one_s103 (e=0) on nodes [] +==================== ('CHECKING SLICES CLEAN UP',) +01:== Cleaning all slices +01:==== Cleaning slice 4 +01:==== Cleaning slice 5 +01:==== Cleaning slice 6 +01: Checking slices: got 2 local (e=2) & 0 foreign (e=0) +02: Checking slices: got 2 local (e=2) & 3 foreign (e=3) +=== refresh After slices clenaup +01:== Refreshing peer got {'new_keys': 0, 'new_slices': 0, 'plcname': 'Thierry plc1', 'new_persons': 0, 'new_nodes': 0} +02:== Refreshing peer got {'new_keys': 0, 'new_slices': -3, 'plcname': 'Thierry plc2 on devbox', 'new_persons': 0, 'new_nodes': 0} +01: Checking slices: got 2 local (e=2) & 0 foreign (e=0) +02: Checking slices: got 2 local (e=2) & 0 foreign (e=0) diff --git a/planetlab4.sql b/planetlab4.sql index 82636c2..54afd43 100644 --- a/planetlab4.sql +++ b/planetlab4.sql @@ -9,7 +9,7 @@ -- -- Copyright (C) 2006 The Trustees of Princeton University -- --- $Id: planetlab4.sql,v 1.42 2006/11/21 10:57:00 thierry Exp $ +-- $Id: planetlab4.sql,v 1.43 2006/11/23 19:35:38 thierry Exp $ -- -------------------------------------------------------------------------------- @@ -79,7 +79,9 @@ CREATE TABLE persons ( -- Timestamps date_created timestamp without time zone NOT NULL DEFAULT CURRENT_TIMESTAMP, - last_updated timestamp without time zone NOT NULL DEFAULT CURRENT_TIMESTAMP + last_updated timestamp without time zone NOT NULL DEFAULT CURRENT_TIMESTAMP, + + peer_id integer REFERENCES peers -- From which peer ) WITH OIDS; CREATE INDEX persons_email_idx ON persons (email) WHERE deleted IS false; @@ -798,6 +800,7 @@ persons.title, persons.phone, persons.url, persons.bio, +persons.peer_id, CAST(date_part('epoch', persons.date_created) AS bigint) AS date_created, CAST(date_part('epoch', persons.last_updated) AS bigint) AS last_updated, COALESCE(person_roles.role_ids, '{}') AS role_ids, -- 2.43.0