dbsession.commit()
self.logger.info("PlImporter: Imported authority (vini site) %s"%auth_record)
+ def locate_by_type_hrn (self, type, hrn):
+ return self.records_by_type_hrn.get ( (type, hrn), None)
+
def run (self, options):
config = Config ()
interface_hrn = config.SFA_INTERFACE_HRN
shell = PlShell (config)
######## retrieve all existing SFA objects
- records = dbsession.query(RegRecord)
+ all_records = dbsession.query(RegRecord).all()
+
# create indexes / hashes by (type,hrn)
- self.records_by_type_hrn = dict ( [ ( (record.type, record.hrn) , record ) for record in records ] )
- # and by (type,pointer)
- self.records_by_type_pointer = \
- dict ( [ ( (record.type, record.pointer) , record ) for record in records if record.pointer != -1 ] )
+ self.records_by_type_hrn = dict ( [ ( (record.type, record.hrn) , record ) for record in all_records ] )
+# and by (type,pointer)
+# the idea was to try and retrieve an sfa record from the pointer data,
+# in case the plc name (e.g. email address) has changed
+# and the new hrn can't be located in the current sfa records
+# however it sounds like it's eventually safer to just create a new sfa record
+# and let the old one get garbage-collected with stale records
+# self.records_by_type_pointer = \
+# dict ( [ ( (record.type, record.pointer) , record ) for record in all_records if record.pointer != -1 ] )
- ######## retrieve PLC records
+ # initialize record.stale to True by default, then mark stale=False on the ones that are in use
+ for record in all_records: record.stale=True
+
+ ######## retrieve PLC data
# Get all plc sites
# retrieve only required stuf
sites = shell.GetSites({'peer_id': None, 'enabled' : True},
['site_id','login_base','node_ids','slice_ids','person_ids',])
# create a hash of sites by login_base
sites_by_login_base = dict ( [ ( site['login_base'], site ) for site in sites ] )
-
# Get all plc users
persons = shell.GetPersons({'peer_id': None, 'enabled': True},
['person_id', 'email', 'key_ids', 'site_ids'])
# create a hash of persons by person_id
persons_by_id = dict ( [ ( person['person_id'], person) for person in persons ] )
-
# Get all plc public keys
# accumulate key ids for keys retrieval
key_ids = []
for person in persons:
key_ids.extend(person['key_ids'])
keys = shell.GetKeys( {'peer_id': None, 'key_id': key_ids} )
-
# create a hash of keys by key_id
keys_by_id = dict ( [ ( key['key_id'], key ) for key in keys ] )
-
# create a dict person_id -> [ (plc)keys ]
keys_by_person_id = {}
for person in persons:
for key_id in person['key_ids']:
pubkeys.append(keys_by_id[key_id])
keys_by_person_id[person['person_id']] = pubkeys
-
# Get all plc nodes
nodes = shell.GetNodes( {'peer_id': None}, ['node_id', 'hostname', 'site_id'])
# create hash by node_id
nodes_by_id = dict ( [ ( node['node_id'], node, ) for node in nodes ] )
-
# Get all plc slices
slices = shell.GetSlices( {'peer_id': None}, ['slice_id', 'name'])
# create hash by slice_id
# import if hrn is not in list of existing hrns or if the hrn exists
# but its not a site record
- if ( 'authority', site_hrn, ) not in self.records_by_type_hrn:
+ site_record=self.locate_by_type_hrn ('authority', site_hrn)
+ if not site_record:
try:
urn = hrn_to_urn(site_hrn, 'authority')
if not self.auth_hierarchy.auth_exists(urn):
self.auth_hierarchy.create_auth(urn)
auth_info = self.auth_hierarchy.get_auth_info(urn)
- auth_record = RegAuthority(hrn=site_hrn, gid=auth_info.get_gid_object(),
+ site_record = RegAuthority(hrn=site_hrn, gid=auth_info.get_gid_object(),
pointer=site['site_id'],
authority=get_authority(site_hrn))
- auth_record.just_created()
- dbsession.add(auth_record)
+ site_record.just_created()
+ dbsession.add(site_record)
dbsession.commit()
- self.logger.info("PlImporter: imported authority (site) : %s" % auth_record)
+ self.logger.info("PlImporter: imported authority (site) : %s" % site_record)
except:
# if the site import fails then there is no point in trying to import the
# site's child records (node, slices, persons), so skip them.
self.logger.log_exc("PlImporter: failed to import site. Skipping child records")
continue
+ site_record.stale=False
# import node records
for node_id in site['node_ids']:
- if node_id not in nodes_by_id:
+ try:
+ node = nodes_by_id[node_id]
+ except:
+ self.logger.warning ("PlImporter: cannot locate node_id %s - ignored"%node_id)
continue
- node = nodes_by_id[node_id]
site_auth = get_authority(site_hrn)
- site_name = get_leaf(site_hrn)
+ site_name = site['login_base']
hrn = hostname_to_hrn(site_auth, site_name, node['hostname'])
- if len(hrn) > 64:
- hrn = hrn[:64]
- if ( 'node', hrn, ) not in self.records_by_type_hrn:
+ # xxx this sounds suspicious
+ if len(hrn) > 64: hrn = hrn[:64]
+ node_record = self.locate_by_type_hrn ( 'node', hrn )
+ if not node_record:
try:
pkey = Keypair(create=True)
urn = hrn_to_urn(hrn, 'node')
self.logger.info("PlImporter: imported node: %s" % node_record)
except:
self.logger.log_exc("PlImporter: failed to import node")
-
+ node_record.stale=False
# import slices
for slice_id in site['slice_ids']:
- if slice_id not in slices_by_id:
- continue
- slice = slices_by_id[slice_id]
- hrn = slicename_to_hrn(interface_hrn, slice['name'])
- if ( 'slice', hrn, ) not in self.records_by_type_hrn:
+ try:
+ slice = slices_by_id[slice_id]
+ except:
+ self.logger.warning ("PlImporter: cannot locate slice_id %s - ignored"%slice_id)
+ slice_hrn = slicename_to_hrn(interface_hrn, slice['name'])
+ slice_record = self.locate_by_type_hrn ('slice', slice_hrn)
+ if not slice_record:
try:
pkey = Keypair(create=True)
- urn = hrn_to_urn(hrn, 'slice')
+ urn = hrn_to_urn(slice_hrn, 'slice')
slice_gid = self.auth_hierarchy.create_gid(urn, create_uuid(), pkey)
- slice_record = RegSlice (hrn=hrn, gid=slice_gid,
+ slice_record = RegSlice (hrn=slice_hrn, gid=slice_gid,
pointer=slice['slice_id'],
- authority=get_authority(hrn))
+ authority=get_authority(slice_hrn))
slice_record.just_created()
dbsession.add(slice_record)
dbsession.commit()
self.logger.info("PlImporter: imported slice: %s" % slice_record)
except:
- self.logger.log_exc("PlImporter: failed to import slice")
+ self.logger.log_exc("PlImporter: failed to import slice")
+ slice_record.stale=False
# import persons
for person_id in site['person_ids']:
- if person_id not in persons_by_id:
- self.logger.warning ("PlImporter: skipping person %s"%person_id)
- continue
- person = persons_by_id[person_id]
- hrn = email_to_hrn(site_hrn, person['email'])
- if len(hrn) > 64:
- hrn = hrn[:64]
-
- previous_record = self.records_by_type_hrn.get( ( 'user', hrn, ) )
- if not previous_record:
- previous_record = self.records_by_type_pointer.get ( ('user', person_id,) )
+ try:
+ person = persons_by_id[person_id]
+ except:
+ self.logger.warning ("PlImporter: cannot locate person_id %s - ignored"%person_id)
+ person_hrn = email_to_hrn(site_hrn, person['email'])
+ # xxx suspicious again
+ if len(person_hrn) > 64: person_hrn = person_hrn[:64]
+
+ person_record = self.locate_by_type_hrn( 'user', person_hrn)
+# see above
+# if not person_record:
+# person_record = self.records_by_type_pointer.get ( ('user', person_id,) )
# if user's primary key has changed then we need to update the
# users gid by forcing an update here
plc_keys = []
sfa_keys = []
- if previous_record:
- sfa_keys = previous_record.reg_keys
+ if person_record:
+ sfa_keys = person_record.reg_keys
if person_id in keys_by_person_id:
plc_keys = keys_by_person_id[person_id]
update_record = False
if not key_in_list (key,sfa_keys):
update_record = True
- if not previous_record or update_record:
+ if not person_record or update_record:
try:
pubkey=None
if 'key_ids' in person and person['key_ids']:
try:
pkey = convert_public_key(pubkey['key'])
except:
- self.logger.warn('PlImporter: unable to convert public key for %s' % hrn)
+ self.logger.warn('PlImporter: unable to convert public key for %s' % person_hrn)
pkey = Keypair(create=True)
else:
# the user has no keys. Creating a random keypair for the user's gid
- self.logger.warn("PlImporter: person %s does not have a PL public key"%hrn)
+ self.logger.warn("PlImporter: person %s does not have a PL public key"%person_hrn)
pkey = Keypair(create=True)
- urn = hrn_to_urn(hrn, 'user')
+ urn = hrn_to_urn(person_hrn, 'user')
person_gid = self.auth_hierarchy.create_gid(urn, create_uuid(), pkey)
- if previous_record:
- previous_record.gid=person_gid
- if pubkey: previous_record.reg_keys=[ RegKey (pubkey['key'], pubkey['key_id'])]
- self.logger.info("PlImporter: updated person: %s" % previous_record)
+ if person_record:
+ person_record.gid=person_gid
+ if pubkey: person_record.reg_keys=[ RegKey (pubkey['key'], pubkey['key_id'])]
+ self.logger.info("PlImporter: updated person: %s" % person_record)
else:
- new_record = RegUser (hrn=hrn, gid=person_gid,
+ person_record = RegUser (hrn=person_hrn, gid=person_gid,
pointer=person['person_id'],
- authority=get_authority(hrn),
+ authority=get_authority(person_hrn),
email=person['email'])
if pubkey:
- new_record.reg_keys=[RegKey (pubkey['key'], pubkey['key_id'])]
+ person_record.reg_keys=[RegKey (pubkey['key'], pubkey['key_id'])]
else:
- logger.warning("No key found for user %s"%new_record)
- dbsession.add (new_record)
+ self.logger.warning("No key found for user %s"%person_record)
+ dbsession.add (person_record)
dbsession.commit()
- self.logger.info("PlImporter: imported person: %s" % new_record)
+ self.logger.info("PlImporter: imported person: %s" % person_record)
except:
- self.logger.log_exc("PlImporter: failed to import person.")
+ self.logger.log_exc("PlImporter: failed to import person %s"%person_id)
+ person_record.stale=False
- # remove stale records
- system_records = [interface_hrn, root_auth, interface_hrn + '.slicemanager']
- for record in records:
- record_hrn=record.hrn
- if record_hrn in system_records:
- continue
+ ### remove stale records
+ # special records must be preserved
+ system_hrns = [interface_hrn, root_auth, interface_hrn + '.slicemanager']
+ for record in all_records:
+ if record.hrn in system_hrns:
+ record.stale=False
if record.peer_authority:
- continue
- type=record.type
- hrn=record.hrn
- # dont delete vini's internet2 placeholdder record
- # normally this would be deleted becuase it does not have a plc record
+ record.stale=False
if ".vini" in interface_hrn and interface_hrn.endswith('vini') and \
- record_hrn.endswith("internet2"):
- continue
-
- found = False
-
- if isinstance (record, RegAuthority):
- for site in sites:
- site_hrn = interface_hrn + "." + site['login_base']
- if site_hrn == record_hrn and site['site_id'] == record.pointer:
- found = True
- break
-
- elif isinstance (record, RegUser):
- login_base = get_leaf(get_authority(record_hrn))
- username = get_leaf(record_hrn)
- if login_base in sites_by_login_base:
- site = sites_by_login_base[login_base]
- for person in persons:
- tmp_username = person['email'].split("@")[0]
- alt_username = person['email'].split("@")[0].replace(".", "_").replace("+", "_")
- if username in [tmp_username, alt_username] and \
- site['site_id'] in person['site_ids'] and \
- person['person_id'] == record.pointer:
- found = True
- break
-
- elif isinstance (record, RegSlice):
- slicename = hrn_to_pl_slicename(record_hrn)
- for slice in slices:
- if slicename == slice['name'] and \
- slice['slice_id'] == record.pointer:
- found = True
- break
-
- elif isinstance (record, RegNode):
- login_base = get_leaf(get_authority(record_hrn))
- nodename = Xrn.unescape(get_leaf(record_hrn))
- if login_base in sites_by_login_base:
- site = sites_by_login_base[login_base]
- for node in nodes:
- tmp_nodename = node['hostname']
- if tmp_nodename == nodename and \
- node['site_id'] == site['site_id'] and \
- node['node_id'] == record.pointer:
- found = True
- break
- else:
- continue
-
- if not found:
- try:
- record_object = self.records_by_type_hrn[ ( type, record_hrn, ) ]
- self.logger.info("PlImporter: deleting record: %s" % record)
- dbsession.delete(record_object)
- dbsession.commit()
- except:
- self.logger.log_exc("PlImporter: failded to delete record")
+ record.hrn.endswith("internet2"):
+ record.stale=False
-# # save pub keys
-# self.logger.info('Import: saving current pub keys')
-# save_keys(keys_filename, person_keys)
-
+ for record in all_records:
+ try: stale=record.stale
+ except:
+ stale=True
+ self.logger.warning("stale not found with %s"%record)
+ if stale:
+ self.logger.info("PlImporter: deleting stale record: %s" % record)
+ dbsession.delete(record)
+ dbsession.commit()