X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=sfa%2Fimporter%2Fplimporter.py;h=d281c61479a09106f1e93202499399ea8c4daee7;hb=b54e90d0c3868fa5afb89ff784cba0d5d36265ad;hp=1e04e54ca13f59c93469073bfeede4a7146205aa;hpb=f38626e863c9e78b95d4b4fd11c9d6e2ccd05205;p=sfa.git diff --git a/sfa/importer/plimporter.py b/sfa/importer/plimporter.py index 1e04e54c..d281c614 100644 --- a/sfa/importer/plimporter.py +++ b/sfa/importer/plimporter.py @@ -24,7 +24,9 @@ from sfa.util.xrn import Xrn, get_leaf, get_authority, hrn_to_urn from sfa.trust.gid import create_uuid from sfa.trust.certificate import convert_public_key, Keypair -from sfa.storage.alchemy import dbsession +# using global alchemy.session() here is fine +# as importer is on standalone one-shot process +from sfa.storage.alchemy import global_dbsession from sfa.storage.model import RegRecord, RegAuthority, RegSlice, RegNode, RegUser, RegKey from sfa.planetlab.plshell import PlShell @@ -115,8 +117,8 @@ class PlImporter: pointer=site['site_id'], authority=get_authority(site_hrn)) auth_record.just_created() - dbsession.add(auth_record) - dbsession.commit() + global_dbsession.add(auth_record) + global_dbsession.commit() self.logger.info("PlImporter: Imported authority (vini site) %s"%auth_record) self.remember_record ( site_record ) @@ -127,7 +129,7 @@ class PlImporter: shell = PlShell (config) ######## retrieve all existing SFA objects - all_records = dbsession.query(RegRecord).all() + all_records = global_dbsession.query(RegRecord).all() # create hash by (type,hrn) # we essentially use this to know if a given record is already known to SFA @@ -145,7 +147,7 @@ class PlImporter: # Get all plc sites # retrieve only required stuf sites = shell.GetSites({'peer_id': None, 'enabled' : True}, - ['site_id','login_base','node_ids','slice_ids','person_ids',]) + ['site_id','login_base','node_ids','slice_ids','person_ids', 'name']) # create a hash of sites by login_base # sites_by_login_base = dict ( [ ( site['login_base'], site ) for site in sites ] ) # Get all plc users @@ -153,6 +155,9 @@ class PlImporter: ['person_id', 'email', 'key_ids', 'site_ids', 'role_ids']) # create a hash of persons by person_id persons_by_id = dict ( [ ( person['person_id'], person) for person in persons ] ) + # also gather non-enabled user accounts so as to issue relevant warnings + disabled_persons = shell.GetPersons({'peer_id': None, 'enabled': False}, ['person_id']) + disabled_person_ids = [ person['person_id'] for person in disabled_persons ] # Get all plc public keys # accumulate key ids for keys retrieval key_ids = [] @@ -167,9 +172,13 @@ class PlImporter: for person in persons: pubkeys = [] for key_id in person['key_ids']: - key = keys_by_id[key_id] - if key['key_type'] == 'ssh': + # by construction all the keys we fetched are ssh keys + # so gpg keys won't be in there + try: + key = keys_by_id[key_id] pubkeys.append(key) + except: + self.logger.warning("Could not spot key %d - probably non-ssh"%key_id) keys_by_person_id[person['person_id']] = pubkeys # Get all plc nodes nodes = shell.GetNodes( {'peer_id': None}, ['node_id', 'hostname', 'site_id']) @@ -185,6 +194,9 @@ class PlImporter: # start importing for site in sites: + if site['name'].startswith('sfa:'): + continue + site_hrn = _get_site_hrn(interface_hrn, site) # import if hrn is not in list of existing hrns or if the hrn exists # but its not a site record @@ -199,14 +211,14 @@ class PlImporter: pointer=site['site_id'], authority=get_authority(site_hrn)) site_record.just_created() - dbsession.add(site_record) - dbsession.commit() + global_dbsession.add(site_record) + global_dbsession.commit() self.logger.info("PlImporter: imported authority (site) : %s" % site_record) self.remember_record (site_record) except: # if the site import fails then there is no point in trying to import the # site's child records (node, slices, persons), so skip them. - self.logger.log_exc("PlImporter: failed to import site. Skipping child records") + self.logger.log_exc("PlImporter: failed to import site %s. Skipping child records"%site_hrn) continue else: # xxx update the record ... @@ -235,12 +247,13 @@ class PlImporter: pointer =node['node_id'], authority=get_authority(node_hrn)) node_record.just_created() - dbsession.add(node_record) - dbsession.commit() + global_dbsession.add(node_record) + global_dbsession.commit() self.logger.info("PlImporter: imported node: %s" % node_record) self.remember_record (node_record) except: - self.logger.log_exc("PlImporter: failed to import node") + self.logger.log_exc("PlImporter: failed to import node %s"%node_hrn) + continue else: # xxx update the record ... pass @@ -249,10 +262,17 @@ class PlImporter: site_pis=[] # import persons for person_id in site['person_ids']: - try: - person = persons_by_id[person_id] - except: - self.logger.warning ("PlImporter: cannot locate person_id %s - ignored"%person_id) + proceed=False + if person_id in persons_by_id: + person=persons_by_id[person_id] + proceed=True + elif person_id in disabled_person_ids: + pass + else: + self.logger.warning ("PlImporter: cannot locate person_id %s in site %s - ignored"%(person_id,site_hrn)) + # make sure to NOT run this if anything is wrong + if not proceed: continue + person_hrn = email_to_hrn(site_hrn, person['email']) # xxx suspicious again if len(person_hrn) > 64: person_hrn = person_hrn[:64] @@ -282,45 +302,66 @@ class PlImporter: plc_keys = keys_by_person_id.get(person['person_id'],[]) if not user_record: (pubkey,pkey) = init_person_key (person, plc_keys ) - person_gid = self.auth_hierarchy.create_gid(person_urn, create_uuid(), pkey) - person_gid.set_email(person['email']) + person_gid = self.auth_hierarchy.create_gid(person_urn, create_uuid(), pkey, email=person['email']) user_record = RegUser (hrn=person_hrn, gid=person_gid, - pointer=person['person_id'], - authority=get_authority(person_hrn), - email=person['email']) + pointer=person['person_id'], + authority=get_authority(person_hrn), + email=person['email']) if pubkey: user_record.reg_keys=[RegKey (pubkey['key'], pubkey['key_id'])] else: self.logger.warning("No key found for user %s"%user_record) user_record.just_created() - dbsession.add (user_record) - dbsession.commit() + global_dbsession.add (user_record) + global_dbsession.commit() self.logger.info("PlImporter: imported person: %s" % user_record) self.remember_record ( user_record ) else: # update the record ? - # if user's primary key has changed then we need to update the + # + # if a user key has changed then we need to update the # users gid by forcing an update here + # + # right now, SFA only has *one* key attached to a user, and this is + # the key that the GID was made with + # so the logic here is, we consider that things are OK (unchanged) if + # all the SFA keys are present as PLC keys + # otherwise we trigger the creation of a new gid from *some* plc key + # and record this on the SFA side + # it would make sense to add a feature in PLC so that one could pick a 'primary' + # key but this is not available on the myplc side for now + # = or = it would be much better to support several keys in SFA but that + # does not seem doable without a major overhaul in the data model as + # a GID is attached to a hrn, but it's also linked to a key, so... + # NOTE: with this logic, the first key entered in PLC remains the one + # current in SFA until it is removed from PLC sfa_keys = user_record.reg_keys - def key_in_list (key,sfa_keys): - for reg_key in sfa_keys: - if reg_key.key==key['key']: return True + def sfa_key_in_list (sfa_key,plc_keys): + for plc_key in plc_keys: + if plc_key['key']==sfa_key.key: + return True return False - # is there a new key in myplc ? + # are all the SFA keys known to PLC ? new_keys=False - for key in plc_keys: - if not key_in_list (key,sfa_keys): - new_keys = True + if not sfa_keys and plc_keys: + new_keys=True + else: + for sfa_key in sfa_keys: + if not sfa_key_in_list (sfa_key,plc_keys): + new_keys = True if new_keys: (pubkey,pkey) = init_person_key (person, plc_keys) person_gid = self.auth_hierarchy.create_gid(person_urn, create_uuid(), pkey) + person_gid.set_email(person['email']) if not pubkey: user_record.reg_keys=[] else: user_record.reg_keys=[ RegKey (pubkey['key'], pubkey['key_id'])] + user_record.gid = person_gid + user_record.just_updated() self.logger.info("PlImporter: updated person: %s" % user_record) user_record.email = person['email'] - dbsession.commit() + global_dbsession.commit() user_record.stale=False # accumulate PIs - PLCAPI has a limitation that when someone has PI role # this is valid for all sites she is in.. @@ -331,7 +372,14 @@ class PlImporter: self.logger.log_exc("PlImporter: failed to import person %d %s"%(person['person_id'],person['email'])) # maintain the list of PIs for a given site - site_record.reg_pis = site_pis + # for the record, Jordan had proposed the following addition as a welcome hotfix to a previous version: + # site_pis = list(set(site_pis)) + # this was likely due to a bug in the above logic, that had to do with disabled persons + # being improperly handled, and where the whole loop on persons + # could be performed twice with the same person... + # so hopefully we do not need to eliminate duplicates explicitly here anymore + site_record.reg_pis = list(set(site_pis)) + global_dbsession.commit() # import slices for slice_id in site['slice_ids']: @@ -350,20 +398,21 @@ class PlImporter: pointer=slice['slice_id'], authority=get_authority(slice_hrn)) slice_record.just_created() - dbsession.add(slice_record) - dbsession.commit() + global_dbsession.add(slice_record) + global_dbsession.commit() self.logger.info("PlImporter: imported slice: %s" % slice_record) self.remember_record ( slice_record ) except: - self.logger.log_exc("PlImporter: failed to import slice") + self.logger.log_exc("PlImporter: failed to import slice %s (%s)"%(slice_hrn,slice['name'])) else: # xxx update the record ... - self.logger.warning ("Slice update not yet implemented") + # given that we record the current set of users anyways, there does not seem to be much left to do here + # self.logger.warning ("Slice update not yet implemented on slice %s (%s)"%(slice_hrn,slice['name'])) pass # record current users affiliated with the slice slice_record.reg_researchers = \ [ self.locate_by_type_pointer ('user',user_id) for user_id in slice['person_ids'] ] - dbsession.commit() + global_dbsession.commit() slice_record.stale=False ### remove stale records @@ -385,5 +434,5 @@ class PlImporter: self.logger.warning("stale not found with %s"%record) if stale: self.logger.info("PlImporter: deleting stale record: %s" % record) - dbsession.delete(record) - dbsession.commit() + global_dbsession.delete(record) + global_dbsession.commit()