X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=sfa%2Fplc%2Fsfa-import-plc.py;h=39cb28c066384c6857f226bf628a2b5597ce6a7f;hb=952322d76247f8991f3c2688ed7e1f5a22ca4572;hp=ae6d4d6e60354b2c36f80511f4a1a4b355101ca6;hpb=0766ddc71af955bb2529bc839094aea26da4190c;p=sfa.git diff --git a/sfa/plc/sfa-import-plc.py b/sfa/plc/sfa-import-plc.py index ae6d4d6e..39cb28c0 100755 --- a/sfa/plc/sfa-import-plc.py +++ b/sfa/plc/sfa-import-plc.py @@ -4,8 +4,8 @@ ### $URL$ # ## -# Import PLC records into the Geni database. It is indended that this tool be -# run once to create Geni records that reflect the current state of the +# Import PLC records into the SFA database. It is indended that this tool be +# run once to create SFA records that reflect the current state of the # planetlab database. # # The import tool assumes that the existing PLC hierarchy should all be part @@ -20,20 +20,19 @@ import getopt import sys import tempfile - +import logging.handlers +import logging from sfa.util.record import * -from sfa.util.genitable import GeniTable -from sfa.util.misc import * +from sfa.util.table import SfaTable +from sfa.util.namespace import * from sfa.util.config import Config -from sfa.util.report import trace, error - from sfa.trust.certificate import convert_public_key, Keypair from sfa.trust.trustedroot import * from sfa.trust.hierarchy import * +from sfa.plc.api import * from sfa.trust.gid import create_uuid from sfa.plc.sfaImport import * - - +from sfa.util.report import trace, error def process_options(): global hrn @@ -43,87 +42,130 @@ def process_options(): name = opt[0] val = opt[1] + +def load_keys(filename): + keys = {} + tmp_dict = {} + try: + execfile(filename, tmp_dict) + if 'keys' in tmp_dict: + keys = tmp_dict['keys'] + return keys + except: + return keys + +def save_keys(filename, keys): + f = open(filename, 'w') + f.write("keys = %s" % str(keys)) + f.close() + def main(): + # setup the logger + LOGFILE='/var/log/sfa_import_plc.log' + logging.basicConfig(level=logging.INFO, + format='%(asctime)s - %(message)s', + filename=LOGFILE) + rotate_handler = logging.handlers.RotatingFileHandler(LOGFILE, maxBytes=1000000, backupCount=5) + logger = logging.getLogger() + logger.addHandler(rotate_handler) + process_options() config = Config() + if not config.SFA_REGISTRY_ENABLED: + sys.exit(0) root_auth = config.SFA_REGISTRY_ROOT_AUTH - level1_auth = config.SFA_REGISTRY_LEVEL1_AUTH - sfaImporter = sfaImport() + interface_hrn = config.SFA_INTERFACE_HRN + keys_filename = config.config_path + os.sep + 'person_keys.py' + sfaImporter = sfaImport(logger) shell = sfaImporter.shell plc_auth = sfaImporter.plc_auth AuthHierarchy = sfaImporter.AuthHierarchy TrustedRoots = sfaImporter.TrustedRoots - table = GeniTable() + table = SfaTable() + if not table.exists(): - table.create() + table.create() - if not level1_auth or level1_auth in ['']: - level1_auth = None - - print "Import: creating top level authorities" - if not level1_auth: - sfaImporter.create_top_level_auth_records(root_auth) - import_auth = root_auth - else: - if not AuthHierarchy.auth_exists(level1_auth): - AuthHierarchy.create_auth(level1_auth) - sfaImporter.create_top_level_auth_records(level1_auth) - import_auth = level1_auth - - print "Import: adding", import_auth, "to trusted list" - authority = AuthHierarchy.get_auth_info(import_auth) + # create root authority + sfaImporter.create_top_level_auth_records(root_auth) + if not root_auth == interface_hrn: + sfaImporter.create_top_level_auth_records(interface_hrn) + + trace("Import: adding " + interface_hrn + " to trusted list", logger) + authority = AuthHierarchy.get_auth_info(interface_hrn) TrustedRoots.add_gid(authority.get_gid_object()) - if ".vini" in import_auth and import_auth.endswith('vini'): + if ".vini" in interface_hrn and interface_hrn.endswith('vini'): # create a fake internet2 site first i2site = {'name': 'Internet2', 'abbreviated_name': 'I2', 'login_base': 'internet2', 'site_id': -1} - sfaImporter.import_site(import_auth, i2site) + sfaImporter.import_site(interface_hrn, i2site) # create dict of all existing sfa records existing_records = {} existing_hrns = [] + key_ids = [] + person_keys = {} results = table.find() for result in results: existing_records[(result['hrn'], result['type'])] = result existing_hrns.append(result['hrn']) # Get all plc sites - sites = shell.GetSites(plc_auth) + sites = shell.GetSites(plc_auth, {'peer_id': None}) + sites_dict = {} + for site in sites: + sites_dict[site['login_base']] = site # Get all plc users - persons = shell.GetPersons(plc_auth, {}, ['person_id', 'email', 'key_ids']) + persons = shell.GetPersons(plc_auth, {'peer_id': None}, ['person_id', 'email', 'key_ids', 'site_ids']) persons_dict = {} for person in persons: persons_dict[person['person_id']] = person + key_ids.extend(person['key_ids']) + + # Get all public keys + keys = shell.GetKeys(plc_auth, {'peer_id': None, 'key_id': key_ids}) + keys_dict = {} + for key in keys: + keys_dict[key['key_id']] = key['key'] + + # create a dict of person keys keyed on key_id + old_person_keys = load_keys(keys_filename) + for person in persons: + pubkeys = [] + for key_id in person['key_ids']: + pubkeys.append(keys_dict[key_id]) + person_keys[person['person_id']] = pubkeys # Get all plc nodes - nodes = shell.GetNodes(plc_auth, {}, ['node_id', 'hostname']) + nodes = shell.GetNodes(plc_auth, {'peer_id': None}, ['node_id', 'hostname', 'site_id']) nodes_dict = {} for node in nodes: nodes_dict[node['node_id']] = node # Get all plc slices - slices = shell.GetSlices(plc_auth, {}, ['slice_id', 'name']) + slices = shell.GetSlices(plc_auth, {'peer_id': None}, ['slice_id', 'name']) slices_dict = {} for slice in slices: slices_dict[slice['slice_id']] = slice - # start importing for site in sites: - site_hrn = import_auth + "." + site['login_base'] + site_hrn = interface_hrn + "." + site['login_base'] + print "Importing site: %s" % site_hrn + # import if hrn is not in list of existing hrns or if the hrn exists # but its not a site record if site_hrn not in existing_hrns or \ (site_hrn, 'authority') not in existing_records: - sfaImporter.import_site(import_auth, site) + site_hrn = sfaImporter.import_site(interface_hrn, site) # import node records for node_id in site['node_ids']: if node_id not in nodes_dict: continue node = nodes_dict[node_id] - hrn = hostname_to_hrn(import_auth, site['login_base'], node['hostname']) + hrn = hostname_to_hrn(interface_hrn, site['login_base'], node['hostname']) if hrn not in existing_hrns or \ (hrn, 'node') not in existing_records: sfaImporter.import_node(site_hrn, node) @@ -133,7 +175,7 @@ def main(): if slice_id not in slices_dict: continue slice = slices_dict[slice_id] - hrn = slicename_to_hrn(import_auth, slice['name']) + hrn = slicename_to_hrn(interface_hrn, slice['name']) if hrn not in existing_hrns or \ (hrn, 'slice') not in existing_records: sfaImporter.import_slice(site_hrn, slice) @@ -144,56 +186,88 @@ def main(): continue person = persons_dict[person_id] hrn = email_to_hrn(site_hrn, person['email']) + old_keys = [] + new_keys = [] + if person_id in old_person_keys: + old_keys = old_person_keys[person_id] + if person_id in person_keys: + new_keys = person_keys[person_id] + update_record = False + for key in new_keys: + if key not in old_keys: + update_record = True + if hrn not in existing_hrns or \ - (hrn, 'user') not in existing_records: + (hrn, 'user') not in existing_records or update_record: sfaImporter.import_person(site_hrn, person) + # remove stale records + for (record_hrn, type) in existing_records.keys(): + record = existing_records[(record_hrn, type)] + # if this is the interface name dont do anything + if record_hrn == interface_hrn or \ + record_hrn == root_auth or \ + record['peer_authority']: + continue + # dont delete vini's internet2 placeholdder record + # normally this would be deleted becuase it does not have a plc record + if ".vini" in interface_hrn and interface_hrn.endswith('vini') and \ + record_hrn.endswith("internet2"): + continue + + found = False - # remove any record in existing_hrns that does not - # have a plc record - site_existing_records_only = lambda (r_hrn, r_type): r_hrn.startswith(site_hrn) - site_existing_records = filter(site_existing_records_only, existing_records.keys()) - for (record_hrn, type) in site_existing_records: - found = False - if type == 'user': - for person in persons_dict.values(): - tmp_hrn = email_to_hrn(site_hrn, person['email']) - if record_hrn == tmp_hrn: - found = True + if type == 'authority': + for site in sites: + site_hrn = interface_hrn + "." + site['login_base'] + if site_hrn == record_hrn and site['site_id'] == record['pointer']: + found = True + break - elif type == 'node': - for node in nodes_dict.values(): - tmp_hrn = hostname_to_hrn(import_auth, site['login_base'], node['hostname']) - if record_hrn == tmp_hrn: - found = True - elif type == 'slice': - for slice in slices_dict.values(): - tmp_hrn = slicename_to_hrn(import_auth, slice['name']) - if record_hrn == tmp_hrn: + elif type == 'user': + login_base = get_leaf(get_authority(record_hrn)) + username = get_leaf(record_hrn) + if login_base in sites_dict: + site = sites_dict[login_base] + for person in persons: + tmp_username = person['email'].split("@")[0] + alt_username = person['email'].split("@")[0].replace(".", "_") + if username in [tmp_username, alt_username] and \ + site['site_id'] in person['site_ids'] and \ + person['person_id'] == record['pointer']: found = True - else: - continue + break + + elif type == 'slice': + slicename = hrn_to_pl_slicename(record_hrn) + for slice in slices: + if slicename == slice['name'] and \ + slice['slice_id'] == record['pointer']: + found = True + break - if not found: - trace("Import: Removing %s %s" % (type, record_hrn)) - record_object = existing_records[(record_hrn, type)] - sfaImporter.delete_record(record_hrn, type) - - # remove stale site_records - site_records_only = lambda(r_hrn, r_type): r_type == 'authority' and r_hrn != import_auth - site_records = filter(site_records_only, existing_records.keys()) - for (record_hrn, type) in site_records: - found = False - for site in sites: - site_hrn = import_auth + "." + site['login_base'] - if site_hrn == record_hrn: - found = True + elif type == 'node': + login_base = get_leaf(get_authority(record_hrn)) + nodename = get_leaf(record_hrn) + if login_base in sites_dict: + site = sites_dict[login_base] + for node in nodes: + tmp_nodename = node['hostname'].split(".")[0] + if tmp_nodename == nodename and \ + node['site_id'] == site['site_id'] and \ + node['node_id'] == record['pointer']: + found = True + break + else: + continue + if not found: - trace("Import: Removing %s %s" % (type, record_hrn)) record_object = existing_records[(record_hrn, type)] sfaImporter.delete_record(record_hrn, type) - + # save pub keys + trace('Import: saving current pub keys', logger) + save_keys(keys_filename, person_keys) if __name__ == "__main__": main()