X-Git-Url: http://git.onelab.eu/?a=blobdiff_plain;f=PLC%2FSites.py;h=04372c997eb3724757c0b6cdeb510523bfdc9231;hb=0c8526883d98762265e1a59fdafeb1922ebc8e92;hp=f452f6c3549b42e3453ae172786403988d8a1256;hpb=4104c224401c26638b0be7ac0d9447fbbc25cc2c;p=plcapi.git diff --git a/PLC/Sites.py b/PLC/Sites.py index f452f6c..04372c9 100644 --- a/PLC/Sites.py +++ b/PLC/Sites.py @@ -15,7 +15,7 @@ class Site(Row): """ Representation of a row in the sites table. To use, optionally instantiate with a dict of values. Update as you would a - dict. Commit to the database with flush(). + dict. Commit to the database with sync(). """ fields = { @@ -27,39 +27,16 @@ class Site(Row): 'latitude': Parameter(float, "Decimal latitude of the site", min = -90.0, max = 90.0), 'longitude': Parameter(float, "Decimal longitude of the site", min = -180.0, max = 180.0), 'url': Parameter(str, "URL of a page that describes the site", max = 254), - 'nodegroup_id': Parameter(int, "Identifier of the nodegroup containing the site's nodes"), - 'organization_id': Parameter(int, "Organizational identifier if the site is part of a larger organization"), - 'ext_consortium_id': Parameter(int, "Consortium identifier if the site is part of an external consortium"), - 'date_created': Parameter(str, "Date and time when node entry was created"), + 'date_created': Parameter(str, "Date and time when site entry was created"), + 'last_updated': Parameter(str, "Date and time when site entry was last updated"), 'deleted': Parameter(bool, "Has been deleted"), - } - - # These fields are derived from join tables and are not actually - # in the sites table. - join_fields = { 'max_slices': Parameter(int, "Maximum number of slices that the site is able to create"), - 'site_share': Parameter(float, "Relative resource share for this site's slices"), - } - - # These fields are derived from join tables and are not returned - # by default unless specified. - extra_fields = { 'person_ids': Parameter([int], "List of account identifiers"), - 'slice_ids': Parameter([int], "List of slice identifiers"), - 'defaultattribute_ids': Parameter([int], "List of default slice attribute identifiers"), - 'pcu_ids': Parameter([int], "List of PCU identifiers"), + # 'slice_ids': Parameter([int], "List of slice identifiers"), + # 'pcu_ids': Parameter([int], "List of PCU identifiers"), 'node_ids': Parameter([int], "List of site node identifiers"), } - default_fields = dict(fields.items() + join_fields.items()) - all_fields = dict(default_fields.items() + extra_fields.items()) - - # Number of slices assigned to each site at the time that the site is created - default_max_slices = 0 - - # XXX Useless, unclear what this value means - default_site_share = 1.0 - def __init__(self, api, fields): Row.__init__(self, fields) self.api = api @@ -90,27 +67,6 @@ class Site(Row): return longitude - def validate_nodegroup_id(self, nodegroup_id): - nodegroups = NodeGroups(self.api) - if nodegroup_id not in nodegroups: - raise PLCInvalidArgument, "No such nodegroup" - - return nodegroup_id - - def validate_organization_id(self, organization_id): - organizations = Organizations(self.api) - if role_id not in organizations: - raise PLCInvalidArgument, "No such organization" - - return organization_id - - def validate_ext_consortium_id(self, organization_id): - consortiums = Consortiums(self.api) - if consortium_id not in consortiums: - raise PLCInvalidArgument, "No such consortium" - - return nodegroup_id - def add_person(self, person, commit = True): """ Add person to existing site. @@ -160,7 +116,7 @@ class Site(Row): if 'site_ids' in person and site_id in person['site_ids']: person['site_ids'].remove(site_id) - def flush(self, commit = True): + def sync(self, commit = True): """ Flush changes back to the database. """ @@ -185,24 +141,9 @@ class Site(Row): else: insert = False - # Create site node group if necessary - if 'nodegroup_id' not in self: - rows = self.api.db.selectall("SELECT NEXTVAL('nodegroups_nodegroup_id_seq') as nodegroup_id") - if not rows: - raise PLCDBError, "Unable to fetch new nodegroup_id" - self['nodegroup_id'] = rows[0]['nodegroup_id'] - - nodegroup_id = self['nodegroup_id'] - # XXX Needs a unique name because we cannot delete site node groups yet - name = self['login_base'] + str(self['site_id']) - description = "Nodes at " + self['login_base'] - is_custom = False - self.api.db.do("INSERT INTO nodegroups (nodegroup_id, name, description, is_custom)" \ - " VALUES (%(nodegroup_id)d, %(name)s, %(description)s, %(is_custom)s)", - locals()) - # Filter out fields that cannot be set or updated directly - fields = dict(filter(lambda (key, value): key in self.fields, + sites_fields = self.api.db.fields('sites') + fields = dict(filter(lambda (key, value): key in sites_fields, self.items())) # Parameterize for safety @@ -211,32 +152,16 @@ class Site(Row): if insert: # Insert new row in sites table - self.api.db.do("INSERT INTO sites (%s) VALUES (%s)" % \ - (", ".join(keys), ", ".join(values)), - fields) - - # Setup default slice site info - # XXX Will go away soon - self['max_slices'] = self.default_max_slices - self['site_share'] = self.default_site_share - self.api.db.do("INSERT INTO dslice03_siteinfo (site_id, max_slices, site_share)" \ - " VALUES (%(site_id)d, %(max_slices)d, %(site_share)f)", - self) + sql = "INSERT INTO sites (%s) VALUES (%s)" % \ + (", ".join(keys), ", ".join(values)) else: - # Update default slice site info - # XXX Will go away soon - if 'max_slices' in self and 'site_share' in self: - self.api.db.do("UPDATE dslice03_siteinfo SET " \ - " max_slices = %(max_slices)d, site_share = %(site_share)f" \ - " WHERE site_id = %(site_id)d", - self) - # Update existing row in sites table columns = ["%s = %s" % (key, value) for (key, value) in zip(keys, values)] - self.api.db.do("UPDATE sites SET " + \ - ", ".join(columns) + \ - " WHERE site_id = %(site_id)d", - fields) + sql = "UPDATE sites SET " + \ + ", ".join(columns) + \ + " WHERE site_id = %(site_id)d" + + self.api.db.do(sql, fields) if commit: self.api.db.commit() @@ -248,12 +173,6 @@ class Site(Row): assert 'site_id' in self - # Make sure extra fields are present - sites = Sites(self.api, [self['site_id']], - ['person_ids', 'slice_ids', 'pcu_ids', 'node_ids']) - assert sites - self.update(sites.values()[0]) - # Delete accounts of all people at the site who are not # members of at least one other non-deleted site. persons = PLC.Persons.Persons(self.api, self['person_ids']) @@ -271,33 +190,23 @@ class Site(Row): person.delete(commit = False) # Delete all site slices - slices = Slices(self.api, self['slice_ids']) - for slice in slices.values(): - slice.delete(commit = False) + # slices = Slices(self.api, self['slice_ids']) + # for slice in slices.values(): + # slice.delete(commit = False) # Delete all site PCUs - pcus = PCUs(self.api, self['pcu_ids']) - for pcu in pcus.values(): - pcu.delete(commit = False) + # pcus = PCUs(self.api, self['pcu_ids']) + # for pcu in pcus.values(): + # pcu.delete(commit = False) # Delete all site nodes nodes = Nodes(self.api, self['node_ids']) for node in nodes.values(): node.delete(commit = False) - # Clean up miscellaneous join tables - for table in ['site_authorized_subnets', - 'dslice03_defaultattribute', - 'dslice03_siteinfo']: - self.api.db.do("DELETE FROM %s" \ - " WHERE site_id = %d" % \ - (table, self['site_id'])) - - # XXX Cannot delete site node groups yet - # Mark as deleted self['deleted'] = True - self.flush(commit) + self.sync(commit) class Sites(Table): """ @@ -306,40 +215,11 @@ class Sites(Table): fields. """ - def __init__(self, api, site_id_or_login_base_list = None, extra_fields = []): + def __init__(self, api, site_id_or_login_base_list = None, fields = Site.fields): self.api = api - sql = "SELECT sites.*" \ - ", dslice03_siteinfo.max_slices" - - # N.B.: Joined IDs may be marked as deleted in their primary tables - join_tables = { - # extra_field: (extra_table, extra_column, join_using) - 'person_ids': ('person_site', 'person_id', 'site_id'), - 'slice_ids': ('dslice03_slices', 'slice_id', 'site_id'), - 'defaultattribute_ids': ('dslice03_defaultattribute', 'defaultattribute_id', 'site_id'), - 'pcu_ids': ('pcu', 'pcu_id', 'site_id'), - 'node_ids': ('nodegroup_nodes', 'node_id', 'nodegroup_id'), - } - - extra_fields = filter(join_tables.has_key, extra_fields) - extra_tables = ["%s USING (%s)" % \ - (join_tables[field][0], join_tables[field][2]) \ - for field in extra_fields] - extra_columns = ["%s.%s" % \ - (join_tables[field][0], join_tables[field][1]) \ - for field in extra_fields] - - if extra_columns: - sql += ", " + ", ".join(extra_columns) - - sql += " FROM sites" \ - " LEFT JOIN dslice03_siteinfo USING (site_id)" - - if extra_tables: - sql += " LEFT JOIN " + " LEFT JOIN ".join(extra_tables) - - sql += " WHERE sites.deleted IS False" + sql = "SELECT %s FROM view_sites WHERE deleted IS False" % \ + ", ".join(fields) if site_id_or_login_base_list: # Separate the list into integers and strings @@ -355,9 +235,12 @@ class Sites(Table): sql += ")" rows = self.api.db.selectall(sql) + for row in rows: - if self.has_key(row['site_id']): - site = self[row['site_id']] - site.update(row) - else: - self[row['site_id']] = Site(api, row) + self[row['site_id']] = site = Site(api, row) + for aggregate in ['person_ids', 'slice_ids', + 'defaultattribute_ids', 'pcu_ids', 'node_ids']: + if not site.has_key(aggregate) or site[aggregate] is None: + site[aggregate] = [] + else: + site[aggregate] = map(int, site[aggregate].split(','))