X-Git-Url: http://git.onelab.eu/?p=plcapi.git;a=blobdiff_plain;f=PLC%2FSites.py;h=2075409243700f25d46a64049e74844ca8745ca3;hp=d3d4dcd1b6496b87b1c9fd5ac6cfacf7dbf5885d;hb=13901cd4465288b634103c1997a5653500f2b5cc;hpb=24d16d18acab3da7bccc3e09df4927e9cf2d3246 diff --git a/PLC/Sites.py b/PLC/Sites.py index d3d4dcd..2075409 100644 --- a/PLC/Sites.py +++ b/PLC/Sites.py @@ -2,87 +2,84 @@ from types import StringTypes import string from PLC.Faults import * -from PLC.Parameter import Parameter +from PLC.Parameter import Parameter, Mixed +from PLC.Filter import Filter from PLC.Debug import profile from PLC.Table import Row, Table -from PLC.Persons import Person, Persons from PLC.Slices import Slice, Slices from PLC.PCUs import PCU, PCUs from PLC.Nodes import Node, Nodes -from PLC.NodeGroups import NodeGroup, NodeGroups +from PLC.Addresses import Address, Addresses +from PLC.Persons import Person, Persons class Site(Row): """ Representation of a row in the sites table. To use, optionally instantiate with a dict of values. Update as you would a - dict. Commit to the database with flush(). + dict. Commit to the database with sync(). """ + table_name = 'sites' + primary_key = 'site_id' + join_tables = ['person_site', 'site_address', 'peer_site'] fields = { 'site_id': Parameter(int, "Site identifier"), - 'name': Parameter(str, "Full site name"), - 'abbreviated_name': Parameter(str, "Abbreviated site name"), - 'login_base': Parameter(str, "Site slice prefix"), + 'name': Parameter(str, "Full site name", max = 254), + 'abbreviated_name': Parameter(str, "Abbreviated site name", max = 50), + 'login_base': Parameter(str, "Site slice prefix", max = 32), 'is_public': Parameter(bool, "Publicly viewable site"), - 'latitude': Parameter(float, "Decimal latitude of the site"), - 'longitude': Parameter(float, "Decimal longitude of the site"), - 'url': Parameter(str, "URL of a page that describes the site"), - 'nodegroup_id': Parameter(int, "Identifier of the nodegroup containing the site's nodes"), - 'organization_id': Parameter(int, "Organizational identifier if the site is part of a larger organization"), - 'ext_consortium_id': Parameter(int, "Consortium identifier if the site is part of an external consortium"), - 'date_created': Parameter(str, "Date and time when node entry was created"), - 'deleted': Parameter(bool, "Has been deleted"), - } - - # These fields are derived from join tables and are not actually - # in the sites table. - join_fields = { + 'enabled': Parameter(bool, "Has been enabled"), + 'latitude': Parameter(float, "Decimal latitude of the site", min = -90.0, max = 90.0, nullok = True), + 'longitude': Parameter(float, "Decimal longitude of the site", min = -180.0, max = 180.0, nullok = True), + 'url': Parameter(str, "URL of a page that describes the site", max = 254, nullok = True), + 'date_created': Parameter(int, "Date and time when site entry was created, in seconds since UNIX epoch", ro = True), + 'last_updated': Parameter(int, "Date and time when site entry was last updated, in seconds since UNIX epoch", ro = True), 'max_slices': Parameter(int, "Maximum number of slices that the site is able to create"), - 'site_share': Parameter(float, "Relative resource share for this site's slices"), - } - - # These fields are derived from join tables and are not returned - # by default unless specified. - extra_fields = { + 'max_slivers': Parameter(int, "Maximum number of slivers that the site is able to create"), 'person_ids': Parameter([int], "List of account identifiers"), 'slice_ids': Parameter([int], "List of slice identifiers"), - 'defaultattribute_ids': Parameter([int], "List of default slice attribute identifiers"), + 'address_ids': Parameter([int], "List of address identifiers"), 'pcu_ids': Parameter([int], "List of PCU identifiers"), 'node_ids': Parameter([int], "List of site node identifiers"), + 'peer_id': Parameter(int, "Peer to which this site belongs", nullok = True), + 'peer_site_id': Parameter(int, "Foreign site identifier at peer", nullok = True), + 'site_tag_ids' : Parameter ([int], "List of tags attached to this site"), + 'ext_consortium_id': Parameter(int, "external consortium id", nullok = True) } + related_fields = { + 'persons': [Mixed(Parameter(int, "Person identifier"), + Parameter(str, "Email address"))], + 'addresses': [Mixed(Parameter(int, "Address identifer"), + Filter(Address.fields))] + } + view_tags_name = "view_site_tags" + # tags are used by the Add/Get/Update methods to expose tags + # this is initialized here and updated by the accessors factory + tags = { } - default_fields = dict(fields.items() + join_fields.items()) - all_fields = dict(default_fields.items() + extra_fields.items()) - - # Number of slices assigned to each site at the time that the site is created - default_max_slices = 0 + def validate_name(self, name): + if not len(name): + raise PLCInvalidArgument, "Name must be specified" - # XXX Useless, unclear what this value means - default_site_share = 1.0 + return name - def __init__(self, api, fields): - Row.__init__(self, fields) - self.api = api + validate_abbreviated_name = validate_name def validate_login_base(self, login_base): - if len(login_base) > 20: - raise PLCInvalidArgument, "Login base must be <= 20 characters" + if not len(login_base): + raise PLCInvalidArgument, "Login base must be specified" - if not set(login_base).issubset(string.ascii_letters): - raise PLCInvalidArgument, "Login base must consist only of ASCII letters" + if not set(login_base).issubset(string.lowercase + string.digits + '.'): + raise PLCInvalidArgument, "Login base must consist only of lowercase ASCII letters or numbers or dots" - login_base = login_base.lower() conflicts = Sites(self.api, [login_base]) - for site_id, site in conflicts.iteritems(): - if not site['deleted'] and ('site_id' not in self or self['site_id'] != site_id): + for site in conflicts: + if 'site_id' not in self or self['site_id'] != site['site_id']: raise PLCInvalidArgument, "login_base already in use" return login_base def validate_latitude(self, latitude): - if latitude < -90.0 or latitude > 90.0: - raise PLCInvalidArgument, "Invalid latitude value" - if not self.has_key('longitude') or \ self['longitude'] is None: raise PLCInvalidArgument, "Longitude must also be specified" @@ -90,116 +87,100 @@ class Site(Row): return latitude def validate_longitude(self, longitude): - if longitude < -180.0 or longitude > 180.0: - raise PLCInvalidArgument, "Invalid longitude value" - if not self.has_key('latitude') or \ self['latitude'] is None: raise PLCInvalidArgument, "Latitude must also be specified" return longitude - def validate_nodegroup_id(self, nodegroup_id): - nodegroups = NodeGroups(self.api) - if nodegroup_id not in nodegroups: - raise PLCInvalidArgument, "No such nodegroup" + validate_date_created = Row.validate_timestamp + validate_last_updated = Row.validate_timestamp + + add_person = Row.add_object(Person, 'person_site') + remove_person = Row.remove_object(Person, 'person_site') + + add_address = Row.add_object(Address, 'site_address') + remove_address = Row.remove_object(Address, 'site_address') + + def update_last_updated(self, commit = True): + """ + Update last_updated field with current time + """ + + assert 'site_id' in self + assert self.table_name + + self.api.db.do("UPDATE %s SET last_updated = CURRENT_TIMESTAMP " % (self.table_name) + \ + " where site_id = %d" % (self['site_id']) ) + self.sync(commit) - return nodegroup_id - def validate_organization_id(self, organization_id): - organizations = Organizations(self.api) - if role_id not in organizations: - raise PLCInvalidArgument, "No such organization" + def associate_persons(self, auth, field, value): + """ + Adds persons found in value list to this site (using AddPersonToSite). + Deletes persons not found in value list from this site (using DeletePersonFromSite). + """ + + assert 'person_ids' in self + assert 'site_id' in self + assert isinstance(value, list) + + (person_ids, emails) = self.separate_types(value)[0:2] - return organization_id + # Translate emails into person_ids + if emails: + persons = Persons(self.api, emails, ['person_id']).dict('person_id') + person_ids += persons.keys() - def validate_ext_consortium_id(self, organization_id): - consortiums = Consortiums(self.api) - if consortium_id not in consortiums: - raise PLCInvalidArgument, "No such consortium" + # Add new ids, remove stale ids + if self['person_ids'] != person_ids: + from PLC.Methods.AddPersonToSite import AddPersonToSite + from PLC.Methods.DeletePersonFromSite import DeletePersonFromSite + new_persons = set(person_ids).difference(self['person_ids']) + stale_persons = set(self['person_ids']).difference(person_ids) - return nodegroup_id + for new_person in new_persons: + AddPersonToSite.__call__(AddPersonToSite(self.api), auth, new_person, self['site_id']) + for stale_person in stale_persons: + DeletePersonFromSite.__call__(DeletePersonFromSite(self.api), auth, stale_person, self['site_id']) - def flush(self, commit = True): + def associate_addresses(self, auth, field, value): """ - Flush changes back to the database. + Deletes addresses_ids not found in value list (using DeleteAddress). + Adds address if slice_fields w/o address_id found in value list (using AddSiteAddress). + Update address if slice_fields w/ address_id found in value list (using UpdateAddress). """ - self.validate() - - try: - if not self['name'] or \ - not self['abbreviated_name'] or \ - not self['login_base']: - raise KeyError - except KeyError: - raise PLCInvalidArgument, "name, abbreviated_name, and login_base must all be specified" - - # Fetch a new site_id if necessary - if 'site_id' not in self: - rows = self.api.db.selectall("SELECT NEXTVAL('sites_site_id_seq') AS site_id") - if not rows: - raise PLCDBError, "Unable to fetch new site_id" - self['site_id'] = rows[0]['site_id'] - insert = True - else: - insert = False - - # Create site node group if necessary - if 'nodegroup_id' not in self: - rows = self.api.db.selectall("SELECT NEXTVAL('nodegroups_nodegroup_id_seq') as nodegroup_id") - if not rows: - raise PLCDBError, "Unable to fetch new nodegroup_id" - self['nodegroup_id'] = rows[0]['nodegroup_id'] - - nodegroup_id = self['nodegroup_id'] - # XXX Needs a unique name because we cannot delete site node groups yet - name = self['login_base'] + str(self['site_id']) - description = "Nodes at " + self['name'] - is_custom = False - self.api.db.do("INSERT INTO nodegroups (nodegroup_id, name, description, is_custom)" \ - " VALUES (%(nodegroup_id)d, %(name)s, %(description)s, %(is_custom)s)", - locals()) - - # Filter out fields that cannot be set or updated directly - fields = dict(filter(lambda (key, value): key in self.fields, - self.items())) - - # Parameterize for safety - keys = fields.keys() - values = [self.api.db.param(key, value) for (key, value) in fields.items()] - - if insert: - # Insert new row in sites table - self.api.db.do("INSERT INTO sites (%s) VALUES (%s)" % \ - (", ".join(keys), ", ".join(values)), - fields) - - # Setup default slice site info - # XXX Will go away soon - self['max_slices'] = self.default_max_slices - self['site_share'] = self.default_site_share - self.api.db.do("INSERT INTO dslice03_siteinfo (site_id, max_slices, site_share)" \ - " VALUES (%(site_id)d, %(max_slices)d, %(site_share)f)", - self) - else: - # Update default slice site info - # XXX Will go away soon - if 'max_slices' in self and 'site_share' in self: - self.api.db.do("UPDATE dslice03_siteinfo SET " \ - " max_slices = %(max_slices)d, site_share = %(site_share)f" \ - " WHERE site_id = %(site_id)d", - self) - - # Update existing row in sites table - columns = ["%s = %s" % (key, value) for (key, value) in zip(keys, values)] - self.api.db.do("UPDATE sites SET " + \ - ", ".join(columns) + \ - " WHERE site_id = %(site_id)d", - fields) - - if commit: - self.api.db.commit() + assert 'address_ids' in self + assert 'site_id' in self + assert isinstance(value, list) + + (address_ids, blank, addresses) = self.separate_types(value) + + for address in addresses: + if 'address_id' in address: + address_ids.append(address['address_id']) + + # Add new ids, remove stale ids + if self['address_ids'] != address_ids: + from PLC.Methods.DeleteAddress import DeleteAddress + stale_addresses = set(self['address_ids']).difference(address_ids) + + for stale_address in stale_addresses: + DeleteAddress.__call__(DeleteAddress(self.api), auth, stale_address) + + if addresses: + from PLC.Methods.AddSiteAddress import AddSiteAddress + from PLC.Methods.UpdateAddress import UpdateAddress + + updated_addresses = filter(lambda address: 'address_id' in address, addresses) + added_addresses = filter(lambda address: 'address_id' not in address, addresses) + + for address in added_addresses: + AddSiteAddress.__call__(AddSiteAddress(self.api), auth, self['site_id'], address) + for address in updated_addresses: + address_id = address.pop('address_id') + UpdateAddress.__call__(UpdateAddress(self.api), auth, address_id, address) def delete(self, commit = True): """ @@ -208,116 +189,85 @@ class Site(Row): assert 'site_id' in self - # Make sure extra fields are present - sites = Sites(self.api, [self['site_id']], - ['person_ids', 'slice_ids', 'pcu_ids', 'node_ids']) - assert sites - self.update(sites.values()[0]) - # Delete accounts of all people at the site who are not # members of at least one other non-deleted site. persons = Persons(self.api, self['person_ids']) - for person_id, person in persons.iteritems(): + for person in persons: delete = True person_sites = Sites(self.api, person['site_ids']) - for person_site_id, person_site in person_sites.iteritems(): - if person_site_id != self['site_id'] and \ - not person_site['deleted']: + for person_site in person_sites: + if person_site['site_id'] != self['site_id']: delete = False break if delete: person.delete(commit = False) + # Delete all site addresses + addresses = Addresses(self.api, self['address_ids']) + for address in addresses: + address.delete(commit = False) + # Delete all site slices slices = Slices(self.api, self['slice_ids']) - for slice in slices.values(): + for slice in slices: slice.delete(commit = False) # Delete all site PCUs pcus = PCUs(self.api, self['pcu_ids']) - for pcu in pcus.values(): + for pcu in pcus: pcu.delete(commit = False) # Delete all site nodes nodes = Nodes(self.api, self['node_ids']) - for node in nodes.values(): + for node in nodes: node.delete(commit = False) # Clean up miscellaneous join tables - for table in ['site_authorized_subnets', - 'dslice03_defaultattribute', - 'dslice03_siteinfo']: - self.api.db.do("DELETE FROM %s" \ - " WHERE site_id = %d" % \ + for table in self.join_tables: + self.api.db.do("DELETE FROM %s WHERE site_id = %d" % \ (table, self['site_id'])) - # XXX Cannot delete site node groups yet - # Mark as deleted self['deleted'] = True - self.flush(commit) + self.sync(commit) class Sites(Table): """ Representation of row(s) from the sites table in the - database. Specify extra_fields to be able to view and modify extra - fields. + database. """ - def __init__(self, api, site_id_or_login_base_list = None, extra_fields = []): - self.api = api - - sql = "SELECT sites.*" \ - ", dslice03_siteinfo.max_slices" - - # N.B.: Joined IDs may be marked as deleted in their primary tables - join_tables = { - # extra_field: (extra_table, extra_column, join_using) - 'person_ids': ('person_site', 'person_id', 'site_id'), - 'slice_ids': ('dslice03_slices', 'slice_id', 'site_id'), - 'defaultattribute_ids': ('dslice03_defaultattribute', 'defaultattribute_id', 'site_id'), - 'pcu_ids': ('pcu', 'pcu_id', 'site_id'), - 'node_ids': ('nodegroup_nodes', 'node_id', 'nodegroup_id'), - } - - extra_fields = filter(join_tables.has_key, extra_fields) - extra_tables = ["%s USING (%s)" % \ - (join_tables[field][0], join_tables[field][2]) \ - for field in extra_fields] - extra_columns = ["%s.%s" % \ - (join_tables[field][0], join_tables[field][1]) \ - for field in extra_fields] - - if extra_columns: - sql += ", " + ", ".join(extra_columns) - - sql += " FROM sites" \ - " LEFT JOIN dslice03_siteinfo USING (site_id)" - - if extra_tables: - sql += " LEFT JOIN " + " LEFT JOIN ".join(extra_tables) - - sql += " WHERE deleted IS False" - - if site_id_or_login_base_list: - # Separate the list into integers and strings - site_ids = filter(lambda site_id: isinstance(site_id, (int, long)), - site_id_or_login_base_list) - login_bases = filter(lambda login_base: isinstance(login_base, StringTypes), - site_id_or_login_base_list) - sql += " AND (False" - if site_ids: - sql += " OR site_id IN (%s)" % ", ".join(map(str, site_ids)) - if login_bases: - sql += " OR login_base IN (%s)" % ", ".join(api.db.quote(login_bases)) - sql += ")" - - rows = self.api.db.selectall(sql) - for row in rows: - if self.has_key(row['site_id']): - site = self[row['site_id']] - site.update(row) + def __init__(self, api, site_filter = None, columns = None): + Table.__init__(self, api, Site, columns) + + view = "view_sites" + for tagname in self.tag_columns: + view= "%s left join %s using (%s)"%(view,Site.tagvalue_view_name(tagname), + Site.primary_key) + + sql = "SELECT %s FROM %s WHERE deleted IS False" % \ + (", ".join(self.columns.keys()+self.tag_columns.keys()),view) + + if site_filter is not None: + if isinstance(site_filter, (list, tuple, set)): + # Separate the list into integers and strings + ints = filter(lambda x: isinstance(x, (int, long)), site_filter) + strs = filter(lambda x: isinstance(x, StringTypes), site_filter) + site_filter = Filter(Site.fields, {'site_id': ints, 'login_base': strs}) + sql += " AND (%s) %s" % site_filter.sql(api, "OR") + elif isinstance(site_filter, dict): + allowed_fields=dict(Site.fields.items()+Site.tags.items()) + site_filter = Filter(allowed_fields, site_filter) + sql += " AND (%s) %s" % site_filter.sql(api, "AND") + elif isinstance (site_filter, StringTypes): + site_filter = Filter(Site.fields, {'login_base':site_filter}) + sql += " AND (%s) %s" % site_filter.sql(api, "AND") + elif isinstance (site_filter, (int, long)): + site_filter = Filter(Site.fields, {'site_id':site_filter}) + sql += " AND (%s) %s" % site_filter.sql(api, "AND") else: - self[row['site_id']] = Site(api, row) + raise PLCInvalidArgument, "Wrong site filter %r"%site_filter + + self.selectall(sql)