diff --git a/LCS/PyCommon/dbcredentials.py b/LCS/PyCommon/dbcredentials.py index 55e79a73cbc3900cb1932ae371697290914379fc..ee244daebb0c3f480629037c00e1b0cc58e6a356 100644 --- a/LCS/PyCommon/dbcredentials.py +++ b/LCS/PyCommon/dbcredentials.py @@ -69,10 +69,10 @@ class Credentials: self.config = {} def __str__(self): - return "type={type} addr={host}:{port} auth={user}:{password} db={database}".format(**self.__dict__) + return "db={database} addr={host}:{port} auth={user}:{password} type={type}".format(**self.__dict__) def stringWithHiddenPassword(self): - return "type={type} addr={host}:{port} auth={user}:XXXXXX db={database}".format(**self.__dict__) + return "db={database} addr={host}:{port} auth={user}:XXXXXX type={type}".format(**self.__dict__) def pg_connect_options(self): """ diff --git a/LCS/PyCommon/postgres.py b/LCS/PyCommon/postgres.py index c43d56541bff70529b308720b8b72fd7197d72ff..13d6cee1db2a03e2b86863a3c08e174a2cc3ad01 100644 --- a/LCS/PyCommon/postgres.py +++ b/LCS/PyCommon/postgres.py @@ -116,6 +116,7 @@ class PostgresDatabaseConnection: query_timeout: float=3600): self._dbcreds = dbcreds self._connection = None + self._cursor = None self.__auto_commit_selects = auto_commit_selects self.__num_connect_retries = num_connect_retries self.__connect_retry_interval = connect_retry_interval @@ -132,7 +133,7 @@ class PostgresDatabaseConnection: for retry_cntr in range(self.__num_connect_retries+1): try: - logger.debug("trying to connect to database: %s", self) + logger.debug("connecting to database: %s", self) self._connection = psycopg2.connect(host=self._dbcreds.host, user=self._dbcreds.user, @@ -172,13 +173,18 @@ class PostgresDatabaseConnection: raise PostgresDBError(error_string) def disconnect(self): - if self.is_connected: - logger.debug("%s disconnecting from db: %s", self.__class__.__name__, self.database) - self._cursor.close() - self._cursor = None - self._connection.close() - self._connection = None - logger.debug("%s disconnected from db: %s", self.__class__.__name__, self.database) + if self._connection is not None or self._cursor is not None: + logger.debug("disconnecting from database: %s", self) + + if self._cursor is not None: + self._cursor.close() + self._cursor = None + + if self._connection is not None: + self._connection.close() + self._connection = None + + logger.info("disconnected from database: %s", self) def _is_recoverable_connection_error(self, error: psycopg2.DatabaseError) -> bool: '''test if psycopg2.DatabaseError is a recoverable connection error''' diff --git a/LCS/PyCommon/test/t_postgres.py b/LCS/PyCommon/test/t_postgres.py index 7a0fdf1bfe5958a65488b26a657fc26ac01db114..29368d849be44e33dde4387c9b4ff9a0397b1fb9 100755 --- a/LCS/PyCommon/test/t_postgres.py +++ b/LCS/PyCommon/test/t_postgres.py @@ -65,7 +65,7 @@ class TestPostgres(unittest.TestCase): # check logging self.assertEqual(NUM_CONNECT_RETRIES, len([ca for ca in mocked_logger.info.call_args_list if 'retrying to connect' in ca[0][0]])) - self.assertEqual(NUM_CONNECT_RETRIES+1, len([ca for ca in mocked_logger.debug.call_args_list if 'trying to connect to database' in ca[0][0]])) + self.assertEqual(NUM_CONNECT_RETRIES+1, len([ca for ca in mocked_logger.debug.call_args_list if 'connecting to database' in ca[0][0]])) self.assertEqual(NUM_CONNECT_RETRIES+1, len([ca for ca in mocked_logger.error.call_args_list if 'could not connect' in ca[0][0]])) def test_reconnect_on_connection_loss(self): diff --git a/MAC/Services/src/ObservationControl2.py b/MAC/Services/src/ObservationControl2.py index 2c4854c7db5b8afb2e0bce490762772aeed72863..d3a10f64eac62d708bc20f5dcbfabcfa4f8e97f4 100644 --- a/MAC/Services/src/ObservationControl2.py +++ b/MAC/Services/src/ObservationControl2.py @@ -36,6 +36,7 @@ logger = logging.getLogger(__name__) class ObservationControlHandler(ServiceMessageHandler): def __init__(self): super(ObservationControlHandler, self).__init__() + self.register_service_method("AbortObservation", self.abort_observation) host = "localhost" diff --git a/SAS/ResourceAssignment/Common/lib/specification.py b/SAS/ResourceAssignment/Common/lib/specification.py index f3018278f3eb2ea056bff7853b722af28531ad96..f674547379642cee263a6270be966a008910278f 100644 --- a/SAS/ResourceAssignment/Common/lib/specification.py +++ b/SAS/ResourceAssignment/Common/lib/specification.py @@ -32,6 +32,8 @@ from lofar.parameterset import parameterset from datetime import datetime, timedelta from lofar.common.datetimeutils import parseDatetime from lofar.sas.resourceassignment.resourceassigner.schedulechecker import movePipelineAfterItsPredecessors +from lofar.common.postgres import PostgresDBQueryExecutionError + import pprint import logging @@ -51,20 +53,20 @@ OUTPUT_PREFIX = "LOFAR.ObsSW." class Specification: - def __init__(self, otdbrpc, momquery, radbrpc): + def __init__(self, otdbrpc, momquery, radb): """Right now we internally represent a subset of a virtual instrument tree in OTDB and manipulate parset type specifications. Therefore we need an otdbid. At some point the OTDB dependency can hopefully be replaced with a proper virtual instrument model, when LOFAR is less dependent on MoM and OTDB and the Scheduler. :param otdbrpc: rpc to talk to otdb :param momquery: rpc to query MoM - :param radbrpc: rpc to talk to radb + :param radb: RADB or RADBRPC instance In practice we need all 3 RPCs as we do consistency checking between the three databases for things like Cluster names. """ self.otdbrpc = otdbrpc - self.radbrpc = radbrpc + self.radb = radb self.momquery = momquery # I have not implemented getter/setter functions for most, maybe in a further refactor if it seems to be useful @@ -166,28 +168,28 @@ class Specification: :param input_dict: Serialized version of a Specification and any predecessors. """ self.otdb_id = input_dict["otdb_id"] - self.mom_id = input_dict["mom_id"] + self.mom_id = input_dict.get("mom_id") self.radb_id = input_dict["task_id"] - self.trigger_id = input_dict["trigger_id"] + self.trigger_id = input_dict.get("trigger_id") self.status = input_dict["status"] self.type = input_dict["task_type"] - self.subtype = input_dict["task_subtype"] - self.starttime = Specification.parse_datetime(input_dict["starttime"]) - self.endtime = Specification.parse_datetime(input_dict["endtime"]) - self.duration = Specification.parse_timedelta(input_dict["duration"]) - self.min_starttime = Specification.parse_datetime(input_dict["min_starttime"]) - self.max_endtime = Specification.parse_datetime(input_dict["max_endtime"]) - self.min_duration = Specification.parse_timedelta(input_dict["min_duration"]) - self.max_duration = Specification.parse_timedelta(input_dict["max_duration"]) - self.cluster = input_dict["cluster"] - self.internal_dict = input_dict["specification"] + self.subtype = input_dict.get("task_subtype") + self.starttime = Specification.parse_datetime(input_dict.get("starttime")) + self.endtime = Specification.parse_datetime(input_dict.get("endtime")) + self.duration = Specification.parse_timedelta(input_dict.get("duration")) + self.min_starttime = Specification.parse_datetime(input_dict.get("min_starttime")) + self.max_endtime = Specification.parse_datetime(input_dict.get("max_endtime")) + self.min_duration = Specification.parse_timedelta(input_dict.get("min_duration")) + self.max_duration = Specification.parse_timedelta(input_dict.get("max_duration")) + self.cluster = input_dict.get("cluster") + self.internal_dict = input_dict.get("specification", {}) self.predecessors = [] - for p in input_dict["predecessors"]: - spec = Specification(self.otdbrpc, self.momquery, self.radbrpc) + for p in input_dict.get("predecessors",[]): + spec = Specification(self.otdbrpc, self.momquery, self.radb) spec.from_dict(p) self.predecessors.append(spec) - self.successor_ids = input_dict["successors"] - self.storagemanager = input_dict["storagemanager"] + self.successor_ids = input_dict.get("successors",[]) + self.storagemanager = input_dict.get("storagemanager") def isObservation(self): """:return if the Specification is for an observation.""" @@ -691,7 +693,7 @@ class Specification: if pred_otdb_id in found_specifications: self.predecessors.append(found_specifications[pred_otdb_id]) else: - spec = Specification(self.otdbrpc, self.momquery, self.radbrpc) + spec = Specification(self.otdbrpc, self.momquery, self.radb) spec.read_from_OTDB_with_predecessors(pred_otdb_id, "otdb", found_specifications) self.predecessors.append(spec) #TODO this needs a try/except somewhere? logger.info("Done reading from OTDB with predecessors") @@ -716,7 +718,7 @@ class Specification: if str() in cluster_name_set or len(cluster_name_set) != 1: # Empty set or name is always an error. # TODO: To support >1 cluster per task is not implemented - # self.radbrpc.insertSpecificationAndTask() as called below and the radb would need to take >1 + # self.radbrpc.insertOrUpdateSpecificationAndTask() as called below and the radb would need to take >1 # cluster name/ Also, there is only 1 processingClusterName in the parset, but we do not always want to # pipeline process all obs outputs, or not on 1 cluster logger.error( @@ -728,7 +730,7 @@ class Specification: # Retrieve known cluster names (not all may be a valid storage target, but we cannot know...) known_cluster_set = {cluster['name'] for cluster in - self.radbrpc.getResourceGroupNames('cluster')} + self.radb.getResourceGroupNames('cluster')} if cluster_name not in known_cluster_set: raise Exception("skipping resource assignment for task with cluster name '" + cluster_name + "' not in known clusters " + str(known_cluster_set)) #TODO better error @@ -973,7 +975,7 @@ class Specification: starttime, endtime, duration, cluster) or None """ #TODO maybe this should read the spec as well? Can especially start/end times be updated outside of here? - task = self.radbrpc.getTask(radb_id) # if radb_id is not None else None + task = self.radb.getTask(radb_id) # if radb_id is not None else None if task: #TODO what is the exact structure of task see schedulerchecker 47 self.radb_id = task["id"] # Should be the same as radb_id, but self.radb_id might not yet be set self.mom_id = task["mom_id"] @@ -1045,7 +1047,7 @@ class Specification: # another service sets the parset spec in OTDB, and updated otdb task status to scheduled, which is then # synced back to RADB if self.radb_id: - self.radbrpc.updateTask(self.radb_id, task_status=new_status) + self.radb.updateTask(self.radb_id, task_status=new_status) # TODO what? else: # self.radbrpc.updateTaskStatusForOtdbId(self.otdb_id, 'error') #TODO, see if we move _send_task_status_notification from resource_assigner to here? @@ -1064,8 +1066,8 @@ class Specification: 'insertSpecification mom_id=%s, otdb_id=%s, status=%s, task_type=%s, start_time=%s, end_time=%s ' 'cluster=%s' % (self.mom_id, self.otdb_id, self.status, self.type, self.starttime, self.endtime, self.cluster) ) - result = self.radbrpc.insertSpecificationAndTask(self.mom_id, self.otdb_id, self.status, self.type, self.starttime, - self.endtime, str(self.as_dict()), self.cluster) #TODO use internal_dict? + result = self.radb.insertOrUpdateSpecificationAndTask(self.mom_id, self.otdb_id, self.status, self.type, self.starttime, + self.endtime, str(self.as_dict()), self.cluster, commit=True) #TODO use internal_dict? specification_id = result['specification_id'] # We never seem to need this again task_id = result['task_id'] @@ -1088,17 +1090,26 @@ class Specification: for predecessor_mom_id in predecessor_mom_ids: # check if the predecessor needs to be linked to this task - predecessor_task = self.radbrpc.getTask(mom_id=predecessor_mom_id) + predecessor_task = self.radb.getTask(mom_id=predecessor_mom_id) if predecessor_task: predecessor_keys = [p.radb_id for p in self.predecessors] if predecessor_task['id'] not in predecessor_keys: logger.info('connecting predecessor task with mom_id=%s otdb_id=%s to its successor with mom_id=%s ' 'otdb_id=%s', predecessor_task['mom_id'], predecessor_task['otdb_id'], self.mom_id, self.otdb_id) - spec = Specification(self.otdbrpc, self.momquery, self.radbrpc) + spec = Specification(self.otdbrpc, self.momquery, self.radb) spec.read_from_radb(predecessor_task['id']) self.predecessors.append(spec) # TODO this needs a try/except somewhere? - self.radbrpc.insertTaskPredecessor(self.radb_id, spec.radb_id) + try: + self.radb.insertTaskPredecessor(self.radb_id, spec.radb_id) + except PostgresDBQueryExecutionError as e: + # task was already connected to predecessor. Log and continue. + if 'task_predecessor_unique' in str(e): + logger.info('predecessor task with mom_id=%s otdb_id=%s was already connected to its successor with mom_id=%s otdb_id=%s', + predecessor_task['mom_id'], predecessor_task['otdb_id'], + self.mom_id, self.otdb_id) + else: + raise else: # Occurs when setting a pipeline to prescheduled while a predecessor has e.g. never been beyond # approved, which is in principle valid. The link in the radb will be made later via processSuccessors() @@ -1123,7 +1134,7 @@ class Specification: for successor_mom_id in successor_mom_ids: # check if the successor needs to be linked to this task - successor_task = self.radbrpc.getTask(mom_id=successor_mom_id) + successor_task = self.radb.getTask(mom_id=successor_mom_id) if successor_task: if successor_task['id'] not in self.successor_ids: logger.info( @@ -1131,9 +1142,18 @@ class Specification: ' otdb_id=%s', successor_task['mom_id'], successor_task['otdb_id'], self.mom_id, self.otdb_id ) self.successor_ids.append(successor_task['id']) - self.radbrpc.insertTaskPredecessor(successor_task['id'], self.radb_id) - - movePipelineAfterItsPredecessors(successor_task, self.radbrpc) + try: + self.radb.insertTaskPredecessor(successor_task['id'], self.radb_id) + except PostgresDBQueryExecutionError as e: + # task was already connected to predecessor. Log and continue. + if 'task_predecessor_unique' in str(e): + logger.info('successor task with mom_id=%s otdb_id=%s was already connected to its predecessor with mom_id=%s otdb_id=%s', + successor_task['mom_id'], successor_task['otdb_id'], + self.mom_id, self.otdb_id) + else: + raise + + movePipelineAfterItsPredecessors(successor_task, self.radb) else: # Occurs when settings a obs or task to prescheduled while a successor has e.g. not yet been beyond # approved, which is quite normal. The link in the radb will be made later via processPredecessors() diff --git a/SAS/ResourceAssignment/ResourceAssigner/lib/resource_assigner.py b/SAS/ResourceAssignment/ResourceAssigner/lib/resource_assigner.py index 45eb4b35ef9a5f0c57394cee5b5dde1b1a29db57..113d6ab2198c5930ae1c6b791c3c4b1961816bd2 100755 --- a/SAS/ResourceAssignment/ResourceAssigner/lib/resource_assigner.py +++ b/SAS/ResourceAssignment/ResourceAssigner/lib/resource_assigner.py @@ -32,7 +32,7 @@ from lofar.messaging.messagebus import ToBus from lofar.messaging import DEFAULT_BROKER, DEFAULT_BUSNAME from lofar.common.util import single_line_with_single_spaces -from lofar.sas.resourceassignment.resourceassignmentservice.rpc import RADBRPC +from lofar.sas.resourceassignment.database.radb import RADatabase from lofar.sas.otdb.otdbrpc import OTDBRPC from lofar.sas.resourceassignment.resourceassigner.config import DEFAULT_RA_NOTIFICATION_PREFIX @@ -72,7 +72,7 @@ class ResourceAssigner(object): :param radb_dbcreds: the credentials to be used for accessing the RADB (default: None, which means default) """ - self.radbrpc = RADBRPC.create(exchange=exchange, broker=broker) + self.radb = RADatabase(dbcreds=radb_dbcreds) self.rerpc = ResourceEstimatorRPC.create(exchange=exchange, broker=broker) self.otdbrpc = OTDBRPC.create(exchange=exchange, broker=broker) self.momrpc = MoMQueryRPC.create(exchange=exchange, broker=broker) @@ -81,7 +81,7 @@ class ResourceAssigner(object): self.ra_notification_bus = ToBus(exchange=exchange, broker=broker) self.obscontrol = ObservationControlRPCClient.create(exchange=exchange, broker=broker) - self.resource_availability_checker = ResourceAvailabilityChecker(self.radbrpc) + self.resource_availability_checker = ResourceAvailabilityChecker(self.radb) # For the DwellScheduler instances created during run-time we store the following variables self.radb_creds = radb_dbcreds @@ -97,8 +97,9 @@ class ResourceAssigner(object): self.close() def open(self): - """Open rpc connections to radb service and resource estimator service""" - self.radbrpc.open() + """Open connections to various service/buses/databases""" + logger.debug("resource_assigner opening all bus/db connections") + self.radb.connect() self.rerpc.open() self.otdbrpc.open() self.momrpc.open() @@ -106,24 +107,27 @@ class ResourceAssigner(object): self.curpc.open() self.ra_notification_bus.open() self.obscontrol.open() + logger.info("resource_assigner opened all bus/db connections") def close(self): - """Close rpc connections to radb service and resource estimator service""" + """Close connections to various service/buses/databases""" + logger.debug("resource_assigner closing all bus/db connections") self.obscontrol.close() - self.radbrpc.close() + self.radb.disconnect() self.rerpc.close() self.otdbrpc.close() self.momrpc.close() self.sqrpc.close() self.curpc.close() self.ra_notification_bus.close() + logger.info("resource_assigner closed all bus/db connections") @property @cache def resource_types(self): """ Returns a dict of all the resource types, to convert name->id. """ - return {rt['name']: rt['id'] for rt in self.radbrpc.getResourceTypes()} + return {rt['name']: rt['id'] for rt in self.radb.getResourceTypes()} def do_assignment(self, otdb_id, specification_tree): """ @@ -140,9 +144,9 @@ class ResourceAssigner(object): :raises an Exception if something unforeseen happened while scheduling """ - logger.info(('do_assignment: otdb_id=%s specification_tree=%s' % (otdb_id, specification_tree))) + logger.info('do_assignment: otdb_id=%s specification_tree=%s', otdb_id, specification_tree) - spec = Specification(self.otdbrpc, self.momrpc, self.radbrpc) + spec = Specification(self.otdbrpc, self.momrpc, self.radb) spec.from_dict(specification_tree) spec.insert_into_radb() # TODO Move this to TaskSpecified? @@ -269,7 +273,7 @@ class ResourceAssigner(object): specification_tree=specification_tree, resource_estimator=self._get_resource_estimates, resource_availability_checker=self.resource_availability_checker, - radbcreds=self.radb_creds, + radb=self.radb, min_starttime=min_starttime, max_starttime=max_starttime, duration=duration) @@ -278,7 +282,7 @@ class ResourceAssigner(object): specification_tree=specification_tree, resource_estimator=self._get_resource_estimates, resource_availability_checker=self.resource_availability_checker, - radbcreds=self.radb_creds) + radb=self.radb) except Exception as e: logger.exception('Error in scheduler._schedule_resources: %s', e) #Why are we mentioning _schedule_resources here? return False @@ -288,13 +292,14 @@ class ResourceAssigner(object): (scheduler_result, changed_tasks) = scheduler.allocate_resources() if not scheduler_result: # try again with basic scheduler to end up with a situation with the 'normal' conflicting resources, which can then be evaluated by users - with BasicScheduler(task_id=spec.radb_id, - specification_tree=specification_tree, - resource_estimator=self._get_resource_estimates, - resource_availability_checker=self.resource_availability_checker, - radbcreds=self.radb_creds) as basic_scheduler: + basic_scheduler = BasicScheduler(task_id=spec.radb_id, + specification_tree=specification_tree, + resource_estimator=self._get_resource_estimates, + resource_availability_checker=self.resource_availability_checker, + radb=self.radb) + with basic_scheduler: (scheduler_result, changed_tasks) = basic_scheduler.allocate_resources() - return scheduler_result + return scheduler_result elif changed_tasks: for t in changed_tasks: if t.status == 'aborted': #MAC_Scheduler can't handle queued right now See also schedulers.py around line 600 diff --git a/SAS/ResourceAssignment/ResourceAssigner/lib/resource_availability_checker.py b/SAS/ResourceAssignment/ResourceAssigner/lib/resource_availability_checker.py index 08bddad5f0a01a38b65e92e10c52d17ad0ad3e55..412982ac92019c1c7967641aedd6216d311c8ce3 100644 --- a/SAS/ResourceAssignment/ResourceAssigner/lib/resource_availability_checker.py +++ b/SAS/ResourceAssignment/ResourceAssigner/lib/resource_availability_checker.py @@ -33,8 +33,8 @@ class CouldNotFindClaimException(Exception): class ResourceAvailabilityChecker(object): - def __init__(self, radbrpc): - self.radbrpc = radbrpc + def __init__(self, radb): + self.radb = radb @property @cache @@ -47,7 +47,7 @@ class ResourceAvailabilityChecker(object): "parent_ids": list of parent resource groups "resource_ids": list of resources in this group. } """ - memberships = self.radbrpc.getResourceGroupMemberships() + memberships = self.radb.getResourceGroupMemberships() return memberships['groups'] # resource-group-to-resource-group relations @property @@ -55,14 +55,14 @@ class ResourceAvailabilityChecker(object): def resource_types(self): """ Returns a dict of all the resource types, to convert name->id. """ - return {rt['name']: rt['id'] for rt in self.radbrpc.getResourceTypes()} + return {rt['name']: rt['id'] for rt in self.radb.getResourceTypes()} @property @cache def resource_claim_property_types(self): """ Returns a dict of all the resource claim property types, to convert name->id. """ - return {rcpt['name']: rcpt['id'] for rcpt in self.radbrpc.getResourceClaimPropertyTypes()} + return {rcpt['name']: rcpt['id'] for rcpt in self.radb.getResourceClaimPropertyTypes()} @cache def _summable_property_type_ids(self): @@ -145,8 +145,8 @@ class ResourceAvailabilityChecker(object): return claims def _get_current_resource_usage(self): - db_resource_list = self.radbrpc.getResources(include_availability=True) - db_resource_max_fill_ratios = self.radbrpc.getResourceAllocationConfig(sql_like_name_pattern='max_fill_ratio_%') + db_resource_list = self.radb.getResources(include_availability=True) + db_resource_max_fill_ratios = self.radb.getResourceAllocationConfig(sql_like_name_pattern='max_fill_ratio_%') self._apply_maximum_fill_ratios(db_resource_list, db_resource_max_fill_ratios) diff --git a/SAS/ResourceAssignment/ResourceAssigner/lib/schedulers.py b/SAS/ResourceAssignment/ResourceAssigner/lib/schedulers.py index 0f17dae22ddc74a22a51b7a79d8860417692480a..334008e611918f6959bcdd166e50f0da12b6cf99 100644 --- a/SAS/ResourceAssignment/ResourceAssigner/lib/schedulers.py +++ b/SAS/ResourceAssignment/ResourceAssigner/lib/schedulers.py @@ -47,7 +47,7 @@ class ScheduleException(Exception): class BasicScheduler(object): """ A Scheduler that allocates resources at a fixed time. Resources are searched for. """ - def __init__(self, task_id, specification_tree, resource_estimator, resource_availability_checker, radbcreds=None): + def __init__(self, task_id, specification_tree, resource_estimator, resource_availability_checker, radb: RADatabase): """ Creates a BasicScheduler instance @@ -55,7 +55,7 @@ class BasicScheduler(object): :param specification_tree: the full specification; will be modified where needed with respect to resources :param resource_estimator: the ResourceEstimator function that turns a specification into resource estimates :param resource_availability_checker: the ResourceAvailabilityScheduler to be used by the BasicScheduler - :param radbcreds: ResourceAssigner database credentials. If None, the default credentials will be used + :param radb: a RADatabase instance. :raises AssertionError if task_id is a negative number or is None """ @@ -64,9 +64,7 @@ class BasicScheduler(object): self.specification_tree = specification_tree self.resource_estimator = resource_estimator self.resource_availability_checker = resource_availability_checker - - # We need a DIRECT connection to the database in order to do client-side (that's us) transactions - self.radb = RADatabase(dbcreds=radbcreds) + self.radb = radb # Ensure a valid task_id is given, since radb.getTasks() will not raise if task_id equals None if task_id < 0: @@ -92,12 +90,10 @@ class BasicScheduler(object): self.close() def open(self): - """ Open radb connection""" - self.radb.connect() + pass def close(self): - """ Close radb connection""" - self.radb.disconnect() + pass def allocate_resources(self): """ @@ -333,7 +329,7 @@ class StationScheduler(BasicScheduler): After scheduling, the get_stations() function returns a list of the allocated stations. """ def __init__(self, task_id, specification_tree, resource_estimator, resource_availability_checker, - radbcreds=None, + radb: RADatabase, broker=DEFAULT_BROKER): """ Creates a StationScheduler instance @@ -342,13 +338,13 @@ class StationScheduler(BasicScheduler): :param specification_tree: the full specification; will be modified where needed with respect to resources :param resource_estimator: the ResourceEstimator function that turns a specification into resource estimates :param resource_availability_checker: the ResourceAvailabilityChecker instance to use - :param radbcreds: the RADB credentials to use + :param radb: a RADatabase instance. :param mom_busname: the MoM Query service bus name (default: 'lofar.ra.command') :param mom_servicename: the MoM Query service name (default: 'momqueryservice') :param broker: the message broker to use for send messages/RPC calls/etc. """ - super(StationScheduler, self).__init__(task_id, specification_tree, resource_estimator, resource_availability_checker, radbcreds) + super(StationScheduler, self).__init__(task_id, specification_tree, resource_estimator, resource_availability_checker, radb) # For observations without a fixed station list, we need to derive one. TODO: this likely isnt the condition we want to decide on self.must_derive_station_list = specification_tree["task_type"] == "observation" and specification_tree["specification"]["Observation.VirtualInstrument.stationList"] == [] and specification_tree["station_requirements"] @@ -489,7 +485,7 @@ class PriorityScheduler(StationScheduler): Conflict resolution is done by killing jobs with lower priority. """ def __init__(self, task_id, specification_tree, resource_estimator, resource_availability_checker, - radbcreds=None, + radb: RADatabase, broker=DEFAULT_BROKER): """ Creates a PriorityScheduler instance @@ -498,11 +494,11 @@ class PriorityScheduler(StationScheduler): :param specification_tree: the full specification; will be modified where needed with respect to resources :param resource_estimator: the ResourceEstimator function that turns a specification into resource estimates :param resource_availability_checker: the ResourceAvailabilityChecker instance to use - :param radbcreds: the RADB credentials to use + :param radb: a RADatabase instance. :param broker: the message broker to use for send messages/RPC calls/etc. """ - super(PriorityScheduler, self).__init__(task_id, specification_tree, resource_estimator, resource_availability_checker, radbcreds) + super(PriorityScheduler, self).__init__(task_id, specification_tree, resource_estimator, resource_availability_checker, radb) self.momqueryservice = MoMQueryRPC.create(broker=broker, timeout=180) @@ -719,7 +715,7 @@ class DwellScheduler(PriorityScheduler): max_starttime, duration, resource_availability_checker, - radbcreds=None, + radb: RADatabase = None, broker=DEFAULT_BROKER): """ Create a DwellScheduler instance @@ -731,7 +727,7 @@ class DwellScheduler(PriorityScheduler): :param max_starttime: the task's desired maximum start time :param duration: the task's duration :param resource_availability_checker: the ResourceAvailabilityChecker to use - :param radbcreds: the RADB credentials to use + :param radb: a RADatabase instance. :param mom_busname: the MoM Query service bus name (default: 'lofar.ra.command') :param mom_servicename: the MoM Query service name (default: 'momqueryservice') :param broker: the message broker to use for send messages/RPC calls/etc. @@ -741,7 +737,7 @@ class DwellScheduler(PriorityScheduler): specification_tree=specification_tree, resource_estimator=resource_estimator, resource_availability_checker=resource_availability_checker, - radbcreds=radbcreds, + radb=radb, broker=broker) self.min_starttime = min_starttime diff --git a/SAS/ResourceAssignment/ResourceAssigner/test/t_resource_availability_checker.py b/SAS/ResourceAssignment/ResourceAssigner/test/t_resource_availability_checker.py index 1a5ba3f2542daafbcd66fb7f74e6974aa995f7fc..db2888dc11c8145df83a4b6b768ebfb3e21210e6 100755 --- a/SAS/ResourceAssignment/ResourceAssigner/test/t_resource_availability_checker.py +++ b/SAS/ResourceAssignment/ResourceAssigner/test/t_resource_availability_checker.py @@ -160,7 +160,7 @@ class ResourceAvailabilityCheckerTest(unittest.TestCase): self.addCleanup(rarpc_patcher.stop) self.rarpc_mock = rarpc_patcher.start() self.rarpc_mock.getTask.side_effect = get_task_side_effect - self.rarpc_mock.insertSpecificationAndTask.return_value = { + self.rarpc_mock.insertOrUpdateSpecificationAndTask.return_value = { 'inserted': True, 'specification_id': self.specification_id, 'task_id': self.task_id diff --git a/SAS/ResourceAssignment/ResourceAssigner/test/t_resourceassigner.py b/SAS/ResourceAssignment/ResourceAssigner/test/t_resourceassigner.py index 014282619181b13083d2c703b72ae5dbe13e0d2a..17cfcb1e8b647371c5a851a43a87f3a1aa475ca1 100755 --- a/SAS/ResourceAssignment/ResourceAssigner/test/t_resourceassigner.py +++ b/SAS/ResourceAssignment/ResourceAssigner/test/t_resourceassigner.py @@ -22,41 +22,24 @@ import unittest from unittest import mock import datetime import sys +from copy import deepcopy import logging logging.basicConfig(format='%(asctime)s %(levelname)s %(process)s %(message)s', level=logging.INFO) -from lofar.sas.resourceassignment.resourceassigner.resource_assigner import ResourceAssigner, DEFAULT_RA_NOTIFICATION_PREFIX -from lofar.sas.resourceassignment.resourceassigner.resource_availability_checker import ResourceAvailabilityChecker +from lofar.sas.resourceassignment.resourceassigner.config import DEFAULT_RA_NOTIFICATION_PREFIX +from lofar.sas.resourceassignment.resourceassigner.resource_assigner import ResourceAssigner +from lofar.messaging.messagebus import TemporaryExchange from lofar.sas.resourceassignment.common.specification import Specification -from lofar.parameterset import parameterset -from lofar.common.datetimeutils import parseDatetime from lofar.common.util import single_line_with_single_spaces +from lofar.messaging.messages import EventMessage +from lofar.sas.resourceassignment.database.testing.radb_common_testing import RADBCommonTestMixin -class TestingResourceAssigner(ResourceAssigner): - def __init__(self, rarpc, rerpc, otdbrpc, momrpc, curpc, sqrpc, ra_notification_bus, dwell_scheduler, obscontrol, - radb_dbcreds=None): - # super gets not done to be able to insert mocks as early as possible otherwise the RPC block unittesting - self.radbrpc = rarpc - self.rerpc = rerpc - self.otdbrpc = otdbrpc - self.momrpc = momrpc - self.curpc = curpc - self.sqrpc = sqrpc - self.ra_notification_bus = ra_notification_bus - # Could mock ResourceAvailabilityChecker, but it is out of play already due to mocked DwellScheduler - self.resource_availability_checker = ResourceAvailabilityChecker(rarpc) - self.dwell_scheduler = dwell_scheduler - self.obscontrol = obscontrol - self.radb_creds = radb_dbcreds - - -class ResourceAssignerTest(unittest.TestCase): +class ResourceAssignerTest(RADBCommonTestMixin, unittest.TestCase): mom_id = 351557 otdb_id = 1290494 specification_id = 2323 - state = 'prescheduled' task_type = 'pipeline' specification_tree = {} @@ -78,36 +61,16 @@ class ResourceAssignerTest(unittest.TestCase): } } - approved_status = 'approved' - approved_otdb_id = 22 - approved_specification_tree = { - 'otdb_id': approved_otdb_id, - 'task_type': 'pipeline', - 'state': approved_status, - 'specification': { - 'Observation.startTime': future_start_time, - 'Observation.stopTime': future_stop_time - } - } - - cep2_specification_tree = { - 'otdb_id': otdb_id, - 'task_type': 'pipeline', - 'state': 'prescheduled', - 'specification': { - 'Observation.startTime': future_start_time, - 'Observation.stopTime': future_stop_time, - 'Observation.DataProducts.Output_Pulsar.enabled': True, - 'Observation.DataProducts.Output_Pulsar.storageClusterName': 'CEP2' - } - } - mom_bug_processing_cluster_name = 'CEP2' mom_bug_otdb_id = 1234 mom_bug_specification_tree = { 'otdb_id': mom_bug_otdb_id, + 'mom_id': None, + 'task_id': None, 'task_type': 'pipeline', 'state': 'prescheduled', + 'starttime': datetime.datetime.utcnow(), + 'endtime': datetime.datetime.utcnow() + datetime.timedelta(minutes=1), 'specification': { 'Observation.startTime': future_start_time, 'Observation.stopTime': future_stop_time, @@ -117,32 +80,6 @@ class ResourceAssignerTest(unittest.TestCase): } } - maintenance_otdb_id = 5678 - maintenance_specification_tree = { - 'otdb_id': maintenance_otdb_id, - 'task_type': 'reservation', - 'task_subtype': 'maintenance', - 'state': 'prescheduled', - 'specification': { - 'Observation.startTime': future_start_time, - 'Observation.stopTime': future_stop_time, - 'Observation.VirtualInstrument.stationList': ['CS001'], - } - } - - projectreservation_otdb_id = 8765 - projectreservation_specification_tree = { - 'otdb_id': projectreservation_otdb_id, - 'task_type': 'reservation', - 'task_subtype': 'project', - 'state': 'prescheduled', - 'specification': { - 'Observation.startTime': future_start_time, - 'Observation.stopTime': future_stop_time, - 'Observation.VirtualInstrument.stationList': ['CS001'], - } - } - task_mom_id = 351543 task_otdb_id = 1290472 task_id = 2299 @@ -159,53 +96,17 @@ class ResourceAssignerTest(unittest.TestCase): predecessor_task_mom_id = 1 predecessor_task_otdb_id = 2 predecessor_task_id = 3 - predecessor_task = { - "mom_id": predecessor_task_mom_id, - "otdb_id": predecessor_task_otdb_id, - "id": predecessor_task_id, - "endtime": datetime.datetime(2016, 3, 25, 22, 47, 31), - "name": "IS HBA_DUAL", - "predecessor_ids": [], - "project_mom_id": 2, - "project_name": "test-lofar", - "specification_id": 2323, - "starttime": datetime.datetime(2016, 3, 25, 21, 47, 31), - "status": "prescheduled", - "status_id": 350, - "successor_ids": [], - "type": "pipeline", - "type_id": 0 - } successor_task_mom_id = 4 successor_task_otdb_id = 5 successor_task_id = 6 - successor_task = { - "mom_id": successor_task_mom_id, - "otdb_id": successor_task_otdb_id, - "id": successor_task_id, - "endtime": datetime.datetime(2016, 3, 25, 22, 47, 31), - "name": "IS HBA_DUAL", - "predecessor_ids": [], - "project_mom_id": 2, - "project_name": "test-lofar", - "specification_id": 2323, - "starttime": datetime.datetime(2016, 3, 25, 21, 47, 31), - "status": "prescheduled", - "status_id": 350, - "successor_ids": [], - "type": "pipeline", - "type_id": 0 - } - resources_with_rcus_otdb_id = 1290495 resources_with_no_resource_types_otdb_id = 1290497 resources_with_negative_estimates_otdb_id = 1290488 resources_with_errors_otdb_id = 1290496 + no_resources_otdb_id = 1290498 resource_error1 = "error 1" resource_error2 = "error 2" - unknown_resource_type_name = "fuel" - unknown_resource_type_otdb_id = 123489 rerpc_status = 0 rerpc_needed_claim_for_bandwidth_size = 2 @@ -269,13 +170,9 @@ class ResourceAssignerTest(unittest.TestCase): } }] }, - str(resources_with_rcus_otdb_id): { - 'errors': [], - 'estimates': [{ - 'resource_types': {'rcu': '111100010111100101101010' }, - 'resource_count': 1, - 'root_resource_group': 'CS001' - }] + str(no_resources_otdb_id): { + 'errors': [], + 'estimates': [] }, str(resources_with_errors_otdb_id): { 'estimates': [{ @@ -309,75 +206,9 @@ class ResourceAssignerTest(unittest.TestCase): }] } }], - }, - str(unknown_resource_type_otdb_id): { - }, - str(maintenance_otdb_id): - {'errors': [], - 'estimates': [{'resource_count': 1, - 'resource_types': {'rcu': '111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111'}, - 'root_resource_group': 'CS001'}, - # TODO: I'm leaving this here to check the order issue in the getResourceGroupMemberships.return_value - # {'resource_count': 1, - # 'resource_types': {'bandwidth': 3000000000, - # 'rsp': 3904}, - # 'root_resource_group': 'CS001RSP0'}, - # {'resource_count': 1, - # 'resource_types': {'bandwidth': 3000000000, - # 'rsp': 3904}, - # 'root_resource_group': 'CS001RSP1'}, - # {'resource_count': 1, - # 'resource_types': {'rcu': '111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111'}, - # 'root_resource_group': 'CS002'}, - # {'resource_count': 1, - # 'resource_types': {'bandwidth': 3000000000, - # 'rsp': 3904}, - # 'root_resource_group': 'CS002RSP0'}, - # {'resource_count': 1, - # 'resource_types': {'bandwidth': 3000000000, - # 'rsp': 3904}, - # 'root_resource_group': 'CS002RSP1'}, - ]} - , - str(projectreservation_otdb_id): - {'errors': [], - 'estimates': [{'resource_count': 1, - 'resource_types': {'rcu': '111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111111'}, - 'root_resource_group': 'CS001'}, - ]} - } - - cep4bandwidth_resource_id = 116 - cep4storage_resource_id = 117 - - storage_claim = { - 'resource_id': cep4storage_resource_id, - 'resource_type_id': 5, - 'starttime': task_start_time, - 'used_rcus': None, - 'endtime': task_end_time + datetime.timedelta(days=365), - 'status': 'tentative', - 'claim_size': rerpc_needed_claim_for_storage_size, - 'properties': [ - {'io_type': 'output', 'type': 15, 'sap_nr': 0, 'value': 0}, - {'io_type': 'output', 'type': 2, 'sap_nr': 0, 'value': 1}, - {'io_type': 'output', 'type': 10, 'sap_nr': 0, 'value': 1073741824} - ] - } - - bandwidth_claim = { - 'resource_id': cep4bandwidth_resource_id, - 'resource_type_id': 3, - 'starttime': task_start_time, - 'used_rcus': None, - 'endtime': task_end_time, - 'status': 'tentative', - 'claim_size': rerpc_needed_claim_for_bandwidth_size, - 'properties': [] + } } - specification_claims = [bandwidth_claim, storage_claim] - def reset_specification_tree(self): self.specification_tree = { 'otdb_id': self.otdb_id, @@ -393,6 +224,9 @@ class ResourceAssignerTest(unittest.TestCase): 'duration': 60, 'cluster': "CEP4", 'task_subtype': 'long baseline pipeline', + 'starttime': self.future_start_time, + 'endtime': self.future_stop_time, + 'storagemanager': None, 'specification': { 'Observation.momID': str(self.mom_id), 'Observation.startTime': self.future_start_time, @@ -431,7 +265,6 @@ class ResourceAssignerTest(unittest.TestCase): 'min_duration': 60, 'max_duration': 60, 'cluster': "CEP4", - 'task_subtype': 'averaging pipeline', 'specification': { 'Observation.DataProducts.Output_InstrumentModel.enabled': False, @@ -543,1079 +376,34 @@ class ResourceAssignerTest(unittest.TestCase): } def setUp(self): + super().setUp() self.reset_task() - def get_task_side_effect(*args, **kwargs): - if 'mom_id' in kwargs: - if kwargs['mom_id'] == self.successor_task_mom_id: - return self.successor_task - elif kwargs['mom_id'] == self.predecessor_task_mom_id: - return self.predecessor_task - elif kwargs['mom_id'] == self.non_existing_task_mom_id: - return None - else: - return self.task - else: - return self.task - - self.successor_task_mom_ids = [self.successor_task_mom_id] - self.predecessor_task_mom_ids = [self.predecessor_task_mom_id] - - rarpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassignmentservice.rpc.RADBRPC') - self.addCleanup(rarpc_patcher.stop) - self.rarpc_mock = rarpc_patcher.start() - self.rarpc_mock.getTask.side_effect = get_task_side_effect - self.rarpc_mock.insertSpecificationAndTask.return_value = { - 'inserted': True, - 'specification_id': self.specification_id, - 'task_id': self.task_id - } - self.rarpc_mock.getResourceClaimPropertyTypes.return_value = [ - {'id': 0, 'name': 'nr_of_is_files'}, - {'id': 1, 'name': 'nr_of_cs_files'}, - {'id': 2, 'name': 'nr_of_uv_files'}, - {'id': 3, 'name': 'nr_of_im_files'}, - {'id': 4, 'name': 'nr_of_img_files'}, - {'id': 5, 'name': 'nr_of_pulp_files'}, - {'id': 6, 'name': 'nr_of_cs_stokes'}, - {'id': 7, 'name': 'nr_of_is_stokes'}, - {'id': 8, 'name': 'is_file_size'}, - {'id': 9, 'name': 'cs_file_size'}, - {'id': 10, 'name': 'uv_file_size'}, - {'id': 11, 'name': 'im_file_size'}, - {'id': 12, 'name': 'img_file_size'}, - {'id': 13, 'name': 'nr_of_pulp_files'}, - {'id': 14, 'name': 'nr_of_cs_parts'}, - {'id': 15, 'name': 'start_sb_nr'}, - {'id': 16, 'name': 'uv_otdb_id'}, - {'id': 17, 'name': 'cs_otdb_id'}, - {'id': 18, 'name': 'is_otdb_id'}, - {'id': 19, 'name': 'im_otdb_id'}, - {'id': 20, 'name': 'img_otdb_id'}, - {'id': 21, 'name': 'pulp_otdb_id'}, - {'id': 22, 'name': 'is_tab_nr'}, - {'id': 23, 'name': 'start_sbg_nr'}, - {'id': 24, 'name': 'pulp_file_size'} - ] - self.rarpc_mock.getResourceTypes.return_value = [ - {'id': 0, 'name': 'rsp', 'unit_id': 0, 'units': 'rsp_channel_bit'}, - {'id': 1, 'name': 'tbb', 'unit_id': 1, 'units': 'bytes'}, - {'id': 2, 'name': 'rcu', 'unit_id': 2, 'units': 'rcu_board'}, - {'id': 3, 'name': 'bandwidth', 'unit_id': 3, 'units': 'bits/second'}, - {'id': 4, 'name': 'processor', 'unit_id': 4, 'units': 'cores'}, - {'id': 5, 'name': 'storage', 'unit_id': 1, 'units': 'bytes'}, - ] - self.rarpc_mock.insertResourceClaims.return_value = {'ids': [1, 2]} - - self.rarpc_mock.getResourceGroupNames.return_value = [{"name": "CEP4"}, {"name": "DRAGNET"}, {"name": "COBALT"}] - - self.rarpc_mock.getResourceGroupMemberships.return_value = {'groups': [ - {'resource_group_parent_id': None, 'resource_group_parent_name': None, 'resource_group_id': 0, - 'resource_group_name': 'CORE', 'child_ids': [1, 2], 'parent_ids': [], 'resource_ids': [0, 1]}, - {'resource_group_parent_id': None, 'resource_group_parent_name': None, 'resource_group_id': 3, - 'resource_group_name': 'CS001', 'child_ids': [], 'parent_ids': [0], 'resource_ids': [212]}, - {'resource_group_parent_id': None, 'resource_group_parent_name': None, 'resource_group_id': 1, - 'resource_group_name': 'CEP4', 'child_ids': [], 'parent_ids': [0], 'resource_ids': [116, 117]}, - # {'resource_group_parent_id': None, 'resource_group_parent_name': None, 'resource_group_id': 4, # TODO: WHY DOES ORDER MATTER IN HERE??? - # 'resource_group_name': 'CS002', 'child_ids': [], 'parent_ids': [0], 'resource_ids': [214]}, # TODO: check what happens when this is moved after e.g. CS001; also comment in CS002 in RE response - ], - 'resources': [{'resource_group_parent_id': 0, - 'resource_group_parent_name': 'CORE', - 'resource_id': 0, - 'resource_name': 'CS001', - 'parent_group_ids': []}, - {'resource_group_parent_id': 0, - 'resource_group_parent_name': 'CORE', - 'resource_id': 1, - 'resource_name': 'CS002', - 'parent_group_ids': []}, - {'resource_group_parent_id': 1, - 'resource_group_parent_name': 'CEP4', - 'resource_id': 2, - 'resource_name': 'CEP4_storage:/data', - 'parent_group_ids': []}]} - - # incomplete response but good enough for tests - self.rarpc_mock.getResources.return_value = [ - {'id': 0, 'name': 'cpunode01_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 1, 'name': 'cpunode01_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 2, 'name': 'cpunode02_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 3, 'name': 'cpunode02_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 4, 'name': 'cpunode03_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 5, 'name': 'cpunode03_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 6, 'name': 'cpunode04_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 7, 'name': 'cpunode04_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 8, 'name': 'cpunode05_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 9, 'name': 'cpunode05_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 10, 'name': 'cpunode06_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 11, 'name': 'cpunode06_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 12, 'name': 'cpunode07_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 13, 'name': 'cpunode07_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 14, 'name': 'cpunode08_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 15, 'name': 'cpunode08_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 16, 'name': 'cpunode09_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 17, 'name': 'cpunode09_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 18, 'name': 'cpunode10_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 19, 'name': 'cpunode10_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 20, 'name': 'cpunode11_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 21, 'name': 'cpunode11_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 22, 'name': 'cpunode12_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 23, 'name': 'cpunode12_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 24, 'name': 'cpunode13_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 25, 'name': 'cpunode13_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 26, 'name': 'cpunode14_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 27, 'name': 'cpunode14_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 28, 'name': 'cpunode15_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 29, 'name': 'cpunode15_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 30, 'name': 'cpunode16_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 31, 'name': 'cpunode16_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 32, 'name': 'cpunode17_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 33, 'name': 'cpunode17_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 34, 'name': 'cpunode18_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 35, 'name': 'cpunode18_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 36, 'name': 'cpunode19_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 37, 'name': 'cpunode19_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 38, 'name': 'cpunode20_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 39, 'name': 'cpunode20_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 40, 'name': 'cpunode21_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 41, 'name': 'cpunode21_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 42, 'name': 'cpunode22_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 43, 'name': 'cpunode22_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 44, 'name': 'cpunode23_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 45, 'name': 'cpunode23_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 46, 'name': 'cpunode24_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 47, 'name': 'cpunode24_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 48, 'name': 'cpunode25_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 49, 'name': 'cpunode25_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 50, 'name': 'cpunode26_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 51, 'name': 'cpunode26_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 52, 'name': 'cpunode27_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 53, 'name': 'cpunode27_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 54, 'name': 'cpunode28_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 55, 'name': 'cpunode28_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 56, 'name': 'cpunode29_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 57, 'name': 'cpunode29_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 58, 'name': 'cpunode30_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 59, 'name': 'cpunode30_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 60, 'name': 'cpunode31_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 61, 'name': 'cpunode31_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 62, 'name': 'cpunode32_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 63, 'name': 'cpunode32_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 64, 'name': 'cpunode33_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 65, 'name': 'cpunode33_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 66, 'name': 'cpunode34_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 67, 'name': 'cpunode34_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 68, 'name': 'cpunode35_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 69, 'name': 'cpunode35_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 70, 'name': 'cpunode36_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 71, 'name': 'cpunode36_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 72, 'name': 'cpunode37_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 73, 'name': 'cpunode37_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 74, 'name': 'cpunode38_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 75, 'name': 'cpunode38_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 76, 'name': 'cpunode39_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 77, 'name': 'cpunode39_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 78, 'name': 'cpunode40_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 79, 'name': 'cpunode40_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 80, 'name': 'cpunode41_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 81, 'name': 'cpunode41_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 82, 'name': 'cpunode42_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 83, 'name': 'cpunode42_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 84, 'name': 'cpunode43_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 85, 'name': 'cpunode43_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 86, 'name': 'cpunode44_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 87, 'name': 'cpunode44_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 88, 'name': 'cpunode45_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 89, 'name': 'cpunode45_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 90, 'name': 'cpunode46_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 91, 'name': 'cpunode46_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 92, 'name': 'cpunode47_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 93, 'name': 'cpunode47_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 94, 'name': 'cpunode48_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 95, 'name': 'cpunode48_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 96, 'name': 'cpunode49_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 97, 'name': 'cpunode49_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 98, 'name': 'cpunode50_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 99, 'name': 'cpunode50_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 100, 'name': 'cbt001_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 101, 'name': 'cbt001_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 102, 'name': 'cbt002_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 103, 'name': 'cbt002_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 104, 'name': 'cbt003_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 105, 'name': 'cbt003_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 106, 'name': 'cbt004_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 107, 'name': 'cbt004_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 108, 'name': 'cbt005_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 109, 'name': 'cbt005_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 110, 'name': 'cbt006_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 111, 'name': 'cbt006_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 112, 'name': 'cbt007_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 113, 'name': 'cbt007_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 114, 'name': 'cbt008_bandwidth', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 115, 'name': 'cbt008_processors', 'type_id': 4, 'type_name': 'processor', 'unit_id': 4, - 'unit': 'cores', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 116, 'name': 'CEP4_bandwidth:/data', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 117, 'name': 'CEP4_storage:/data', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 118, 'name': 'dragproc_bandwidth:/data', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 119, 'name': 'dragproc_storage:/data', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 120, 'name': 'drg01_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 121, 'name': 'drg01_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 122, 'name': 'drg01_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 123, 'name': 'drg01_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 124, 'name': 'drg02_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 125, 'name': 'drg02_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 126, 'name': 'drg02_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 127, 'name': 'drg02_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 128, 'name': 'drg03_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 129, 'name': 'drg03_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 130, 'name': 'drg03_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 131, 'name': 'drg03_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 132, 'name': 'drg04_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 133, 'name': 'drg04_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 134, 'name': 'drg04_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 135, 'name': 'drg04_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 136, 'name': 'drg05_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 137, 'name': 'drg05_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 138, 'name': 'drg05_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 139, 'name': 'drg05_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 140, 'name': 'drg06_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 141, 'name': 'drg06_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 142, 'name': 'drg06_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 143, 'name': 'drg06_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 144, 'name': 'drg07_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 145, 'name': 'drg07_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 146, 'name': 'drg07_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 147, 'name': 'drg07_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 148, 'name': 'drg08_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 149, 'name': 'drg08_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 150, 'name': 'drg08_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 151, 'name': 'drg08_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 152, 'name': 'drg09_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 153, 'name': 'drg09_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 154, 'name': 'drg09_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 155, 'name': 'drg09_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 156, 'name': 'drg10_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 157, 'name': 'drg10_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 158, 'name': 'drg10_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 159, 'name': 'drg10_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 160, 'name': 'drg11_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 161, 'name': 'drg11_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 162, 'name': 'drg11_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 163, 'name': 'drg11_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 164, 'name': 'drg12_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 165, 'name': 'drg12_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 166, 'name': 'drg12_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 167, 'name': 'drg12_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 168, 'name': 'drg13_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 169, 'name': 'drg13_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 170, 'name': 'drg13_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 171, 'name': 'drg13_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 172, 'name': 'drg14_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 173, 'name': 'drg14_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 174, 'name': 'drg14_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 175, 'name': 'drg14_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 176, 'name': 'drg15_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 177, 'name': 'drg15_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 178, 'name': 'drg15_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 179, 'name': 'drg15_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 180, 'name': 'drg16_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 181, 'name': 'drg16_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 182, 'name': 'drg16_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 183, 'name': 'drg16_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 184, 'name': 'drg17_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 185, 'name': 'drg17_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 186, 'name': 'drg17_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 187, 'name': 'drg17_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 188, 'name': 'drg18_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 189, 'name': 'drg18_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 190, 'name': 'drg18_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 191, 'name': 'drg18_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 192, 'name': 'drg19_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 193, 'name': 'drg19_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 194, 'name': 'drg19_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 195, 'name': 'drg19_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 196, 'name': 'drg20_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 197, 'name': 'drg20_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 198, 'name': 'drg20_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 199, 'name': 'drg20_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 200, 'name': 'drg21_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 201, 'name': 'drg21_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 202, 'name': 'drg21_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 203, 'name': 'drg21_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 204, 'name': 'drg22_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 205, 'name': 'drg22_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 206, 'name': 'drg22_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 207, 'name': 'drg22_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 208, 'name': 'drg23_bandwidth:/data1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 209, 'name': 'drg23_bandwidth:/data2', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, - 'unit': 'bits/second', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 210, 'name': 'drg23_storage:/data1', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 211, 'name': 'drg23_storage:/data2', 'type_id': 5, 'type_name': 'storage', 'unit_id': 1, - 'unit': 'bytes', 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 212, 'name': 'CS001rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 213, 'name': 'CS001tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 214, 'name': 'CS002rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 215, 'name': 'CS002tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 216, 'name': 'CS003rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 217, 'name': 'CS003tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 218, 'name': 'CS004rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 219, 'name': 'CS004tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 220, 'name': 'CS005rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 221, 'name': 'CS005tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 222, 'name': 'CS006rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 223, 'name': 'CS006tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 224, 'name': 'CS007rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 225, 'name': 'CS007tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 226, 'name': 'CS011rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 227, 'name': 'CS011tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 228, 'name': 'CS013rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 229, 'name': 'CS013tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 230, 'name': 'CS017rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 231, 'name': 'CS017tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 232, 'name': 'CS021rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 233, 'name': 'CS021tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 234, 'name': 'CS024rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 235, 'name': 'CS024tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 236, 'name': 'CS026rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 237, 'name': 'CS026tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 238, 'name': 'CS028rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 239, 'name': 'CS028tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 240, 'name': 'CS030rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 241, 'name': 'CS030tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 242, 'name': 'CS031rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 243, 'name': 'CS031tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 244, 'name': 'CS032rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 245, 'name': 'CS032tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 246, 'name': 'CS101rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 247, 'name': 'CS101tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 248, 'name': 'CS103rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 249, 'name': 'CS103tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 250, 'name': 'CS201rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 251, 'name': 'CS201tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 252, 'name': 'CS301rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 253, 'name': 'CS301tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 254, 'name': 'CS302rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 255, 'name': 'CS302tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 256, 'name': 'CS401rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 257, 'name': 'CS401tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 258, 'name': 'CS501rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 259, 'name': 'CS501tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 260, 'name': 'RS106rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 261, 'name': 'RS106tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 262, 'name': 'RS205rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 263, 'name': 'RS205tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 264, 'name': 'RS208rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 265, 'name': 'RS208tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 266, 'name': 'RS210rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 267, 'name': 'RS210tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 268, 'name': 'RS305rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 269, 'name': 'RS305tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 270, 'name': 'RS306rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 271, 'name': 'RS306tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 272, 'name': 'RS307rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 273, 'name': 'RS307tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 274, 'name': 'RS310rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 275, 'name': 'RS310tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 276, 'name': 'RS406rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 277, 'name': 'RS406tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 278, 'name': 'RS407rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 279, 'name': 'RS407tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 280, 'name': 'RS408rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 281, 'name': 'RS408tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 282, 'name': 'RS409rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 283, 'name': 'RS409tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 284, 'name': 'RS503rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 285, 'name': 'RS503tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 286, 'name': 'RS508rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 287, 'name': 'RS508tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 288, 'name': 'RS509rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 289, 'name': 'RS509tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 290, 'name': 'DE601rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 291, 'name': 'DE601tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 292, 'name': 'DE602rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 293, 'name': 'DE602tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 294, 'name': 'DE603rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 295, 'name': 'DE603tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 296, 'name': 'DE604rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 297, 'name': 'DE604tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 298, 'name': 'DE605rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 299, 'name': 'DE605tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 300, 'name': 'FR606rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 301, 'name': 'FR606tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 302, 'name': 'SE607rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 303, 'name': 'SE607tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 304, 'name': 'UK608rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 305, 'name': 'UK608tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 306, 'name': 'DE609rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 307, 'name': 'DE609tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 308, 'name': 'PL610rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 309, 'name': 'PL610tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 310, 'name': 'PL611rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 311, 'name': 'PL611tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 312, 'name': 'PL612rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 313, 'name': 'PL612tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 314, 'name': 'IE613rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 315, 'name': 'IE613tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 316, 'name': 'IS614rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 317, 'name': 'IS614tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 318, 'name': 'TEST1rcu', 'type_id': 2, 'type_name': 'rcu', 'unit_id': 2, 'unit': 'rcu_board', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 319, 'name': 'TEST1tbb', 'type_id': 1, 'type_name': 'tbb', 'unit_id': 1, 'unit': 'bytes', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 320, 'name': 'CS001chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 321, 'name': 'CS001bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 322, 'name': 'CS001chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 323, 'name': 'CS001bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 324, 'name': 'CS002chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 325, 'name': 'CS002bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 326, 'name': 'CS002chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 327, 'name': 'CS002bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 328, 'name': 'CS003chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 329, 'name': 'CS003bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 330, 'name': 'CS003chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 331, 'name': 'CS003bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 332, 'name': 'CS004chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 333, 'name': 'CS004bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 334, 'name': 'CS004chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 335, 'name': 'CS004bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 336, 'name': 'CS005chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 337, 'name': 'CS005bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 338, 'name': 'CS005chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 339, 'name': 'CS005bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 340, 'name': 'CS006chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 341, 'name': 'CS006bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 342, 'name': 'CS006chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 343, 'name': 'CS006bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 344, 'name': 'CS007chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 345, 'name': 'CS007bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 346, 'name': 'CS007chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 347, 'name': 'CS007bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 348, 'name': 'CS011chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 349, 'name': 'CS011bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 350, 'name': 'CS011chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 351, 'name': 'CS011bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 352, 'name': 'CS013chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 353, 'name': 'CS013bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 354, 'name': 'CS013chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 355, 'name': 'CS013bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 356, 'name': 'CS017chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 357, 'name': 'CS017bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 358, 'name': 'CS017chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 359, 'name': 'CS017bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 360, 'name': 'CS021chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 361, 'name': 'CS021bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 362, 'name': 'CS021chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 363, 'name': 'CS021bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 364, 'name': 'CS024chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 365, 'name': 'CS024bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 366, 'name': 'CS024chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 367, 'name': 'CS024bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 368, 'name': 'CS026chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 369, 'name': 'CS026bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 370, 'name': 'CS026chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 371, 'name': 'CS026bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 372, 'name': 'CS028chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 373, 'name': 'CS028bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 374, 'name': 'CS028chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 375, 'name': 'CS028bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 376, 'name': 'CS030chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 377, 'name': 'CS030bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 378, 'name': 'CS030chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 379, 'name': 'CS030bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 380, 'name': 'CS031chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 381, 'name': 'CS031bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 382, 'name': 'CS031chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 383, 'name': 'CS031bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 384, 'name': 'CS032chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 385, 'name': 'CS032bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 386, 'name': 'CS032chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 387, 'name': 'CS032bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 388, 'name': 'CS101chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 389, 'name': 'CS101bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 390, 'name': 'CS101chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 391, 'name': 'CS101bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 392, 'name': 'CS103chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 393, 'name': 'CS103bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 394, 'name': 'CS103chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 395, 'name': 'CS103bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 396, 'name': 'CS201chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 397, 'name': 'CS201bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 398, 'name': 'CS201chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 399, 'name': 'CS201bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 400, 'name': 'CS301chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 401, 'name': 'CS301bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 402, 'name': 'CS301chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 403, 'name': 'CS301bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 404, 'name': 'CS302chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 405, 'name': 'CS302bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 406, 'name': 'CS302chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 407, 'name': 'CS302bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 408, 'name': 'CS401chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 409, 'name': 'CS401bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 410, 'name': 'CS401chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 411, 'name': 'CS401bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 412, 'name': 'CS501chan0', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 413, 'name': 'CS501bw0', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 414, 'name': 'CS501chan1', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 415, 'name': 'CS501bw1', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 416, 'name': 'RS106chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 417, 'name': 'RS106bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 418, 'name': 'RS205chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 419, 'name': 'RS205bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 420, 'name': 'RS208chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 421, 'name': 'RS208bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 422, 'name': 'RS210chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 423, 'name': 'RS210bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 424, 'name': 'RS305chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 425, 'name': 'RS305bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 426, 'name': 'RS306chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 427, 'name': 'RS306bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 428, 'name': 'RS307chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 429, 'name': 'RS307bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 430, 'name': 'RS310chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 431, 'name': 'RS310bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 432, 'name': 'RS406chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 433, 'name': 'RS406bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 434, 'name': 'RS407chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 435, 'name': 'RS407bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 436, 'name': 'RS408chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 437, 'name': 'RS408bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 438, 'name': 'RS409chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 439, 'name': 'RS409bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 440, 'name': 'RS503chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 441, 'name': 'RS503bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 442, 'name': 'RS508chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 443, 'name': 'RS508bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 444, 'name': 'RS509chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 445, 'name': 'RS509bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 446, 'name': 'DE601chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 447, 'name': 'DE601bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 448, 'name': 'DE602chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 449, 'name': 'DE602bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 450, 'name': 'DE603chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 451, 'name': 'DE603bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 452, 'name': 'DE604chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 453, 'name': 'DE604bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 454, 'name': 'DE605chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 455, 'name': 'DE605bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 456, 'name': 'FR606chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 457, 'name': 'FR606bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 458, 'name': 'SE607chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 459, 'name': 'SE607bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 460, 'name': 'UK608chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 461, 'name': 'UK608bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 462, 'name': 'DE609chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 463, 'name': 'DE609bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 464, 'name': 'PL610chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 465, 'name': 'PL610bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 466, 'name': 'PL611chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 467, 'name': 'PL611bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 468, 'name': 'PL612chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 469, 'name': 'PL612bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 470, 'name': 'IE613chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 471, 'name': 'IE613bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 472, 'name': 'IS614chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 473, 'name': 'IS614bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 474, 'name': 'TEST1chan', 'type_id': 0, 'type_name': 'rsp', 'unit_id': 0, 'unit': 'rsp_channel_bit', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1}, - {'id': 475, 'name': 'TEST1bw', 'type_id': 3, 'type_name': 'bandwidth', 'unit_id': 3, 'unit': 'bits/second', - 'available_capacity': 10, 'used_capacity': 0, 'total_capacity': 10, 'active': 1} - ] - - self.rarpc_mock.getResourceClaims.return_value = [] - - self.rarpc_mock.getResourceAllocationConfig.return_value = [ - {'name': 'max_fill_ratio_CEP4_storage', 'value': 0.85}, {'name': 'claim_timeout', 'value': 172800}, - {'name': 'min_inter_task_delay', 'value': 60}, {'name': 'max_fill_ratio_CEP4_bandwidth', 'value': 0.75} - ] + self.tmp_exchange = TemporaryExchange(__package__) + self.tmp_exchange.open() + self.addCleanup(self.tmp_exchange.close) def rerpc_mock_get_estimated_resources(specification_tree): otdb_id = specification_tree['otdb_id'] return self.rerpc_replymessage[str(otdb_id)] - rerpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassignmentestimator.rpc.ResourceEstimatorRPC') + rerpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.ResourceEstimatorRPC') self.addCleanup(rerpc_patcher.stop) self.rerpc_mock = rerpc_patcher.start() + self.rerpc_mock.create.side_effect = lambda **kwargs: self.rerpc_mock # make factory method 'create' return the mock instance self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources - otdbrpc_patcher = mock.patch('lofar.sas.otdb.otdbrpc') + otdbrpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.OTDBRPC') self.addCleanup(otdbrpc_patcher.stop) self.otdbrpc_mock = otdbrpc_patcher.start() + self.otdbrpc_mock.create.side_effect = lambda **kwargs: self.otdbrpc_mock # make factory method 'create' return the mock instance - momrpc_patcher = mock.patch('lofar.mom.momqueryservice.momqueryrpc') + momrpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.MoMQueryRPC') self.addCleanup(momrpc_patcher.stop) self.momrpc_mock = momrpc_patcher.start() - self.momrpc_mock.getPredecessorIds.return_value = {str(self.task_mom_id): self.predecessor_task_mom_ids} - self.momrpc_mock.getSuccessorIds.return_value = {str(self.task_mom_id): self.successor_task_mom_ids} + self.momrpc_mock.create.side_effect = lambda **kwargs: self.momrpc_mock # make factory method 'create' return the mock instance + self.momrpc_mock.getPredecessorIds.return_value = {str(self.task_mom_id): [self.predecessor_task_mom_id]} + self.momrpc_mock.getSuccessorIds.return_value = {str(self.task_mom_id): [self.successor_task_mom_id]} self.momrpc_mock.get_time_restrictions.return_value = { "minStartTime": self.task_minstarttime.strftime('%Y-%m-%dT%H:%M:%S'), "maxEndTime": self.task_maxendtime.strftime('%Y-%m-%dT%H:%M:%S'), @@ -1623,96 +411,56 @@ class ResourceAssignerTest(unittest.TestCase): "maxDuration": str(self.task_maxduration) } - curpc_patcher = mock.patch('lofar.sas.datamanagement.cleanup.rpc') + curpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.CleanupRPC') self.addCleanup(curpc_patcher.stop) self.curpc_mock = curpc_patcher.start() + self.curpc_mock.create.side_effect = lambda **kwargs: self.curpc_mock # make factory method 'create' return the mock instance + self.curpc_mock.removeTaskData.return_value = {'deleted': True, 'message': ""} - sqrpc_patcher = mock.patch('lofar.sas.datamanagement.storagequery.rpc') + sqrpc_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.StorageQueryRPC') self.addCleanup(sqrpc_patcher.stop) self.sqrpc_mock = sqrpc_patcher.start() + self.sqrpc_mock.create.side_effect = lambda **kwargs: self.sqrpc_mock # make factory method 'create' return the mock instance + self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10} - ra_notification_bus_patcher = mock.patch('lofar.messaging.messagebus') - self.addCleanup(ra_notification_bus_patcher.stop) - self.ra_notification_bus_mock = ra_notification_bus_patcher.start() - - logger_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.logger') - self.addCleanup(logger_patcher.stop) - self.logger_mock = logger_patcher.start() - - dwell_scheduler_patcher = mock.patch( - 'lofar.sas.resourceassignment.resourceassigner.resource_assigner.DwellScheduler' - ) - self.addCleanup(dwell_scheduler_patcher.stop) - self.dwell_scheduler_mock = dwell_scheduler_patcher.start() - self.dwell_scheduler_mock().allocate_resources.return_value = (True, []) - - prio_scheduler_patcher = mock.patch( - 'lofar.sas.resourceassignment.resourceassigner.resource_assigner.PriorityScheduler' - ) - self.addCleanup(prio_scheduler_patcher.stop) - self.prio_scheduler_mock = prio_scheduler_patcher.start() - self.prio_scheduler_mock().allocate_resources.return_value = (True, []) - - obscontrol_patcher = mock.patch( - 'lofar.sas.resourceassignment.resourceassigner.resource_assigner.ObservationControlRPCClient.abort_observation' - ) + obscontrol_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.ObservationControlRPCClient') self.addCleanup(obscontrol_patcher.stop) self.obscontrol_mock = obscontrol_patcher.start() + self.obscontrol_mock.create.side_effect = lambda **kwargs: self.obscontrol_mock # make factory method 'create' return the mock instance - self.spec_mock = mock.MagicMock() - self.spec_mock.status = 'prescheduled' - self.spec_mock.radb_id = self.task_id - self.spec_mock.mom_id = self.task_mom_id - self.spec_mock.otdb_id = self.task_otdb_id - self.spec_mock.type = self.task_type - self.spec_mock.duration = datetime.timedelta(seconds=60) - self.spec_mock.starttime = datetime.datetime.utcnow() - self.spec_mock.endtime = datetime.datetime.utcnow() - self.spec_mock.min_starttime = datetime.datetime.utcnow() - self.spec_mock.max_endtime = datetime.datetime.utcnow() - self.spec_mock.calculate_dwell_values.return_value = (self.spec_mock.min_starttime, - self.spec_mock.max_endtime, - self.spec_mock.duration) - - specification_patcher = mock.patch( - 'lofar.sas.resourceassignment.resourceassigner.resource_assigner.Specification') - self.addCleanup(specification_patcher.stop) - self.specification_mock = specification_patcher.start() - self.specification_mock.return_value = self.spec_mock + ra_notification_bus_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.ToBus') + self.addCleanup(ra_notification_bus_patcher.stop) + self.ra_notification_bus_mock = ra_notification_bus_patcher.start() # Select logger output to see def myprint(s, *args): print(s % args if args else s, file=sys.stderr) - #self.logger_mock.debug.side_effect = myprint + logger_patcher = mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.logger') + self.addCleanup(logger_patcher.stop) + self.logger_mock = logger_patcher.start() self.logger_mock.info.side_effect = myprint self.logger_mock.warn.side_effect = myprint self.logger_mock.error.side_effect = myprint - self.resource_assigner = TestingResourceAssigner(self.rarpc_mock, self.rerpc_mock, - self.otdbrpc_mock, self.momrpc_mock, - self.curpc_mock, self.sqrpc_mock, - self.ra_notification_bus_mock, - self.dwell_scheduler_mock, - self.obscontrol_mock) + self.resource_assigner = ResourceAssigner(exchange=self.tmp_exchange.address, radb_dbcreds=self.dbcreds) + self.addCleanup(self.resource_assigner.close) self.reset_specification_tree() def assert_all_services_opened(self): - self.assertTrue(self.rarpc_mock.open.called, "RARPC.open was not called") - self.assertTrue(self.rerpc_mock.open.called, "RPC.open was not called") + self.assertTrue(self.rerpc_mock.open.called, "ResourceEstimatorRPC.open was not called") self.assertTrue(self.otdbrpc_mock.open.called, "OTDBRPC.open was not called") self.assertTrue(self.momrpc_mock.open.called, "MOMRPC.open was not called") self.assertTrue(self.curpc_mock.open.called, "CURPC.open was not called") - self.assertTrue(self.ra_notification_bus_mock.open.called, "ra_notification_bus.open was not called") + self.assertTrue(any(mc for mc in self.ra_notification_bus_mock.mock_calls if 'open()' in str(mc)), "ra_notification_bus.open was not called") def assert_all_services_closed(self): - self.assertTrue(self.rarpc_mock.close.called, "RARPC.close was not called") - self.assertTrue(self.rerpc_mock.close.called, "RPC.close was not called") + self.assertTrue(self.rerpc_mock.close.called, "ResourceEstimatorRPC.close was not called") self.assertTrue(self.otdbrpc_mock.close.called, "OTDBRPC.close was not called") self.assertTrue(self.momrpc_mock.close.called, "MOMRPC.close was not called") self.assertTrue(self.curpc_mock.close.called, "CURPC.close was not called") - self.assertTrue(self.ra_notification_bus_mock.close.called, "ra_notification_bus.close was not called") + self.assertTrue(any(mc for mc in self.ra_notification_bus_mock.mock_calls if 'close()' in str(mc)), "ra_notification_bus.close was not called") def test_open_opens_all_services(self): self.resource_assigner.open() @@ -1725,9 +473,7 @@ class ResourceAssignerTest(unittest.TestCase): self.assert_all_services_closed() def test_contextManager_opens_and_closes_all_services(self): - with TestingResourceAssigner(self.rarpc_mock, self.rerpc_mock, self.otdbrpc_mock, self.momrpc_mock, - self.curpc_mock, self.sqrpc_mock, self.ra_notification_bus_mock, - self.dwell_scheduler_mock, self.obscontrol_mock): + with self.resource_assigner: self.assert_all_services_opened() self.assert_all_services_closed() @@ -1735,10 +481,8 @@ class ResourceAssignerTest(unittest.TestCase): def test_do_assignment_logs_specification(self): self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - self.logger_mock.info.assert_any_call('do_assignment: otdb_id=%s specification_tree=%s' % ( - self.specification_tree['otdb_id'], - self.specification_tree - )) + self.logger_mock.info.assert_any_call('do_assignment: otdb_id=%s specification_tree=%s', + self.specification_tree['otdb_id'], self.specification_tree) def test_do_assignment_log_non_approved_or_prescheduled_states(self): otdb_id = self.non_approved_or_prescheduled_otdb_id @@ -1754,22 +498,18 @@ class ResourceAssignerTest(unittest.TestCase): (otdb_id, status, assignable_task_states_str)) def test_do_assignment_approved_task_should_not_be_rescheduled(self): - self.spec_mock.status = 'approved' - - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - - self.logger_mock.info.assert_any_call('Task otdb_id=%s is only approved, no resource assignment needed yet' % - self.specification_tree['otdb_id']) - - def test_do_assignment_inserts_specification_and_task_in_radb(self): - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) + otdb_id = self.specification_tree['otdb_id'] + # assure task is not known yet + self.assertIsNone(self.radb.getTask(otdb_id=otdb_id)) - self.spec_mock.insert_into_radb.assert_called() + self.specification_tree['status'] = 'approved' + self.resource_assigner.do_assignment(otdb_id, self.specification_tree) - def test_do_assignment_fills_specification_based_on_specication_dict(self): - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) + # assure task is known now, and scheduled + self.assertIsNotNone(self.radb.getTask(otdb_id=otdb_id)) + self.assertEqual('approved', self.radb.getTask(otdb_id=otdb_id)['status']) - self.spec_mock.from_dict.assert_called_with(self.specification_tree) + self.logger_mock.info.assert_any_call('Task otdb_id=%s is only approved, no resource assignment needed yet' % otdb_id) def freeze_time_one_day_in_the_future(self, datetime_mock): now = datetime.datetime.utcnow() + datetime.timedelta(days=1) @@ -1787,52 +527,35 @@ class ResourceAssignerTest(unittest.TestCase): self.rerpc_mock.get_estimated_resources.any_calls_with(self.specification_tree) - # def test_do_assignment_logs_when_otdb_id_not_needed_resources(self): - # self.spec_mock.radb_id = self.otdb_id + 11 - # self.specification_tree["otdb_id"] = self.otdb_id + 11 - # - # self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - # - # self.logger_mock.error.assert_any_call( - # "An exception occurred while obtaining resource estimates. Exception=no otdb_id %s found in estimator results %s" % - # (self.otdb_id + 11, self.rerpc_replymessage) - # ) - def test_do_assignment_puts_spec_to_error_when_resource_estimation_gives_an_error(self): - self.dwell_scheduler_mock().allocate_resources.return_value = False - self.specification_tree["otdb_id"] = self.otdb_id + 11 + with mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.PriorityScheduler') as scheduler_mock: + scheduler_mock.allocate_resources.return_value = False - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) + self.specification_tree["otdb_id"] = self.otdb_id + 11 + self.specification_tree['status'] = 'prescheduled' + self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - self.spec_mock.set_status.assert_called_with('error') + task = self.radb.getTask(otdb_id=self.specification_tree["otdb_id"]) + self.assertEqual('error', task['status']) def test_do_assignment_should_not_claim_resouces_when_otdb_id_not_needed_resources(self): - self.specification_tree["otdb_id"] = self.otdb_id + 1 + self.specification_tree["otdb_id"] = self.no_resources_otdb_id + self.specification_tree['status'] = 'prescheduled' self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - self.rarpc_mock.insertResourceClaims.assert_not_called() - - # def test_do_assignment_logs_when_task_type_not_in_needed_resources(self): - # wrong_task_type = "observation" - # self.spec_mock.type = wrong_task_type - # self.specification_tree["task_type"] = wrong_task_type - # - # self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - # - # self.logger_mock.error.assert_any_call( - # "An exception occurred while obtaining resource estimates. Exception=no task type %s found in estimator results %s" % - # (wrong_task_type, self.rerpc_replymessage[str(self.otdb_id)]) - # ) + task = self.radb.getTask(otdb_id=self.specification_tree["otdb_id"]) + self.assertEqual([], self.radb.getResourceClaims(task_ids=task['id'])) def test_do_assignment_should_not_claim_resources_when_task_type_not_in_needed_resources(self): wrong_task_type = "observation" - self.spec_mock.type = wrong_task_type self.specification_tree["task_type"] = wrong_task_type - + self.specification_tree['status'] = 'prescheduled' self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - self.rarpc_mock.insertResourceClaims.assert_not_called() + task = self.radb.getTask(otdb_id=self.specification_tree["otdb_id"]) + self.assertEqual('error', task['status']) + self.assertEqual([], self.radb.getResourceClaims(task_ids=task['id'])) def test_do_assignment_should_log_single_errors_in_needed_resources(self): self.specification_tree["otdb_id"] = self.resources_with_errors_otdb_id @@ -1859,76 +582,122 @@ class ResourceAssignerTest(unittest.TestCase): self.resource_assigner._get_resource_estimates(self.specification_tree) def ra_notification_bus_send_called_with(self, content, subject): - for call in self.ra_notification_bus_mock.send.call_args_list: - if call[0][0].subject == subject and call[0][0].content == content: - return True + for call in self.ra_notification_bus_mock().send.call_args_list: + if isinstance(call[0][0], EventMessage): + msg = call[0][0] + if msg.subject == subject and msg.content == content: + return True return False def test_do_assignment_notifies_bus_when_it_was_unable_to_schedule_Conflict(self): - content = {'radb_id': self.task_id, 'otdb_id': self.task_otdb_id, 'mom_id': self.task_mom_id} - subject = 'Task' + 'Conflict' - - self.spec_mock.status = 'conflict' - self.dwell_scheduler_mock().allocate_resources.return_value = False - - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - - self.assertBusNotificationAndLogging(content, subject) - - def test_do_assignment_should_read_spec_from_radb(self): - self.spec_mock.status = 'conflict' - self.dwell_scheduler_mock().allocate_resources.return_value = False + # prepare: insert a blocking task with a huge claim on storage (directly via the radb, not via the resource_assigner) + task_id = self.radb.insertOrUpdateSpecificationAndTask(9876, 9876, 'prescheduled', 'observation', + datetime.datetime.utcnow()-datetime.timedelta(days=1), + datetime.datetime.utcnow()+datetime.timedelta(days=1), + "", "CEP4")['task_id'] + task = self.radb.getTask(task_id) + cep_storage_resource = next(r for r in self.radb.getResources(resource_types='storage', include_availability=True) if 'CEP4' in r['name']) + self.radb.insertResourceClaim(cep_storage_resource['id'], task_id, task['starttime'], task['endtime'], + 0.75*cep_storage_resource['total_capacity'], "", 0) + self.radb.updateTaskAndResourceClaims(task_id, claim_status='claimed', task_status='scheduled') + self.assertEqual('scheduled', self.radb.getTask(task_id)['status']) + + # make sure the estimater mock asks for too much storage which wont fit during scheduling + def rerpc_mock_get_estimated_resources(specification_tree): + otdb_id = specification_tree['otdb_id'] + estimates = deepcopy(self.rerpc_replymessage[str(otdb_id)]) + estimates['estimates'][0]['resource_types']['storage'] = 0.75*cep_storage_resource['total_capacity'] + return estimates + self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources + # now test the resource_assigner.do_assignment. Should not succeed. Task and claims should go to conflict status. + self.specification_tree['status'] = 'prescheduled' self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - self.spec_mock.read_from_radb.assert_called_with(self.spec_mock.radb_id) - - def test_do_assignment_should_set_status_to_conflict_again_when_cant_schedule_and_in_conflict(self): - self.spec_mock.status = 'conflict' - self.dwell_scheduler_mock().allocate_resources.return_value = False + # check if task is in the radb, and if status is in conflict + resulting_task = self.radb.getTask(otdb_id=self.specification_tree['otdb_id']) + self.assertIsNotNone(resulting_task) + self.assertEqual('conflict', resulting_task['status']) - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - - self.spec_mock.set_status.assert_called_with('conflict') + # check if TaskConflict notification was logged and send + content = {'radb_id': resulting_task['id'], 'otdb_id': resulting_task['otdb_id'], 'mom_id': resulting_task['mom_id']} + subject = 'TaskConflict' + self.assertBusNotificationAndLogging(content, subject) def test_do_assignment_notifies_bus_when_it_was_unable_to_schedule_Error(self): - content = {'radb_id': self.task_id, 'otdb_id': self.task_otdb_id, 'mom_id': self.task_mom_id} - subject = 'Task' + 'Error' - - self.spec_mock.status = 'error' - self.dwell_scheduler_mock().allocate_resources.return_value = False + # make sure the estimater mock asks for more storage than available resulting in TaskError + def rerpc_mock_get_estimated_resources(specification_tree): + otdb_id = specification_tree['otdb_id'] + estimates = deepcopy(self.rerpc_replymessage[str(otdb_id)]) + cep_storage_resource = next(r for r in + self.radb.getResources(resource_types='storage', include_availability=True) + if 'CEP4' in r['name']) + estimates['estimates'][0]['resource_types']['storage'] = 1+cep_storage_resource['total_capacity'] + return estimates + self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources + # now test the resource_assigner.do_assignment. Should not succeed. Task should go to error status. + self.specification_tree['status'] = 'prescheduled' self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) + # check if task is in the radb, and if status is in error + resulting_task = self.radb.getTask(otdb_id=self.specification_tree['otdb_id']) + self.assertIsNotNone(resulting_task) + self.assertEqual('error', resulting_task['status']) + + # check if TaskError notification was logged and send + content = {'radb_id': resulting_task['id'], 'otdb_id': resulting_task['otdb_id'], 'mom_id': resulting_task['mom_id']} + subject = 'TaskError' self.assertBusNotificationAndLogging(content, subject) def test_do_assignment_should_set_status_to_error_again_when_cant_schedule_and_not_in_conflict(self): - self.spec_mock.status = 'error' - self.dwell_scheduler_mock().allocate_resources.return_value = False + # make sure the estimater mock asks for more storage than available resulting in TaskError + def rerpc_mock_get_estimated_resources(specification_tree): + otdb_id = specification_tree['otdb_id'] + estimates = deepcopy(self.rerpc_replymessage[str(otdb_id)]) + cep_storage_resource = next(r for r in + self.radb.getResources(resource_types='storage', include_availability=True) + if 'CEP4' in r['name']) + estimates['estimates'][0]['resource_types']['storage'] = 1+cep_storage_resource['total_capacity'] + return estimates + self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) + # check if the task assignment results in an error twice (apparently it didn't someday for whatever reason) + for i in range(2): + self.specification_tree['status'] = 'prescheduled' + self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - self.spec_mock.set_status.assert_called_with('error') + # check if task is in the radb, and if status is in error + resulting_task = self.radb.getTask(otdb_id=self.specification_tree['otdb_id']) + self.assertIsNotNone(resulting_task) + self.assertEqual('error', resulting_task['status']) - def test_do_assignment_logs_task_data_removal_if_task_is_pipeline(self): - self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10} + # check if TaskError notification was logged and send + content = {'radb_id': resulting_task['id'], 'otdb_id': resulting_task['otdb_id'], 'mom_id': resulting_task['mom_id']} + subject = 'TaskError' + self.assertBusNotificationAndLogging(content, subject) + + # reset the mock calls for next round + self.ra_notification_bus_mock().send.reset_mock() + self.logger_mock.info.send.reset_mock() + def test_do_assignment_logs_task_data_removal_if_task_is_pipeline(self): + self.specification_tree['status'] = 'prescheduled' self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) self.logger_mock.info.assert_any_call("removing data on disk from previous run for otdb_id %s", self.otdb_id) def test_do_assignment_removes_task_data_if_task_is_pipeline(self): - self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10} - + self.specification_tree['status'] = 'prescheduled' self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - self.curpc_mock.removeTaskData.assert_any_call(self.task_otdb_id) + self.curpc_mock.removeTaskData.assert_any_call(self.specification_tree['otdb_id']) def test_do_assignment_logs_when_taks_data_could_not_be_deleted(self): message = "file was locked" - self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10} self.curpc_mock.removeTaskData.return_value = {'deleted': False, 'message': message} + self.specification_tree['status'] = 'prescheduled' self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) self.logger_mock.warning.assert_any_call( @@ -1936,26 +705,42 @@ class ResourceAssignerTest(unittest.TestCase): def test_do_assignment_logs_exception_from_curcp_removeTaskData(self): exception_str = "Error something went wrong" - self.sqrpc_mock.getDiskUsageForOTDBId.return_value = {'found': True, 'disk_usage': 10} self.curpc_mock.removeTaskData.side_effect = Exception(exception_str) - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], - self.specification_tree) + otdb_id = self.specification_tree['otdb_id'] + # assure task is not known yet + self.assertIsNone(self.radb.getTask(otdb_id=otdb_id)) + + self.specification_tree['status'] = 'prescheduled' + self.resource_assigner.do_assignment(otdb_id, self.specification_tree) self.logger_mock.error.assert_any_call("Exception in cleaning up earlier data: %s", exception_str) def test_do_assignment_notifies_bus_when_task_is_scheduled(self): - content = {'radb_id': self.task_id, 'otdb_id': self.task_otdb_id, 'mom_id': self.task_mom_id} - subject = 'Task' + 'Scheduled' - + self.specification_tree['status'] = 'prescheduled' self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) + # check if task is in the radb, and if status is scheduled + resulting_task = self.radb.getTask(otdb_id=self.specification_tree['otdb_id']) + self.assertIsNotNone(resulting_task) + self.assertEqual('scheduled', resulting_task['status']) + + # check if TaskScheduled notification was logged and send + content = {'radb_id': resulting_task['id'], 'otdb_id': resulting_task['otdb_id'], 'mom_id': resulting_task['mom_id']} + subject = 'TaskScheduled' self.assertBusNotificationAndLogging(content, subject) def test_do_assignement_set_status_on_spec_when_scheduleable(self): - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) + otdb_id = self.specification_tree['otdb_id'] + # assure task is not known yet + self.assertIsNone(self.radb.getTask(otdb_id=otdb_id)) + + self.specification_tree['status'] = 'prescheduled' + self.resource_assigner.do_assignment(otdb_id, self.specification_tree) - self.spec_mock.set_status.assert_called_with('scheduled') + # assure task is known now, and scheduled + self.assertIsNotNone(self.radb.getTask(otdb_id=otdb_id)) + self.assertEqual('scheduled', self.radb.getTask(otdb_id=otdb_id)['status']) def assertBusNotificationAndLogging(self, content, subject): self.assertTrue(self.ra_notification_bus_send_called_with(content, "%s.%s" %(DEFAULT_RA_NOTIFICATION_PREFIX, subject))) @@ -1967,6 +752,7 @@ class ResourceAssignerTest(unittest.TestCase): self.otdbrpc_mock.taskSetSpecification.side_effect = Exception(exception_str) with self.assertRaisesRegexp(Exception, exception_str): + self.mom_bug_specification_tree['status'] = 'prescheduled' self.resource_assigner.do_assignment(self.mom_bug_specification_tree['otdb_id'], self.mom_bug_specification_tree) @@ -2019,14 +805,72 @@ class ResourceAssignerTest(unittest.TestCase): self.obscontrol_mock.abort_observation.assert_called_with(spec.otdb_id) # SW-800 The schedulers need open and close called (using context manager) - def test_do_assignement_uses_context_manager_on_dwell_scheduler(self): - self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) + def test_do_assignement_uses_context_manager_on_schedulers(self): + with mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.BasicScheduler') as basic_scheduler_mock: + with mock.patch('lofar.sas.resourceassignment.resourceassigner.resource_assigner.PriorityScheduler') as prio_scheduler_mock: + basic_scheduler_mock().allocate_resources.return_value = (False, None) + prio_scheduler_mock().allocate_resources.return_value = (False, None) + + self.specification_tree['status'] = 'prescheduled' + self.resource_assigner.do_assignment(self.specification_tree['otdb_id'], self.specification_tree) - self.dwell_scheduler_mock().__enter__.assert_called() - self.dwell_scheduler_mock().__exit__.assert_called() + basic_scheduler_mock().__enter__.assert_called() + basic_scheduler_mock().__exit__.assert_called() + + prio_scheduler_mock().__enter__.assert_called() + prio_scheduler_mock().__exit__.assert_called() #This class is currently missing any tests of interaction between tasks already scheduled and new tasks, # e.g. triggered ones. It would require a totally different way to set up the tests to be able to test this. + def test_do_assignment_does_not_raise_on_inserting_predecessors(self): + '''SW-816: When scheduling a successor task, it failed on 'duplicate key value violates unique constraint "task_predecessor_unique"' error in the radb. + This test proves correct/expected behaviour.''' + predecessor_spec = deepcopy(self.specification_tree) + predecessor_otdb_id = predecessor_spec['otdb_id'] + predecessor_spec['status'] = 'prescheduled' + predecessor_spec['predecessors'] = [] + self.resource_assigner.do_assignment(predecessor_otdb_id, predecessor_spec) + + # check if task is in the radb, and if status is scheduled + predecessor_task = self.radb.getTask(otdb_id=predecessor_otdb_id) + self.assertIsNotNone(predecessor_task) + self.assertEqual('scheduled', predecessor_task['status']) + + + successor_spec = deepcopy(self.specification_tree) + successor_spec['otdb_id'] += 1000 + successor_spec['mom_id'] += 1000 + successor_otdb_id = successor_spec['otdb_id'] + successor_spec['status'] = 'prescheduled' + successor_spec['predecessors'] = [predecessor_spec] + + # let the mocked resource estimator return the same estimates for this new otdb_id+1000 + def rerpc_mock_get_estimated_resources(specification_tree): + otdb_id = specification_tree['otdb_id']-1000 + return self.rerpc_replymessage[str(otdb_id)] + self.rerpc_mock.get_estimated_resources.side_effect = rerpc_mock_get_estimated_resources + + # let the momrpc_mock provide the proper linkage between the tasks + self.momrpc_mock.getPredecessorIds.return_value = {str(successor_spec['mom_id']): [predecessor_spec['mom_id']]} + self.momrpc_mock.getSuccessorIds.return_value = {str(predecessor_spec['mom_id']): [successor_spec['mom_id']]} + + # it should be possible to scheduled the successor twice and link it twice to the predecessor. + # the second time, it should just be 'cleaned-up' and rescheduled/relinked. + for i in range(2): + self.resource_assigner.do_assignment(successor_otdb_id, successor_spec) + + # check if task is in the radb, and if status is scheduled + successor_task = self.radb.getTask(otdb_id=successor_otdb_id) + self.assertIsNotNone(successor_task) + self.assertEqual('scheduled', successor_task['status']) + self.assertEqual([predecessor_task['id']], successor_task['predecessor_ids'], ) + + # check if predecessor_task is also linked to its successor + predecessor_task = self.radb.getTask(otdb_id=predecessor_otdb_id) + self.assertEqual([successor_task['id']], predecessor_task['successor_ids'], ) + + + if __name__ == '__main__': unittest.main() diff --git a/SAS/ResourceAssignment/ResourceAssigner/test/t_schedulers.py b/SAS/ResourceAssignment/ResourceAssigner/test/t_schedulers.py index c03af9a1bfe7d7be28c31594405840211b4bc76d..85efbd6aa7c5be61a49cfda0c5c576349fcdb40e 100755 --- a/SAS/ResourceAssignment/ResourceAssigner/test/t_schedulers.py +++ b/SAS/ResourceAssignment/ResourceAssigner/test/t_schedulers.py @@ -37,41 +37,34 @@ from lofar.sas.resourceassignment.database.radb import FETCH_ONE import logging logger = logging.getLogger(__name__) -try: - from . import radb_common_testing -except (ImportError, SystemError): - import radb_common_testing +from lofar.sas.resourceassignment.database.testing.radb_common_testing import RADBCommonTestMixin - -def setUpModule(): - return radb_common_testing.setUpModule() - - -def tearDownModule(): - return radb_common_testing.tearDownModule() - - -class SchedulerTest(radb_common_testing.RADBCommonTest): +class SchedulerTest(RADBCommonTestMixin, unittest.TestCase): """ create test radb postgres instance, and use that in a ResourceAvailabilityChecker""" - def setUp(self): - super(SchedulerTest, self).setUp() - self.resource_availability_checker = ResourceAvailabilityChecker(self.radb) - self._enforce_limited_station_group_list() + @classmethod + def setUpClass(cls): + super().setUpClass() + cls._enforce_limited_station_group_list() - def _enforce_limited_station_group_list(self): + @classmethod + def _enforce_limited_station_group_list(cls): # for test simplicity, create a simple virtual instrument which makes debugging easier. # this is safe, because we are working on a test database LIMITED_STATION_GROUP_LIST = ('CS001', 'CS002', 'RS106', 'RS205') - unwanted_resource_group_ids = [rg['id'] for rg in self.radb.getResourceGroups() + unwanted_resource_group_ids = [rg['id'] for rg in cls.radb.getResourceGroups() if rg['type'] == 'station' and rg['name'] not in LIMITED_STATION_GROUP_LIST] - self.radb.executeQuery("DELETE FROM virtual_instrument.resource_group rg WHERE rg.id in (%s)" % ( - ', '.join([str(id) for id in unwanted_resource_group_ids])),) - self.radb.commit() + if unwanted_resource_group_ids: + cls.radb.executeQuery("DELETE FROM virtual_instrument.resource_group rg WHERE rg.id in (%s)" % ( + ', '.join([str(id) for id in unwanted_resource_group_ids])),) + cls.radb.commit() + def setUp(self): + super().setUp() + self.resource_availability_checker = ResourceAvailabilityChecker(self.radb) class BasicSchedulerTest(SchedulerTest): def new_task(self, mom_otdb_id=0, starttime=None, endtime=None): @@ -89,7 +82,7 @@ class BasicSchedulerTest(SchedulerTest): if endtime is None: endtime = datetime.datetime(2017, 1, 1, 2, 0, 0) - return self.radb.insertSpecificationAndTask(mom_id=mom_otdb_id, + return self.radb.insertOrUpdateSpecificationAndTask(mom_id=mom_otdb_id, otdb_id=mom_otdb_id, task_status='approved', task_type='observation', @@ -111,7 +104,7 @@ class BasicSchedulerTest(SchedulerTest): return BasicScheduler(task_id, specification_tree if specification_tree else self.get_specification_tree(task_id), resource_estimator if resource_estimator else lambda _:[], - self.resource_availability_checker, self.radb.dbcreds) + self.resource_availability_checker, self.radb) def get_station_bandwidth_max_capacity(self): resource_CS001bw0 = [r for r in self.radb.getResources(resource_types="bandwidth", include_availability=True) @@ -251,7 +244,7 @@ class StationSchedulerTest(BasicSchedulerTest): return StationScheduler(task_id, specification_tree if specification_tree else self.get_specification_tree(task_id), resource_estimator if resource_estimator else self.fake_resource_estimator, - self.resource_availability_checker, self.radb.dbcreds) + self.resource_availability_checker, self.radb) def fake_resource_estimator(self, specification_tree): """ Return an estimate for each station, plus a fixed storage claim of half the available storage capacity. """ @@ -496,7 +489,7 @@ class PrioritySchedulerTest(StationSchedulerTest): self.mock_datetime() def new_task_without_momid(self, otdb_id): - return self.radb.insertSpecificationAndTask(mom_id=None, + return self.radb.insertOrUpdateSpecificationAndTask(mom_id=None, otdb_id=otdb_id, task_status='approved', task_type='observation', @@ -519,7 +512,7 @@ class PrioritySchedulerTest(StationSchedulerTest): return PriorityScheduler(task_id, specification_tree if specification_tree else self.get_specification_tree(task_id), resource_estimator if resource_estimator else self.fake_resource_estimator, - self.resource_availability_checker, self.radb.dbcreds) + self.resource_availability_checker, self.radb) def test_unschedule_lower_priority_future_task(self): """ @@ -910,7 +903,7 @@ class DwellSchedulerTest(PrioritySchedulerTest): min_starttime, max_starttime, datetime.timedelta(hours=1), # duration - self.resource_availability_checker, self.radb.dbcreds) + self.resource_availability_checker, self.radb) def test_no_dwell(self): """ Whether a task will not dwell unnecessarily on an empty system. """ diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/CMakeLists.txt b/SAS/ResourceAssignment/ResourceAssignmentDatabase/CMakeLists.txt index 669f6c8fcc4e18824c5f383c33f834499b9639fc..dc5157d4721e30a115e9c35336aed26a68007822 100644 --- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/CMakeLists.txt +++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/CMakeLists.txt @@ -10,9 +10,7 @@ set(_py_files config.py radb.py radbpglistener.py - radbbuslistener.py - tests/radb_common_testing.py -) + radbbuslistener.py) python_install(${_py_files} DESTINATION lofar/sas/resourceassignment/database) diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb.py b/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb.py index 9fa72c99b29f066c3efdbfce4ebf764b9dc43527..43a7c7674b9666c27efe62c574796e425ef6e7ff 100644 --- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb.py +++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb.py @@ -35,9 +35,14 @@ class RADBError(Exception): pass class RADatabase(PostgresDatabaseConnection): - def __init__(self, dbcreds: dbcredentials.DBCredentials, + def __init__(self, dbcreds: dbcredentials.DBCredentials=None, num_connect_retries: int=5, connect_retry_interval: float=1.0): + + if dbcreds is None: + dbcreds = dbcredentials.DBCredentials().get("RADB") + logger.info("Read default RADB dbcreds from disk: %s" % dbcreds.stringWithHiddenPassword()) + super().__init__(dbcreds=dbcreds, auto_commit_selects=False, num_connect_retries=num_connect_retries, @@ -1315,6 +1320,16 @@ class RADatabase(PostgresDatabaseConnection): return self._cursor.rowcount > 0 return True + def deleteResourceClaimForTask(self, task_id, commit=True): + query = '''LOCK TABLE resource_allocation.resource_claim, resource_allocation.resource_usage, resource_allocation.task IN EXCLUSIVE MODE; '''\ + '''DELETE FROM resource_allocation.resource_claim + WHERE resource_allocation.resource_claim.task_id = %s;''' + + self.executeQuery(query, (task_id,)) + if commit: + self.commit() + return self._cursor.rowcount > 0 + def updateResourceClaim(self, resource_claim_id, resource_id=None, task_id=None, starttime=None, endtime=None, status=None, claim_size=None, username=None, used_rcus=None, user_id=None, commit=True): @@ -1552,41 +1567,31 @@ class RADatabase(PostgresDatabaseConnection): else: self.executeQuery('SELECT * from resource_allocation.rebuild_resource_usages_from_claims_for_resource_of_status(%s, %s)', (resource_id, claim_status_id), fetch=FETCH_NONE) - def insertSpecificationAndTask(self, mom_id, otdb_id, task_status, task_type, starttime, endtime, content, cluster, commit=True): + def insertOrUpdateSpecificationAndTask(self, mom_id, otdb_id, task_status, task_type, starttime, endtime, content, cluster, commit=True): ''' Insert a new specification and task in one transaction. - Removes existing task with same otdb_id if present in the same transaction. - Removes existing task with same mom_id if present in the same transaction. + Removes resource_claims for existing task with same otdb_id if present in the same transaction. ''' try: existing_task = self.getTask(otdb_id=otdb_id) - if existing_task: - # delete old specification, task, and resource claims using cascaded delete - self.deleteSpecification(existing_task['specification_id'], False) - else: - if mom_id: - existing_task = self.getTask(mom_id=mom_id) - if existing_task: - # delete old specification, task, and resource claims using cascaded delete - self.deleteSpecification(existing_task['specification_id'], False) - - specId = self.insertSpecification(starttime, endtime, content, cluster, False) - taskId = self.insertTask(mom_id, otdb_id, task_status, task_type, specId, False) - - if specId >= 0 and taskId >= 0: - # restore "old" predecessor/successor relationships if needed - if existing_task: - if existing_task['predecessor_ids']: - self.insertTaskPredecessors(taskId, existing_task['predecessor_ids'], False) - if existing_task['successor_ids']: - for suc_id in existing_task['successor_ids']: - self.insertTaskPredecessor(suc_id, taskId, False) - - if commit: - self.commit() - return {'inserted': True, 'specification_id': specId, 'task_id': taskId} + if existing_task is None and mom_id is not None: + existing_task = self.getTask(mom_id=mom_id) + + if existing_task is not None: + # delete any existing resource_claims and update the properties of the spec and task + logger.info("insertOrUpdateSpecificationAndTask: a task with the same mom/otdb id is already known: existing_task=%s. Clearing current resource claims (if any) and updating specification and task.", existing_task) + specId = existing_task['specification_id'] + taskId = existing_task['id'] + self.deleteResourceClaimForTask(existing_task['id'], False) + self.updateSpecification(specId, starttime=starttime, endtime=endtime, content=content, cluster=cluster, commit=False) + self.updateTask(taskId, mom_id=mom_id, otdb_id=otdb_id, task_status=task_status, task_type=task_type, commit=False) else: - self.rollback() + specId = self.insertSpecification(starttime, endtime, content, cluster, False) + taskId = self.insertTask(mom_id, otdb_id, task_status, task_type, specId, False) + + if commit: + self.commit() + return {'inserted': True, 'specification_id': specId, 'task_id': taskId} except Exception as e: logger.error(e) self.rollback() diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb/sql/add_functions_and_triggers.sql b/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb/sql/add_functions_and_triggers.sql index e89ea38d82e0b932e5403cf4dc27360b2d828b9c..a24822796b807ce3f5963758bd38dd0034a77566 100644 --- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb/sql/add_functions_and_triggers.sql +++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/radb/sql/add_functions_and_triggers.sql @@ -916,5 +916,31 @@ CREATE TRIGGER T_after_claim_insertupdatedelete --------------------------------------------------------------------------------------------------------------------- +CREATE OR REPLACE FUNCTION resource_allocation.after_claim_truncate() + RETURNS trigger AS +$BODY$ +DECLARE +BEGIN + TRUNCATE resource_allocation.resource_usage CASCADE; + TRUNCATE resource_allocation.resource_usage_delta CASCADE; + RETURN NEW; +END; +$BODY$ + LANGUAGE plpgsql VOLATILE + COST 100; +ALTER FUNCTION resource_allocation.after_claim_truncate() + OWNER TO resourceassignment; +COMMENT ON FUNCTION resource_allocation.after_claim_truncate() + IS 'tables resource_allocation.resource_usage and resource_allocation.resource_usage do not have references to the resource_allocation.resource_claim table. So there is no automatic cascading truncate. This function and the truncate trigger on resource_allocation.resource_claim makes sure these tables are truncated as well.'; + +DROP TRIGGER IF EXISTS T_after_claim_truncate ON resource_allocation.resource_claim CASCADE; +CREATE TRIGGER T_after_claim_truncate + AFTER TRUNCATE + ON resource_allocation.resource_claim + FOR EACH STATEMENT + EXECUTE PROCEDURE resource_allocation.after_claim_truncate(); + +--------------------------------------------------------------------------------------------------------------------- + COMMIT; diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/CMakeLists.txt b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/CMakeLists.txt index cd4293dd4bc5d3cee8357e9cbb747ce9bb24058c..ccac9fb675a0aa71459daa90738baa0ccb2b9256 100644 --- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/CMakeLists.txt +++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/CMakeLists.txt @@ -1,12 +1,18 @@ # $Id: CMakeLists.txt 32679 2015-10-26 09:31:56Z schaap $ include(LofarCTest) -include(FindPythonModule) -find_python_module(testing.postgresql) -find_python_module(dateutil) +if(BUILD_TESTING) + include(PythonInstall) + include(FindPythonModule) -lofar_add_test(t_radb_functionality) -lofar_add_test(t_radb_performance) + find_python_module(testing.postgresql) + find_python_module(dateutil) -set_tests_properties(t_radb_functionality PROPERTIES TIMEOUT 300) -set_tests_properties(t_radb_performance PROPERTIES TIMEOUT 300) + python_install(radb_common_testing.py DESTINATION lofar/sas/resourceassignment/database/testing) + + lofar_add_test(t_radb_functionality) + lofar_add_test(t_radb_performance) + + set_tests_properties(t_radb_functionality PROPERTIES TIMEOUT 300) + set_tests_properties(t_radb_performance PROPERTIES TIMEOUT 300) +endif() \ No newline at end of file diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/radb_common_testing.py b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/radb_common_testing.py index 7487aed9b10d20fb1bcf255bcb960e68e3180e0f..e1d89daaf30d188d0da93daa25011c5c9e87802c 100755 --- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/radb_common_testing.py +++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/radb_common_testing.py @@ -29,84 +29,76 @@ logger = logging.getLogger(__name__) import testing.postgresql from lofar.common.dbcredentials import Credentials +from lofar.common.postgres import PostgresDatabaseConnection, FETCH_ALL from lofar.sas.resourceassignment.database.radb import RADatabase -# Create shared test database for better performance -database_credentials = None -Postgresql = None +class RADBCommonTestMixin(): + ''' + A common test mixin class from which you can derive to get a freshly setup postgres testing instance with the latest RADB sql setup scripts applied. + ''' -def setUpModule(): - global database_credentials, Postgresql - database_credentials = Credentials() - Postgresql = testing.postgresql.PostgresqlFactory(cache_initialized_db = True) - -def tearDownModule(): - # clear cached database at end of tests - logger.info('tearDownModule') - Postgresql.clear_cache() - -class RADBCommonTest(unittest.TestCase): - - def setUp(self): - logger.info('setting up test RA database...') + @classmethod + def setUpClass(cls): + logger.info('setting up test database instance...') # connect to shared test db - self.postgresql = Postgresql() # fresh db instead of shared one: self.postgresql = testing.postgresql.Postgresql() - - # set up fixtures - # Note: In theory, this can be moved to the PostgresqlFactory call as kwarg 'on_initialized=populatedb' - # ...but for some reason that was much slower than keeping it here. - self._setup_database() + cls.postgresql = testing.postgresql.PostgresqlFactory(cache_initialized_db=True)() + cls.dbcreds = Credentials() # update credentials (e.g. port changes for each test) - database_credentials.host = self.postgresql.dsn()['host'] - database_credentials.database = self.postgresql.dsn()['database'] - database_credentials.port = self.postgresql.dsn()['port'] - - # connect with useradministration role for tests - self.connection = psycopg2.connect(host = database_credentials.host, - user = database_credentials.user, - password = database_credentials.password, - dbname = database_credentials.database, - port = database_credentials.port) - - # set up radb python module - self.radb = RADatabase(database_credentials) - self.radb.connect() - logger.info('...finished setting up test RA database') - - def tearDown(self): - self.radb.disconnect() - - db_log_file_name = os.path.join(self.postgresql.base_dir, '%s.log' % self.postgresql.name) - logger.info('Printing test-postgress-database server log: %s', db_log_file_name) - with open(db_log_file_name, 'r') as db_log_file: - for line in db_log_file.readlines(): - print(" postgres log: %s" % line.strip(), file=sys.stderr) - - logger.info('removing test RA database...') - self.connection.close() - - # self.Postgresql.clear_cache() # for fresh db during setUp, do instead: - self.postgresql.stop() - - def _setup_database(self): + cls.dbcreds.host = cls.postgresql.dsn()['host'] + cls.dbcreds.database = cls.postgresql.dsn()['database'] + cls.dbcreds.port = cls.postgresql.dsn()['port'] # connect to db as root - conn = psycopg2.connect(**self.postgresql.dsn()) + conn = psycopg2.connect(**cls.postgresql.dsn()) cursor = conn.cursor() # set credentials to be used during tests - database_credentials.user = 'resourceassignment' - database_credentials.password = 'blabla' # cannot be empty... + cls.dbcreds.user = 'resourceassignment' + cls.dbcreds.password = 'secret' # cannot be empty... # create user role # Note: NOSUPERUSER currently raises "permission denied for schema virtual_instrument" # Maybe we want to sort out user creation and proper permissions in the sql scripts? - query = "CREATE USER %s WITH SUPERUSER PASSWORD '%s'" % ( - database_credentials.user, - database_credentials.password) + query = "CREATE USER %s WITH SUPERUSER PASSWORD '%s'" % (cls.dbcreds.user, cls.dbcreds.password) cursor.execute(query) + cursor.close() + conn.commit() + conn.close() + + logger.info('Finished setting up test database instance. It is avaiblable at: %s', cls.dbcreds.stringWithHiddenPassword()) + + cls.radb = RADatabase(cls.dbcreds) + cls.radb.connect() + + # set up a fresh copy of the RADB sql schema + cls._setup_database(cls.radb) + + def setUp(self): + # wipe all tables by truncating specification which cascades into the rest. + logger.debug("setUp: Wiping radb tables for each unittest.") + self.radb.executeQuery("TRUNCATE TABLE resource_allocation.specification CASCADE;") + self.radb.commit() + + @classmethod + def tearDownClass(cls): + cls.radb.disconnect() + + db_log_file_name = os.path.join(cls.postgresql.base_dir, '%s.log' % cls.postgresql.name) + logger.info('Printing test-postgress-database server log: %s', db_log_file_name) + with open(db_log_file_name, 'r') as db_log_file: + for line in db_log_file.readlines(): + print(" postgres log: %s" % line.strip(), file=sys.stderr) + + logger.info('removing test RA database at %s', cls.dbcreds.stringWithHiddenPassword()) + cls.postgresql.stop() + logger.info('test RA removed') + + @staticmethod + def _setup_database(db: PostgresDatabaseConnection): + logger.info('applying RADB sql schema to %s', db) + # populate db tables # These are applied in given order to set up test db # Note: cannot use create_and_populate_database.sql since '\i' is not understood by cursor.execute() @@ -120,50 +112,34 @@ class RADBCommonTest(unittest.TestCase): for sql_path in sql_createdb_paths: logger.debug("setting up database. applying sql file: %s", sql_path) with open(sql_path) as sql: - cursor.execute(sql.read()) - - cursor.close() - conn.commit() - conn.close() - - def _execute_query(self, query, fetch = False): - cursor = self.connection.cursor() - cursor.execute(query) - ret = None - if fetch: - ret = cursor.fetchall() - cursor.close() - self.connection.commit() - return ret - - # --- tests start here - - # integrity tests of postgres database itself - # - # Note: These are meant to make sure the setup generally works and all sql scripts were applied. - # I don't see much benefit in full coverage here since it should be all be tested through RADataBase functionality. - # Of course new tests can be added here where db functionality like triggers should be tested separately from the - # Python part of the job. + db.executeQuery(sql.read()) + db.commit() +class RADBCommonTest(RADBCommonTestMixin, unittest.TestCase): # database created? def test_select_tables_contains_tables_for_each_schema(self): - query = "SELECT table_schema,table_name FROM information_schema.tables" - fetch = self._execute_query(query, fetch = True) - self.assertTrue('resource_allocation' in str(fetch)) - self.assertTrue('resource_monitoring' in str(fetch)) - self.assertTrue('virtual_instrument' in str(fetch)) + with PostgresDatabaseConnection(self.dbcreds) as connection: + query = "SELECT table_schema,table_name FROM information_schema.tables" + result = connection.executeQuery(query, fetch=FETCH_ALL) + self.assertTrue('resource_allocation' in str(result)) + self.assertTrue('resource_monitoring' in str(result)) + self.assertTrue('virtual_instrument' in str(result)) # resource allocation_statics there? def test_select_task_types_contains_obervation(self): - query = "SELECT * FROM resource_allocation.task_type" - fetch = self._execute_query(query, fetch = True) - self.assertTrue('observation' in str(fetch)) + with PostgresDatabaseConnection(self.dbcreds) as connection: + query = "SELECT * FROM resource_allocation.task_type" + result = connection.executeQuery(query, fetch=FETCH_ALL) + self.assertTrue('observation' in str(result)) # virtual instrument there? def test_select_virtualinstrument_units_contain_rcuboard(self): - query = "SELECT * FROM virtual_instrument.unit" - fetch = self._execute_query(query, fetch = True) - self.assertTrue('rcu_board' in str(fetch)) + with PostgresDatabaseConnection(self.dbcreds) as connection: + query = "SELECT * FROM virtual_instrument.unit" + result = connection.executeQuery(query, fetch=FETCH_ALL) + self.assertTrue('rcu_board' in str(result)) + +__all__ = ['RADBCommonTestMixin'] if __name__ == "__main__": os.environ['TZ'] = 'UTC' diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_functionality.py b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_functionality.py index 417b891b08286104ba5f3f80f4513d1ac23db780..251bf9abe62045286cbfca4f7faed632ddd1a5f6 100755 --- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_functionality.py +++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_functionality.py @@ -31,18 +31,13 @@ logger = logging.getLogger(__name__) import unittest.mock as mock from multiprocessing import Process, Event -import radb_common_testing +from lofar.sas.resourceassignment.database.testing.radb_common_testing import RADBCommonTestMixin -from lofar.sas.resourceassignment.database.radb import RADatabase, PostgresDBQueryExecutionError, FETCH_ONE, FETCH_ALL +from lofar.sas.resourceassignment.database.radb import RADatabase +from lofar.common.postgres import PostgresDatabaseConnection, PostgresDBQueryExecutionError, FETCH_ONE, FETCH_ALL from time import sleep -def setUpModule(): - return radb_common_testing.setUpModule() - -def tearDownModule(): - return radb_common_testing.tearDownModule() - -class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): +class ResourceAssignmentDatabaseTest(RADBCommonTestMixin, unittest.TestCase): class test_task: """ A lot of tests involve manipulation of a task (and its corresponding specification) in the RADB. A test task @@ -62,8 +57,10 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): cluster='CEP4'): query = "INSERT INTO resource_allocation.specification (starttime, endtime, content, cluster) " \ "VALUES ('%s', '%s', '%s', '%s') RETURNING id" % (starttime, endtime, content, cluster) - res = self._execute_query(query, fetch=True) - return res[0][0] + with PostgresDatabaseConnection(self.dbcreds) as connection: + res = connection.executeQuery(query, fetch=FETCH_ALL) + connection.commit() + return res[0]['id'] # def test_insert_specification_creates_new_entry(self): # insert spec @@ -72,86 +69,53 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # check it is there query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident - fetch = self._execute_query(query, fetch=True) - self.assertTrue(content in str(fetch)) + with PostgresDatabaseConnection(self.dbcreds) as connection: + res = connection.executeQuery(query, fetch=FETCH_ALL) + self.assertTrue(content in str(res)) def test_update_specification_changes_entry(self): # insert spec ident = self._insert_test_spec() - # update existing spec content - newcontent = 'testcontent_new' - query = "UPDATE resource_allocation.specification SET content = '%s'" % newcontent - self._execute_query(query) + with PostgresDatabaseConnection(self.dbcreds) as connection: + # update existing spec content + newcontent = 'testcontent_new' + query = "UPDATE resource_allocation.specification SET content = '%s'" % newcontent + connection.executeQuery(query) - # check updated content - query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident - fetch = self._execute_query(query, fetch=True) - self.assertTrue(newcontent in str(fetch)) + # check updated content + query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident + res = connection.executeQuery(query, fetch=FETCH_ALL) + self.assertTrue(newcontent in str(res)) def test_delete_specification(self): # insert spec content = 'deletecontent' ident = self._insert_test_spec(content=content) - # make sure it's there - query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident - fetch = self._execute_query(query, fetch=True) - self.assertTrue(content in str(fetch)) + with PostgresDatabaseConnection(self.dbcreds) as connection: + # make sure it's there + query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident + res = connection.executeQuery(query, fetch=FETCH_ALL) + self.assertTrue(content in str(res)) - # delete testspec again - query = "DELETE FROM resource_allocation.specification WHERE id = %s" % ident - self._execute_query(query) + # delete testspec again + query = "DELETE FROM resource_allocation.specification WHERE id = %s" % ident + connection.executeQuery(query) - # make sure it's gone - query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident - fetch = self._execute_query(query, fetch=True) - self.assertFalse(content in str(fetch)) + # make sure it's gone + query = "SELECT content FROM resource_allocation.specification WHERE id=%s" % ident + res = connection.executeQuery(query, fetch=FETCH_ALL) + self.assertFalse(content in str(res)) # triggers in place? def test_insert_specification_swaps_startendtimes_if_needed(self): #when inserting spec with start>endtime, should raise error - with self.assertRaises(psycopg2.InternalError) as context: + with self.assertRaises(PostgresDBQueryExecutionError) as context: # insert spec starttime = '2017-05-10 12:00:00' endtime = '2017-05-10 10:00:00' - ident = self._insert_test_spec(starttime=starttime, endtime=endtime) - - # notifications in place? - def test_insert_task_triggers_notification(self): - # insert specification to not raise INtegrityError - ident = self._insert_test_spec() - - # listen on notification - cursor = self.connection.cursor() - cursor.execute("LISTEN %s;", (psycopg2.extensions.AsIs('task_insert'),)) - - # todo: fix this and use this instead to listen for notifications. - # todo: ...Problem: For some reason callback function is not called. - # set up listener in a way we can check it was called - # callback = mock.Mock() - # callback.listen.return_value = 42 - # self.listener.subscribe('task_insert', callback.listen) - - # trigger notification - query = "INSERT INTO resource_allocation.task (mom_id, otdb_id, status_id, type_id, specification_id)" \ - "VALUES (%s, %s, %s, %s, %s)" % (1, 1, 200, 0, ident) - self._execute_query(query) - - # wait for notification - notification = '' - self.connection.poll() - while self.connection.notifies: - try: - notification = self.connection.notifies.pop(0) - break - except Exception: - pass - - self.assertTrue('task_insert' in str(notification)) - - # todo: fix listener and use instead of polling: - # callback.listen.assert_called() + self._insert_test_spec(starttime=starttime, endtime=endtime) # # radb functionality tests @@ -168,7 +132,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): :returns 2-tuple (task_id, spec_id) or None if task wasn't inserted """ - task = self.radb.insertSpecificationAndTask(mom_id=mom_id, + task = self.radb.insertOrUpdateSpecificationAndTask(mom_id=mom_id, otdb_id=otdb_id, task_status=self.test_task.task_status, task_type=self.test_task.task_type, @@ -461,7 +425,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): task_id = self.radb.insertTask(mom_id, otdb_id, 'conflict', 'observation', spec_id) - self.assertEqual(task_id, 1) + self.assertIsNotNone(task_id) def test_insertTask_duplicate_mom_ids_fails(self): """ Verify if radb.insertTask() raises exception when called with already occupied mom_id """ @@ -537,6 +501,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): task_id = self.radb.insertTask(sample_task['mom_id'], sample_task['otdb_id'], sample_task['status'], sample_task['type'], sample_task['specification_id']) + sample_task['id'] = task_id task = self.radb.getTask(id=task_id) self.assertEqual(task, sample_task) @@ -831,6 +796,25 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): self.assertIs(len(ids), 2) + def test_reinsert_task_with_predecessor(self): + """ Verify if radb.insertTaskPredecessor() returns an ID when called with valid and existing task and + predecessor IDs. """ + + # Shoot 2 unique tasks and corresponding specifications into RADB + task_id_a, _ = self._insert_test_task_and_specification(mom_id=1, otdb_id=10) + task_id_b, _ = self._insert_test_task_and_specification(mom_id=2, otdb_id=11) + + # link b to predecessor a and check it + self.radb.insertTaskPredecessor(task_id_b, task_id_a) + self.assertEqual([], self.radb.getTask(task_id_a)['predecessor_ids']) + self.assertEqual([task_id_a], self.radb.getTask(task_id_b)['predecessor_ids']) + + # reinsert task b + # check if b is still linked to a + task_id_b, _ = self._insert_test_task_and_specification(mom_id=2, otdb_id=11) + self.assertEqual([], self.radb.getTask(task_id_a)['predecessor_ids']) + self.assertEqual([task_id_a], self.radb.getTask(task_id_b)['predecessor_ids']) + def test_getSpecifications_select_all_on_empty_db_succeeds(self): """ Verify if radb.getSpecifications() returns an empty list on an empty RADB """ @@ -885,7 +869,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): now = datetime.utcnow() now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to full hour - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1), + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1), 'foo', 'CEP4') self.assertTrue(result['inserted']) spec_id1 = result['specification_id'] @@ -977,7 +961,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): logger.info('-- now test with a 2nd task, and test resource availability, conflicts etc. --') # another task, fully overlapping with task1 - result = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', now, now+timedelta(hours=1), 'foo', 'CEP4') + result = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', now, now+timedelta(hours=1), 'foo', 'CEP4') self.assertTrue(result['inserted']) spec_id2 = result['specification_id'] task_id2 = result['task_id'] @@ -1102,7 +1086,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): task2 = self.radb.getTask(task_id2) # another task, partially overlapping with both task1 & task3 - result = self.radb.insertSpecificationAndTask(2, 2, 'approved', 'observation', + result = self.radb.insertOrUpdateSpecificationAndTask(2, 2, 'approved', 'observation', task1['starttime'] + (task1['endtime']-task1['starttime'])/2, task2['starttime'] + (task2['endtime']-task2['starttime'])/2, 'foo', 'CEP4') @@ -1187,7 +1171,8 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # suppose the resource_usages table is broken for some reason, fix it.... # break it first... - self._execute_query('TRUNCATE TABLE resource_allocation.resource_usage;') + self.radb.executeQuery('TRUNCATE TABLE resource_allocation.resource_usage;') + self.radb.commit() #check that it's broken self.assertNotEqual(40, self.radb.get_max_resource_usage_between(cep4_id, task1['starttime'], task1['starttime'], 'claimed')['usage']) #fix it @@ -1209,7 +1194,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): future = now + timedelta(hours=2) # insert one task, and reuse that for multiple claims - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'content', 'CEP4') self.assertTrue(result['inserted']) task_id = result['task_id'] @@ -1406,7 +1391,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): future = now + timedelta(hours=2) #insert one task, and reuse that for multiple overlapping claims - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1), + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1), 'foo', 'CEP4') self.assertTrue(result['inserted']) task_id = result['task_id'] @@ -1528,7 +1513,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): future = now + timedelta(hours=2) # insert one task, and reuse that for multiple overlapping claims - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'first content', + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'first content', 'CEP4') self.assertTrue(result['inserted']) task_id = result['task_id'] @@ -1538,11 +1523,11 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): self.assertEqual(task_id, task['id']) self.assertEqual('first content', self.radb.getSpecification(task['specification_id'])['content']) - # prove that we can re-insert the spec/task, and that the new task is indeed inserted and new - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'second content', + # prove that we can re-insert the spec/task, and that the new task is indeed updated + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'second content', 'CEP4') self.assertTrue(result['inserted']) - self.assertNotEqual(task_id, result['task_id']) # we should have a new id because it was re-inserted + self.assertEqual(task_id, result['task_id']) # as per 20190916 inserting a task again should not yield a new task id. task_id = result['task_id'] task = self.radb.getTask(task_id) @@ -1576,10 +1561,10 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # prove again that we can re-insert the spec/task (future with claims), and that the new task is indeed inserted and new, # and that the claim(s) and usage(s) were actually deleted (via cascading deletes) - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'third content', + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'third content', 'CEP4') self.assertTrue(result['inserted']) - self.assertNotEqual(task_id, result['task_id']) # we should have a new id because it was re-inserted + self.assertEqual(task_id, result['task_id']) # as per 20190916 inserting a task again should not yield a new task id. task_id = result['task_id'] task = self.radb.getTask(task_id) @@ -1629,10 +1614,10 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # and prove again that we can re-insert the spec/task (future with claims and a corrupted usage table), and that the new task is indeed inserted and new, # and that the claim(s) and usage(s) were actually deleted (via cascading deletes) # 2017-08-29: YEAH! the insert fails just like on production. Now we can start making a fix! - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'fourth content', + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', future, future + timedelta(hours=1), 'fourth content', 'CEP4') self.assertTrue(result['inserted']) - self.assertNotEqual(task_id, result['task_id']) # we should have a new id because it was re-inserted + self.assertEqual(task_id, result['task_id']) # as per 20190916 inserting a task again should not yield a new task id. task_id = result['task_id'] task = self.radb.getTask(task_id) @@ -1670,7 +1655,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): start = now - timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to current full hour #insert a task - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start, start+timedelta(hours=2), 'foo', 'CEP4') + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start, start+timedelta(hours=2), 'foo', 'CEP4') self.assertTrue(result['inserted']) task_id = result['task_id'] @@ -1795,7 +1780,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): for spec in self.radb.getSpecifications(): self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims - result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start, + result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start, start + timedelta(hours=2), 'foo', 'CEP4') task1_id = result1['task_id'] @@ -1811,7 +1796,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # claim same - result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=5), + result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=5), start + timedelta(hours=2, minutes=5), 'foo', 'CEP4') task2_id = result2['task_id'] @@ -1837,7 +1822,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): for spec in self.radb.getSpecifications(): self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims - result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start, + result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start, start + timedelta(hours=2), 'foo', 'CEP4') task1_id = result1['task_id'] @@ -1853,7 +1838,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # claim same - result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=5), + result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=5), start + timedelta(hours=1, minutes=50), 'foo', 'CEP4') task2_id = result2['task_id'] @@ -1879,7 +1864,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): for spec in self.radb.getSpecifications(): self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims - result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start, + result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start, start + timedelta(hours=2), 'foo', 'CEP4') task1_id = result1['task_id'] @@ -1895,7 +1880,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # claim same - result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=-5), + result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=-5), start + timedelta(hours=1, minutes=55), 'foo', 'CEP4') task2_id = result2['task_id'] @@ -1925,7 +1910,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): for spec in self.radb.getSpecifications(): self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims - result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start, + result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start, start + timedelta(hours=2), 'foo', 'CEP4') task1_id = result1['task_id'] @@ -1941,7 +1926,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # claim same - result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=-5), + result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(minutes=-5), start + timedelta(hours=2, minutes=5), 'foo', 'CEP4') task2_id = result2['task_id'] @@ -1967,7 +1952,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): for spec in self.radb.getSpecifications(): self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims - result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start, + result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start, start + timedelta(hours=2), 'foo', 'CEP4') task1_id = result1['task_id'] @@ -1983,7 +1968,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # claim same - result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start, + result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start, start + timedelta(hours=2), 'foo', 'CEP4') task2_id = result2['task_id'] @@ -2009,7 +1994,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): for spec in self.radb.getSpecifications(): self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims - result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start, + result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start, start + timedelta(hours=2), 'foo', 'CEP4') task1_id = result1['task_id'] @@ -2025,7 +2010,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # claim same - result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(hours=3), + result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start + timedelta(hours=3), start + timedelta(hours=5), 'foo', 'CEP4') task2_id = result2['task_id'] @@ -2049,7 +2034,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): for spec in self.radb.getSpecifications(): self.radb.deleteSpecification(spec['id']) # cascades into tasks and claims - result1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', start + timedelta(hours=3), + result1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', start + timedelta(hours=3), start + timedelta(hours=5), 'foo', 'CEP4') task1_id = result1['task_id'] @@ -2065,7 +2050,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # claim same - result2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', start, + result2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', start, start + timedelta(hours=2), 'foo', 'CEP4') task2_id = result2['task_id'] @@ -2104,7 +2089,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): resource_max_cap = self.radb.get_resource_claimable_capacity(RESOURCE_ID, base_time, base_time) # insert the 'low prio' spec, task... - spec_task_low = self.radb.insertSpecificationAndTask(1, 1, 'prescheduled', 'observation', + spec_task_low = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'prescheduled', 'observation', base_time + timedelta(minutes=5), base_time + timedelta(minutes=10), 'foo', 'CEP4') task_low_id = spec_task_low['task_id'] @@ -2145,7 +2130,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # overlapping with the beginning of task_low # so, the dwellscheduler finds task_low in task_high's higway # so, task_low is aborted by the dwellscheduler (later in the code). - spec_task_high1 = self.radb.insertSpecificationAndTask(2, 2, 'approved', 'observation', + spec_task_high1 = self.radb.insertOrUpdateSpecificationAndTask(2, 2, 'approved', 'observation', base_time, base_time + timedelta(minutes=7), 'foo', 'CEP4') task_high1_id = spec_task_high1['task_id'] @@ -2214,7 +2199,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): now = datetime.utcnow() now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to full hour - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1), 'foo', 'CEP4') self.assertTrue(result['inserted']) self.assertIsNotNone(result['task_id']) @@ -2257,7 +2242,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): now = datetime.utcnow() now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to full hour - result = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', + result = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', now, now+timedelta(hours=1), 'foo', 'CEP4') self.assertTrue(result['inserted']) self.assertIsNotNone(result['task_id']) @@ -2332,7 +2317,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): now = now - timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) now = now + timedelta(hours=1) - spec_task = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', + spec_task = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', now, now + timedelta(minutes=10), 'foo', 'CEP4') @@ -2405,7 +2390,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): now = datetime.utcnow() - spec_task = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', + spec_task = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', now, now, # tasks can have zero duration 'foo', 'CEP4') @@ -2454,7 +2439,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): resource_max_cap = self.radb.get_resource_claimable_capacity(RESOURCE_ID, base_time, base_time) # insert a first task and full claim on a resource... - spec_task1 = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', + spec_task1 = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', base_time + timedelta(minutes=+0), base_time + timedelta(minutes=+10), 'foo', 'CEP4') self.assertTrue(spec_task1['inserted']) @@ -2471,7 +2456,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # insert second (partially overlapping) task and claim on same resource, which we expect to get a conflict status # because the first claim already claims the resource fully. - spec_task2 = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', + spec_task2 = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', base_time + timedelta(minutes=+5), base_time + timedelta(minutes=+15), 'foo', 'CEP4') self.assertTrue(spec_task2['inserted']) @@ -2526,7 +2511,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): microseconds=base_time.microsecond) # insert a first task and full claim on a resource... - spec_task = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'observation', + spec_task = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'observation', base_time + timedelta(minutes=-20), base_time + timedelta(minutes=-10), 'foo', 'CEP4') self.assertTrue(spec_task['inserted']) @@ -2583,7 +2568,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): RESOURCE_ID = 0 resource_max_cap = int(self.radb.get_resource_claimable_capacity(RESOURCE_ID, now, now)) - task1_id = self.radb.insertSpecificationAndTask(1, 1, 'approved', 'observation', + task1_id = self.radb.insertOrUpdateSpecificationAndTask(1, 1, 'approved', 'observation', now+timedelta(hours=1), now + timedelta(hours=2), 'content', 'CEP4')['task_id'] @@ -2629,7 +2614,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): self.radb.getResourceUsages(task1['starttime'], task1['endtime'], RESOURCE_ID)[RESOURCE_ID]['claimed']) # insert second task after the first one (not overlapping) - task2_id = self.radb.insertSpecificationAndTask(2, 2, 'approved', 'observation', + task2_id = self.radb.insertOrUpdateSpecificationAndTask(2, 2, 'approved', 'observation', now + timedelta(hours=3), now + timedelta(hours=4), 'content', 'CEP4')['task_id'] @@ -2680,9 +2665,9 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): now = datetime.utcnow() now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to full hour for i in [1,2]: - task_id = self.radb.insertSpecificationAndTask(i, i, 'approved', 'observation', - now+timedelta(hours=1), now + timedelta(hours=2), - 'content', 'CEP4')['task_id'] + task_id = self.radb.insertOrUpdateSpecificationAndTask(i, i, 'approved', 'observation', + now+timedelta(hours=1), now + timedelta(hours=2), + 'content', 'CEP4')['task_id'] task = self.radb.getTask(task_id) claim = {'resource_id': 0, @@ -2737,10 +2722,6 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): logger.error("exiting updateTask2StartStopTimesLoop because of concurrency issue in updateTask1StatusLoop") return - # updateTask1StatusLoop() - # updateTask2StartStopTimesLoop() - # return - # use multiprocessing.Process instead of threading.Thread for real concurrency p1 = Process(target=updateTask1StatusLoop, daemon=True) p2 = Process(target=updateTask2StartStopTimesLoop, daemon=True) @@ -2766,7 +2747,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): # we need a task.... now = datetime.utcnow() now -= timedelta(minutes=now.minute, seconds=now.second, microseconds=now.microsecond) # round to full hour - task_id = self.radb.insertSpecificationAndTask(0, 0, 'approved', 'reservation', + task_id = self.radb.insertOrUpdateSpecificationAndTask(0, 0, 'approved', 'reservation', now + timedelta(hours=1), now + timedelta(hours=2), 'content', 'CEP4')['task_id'] diff --git a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_performance.py b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_performance.py index 13ade9788c9c00d4558f80c6733fe60c4176e165..91f1a8cec6576e99706168e308a37a8340fef738 100755 --- a/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_performance.py +++ b/SAS/ResourceAssignment/ResourceAssignmentDatabase/tests/t_radb_performance.py @@ -30,16 +30,10 @@ import logging logger = logging.getLogger(__name__) -import radb_common_testing +from lofar.sas.resourceassignment.database.testing.radb_common_testing import RADBCommonTestMixin from lofar.sas.resourceassignment.database.radb import RADatabase, FETCH_ONE -def setUpModule(): - return radb_common_testing.setUpModule() - -def tearDownModule(): - return radb_common_testing.tearDownModule() - -class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): +class ResourceAssignmentDatabaseTest(RADBCommonTestMixin, unittest.TestCase): def test_resource_usages_performance(self): ELAPSED_TRESHOLD = 2.0 #max allowed insert/update/delete time in seconds @@ -77,7 +71,7 @@ class ResourceAssignmentDatabaseTest(radb_common_testing.RADBCommonTest): logger.info('starting task and claim scheduling: counter=%s num_claims_per_resource=%s num_claims_to_insert=%s oversubscription_factor=%s', counter, num_claims_per_resource, num_claims_to_insert, oversubscription_factor) - result = self.radb.insertSpecificationAndTask(counter, counter, 'approved', 'observation', + result = self.radb.insertOrUpdateSpecificationAndTask(counter, counter, 'approved', 'observation', now+timedelta(hours=3*counter), now + timedelta(hours=3*counter + 1), 'content', 'CEP4') diff --git a/SAS/ResourceAssignment/ResourceAssignmentService/rpc.py b/SAS/ResourceAssignment/ResourceAssignmentService/rpc.py index 6307d9cb4f2eb8b2244446a31e56b68f81363bf7..2e2f67b91c069b9d22fd81f3212106e539b6ba09 100644 --- a/SAS/ResourceAssignment/ResourceAssignmentService/rpc.py +++ b/SAS/ResourceAssignment/ResourceAssignmentService/rpc.py @@ -240,8 +240,8 @@ class RADBRPC(RPCClientContextManagerMixin): def getSpecification(self, id): return self._rpc_client.execute('GetSpecification', id=id) - def insertSpecificationAndTask(self, mom_id, otdb_id, task_status, task_type, starttime, endtime, content, cluster): - return self._rpc_client.execute('InsertSpecificationAndTask', + def insertOrUpdateSpecificationAndTask(self, mom_id, otdb_id, task_status, task_type, starttime, endtime, content, cluster): + return self._rpc_client.execute('insertOrUpdateSpecificationAndTask', mom_id=mom_id, otdb_id=otdb_id, task_status=task_status, diff --git a/SAS/ResourceAssignment/ResourceAssignmentService/service.py b/SAS/ResourceAssignment/ResourceAssignmentService/service.py index 888753b4fca36fb8c4d0003b68c330b31fe5b6cc..ddfd4fad090f93ddc6a54e1c870c1a2fb7b4da3c 100644 --- a/SAS/ResourceAssignment/ResourceAssignmentService/service.py +++ b/SAS/ResourceAssignment/ResourceAssignmentService/service.py @@ -60,7 +60,7 @@ class RADBServiceMessageHandler(ServiceMessageHandler): self.register_service_method('GetTaskTypes', self._getTaskTypes) self.register_service_method('GetSpecifications', self._getSpecifications) self.register_service_method('GetSpecification', self._getSpecification) - self.register_service_method('InsertSpecificationAndTask', self._insertSpecificationAndTask) + self.register_service_method('insertOrUpdateSpecificationAndTask', self._insertOrUpdateSpecificationAndTask) self.register_service_method('InsertSpecification', self._insertSpecification) self.register_service_method('DeleteSpecification', self._deleteSpecification) self.register_service_method('UpdateSpecification', self._updateSpecification) @@ -317,9 +317,9 @@ class RADBServiceMessageHandler(ServiceMessageHandler): specification = self.radb.getSpecification(kwargs['id']) return specification - def _insertSpecificationAndTask(self, **kwargs): - logger.info('InsertSpecificationAndTask: %s' % dict({k:v for k,v in list(kwargs.items()) if v != None and k != 'content'})) - return self.radb.insertSpecificationAndTask(kwargs['mom_id'], + def _insertOrUpdateSpecificationAndTask(self, **kwargs): + logger.info('insertOrUpdateSpecificationAndTask: %s' % dict({k:v for k,v in list(kwargs.items()) if v != None and k != 'content'})) + return self.radb.insertOrUpdateSpecificationAndTask(kwargs['mom_id'], kwargs['otdb_id'], kwargs['task_status'], kwargs['task_type'], diff --git a/SAS/ResourceAssignment/TaskPrescheduler/taskprescheduler.py b/SAS/ResourceAssignment/TaskPrescheduler/taskprescheduler.py index a5e6341fccba4509384dc7323f8e09b0c99cee28..760096dd32daec4b34648fd2aa4bd8825686b38e 100644 --- a/SAS/ResourceAssignment/TaskPrescheduler/taskprescheduler.py +++ b/SAS/ResourceAssignment/TaskPrescheduler/taskprescheduler.py @@ -32,7 +32,7 @@ from lofar.sas.otdb.OTDBBusListener import OTDBBusListener, OTDBEventMessageHand from lofar.sas.otdb.otdbrpc import OTDBRPC from lofar.mom.momqueryservice.momqueryrpc import MoMQueryRPC from lofar.sas.resourceassignment.taskprescheduler.cobaltblocksize import CorrelatorSettings, StokesSettings, BlockConstraints, BlockSize -from lofar.sas.resourceassignment.resourceassignmentservice.rpc import RADBRPC +from lofar.sas.resourceassignment.database.radb import RADatabase from lofar.sas.resourceassignment.common.specification import Specification from lofar.sas.resourceassignment.common.specification import OUTPUT_PREFIX @@ -88,18 +88,18 @@ class TaskPrescheduler(OTDBEventMessageHandler): super().__init__() self.otdbrpc = OTDBRPC.create(exchange=exchange, broker=broker) self.momquery = MoMQueryRPC.create(exchange=exchange, broker=broker) - self.radbrpc = RADBRPC.create(exchange=exchange, broker=broker) + self.radb = RADatabase() def start_handling(self): self.otdbrpc.open() self.momquery.open() - self.radbrpc.open() + self.radb.connect() super().start_handling() def stop_handling(self): self.otdbrpc.close() self.momquery.close() - self.radbrpc.close() + self.radb.disconnect() super().stop_handling() def onObservationApproved(self, treeId, modificationTime): @@ -125,7 +125,7 @@ class TaskPrescheduler(OTDBEventMessageHandler): # We get the parset for all tasks we receive instead of just for the ones with # a trigger. status = "approved" - spec = Specification(self.otdbrpc, self.momquery, self.radbrpc) + spec = Specification(self.otdbrpc, self.momquery, self.radb) spec.set_status(status) spec.read_from_OTDB_with_predecessors(treeId, "otdb", {}) #Now checks predecessors, which theoretically could cause race contitions spec.read_from_mom() @@ -165,7 +165,7 @@ class TaskPrescheduler(OTDBEventMessageHandler): except Exception as e: logger.exception(e) logger.error("Problem setting specification or status in OTDB for otdb_id=%s", otdb_id) - self.radbrpc.updateTaskStatusForOtdbId(otdb_id, 'error') # We don't catch an exception if this fails. + self.radb.updateTaskStatusForOtdbId(otdb_id, 'error') # We don't catch an exception if this fails. def main(): from optparse import OptionParser